On Thu, Sep 1, 2011 at 6:38 AM, Rik Bellens <rik.bell...@telin.ugent.be> wrote:
> Op 01-09-11 14:22, Scott Marlowe schreef:

>> Yeah, could be.  Take a look at this page:
>> http://wiki.postgresql.org/wiki/Show_database_bloat and see if the
>> query there sheds some light on your situ.
>
> thanks for this answer
>
> if i run the query, I get 12433752064 wasted bytes on stats_count_pkey, so I
> suppose that is the reason

Also look into installing something like nagios and the
check_postgresql.pl plugin to keep track of these things before they
get out of hand.

csb time:  Back in the day when pg 6.5.3 and 7.0 was new and
interesting, I had a table that was 80k or so, and an index that was
about 100M.  Back when dual core machines were servers, and 1G ram was
an extravagance.   I had a process that deleted everything from the
table each night and replaced it, and the index was so huge that
lookups were taking something like 10 seconds each.  A simple drop /
create index fixed it right up.  The check_postgresql.pl script is a
god sent tool to keep your db healthy and happy.

-- 
Sent via pgsql-general mailing list (pgsql-general@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general

Reply via email to