I truly hoping I'm missing something silly here. I've got a cron job to run a dumpall each early am. It fails, and I get a handful of emails. The first reads like this:

pg_dump: [archiver (db)] connection to database "adocs" failed: FATAL:  sorry, 
too many clients already
pg_dumpall: pg_dump failed on database "adocs", exiting


...and then as we go along we get this one repeating for each database:

pg_dump: WARNING:  out of shared memory
pg_dump: SQL command failed
pg_dump: Error message from server: ERROR:  out of shared memory
HINT:  You may need to increase max_locks_per_transaction.
pg_dump: The command was: SELECT sequence_name, last_value, increment_by, CASE WHEN 
increment_by > 0 AND max_value = 9223372036854775807 THEN NULL      WHEN increment_by 
< 0 AND max_value = -1 THEN NULL      ELSE max_value END AS max_value, CASE WHEN 
increment_by > 0 AND min_value = 1 THEN NULL      WHEN increment_by < 0 AND min_value 
= -9223372036854775807 THEN NULL      ELSE min_value END AS min_value, cache_value, 
is_cycled, is_called from tabproj_skey
pg_dumpall: pg_dump failed on database "XXXXX", exiting




The cron entry (for user root) is

* 1 * * * /root/dumpall.sh > /dev/null

and the routine in question is this:

pg_dumpall -U postgres  > /home/bups/bsource/pg/dhost2.dumpall
chown bups:root /home/bups/bsource/pg/dhost2.dumpall
chmod 600       /home/bups/bsource/pg/dhost2.dumpall




---------------------------(end of broadcast)---------------------------
TIP 4: Have you searched our list archives?

              http://archives.postgresql.org/

Reply via email to