I have a server of which the main contents are ~10,000 tables with several
thousand records each.
I was able to backup this server using PGAdmin in postgresql 8.4 but now
that I've migrated to 9.0, I get:
E:\Program Files\PostgreSQL\9.0\bin\pg_dumpall.exe --host localhost --port
5433 --username "postgres" --verbose --file
pg_dump: reading constraints
pg_dump: reading triggers
pg_dump: reading large objects
pg_dump: reading dependency data
pg_dump: saving encoding = UTF8
pg_dump: saving standard_conforming_strings = off
pg_dump: saving database definition
pg_dump: WARNING: out of shared memory
pg_dump: SQL command failed
pg_dump: Error message from server: ERROR: out of shared memory
HINT: You might need to increase max_locks_per_transaction.
pg_dump: The command was: SELECT sequence_name, start_value, last_value,
increment_by, CASE WHEN increment_by > 0 AND max_value = 9223372036854775807
THEN NULL WHEN increment_by < 0 AND max_value = -1 THEN NULL ELSE
max_value END AS max_value, CASE WHEN increment_by > 0 AND min_value = 1
THEN NULL WHEN increment_by < 0 AND min_value = -9223372036854775807
THEN NULL ELSE min_value END AS min_value, cache_value, is_cycled,
is_called from arl_record_seq
pg_dump: *** aborted because of error
pg_dumpall: pg_dump failed on database "FinancialData", exiting
Process returned exit code 1.
I have tried increasing max_locks_per_transaction first to 1500, then 10000
and then 50000 with no success.
This seems like it might be a bug since it worked in 8.4.
Thanks for any help;