"Kevin Grittner" <[email protected]> writes:
> Tom Lane <[email protected]> wrote:
>> Do you have the opportunity to try an experiment on hardware
>> similar to what you're running that on? Create a database with 7
>> million tables and see what the dump/restore times are like, and
>> whether pg_dump/pg_restore appear to be CPU-bound or
>> memory-limited when doing it.
> If these can be empty (or nearly empty) tables, I can probably swing
> it as a background task. You didn't need to match the current 1.3
> TB database size I assume?
Empty is fine.
>> If they aren't, we could conclude that millions of TOC entries
>> isn't a problem.
> I'd actually be rather more concerned about the effects on normal
> query plan times, or are you confident that won't be an issue?
This is only a question of what happens internally in pg_dump and
pg_restore --- I'm not suggesting we change anything on the database
side.
regards, tom lane
--
Sent via pgsql-hackers mailing list ([email protected])
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-hackers