Hi All,

I am writing a program that needs time-series-based insert mostly workload.
I need to make the system scaleable with many thousand of inserts/s. One of
the techniques I plan to use is time-based table partitioning and I am
trying to figure out how large to make my time tables.

Does anybody have any hints on optimal table sizes either in terms of rows
or in terms of size? Any rule of thumbs I can use for table size in
relation to amount of memory on the machine? Is the size of the index more
important than the size of the table (if queries mostly use indexes)?

Basically, I am asking for pointers about how to think about this problem
and any experiences people have had.

Thanks,
Mat

P.S. I am aware of limits listed here: http://www.postgresql.org/about/. I
am asking about practical size limits for performance consideration.

Reply via email to