One other option is to shut the database down competely, and then do a copy of the file system the new server. I have done this when I need to move a very large database to a new server. I can copy 500GB's in a couple of hours, where restoring my large databases backups would take 10+ hours. Just make sure you are keeping postgres at the same version level.
HTH, Chris On 12/19/06, Arnau <[EMAIL PROTECTED]> wrote:
Hi all, I've got a DB in production that is bigger than 2GB that dumping it takes more than 12 hours. I have a new server to replace this old one where I have restore the DB's dump. The problem is I can't afford to have the server out of business for so long, so I need your advice about how you'd do this dump/restore. The big amount of data is placed in two tables (statistics data), so I was thinking in dump/restore all except this two tables and once the server is running again I'd dump/restore this data. The problem is I don't know how exactly do this. Any suggestion? Thanks -- Arnau ---------------------------(end of broadcast)--------------------------- TIP 9: In versions below 8.0, the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match