Not that my DB is that big..  but if it was and it contained any sort of
financial data (something that you might want to dispute 2 years down
the road) then I would have multiple replicated systems (which I do have
.. but they are MSSQL) and I would also be backing the data up to an
offsite storage.. either via tape or another box with enough storage
space.  Your best bet is to have geographical redundancy.

Travis

-----Original Message-----
From: Lincoln Yeoh [mailto:[EMAIL PROTECTED]
Sent: Sunday, September 14, 2003 10:20 AM
To: Lamar Owen
Cc: PgSQL General ML
Subject: Re: need for in-place upgrades (was Re: [GENERAL] State of



>At 07:16 PM 9/13/2003 -0400, Lamar Owen wrote:
>'migration' server.  And I really don't want to think about
dump/restore 
>of 100TB (if PostgreSQL actually stores the image files, which it
might).

Hmm. Just curious, do people generally backup 100TB of data, or once
most 
reach this point they have to hope that it's just hardware failures
they'll 
deal with and not software/other issues?

100TB sounds like a lot of backup media and time... Not to mention
ensuring 
that the backups will work with available and functioning backup
hardware.

Head hurts just to think about it,

Link.

---------------------------(end of broadcast)---------------------------
TIP 1: subscribe and unsubscribe commands go to [EMAIL PROTECTED]

---------------------------(end of broadcast)---------------------------
TIP 9: the planner will ignore your desire to choose an index scan if your
      joining column's datatypes do not match

Reply via email to