Re: [GENERAL] pg_dump's over 2GB

2000-09-29 Thread Ross J. Reedstrom

On Fri, Sep 29, 2000 at 11:41:51AM -0500, Jeff Hoffmann wrote:
> Bryan White wrote:
> > 
> > I am thinking that
> > instead I will need to pipe pg_dumps output into gzip thus avoiding the
> > creation of a file of that size.
> 
> sure, i do it all the time.  unfortunately, i've had it happen a few
> times where even gzipping a database dump goes over 2GB, which is a real
> PITA since i have to dump some tables individually.  generally, i do


> something like 
>   pg_dump database | gzip > database.pgz 

Hmm, how about:

pg_dump database | gzip | split -b 1024m - database_

Which will give you 1GB files, named database_aa, database_ab, etc.

> to dump the database and 
>   gzip -dc database.pgz | psql database

cat database_* | gunzip | psql database

Ross Reedstrom
-- 
Open source code is like a natural resource, it's the result of providing
food and sunshine to programmers, and then staying out of their way.
[...] [It] is not going away because it has utility for both the developers 
and users independent of economic motivations.  Jim Flynn, Sunnyvale, Calif.



Re: [GENERAL] pg_dump's over 2GB

2000-09-29 Thread Jeff Hoffmann

Bryan White wrote:
> 
> I am thinking that
> instead I will need to pipe pg_dumps output into gzip thus avoiding the
> creation of a file of that size.
> 

sure, i do it all the time.  unfortunately, i've had it happen a few
times where even gzipping a database dump goes over 2GB, which is a real
PITA since i have to dump some tables individually.  generally, i do
something like 
pg_dump database | gzip > database.pgz 
to dump the database and 
gzip -dc database.pgz | psql database
to restore it.  i've always thought that compress should be an option
for pg_dump, but it's really not that much more work to just pipe the
input and output through gzip.

-- 

Jeff Hoffmann
PropertyKey.com



Re: [GENERAL] pg_dump's over 2GB

2000-09-29 Thread Steve Wolfe

> My current backups made with pg_dump are currently 1.3GB.  I am wondering
> what kind of headaches I will have to deal with once they exceed 2GB.
>
> What will happen with pg_dump on a Linux 2.2.14 i386 kernel when the
output
> exceeds 2GB?

  There are some ways around it if your program supports it, I'm not sure if
it works with redirects...

> Currently the dump file is later fed to a 'tar cvfz'.  I am thinking that
> instead I will need to pipe pg_dumps output into gzip thus avoiding the
> creation of a file of that size.

   Why not just pump the data right into gzip?  Something like:

pg_dumpall | gzip --stdout > pgdump.gz

  (I'm sure that the more efficient shell scripters will know a better way)

  If your data is anything like ours, you will get at least a 5:1
compression ratio, meaning you can actually dump around 10 gigs of data
before you hit the 2 gig file limit.

steve




Re: [GENERAL] pg_dump's over 2GB

2000-09-29 Thread Adam Haberlach

On Fri, Sep 29, 2000 at 12:15:26PM -0400, Bryan White wrote:
> My current backups made with pg_dump are currently 1.3GB.  I am wondering
> what kind of headaches I will have to deal with once they exceed 2GB.
> 
> What will happen with pg_dump on a Linux 2.2.14 i386 kernel when the output
> exceeds 2GB?
> Currently the dump file is later fed to a 'tar cvfz'.  I am thinking that
> instead I will need to pipe pg_dumps output into gzip thus avoiding the
> creation of a file of that size.
> 
> Does anyone have experince with this sort of thing?

We have had some problems with tar silently truncated some > 2Gb files
during a backup.  We also had to move the perforce server from Linux to
BSD because some checkpoint files were truncating at 2Gb (not a perforce
problem, but a Linux one).

Be careful, test frequently, etc...

-- 
Adam Haberlach| A billion hours ago, human life appeared on
[EMAIL PROTECTED]   | earth.  A billion minutes ago, Christianity
http://www.newsnipple.com | emerged.  A billion Coca-Colas ago was
'88 EX500 | yesterday morning. -1996 Coca-Cola Ann. Rpt.