Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-25 Thread Greg
uep,
This solution seems like the best and most efficient way of handling large 
filesystems. My biggest question however is, when backing this up to tape, can 
it be split across several tapes? I will be using bacula to back this up. Will 
i need to tar or star this filesystem before writing it to tape? The next 
question I have is since I am using a primary and a secondary server, using zfs 
send/recv how would I incorporate this solution to then backing the secondary 
SAN to tape? Will I need double the hard disk space as I will need a file based 
copy of the ZFS filesystem? I know it is a lot of questions but I thought the 
solution would work perfect in my environment.

Thanks,
Greg
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-25 Thread Robert Milkowski

On 21/01/2010 11:55, Julian Regel wrote:


>> Until you try to pick one up and put it in a fire safe!

>Then you backup to tape from x4540 whatever data you need.
>In case of enterprise products you save on licensing here as you need 
a one client license per x4540 but in fact can >backup data from many 
clients which are there.


Which brings up full circle...

What do you then use to backup to tape bearing in mind that the 
Sun-provided tools all have significant limitations?


I guess you need to use a third party tool and watch carefully that 
they provide complete backups.




in this specific environment I was talking about NetBackup is used.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-22 Thread Ian Collins

A Darren Dunham wrote:

On Wed, Jan 20, 2010 at 08:11:27AM +1300, Ian Collins wrote:
  

True, but I wonder how viable its future is.  One of my clients
requires 17 LT04 types for a full backup, which cost more and takes
up more space than the equivalent in removable hard drives.



What kind of removable hard drives are you getting that are cheaper than
tape?

  
It's not the raw cost per GB, it's the way the tapes are used.  To aid 
recovery times, a number of different backup sets (groups of 
filesystems) are written, so the tapes aren't all used to capacity.


--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-22 Thread A Darren Dunham
On Thu, Jan 21, 2010 at 12:38:56AM +0100, Ragnar Sundblad wrote:
> On 21 jan 2010, at 00.20, Al Hopper wrote:
> > I remember for about 5 years ago (before LT0-4 days) that streaming
> > tape drives would go to great lengths to ensure that the drive kept
> > streaming - because it took so much time to stop, backup and stream
> > again.  And one way the drive firmware accomplished that was to write
> > blocks of zeros when there was no data available.

> I haven't seen drives that fill out with zeros. Sounds like an ugly
> solution, but maybe it could be useful in some strange case.

It was closer to 15 years ago than 5, but this may be a reference to the
first release of the DLT 7000.  That version came out with only 4MB as a
RAM buffer, which is insufficient to buffer at speed during a stop/start
cycle.  It didn't write zeros, but it would disable the on-drive
compression to try to keep the speed of bits being written to tape up.
So the effect was similar in that the capacity of the media was reduced.
The later versions had 8MB buffers and that behavior was removed.

-- 
Darren
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-22 Thread A Darren Dunham
On Wed, Jan 20, 2010 at 08:11:27AM +1300, Ian Collins wrote:
> True, but I wonder how viable its future is.  One of my clients
> requires 17 LT04 types for a full backup, which cost more and takes
> up more space than the equivalent in removable hard drives.

What kind of removable hard drives are you getting that are cheaper than
tape?

LTO4 media is less than 2.5 cents/GB for us (before compression,
acquisition cost only).

-- 
Darren
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-22 Thread Mike Gerdts
On Thu, Jan 21, 2010 at 11:28 AM, Richard Elling
 wrote:
> On Jan 21, 2010, at 3:55 AM, Julian Regel wrote:
>> >> Until you try to pick one up and put it in a fire safe!
>>
>> >Then you backup to tape from x4540 whatever data you need.
>> >In case of enterprise products you save on licensing here as you need a one 
>> >client license per x4540 but in fact can >backup data from many clients 
>> >which are there.
>>
>> Which brings up full circle...
>>
>> What do you then use to backup to tape bearing in mind that the Sun-provided 
>> tools all have significant limitations?
>
> Poor choice of words.  Sun resells NetBackup and (IIRC) that which was
> formerly called NetWorker.  Thus, Sun does provide enterprise backup
> solutions.

(Symantec nee Veritas) NetBackup and (EMC nee Legato) Networker are
different products that compete in the enterprise backup space.

Under the covers NetBackup uses gnu tar to gather file data for the
backup stream.  At one point (maybe still the case), one of the
claimed features of netbackup is that if a tape is written without
multiplexing, you can use gnu tar to extract data.  This seems to be
most useful when you need to recover master and/or media servers and
to be able to extract your data after you no longer use netbackup.

-- 
Mike Gerdts
http://mgerdts.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-22 Thread Joerg Schilling
Ian Collins  wrote:

> In addition to Richard's comments, I doubt many medium to large 
> businesses would use ufsdump/restore as their backup solution.

ufsdump/restore is a reliable backup sulution as long as you install
a mirrored root fs (in case you are using UFS as root fs).
In fact many sites now use a basic backup bethod that is much less
reliable than ufsdump/restore. The fact that they use a high level front 
end does not necessarily remove limitations from the software that is used
as backend. I would never trust a backup sulution that uses e.g. GNU tar
as backend.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-21 Thread Ian Collins

Julian Regel wrote:


>> Until you try to pick one up and put it in a fire safe!

>Then you backup to tape from x4540 whatever data you need.
>In case of enterprise products you save on licensing here as you need 
a one client license per x4540 but in fact can >backup data from many 
clients which are there.


Which brings up full circle...

What do you then use to backup to tape bearing in mind that the 
Sun-provided tools all have significant limitations?


In addition to Richard's comments, I doubt many medium to large 
businesses would use ufsdump/restore as their backup solution.


--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-21 Thread Richard Elling
On Jan 21, 2010, at 3:55 AM, Julian Regel wrote:
> >> Until you try to pick one up and put it in a fire safe!
> 
> >Then you backup to tape from x4540 whatever data you need.
> >In case of enterprise products you save on licensing here as you need a one 
> >client license per x4540 but in fact can >backup data from many clients 
> >which are there.
> 
> Which brings up full circle...
> 
> What do you then use to backup to tape bearing in mind that the Sun-provided 
> tools all have significant limitations?

Poor choice of words.  Sun resells NetBackup and (IIRC) that which was 
formerly called NetWorker.  Thus, Sun does provide enterprise backup
solutions.

If I may put on my MBA hat, the competition is not ufsdump.  ufsdump has
nearly zero market penetration and no prospects for improving its market
share. Making another ufsdump will also gain no market share.  The market
leaders are the likes of EMC, IBM, and Symantec with their heterogenous
backup support. If Sun wanted to provide a better solution that might gain
market share against the others, then it would also need to be heterogenous.
So I think it would be hard to make a business case for a whole new backup
solution. A less costly and less risky approach is to work with the market 
leaders to better integrate with dataset replication.  Caveat: this may already
be available, I haven't looked recently.

> I guess you need to use a third party tool and watch carefully that they 
> provide complete backups.

This is a good idea anyway.
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-21 Thread Julian Regel


>> Until you try to pick one up and put it in a fire safe!

>Then you backup to tape from x4540 whatever data you need.
>In case of enterprise products you save on licensing here as you need a one 
>client license per x4540 but in fact can >backup data from many clients which 
>are there.

Which brings up full circle...

What do you then use to backup to tape bearing in mind that the Sun-provided 
tools all have significant limitations?

I guess you need to use a third party tool and watch carefully that they 
provide complete backups.

JR
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



  ___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-21 Thread Robert Milkowski

On 21/01/2010 09:07, Ian Collins wrote:

Robert Milkowski wrote:

On 20/01/2010 19:20, Ian Collins wrote:

Julian Regel wrote:

>It is actually not that easy.
>
>Compare a cost of 2x x4540 with 1TB disks to equivalent solution 
on LTO.

>
>Each x4540 could be configured as: 4x 11 disks in raidz-2 + 2x hot 
spare

>+ 2x OS disks.
>The four raidz2 group form a single pool. This would provide well 
over

>30TB of logical storage per each box.
>
>Now you rsync all the data from your clients to a dedicated 
filesystem

>per client, then create a snapshot.
>All snapshots are replicated to a 2nd x4540 so even if you would 
loose

>entire box/data for some reason you would still have a spare copy.
>
>Now compare it to a cost of a library, lto drives, tapes, software +
>licenses, support costs, ...
>
>See more details at
>http://milek.blogspot.com/2009/12/my-presentation-at-losug.html

I've just read your presentation Robert. Interesting stuff.

I've also just done a pen and paper exercise to see how much 30TB 
of tape would cost as a comparison to your disk based solution.


Using list prices from Sun's website (and who pays list..?), an 
SL48 with 2 x LTO3 drives would cost £14000. I couldn't see a price 
on an LTO4 equipped SL48 despite the Sun website saying it's a 
supported option. Each LTO3 has a native capacity of 300GB and the 
SL48 can hold up to 48 tapes in the library (14.4TB native per 
library). To match the 30TB in your solution, we'd need two 
libraries totalling £28000.


You would also need 100 LTO3 tapes to provide 30TB of native 
storage. I recently bought a pack of 20 tapes for £340, so five 
packs would be £1700.


So you could provision a tape backup for just under £3 
(~$49000). In comparison, the cost of one X4540 with ~ 36TB usable 
storage is UK list price £30900. I've not factored in backup 
software since you could use an open source solution such as Amanda 
or Bacula.


A more apples to apples comparison would be to compare the storage 
only.  Both removable drive and tape options require a server with 
FC or SCSI ports, so that can be excluded from the comparison.




I think one should actually compare whole solutions - including 
servers, fc infrastructure, tape drives, robots, software costs, rack 
space, ...


Servers like x4540 are ideal for zfs+rsync backup solution - very 
compact, good $/GB ratio, enough CPU power for its capacity, allow to 
easily scale it horizontally, and it is not too small and not too 
big. Then thanks to its compactness they are very easy to administer.


Until you try to pick one up and put it in a fire safe!


Then you backup to tape from x4540 whatever data you need.
In case of enterprise products you save on licensing here as you need a 
one client license per x4540 but in fact can backup data from many 
clients which are there.


:)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-21 Thread Andrew Gabriel

Robert Milkowski wrote:
I think one should actually compare whole solutions - including servers, 
fc infrastructure, tape drives, robots, software costs, rack space, ...


Servers like x4540 are ideal for zfs+rsync backup solution - very 
compact, good $/GB ratio, enough CPU power for its capacity, allow to 
easily scale it horizontally, and it is not too small and not too big. 
Then thanks to its compactness they are very easy to administer.


Depending on an anvironment one could deploy them always in paris - one 
in one datacenter and 2nd one in other datacanter with ZFS send based 
replication of all backups (snapshots). Or one may replicate 
(cross-replicate) only selected clients if needed.


Something else that often sells the 4500/4540 relates to internal 
company politics. Often, inside a company storage has to be provisioned 
from the company's storage group, using very expensive SAN based 
storage, indeed so expensive by the time the company's storage group 
have added their overhead onto the already expensive SAN, that whole 
projects become unviable. Instead, teams find they can order 4500/4540's 
which slip under the radar as servers (or even PCs), and they now have 
affordable storage for their projects, which makes them viable once more.


--
Andrew
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-21 Thread Ian Collins

Robert Milkowski wrote:

On 20/01/2010 19:20, Ian Collins wrote:

Julian Regel wrote:

>It is actually not that easy.
>
>Compare a cost of 2x x4540 with 1TB disks to equivalent solution on 
LTO.

>
>Each x4540 could be configured as: 4x 11 disks in raidz-2 + 2x hot 
spare

>+ 2x OS disks.
>The four raidz2 group form a single pool. This would provide well over
>30TB of logical storage per each box.
>
>Now you rsync all the data from your clients to a dedicated filesystem
>per client, then create a snapshot.
>All snapshots are replicated to a 2nd x4540 so even if you would loose
>entire box/data for some reason you would still have a spare copy.
>
>Now compare it to a cost of a library, lto drives, tapes, software +
>licenses, support costs, ...
>
>See more details at
>http://milek.blogspot.com/2009/12/my-presentation-at-losug.html

I've just read your presentation Robert. Interesting stuff.

I've also just done a pen and paper exercise to see how much 30TB of 
tape would cost as a comparison to your disk based solution.


Using list prices from Sun's website (and who pays list..?), an SL48 
with 2 x LTO3 drives would cost £14000. I couldn't see a price on an 
LTO4 equipped SL48 despite the Sun website saying it's a supported 
option. Each LTO3 has a native capacity of 300GB and the SL48 can 
hold up to 48 tapes in the library (14.4TB native per library). To 
match the 30TB in your solution, we'd need two libraries totalling 
£28000.


You would also need 100 LTO3 tapes to provide 30TB of native 
storage. I recently bought a pack of 20 tapes for £340, so five 
packs would be £1700.


So you could provision a tape backup for just under £3 
(~$49000). In comparison, the cost of one X4540 with ~ 36TB usable 
storage is UK list price £30900. I've not factored in backup 
software since you could use an open source solution such as Amanda 
or Bacula.


A more apples to apples comparison would be to compare the storage 
only.  Both removable drive and tape options require a server with FC 
or SCSI ports, so that can be excluded from the comparison.




I think one should actually compare whole solutions - including 
servers, fc infrastructure, tape drives, robots, software costs, rack 
space, ...


Servers like x4540 are ideal for zfs+rsync backup solution - very 
compact, good $/GB ratio, enough CPU power for its capacity, allow to 
easily scale it horizontally, and it is not too small and not too big. 
Then thanks to its compactness they are very easy to administer.


Until you try to pick one up and put it in a fire safe!

Depending on an anvironment one could deploy them always in paris - 
one in one datacenter and 2nd one in other datacanter with ZFS send 
based replication of all backups (snapshots). Or one may replicate 
(cross-replicate) only selected clients if needed.


Yes, I agree.  That's how my client's systems are configured (pairs).  
We also have another with an attached tape library.


--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-21 Thread Robert Milkowski

On 20/01/2010 19:20, Ian Collins wrote:

Julian Regel wrote:

>It is actually not that easy.
>
>Compare a cost of 2x x4540 with 1TB disks to equivalent solution on 
LTO.

>
>Each x4540 could be configured as: 4x 11 disks in raidz-2 + 2x hot 
spare

>+ 2x OS disks.
>The four raidz2 group form a single pool. This would provide well over
>30TB of logical storage per each box.
>
>Now you rsync all the data from your clients to a dedicated filesystem
>per client, then create a snapshot.
>All snapshots are replicated to a 2nd x4540 so even if you would loose
>entire box/data for some reason you would still have a spare copy.
>
>Now compare it to a cost of a library, lto drives, tapes, software +
>licenses, support costs, ...
>
>See more details at
>http://milek.blogspot.com/2009/12/my-presentation-at-losug.html

I've just read your presentation Robert. Interesting stuff.

I've also just done a pen and paper exercise to see how much 30TB of 
tape would cost as a comparison to your disk based solution.


Using list prices from Sun's website (and who pays list..?), an SL48 
with 2 x LTO3 drives would cost £14000. I couldn't see a price on an 
LTO4 equipped SL48 despite the Sun website saying it's a supported 
option. Each LTO3 has a native capacity of 300GB and the SL48 can 
hold up to 48 tapes in the library (14.4TB native per library). To 
match the 30TB in your solution, we'd need two libraries totalling 
£28000.


You would also need 100 LTO3 tapes to provide 30TB of native storage. 
I recently bought a pack of 20 tapes for £340, so five packs would be 
£1700.


So you could provision a tape backup for just under £3 (~$49000). 
In comparison, the cost of one X4540 with ~ 36TB usable storage is UK 
list price £30900. I've not factored in backup software since you 
could use an open source solution such as Amanda or Bacula.


A more apples to apples comparison would be to compare the storage 
only.  Both removable drive and tape options require a server with FC 
or SCSI ports, so that can be excluded from the comparison.






I think one should actually compare whole solutions - including servers, 
fc infrastructure, tape drives, robots, software costs, rack space, ...


Servers like x4540 are ideal for zfs+rsync backup solution - very 
compact, good $/GB ratio, enough CPU power for its capacity, allow to 
easily scale it horizontally, and it is not too small and not too big. 
Then thanks to its compactness they are very easy to administer.


Depending on an anvironment one could deploy them always in paris - one 
in one datacenter and 2nd one in other datacanter with ZFS send based 
replication of all backups (snapshots). Or one may replicate 
(cross-replicate) only selected clients if needed.



--
Robert Milkowski
http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-21 Thread Robert Milkowski

On 20/01/2010 15:45, David Dyer-Bennet wrote:

On Wed, January 20, 2010 09:23, Robert Milkowski wrote:

   

Now you rsync all the data from your clients to a dedicated filesystem
per client, then create a snapshot.
 


Is there an rsync out there that can reliably replicate all file
characteristics between two ZFS/Solaris systems?  I haven't found one.
The ZFS ACLs seem to be beyond all of them, in particular.

   


No it doesn't support ZFS ACLs - fortunately it is not an issue for us.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Joerg Schilling
Ragnar Sundblad  wrote:

> Yes! Modern LTO drives can typically vary their speed about a factor four
> or so, so even if you can't keep up with the tape drive maximum speed,
> it will typically work pretty good anyway. If you can't keep up even then,
> it will have to stop, back up a bit, and restart, which will be _very_
> slow. Having a disk system deliver data at 240 MB/s at the same time
> as you are writing to it can be a bit of a challenge. 

And star implements a FIFO that is written in a way that dramatically reduces 
the sawtooth behavior seen typically with other applications. You just need to 
tell star to use a say 2 GB FIFO and star will be able to keep the tame 
streaming for a longer time before it waits until there is enough data for the 
next longer streaming period.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Ragnar Sundblad

On 21 jan 2010, at 00.20, Al Hopper wrote:

> I remember for about 5 years ago (before LT0-4 days) that streaming
> tape drives would go to great lengths to ensure that the drive kept
> streaming - because it took so much time to stop, backup and stream
> again.  And one way the drive firmware accomplished that was to write
> blocks of zeros when there was no data available.  This also occurred
> when the backup source was sending a bunch of small files, which took
> longer to stream and did'nt produce enough data to keep the drive
> writing useful data.  And if you had the tape hardware setup to do
> compression, then, assuming a normal 2:1 compression ratio, you'd need
> to source 240Mb/Sec in order to keep the tape writing 120Mb/Sec.  The
> net result was the consumption of a lot more tape than a
> back-of-the-napkin calculation told you was required.
> 
> Obviously at higher compression ratios or with the higher stream data
> write rates you quote below - this problem becomes more troublesome.
> So I agree with your conclusion: "The only way to realistically feed
> that is from disk."

Yes! Modern LTO drives can typically vary their speed about a factor four
or so, so even if you can't keep up with the tape drive maximum speed,
it will typically work pretty good anyway. If you can't keep up even then,
it will have to stop, back up a bit, and restart, which will be _very_
slow. Having a disk system deliver data at 240 MB/s at the same time
as you are writing to it can be a bit of a challenge. 

I haven't seen drives that fill out with zeros. Sounds like an ugly
solution, but maybe it could be useful in some strange case.

/ragge s

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Ragnar Sundblad

On 20 jan 2010, at 17.22, Julian Regel wrote:

> >It is actually not that easy.
> >
> >Compare a cost of 2x x4540 with 1TB disks to equivalent solution on LTO.
> >
> >Each x4540 could be configured as: 4x 11 disks in raidz-2 + 2x hot spare 
> >+ 2x OS disks.
> >The four raidz2 group form a single pool. This would provide well over 
> >30TB of logical storage per each box.
> >
> >Now you rsync all the data from your clients to a dedicated filesystem 
> >per client, then create a snapshot.
> >All snapshots are replicated to a 2nd x4540 so even if you would loose 
> >entire box/data for some reason you would still have a spare copy.
> >
> >Now compare it to a cost of a library, lto drives, tapes, software + 
> >licenses, support costs, ...
> >
> >See more details at 
> >http://milek.blogspot.com/2009/12/my-presentation-at-losug.html
> 
> I've just read your presentation Robert. Interesting stuff.
> 
> I've also just done a pen and paper exercise to see how much 30TB of tape 
> would cost as a comparison to your disk based solution.
> 
> Using list prices from Sun's website (and who pays list..?), an SL48 with 2 x 
> LTO3 drives would cost £14000. I couldn't see a price on an LTO4 equipped 
> SL48 despite the Sun website saying it's a supported option. Each LTO3 has a 
> native capacity of 300GB and the SL48 can hold up to 48 tapes in the library 
> (14.4TB native per library). To match the 30TB in your solution, we'd need 
> two libraries totalling £28000.

LTO3 has native capacity 400 GB, LTO4 has 800. The price is about the same per 
tape and per drive, a little higher for LTO4.

> You would also need 100 LTO3 tapes to provide 30TB of native storage. I 
> recently bought a pack of 20 tapes for £340, so five packs would be £1700.

Or rather 37 LTO4 tapes, and only one 48 tape library. But that doesn't matter, 
the interesting part is that one now can use whatever best solves the problem 
at hand.

> So you could provision a tape backup for just under £3 (~$49000). In 
> comparison, the cost of one X4540 with ~ 36TB usable storage is UK list price 
> £30900. I've not factored in backup software since you could use an open 
> source solution such as Amanda or Bacula.
> 
> Which isn't to say tape would be a "better" solution since it's going to be 
> slower to restore etc. But it does show that tape can work out cheaper, 
> especially since the cost of a high speed WAN link isn't required.

Reading from tape is normally faster than reading from (a single) disk. Seek 
time of course isn't.

/ragge

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Al Hopper
On Wed, Jan 20, 2010 at 2:52 PM, David Magda  wrote:
>
> On Jan 20, 2010, at 12:21, Robert Milkowski wrote:
>
>> On 20/01/2010 16:22, Julian Regel wrote:
>>>
>> [...]
>>>
>>> So you could provision a tape backup for just under £3 (~$49000). In
>>> comparison, the cost of one X4540 with ~ 36TB usable storage is UK list
>>> price £30900. I've not factored in backup software since you could use an
>>> open source solution such as Amanda or Bacula.
>>
>> [...]
>> You would also need to add at least one server to your library with fc
>> cards.
>> Then with most software you would need more tapes due to data
>> fragmentation and a need to do regular full backups (with zfs+rsync you only
>> do a full backup once).
>>
>> So in best case a library will cost about the same as disk based solution
>> but generally will be less flexible, etc. If you would add any enterprise
>> software on top of it (Legato, NetBackup, ...) then the price would change
>> dramaticallly. Additionally with ZFS one could start using deduplication (in
>> testing already).
>
> Regardless of the economics of tape, nowadays you generally need to go to
> disk first because trying to stream at 120 MB/s (LTO-4) really isn't
> practical over the network, directly from the client.

I remember for about 5 years ago (before LT0-4 days) that streaming
tape drives would go to great lengths to ensure that the drive kept
streaming - because it took so much time to stop, backup and stream
again.  And one way the drive firmware accomplished that was to write
blocks of zeros when there was no data available.  This also occurred
when the backup source was sending a bunch of small files, which took
longer to stream and did'nt produce enough data to keep the drive
writing useful data.  And if you had the tape hardware setup to do
compression, then, assuming a normal 2:1 compression ratio, you'd need
to source 240Mb/Sec in order to keep the tape writing 120Mb/Sec.  The
net result was the consumption of a lot more tape than a
back-of-the-napkin calculation told you was required.

Obviously at higher compression ratios or with the higher stream data
write rates you quote below - this problem becomes more troublesome.
So I agree with your conclusion: "The only way to realistically feed
that is from disk."

> So in the end you'll be starting with disk (either DAS or VTL or whatever),
> and generally going to tape if you need to keep stuff that's older than
> (say) 3-6 months. Tape also doesn't rotate while it's sitting there, so if
> it's going to be sitting around for a while (e.g., seven years) better to
> use tape than something that sucks up power.
>
> LTO-5 is expected to be released RSN, with a native capacity of 1.6 TB and
> (uncompressed) writes at 180 MB/s. The only way to realistically feed that
> is from disk.
>
> ___

Regards,

-- 
Al Hopper  Logical Approach Inc,Plano,TX a...@logical-approach.com
   Voice: 972.379.2133 Timezone: US CDT
OpenSolaris Governing Board (OGB) Member - Apr 2005 to Mar 2007
http://www.opensolaris.org/os/community/ogb/ogb_2005-2007/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Joerg Schilling
Miles Nordin  wrote:

> From the perspective of MY business, I would much rather have the dark
> OOB acl/fork/whatever-magic that's gone into ZFS and NFSv4 supported
> in standard tools like rsync and GNUtar.  This is, for example, what

GNU tar does not support any platform speficic feature on any OS.
Don't expect that GNU tar will ever add such properties..

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Joerg Schilling
Ian Collins  wrote:

> > We are talking about TAR and I did give a pointer to the star archive 
> > format 
> > documentation, so it is obvious that I was talking about the ACL format from
> > Sun tar. This format is not documented.
> >
> >   
> It is, Sun's ZFS ACL aware tools use acltotext() to format ACLs.

Please don't reply without checking facts.

The fact that you know that there is salt in the soup does not give
you the whole list of ingredients Please look into the Sun tar format
to understand that you are wrong.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread David Magda


On Jan 20, 2010, at 12:21, Robert Milkowski wrote:


On 20/01/2010 16:22, Julian Regel wrote:



[...]
So you could provision a tape backup for just under £3 (~ 
$49000). In comparison, the cost of one X4540 with ~ 36TB usable  
storage is UK list price £30900. I've not factored in backup  
software since you could use an open source solution such as Amanda  
or Bacula.

[...]
You would also need to add at least one server to your library with  
fc cards.
Then with most software you would need more tapes due to data  
fragmentation and a need to do regular full backups (with zfs+rsync  
you only do a full backup once).


So in best case a library will cost about the same as disk based  
solution but generally will be less flexible, etc. If you would add  
any enterprise software on top of it (Legato, NetBackup, ...) then  
the price would change dramaticallly. Additionally with ZFS one  
could start using deduplication (in testing already).


Regardless of the economics of tape, nowadays you generally need to go  
to disk first because trying to stream at 120 MB/s (LTO-4) really  
isn't practical over the network, directly from the client.


So in the end you'll be starting with disk (either DAS or VTL or  
whatever), and generally going to tape if you need to keep stuff  
that's older than (say) 3-6 months. Tape also doesn't rotate while  
it's sitting there, so if it's going to be sitting around for a while  
(e.g., seven years) better to use tape than something that sucks up  
power.


LTO-5 is expected to be released RSN, with a native capacity of 1.6 TB  
and (uncompressed) writes at 180 MB/s. The only way to realistically  
feed that is from disk.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Miles Nordin
> "jr" == Julian Regel  writes:

jr> While I am sure that star is technically a fine utility, the
jr> problem is that it is effectively an unsupported product.

I have no problems with this whatsoever.

jr> If our customers find a bug in their backup that is caused by
jr> a failure in a Sun supplied utility, then they have a legal
jr> course of action. The customer's system administrators are
jr> covered because they were using tools provided by the
jr> vendor. The wrath of the customer would be upon Sun, not the
jr> supplier (us) or the supplier's technical lead (me).

We were just talking about this somewhere else, actually: ``if
something goes wrong, its their ass. but if nothing ever gets done,
its nobody's fault.''  It's sad for me how much money is to be made
supporting broken corporate cultures like that.

I'm not saying you're wrong, just that you might not want to
contribute to such a culture because you've chosen to endure it for a
scratch.  You need to have a better way to evaluate employees than
micromanagement-by-the-clueless and vindictive hindsight.  But the
point that there's money to be made by bleeding it out of ossified
broken American companies is well-taken.

jr> From the perspective of the business, the system administrator
jr> will have acted irresponsibly by choosing a tool that has no
jr> vendor support.

From the perspective of MY business, I would much rather have the dark
OOB acl/fork/whatever-magic that's gone into ZFS and NFSv4 supported
in standard tools like rsync and GNUtar.  This is, for example, what
Apple achieved with CUPS and why I can share printers between Ubuntu
and Mac OS effortlessly, and this increases the amount of money I'm
willing to give Apple for their proprietary platform.  The purpose of
the tool I'm discussing definitely includes the same level of
cooperation, so working with the existing best-in-class and
most-popular tools, and reasonableness, might be better than brittle
CYA support in some fringey '/opt/SUNWbkpkit/bin/VendorCP -Rf' tool.

Even if you get your cyaCP tool you may find it doesn't achieve the
ass-covering you wanted because these tools can be cheeky little
bastards.  Most of the other quirky little balkanized-platform
Solaris-only tools are littered with straightjacketing assertions to
avoid ``call generators'' and push the blame back onto the sysadmin,
then there is some ``all bets are off'' flag to allowe you to actually
accomplish job, like 'NOINUSECHECK=1 format -e'.  

Honestly...why bother playing this game?


pgpBTS02xCBZW.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Ian Collins

Joerg Schilling wrote:

Ian Collins  wrote:

  

The correct way to archivbe ACLs would be to put them into extended POSIX tar
attrubutes as star does.

See http://cdrecord.berlios.de/private/man/star/star.4.html for a format 
documentation or have a look at ftp://ftp.berlios.de/pub/star/alpha, e.g.

ftp://ftp.berlios.de/pub/star/alpha/acl-test.tar.gz

The ACL format used by Sun is undocumented..

  
  

man acltotext



We are talking about TAR and I did give a pointer to the star archive format 
documentation, so it is obvious that I was talking about the ACL format from

Sun tar. This format is not documented.

  

It is, Sun's ZFS ACL aware tools use acltotext() to format ACLs.

--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Ian Collins

Julian Regel wrote:

>It is actually not that easy.
>
>Compare a cost of 2x x4540 with 1TB disks to equivalent solution on LTO.
>
>Each x4540 could be configured as: 4x 11 disks in raidz-2 + 2x hot spare
>+ 2x OS disks.
>The four raidz2 group form a single pool. This would provide well over
>30TB of logical storage per each box.
>
>Now you rsync all the data from your clients to a dedicated filesystem
>per client, then create a snapshot.
>All snapshots are replicated to a 2nd x4540 so even if you would loose
>entire box/data for some reason you would still have a spare copy.
>
>Now compare it to a cost of a library, lto drives, tapes, software +
>licenses, support costs, ...
>
>See more details at
>http://milek.blogspot.com/2009/12/my-presentation-at-losug.html

I've just read your presentation Robert. Interesting stuff.

I've also just done a pen and paper exercise to see how much 30TB of 
tape would cost as a comparison to your disk based solution.


Using list prices from Sun's website (and who pays list..?), an SL48 
with 2 x LTO3 drives would cost £14000. I couldn't see a price on an 
LTO4 equipped SL48 despite the Sun website saying it's a supported 
option. Each LTO3 has a native capacity of 300GB and the SL48 can hold 
up to 48 tapes in the library (14.4TB native per library). To match 
the 30TB in your solution, we'd need two libraries totalling £28000.


You would also need 100 LTO3 tapes to provide 30TB of native storage. 
I recently bought a pack of 20 tapes for £340, so five packs would be 
£1700.


So you could provision a tape backup for just under £3 (~$49000). 
In comparison, the cost of one X4540 with ~ 36TB usable storage is UK 
list price £30900. I've not factored in backup software since you 
could use an open source solution such as Amanda or Bacula.


A more apples to apples comparison would be to compare the storage 
only.  Both removable drive and tape options require a server with FC or 
SCSI ports, so that can be excluded from the comparison.


So for 30TB, assuming 2TB drives @ ~£100 with a pool built of 6 drive 
raidz vdevs 18 drives would be required plus 2 16 drive shelves.  So 
each backup set would cost about £1800.  So there's not a great deal of 
difference.  With drives you also get the added benefit of keeping all 
your incrementals (as snapshots) on the archive set.


HDD price per GB will continue to drop faster than tape, so it will be 
interesting to do the same comparison in 12 months.


--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Richard Elling
On Jan 20, 2010, at 3:15 AM, Joerg Schilling wrote:

> Richard Elling  wrote:
> 
>>> 
>>> ufsdump/restore was perfect in that regard.  The lack of equivalent 
>>> functionality is a big problem for the situations where this functionality 
>>> is a business requirement.
>> 
>> How quickly we forget ufsdump's limitations :-).  For example, it is not 
>> supported
>> for use on an active file system (known data corruption possibility) and 
>> UFS snapshots are, well, a poor hack and often not usable for backups.
>> As the ufsdump(1m) manpage says,
> 
> It seems you forgot that zfs also needs snapshots. There is nothing bad with 
> snapshots.

Yes, snapshots are a good thing. But most people who try fssnap 
on the UFS root file system will discover that it doesn't work; for 
reasons mentioned in the NOTES section of fssnap_ufs(1m). 
fssnap_ufs is simply a butt-ugly hack. So if you believe you can
reliably use ufsdump to store a DR copy of root for a 7x24x365 
production environment, then you probably believe the Backup
Fairy will leave a coin under your pillow when your restore fails :-)

Fortunately, ZFS snapshot do the right thing.
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Robert Milkowski

On 20/01/2010 17:21, Robert Milkowski wrote:

On 20/01/2010 16:22, Julian Regel wrote:

>It is actually not that easy.
>
>Compare a cost of 2x x4540 with 1TB disks to equivalent solution on LTO.
>
>Each x4540 could be configured as: 4x 11 disks in raidz-2 + 2x hot 
spare

>+ 2x OS disks.
>The four raidz2 group form a single pool. This would provide well over
>30TB of logical storage per each box.
>
>Now you rsync all the data from your clients to a dedicated filesystem
>per client, then create a snapshot.
>All snapshots are replicated to a 2nd x4540 so even if you would loose
>entire box/data for some reason you would still have a spare copy.
>
>Now compare it to a cost of a library, lto drives, tapes, software +
>licenses, support costs, ...
>
>See more details at
>http://milek.blogspot.com/2009/12/my-presentation-at-losug.html

I've just read your presentation Robert. Interesting stuff.

I've also just done a pen and paper exercise to see how much 30TB of 
tape would cost as a comparison to your disk based solution.


Using list prices from Sun's website (and who pays list..?), an SL48 
with 2 x LTO3 drives would cost £14000. I couldn't see a price on an 
LTO4 equipped SL48 despite the Sun website saying it's a supported 
option. Each LTO3 has a native capacity of 300GB and the SL48 can 
hold up to 48 tapes in the library (14.4TB native per library). To 
match the 30TB in your solution, we'd need two libraries totalling 
£28000.


You would also need 100 LTO3 tapes to provide 30TB of native storage. 
I recently bought a pack of 20 tapes for £340, so five packs would be 
£1700.


So you could provision a tape backup for just under £3 (~$49000). 
In comparison, the cost of one X4540 with ~ 36TB usable storage is UK 
list price £30900. I've not factored in backup software since you 
could use an open source solution such as Amanda or Bacula.


Which isn't to say tape would be a "better" solution since it's going 
to be slower to restore etc. But it does show that tape can work out 
cheaper, especially since the cost of a high speed WAN link isn't 
required.


JR

You would also need to add at least one server to your library with fc 
cards.
Then with most software you would need more tapes due to data 
fragmentation and a need to do regular full backups (with zfs+rsync 
you only do a full backup once).


So in best case a library will cost about the same as disk based 
solution but generally will be less flexible, etc. If you would add 
any enterprise software on top of it (Legato, NetBackup, ...) then the 
price would change dramaticallly. Additionally with ZFS one could 
start using deduplication (in testing already).





What I really mean is that a disk based solution used to be much more 
expensive than tapes but currently they are comparable in costs while 
often the disk based is more flexible.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Robert Milkowski

On 20/01/2010 16:22, Julian Regel wrote:

>It is actually not that easy.
>
>Compare a cost of 2x x4540 with 1TB disks to equivalent solution on LTO.
>
>Each x4540 could be configured as: 4x 11 disks in raidz-2 + 2x hot spare
>+ 2x OS disks.
>The four raidz2 group form a single pool. This would provide well over
>30TB of logical storage per each box.
>
>Now you rsync all the data from your clients to a dedicated filesystem
>per client, then create a snapshot.
>All snapshots are replicated to a 2nd x4540 so even if you would loose
>entire box/data for some reason you would still have a spare copy.
>
>Now compare it to a cost of a library, lto drives, tapes, software +
>licenses, support costs, ...
>
>See more details at
>http://milek.blogspot.com/2009/12/my-presentation-at-losug.html

I've just read your presentation Robert. Interesting stuff.

I've also just done a pen and paper exercise to see how much 30TB of 
tape would cost as a comparison to your disk based solution.


Using list prices from Sun's website (and who pays list..?), an SL48 
with 2 x LTO3 drives would cost £14000. I couldn't see a price on an 
LTO4 equipped SL48 despite the Sun website saying it's a supported 
option. Each LTO3 has a native capacity of 300GB and the SL48 can hold 
up to 48 tapes in the library (14.4TB native per library). To match 
the 30TB in your solution, we'd need two libraries totalling £28000.


You would also need 100 LTO3 tapes to provide 30TB of native storage. 
I recently bought a pack of 20 tapes for £340, so five packs would be 
£1700.


So you could provision a tape backup for just under £3 (~$49000). 
In comparison, the cost of one X4540 with ~ 36TB usable storage is UK 
list price £30900. I've not factored in backup software since you 
could use an open source solution such as Amanda or Bacula.


Which isn't to say tape would be a "better" solution since it's going 
to be slower to restore etc. But it does show that tape can work out 
cheaper, especially since the cost of a high speed WAN link isn't 
required.


JR

You would also need to add at least one server to your library with fc 
cards.
Then with most software you would need more tapes due to data 
fragmentation and a need to do regular full backups (with zfs+rsync you 
only do a full backup once).


So in best case a library will cost about the same as disk based 
solution but generally will be less flexible, etc. If you would add any 
enterprise software on top of it (Legato, NetBackup, ...) then the price 
would change dramaticallly. Additionally with ZFS one could start using 
deduplication (in testing already).



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Miles Nordin
> "ae" == Allen Eastwood  writes:
> "ic" == Ian Collins  writes:

 >> If people are really still backing up to tapes or DVD's, just
 >> use file vdev's, export the pool, and then copy the unmounted
 >> vdev onto the tape or DVD.

ae> And some of those enterprises require backup mechanism that
ae> can be easily used in a DR situation.

ae> ufsdump/restore was perfect in that regard.  The lack of
ae> equivalent functionality is a big problem for the situations
ae> where this functionality is a business requirement.

ae> For example, one customer, local government, requires a backup
ae> that can be taken offsite and used in a DR situation.

Were you confused by some part of:

 Use file vdevs, epxort the pool, and then copy the unmounted vdev
 onto the tape.

or do you find that this doesn't do what you want?  because it seems
fine to me.  And the fact that it doesn't need any extra tools means
it's unlikely to break (1) far into the future or (2) for a few
unlucky builds, and (3) that the restore environment is simple and
doesn't involve prepopulating some Legato Database with the TOC of
every tape in the library or some such nonsense, which ought to all be
among your ``requirements'' but if you're substituting for those,
``works in exactly the way we were used to it working before'' then
you may as well use 'zfs send' since you're more concerned with
identical-feeling invocatoin syntax than the problems I mentioned.

ic> For a full recovery, you can archive a send stream and receive
ic> it back.

You can send the stream to the tape, transport the tape to the DR
site, and receive it.  You can do this weekly as part of your offsite
backup plan provided that you receive each tape you transport
immediately.  Then the data should be permanently stored on disk at
the DR site, and the tapes used only for transport.

If you store the backup permanently on tape then it's a step backwards
from tar/cpio/ufsrestore because the 'zfs send' format is more fragile
and has to be restored entire.  If you receive the tape immediately
this is an improvement because under the old convention tapes could be
damaged in transit, or over the years by heat/dust/sunlight, without
your knowledge, while on disks it's simple to scrub periodically.

I am not trying to take away your tapes, Allen, so please quote Ian
instead if that's the thing you object to.  I've instead suggested a
different way to use them if you really do need them archivally: store
file vdev's on them.  If you're just using them to replicate data to
the DR site then you needn't even go as far as my workaround.

I do agree that there's a missing tool: it's not possible to copy one
subdirectory to another while preserving holes, forkey extended
attributes, and ACL's.  Also if Windows ACL's are going to be stored
right in the filesystem, then Windows ACL's probably ought to be
preserved over an rsync pipe between Solaris and EnTee, or a
futuristic tarball written on one and extracted on the other.  I don't
agree that the missing tool is designed primarily for the narrow
use-case of writing to ancient backup tapes: it's a more general tool.
or, really, it's just a matter of documenting and committing the
extra-OOB-gunk APIs and then fixing rsync and GNUtar.


pgpk5wk2koJcZ.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Julian Regel
>It is actually not that easy.
>
>Compare a cost of 2x x4540 with 1TB disks to equivalent solution on LTO.
>
>Each x4540 could be configured as: 4x 11 disks in raidz-2 + 2x hot spare 
>+ 2x OS disks.
>The four raidz2 group form a single pool. This would provide well over 
>30TB of logical storage per each box.
>
>Now you rsync all the data from your clients to a dedicated filesystem 
>per client, then create a snapshot.
>All snapshots are replicated to a 2nd x4540 so even if you would loose 
>entire box/data for some reason you would still have a spare copy.
>
>Now compare it to a cost of a library, lto drives, tapes, software + 
>licenses, support costs, ...
>
>See more details at 
>http://milek.blogspot.com/2009/12/my-presentation-at-losug.html

I've just read your presentation Robert. Interesting stuff.

I've also just done a pen and paper exercise to see how much 30TB of tape would 
cost as a comparison to your disk based solution.

Using list prices from Sun's website (and who pays list..?), an SL48 with 2 x 
LTO3 drives would cost £14000. I couldn't see a price on an LTO4 equipped SL48 
despite the Sun website saying it's a supported option. Each LTO3 has a native 
capacity of 300GB and the SL48 can hold up to 48 tapes in the library (14.4TB 
native per library). To match the 30TB in your solution, we'd need two 
libraries totalling £28000.

You would also need 100 LTO3 tapes to provide 30TB of native storage. I 
recently bought a pack of 20 tapes for £340, so five packs would be £1700.


So you could provision a tape backup for just under £3 (~$49000). In 
comparison, the cost of one X4540 with ~ 36TB usable storage is UK list price 
£30900. I've not factored in backup software since you could use an open source 
solution such as Amanda or Bacula.

Which isn't to say tape would be a "better" solution since it's going to be 
slower to restore etc. But it does show that tape can work out cheaper, 
especially since the cost of a high speed WAN link isn't required.

JR



  ___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Bob Friesenhahn

On Wed, 20 Jan 2010, Julian Regel wrote:


If our customers find a bug in their backup that is caused by a 
failure in a Sun supplied utility, then they have a legal course of 
action. The customer's system administrators are covered because 
they were using tools provided by the vendor. The wrath of the 
customer would be upon Sun, not the supplier (us) or the supplier's 
technical lead (me).


I would love to try whatever you are smoking because it must be really 
good stuff.  It would be a bold new step for me, but the benefits are 
clear.


While your notions of the transitive protection offered by vendor 
"support" are interesting, I will be glad to meet you in the 
unemployment line then we can share some coffee and discuss the good 
old days.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread David Dyer-Bennet

On Wed, January 20, 2010 04:48, Ragnar Sundblad wrote:

> LTO media is still cheaper than equivalent sized disks, maybe a factor 5
> or so. LTO drivers cost a little, but so do disk shelves. So, now that
> there is no big price issue, there is choice instead. Use it!

Depends on the scale you're operating at.

Backing up my 800GB home data pool onto a couple of external 1TB USB
drives is *immensely* cheaper than buying tape equipment.

At enough bigger scales, I accept that tape is still cheaper.  Makes
sense, since the tapes are relatively simple compared to drives, and you
only need a small number of drives to use a large number of tapes.

I think hard drives are still cheaper at small-enterprise levels, actually.

-- 
David Dyer-Bennet, d...@dd-b.net; http://dd-b.net/
Snapshots: http://dd-b.net/dd-b/SnapshotAlbum/data/
Photos: http://dd-b.net/photography/gallery/
Dragaera: http://dragaera.info

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread David Dyer-Bennet

On Wed, January 20, 2010 09:23, Robert Milkowski wrote:

> Now you rsync all the data from your clients to a dedicated filesystem
> per client, then create a snapshot.


Is there an rsync out there that can reliably replicate all file
characteristics between two ZFS/Solaris systems?  I haven't found one. 
The ZFS ACLs seem to be beyond all of them, in particular.

(Losing just that, and preserving the data, is clearly far, far better
than losing everything!  And a system build *knowing* it was losing the
protections could preserve them some other way.)

-- 
David Dyer-Bennet, d...@dd-b.net; http://dd-b.net/
Snapshots: http://dd-b.net/dd-b/SnapshotAlbum/data/
Photos: http://dd-b.net/photography/gallery/
Dragaera: http://dragaera.info

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Robert Milkowski

On 20/01/2010 11:26, Joerg Schilling wrote:

Edward Ned Harvey  wrote:

   

Star implements this in a very effective way (by using libfind) that is
even
faster that the find(1) implementation from Sun.
   

Even if I just "find" my filesystem, it will run for 7 hours.  But zfs can
create my whole incremental snapshot in a minute or two.  There is no way
star or any other user-space utility that walks the filesystem can come
remotely close to this performance.  Such performance can only be
implemented at the filesystem level, or lower.
 

You claim that it is fast for you but this is because it is block oriented and
because you probably changed only few data.

If you like to have a backup that allows to access files, you need a file based
backup and I am sure that even a filesystem level scan for recently changed
files will not be much faster than what you may achive with e.g. star.

   


Not really - I mean if you do an incremental zfs send and on the other 
side receive it so a filesystem/snapshot is created then you have an 
access to each individual file while having the benefit of doing block 
level incremental zfs send which in many environments will be way 
faster then rsync/start/...


--
Robert Milkowski
http://milek.blogspot.com


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Robert Milkowski

On 20/01/2010 10:48, Ragnar Sundblad wrote:

On 19 jan 2010, at 20.11, Ian Collins wrote:

   

Julian Regel wrote:
 

Based on what I've seen in other comments, you might be right. Unfortunately, I 
don't feel comfortable backing up ZFS filesystems because the tools aren't 
there to do it (built into the operating system or using Zmanda/Amanda).

   

Commercial backup solutions are available for ZFS.
 

I know tape backup isn't sexy, but it's a reality for many of us and it's not 
going away anytime soon.

   

True, but I wonder how viable its future is.  One of my clients requires 17 
LT04 types for a full backup, which cost more and takes up more space than the 
equivalent in removable hard drives.

In the past few years growth in hard drive capacities has outstripped tapes to 
the extent that removable hard drives and ZFS snapshots have become a more cost 
effective and convenient backup media.
 

LTO media is still cheaper than equivalent sized disks, maybe a factor 5 or so. 
LTO drivers cost a little, but so do disk shelves. So, now that there is no big 
price issue, there is choice instead. Use it!

Hard drives are good for random access - both restore of individual files and 
partial rewrite.

Hard drivers aren't faster than tape for data transfer, but they might be 
cheaper to run in parallel and therefore you could potentially gain speed. Hard 
drives have shorter seek time, which may be important.

Hard drives are probably bad for storing for longer times - especially - you 
will never know how long it could be stored before it will fail. A month? 
Probably. A year? Maybe. Five years? Well... Ten years? Probably not. LTO tapes 
are supposed to be able to keep it's data for at least 30 years of stored 
properly. Hard drives are probably best when used online or at least very often.

So - it is wrong to say that one is better or cheaper than the other. They have 
different properties, and could be used to solve different problems.


   


It is actually not that easy.

Compare a cost of 2x x4540 with 1TB disks to equivalent solution on LTO.

Each x4540 could be configured as: 4x 11 disks in raidz-2 + 2x hot spare 
+ 2x OS disks.
The four raidz2 group form a single pool. This would provide well over 
30TB of logical storage per each box.


Now you rsync all the data from your clients to a dedicated filesystem 
per client, then create a snapshot.
All snapshots are replicated to a 2nd x4540 so even if you would loose 
entire box/data for some reason you would still have a spare copy.


Now compare it to a cost of a library, lto drives, tapes, software + 
licenses, support costs, ...


See more details at 
http://milek.blogspot.com/2009/12/my-presentation-at-losug.html


--
Robert Milkowski
http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Robert Milkowski

On 19/01/2010 19:11, Ian Collins wrote:

Julian Regel wrote:


Based on what I've seen in other comments, you might be right. 
Unfortunately, I don't feel comfortable backing up ZFS filesystems 
because the tools aren't there to do it (built into the operating 
system or using Zmanda/Amanda).



Commercial backup solutions are available for ZFS.
I know tape backup isn't sexy, but it's a reality for many of us and 
it's not going away anytime soon.


True, but I wonder how viable its future is.  One of my clients 
requires 17 LT04 types for a full backup, which cost more and takes up 
more space than the equivalent in removable hard drives.


In the past few years growth in hard drive capacities has outstripped 
tapes to the extent that removable hard drives and ZFS snapshots have 
become a more cost effective and convenient backup media.


What do people with many tens of TB use for backup these days?


http://milek.blogspot.com/2009/12/my-presentation-at-losug.html

--
Robert Milkowski
http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Michael Schuster

Joerg Schilling wrote:

Julian Regel  wrote:

If you like to have a backup that allows to access files, you need a file based 
backup and I am sure that even a filesystem level scan for recently changed 
files will not be much faster than what you may achive with e.g. star.


Note that ufsdump directly accesees the raw disk device and thus _is_ at 
filesystem leven but still is slower than star on UFS.

While I am sure that star is technically a fine utility, the problem is that it 
is effectively an unsupported product.


From this viewpoint, you may call most of Solaris "unsupported".


what is that supposed to mean?

Michael
--
Michael Schusterhttp://blogs.sun.com/recursion
Recursion, n.: see 'Recursion'
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Joerg Schilling
Julian Regel  wrote:

> >> While I am sure that star is technically a fine utility, the problem is 
> >> that it is effectively an unsupported product.
>
> >From this viewpoint, you may call most of Solaris "unsupported".
>
> From the perspective of the business, the contract with Sun provides that 
> support.

>From a perspective of reality, such a contract will not help.

> >Do you really believe that Sun will help such a customer?
> >There are many bugs in Solaris (I remember e.g. some showstopper
> >bugs in the multimedia area) that are not fixed although they are known
> >since a very long time (more than a year).
> >There is a bug in ACL handling in Sun's tar (reported by me in 2004 or even 
> >before) that is not fixed. As a result in many cases ACLs are not restored.
>
> If Sun don't fix a critical bug that is affecting the availability of
> server that is under support, then it becomes a problem for the legal
> department. In the ACL example, it's possible the effected users didn't have 
> a support contract.

What you seem to point out is that in case of a problem for a customer with a 
contract, the legal department gets involved. Unfortunately, laywers do not fix 
bugs.

> >Note that bugs in star are fixed much faster and looking back at the 28 years
> >of history with star, I know of not a single bug that took more than 3 months
> >to get a fix. Typically, bugs are fixed withing less than a week - many bugs
> >even within a few hours. This is a support quality that Sun does not offer.
>
> Possibly, but there is no guarantee that it will be fixed, no-one to call 
> when there is a problem, no-one to escalate the problem to if it is ignored, 
> and no company to sue if it all goes wrong.

Escalating a problem does not fix it. 

> >So please explain us where you see a problem with star..
>
> Hopefully my above comments explain sufficiently. It's not a technical issue 
> with star, it's a business issue. The rules there are very different and not 
> based on merit (this is also why many companies prefer running their mission 
> critical apps on Red Hat Enterprise Linux instead of CentOS, even though 
> technically they are almost identical).

Now we are back to reality.

A person that is interested in a solution will usually check what happened
in similar cases before. If you compare star with Sun supplied tools with
this background, Sun cannot outperform star.

"Red Hat Enterprise Linux" may offer something you cannot get with CentOS.
But I don't see that Sun can offer something you don't get with star.

Let me make another reality check:

Many people use GNU tar for backup purposes, but my first automated test case
with incremental backups using GNU tar did fail so miserably that I was unable
to use GNU tar as a test reference at all.

On the other side, I am doing incremental backup _and_ restore tests with 
gigabytes of real delta data on a dayly base since 2004 and I did hot see any 
problem since April 2005.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Julian Regel
>> While I am sure that star is technically a fine utility, the problem is that 
>> it is effectively an unsupported product.

>From this viewpoint, you may call most of Solaris "unsupported".

From the perspective of the business, the contract with Sun provides that 
support.

>> If our customers find a bug in their backup that is caused by a failure in a 
>> Sun supplied utility, then they have a legal course of action. The 
>> >>customer's system administrators are covered because they were using tools 
>> provided by the vendor. The wrath of the customer would be upon >>Sun, not 
>> the supplier (us) or the supplier's technical lead (me).

>Do you really believe that Sun will help such a customer?
>There are many bugs in Solaris (I remember e.g. some showstopper
>bugs in the multimedia area) that are not fixed although they are known
>since a very long time (more than a year).
>There is a bug in ACL handling in Sun's tar (reported by me in 2004 or even 
>before) that is not fixed. As a result in many cases ACLs are not restored.

If Sun don't fix a critical bug that is affecting the availability of
server that is under support, then it becomes a problem for the legal
department. In the ACL example, it's possible the effected users didn't have a 
support contract.

>Note that bugs in star are fixed much faster and looking back at the 28 years
>of history with star, I know of not a single bug that took more than 3 months
>to get a fix. Typically, bugs are fixed withing less than a week - many bugs
>even within a few hours. This is a support quality that Sun does not offer.

Possibly, but there is no guarantee that it will be fixed, no-one to call when 
there is a problem, no-one to escalate the problem to if it is ignored, and no 
company to sue if it all goes wrong.

>So please explain us where you see a problem with star..

Hopefully my above comments explain sufficiently. It's not a technical issue 
with star, it's a business issue. The rules there are very different and not 
based on merit (this is also why many companies prefer running their mission 
critical apps on Red Hat Enterprise Linux instead of CentOS, even though 
technically they are almost identical).

JR



  ___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Joerg Schilling
Julian Regel  wrote:

> > If you like to have a backup that allows to access files, you need a file 
> > based 
>
> > backup and I am sure that even a filesystem level scan for recently changed 
> > files will not be much faster than what you may achive with e.g. star.
> >
> > Note that ufsdump directly accesees the raw disk device and thus _is_ at 
> > filesystem leven but still is slower than star on UFS.
>
> While I am sure that star is technically a fine utility, the problem is that 
> it is effectively an unsupported product.

>From this viewpoint, you may call most of Solaris "unsupported".

> If our customers find a bug in their backup that is caused by a failure in a 
> Sun supplied utility, then they have a legal course of action. The customer's 
> system administrators are covered because they were using tools provided by 
> the vendor. The wrath of the customer would be upon Sun, not the supplier 
> (us) or the supplier's technical lead (me).

Do you really believe that Sun will help such a customer?
There are many bugs in Solaris (I remember e.g. some showstopper
bugs in the multimedia area) that are not fixed although they are known
since a very long time (more than a year).

There is a bug in ACL handling in Sun's tar (reported by me in 2004 or even 
before) that is not fixed. As a result in many cases ACLs are not restored.

Note that bugs in star are fixed much faster and looking back at the 28 years
of history with star, I know of not a single bug that took more than 3 months
to get a fix. Typically, bugs are fixed withing less than a week - many bugs
even within a few hours. This is a support quality that Sun does not offer.

So please explain us where you see a problem with star..

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Julian Regel
> If you like to have a backup that allows to access files, you need a file 
> based 

> backup and I am sure that even a filesystem level scan for recently changed 
> files will not be much faster than what you may achive with e.g. star.
>
> Note that ufsdump directly accesees the raw disk device and thus _is_ at 
> filesystem leven but still is slower than star on UFS.

While I am sure that star is technically a fine utility, the problem is that it 
is effectively an unsupported product.

If our customers find a bug in their backup that is caused by a failure in a 
Sun supplied utility, then they have a legal course of action. The customer's 
system administrators are covered because they were using tools provided by the 
vendor. The wrath of the customer would be upon Sun, not the supplier (us) or 
the supplier's technical lead (me).

If the system administrator has chosen star (or if the supplier recommends 
star), then the conversation becomes a lot more awkward. From the perspective 
of the business, the system administrator will have acted irresponsibly by 
choosing a tool that has no vendor support. Alternatively, the supplier will be 
held responsible for recommending a product that has broken the customer's 
ability to restore, and with no legal recourse, I wouldn't dare touch it. Sorry.

This is why Sun need to provide the solution themselves (or adopt and provide 
support for star or similar third party products).

JR



  ___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Julian Regel
While I can appreciate that ZFS snapshots are very useful in being able to 
recover files that users might have deleted, they do not do much to help when 
the entire disk array experiences a crash/corruption or catches fire. Backing 
up to a second array helps if a) the array is off-site and for many of us the 
cost of remote links with sufficient bandwidth is still prohibitive,or b) on 
the local network but sufficiently far away from the original array such that 
the fire does not corrupt damage the backup as well.

This leaves some form of removable storage. I'm not sure I'm aware of any 
enterprise-level removable disk solution, primarily because disk isn't really 
designed to be used for offsite backup whereas tape is.

The biggest problem with tape was finding a sufficiently large window in which 
to perform the backup. ZFS snapshots completely solves this issue, but Sun have 
failed to provide the mechanism to protect the data off-site.



  ___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Joerg Schilling
Edward Ned Harvey  wrote:

> > Star implements this in a very effective way (by using libfind) that is
> > even
> > faster that the find(1) implementation from Sun.
>
> Even if I just "find" my filesystem, it will run for 7 hours.  But zfs can
> create my whole incremental snapshot in a minute or two.  There is no way
> star or any other user-space utility that walks the filesystem can come
> remotely close to this performance.  Such performance can only be
> implemented at the filesystem level, or lower.

You claim that it is fast for you but this is because it is block oriented and 
because you probably changed only few data. 

If you like to have a backup that allows to access files, you need a file based 
backup and I am sure that even a filesystem level scan for recently changed 
files will not be much faster than what you may achive with e.g. star.

Note that ufsdump directly accesees the raw disk device and thus _is_ at 
filesystem leven but still is slower than star on UFS.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Joerg Schilling
Ian Collins  wrote:

> > The correct way to archivbe ACLs would be to put them into extended POSIX 
> > tar
> > attrubutes as star does.
> >
> > See http://cdrecord.berlios.de/private/man/star/star.4.html for a format 
> > documentation or have a look at ftp://ftp.berlios.de/pub/star/alpha, e.g.
> > ftp://ftp.berlios.de/pub/star/alpha/acl-test.tar.gz
> >
> > The ACL format used by Sun is undocumented..
> >
> >   
> man acltotext

We are talking about TAR and I did give a pointer to the star archive format 
documentation, so it is obvious that I was talking about the ACL format from
Sun tar. This format is not documented.



Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Joerg Schilling
Richard Elling  wrote:

> > 
> > ufsdump/restore was perfect in that regard.  The lack of equivalent 
> > functionality is a big problem for the situations where this functionality 
> > is a business requirement.
>
> How quickly we forget ufsdump's limitations :-).  For example, it is not 
> supported
> for use on an active file system (known data corruption possibility) and 
> UFS snapshots are, well, a poor hack and often not usable for backups.
> As the ufsdump(1m) manpage says,

It seems you forgot that zfs also needs snapshots. There is nothing bad with 
snapshots.

When I was talking with Jeff Bonwick in September 2004 (before ZFS became 
public), the only feature that was missing in Solaris for a 100% correct
backup based on star was an interface for holey files, so we designed it.

I believe the only mistake from ufsdumps is that is does not use standard
OS interfaces and that it does not use a standard archive format. You get
both with star and star is even faster than ufsdump.


Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Ragnar Sundblad

On 19 jan 2010, at 20.11, Ian Collins wrote:

> Julian Regel wrote:
>> 
>> Based on what I've seen in other comments, you might be right. 
>> Unfortunately, I don't feel comfortable backing up ZFS filesystems because 
>> the tools aren't there to do it (built into the operating system or using 
>> Zmanda/Amanda).
>> 
> Commercial backup solutions are available for ZFS.
>> I know tape backup isn't sexy, but it's a reality for many of us and it's 
>> not going away anytime soon.
>> 
> True, but I wonder how viable its future is.  One of my clients requires 17 
> LT04 types for a full backup, which cost more and takes up more space than 
> the equivalent in removable hard drives.
> 
> In the past few years growth in hard drive capacities has outstripped tapes 
> to the extent that removable hard drives and ZFS snapshots have become a more 
> cost effective and convenient backup media.

LTO media is still cheaper than equivalent sized disks, maybe a factor 5 or so. 
LTO drivers cost a little, but so do disk shelves. So, now that there is no big 
price issue, there is choice instead. Use it!

Hard drives are good for random access - both restore of individual files and 
partial rewrite. 

Hard drivers aren't faster than tape for data transfer, but they might be 
cheaper to run in parallel and therefore you could potentially gain speed. Hard 
drives have shorter seek time, which may be important.

Hard drives are probably bad for storing for longer times - especially - you 
will never know how long it could be stored before it will fail. A month? 
Probably. A year? Maybe. Five years? Well... Ten years? Probably not. LTO tapes 
are supposed to be able to keep it's data for at least 30 years of stored 
properly. Hard drives are probably best when used online or at least very often.

So - it is wrong to say that one is better or cheaper than the other. They have 
different properties, and could be used to solve different problems.

/ragge s

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-20 Thread Ian Collins

Allen Eastwood wrote:

On Jan 19, 2010, at 22:54 , Ian Collins wrote:

  

Allen Eastwood wrote:


On Jan 19, 2010, at 18:48 , Richard Elling wrote:

 
  

Many people use send/recv or AVS for disaster recovery on the inexpensive
side. Obviously, enterprise backup systems also provide DR capabilities.
Since ZFS has snapshots that actually work, and you can use send/receive
or other backup solutions on snapshots, I assert the "problem" is low priority.

   

What I have issue with is the idea that no one uses/should use tape any more.  There are places for tape and it still has value as a backup device.  In many cases in the past, ufsdump, despite it's many issues, was able to restore working OS's, or individual files.  Perfect, not by a long shot.  But it did get the job done. 
  
As was pointed out earlier, all I needed was a Solaris CD (or network boot) and I could restore.  Entire OS gone, boot and ufsrestore.  Critical files deleted, same thing…and I can restore just the file(s) I need.  And while it's been a few years since I've read the man page on ufsdump, ufsrestore and fssnap, those tools have proven useful when dealing with a downed system.
 
  

For a full recovery, you can archive a send stream and receive it back.  With 
ZFS snapshots, the need for individual file recovery from tape is much reduced. 
 The backup server I manage for a large client has 60 days of snaps and I can't 
remember when they had to go to tape to recover a file.

--
Ian.




Let's see…

For full recovery, I have to zfs send to something, preferably that understands tape (yes, I know I can send to tape directly, but how well does zfs send handle the end of the tape? auto-changers?).  
I keep a stream (as a file) of my root pool on a USB stick.  It could be 
on tape, but root pools are small.



Then for individual file recovery, I have snaphots…which I also have to get on 
to tape…if I want to have them available on something other than the boot 
devices.

  
No, just keep the snapshots in place.  If a file is lost, just gab it 
form the snapshot directory.  If the root filesystem is munted, roll 
back to the last snapshot.



Now…to recover the entire OS, perhaps not so bad…but that's one tool.  And to 
recover the one file, say a messed up /etc/system, that's preventing my OS from 
booting?  Have to get that snapshot where I can use it first…oh and restoring 
individual files and not the entire snapshot?

  
As I said, roll back.  Boot form install media, import the root pool, 
get the file from a snapshot, or roll back to the last good snapshot, 
export and reboot.



At best, it's an unwieldy process.  But does it offer the simplicity that 
ufsdump/ufsrestore (or dump/restore on how many Unix variants…) did?  No way.

  
It certainly does for file recovery.  Do you run incremental dumps every 
hour, or every 15 minutes?  Periodic snapshots are quick and cheep.  As 
I said before, careful use of snapshots all be removes the need to 
recover files from tape.  We have 60 days of 4 hourly and 24 hourly 
snapshots in place, so the odds on finding a recent copy of a lost file 
are way better than they would be on daily incrementals.  I certainly 
don't miss the pain of loading a sequence of incrementals to recover 
lost data.


So ZFS solves the problems in a different way.  For fat-finger recovery, 
it's way better than ufsdump/ufsrestore.

A simple, effective dump/restore that deals with all the supported file 
systems, can deal with tape or disk, and allows for complete OS restore or 
individual file restore, and can be run from a install CD/DVD.  As much as I 
love ZFS and as many problems as it does solve, leaving this out was a mistake, 
IMO.
  

It possibly was, but it has encouraged us to find better solutions.

--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Allen Eastwood
On Jan 19, 2010, at 22:54 , Ian Collins wrote:

> Allen Eastwood wrote:
>> On Jan 19, 2010, at 18:48 , Richard Elling wrote:
>> 
>>  
>>> Many people use send/recv or AVS for disaster recovery on the inexpensive
>>> side. Obviously, enterprise backup systems also provide DR capabilities.
>>> Since ZFS has snapshots that actually work, and you can use send/receive
>>> or other backup solutions on snapshots, I assert the "problem" is low 
>>> priority.
>>> 
>>>
>> 
>> 
>> What I have issue with is the idea that no one uses/should use tape any 
>> more.  There are places for tape and it still has value as a backup device.  
>> In many cases in the past, ufsdump, despite it's many issues, was able to 
>> restore working OS's, or individual files.  Perfect, not by a long shot.  
>> But it did get the job done. 
> 
>> As was pointed out earlier, all I needed was a Solaris CD (or network boot) 
>> and I could restore.  Entire OS gone, boot and ufsrestore.  Critical files 
>> deleted, same thing…and I can restore just the file(s) I need.  And while 
>> it's been a few years since I've read the man page on ufsdump, ufsrestore 
>> and fssnap, those tools have proven useful when dealing with a downed system.
>>  
> For a full recovery, you can archive a send stream and receive it back.  With 
> ZFS snapshots, the need for individual file recovery from tape is much 
> reduced.  The backup server I manage for a large client has 60 days of snaps 
> and I can't remember when they had to go to tape to recover a file.
> 
> -- 
> Ian.
> 

Let's see…

For full recovery, I have to zfs send to something, preferably that understands 
tape (yes, I know I can send to tape directly, but how well does zfs send 
handle the end of the tape? auto-changers?).  Then for individual file 
recovery, I have snaphots…which I also have to get on to tape…if I want to have 
them available on something other than the boot devices.

Now…to recover the entire OS, perhaps not so bad…but that's one tool.  And to 
recover the one file, say a messed up /etc/system, that's preventing my OS from 
booting?  Have to get that snapshot where I can use it first…oh and restoring 
individual files and not the entire snapshot?

At best, it's an unwieldy process.  But does it offer the simplicity that 
ufsdump/ufsrestore (or dump/restore on how many Unix variants…) did?  No way.

I suppose it might be fair to say, strictly speaking, that backup/restore 
probably should be dealt with as a "ZFS" issue.  It's kind of ironic that 
Microsoft offers a backup utility and Sun is basically dropping theirs.  It 
might be better to frame the debate somewhere else, but zfs send/receive and 
snapshots are not the same as a built-in OS utility for backing up and 
restoring the OS and OS files.

A simple, effective dump/restore that deals with all the supported file 
systems, can deal with tape or disk, and allows for complete OS restore or 
individual file restore, and can be run from a install CD/DVD.  As much as I 
love ZFS and as many problems as it does solve, leaving this out was a mistake, 
IMO.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Edward Ned Harvey
> Star implements this in a very effective way (by using libfind) that is
> even
> faster that the find(1) implementation from Sun.

Even if I just "find" my filesystem, it will run for 7 hours.  But zfs can
create my whole incremental snapshot in a minute or two.  There is no way
star or any other user-space utility that walks the filesystem can come
remotely close to this performance.  Such performance can only be
implemented at the filesystem level, or lower.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Ian Collins

Allen Eastwood wrote:

On Jan 19, 2010, at 18:48 , Richard Elling wrote:

  

Many people use send/recv or AVS for disaster recovery on the inexpensive
side. Obviously, enterprise backup systems also provide DR capabilities.
Since ZFS has snapshots that actually work, and you can use send/receive
or other backup solutions on snapshots, I assert the "problem" is low priority.





 What I have issue with is the idea that no one uses/should use tape any more.  There are places for tape and it still has value as a backup device.  In many cases in the past, ufsdump, despite it's many issues, was able to restore working OS's, or individual files.  Perfect, not by a long shot.  But it did get the job done. 



As was pointed out earlier, all I needed was a Solaris CD (or network boot) and 
I could restore.  Entire OS gone, boot and ufsrestore.  Critical files deleted, 
same thing…and I can restore just the file(s) I need.  And while it's been a 
few years since I've read the man page on ufsdump, ufsrestore and fssnap, those 
tools have proven useful when dealing with a downed system.
  
For a full recovery, you can archive a send stream and receive it back.  
With ZFS snapshots, the need for individual file recovery from tape is 
much reduced.  The backup server I manage for a large client has 60 days 
of snaps and I can't remember when they had to go to tape to recover a file.


--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Ian Collins

Joerg Schilling wrote:

Ian Collins  wrote:

  
Sun's tar also writes ACLs in a way that is 100% non-portable. Star cannot 
understand them and probably never will be able to understand this format as it

is not well defined for a portable program like star.

  
  
Is that because they are NFSv4 ACLs?  tar uses the formatting routines 
from libacl to archive them.



The correct way to archivbe ACLs would be to put them into extended POSIX tar
attrubutes as star does.

See http://cdrecord.berlios.de/private/man/star/star.4.html for a format 
documentation or have a look at ftp://ftp.berlios.de/pub/star/alpha, e.g.

ftp://ftp.berlios.de/pub/star/alpha/acl-test.tar.gz

The ACL format used by Sun is undocumented..

  

man acltotext

--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Mike La Spina
I use zfs send/recv in the enterprise and in smaller environments all time and 
it's is excellent.

Have a look at how awesome the functionally is in this example.

http://blog.laspina.ca/ubiquitous/provisioning_disaster_recovery_with_zfs

Regards,

Mike
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Allen Eastwood

On Jan 19, 2010, at 18:48 , Richard Elling wrote:

> 
> Many people use send/recv or AVS for disaster recovery on the inexpensive
> side. Obviously, enterprise backup systems also provide DR capabilities.
> Since ZFS has snapshots that actually work, and you can use send/receive
> or other backup solutions on snapshots, I assert the "problem" is low 
> priority.
> 


 What I have issue with is the idea that no one uses/should use tape any more.  
There are places for tape and it still has value as a backup device.  In many 
cases in the past, ufsdump, despite it's many issues, was able to restore 
working OS's, or individual files.  Perfect, not by a long shot.  But it did 
get the job done. 

As was pointed out earlier, all I needed was a Solaris CD (or network boot) and 
I could restore.  Entire OS gone, boot and ufsrestore.  Critical files deleted, 
same thing…and I can restore just the file(s) I need.  And while it's been a 
few years since I've read the man page on ufsdump, ufsrestore and fssnap, those 
tools have proven useful when dealing with a downed system.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Richard Elling

On Jan 19, 2010, at 4:26 PM, Allen Eastwood wrote:

>> Message: 3
>> Date: Tue, 19 Jan 2010 15:48:52 -0500
>> From: Miles Nordin 
>> To: zfs-discuss@opensolaris.org
>> Subject: Re: [zfs-discuss] zfs send/receive as backup - reliability?
>> Message-ID: 
>> Content-Type: text/plain; charset="us-ascii"
>> 
>> I don't think a replacement for the ufsdump/ufsrestore tool is really
>> needed.  From now on, backups just go into Another Zpool.
>> 
>> We need the zfs send stream format commitment (stream format depends
>> only on zfs version, not on zpool version or kernel version), but
>> that's enough.
>> 
>> 
>> If people are really still backing up to tapes or DVD's, just use file
>> vdev's, export the pool, and then copy the unmounted vdev onto the
>> tape or DVD.  The requirement that your backup be staged on a disk
>> isn't an obstacle in the backup direction: Amanda already has this
>> requirement.  Amanda, when I used it, did *not* have this requirement
>> in the restore direction so one would have to keep that in mind: if
>> using tapes, he needs a scratch disk as big as the biggest tape file
>> on the DR site or the development environment or Compliance Extractor
>> Station or wherever the restore is happening.
> 
> 
> While this idea may be fine for a home user…those us of who have customers in 
> the enterprise still have a need for tape backups.
> 
> And some of those enterprises require backup mechanism that can be easily 
> used in a DR situation.
> 
> ufsdump/restore was perfect in that regard.  The lack of equivalent 
> functionality is a big problem for the situations where this functionality is 
> a business requirement.

How quickly we forget ufsdump's limitations :-).  For example, it is not 
supported
for use on an active file system (known data corruption possibility) and 
UFS snapshots are, well, a poor hack and often not usable for backups.
As the ufsdump(1m) manpage says,
 The ufsdump command can only be used on unmounted file  sys-
 tems,  or  those  mounted  read-only.  Attempting  to dump a
 mounted, read-write file system might  result  in  a  system
 disruption  or the inability to restore files from the dump.
 Consider using the fssnap(1M) command to create a file  sys-
 tem  snapshot  if  you  need a point-in-time image of a file
 system that is mounted.

So, if you are taking ufsdumps of an active file system for business 
requirements,
then perhaps you should rethink the solution.

> For example, one customer, local government, requires a backup that can be 
> taken offsite and used in a DR situation.  They have 2 sun servers.  They 
> have very little Solaris knowledge. So whatever I provide them has to be easy 
> and easily documented.  Lack of "zfsdump/restore" means I either have to 
> forgo moving them to ZFS root, or have to add in a third party 
> application…like I'm really going to meet their requirements with Amanda or 
> Backula…

Many people use send/recv or AVS for disaster recovery on the inexpensive
side. Obviously, enterprise backup systems also provide DR capabilities.
Since ZFS has snapshots that actually work, and you can use send/receive
or other backup solutions on snapshots, I assert the "problem" is low priority.

> This lack in Solaris is just plain ludicrous to at least some parts of the 
> business/enterprise world.

Methinks you never read the ufsdump man page? :-)
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Allen Eastwood
> Message: 3
> Date: Tue, 19 Jan 2010 15:48:52 -0500
> From: Miles Nordin 
> To: zfs-discuss@opensolaris.org
> Subject: Re: [zfs-discuss] zfs send/receive as backup - reliability?
> Message-ID: 
> Content-Type: text/plain; charset="us-ascii"
> 
> I don't think a replacement for the ufsdump/ufsrestore tool is really
> needed.  From now on, backups just go into Another Zpool.
> 
> We need the zfs send stream format commitment (stream format depends
> only on zfs version, not on zpool version or kernel version), but
> that's enough.
> 
> 
> If people are really still backing up to tapes or DVD's, just use file
> vdev's, export the pool, and then copy the unmounted vdev onto the
> tape or DVD.  The requirement that your backup be staged on a disk
> isn't an obstacle in the backup direction: Amanda already has this
> requirement.  Amanda, when I used it, did *not* have this requirement
> in the restore direction so one would have to keep that in mind: if
> using tapes, he needs a scratch disk as big as the biggest tape file
> on the DR site or the development environment or Compliance Extractor
> Station or wherever the restore is happening.


While this idea may be fine for a home user…those us of who have customers in 
the enterprise still have a need for tape backups.

And some of those enterprises require backup mechanism that can be easily used 
in a DR situation.

ufsdump/restore was perfect in that regard.  The lack of equivalent 
functionality is a big problem for the situations where this functionality is a 
business requirement.

For example, one customer, local government, requires a backup that can be 
taken offsite and used in a DR situation.  They have 2 sun servers.  They have 
very little Solaris knowledge. So whatever I provide them has to be easy and 
easily documented.  Lack of "zfsdump/restore" means I either have to forgo 
moving them to ZFS root, or have to add in a third party application…like I'm 
really going to meet their requirements with Amanda or Backula…

This lack in Solaris is just plain ludicrous to at least some parts of the 
business/enterprise world.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Daniel Carosone
There is a tendency to conflate "backup" and "archive", both generally
and in this thread. They have different requirements.  

Backups should enable quick restore of a full operating image with all
the necessary system level attributes. They concerned with SLA and
uptime and outage and data loss window criteria. 

Archives are usually much more concerned with application data, which
may have been specially prepared for the future user of the
archive. Portability is often more important here, both in archive
format but also in data schema within, even at the expense of some
kinds of technical integrity or completeness (e.g. ACLs).

Adding regulatory requirements into the mix further complicates matters.

If one solution can address the union of all requirements, then you're
in luck, but that's not always the case.  Sometimes the best way to
resolve the tension is to save the data differently for different
purposes. 

--
Dan.

pgpfsQPyUEOVE.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Daniel Carosone
On Tue, Jan 19, 2010 at 12:16:01PM +0100, Joerg Schilling wrote:
> Daniel Carosone  wrote:
> 
> > I also don't recommend files >1Gb in size for DVD media, due to
> > iso9660 limitations.  I haven't used UDF enough to say much about any
> > limitations there.
> 
> ISO-9660 supports files up to 8 TB. 

However, some implementations have (or had?) limitations which meant
that they couldn't read such disks.  Some of those platforms even had
problems (or required remembering arcane options and flags for "large
file support") when copying these files back to hard disk.  

Linux was a prime culprit, though not the only one.  I didn't use
linux, but I never knew what might be at hand when needing to read
disks for recovery, so I wanted them to be portable.  I envisaged
using multiple machines to read more disks in parallel, for example.

It may not be as relevant for this use case, nor perhaps for more
modern platforms, but it was a habit I developed earlier for other
kinds of archives. 

I apologise, though, for misattributing this limitation to the
filesystem spec.  I misremembered the reason why I chose the limit.
It was a practical limit, and the practicalities may well have changed
since.

> Do you have a bigger pool?

I certainly don't have a bigger DVD media :-)

(Yes, I know iso9660 also has multivolume support, but I'm similarly
wary of being able to read it back on all platforms, since it's not
encountered often, and there's no need when the source files can be
split to better effect.)

--
Dan.


pgpqa8vKafJ1t.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Miles Nordin
> "ic" == Ian Collins  writes:

 >> I know tape backup isn't sexy, but it's a reality for many of
 >> us and it's not going away anytime soon.

ic> True, but I wonder how viable its future is.  One of my
ic> clients requires 17 LT04 types for a full backup, which cost
ic> more and takes up more space than the

but Ian, aiui SOX requires backups onto a write-once non-eraseable
medium, because of all the discovery shennanigans and
claimed-incompetence around the time of the Worldcomm and Enron
bankruptcy near the start of Bush II.  The law in practice is sort of
like a tiger that chases the herd and eats the sick animal that's
lagging behind, and currently I think the herd is using ``WORM tape''
which is a tape that's made to appear to be append-only by the tape
drive's firmware.  These tapes exist now and must be retained for
seven years, so even if you invented a newfangled gadget meeting the
WORM requirements TODAY so that you can stay in the middle of the herd
without using tape, you still will not get rid of tape for seven
years.


pgp2HKJuL5cLI.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Miles Nordin
> "jk" == Jerry K  writes:

jk> +1 
jk> for zfsdump/zfsrestore

-1

I don't think a replacement for the ufsdump/ufsrestore tool is really
needed.  From now on, backups just go into Another Zpool.

We need the zfs send stream format commitment (stream format depends
only on zfs version, not on zpool version or kernel version), but
that's enough.


If people are really still backing up to tapes or DVD's, just use file
vdev's, export the pool, and then copy the unmounted vdev onto the
tape or DVD.  The requirement that your backup be staged on a disk
isn't an obstacle in the backup direction: Amanda already has this
requirement.  Amanda, when I used it, did *not* have this requirement
in the restore direction so one would have to keep that in mind: if
using tapes, he needs a scratch disk as big as the biggest tape file
on the DR site or the development environment or Compliance Extractor
Station or wherever the restore is happening.

File vdev's have all the wanted characteristics: bitflip resiliency,
checksums, ability to represent all the opaque mysteryfeatures of
inner ZFS's, snapshot/clone efficiency, and importability by future
ZFS kernels ~forever.  It still has the all-or-nothing-restore
downside, but maybe we can live with that.  Those who can't might
stick to spinning-disk backups.

It might be nice to have a ``read-only vdev'' like a mac os compressed
disk image that can occupy the same pool next to a read-write
vdev---this would be neat for livecd's and incrementals---but it's a
neat feature, not something blocking a frequent/unavoidable sysadmin
task.

What's needed IMHO is non-ZFS-specific tools like tar/pax/cpio/rsync
that understand all this new ACL and opaque-fork garbage that
filesystems are growing (and not just Sun's either).  It's needed to
copy between NFSv4 and ZFS, and also to split/combine subdirectories
into separate/common ZFS filesystems without losing any of the
newfangled mysterybaggage.  It has as much to do with shaking corner
cases and awkwardnesses out of the newfangled API's as it does with
actually accomplishing a sysadmin task: the best documentation for the
weird appendages in various modern Unix filesystems would probably be
the tool itself, if we had it.  

Without this kind of tool and the API shakeout that making it would
require, our systems are slowly turning into shitty Windows boxes that
need some black magic proprietary tool like Ghost or PQDI to capture
all the silly out-of-band garbage their installations depend on and/or
accumulate.


pgpnDeqTYeOuV.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Joerg Schilling
Ian Collins  wrote:

> > Sun's tar also writes ACLs in a way that is 100% non-portable. Star cannot 
> > understand them and probably never will be able to understand this format 
> > as it
> > is not well defined for a portable program like star.
> >
> >   
> Is that because they are NFSv4 ACLs?  tar uses the formatting routines 
> from libacl to archive them.

The correct way to archivbe ACLs would be to put them into extended POSIX tar
attrubutes as star does.

See http://cdrecord.berlios.de/private/man/star/star.4.html for a format 
documentation or have a look at ftp://ftp.berlios.de/pub/star/alpha, e.g.
ftp://ftp.berlios.de/pub/star/alpha/acl-test.tar.gz

The ACL format used by Sun is undocumented..


Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Ian Collins

Joerg Schilling wrote:

Julian Regel  wrote:

  

When I look at the documentation for Zmanda 
(http://docs.zmanda.com/Project:Amanda_Enterprise_3.0/ZMC_Users_Manual/Appendix_A:_Backing_Up_and_Restoring_Solaris_ZFS),
 it states that the command used to backup a ZFS filesystem depends on whether 
backing up ACLs are required (the default is not to(!)). If backing up ACLs are 
required - which they are for us - then the native /usr/bin/tar command is 
used. Given that /usr/bin/tar doesn't handle sparse files correctly and updates 
atime when creating archives, it's not possible to create a complete copy of a 
filesystem without making intrusive changes to the structure of the data and/or 
metadata.

So it's arguable that ufsdump was a significant improvement over tar, and that 
the current archiving solutions for ZFS are actually a step backwards.



From my perspective, Sun really need to create a zfsdump/zfsrestore set of 
commands that perform block level backups of a specified filesystem (or 
snapshot) and that preserves ACLs, atime, sparse files, handles multiple tapes 
(many of us still use and rely on tape!) etc.
  


Sun's tar is not able to archive files > 8 GB in a way that can be read on any 
other OS unless you use star. This is because Sun's tar does not implement 
support for the POSIX.1-2001 extended tar format.


Sun's tar also writes ACLs in a way that is 100% non-portable. Star cannot 
understand them and probably never will be able to understand this format as it

is not well defined for a portable program like star.

  
Is that because they are NFSv4 ACLs?  tar uses the formatting routines 
from libacl to archive them.


--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Bob Friesenhahn

On Wed, 20 Jan 2010, Ian Collins wrote:

Commercial backup solutions are available for ZFS.
I know tape backup isn't sexy, but it's a reality for many of us and it's 
not going away anytime soon.


True, but I wonder how viable its future is.  One of my clients requires 17 
LT04 types for a full backup, which cost more and takes up more space than 
the equivalent in removable hard drives.


In the past few years growth in hard drive capacities has outstripped tapes 
to the extent that removable hard drives and ZFS snapshots have become a more 
cost effective and convenient backup media.


In all of these discussions, people seem to forget some very important 
criteria.  That important criteria is the time required for a full 
recovery from backup.


If the company is not able to survive more than a week with total 
disruption of business, but the full restore will take three weeks, 
then the backup has been rendered 100% ineffective.


When planning for recovery from disaster, the time to recover is 
exceedingly important.


When using the backup to filesystem approach, that backup could be 
done to a remote mirrored system (filesystems + hardware) which is 
sufficiently distant to protect against any local disaster.  If there 
is a disaster in the local data center, that system could be 
immediately put on line (assuming adequate connectivity), or that 
system could be loaded on a truck for overnight delivery as a 
replacement to the data center.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Ian Collins

Joerg Schilling wrote:

Ian Collins  wrote:

  

Julian Regel wrote:

Based on what I've seen in other comments, you might be right. 
Unfortunately, I don't feel comfortable backing up ZFS filesystems 
because the tools aren't there to do it (built into the operating 
system or using Zmanda/Amanda).


  

Commercial backup solutions are available for ZFS.



On what archiving programs are these solutions based on?

  
I'm sure they use their own formats.  The only one I have crossed paths 
with is NetVault.


--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Joerg Schilling
Ian Collins  wrote:

> Julian Regel wrote:
> >
> > Based on what I've seen in other comments, you might be right. 
> > Unfortunately, I don't feel comfortable backing up ZFS filesystems 
> > because the tools aren't there to do it (built into the operating 
> > system or using Zmanda/Amanda).
> >
> Commercial backup solutions are available for ZFS.

On what archiving programs are these solutions based on?

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Joerg Schilling
Julian Regel  wrote:

> When I look at the documentation for Zmanda 
> (http://docs.zmanda.com/Project:Amanda_Enterprise_3.0/ZMC_Users_Manual/Appendix_A:_Backing_Up_and_Restoring_Solaris_ZFS),
>  it states that the command used to backup a ZFS filesystem depends on 
> whether backing up ACLs are required (the default is not to(!)). If backing 
> up ACLs are required - which they are for us - then the native /usr/bin/tar 
> command is used. Given that /usr/bin/tar doesn't handle sparse files 
> correctly and updates atime when creating archives, it's not possible to 
> create a complete copy of a filesystem without making intrusive changes to 
> the structure of the data and/or metadata.
>
> So it's arguable that ufsdump was a significant improvement over tar, and 
> that the current archiving solutions for ZFS are actually a step backwards.
>
> > From my perspective, Sun really need to create a zfsdump/zfsrestore set of 
> > commands that perform block level backups of a specified filesystem (or 
> > snapshot) and that preserves ACLs, atime, sparse files, handles multiple 
> > tapes (many of us still use and rely on tape!) etc.

Sun's tar is not able to archive files > 8 GB in a way that can be read on any 
other OS unless you use star. This is because Sun's tar does not implement 
support for the POSIX.1-2001 extended tar format.

Sun's tar also writes ACLs in a way that is 100% non-portable. Star cannot 
understand them and probably never will be able to understand this format as it
is not well defined for a portable program like star.

Star supports to do backups without affecting atime (on Solaris) since 15 years 
already and star supports the withdrawn POSIX draft ACLs in a OS independent 
way. This allows to use star for platform independent ACL support in case you 
are using UFS on Solaris.

Star will in future support NTFS ACLs in a OS independent way. If someone likes 
to contribute, he is of course welcome. As I am currently working on 
cdrtools-3.0-final, star is currently on maintenance only.

What we need for the future is a OS/FS independent program that implements the
needed features.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Ian Collins

Julian Regel wrote:


Based on what I've seen in other comments, you might be right. 
Unfortunately, I don't feel comfortable backing up ZFS filesystems 
because the tools aren't there to do it (built into the operating 
system or using Zmanda/Amanda).



Commercial backup solutions are available for ZFS.
I know tape backup isn't sexy, but it's a reality for many of us and 
it's not going away anytime soon.


True, but I wonder how viable its future is.  One of my clients requires 
17 LT04 types for a full backup, which cost more and takes up more space 
than the equivalent in removable hard drives.


In the past few years growth in hard drive capacities has outstripped 
tapes to the extent that removable hard drives and ZFS snapshots have 
become a more cost effective and convenient backup media.


What do people with many tens of TB use for backup these days?

--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Julian Regel


>> The beauty of ufsdump/ufsrestore is that because it's bundled with the 
>> operating system, I can perform bare metal recovery using a Solaris DVD and 
>> locally attached tape drive. It's simple and arguably essential for system 
>> administrators.

> Yep. And it was invented because there was no option other than tar
> at the time. Today, there are many very comprehensive backup solutions
> on the market including open source. Some are nicely integrated, such 
> as NexentaStor and Zmanda.

When I look at the documentation for Zmanda 
(http://docs.zmanda.com/Project:Amanda_Enterprise_3.0/ZMC_Users_Manual/Appendix_A:_Backing_Up_and_Restoring_Solaris_ZFS),
 it states that the command used to backup a ZFS filesystem depends on whether 
backing up ACLs are required (the default is not to(!)). If backing up ACLs are 
required - which they are for us - then the native /usr/bin/tar command is 
used. Given that /usr/bin/tar doesn't handle sparse files correctly and updates 
atime when creating archives, it's not possible to create a complete copy of a 
filesystem without making intrusive changes to the structure of the data and/or 
metadata.

So it's arguable that ufsdump was a significant improvement over tar, and that 
the current archiving solutions for ZFS are actually a step backwards.

> From my perspective, Sun really need to create a zfsdump/zfsrestore set of 
> commands that perform block level backups of a specified filesystem (or 
> snapshot) and that preserves ACLs, atime, sparse files, handles multiple 
> tapes (many of us still use and rely on tape!) etc.

> In my crystal ball, I see a divergence away from traditional backup 
> solution approaches. Today you can feel comfortable with backing up
> ZFS file systems because they provide a POSIX file system interface.

Based on what I've seen in other comments, you might be right. Unfortunately, I 
don't feel comfortable backing up ZFS filesystems because the tools aren't 
there to do it (built into the operating system or using Zmanda/Amanda).

I know tape backup isn't sexy, but it's a reality for many of us and it's not 
going away anytime soon.

JR


  ___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Tim Cook
On Tue, Jan 19, 2010 at 11:36 AM, Richard Elling
wrote:

> On Jan 19, 2010, at 1:53 AM, Julian Regel wrote:
> > > When we brought it up last time, I think we found no one knows of a
> > > userland tool similar to 'ufsdump' that's capable of serializing a ZFS
> > > along with holes, large files, ``attribute'' forks, windows ACL's, and
> > > checksums of its own, and then restoring the stream in a
> > > filesystem-agnostic read/write/lseek/... manner like 'ufsrestore'.
> >
> > This has been something I've been working on for the last few days and
> I've not yet found an adequate solution. I've looked at tar, cpio and pax as
> potential tools for backing up a ZFS filesystem to tape, but each has
> limitations. As of today, Sun do not provide the ability to reliably backup
> a Solaris installation without resorting to third party tools. This is
> crazy!
> >
> > The beauty of ufsdump/ufsrestore is that because it's bundled with the
> operating system, I can perform bare metal recovery using a Solaris DVD and
> locally attached tape drive. It's simple and arguably essential for system
> administrators.
>
> Yep. And it was invented because there was no option other than tar
> at the time. Today, there are many very comprehensive backup solutions
> on the market including open source. Some are nicely integrated, such
> as NexentaStor and Zmanda.
>
> > From my perspective, Sun really need to create a zfsdump/zfsrestore set
> of commands that perform block level backups of a specified filesystem (or
> snapshot) and that preserves ACLs, atime, sparse files, handles multiple
> tapes (many of us still use and rely on tape!) etc.
>
> In my crystal ball, I see a divergence away from traditional backup
> solution approaches. Today you can feel comfortable with backing up
> ZFS file systems because they provide a POSIX file system interface.
> Other datasets don't resemble POSIX file systems at all, and I see several
> different types of datasets with the opportunity to become popular with
> the basic ZVol getting a lot of attention lately. The win will go to the
> vendor who can provide the best value for the various datasets in the ZFS
> environment.
>
> > Does anyone know the process to formally request this (we have a support
> contract), or would I be wasting my time?
>
> You can pile onto CR 5004379, want comprehensive backup strategy
> http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=5004379
>
> I can't answer the latter, but judging by the date of the CR, I wouldn't
> hold my
> breath. Give the fine folks at Zmanda a look.
>
> Also, the ADM project seems to be dead. Unfortunate?
> http://hub.opensolaris.org/bin/view/Project+adm/WhatisADM
>  -- richard
>
>

I'm likely missing something, so please, fill me in.  Aren't they simply
asking for the functionality already provided by NDMP?

-- 
--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Richard Elling
On Jan 19, 2010, at 1:53 AM, Julian Regel wrote:
> > When we brought it up last time, I think we found no one knows of a
> > userland tool similar to 'ufsdump' that's capable of serializing a ZFS
> > along with holes, large files, ``attribute'' forks, windows ACL's, and
> > checksums of its own, and then restoring the stream in a
> > filesystem-agnostic read/write/lseek/... manner like 'ufsrestore'.
> 
> This has been something I've been working on for the last few days and I've 
> not yet found an adequate solution. I've looked at tar, cpio and pax as 
> potential tools for backing up a ZFS filesystem to tape, but each has 
> limitations. As of today, Sun do not provide the ability to reliably backup a 
> Solaris installation without resorting to third party tools. This is crazy!
> 
> The beauty of ufsdump/ufsrestore is that because it's bundled with the 
> operating system, I can perform bare metal recovery using a Solaris DVD and 
> locally attached tape drive. It's simple and arguably essential for system 
> administrators.

Yep. And it was invented because there was no option other than tar
at the time. Today, there are many very comprehensive backup solutions
on the market including open source. Some are nicely integrated, such 
as NexentaStor and Zmanda.

> From my perspective, Sun really need to create a zfsdump/zfsrestore set of 
> commands that perform block level backups of a specified filesystem (or 
> snapshot) and that preserves ACLs, atime, sparse files, handles multiple 
> tapes (many of us still use and rely on tape!) etc.

In my crystal ball, I see a divergence away from traditional backup 
solution approaches. Today you can feel comfortable with backing up
ZFS file systems because they provide a POSIX file system interface.
Other datasets don't resemble POSIX file systems at all, and I see several
different types of datasets with the opportunity to become popular with
the basic ZVol getting a lot of attention lately. The win will go to the 
vendor who can provide the best value for the various datasets in the ZFS
environment.

> Does anyone know the process to formally request this (we have a support 
> contract), or would I be wasting my time?

You can pile onto CR 5004379, want comprehensive backup strategy
http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=5004379

I can't answer the latter, but judging by the date of the CR, I wouldn't hold my
breath. Give the fine folks at Zmanda a look.

Also, the ADM project seems to be dead. Unfortunate?
http://hub.opensolaris.org/bin/view/Project+adm/WhatisADM
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Jerry K

+1

for zfsdump/zfsrestore



Julian Regel wrote:



When we brought it up last time, I think we found no one knows of a
userland tool similar to 'ufsdump' that's capable of serializing a ZFS
along with holes, large files, ``attribute'' forks, windows ACL's, and
checksums of its own, and then restoring the stream in a
filesystem-agnostic read/write/lseek/... manner like 'ufsrestore'.


This has been something I've been working on for the last few days and I've not 
yet found an adequate solution. I've looked at tar, cpio and pax as potential 
tools for backing up a ZFS filesystem to tape, but each has limitations. As of 
today, Sun do not provide the ability to reliably backup a Solaris installation 
without resorting to third party tools. This is crazy!

The beauty of ufsdump/ufsrestore is that because it's bundled with the 
operating system, I can perform bare metal recovery using a Solaris DVD and 
locally attached tape drive. It's simple and arguably essential for system 
administrators.

From my perspective, Sun really need to create a zfsdump/zfsrestore set of 
commands that perform block level backups of a specified filesystem (or 
snapshot) and that preserves ACLs, atime, sparse files, handles multiple tapes 
(many of us still use and rely on tape!) etc.

Does anyone know the process to formally request this (we have a support 
contract), or would I be wasting my time?

Thanks

JR

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



  





___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Joerg Schilling
Daniel Carosone  wrote:

> I also don't recommend files >1Gb in size for DVD media, due to
> iso9660 limitations.  I haven't used UDF enough to say much about any
> limitations there.

ISO-9660 supports files up to 8 TB. Do you have a bigger pool?

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Joerg Schilling
Richard Elling  wrote:

> OOB, the default OpenSolaris PATH places /usr/gnu/bin ahead
> of /usr/bin, so gnu tar is "the default."  As of b130 (I'm not running
> an older build currently) the included gnu tar is version 1.22 which
> is the latest as released March 2009 at http://www.gnu.org/software/tar/

This causes Indiana to by default offer tar implementation that creates archives
that are in conflict with the POSIX standard.

Also note that GNU tar is unable to be used as a backup utility:

-   It does neither support ACLs not any other file attributes.

-   It failes even with simples modifications in the original filesystem
and people will become aware of this problem at the time when
they try to restore a backup that will not work in many cases.


> That is my understanding as well.  It seems that the backup vendors
> are moving forward in a more-or-less vendor-specific way. This is
> not necessarily a bad thing, since there are open source solutions.
> But I see the requirements for backups being much more sophisticated
> than ufsdump was 25 years ago.  hmmm... has ufsdump changed over 
> the past 25 years? ;-)

ufsdump did (slightly) change as it now allows you to specify a subdirectory
instead of the whole filesystem.

But what features are you missing?

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Joerg Schilling
Miles Nordin  wrote:

> When we brought it up last time, I think we found no one knows of a
> userland tool similar to 'ufsdump' that's capable of serializing a ZFS
> along with holes, large files, ``attribute'' forks, windows ACL's, and
> checksums of its own, and then restoring the stream in a
> filesystem-agnostic read/write/lseek/... manner like 'ufsrestore'.

Star is such a tool. It privides all the features known from ufsdump
but it is OS and filesystem indepentent. 

Once Sun shows interest in star, there will of course be also support 
for NTFS-ACLs.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Joerg Schilling
Lassi Tuura  wrote:

> I guess what I am after is, for data which really matters to its owners and 
> which they actually had to recover, did people use tar/pax archives (~ file 
> level standard archive format), dump/restore (~ semi-standard format based on 
> files/inodes) or zfs send/receive (~ file system block level dump), or 
> something else, and why? (Regardless of how these are implemented, hobby 
> scripts or enterprise tools, how they dealt with possible media failure 
> issues, etc.)

star combines the advantages from ufsdump/ufsrestore (true incremental dumps)
with the advantages of a POSIX standard achive format.

Note that star is even at least 30% faster that ufsdump (although ufsdump
reads the raw disk device while star uses the official OS filesystem interface).

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Joerg Schilling
Edward Ned Harvey  wrote:

> > I still believe that a set of compressed incremental star archives give
> > you
> > more features.
>
> Big difference there is that in order to create an incremental star archive,
> star has to walk the whole filesystem or folder that's getting backed up,
> and do a "stat" on every file to see which files have changed since the last
> backup run.  If you have a large filesystem, that can take a very long time.

Star implements this in a very effective way (by using libfind) that is even 
faster that the find(1) implementation from Sun.

The big advantage with star is that you get OS and filesystem independent
archives that are compatible with recent POSIX standards and thus can be read 
on any POSIX.1-2001 compliant OS even if you don't have a star binary handy.

> I run incremental zfs send & receive, and it completes typically in a minute
> or two.

Did you make a test with star?

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Joerg Schilling
Thomas Burgess  wrote:

> so star is better than tar?

Star is the oldest OSS tar implementation (it started in 1982).
Star is (in contrary to Sun's tar and to GNU tar) able to create
archives with attributes from recent POSIX standards and it implements
aprox. twice as many features than GNU tar without forcing you to learn
twice as many options ;-)

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-19 Thread Julian Regel


> When we brought it up last time, I think we found no one knows of a
> userland tool similar to 'ufsdump' that's capable of serializing a ZFS
> along with holes, large files, ``attribute'' forks, windows ACL's, and
> checksums of its own, and then restoring the stream in a
> filesystem-agnostic read/write/lseek/... manner like 'ufsrestore'.

This has been something I've been working on for the last few days and I've not 
yet found an adequate solution. I've looked at tar, cpio and pax as potential 
tools for backing up a ZFS filesystem to tape, but each has limitations. As of 
today, Sun do not provide the ability to reliably backup a Solaris installation 
without resorting to third party tools. This is crazy!

The beauty of ufsdump/ufsrestore is that because it's bundled with the 
operating system, I can perform bare metal recovery using a Solaris DVD and 
locally attached tape drive. It's simple and arguably essential for system 
administrators.

From my perspective, Sun really need to create a zfsdump/zfsrestore set of 
commands that perform block level backups of a specified filesystem (or 
snapshot) and that preserves ACLs, atime, sparse files, handles multiple tapes 
(many of us still use and rely on tape!) etc.

Does anyone know the process to formally request this (we have a support 
contract), or would I be wasting my time?

Thanks

JR

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



  ___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Daniel Carosone
On Mon, Jan 18, 2010 at 01:38:16PM -0800, Richard Elling wrote:
> The Solaris 10 10/09 zfs(1m) man page says:
> 
>  The format of the stream is committed. You will be  able
>  to receive your streams on future versions of ZFS.
> 
> I'm not sure when that hit snv, but obviously it was after b112.
> ...

I only noticed it recently.  My guess was that it arrived together
with the format changes for "zfs send -D".   As handy as that change
is, it doesn't address some of the other deficiencies involved in
trying to repurpose a replication stream as an archive format.

> > When we brought it up last time, I think we found no one knows of a
> > userland tool similar to 'ufsdump' that's capable of serializing a ZFS
> > along with holes, large files, ``attribute'' forks, windows ACL's, and
> > checksums of its own, and then restoring the stream in a
> > filesystem-agnostic read/write/lseek/... manner like 'ufsrestore'.

It was precisely the realisation that the zpool-in-a-file format had
all of the desired characteristics that led to the scheme I described
elsewhere in this thread.  

(It wasn't one of my criteria, but this goes up to and including the
"userland tool" component - although I've not tried, I understand
there are userland versions of the zfs tools in the test suite.)

It has the advantage of being entirely common with the original, so it
will be well tested and keep in sync with new features (dedup, crypto,
next+N).   A focus on formalising this usage pattern would bring
further benefits in terms of features  (e.g, as a worthwhile use case
for "read only import") and of practices (documentation and process).

--
Dan.



pgpmzgIoI2655.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Robert Milkowski

On 18/01/2010 18:28, Lassi Tuura wrote:

Hi,

   

Here is the big difference. For a professional backup people still typically
use tapes although tapes have become expensive.

I still believe that a set of compressed incremental star archives give you
more features.
 

Thanks for your comments!

I think I roughly understand the feature trade-offs, and have some experience 
with back-up solutions ranging from simple to enterprise.

I guess what I am after is, for data which really matters to its owners and 
which they actually had to recover, did people use tar/pax archives (~ file 
level standard archive format), dump/restore (~ semi-standard format based on 
files/inodes) or zfs send/receive (~ file system block level dump), or 
something else, and why? (Regardless of how these are implemented, hobby 
scripts or enterprise tools, how they dealt with possible media failure issues, 
etc.)

Other than the file system vs. file restore, is there any concern in doing the block level thing? 
"Concerns" as in "mind being in the line of fire if it fails?" :-)

   


What we are doing basically is:

1. incremental rsync from a client to a dedicated filesystem for that client
2. snapshot after rsync finished
3. go to #1 for next backup

a pool is a dynamic stripe across raidz-2 groups + hot spares.
Then selected clients along with all their backups (snapshots) are 
replicated to another device which is exactly the same hardware/software 
configuration.


Add to it a management of snapshots (retention policies), reporting, 
etc. and you have a pretty good backup solution which allows you to 
restore a single file or entire filesystem with a very easy access to 
any backup. And it works for different OSes as clients.


See more details at
http://milek.blogspot.com/2009/12/my-presentation-at-losug.html

--
Robert Milkowski
http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Daniel Carosone
On Mon, Jan 18, 2010 at 07:34:51PM +0100, Lassi Tuura wrote:
> > Consider then, using a zpool-in-a-file as the file format, rather than
> > zfs send streams.
> 
> This is an interesting suggestion :-)
> 
> Did I understand you correctly that once a slice is written, zfs
> won't rewrite it? In other words, I can have an infinitely growing
> pile of these slices, and once zfs fills one file up (or one
> "raidz2" set of files), I flush it to tape/optical disk/whatever,
> and zfs won't change it "ever after"? When I need more space, I just
> add more slices, but old slices are effectively read-only?

No.  The intention of this scheme is to have an ongoing backup pool,
with a rolling set of historical snapshots collected together.  I
wouldn't suggest growing the pool indefinately, because you'll need an
indefinite number of pool backing files online when it comes time for
a distant-future restore.

Even if you were never to delete old snapshots, zfs will still
update some of the metadata in each file, or raidz set of files,
whenever you make changes to your backup pool.  You will need to sync
each of these files to your backup media or offsite storage, to have a
consistent pool to restore from.  

However, you can add new top-level vdevs (e.g. raidz set of files) as
your backup pool starts to fill.  Until you start to recycle space (by
deleting old snapshots from the backup pool) there won't be much new
data written to the original full files.  Therefore, if your offsite
replication of these files is efficient for incremental changes
(e.g. rsync) then they will update quickly.

The same will happen for changes within the existing files, of course,
if they're not yet full or if you are recycling space.  It's the same
data, all that changes is how it's spread among the files.

If you're writing to tapes or dvd's, you'll either need to rewrite the
files completely every time, or layer some *other* incremental
mechanism on top (perhaps your existing large-scale enterprise VTL,
which you'd like to continue using, or are forced to continue using :).

I don't mind rewriting tape media every time.. keep 2 or three sets
and just roll between them each time.  There are advantages to
having a predictable cycle - knowing that you're going to need exactly
N tapes this week and they will take T time to write out.  You also
get the benefit of all other D-to-D-to-T schemes, that this can be
done at leisure asynchronously to host backup windows. For DVD's its a
little more annoying, but at least the media is cheap. 

For these purposes, you can also consider removable hard disks as
tapes.  As I replace one generation of drives with the next, higher
capacity, I intend for the previous ones can move to backup service. 

> And perhaps most importantly: has anyone actually done this for
> their back-ups and has success stories on restore after media
> failure? 

For me it's somewhere between concept and regular practice. It works,
I've tinkered with it and done intermittent runs at archiving off the
pool files and test imports and restores, but not written automation
or made it as regular practice as I should.  I've used it to drop
backups of my opensolaris hosts onto enterprise backed-up servers, in 
non-solaris customer environments.  

There are some wrinkles, like you can't mount zpool files off
read-only DVD media directly - you have to copy them to r/w scratch
space because import wants to update metadata (is there an RFE for
read-only import? I forget).  This is mildly irritating at worst -
copying one or two dvd's worth of files is easy, yet any more than
this in your backup pool, you'll need to copy anyway for lack of many
dvd readers at once. 

I also don't recommend files >1Gb in size for DVD media, due to
iso9660 limitations.  I haven't used UDF enough to say much about any
limitations there.

--
Dan.

pgpwXMKusfxKP.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Richard Elling
On Jan 18, 2010, at 11:04 AM, Miles Nordin wrote:
...
> Another problem is that the snv_112 man page says this:
> 
> -8<-
> The format of the stream is evolving. No backwards  com-
> patibility is guaranteed. You may not be able to receive
> your streams on future versions of ZFS.
> -8<-
> 
> I think the new man page says something more generous. 


The Solaris 10 10/09 zfs(1m) man page says:

 The format of the stream is committed. You will be  able
 to receive your streams on future versions of ZFS.

I'm not sure when that hit snv, but obviously it was after b112.
...

> the solaris userland is ancient, so if you use gtar you'll be
> surprised less often, like IMHO you are less likely to have silently
> truncated files, but then it's maintained upstream so it's missing
> support for all these forked files, mandatory labels, and windows
> ACL's that they're piling into ZFS now for NAS features.

OOB, the default OpenSolaris PATH places /usr/gnu/bin ahead
of /usr/bin, so gnu tar is "the default."  As of b130 (I'm not running
an older build currently) the included gnu tar is version 1.22 which
is the latest as released March 2009 at http://www.gnu.org/software/tar/

> When we brought it up last time, I think we found no one knows of a
> userland tool similar to 'ufsdump' that's capable of serializing a ZFS
> along with holes, large files, ``attribute'' forks, windows ACL's, and
> checksums of its own, and then restoring the stream in a
> filesystem-agnostic read/write/lseek/... manner like 'ufsrestore'.

That is my understanding as well.  It seems that the backup vendors
are moving forward in a more-or-less vendor-specific way. This is
not necessarily a bad thing, since there are open source solutions.
But I see the requirements for backups being much more sophisticated
than ufsdump was 25 years ago.  hmmm... has ufsdump changed over 
the past 25 years? ;-)
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Miles Nordin
> "mg" == Mike Gerdts  writes:
> "tt" == Toby Thain  writes:
> "tb" == Thomas Burgess  writes:

mg> Yet it is used in ZFS flash archives on Solaris 10 and are
mg> slated for use in the successor to flash archives.

in FLAR, ``if a single bit's flipped, the whole archive and any
incrementals descended from it are toast'' is a desireable
characteristic.  FLAR is a replication tool, just like 'zfs send | zfs
receive' together are, and the sensitivity's desireable because you
want a deterministicaly exact replica.

If zpools themselves had the characteristic, ``single bit flipped,
entire pool lost,'' few would be happy with that.  In fact it's part
of the ZFS kool-aid cocktail that bitflips are detected and reported,
and that critical metadata is written several times far apart on the
pool so the damage of a single bitflip, even on an unredundant pool,
is limited.  Other things that aren't ``enterprise backup solutions'',
like tarballs, are also resilient to single bit flips.  so, why
tolerate this fragility from a backup format?  It's not appropriate.

The other problems are lack of a verification tool that doesn't
involve ponderous amounts of kernel code revision-bound to hardware
drivers and such that may one day become unobtainable, the possibility
of painting yourself into a corner (ex. if you are backing up a 17TB
filesystem, you must provide 17TB of restore space or else there is no
way to access a 2kB file), and the stream format which is
intentionally tied to the relatively often-changing zfs version so
that it can make exact copies but at the cost of constraining the
restore environment which may be separated from the backup environment
by seven years and/or in the midst of a disaster in a way that
tarballs and cpios and so on don't constrain.

Another problem is that the snv_112 man page says this:

-8<-
 The format of the stream is evolving. No backwards  com-
 patibility is guaranteed. You may not be able to receive
 your streams on future versions of ZFS.
-8<-

I think the new man page says something more generous.  The use case
here is, we should be able to:

   old solarisnew solaris   newer solaris

   zfs send  ->   zfs recv
  \
   \_ zpool upgrade -->
  (not 'zfs upgrade')


 -- zfs send
|
 -> zfs recv


   zfs recv  <- zfs send


That is, the stream format should be totally deterministic given the
zfs version, and not depend on the zpool version nor the
kernel/userland version.  This way a single backup pool can hold the
backups of several different versions of solaris.  The alternative to
this level of guarantee is for your archival backup to involve a ball
of executable code like a LiveCD or a .VDI, and that's just not on.
Seven to ten years, people---there are laws TODAY that require
archiving WORM media that long, and being able to read it, which means
no rewriting for a decade.  That's a ``hard requirement'' as
Christopher says, _today,_ never mind what's desireable or useful.
I'm not sure the new man page makes a promise quite that big as
accomodating the above chart, but IIRC it's much better than the old
one.


anyway, zfs send and receive are very good replication tools, and
replication can be used for backup (by replicating into another zpool
*NOT* by storing the stream modulo the caveat above) or for system
install (by recv'ing multiple times like FLAR).

If you choose to store zfs send streams as backups, the least you can
do is warn those you advise that zfs send streams are different from
other kidns of backup stream, because sysadmin experience in how to
write backups is old and hard-won, and these tools diverge from it.
they diverge usefully---I'm not putting them down---I'm saying *don't
use them that way* unless you are able to think through
$subtleproblems, after which you'll probably not want to use them that
way.

tt> I can see the temptation, but isn't it a bit under-designed? I
tt> think Mr Nordin might have ranted about this in the past...

no, I think it's great for FLAR.  single-bit-self-destruct is exactly
what's wanted for FLAR.

for those case where you could somehow magically capture and replay an
rsync stream it's great.  It's not a dumb design because IMHO a single
format probably can't do well for both replication and backup, but
it's not a backup format, enterprise or otherwise.  What escalates my
objection into a rant is the way ``enterprise backup solution'' gets
tossed around as some kind of soundbite hatetorial prole phrase
substituting and blocking off exploring the actual use cases which
we've done over and over again on this list.  I 

Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Lassi Tuura
Hi,

> .. it's hard to beat the convenience of a "backup file" format, for
> all sorts of reasons, including media handling, integration with other
> services, and network convenience. 

Yes.

> Consider then, using a zpool-in-a-file as the file format, rather than
> zfs send streams.

This is an interesting suggestion :-)

Did I understand you correctly that once a slice is written, zfs won't rewrite 
it? In other words, I can have an infinitely growing pile of these slices, and 
once zfs fills one file up (or one "raidz2" set of files), I flush it to 
tape/optical disk/whatever, and zfs won't change it "ever after"? When I need 
more space, I just add more slices, but old slices are effectively read-only?

Or did I misunderstand how you meant it work? It sounded very interesting but 
my understanding on zfs is currently limited :-)

And perhaps most importantly: has anyone actually done this for their back-ups 
and has success stories on restore after media failure?

Regards,
Lassi
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Lassi Tuura
Hi,

> Here is the big difference. For a professional backup people still typically
> use tapes although tapes have become expensive.
> 
> I still believe that a set of compressed incremental star archives give you 
> more features.

Thanks for your comments!

I think I roughly understand the feature trade-offs, and have some experience 
with back-up solutions ranging from simple to enterprise.

I guess what I am after is, for data which really matters to its owners and 
which they actually had to recover, did people use tar/pax archives (~ file 
level standard archive format), dump/restore (~ semi-standard format based on 
files/inodes) or zfs send/receive (~ file system block level dump), or 
something else, and why? (Regardless of how these are implemented, hobby 
scripts or enterprise tools, how they dealt with possible media failure issues, 
etc.)

Other than the file system vs. file restore, is there any concern in doing the 
block level thing? "Concerns" as in "mind being in the line of fire if it 
fails?" :-)

Regards,
Lassi
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Lassi Tuura
Hi,

>> I am considering building a modest sized storage system with zfs. Some
>> of the data on this is quite valuable, some small subset to be backed
>> up "forever", and I am evaluating back-up options with that in mind.
> 
> You don't need to store the "zfs send" data stream on your backup media.
> This would be annoying for the reasons mentioned - some risk of being able
> to restore in future (although that's a pretty small risk) and inability to
> restore with any granularity, i.e. you have to restore the whole FS if you
> restore anything at all.
> 
> A better approach would be "zfs send" and pipe directly to "zfs receive" on
> the external media.  This way, in the future, anything which can read ZFS
> can read the backup media, and you have granularity to restore either the
> whole FS, or individual things inside there.
> 
> Plus, the only way to guarantee the integrity of a "zfs send" data stream is
> to perform a "zfs receive" on that data stream.  So by performing a
> successful receive, you've guaranteed the datastream is not corrupt.  Yet.

Thanks for your feedback!

My plan is to have near-line back-up on zfs on another physically independent 
media, much as you describe. Then another copy off-site on separate system 
(disk), and third copy on WORM-like media in "chunks" somewhere else. I am 
really looking to have the latter two sliced to fit non-disk media (tape, 
optical disks, ...), or external storage not really usable as a filesystem (S3, 
web hosting, ...).

The off-site disk is not in zfs-capable system, unless I set up a virtual 
machine; but it won't have enough physical disks for raidz2 anyway. I am 
primarily looking for something I can store encrypted in dvd- or tape-sized 
slices on at least two physically separate media. This backup would be used 
when at least two other media have already been lost, so while convenience is a 
plus I really desire reliability and longevity. (Yes I know tapes and DVDs die 
too.)

I appreciate the comments by others on zfs send not allowing individual file 
restore, but as I wrote before this is not very important to me, at least not 
as important than my other questions. Apropos, is anyone able to respond to 
those? (Is format documented, independent tools, bugs known to have affected 
send/restore, would you recommend over tar/pax in real life, etc.)

I am very interested in what people have actually done and experienced in real 
life, including disaster recovery from multiple media failure.

Regards,
Lassi
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Robert Milkowski
or you might do something like: 
http://milek.blogspot.com/2009/12/my-presentation-at-losug.html


however in your case if all your clients are running zfs only 
filesystems then relaying just on zfs send|recv might be a good idea.



--
Robert Milkowski
http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Robert Milkowski

On 18/01/2010 08:59, Phil Harman wrote:
YMMV. At a recent LOSUG meeting we were told of a case where rsync was 
faster than an incremental zfs send/recv. But I think that was for a 
mail server with many tiny files (i.e. changed blocks are very easy to 
find in files with very few blocks).





After changes around build 114 (iirc) this shouldn't be the case anymore 
znd an incremental zfs send should always be able to at least match the 
performance of incremental rsync. Now with lots of files where only a 
small subset of them changes incremental zfs send should be much faster 
and that's my observation.


--
Robert Milkowski
http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Thomas Burgess
On Mon, Jan 18, 2010 at 3:59 AM, Phil Harman  wrote:

> YMMV. At a recent LOSUG meeting we were told of a case where rsync was
> faster than an incremental zfs send/recv. But I think that was for a mail
> server with many tiny files (i.e. changed blocks are very easy to find in
> files with very few blocks).
>
> However, I don't see why further ZFS perfomance work couldn't close that
> gap, since rsync will always need to compare directories and timestamps.
>
> Phil
>
> The best info i've read on this was on this blog:
http://richardelling.blogspot.com/2009/01/parallel-zfs-sendreceive.html


>
> On 18 Jan 2010, at 08:07, Edward Ned Harvey  wrote:
>
>  I still believe that a set of compressed incremental star archives give
>>> you
>>> more features.
>>>
>>
>> Big difference there is that in order to create an incremental star
>> archive,
>> star has to walk the whole filesystem or folder that's getting backed up,
>> and do a "stat" on every file to see which files have changed since the
>> last
>> backup run.  If you have a large filesystem, that can take a very long
>> time.
>>
>> I recently switched to ZFS specifically for this reason.  Previously, I
>> was
>> doing a nightly rsync on 1Tb of data.  It required 10 hrs every night to
>> run, and copy typically a few hundred megs that had changed that day.
>>  Now,
>> I run incremental zfs send & receive, and it completes typically in a
>> minute
>> or two.
>>
>> ___
>> zfs-discuss mailing list
>> zfs-discuss@opensolaris.org
>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>>
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Phil Harman
YMMV. At a recent LOSUG meeting we were told of a case where rsync was  
faster than an incremental zfs send/recv. But I think that was for a  
mail server with many tiny files (i.e. changed blocks are very easy to  
find in files with very few blocks).


However, I don't see why further ZFS perfomance work couldn't close  
that gap, since rsync will always need to compare directories and  
timestamps.


Phil

On 18 Jan 2010, at 08:07, Edward Ned Harvey   
wrote:


I still believe that a set of compressed incremental star archives  
give

you
more features.


Big difference there is that in order to create an incremental star  
archive,
star has to walk the whole filesystem or folder that's getting  
backed up,
and do a "stat" on every file to see which files have changed since  
the last
backup run.  If you have a large filesystem, that can take a very  
long time.


I recently switched to ZFS specifically for this reason.   
Previously, I was
doing a nightly rsync on 1Tb of data.  It required 10 hrs every  
night to
run, and copy typically a few hundred megs that had changed that  
day.  Now,
I run incremental zfs send & receive, and it completes typically in  
a minute

or two.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Edward Ned Harvey
> Consider then, using a zpool-in-a-file as the file format, rather than
> zfs send streams.

That's a pretty cool idea.  Then you've still got the entire zfs volume
inside of a file, but you're able to mount and extract individual files if
you want, and you're able to pipe your zfs send directly to zfs receive, and
basically you get all the benefits of zfs send & receive, but also get all
the benefits of being able to treat your backup like a file or a set of
files.

Pretty cool.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-18 Thread Edward Ned Harvey
> I still believe that a set of compressed incremental star archives give
> you
> more features.

Big difference there is that in order to create an incremental star archive,
star has to walk the whole filesystem or folder that's getting backed up,
and do a "stat" on every file to see which files have changed since the last
backup run.  If you have a large filesystem, that can take a very long time.

I recently switched to ZFS specifically for this reason.  Previously, I was
doing a nightly rsync on 1Tb of data.  It required 10 hrs every night to
run, and copy typically a few hundred megs that had changed that day.  Now,
I run incremental zfs send & receive, and it completes typically in a minute
or two.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-17 Thread Thomas Burgess
>
>
>
> What dou you use instead?
>
*
*
**tar cvf - /some/dir | (cd /some/other/dir; tar xf -)


>
> BTW: I recommend "star" and to use either the H=exustar or -dump option.
>
> Jörg
>
>
i will have to check it out.  I recently migrated to opensolaris from
FreeBSD and i have a LOT to learn.  I am really enjoying Opensolaris so far.


so star is better than tar?
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-17 Thread Daniel Carosone
On Sun, Jan 17, 2010 at 05:31:39AM -0500, Edward Ned Harvey wrote:
> Instead, it is far preferable to "zfs send | zfs receive"  ...  That is,
> receive the data stream on external media as soon as you send it. 

Agree 100% - but..

.. it's hard to beat the convenience of a "backup file" format, for
all sorts of reasons, including media handling, integration with other
services, and network convenience. 

Consider then, using a zpool-in-a-file as the file format, rather than
zfs send streams.

Make a backup pool out of N conveniently-sized files, sparse to start
with if it suits you.  Go ahead and make it a raidz[23] pool, in case
some of your backup media goes bad. zfs recv your backups into this
pool.  zfs export the pool, to quiesce the file contents. 

If the files are themselves in zfs, snapshot that filesystem now for
good measure; then you can immediately bring your backup pool online
again, as well as have local reference copies of old backups, as
storage space allows. 

Send these files to whatever backup system/service you want to use.  
Handle them like any other large archive file format.

If you choose file sizes well, you can burn them to dvd's or write
them to tapes.  Multiple smaller files (rather than one file per
medium) can be easier to handle, but its best to make multiple raidz
vdevs and arrange a file from each vdev per medium (so you can still
recover from a lost dvd/tape). 

If you have a non-zfs remote storage server, rsync works well to
update the remote copies incrementally after each backup cycle (and to
bring back older versions later if needed).  Lots of cloud backup
providers exist now that can do similar incremental replication.

gpg sign and encrypt the files before sending, if you need to. Some
day soon, zfs crypto will allow backups encrypted within the pool,
without defeating incremental replication of the files (as gpg will).

Another option is to mount these files directly on whatever generic NAS
devices you want to hold the backups, and import the pool from there.
I'd be wary, but if you were to consider that, fortunately there's a
good testing tool (zfs scrub) to help you be sure the NAS service is
reliable.  

You do need all (or at least most) of your files available in order to
do a restore.  Given that you should be preferring to use local
snapshots for small recovery jobs anyway, that's not really a burden
for a full recovery.  At this point, you get back all the snapshots
that were in the backup pool at the time it was saved.

The zfs send stream format used to not be committed, though it appears
this has recently changed.  It still has all the other drawbacks
previously noted. The zpool format is committed and does have a tested
and supported backwards compatibility/upgrade path. 

--
Dan.

pgpmu7bvoOxnf.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-17 Thread Joerg Schilling
Thomas Burgess  wrote:

>  I found this out the hard way last time i used it.  I was backing up all my
> data from one system to another using cpio and i had a bunch of movies over
> 8GB (720p and 1080p mkv files)
>
> none of them worked.  I never use cpio anymore.

What dou you use instead?

BTW: I recommend "star" and to use either the H=exustar or -dump option.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-17 Thread Thomas Burgess
>
> Cpio coes not
> support sparse files and is unable to archive files > 8 GB.
>
> Jörg
>
>
 I found this out the hard way last time i used it.  I was backing up all my
data from one system to another using cpio and i had a bunch of movies over
8GB (720p and 1080p mkv files)

none of them worked.  I never use cpio anymore.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-17 Thread Joerg Schilling
Edward Ned Harvey  wrote:

> > NO, zfs send is not a backup.
>
> Understood, but perhaps you didn't read my whole message.  Here, I will
> spell out the whole discussion:
...
> Instead, it is far preferable to "zfs send | zfs receive"  ...  That is,
> receive the data stream on external media as soon as you send it.  By
> receiving the data stream onto external media, instead of just saving the
> datastream as a file on external media ... You solve both of the above
> problems.  Obviously this is only possible with external disks, and not
> possible with tapes.

Here is the big difference. For a professional backup people still typically
use tapes although tapes have become expensive.

I still believe that a set of compressed incremental star archives give you 
more features.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-17 Thread Joerg Schilling
Toby Thain  wrote:

> > Yet it is used in ZFS flash archives on Solaris 10
>
> I can see the temptation, but isn't it a bit under-designed? I think  
> Mr Nordin might have ranted about this in the past...

Isn't flash cpio based and thus not prepared for the future? Cpio coes not 
support sparse files and is unable to archive files > 8 GB.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-17 Thread Edward Ned Harvey
> NO, zfs send is not a backup.

Understood, but perhaps you didn't read my whole message.  Here, I will
spell out the whole discussion:

If you "zfs send > somefile" it is well understood there are two big
problems with this method of backup.  #1 If a single bit error is introduced
into the file, then the whole data stream is corrupt.  #2 If you want to
restore just a subset of the filesystem, you cannot.  The only option
available is to restore the whole filesystem.

Instead, it is far preferable to "zfs send | zfs receive"  ...  That is,
receive the data stream on external media as soon as you send it.  By
receiving the data stream onto external media, instead of just saving the
datastream as a file on external media ... You solve both of the above
problems.  Obviously this is only possible with external disks, and not
possible with tapes.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-16 Thread Toby Thain


On 16-Jan-10, at 6:51 PM, Mike Gerdts wrote:

On Sat, Jan 16, 2010 at 5:31 PM, Toby Thain  
 wrote:

On 16-Jan-10, at 7:30 AM, Edward Ned Harvey wrote:

I am considering building a modest sized storage system with  
zfs. Some
of the data on this is quite valuable, some small subset to be  
backed
up "forever", and I am evaluating back-up options with that in  
mind.


You don't need to store the "zfs send" data stream on your backup  
media.
This would be annoying for the reasons mentioned - some risk of  
being able
to restore in future (although that's a pretty small risk) and  
inability

to
restore with any granularity, i.e. you have to restore the whole  
FS if you

restore anything at all.

A better approach would be "zfs send" and pipe directly to "zfs  
receive"

on
the external media.  This way, in the future, anything which can  
read ZFS
can read the backup media, and you have granularity to restore  
either the

whole FS, or individual things inside there.


There have also been comments about the extreme fragility of the  
data stream
compared to other archive formats. In general it is strongly  
discouraged for

these purposes.



Yet it is used in ZFS flash archives on Solaris 10


I can see the temptation, but isn't it a bit under-designed? I think  
Mr Nordin might have ranted about this in the past...


--Toby



and are slated for
use in the successor to flash archives.  This initial proposal seems
to imply using the same mechanism for a system image backup (instead
of just system provisioning).

http://mail.opensolaris.org/pipermail/caiman-discuss/2010-January/ 
015909.html


--
Mike Gerdts
http://mgerdts.blogspot.com/


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send/receive as backup - reliability?

2010-01-16 Thread Mike Gerdts
On Sat, Jan 16, 2010 at 5:31 PM, Toby Thain  wrote:
> On 16-Jan-10, at 7:30 AM, Edward Ned Harvey wrote:
>
>>> I am considering building a modest sized storage system with zfs. Some
>>> of the data on this is quite valuable, some small subset to be backed
>>> up "forever", and I am evaluating back-up options with that in mind.
>>
>> You don't need to store the "zfs send" data stream on your backup media.
>> This would be annoying for the reasons mentioned - some risk of being able
>> to restore in future (although that's a pretty small risk) and inability
>> to
>> restore with any granularity, i.e. you have to restore the whole FS if you
>> restore anything at all.
>>
>> A better approach would be "zfs send" and pipe directly to "zfs receive"
>> on
>> the external media.  This way, in the future, anything which can read ZFS
>> can read the backup media, and you have granularity to restore either the
>> whole FS, or individual things inside there.
>
> There have also been comments about the extreme fragility of the data stream
> compared to other archive formats. In general it is strongly discouraged for
> these purposes.
>

Yet it is used in ZFS flash archives on Solaris 10 and are slated for
use in the successor to flash archives.  This initial proposal seems
to imply using the same mechanism for a system image backup (instead
of just system provisioning).

http://mail.opensolaris.org/pipermail/caiman-discuss/2010-January/015909.html

-- 
Mike Gerdts
http://mgerdts.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


  1   2   >