Re: [zfs-discuss] How well does zfs mirror handle temporary disk offlines?

2011-01-18 Thread Edward Ned Harvey
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Erik Trimble
> 
> As far as what the resync does:  ZFS does "smart" resilvering, in that
> it compares what the "good" side of the mirror has against what the
> "bad" side has, and only copies the differences over to sync them up.
> This is one of ZFS's great strengths, in that most other RAID systems
> can't do this.

It's also one of ZFS's great weaknesses.  It's a strength as long as not
much data has changed, or it was highly sequential in nature, or the drives
in the pool have extremely high IOPS (SSD's etc) because then resilvering
just the changed parts can be done very quickly.  Much quicker than
resilvering the whole drive sequentially as a typical hardware raid would
do.  However, as is often the case, a large percentage of the drive may have
changed, in essentially random order.  There are many situations where
something like 3% of the drive has changed, yet the resilver takes 100% as
long as rewriting the entire drive sequentially would have taken.  10% of
the drive changed  ZFS resilver might be 4x slower than sequentially
overwriting the entire disk as a hardware raid would have done.

Ultimately, your performance depends entirely on your usage patterns, your
pool configuration, and type of hardware.

To the OP:  If you've got one device on one SAN, mirrored to another device
on another SAN, you're probably only expecting very brief outages on either
SAN.  As such, you probably won't see any large percentage of the online SAN
change, and when the temporarily failed SAN comes back online, you can
probably expect a very fast resilver.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How well does zfs mirror handle temporary disk offlines?

2011-01-18 Thread Erik Trimble
On Tue, 2011-01-18 at 13:34 -0800, Philip Brown wrote:
> > On Tue, 2011-01-18 at 14:51 -0500, Torrey McMahon
> > wrote:
> >
> > ZFS's ability to handle "short-term" interruptions
> > depend heavily on the
> > underlying device driver.
> > 
> > If the device driver reports the device as
> > "dead/missing/etc" at any
> > point, then ZFS is going to require a "zpool replace"
> > action before it
> > re-accepts the device.  If the underlying driver
> > simply stalls, then
> > it's more graceful (and no user interaction is
> > required).
> > 
> > As far as what the resync does:  ZFS does "smart"
> > resilvering, in that
> > it compares what the "good" side of the mirror has
> > against what the
> > "bad" side has, and only copies the differences over
> > to sync them up.
> >
> 
> Hmm. Well, we're talking fibre, so we're very concerned with the recovery  
> mode when the fibre drivers have marked it as "failed". (except it hasnt 
> "really" failed, we've just had a switch drop out)
> 
> I THINK what you are saying, is that we could, in this situation, do:
> 
> zpool replace (old drive) (new drive)
> 
> and then your "smart" recovery, should do the limited resilvering only. Even 
> for potentially long outages.
> 
> Is that what you are saying?


Yes. It will always look at the "replaced" drive to see if it was a
prior member of the mirror, and do smart resilvering if possible.

If the device path stays the same (which, hopefully, it should), you can
even do:

zpool replace (old device) (old device)




-- 
Erik Trimble
Java System Support
Mailstop:  usca22-317
Phone:  x67195
Santa Clara, CA
Timezone: US/Pacific (GMT-0800)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How well does zfs mirror handle temporary disk offlines?

2011-01-18 Thread Philip Brown
> On Tue, 2011-01-18 at 14:51 -0500, Torrey McMahon
> wrote:
>
> ZFS's ability to handle "short-term" interruptions
> depend heavily on the
> underlying device driver.
> 
> If the device driver reports the device as
> "dead/missing/etc" at any
> point, then ZFS is going to require a "zpool replace"
> action before it
> re-accepts the device.  If the underlying driver
> simply stalls, then
> it's more graceful (and no user interaction is
> required).
> 
> As far as what the resync does:  ZFS does "smart"
> resilvering, in that
> it compares what the "good" side of the mirror has
> against what the
> "bad" side has, and only copies the differences over
> to sync them up.
>

Hmm. Well, we're talking fibre, so we're very concerned with the recovery  mode 
when the fibre drivers have marked it as "failed". (except it hasnt "really" 
failed, we've just had a switch drop out)

I THINK what you are saying, is that we could, in this situation, do:

zpool replace (old drive) (new drive)

and then your "smart" recovery, should do the limited resilvering only. Even 
for potentially long outages.

Is that what you are saying?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How well does zfs mirror handle temporary disk offlines?

2011-01-18 Thread Chris Banal

Erik Trimble wrote:

On Tue, 2011-01-18 at 14:51 -0500, Torrey McMahon wrote:

On 1/18/2011 2:46 PM, Philip Brown wrote:

My specific question is, how easily does ZFS handle*temporary*  SAN 
disconnects, to one side of the mirror?
What if the outage is only 60 seconds?
3 minutes?
10 minutes?
an hour?


No idea how well it will reconnect the device but we had an X4500 that 
would randomly boot up and one or two disks would be missing. Reboot 
again and one or two other disks would be missing. While we were trouble 
shooting this problem this happened dozens and dozens of times and zfs 
had no trouble with it as far as I could tell. Would only resliver the 
data that was changed while that drive was offline. We had no data loss.



Thank you,
Chris Banal

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How well does zfs mirror handle temporary disk offlines?

2011-01-18 Thread Erik Trimble
On Tue, 2011-01-18 at 14:51 -0500, Torrey McMahon wrote:
> 
> On 1/18/2011 2:46 PM, Philip Brown wrote:
> > My specific question is, how easily does ZFS handle*temporary*  SAN 
> > disconnects, to one side of the mirror?
> > What if the outage is only 60 seconds?
> > 3 minutes?
> > 10 minutes?
> > an hour?
> 
> Depends on the multipath drivers and the failure mode. For example, if 
> the link drops completely at the host hba connection some failover 
> drivers will mark the path down immediately which will propagate up the 
> stack faster than an intermittent connection or something father down 
> stream failing.
> 
> > If we have 2x1TB drives, in a simple zfs mirror if one side goes 
> > temporarily off line, will zfs attempt to resync **1 TB** when it comes 
> > back? Or does it have enough intelligence to say, "oh hey I know this 
> > disk..and I know [these bits] are still good, so I just need to resync 
> > [that bit]" ?
> 
> My understanding is yes though I can't find the reference for this. (I'm 
> sure someone else will find it in short order.)


ZFS's ability to handle "short-term" interruptions depend heavily on the
underlying device driver.

If the device driver reports the device as "dead/missing/etc" at any
point, then ZFS is going to require a "zpool replace" action before it
re-accepts the device.  If the underlying driver simply stalls, then
it's more graceful (and no user interaction is required).

As far as what the resync does:  ZFS does "smart" resilvering, in that
it compares what the "good" side of the mirror has against what the
"bad" side has, and only copies the differences over to sync them up.
This is one of ZFS's great strengths, in that most other RAID systems
can't do this.



-- 
Erik Trimble
Java System Support
Mailstop:  usca22-317
Phone:  x67195
Santa Clara, CA
Timezone: US/Pacific (GMT-0800)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How well does zfs mirror handle temporary disk offlines?

2011-01-18 Thread Torrey McMahon



On 1/18/2011 2:46 PM, Philip Brown wrote:

My specific question is, how easily does ZFS handle*temporary*  SAN 
disconnects, to one side of the mirror?
What if the outage is only 60 seconds?
3 minutes?
10 minutes?
an hour?


Depends on the multipath drivers and the failure mode. For example, if 
the link drops completely at the host hba connection some failover 
drivers will mark the path down immediately which will propagate up the 
stack faster than an intermittent connection or something father down 
stream failing.



If we have 2x1TB drives, in a simple zfs mirror if one side goes temporarily off 
line, will zfs attempt to resync **1 TB** when it comes back? Or does it have enough 
intelligence to say, "oh hey I know this disk..and I know [these bits] are still 
good, so I just need to resync [that bit]" ?


My understanding is yes though I can't find the reference for this. (I'm 
sure someone else will find it in short order.)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] How well does zfs mirror handle temporary disk offlines?

2011-01-18 Thread Philip Brown
Sorry if this is well known.. I tried a bunch of googles, but didnt get 
anywhere useful. Closest I came, was 
http://mail.opensolaris.org/pipermail/zfs-discuss/2009-April/028090.html  but 
that doesnt answer my question, below, reguarding zfs mirror recovery.
Details of our needs follow.


We normally are very into redundancy. Pretty much all our SAN storage is dual 
ported, along with all our production hosts. Two completely redundant paths to 
storage. Two independant SANs.

However, now, we are encountering a need for "tier 3" storage, aka "not that 
important, we're going to go cheap on it" ;-)
That being said, we'd still like to make it as reliable and robust as possible. 
So I was wondering just how robust it would be to do ZFS mirroring, across 2 
sans.

My specific question is, how easily does ZFS handle *temporary* SAN 
disconnects, to one side of the mirror?
What if the outage is only 60 seconds?
3 minutes?
10 minutes?
an hour?

If we have 2x1TB drives, in a simple zfs mirror if one side goes 
temporarily off line, will zfs attempt to resync **1 TB** when it comes back? 
Or does it have enough intelligence to say, "oh hey I know this disk..and I 
know [these bits] are still good, so I just need to resync [that bit]" ?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss