Anyone who has an Xraid should have one (or 2) of these BBC modules. good mojo.
http://store.apple.com/1-800-MY-APPLE/WebObjects/AppleStore.woa/wa/RSLID ?mco=6C04E0D7&nplm=M8941G/B Can you tell I <3 apple? -----Original Message----- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Wee Yeh Tan Sent: Thursday, April 26, 2007 9:40 PM To: cedric briner Cc: [EMAIL PROTECTED] Subject: Re: [zfs-discuss] HowTo: UPS + ZFS & NFS + no fsync Cedric, On 4/26/07, cedric briner <[EMAIL PROTECTED]> wrote: > >> okay let'say that it is not. :) > >> Imagine that I setup a box: > >> - with Solaris > >> - with many HDs (directly attached). > >> - use ZFS as the FS > >> - export the Data with NFS > >> - on an UPS. > >> > >> Then after reading the : > >> http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_G > >> uide#ZFS_and_Complex_Storage_Considerations > >> > >> I wonder if there is a way to tell the OS to ignore the fsync flush > >> commands since they are likely to survive a power outage. > > > > Cedric, > > > > You do not want to ignore syncs from ZFS if your harddisk is > > directly attached to the server. As the document mentioned, that is > > really for Complex Storage with NVRAM where flush is not necessary. > > This post follows : `XServe Raid & Complex Storage Considerations' > http://www.opensolaris.org/jive/thread.jspa?threadID=29276&tstart=0 Ah... I wasn't aware the other thread was started by you :). If your storage device features NVRAM, you should in fact configure it as discussed in the stated thread. However, if your storage device(s) are directly attached disks (or anything without an NVRAM controller), zfs_noflush=1 is potentially fatal (see link below). > Where we have made the assumption (*1) if the XServe Raid is connected > to an UPS that we can consider the RAM in the XServe Raid as it was NVRAM. I'm not sure about the interaction between XServe and the UPS but I'd imagine that the UPS can probably power the XServe for a few minutes after a power outage. That should be enough time for the XServe to drain stuff in its RAM to disk. > (*1) > This assumption is even pointed by Roch : > http://blogs.sun.com/roch/#zfs_to_ufs_performance_comparison > >> Intelligent Storage > through: `the Shenanigans with ZFS flushing and intelligent arrays...' > http://blogs.digitar.com/jjww/?itemid=44 > >> Tell your array to ignore ZFS' flush commands > > So in this way, when we export it with NFS we get a boost in the BW. Indeed. This is especially true if you consider that expensive storage are likely to be shared by more than 1 host. A flush command likely flushes the entire cache rather than just parts relevant to the requesting host. > Okay, then is there any difference that I do not catch between : > - the Shenanigans with ZFS flushing and intelligent arrays... > - and my situation > > I mean, I want to have a cheap and reliable nfs service. Why should I > buy expensive `Complex Storage with NVRAM' and not just buying a > machine with 8 IDE HD's ? Your 8 IDE HD may not benefit from zfs_noflush=1 since their caches are small anyway but the potential impact on reliability will be fairly severe. http://www.opensolaris.org/jive/thread.jspa?messageID=91730 Nothing is stopping you though from getting decent performance from 8 IDE HDD. You just should not treat them like they are NVRAM backed array. -- Just me, Wire ... Blog: <prstat.blogspot.com> _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss