Re: [zfs-discuss] dedupe is in

2009-11-02 Thread Ross Smith
Ok, thanks everyone then (but still thanks to Victor for the heads up) :-) On Mon, Nov 2, 2009 at 4:03 PM, Victor Latushkin wrote: > On 02.11.09 18:38, Ross wrote: >> >> Double WOHOO!  Thanks Victor! > > Thanks should go to Tim Haley, Jeff Bonwick and George Wilson ;-) > ___

Re: [zfs-discuss] Tunable iSCSI timeouts - ZFS over iSCSI fix

2009-07-29 Thread Ross Smith
Yup, somebody pointed that out to me last week and I can't wait :-) On Wed, Jul 29, 2009 at 7:48 PM, Dave wrote: > Anyone (Ross?) creating ZFS pools over iSCSI connections will want to pay > attention to snv_121 which fixes the 3 minute hang after iSCSI disk > problems: > > http://bugs.opensolari

Re: [zfs-discuss] ZFS: unreliable for professional usage?

2009-02-14 Thread Ross Smith
where somebody used mdb to search a corrupt pool to try to recover data: http://opensolaris.org/jive/message.jspa?messageID=318009 On Fri, Feb 13, 2009 at 11:09 PM, Richard Elling wrote: > Tim wrote: >> >> >> On Fri, Feb 13, 2009 at 4:21 PM, Bob Friesenhahn >> mailto:bf

Re: [zfs-discuss] ZFS: unreliable for professional usage?

2009-02-13 Thread Ross Smith
t more time should recovery be needed. On Fri, Feb 13, 2009 at 8:59 PM, Bob Friesenhahn wrote: > On Fri, 13 Feb 2009, Ross Smith wrote: > >> Thinking about this a bit more, you've given me an idea: Would it be >> worth ZFS occasionally reading previous uberblocks from the

Re: [zfs-discuss] ZFS: unreliable for professional usage?

2009-02-13 Thread Ross Smith
On Fri, Feb 13, 2009 at 8:24 PM, Bob Friesenhahn wrote: > On Fri, 13 Feb 2009, Ross Smith wrote: >> >> You have to consider that even with improperly working hardware, ZFS >> has been checksumming data, so if that hardware has been working for >> any length of time, you

Re: [zfs-discuss] ZFS: unreliable for professional usage?

2009-02-13 Thread Ross Smith
On Fri, Feb 13, 2009 at 8:24 PM, Bob Friesenhahn wrote: > On Fri, 13 Feb 2009, Ross Smith wrote: >> >> You have to consider that even with improperly working hardware, ZFS >> has been checksumming data, so if that hardware has been working for >> any length of time, you

Re: [zfs-discuss] ZFS: unreliable for professional usage?

2009-02-13 Thread Ross Smith
On Fri, Feb 13, 2009 at 7:41 PM, Bob Friesenhahn wrote: > On Fri, 13 Feb 2009, Ross wrote: >> >> Something like that will have people praising ZFS' ability to safeguard >> their data, and the way it recovers even after system crashes or when >> hardware has gone wrong. You could even have a "comm

Re: [zfs-discuss] ZFS: unreliable for professional usage?

2009-02-12 Thread Ross Smith
That would be the ideal, but really I'd settle for just improved error handling and recovery for now. In the longer term, disabling write caching by default for USB or Firewire drives might be nice. On Thu, Feb 12, 2009 at 8:35 PM, Gary Mills wrote: > On Thu, Feb 12, 2009 at 11:53:40AM -0500, G

Re: [zfs-discuss] FW: Supermicro AOC-SAT2-MV8 hang when drive removed

2009-02-12 Thread Ross Smith
Heh, yeah, I've thought the same kind of thing in the past. The problem is that the argument doesn't really work for system admins. As far as I'm concerned, the 7000 series is a new hardware platform, with relatively untested drivers, running a software solution that I know is prone to locking up

Re: [zfs-discuss] Data loss bug - sidelined??

2009-02-06 Thread Ross Smith
ve that is how often the cache should be writing). On Fri, Feb 6, 2009 at 7:04 PM, Brent Jones wrote: > On Fri, Feb 6, 2009 at 10:50 AM, Ross Smith wrote: >> I can check on Monday, but the system will probably panic... which >> doesn't really help :-) >> >> Am I

Re: [zfs-discuss] Data loss bug - sidelined??

2009-02-06 Thread Ross Smith
I can check on Monday, but the system will probably panic... which doesn't really help :-) Am I right in thinking failmode=wait is still the default? If so, that should be how it's set as this testing was done on a clean install of snv_106. From what I've seen, I don't think this is a problem wi

Re: [zfs-discuss] Any way to set casesensitivity=mixed on the main pool?

2009-02-04 Thread Ross Smith
It's not intuitive because when you know that -o sets options, an error message saying that it's not a valid property makes you think that it's not possible to do what you're trying. Documented and intuitive are very different things. I do appreciate that the details are there in the manuals, but

Re: [zfs-discuss] SSD drives in Sun Fire X4540 or X4500 for dedicated ZIL device

2009-01-23 Thread Ross Smith
That's my understanding too. One (STEC?) drive as a write cache, basically a write optimised SSD. And cheaper, larger, read optimised SSD's for the read cache. I thought it was an odd strategy until I read into SSD's a little more and realised you really do have to think about your usage cases w

Re: [zfs-discuss] zfs list improvements?

2009-01-10 Thread Ross Smith
Hmm... that's a tough one. To me, it's a trade off either way, using a -r parameter to specify the depth for zfs list feels more intuitive than adding extra commands to modify the -r behaviour, but I can see your point. But then, using -c or -d means there's an optional parameter for zfs list tha

Re: [zfs-discuss] Using zfs mirror as a simple backup mechanism for time-slider.

2008-12-22 Thread Ross Smith
On Fri, Dec 19, 2008 at 6:47 PM, Richard Elling wrote: > Ross wrote: >> >> Well, I really like the idea of an automatic service to manage >> send/receives to backup devices, so if you guys don't mind, I'm going to >> share some other ideas for features I think would be useful. >> > > cool. > >> On

Re: [zfs-discuss] Using zfs mirror as a simple backup mechanism for time-slider.

2008-12-18 Thread Ross Smith
I was thinking more something like: - find all disk devices and slices that have ZFS pools on them - show users the devices and pool names (and UUIDs and device paths in case of conflicts).. >>> >>> I was thinking that device & pool names are too variable, you need

Re: [zfs-discuss] Using zfs mirror as a simple backup mechanism for time-slider.

2008-12-18 Thread Ross Smith
>> Of course, you'll need some settings for this so it's not annoying if >> people don't want to use it. A simple tick box on that pop up dialog >> allowing people to say "don't ask me again" would probably do. > > I would like something better than that. "Don't ask me again" sucks > when much, m

Re: [zfs-discuss] Using zfs mirror as a simple backup mechanism for time-slider.

2008-12-18 Thread Ross Smith
On Thu, Dec 18, 2008 at 7:11 PM, Nicolas Williams wrote: > On Thu, Dec 18, 2008 at 07:05:44PM +0000, Ross Smith wrote: >> > Absolutely. >> > >> > The tool shouldn't need to know that the backup disk is accessed via >> > USB, or whateve

Re: [zfs-discuss] Using zfs mirror as a simple backup mechanism for time-slider.

2008-12-18 Thread Ross Smith
> Absolutely. > > The tool shouldn't need to know that the backup disk is accessed via > USB, or whatever. The GUI should, however, present devices > intelligently, not as cXtYdZ! Yup, and that's easily achieved by simply prompting for a user friendly name as devices are attached. Now you could

Re: [zfs-discuss] Need Help Invalidating Uberblock

2008-12-16 Thread Ross Smith
It sounds to me like there are several potentially valid filesystem uberblocks available, am I understanding this right? 1. There are four copies of the current uberblock. Any one of these should be enough to load your pool with no data loss. 2. There are also a few (would love to know how many)

Re: [zfs-discuss] Split responsibility for data with ZFS

2008-12-15 Thread Ross Smith
I'm not sure I follow how that can happen, I thought ZFS writes were designed to be atomic? They either commit properly on disk or they don't? On Mon, Dec 15, 2008 at 6:34 PM, Bob Friesenhahn wrote: > On Mon, 15 Dec 2008, Ross wrote: > >> My concern is that ZFS has all this information on disk,

Re: [zfs-discuss] Split responsibility for data with ZFS

2008-12-15 Thread Ross Smith
Forgive me for not understanding the details, but couldn't you also work backwards through the blocks with ZFS and attempt to recreate the uberblock? So if you lost the uberblock, could you (memory and time allowing) start scanning the disk, looking for orphan blocks that aren't refernced anywhere

Re: [zfs-discuss] zfs not yet suitable for HA applications?

2008-12-05 Thread Ross Smith
Hi Dan, replying in line: On Fri, Dec 5, 2008 at 9:19 PM, David Anderson <[EMAIL PROTECTED]> wrote: > Trying to keep this in the spotlight. Apologies for the lengthy post. Heh, don't apologise, you should see some of my posts... o_0 > I'd really like to see features as described by Ross in his s

Re: [zfs-discuss] Availability: ZFS needs to handle disk removal / driver failure better

2008-12-03 Thread Ross Smith
Yeah, thanks Maurice, I just saw that one this afternoon. I guess you can't reboot with iscsi full stop... o_0 And I've seen the iscsi bug before (I was just too lazy to look it up lol), I've been complaining about that since February. In fact it's been a bad week for iscsi here, I've managed to

Re: [zfs-discuss] Availability: ZFS needs to handle disk removal / driver failure better

2008-12-02 Thread Ross Smith
Hi Richard, Thanks, I'll give that a try. I think I just had a kernel dump while trying to boot this system back up though, I don't think it likes it if the iscsi targets aren't available during boot. Again, that rings a bell, so I'll go see if that's another known bug. Changing that setting on

Re: [zfs-discuss] Availability: ZFS needs to handle disk removal / driver failure better

2008-12-02 Thread Ross Smith
Hey folks, I've just followed up on this, testing iSCSI with a raided pool, and it still appears to be struggling when a device goes offline. >>> I don't see how this could work except for mirrored pools. Would that >>> carry enough market to be worthwhile? >>> -- richard >>> >> >> I have to adm

Re: [zfs-discuss] Availability: ZFS needs to handle disk removal / driver failure better

2008-11-27 Thread Ross Smith
On Fri, Nov 28, 2008 at 5:05 AM, Richard Elling <[EMAIL PROTECTED]> wrote: > Ross wrote: >> >> Well, you're not alone in wanting to use ZFS and iSCSI like that, and in >> fact my change request suggested that this is exactly one of the things that >> could be addressed: >> >> "The idea is really a

Re: [zfs-discuss] "ZFS, Smashing Baby" a fake???

2008-11-25 Thread Ross Smith
t and parcel of that. On Tue, Nov 25, 2008 at 3:57 PM, Bob Friesenhahn <[EMAIL PROTECTED]> wrote: > On Tue, 25 Nov 2008, Ross Smith wrote: >> >> Good to hear there's work going on to address this. >> >> What did you guys think to my idea of ZFS supporting

Re: [zfs-discuss] "ZFS, Smashing Baby" a fake???

2008-11-25 Thread Ross Smith
> The shortcomings of timeouts have been discussed on this list before. How do > you tell the difference between a drive that is dead and a path that is just > highly loaded? A path that is dead is either returning bad data, or isn't returning anything. A highly loaded path is by definition readi

Re: [zfs-discuss] "ZFS, Smashing Baby" a fake???

2008-11-25 Thread Ross Smith
Hmm, true. The idea doesn't work so well if you have a lot of writes, so there needs to be some thought as to how you handle that. Just thinking aloud, could the missing writes be written to the log file on the rest of the pool? Or temporarily stored somewhere else in the pool? Would it be an o

Re: [zfs-discuss] "ZFS, Smashing Baby" a fake???

2008-11-25 Thread Ross Smith
No, I count that as "doesn't return data ok", but my post wasn't very clear at all on that. Even for a write, the disk will return something to indicate that the action has completed, so that can also be covered by just those two scenarios, and right now ZFS can lock the whole pool up if it's wait

Re: [zfs-discuss] "ZFS, Smashing Baby" a fake???

2008-11-25 Thread Ross Smith
PS. I think this also gives you a chance at making the whole problem much simpler. Instead of the hard question of "is this faulty", you're just trying to say "is it working right now?". In fact, I'm now wondering if the "waiting for a response" flag wouldn't be better as "possibly faulty". Tha

Re: [zfs-discuss] "ZFS, Smashing Baby" a fake???

2008-11-25 Thread Ross Smith
Hey Jeff, Good to hear there's work going on to address this. What did you guys think to my idea of ZFS supporting a "waiting for a response" status for disks as an interim solution that allows the pool to continue operation while it's waiting for FMA or the driver to fault the drive? I do appre

Re: [zfs-discuss] questions on zfs send,receive,backups

2008-11-03 Thread Ross Smith
>> If the file still existed, would this be a case of redirecting the >> file's top level block (dnode?) to the one from the snapshot? If the >> file had been deleted, could you just copy that one block? >> >> Is it that simple, or is there a level of interaction between files >> and snapshots tha

Re: [zfs-discuss] questions on zfs send,receive,backups

2008-11-03 Thread Ross Smith
> Snapshots are not replacements for traditional backup/restore features. > If you need the latter, use what is currently available on the market. > -- richard I'd actually say snapshots do a better job in some circumstances. Certainly they're being used that way by the desktop team: http://blogs.

Re: [zfs-discuss] questions on zfs send,receive,backups

2008-11-03 Thread Ross Smith
Hi Darren, That's storing a dump of a snapshot on external media, but files within it are not directly accessible. The work Tim et all are doing is actually putting a live ZFS filesystem on external media and sending snapshots to it. A live ZFS filesystem is far more useful (and reliable) than a

Re: [zfs-discuss] Disabling COMMIT at NFS level, or disabling ZIL on a per-filesystem basis

2008-10-23 Thread Ross Smith
No problem. I didn't use mirrored slogs myself, but that's certainly a step up for reliability. It's pretty easy to create a boot script to re-create the ramdisk and re-attach it to the pool too. So long as you use the same device name for the ramdisk you can add it each time with a simple "zpoo

Re: [zfs-discuss] Improving zfs send performance

2008-10-16 Thread Ross Smith
Oh dear god. Sorry folks, it looks like the new hotmail really doesn't play well with the list. Trying again in plain text: > Try to separate the two things: > > (1) Try /dev/zero -> mbuffer --- network ---> mbuffer> /dev/null > That should give you wirespeed I tried that already. It s

Re: [zfs-discuss] Improving zfs send performance

2008-10-16 Thread Ross Smith
> Try to separate the two things:> > (1) Try /dev/zero -> mbuffer --- network > ---> mbuffer > /dev/null > That should give you wirespeed I tried that already. It still gets just 10-11MB/s from this server. I can get zfs send / receive and mbuffer working at 30MB/s though from a couple of test

Re: [zfs-discuss] Improving zfs send performance

2008-10-15 Thread Ross Smith
I'm using 2008-05-07 (latest stable), am I right in assuming that one is ok? > Date: Wed, 15 Oct 2008 13:52:42 +0200 > From: [EMAIL PROTECTED] > To: [EMAIL PROTECTED]; zfs-discuss@opensolaris.org > Subject: Re: [zfs-discuss] Improving zfs send performance

Re: [zfs-discuss] Improving zfs send performance

2008-10-15 Thread Ross Smith
Thanks, that got it working. I'm still only getting 10MB/s, so it's not solved my problem - I've still got a bottleneck somewhere, but mbuffer is a huge improvement over standard zfs send / receive. It makes such a difference when you can actually see what's going on. --

Re: [zfs-discuss] ZFS Mirrors braindead?

2008-10-07 Thread Ross Smith
Oh cool, that's great news. Thanks Eric. > Date: Tue, 7 Oct 2008 11:50:08 -0700 > From: [EMAIL PROTECTED] > To: [EMAIL PROTECTED] > CC: zfs-discuss@opensolaris.org > Subject: Re: [zfs-discuss] ZFS Mirrors braindead? > > On Tue, Oct 07, 2008 at 11:42:57A

Re: [zfs-discuss] Scripting zfs send / receive

2008-09-26 Thread Ross Smith
Hi Mertol, Yes, I'm using zfs send -i to just send the changes rather than the whole thing. I'll have a think about your suggestion for deleting snapshots too, that does sound like a good idea. Unfortunately I won't be able to synchronise any applications with this script. It's backing up

Re: [zfs-discuss] Availability: ZFS needs to handle disk removal / driver failure better

2008-09-02 Thread Ross Smith
Thinking about it, we could make use of this too. The ability to add a remote iSCSI mirror to any pool without sacrificing local performance could be a huge benefit. > From: [EMAIL PROTECTED] > To: [EMAIL PROTECTED] > CC: [EMAIL PROTECTED]; zfs-discuss@opensolaris.org > Subject: Re: Availabilit

Re: [zfs-discuss] EMC - top of the table for efficiency, how well would ZFS do?

2008-08-31 Thread Ross Smith
EMAIL PROTECTED] To: [EMAIL PROTECTED] Subject: Re: [zfs-discuss] EMC - top of the table for efficiency, how well would ZFS do? CC: zfs-discuss@opensolaris.org On Sun, Aug 31, 2008 at 10:39 AM, Ross Smith <[EMAIL PROTECTED]> wrote: Hey Tim, I'll admit I just quoted the blog w

Re: [zfs-discuss] EMC - top of the table for efficiency, how well would ZFS do?

2008-08-31 Thread Ross Smith
Hey Tim, I'll admit I just quoted the blog without checking, I seem to remember the sales rep I spoke to recommending putting aside 20-50% of my disk for snapshots. Compared to ZFS where I don't need to reserve any space it feels very old fashioned. With ZFS, snapshots just take up as much s

Re: [zfs-discuss] Availability: ZFS needs to handle disk removal / driver failure better

2008-08-30 Thread Ross Smith
Triple mirroring you say? That'd be me then :D The reason I really want to get ZFS timeouts sorted is that our long term goal is to mirror that over two servers too, giving us a pool mirrored across two servers, each of which is actually a zfs iscsi volume hosted on triply mirrored disks. Oh

Re: [zfs-discuss] Availability: ZFS needs to handle disk removal / driver failure better

2008-08-28 Thread Ross Smith
Hi guys, Bob, my thought was to have this timeout as something that can be optionally set by the administrator on a per pool basis. I'll admit I was mainly thinking about reads and hadn't considered the write scenario, but even having thought about that it's still a feature I'd like. After a

Re: [zfs-discuss] ZFS automatic snapshots 0.11 Early Access

2008-08-27 Thread Ross Smith
That sounds absolutely perfect Tim, thanks. Yes, we'll be sending these to other zfs filesystems, although I haven't looked at the send/receive part of your service yet. What I'd like to do is stage the send/receive as files on an external disk, and then receive them remotely from that. I'v

Re: [zfs-discuss] Best layout for 15 disks?

2008-08-22 Thread Ross Smith
Yup, you got it, and an 8 disk raid-z2 array should still fly for a home system :D I'm guessing you're on gigabit there? I don't see you having any problems hitting the bandwidth limit on it. Ross > Date: Fri, 22 Aug 2008 11:11:21 -0700 > From: [EMAIL PROTECTED] > To: [EMAIL PROTECTED] > Su

Re: [zfs-discuss] FW: Supermicro AOC-SAT2-MV8 hang when drive removed

2008-08-20 Thread Ross Smith
> > Without fail, cfgadm changes the status from "disk" to "sata-port" when I > > unplug a device attached to port 6 or 7, but most of the time unplugging > > disks 0-5 results in no change in cfgadm, until I also attach disk 6 or 7. > > That does seem inconsistent, or at least, it's not what I'd

Re: [zfs-discuss] FW: Supermicro AOC-SAT2-MV8 hang when drive removed

2008-08-15 Thread Ross Smith
Oh god no, I'm already learning three new operating systems, now is not a good time to add a fourth. Ross<-- Windows admin now working with Ubuntu, OpenSolaris and ESX Date: Fri, 15 Aug 2008 10:07:31 -0500From: [EMAIL PROTECTED]: [EMAIL PROTECTED]: Re: [zfs-discuss] FW: Supermicro AOC-S

Re: [zfs-discuss] Zpool import not working - I broke my pool...

2008-08-06 Thread Ross Smith
Hmm... got a bit more information for you to add to that bug I think. Zpool import also doesn't work if you have mirrored log devices and either one of them is offline. I created two ramdisks with: # ramdiskadm -a rc-pool-zil-1 256m # ramdiskadm -a rc-pool-zil-2 256m And added them to the p

Re: [zfs-discuss] Zpool import not working - I broke my pool...

2008-08-05 Thread Ross Smith
> PROTECTED]> CC: [EMAIL PROTECTED]; zfs-discuss@opensolaris.org> > Ross Smith > wrote:> > Just a thought, before I go and wipe this zpool, is there any way > to > > manually recreate the /etc/zfs/zpool.cache file?> > Do you have a copy > in a snapshot? ZF

Re: [zfs-discuss] Zpool import not working - I broke my pool...

2008-08-05 Thread Ross Smith
Just a thought, before I go and wipe this zpool, is there any way to manually recreate the /etc/zfs/zpool.cache file? Ross> Date: Mon, 4 Aug 2008 10:42:43 -0600> From: [EMAIL PROTECTED]> Subject: Re: [zfs-discuss] Zpool import not working - I broke my pool...> To: [EMAIL PROTECTED]; [EMAIL PR

Re: [zfs-discuss] are these errors dangerous

2008-08-03 Thread Ross Smith
Hi Matt, If it's all 3 disks, I wouldn't have thought it likely to be disk errors, and I don't think it's a ZFS fault as such. You might be better posting the question in the storage or help forums to see if anybody there can shed more light on this. Ross > Date: Sun, 3 Aug 2008 16:48:03 +

Re: [zfs-discuss] Replacing the boot HDDs in x4500

2008-08-01 Thread Ross Smith
Sorry Ian, I was posting on the forum and missed the word "disks" from my previous post. I'm still not used to Sun's mutant cross of a message board / mailing list. Ross > Date: Fri, 1 Aug 2008 21:08:08 +1200> From: [EMAIL PROTECTED]> To: [EMAIL > PROTECTED]> CC: zfs-discuss@opensolaris.org>

Re: [zfs-discuss] Can I trust ZFS?

2008-07-31 Thread Ross Smith
Hey Brent, On the Sun hardware like the Thumper you do get a nice bright blue "ready to remove" led as soon as you issue the "cfgadm -c unconfigure xxx" command. On other hardware it takes a little more care, I'm labelling our drive bays up *very* carefully to ensure we always remove the rig

Re: [zfs-discuss] Supermicro AOC-SAT2-MV8 hang when drive removed

2008-07-31 Thread Ross Smith
Gave up on ZFS ever recovering. A shutdown attempt hung as expected. I hard-reset the computer. Ross > Date: Wed, 30 Jul 2008 11:17:08 -0700> From: [EMAIL PROTECTED]> Subject: Re: > [zfs-discuss] Supermicro AOC-SAT2-MV8 hang when drive removed> To: [EMAIL > PROTECTED]&

Re: [zfs-discuss] Supermicro AOC-SAT2-MV8 hang when drive removed

2008-07-30 Thread Ross Smith
I agree that device drivers should perform the bulk of the fault monitoring, however I disagree that this absolves ZFS of any responsibility for checking for errors. The primary goal of ZFS is to be a filesystem and maintain data integrity, and that entails both reading and writing data to the

Re: [zfs-discuss] Supermicro AOC-SAT2-MV8 hang when drive removed

2008-07-29 Thread Ross Smith
A little more information today. I had a feeling that ZFS would continue quite some time before giving an error, and today I've shown that you can carry on working with the filesystem for at least half an hour with the disk removed. I suspect on a system with little load you could carry on wo

Re: [zfs-discuss] Supermicro AOC-SAT2-MV8 hang when drive removed

2008-07-28 Thread Ross Smith
ubject: RE: [zfs-discuss] Supermicro AOC-SAT2-MV8 hang when > drive removed> > On Mon, 28 Jul 2008, Ross Smith wrote:> > >> > "File > Browser" is the name of the program that Solaris opens when > > you open > "Computer" on the desktop. It'

Re: [zfs-discuss] Supermicro AOC-SAT2-MV8 hang when drive removed

2008-07-28 Thread Ross Smith
snv_91. I downloaded snv_94 today so I'll be testing with that tomorrow. > Date: Mon, 28 Jul 2008 09:58:43 -0700> From: [EMAIL PROTECTED]> Subject: Re: > [zfs-discuss] Supermicro AOC-SAT2-MV8 hang when drive removed> To: [EMAIL > PROTECTED]> > Which OS and revision?> -- richard> > > Ross wrote:

Re: [zfs-discuss] Supermicro AOC-SAT2-MV8 hang when drive removed

2008-07-28 Thread Ross Smith
"File Browser" is the name of the program that Solaris opens when you open "Computer" on the desktop. It's the default graphical file manager. It does eventually stop copying with an error, but it takes a good long while for ZFS to throw up that error, and even when it does, the pool doesn't

Re: [zfs-discuss] J4500 device renumbering

2008-07-15 Thread Ross Smith
It sounds like you might be interested to read up on Eric Schrock's work. I read today about some of the stuff he's been doing to bring integrated fault management to Solaris: http://blogs.sun.com/eschrock/entry/external_storage_enclosures_in_solaris His last paragraph is great to see, Sun real

[zfs-discuss] FW: please help with raid / failure / rebuild calculations

2008-07-15 Thread Ross Smith
bits vs bytes D'oh! again. It's a good job I don't do these calculations professionally. :-)> Date: Tue, 15 Jul 2008 02:30:33 -0400> From: [EMAIL PROTECTED]> To: [EMAIL PROTECTED]> Subject: Re: [zfs-discuss] please help with raid / failure / rebuild calculations> CC: zfs-discuss@opensola