Re: [zfs-discuss] 'zfs recv' is very slow
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 Brent Jones: > My results are much improved, on the order of 5-100 times faster > (either over Mbuffer or SSH). this is good news - although not quite soon enough for my current 5TB zfs send ;-) have you tested if this also improves the performance of incremental sends? - river. -BEGIN PGP SIGNATURE- iD4DBQFJeZ8GIXd7fCuc5vIRAjPeAJ9Ed9AdwcTWdqkAizVqIPp1qUyNtACY8DA6 a9zguVE8f/TZ6pH/Haa4/Q== =vf9T -END PGP SIGNATURE- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Is scrubbing "safe" in 101b? (OpenSolaris 2008.11)
I thought I'd noticed that my crashes tended to occur when I was running a scrub, and saw at least one open bug that was scrub-related that could cause such a crash. However, I eventually tracked my problem down (as it got worse) to a bad piece of memory (been nearly a week since I replaced the memory, and no more problems). Which leaves me wondering, how safe is running a scrub? Scrub is one of the things that made ZFS so attractive to me, and my automatic reaction when I first hook up the data disks during a recovery is "run a scrub!". -- David Dyer-Bennet, d...@dd-b.net; http://dd-b.net/ Snapshots: http://dd-b.net/dd-b/SnapshotAlbum/data/ Photos: http://dd-b.net/photography/gallery/ Dragaera: http://dragaera.info ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Is scrubbing "safe" in 101b? (OpenSolaris 2008.11)
>I thought I'd noticed that my crashes tended to occur when I was running a >scrub, and saw at least one open bug that was scrub-related that could >cause such a crash. However, I eventually tracked my problem down (as it >got worse) to a bad piece of memory (been nearly a week since I replaced >the memory, and no more problems). I had a problem and I think that it was a bad motherboard; it too panic'ed during scrub and it even said that "scrub finished" (it went from 50% to finished, immediately). I replaced the system (motherboard, harddisk) and I re-ran scrub; no problem with scrub that time but it took the amount it should have taken. >Which leaves me wondering, how safe is running a scrub? Scrub is one of >the things that made ZFS so attractive to me, and my automatic reaction >when I first hook up the data disks during a recovery is "run a scrub!". If your memory is bad, anything can happen. A scrub can rewrite bad data; but it can be the case that the disk is fine but the memory is bad. Then, if the data is replicated it can be copied and rewritten; it is then possible to write incorrect data (and if they need to recompute the checksum, then oops) Casper ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] SSD drives in Sun Fire X4540 or X4500 for dedicated ZIL device
On Thu, 22 Jan 2009, Ross wrote: > However, now I've written that, Sun use SATA (SAS?) SSD's in their > high end fishworks storage, so I guess it definately works for some > use cases. But the "fishworks" (Fishworks is a development team, not a product) write cache device is not based on FLASH. It is based on DRAM. The difference is like night and day. Apparently there can also be a read cache which is based on FLASH. Bob == Bob Friesenhahn bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/ GraphicsMagick Maintainer,http://www.GraphicsMagick.org/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] SSD drives in Sun Fire X4540 or X4500 for dedicated ZIL device
That's my understanding too. One (STEC?) drive as a write cache, basically a write optimised SSD. And cheaper, larger, read optimised SSD's for the read cache. I thought it was an odd strategy until I read into SSD's a little more and realised you really do have to think about your usage cases with these. SSD's are very definitely not all alike. On Fri, Jan 23, 2009 at 4:33 PM, Greg Mason wrote: > If i'm not mistaken (and somebody please correct me if i'm wrong), the Sun > 7000 series storage appliances (the Fishworks boxes) use enterprise SSDs, > with dram caching. One such product is made by STEC. > > My understanding is that the Sun appliances use one SSD for the ZIL, and one > as a read cache. For the 7210 (which is basically a Sun Fire X4540), that > gives you 46 disks and 2 SSDs. > > -Greg > > > Bob Friesenhahn wrote: >> >> On Thu, 22 Jan 2009, Ross wrote: >> >>> However, now I've written that, Sun use SATA (SAS?) SSD's in their high >>> end fishworks storage, so I guess it definately works for some use cases. >> >> But the "fishworks" (Fishworks is a development team, not a product) write >> cache device is not based on FLASH. It is based on DRAM. The difference is >> like night and day. Apparently there can also be a read cache which is based >> on FLASH. >> >> Bob >> == >> Bob Friesenhahn >> bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/ >> GraphicsMagick Maintainer,http://www.GraphicsMagick.org/ >> >> ___ >> zfs-discuss mailing list >> zfs-discuss@opensolaris.org >> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss >> >> > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Is scrubbing "safe" in 101b? (OpenSolaris 2008.11)
On Fri, January 23, 2009 09:52, casper@sun.com wrote: >>Which leaves me wondering, how safe is running a scrub? Scrub is one of >>the things that made ZFS so attractive to me, and my automatic reaction >>when I first hook up the data disks during a recovery is "run a scrub!". > > > If your memory is bad, anything can happen. A scrub can rewrite bad > data; but it can be the case that the disk is fine but the memory is > bad. Then, if the data is replicated it can be copied and rewritten; > it is then possible to write incorrect data (and if they need to recompute > the checksum, then oops) The memory was ECC, so it *should* have mostly detected problems early enough to avoid writing bad data. And so far nothing has been detected as bad in the pool during light use. But I haven't yet run a scrub since fixing the memory, so I have no idea what horrors may be lurking in wait. The pool is two mirror vdevs, and then I have two backups on external hard drives, and then I have two sets of optical disks of the photos, one of them off-site (I'd lose several months of photos if I had to fall back to the optical disks, I'm a bit behind there). So I'm not yet in great fear of actually losing anything, and have very little risk of actually losing a LOT. But what I'm wondering is, are there known bugs in 101b that make scrubbing inadvisable with that code? I'd love to *find out* what horrors may be lurking. -- David Dyer-Bennet, d...@dd-b.net; http://dd-b.net/ Snapshots: http://dd-b.net/dd-b/SnapshotAlbum/data/ Photos: http://dd-b.net/photography/gallery/ Dragaera: http://dragaera.info ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] SSD drives in Sun Fire X4540 or X4500 for dedicated ZIL device
This is correct, and you can read about it here: http://blogs.sun.com/ahl/entry/fishworks_launch Adam On Fri, Jan 23, 2009 at 05:03:57PM +, Ross Smith wrote: > That's my understanding too. One (STEC?) drive as a write cache, > basically a write optimised SSD. And cheaper, larger, read optimised > SSD's for the read cache. > > I thought it was an odd strategy until I read into SSD's a little more > and realised you really do have to think about your usage cases with > these. SSD's are very definitely not all alike. > > > On Fri, Jan 23, 2009 at 4:33 PM, Greg Mason wrote: > > If i'm not mistaken (and somebody please correct me if i'm wrong), the Sun > > 7000 series storage appliances (the Fishworks boxes) use enterprise SSDs, > > with dram caching. One such product is made by STEC. > > > > My understanding is that the Sun appliances use one SSD for the ZIL, and one > > as a read cache. For the 7210 (which is basically a Sun Fire X4540), that > > gives you 46 disks and 2 SSDs. > > > > -Greg > > > > > > Bob Friesenhahn wrote: > >> > >> On Thu, 22 Jan 2009, Ross wrote: > >> > >>> However, now I've written that, Sun use SATA (SAS?) SSD's in their high > >>> end fishworks storage, so I guess it definately works for some use cases. > >> > >> But the "fishworks" (Fishworks is a development team, not a product) write > >> cache device is not based on FLASH. It is based on DRAM. The difference > >> is > >> like night and day. Apparently there can also be a read cache which is > >> based > >> on FLASH. > >> > >> Bob > >> == > >> Bob Friesenhahn > >> bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/ > >> GraphicsMagick Maintainer,http://www.GraphicsMagick.org/ > >> > >> ___ > >> zfs-discuss mailing list > >> zfs-discuss@opensolaris.org > >> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > >> > >> > > > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Adam Leventhal, Fishworks http://blogs.sun.com/ahl ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] SSD drives in Sun Fire X4540 or X4500 for dedicated ZIL device
If i'm not mistaken (and somebody please correct me if i'm wrong), the Sun 7000 series storage appliances (the Fishworks boxes) use enterprise SSDs, with dram caching. One such product is made by STEC. My understanding is that the Sun appliances use one SSD for the ZIL, and one as a read cache. For the 7210 (which is basically a Sun Fire X4540), that gives you 46 disks and 2 SSDs. -Greg Bob Friesenhahn wrote: > On Thu, 22 Jan 2009, Ross wrote: > >> However, now I've written that, Sun use SATA (SAS?) SSD's in their >> high end fishworks storage, so I guess it definately works for some >> use cases. > > But the "fishworks" (Fishworks is a development team, not a product) > write cache device is not based on FLASH. It is based on DRAM. The > difference is like night and day. Apparently there can also be a read > cache which is based on FLASH. > > Bob > == > Bob Friesenhahn > bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/ > GraphicsMagick Maintainer,http://www.GraphicsMagick.org/ > > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Is scrubbing "safe" in 101b? (OpenSolaris 2008.11)
* David Dyer-Bennet (d...@dd-b.net) wrote: > > On Fri, January 23, 2009 09:52, casper@sun.com wrote: > > >>Which leaves me wondering, how safe is running a scrub? Scrub is one of > >>the things that made ZFS so attractive to me, and my automatic reaction > >>when I first hook up the data disks during a recovery is "run a scrub!". > > > > > > If your memory is bad, anything can happen. A scrub can rewrite bad > > data; but it can be the case that the disk is fine but the memory is > > bad. Then, if the data is replicated it can be copied and rewritten; > > it is then possible to write incorrect data (and if they need to recompute > > the checksum, then oops) > > The memory was ECC, so it *should* have mostly detected problems early > enough to avoid writing bad data. And so far nothing has been detected as > bad in the pool during light use. But I haven't yet run a scrub since > fixing the memory, so I have no idea what horrors may be lurking in wait. > > The pool is two mirror vdevs, and then I have two backups on external hard > drives, and then I have two sets of optical disks of the photos, one of > them off-site (I'd lose several months of photos if I had to fall back to > the optical disks, I'm a bit behind there). So I'm not yet in great fear > of actually losing anything, and have very little risk of actually losing > a LOT. > > But what I'm wondering is, are there known bugs in 101b that make > scrubbing inadvisable with that code? I'd love to *find out* what horrors > may be lurking. There's nothing in the release notes for 2008.11 (based on 101b) about issues running scrub. I've been using 101b for some time now and haven't seen or heard of any issues running scrub. There's always bugs. But I'm pretty certain there isn't a known 'zfs scrub is inadvisable under any and all conditions' bug laying about. I've certainly not heard of such a thing (and it would be pretty big news for 2008.11 if true). Cheers, -- Glenn ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Is scrubbing "safe" in 101b? (OpenSolaris 2008.11)
On Fri, January 23, 2009 12:01, Glenn Lagasse wrote: > * David Dyer-Bennet (d...@dd-b.net) wrote: >> But what I'm wondering is, are there known bugs in 101b that make >> scrubbing inadvisable with that code? I'd love to *find out* what >> horrors >> may be lurking. > > There's nothing in the release notes for 2008.11 (based on 101b) about > issues running scrub. I've been using 101b for some time now and > haven't seen or heard of any issues running scrub. > > There's always bugs. But I'm pretty certain there isn't a known 'zfs > scrub is inadvisable under any and all conditions' bug laying about. > I've certainly not heard of such a thing (and it would be pretty big > news for 2008.11 if true). Thanks. I appreciate the qualifications, and agree that there are, indeed, always bugs. Okay; I think I'll make one last check of my own through the buglist (so it won't be your fault if I muck up my data :-)), and if nothing turns up that scares me I will run the scrub and see what happens. Hmmm; maybe update ONE of the two backups first. It's at times like this when I really miss relatively cheap backup tapes (current tapes that I've seen aren't relatively cheap). -- David Dyer-Bennet, d...@dd-b.net; http://dd-b.net/ Snapshots: http://dd-b.net/dd-b/SnapshotAlbum/data/ Photos: http://dd-b.net/photography/gallery/ Dragaera: http://dragaera.info ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Bug report: disk replacement confusion
Yes, everything seems to be fine, but that was still scary, and the fix was not completely obvious. At the very least, I would suggest adding text such as the following to the page at http://www.sun.com/msg/ZFS-8000-FD : When physically replacing the failed device, it is best to use the same controller port, so that the new device will have the same device name as the failed one. -- Scott -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS Import Problem
--- On Tue, 12/30/08, Andrew Gabriel wrote: >If you were doing a rolling upgrade, I suspect the old disks are all >horribly out of sync with each other? > >If that is the problem, then if the filesystem(s) have a snapshot that >existed when all the old disks were still online, I wonder if it might >be possible to roll them back to it by hand, so it looks like the >current live filesystem? I don't know if this is possible with zdb. What I meant is that I rolled the new drives in one at a time... replace a drive, let it resilver, replace another, let it resilver. etc. The filesystems were not mounted at the time so the data should have been static. The only thing that changed (from what I can tell) is the zpool labels. Thanks, Michael ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS Import Problem
Yes, but I can't export a pool that has never been imported. These drives are no longer connected to their original system, and at this point, when I connect them to their original system, the results are the same. Thanks, Michael --- On Tue, 12/30/08, Weldon S Godfrey 3 wrote: > > Did you try zpool export 1st? ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs smb public share, files created not public
Roger wrote: > Hi! > Im running popensolaris b101 and ive made a zfs pool called tank and an fs > inside of it tank/public, ive shared it with smb. > > zfs set sharesmb=on tank/public > > im using solaris smb and not samba. > > The problem is this. When i connect and create a file its readable to > everyone but it should be 777. So if i create a file someone else cant erase > it. How do i do to get this to work? > > cheers > -gob It would be better to ask this question in cifs-disc...@opensolaris.org I believe the following should get you the results you want. # chmod A=everyone@:full_set:fd:allow /tank/public then all cifs created files will have full access and anyone can read,write,delete,... -Mark ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] help please - Zpool import : I/O error
Hi, I have a big problem with my ZFS drive. After a kernel panic, I cannot import the pool anymore : -- => zpool status no pools available => zpool list no pools available -- => zpool import pool: ZFS id: 9004030332584099627 state: FAULTED status: The pool metadata is corrupted. action: The pool cannot be imported due to damaged devices or data. The pool may be active on on another system, but can be imported using the '-f' flag. see: http://www.sun.com/msg/ZFS-8000-72 config: ZFSFAULTED corrupted data disk2s2 ONLINE pool: ZFS id: 5050959592823553345 state: FAULTED status: The pool was last accessed by another system. action: The pool cannot be imported due to damaged devices or data. see: http://www.sun.com/msg/ZFS-8000-EY config: ZFS UNAVAIL insufficient replicas disk1 UNAVAIL cannot open -- => zpool import -f 9004030332584099627 cannot import 'ZFS': I/O error -- I am despaired, all my data are on this drive and I hadn't time to make a backup. Is there anything I can do ? Thank you very much for your help, Mathieu ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] OpenSolaris Storage Summit Feb 23, 2009 San Francisco
Hi All, sorry for all the duplicates. Feel free to pass on to other interested parties. The OpenSolaris Storage Community is holding a Storage Summit on February 23 at the Grand Hyatt San Francisco, prior to the FAST conference. The registration wiki is here: https://wikis.sun.com/display/OpenSolaris/OpenSolaris+Storage+Summit+200902 We are still sorting out keynote speakers and the exact format. On the wiki it's possible to suggest topics. We will also be hosting open Community Calls on 1/26 and 2/17, details to follow. If you are interested in helping organise or anything related to this event contact me: p...@sun.com So once again the details: OpenSolaris Storage Summit February 23, 2009 Grand Hyatt San Francisco 345 Stockton Street, San Francisco, California, USA 94108 thanks and see you there! peter ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] moving zfs root rpool between systems
I have two 280R systems. System A has Solaris 10u6, and its (2) drives are configured as a ZFS rpool, and are mirrored. I would like to pull these drives, and move them to my other 280, system B, which is currently hard drive-less. Although unsupported by Sun, I have done this before without issue using UFS. I have also successfully moved zpool's from system to system, using zpool export & zpool import. This is unique, at least to me as the root file system and the OS are in rpool. Is this possible? John ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] ZFS upgrade mangled my share
I was having CIFs problems on my Mac so I upgrade to build 105. After getting all my shares populated with data I ran zpool scrub on the raidz array and it told me the version was out of date so I upgraded. One of my shares is now inaccessible and I cannot even delete it :( > > r...@bitchko:/nas1/backups# zfs list -r /nas1/backups/ > NAME USED AVAIL REFER MOUNTPOINT > nas1/backups 33.3K 1.04T 33.3K /nas1/backups > r...@bitchko:/# zfs destroy /nas1/backups/ > cannot open '/nas1/backups/': invalid dataset name > r...@bitchko:/nas1# rm -Rf backups/ > rm: cannot remove directory `backups/': Device busy > This is pretty scary for something that's storing valuable data. Any ideas how I can destroy the dataset and recreate it? Thanks Colin ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] ZFS encryption?? - [Fwd: [osol-announce] SXCE Build 105 available
It was rumored that Nevada build 105 would have ZFS encrypted file systems integrated into the main source. In reviewing the Change logs (URL's below) I did not see anything mentioned that this had come to pass. Its going to be another week before I have a chance to play with b105. Does anyone know specifically if b105 has ZFS encryption? Thanks, Jerry Original Message Subject: [osol-announce] SXCE Build 105 available Date: Fri, 09 Jan 2009 08:58:40 -0800 Please find the links to SXCE Build 105 at: http://www.opensolaris.org/os/downloads/ This is still a DVD only release. - wget work around: http://wikis.sun.com/pages/viewpage.action?pageId=28448383 --- Changelogs: ON (The kernel, drivers, and core utilities): http://dlc.sun.com/osol/on/downloads/b105/on-changelog-b105.html X Window System: http://opensolaris.org/os/community/x_win/changelogs/changelogs-nv_100/ - Derek ___ opensolaris-announce mailing list opensolaris-annou...@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/opensolaris-announce ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] mirror rpool
Richard Elling wrote: > mijenix wrote: >> yes, that's the way zpool likes it >> >> I think I've to understand how (Open)Solaris create disks or how the >> partition thing works under OSol. Do you know any guide or howto? >> > > We've tried to make sure the ZFS Admin Guide covers these things, including > the procedure for mirroring the root pool by hand (Jumpstart can do it > automatically :-) > http://www.opensolaris.org/os/community/zfs/docs/zfsadmin.pdf > > If it does not meet your needs, please let us know. > -- richard > ok, ok I posted a newbie question, I got it. Holy RTFM ;-) Thanks for the links and suggestions. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Changing from ZFS back to HFS+
Hi All, Since switching to ZFS I get a lot of ³beach balls². I think for productivity sake I should switch back to HFS+. My home directory was on this ZFS parition. I backed up my data to another drive and tried using Disk Utility to select my ZFS partition, un-mount it and format just that partition HFS+. This failed. I manually unmounted my ZFS Partition using zfs unmount and it unmounted but Disk Utility still gives an error. Can I do this or am I looking at re-partitioning the whole drive and starting from scratch with Leopard, BootCamp partition and an HFS+ partition for my data? Or, is a new release coming out that might relieve me of some of these issues? Thanks, -Jason ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] replace same sized disk fails with too small error
It also wouldn't be a bad idea for ZFS to also verify drives designated as hot spares in fact have sufficient capacity to be compatible replacements for particular configurations, prior to actually being critically required (as if drives otherwise appearing to have equivalent capacity may not, it wouldn't be a nice thing to first discover upon attempted replacement of a failed drive). ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zpool import fails to find pool
Hi all, I moved from Sol 10 Update4 to update 6. Before doing this I exported both of my zpools, and replace the discs containing the ufs root on with two new discs (these discs did not have any zpool /zfs info and are raid mirrored in hardware) Once I had installed update6 I did a zpool import, but it only shows (and was able to) import one of the two pools. Looking at dmesg it appears as though I have lost one drive (although they are all physically accounted for). However both my pools are running raidz2 so even if I had lost one disc (another matter) I would still be able to expect to import them. If I swap back to my ufs root discs for update 4 then I can see the pool with import... I'm stuck as to what to do next? Anyone any ideas? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS upgrade mangled my share
Colin Johnson wrote: > I was having CIFs problems on my Mac so I upgrade to build 105. > After getting all my shares populated with data I ran zpool scrub on > the raidz array and it told me the version was out of date so I > upgraded. > > One of my shares is now inaccessible and I cannot even delete it :( > >> r...@bitchko:/nas1/backups# zfs list -r /nas1/backups/ >> NAME USED AVAIL REFER MOUNTPOINT >> nas1/backups 33.3K 1.04T 33.3K /nas1/backups > >> r...@bitchko:/# zfs destroy /nas1/backups/ >> cannot open '/nas1/backups/': invalid dataset name > There's a typo there, you would have to do zfs destroy nas1/backups Unfortunately, you can't use the mountpoint, you have to name the dataset. -tim >> r...@bitchko:/nas1# rm -Rf backups/ >> rm: cannot remove directory `backups/': Device busy >> > > > This is pretty scary for something that's storing valuable data. > > Any ideas how I can destroy the dataset and recreate it? > > Thanks > > Colin > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS encryption?? - [Fwd: [osol-announce] SXCE Build 105 available
Jerry K wrote: > It was rumored that Nevada build 105 would have ZFS encrypted file > systems integrated into the main source. > > In reviewing the Change logs (URL's below) I did not see anything > mentioned that this had come to pass. Its going to be another week > before I have a chance to play with b105. > > Does anyone know specifically if b105 has ZFS encryption? > It does not. -tim > Thanks, > > Jerry > > > Original Message > Subject: [osol-announce] SXCE Build 105 available > Date: Fri, 09 Jan 2009 08:58:40 -0800 > > > Please find the links to SXCE Build 105 at: > > http://www.opensolaris.org/os/downloads/ > > This is still a DVD only release. > > - > wget work around: > > http://wikis.sun.com/pages/viewpage.action?pageId=28448383 > > --- > Changelogs: > > ON (The kernel, drivers, and core utilities): > > http://dlc.sun.com/osol/on/downloads/b105/on-changelog-b105.html > > X Window System: > > http://opensolaris.org/os/community/x_win/changelogs/changelogs-nv_100/ > > - Derek > ___ > opensolaris-announce mailing list > opensolaris-annou...@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/opensolaris-announce > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import fails to find pool
James Nord wrote: > Hi all, > > I moved from Sol 10 Update4 to update 6. > > Before doing this I exported both of my zpools, and replace the discs > containing the ufs root on with two new discs (these discs did not have any > zpool /zfs info and are raid mirrored in hardware) > > Once I had installed update6 I did a zpool import, but it only shows (and was > able to) import one of the two pools. > > Looking at dmesg it appears as though I have lost one drive (although they > are all physically accounted for). However both my pools are running raidz2 > so even if I had lost one disc (another matter) I would still be able to > expect to import them. > > If I swap back to my ufs root discs for update 4 then I can see the pool with > import... > > I'm stuck as to what to do next? Anyone any ideas? > Look at the list of disks in the output of format(1M) in Update 4, compared with Update 6. Are there any disks missing? (It shouldn't matter if the order is different.) If there is a disk missing, is it on a different type of controller from the others? (Just wondering if a disk driver might have stopped working for some reason between the two releases.) -- Andrew ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool status -x strangeness
A little gotcha that I found in my 10u6 update process was that 'zpool upgrade [poolname]' is not the same as 'zfs upgrade [poolname]/[filesystem(s)]' What does 'zfs upgrade' say? I'm not saying this is the source of your problem, but it's a detail that seemed to affect stability for me. On Thu, Jan 22, 2009 at 7:25 AM, Ben Miller wrote: > The pools are upgraded to version 10. Also, this is on Solaris 10u6. > > # zpool upgrade > This system is currently running ZFS pool version 10. > > All pools are formatted using this version. > > Ben > >> What's the output of 'zfs upgrade' and 'zpool >> upgrade'? (I'm just >> curious - I had a similar situation which seems to be >> resolved now >> that I've gone to Solaris 10u6 or OpenSolaris >> 2008.11). >> >> >> >> On Wed, Jan 21, 2009 at 2:11 PM, Ben Miller >> wrote: >> > Bug ID is 6793967. >> > >> > This problem just happened again. >> > % zpool status pool1 >> > pool: pool1 >> > state: DEGRADED >> > scrub: resilver completed after 0h48m with 0 >> errors on Mon Jan 5 12:30:52 2009 >> > config: >> > >> >NAME STATE READ WRITE CKSUM >> >pool1 DEGRADED 0 0 0 >> > raidz2 DEGRADED 0 0 0 >> >c4t8d0s0 ONLINE 0 0 0 >> >c4t9d0s0 ONLINE 0 0 0 >> >c4t10d0s0 ONLINE 0 0 0 >> >c4t11d0s0 ONLINE 0 0 0 >> >c4t12d0s0 REMOVED 0 0 0 >> >c4t13d0s0 ONLINE 0 0 0 >> > >> > errors: No known data errors >> > >> > % zpool status -x >> > all pools are healthy >> > % >> > # zpool online pool1 c4t12d0s0 >> > % zpool status -x >> > pool: pool1 >> > state: ONLINE >> > status: One or more devices is currently being >> resilvered. The pool will >> >continue to function, possibly in a degraded >> state. >> > action: Wait for the resilver to complete. >> > scrub: resilver in progress for 0h0m, 0.12% done, >> 2h38m to go >> > config: >> > >> >NAME STATE READ WRITE CKSUM >> >pool1 ONLINE 0 0 0 >> > raidz2 ONLINE 0 0 0 >> >c4t8d0s0 ONLINE 0 0 0 >> >c4t9d0s0 ONLINE 0 0 0 >> >c4t10d0s0 ONLINE 0 0 0 >> >c4t11d0s0 ONLINE 0 0 0 >> >c4t12d0s0 ONLINE 0 0 0 >> >c4t13d0s0 ONLINE 0 0 0 >> > >> > errors: No known data errors >> > % >> > >> > Ben >> > > -- > This message posted from opensolaris.org > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Raidz1 faulted with single bad disk. Requesting assistance.
I've seen reports of a recent Seagate firmware update bricking drives again. What's the output of 'zpool import' from the LiveCD? It sounds like more than 1 drive is dropping off. On Thu, Jan 22, 2009 at 10:52 PM, Brad Hill wrote: >> I would get a new 1.5 TB and make sure it has the new >> firmware and replace >> c6t3d0 right away - even if someone here comes up >> with a magic solution, you >> don't want to wait for another drive to fail. > > The replacement disk showed up today but I'm unable to replace the one marked > UNAVAIL: > > r...@blitz:~# zpool replace tank c6t3d0 > cannot open 'tank': pool is unavailable > >> I would in this case also immediately export the pool (to prevent any >> write attempts) and see about a firmware update for the failed drive >> (probably need windows for this). > > While I didn't export first, I did boot with a livecd and tried to force the > import with that: > > r...@opensolaris:~# zpool import -f tank > internal error: Bad exchange descriptor > Abort (core dumped) > > Hopefully someone on this list understands what situation I am in and how to > resolve it. Again, many thanks in advance for any suggestions you all have to > offer. > -- > This message posted from opensolaris.org > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Changing from ZFS back to HFS+
This is primarily a list for OpenSolaris ZFS - OS X is a little different ;) However, I think you need to do a 'sudo zpool destroy [poolname]' from Terminal.app Be warned, you can't go back once you have done this! On Sun, Jan 18, 2009 at 4:42 PM, Jason Todd Slack-Moehrle wrote: > Hi All, > > Since switching to ZFS I get a lot of "beach balls". I think for > productivity sake I should switch back to HFS+. My home directory was on > this ZFS parition. > > I backed up my data to another drive and tried using Disk Utility to select > my ZFS partition, un-mount it and format just that partition HFS+. This > failed. I manually unmounted my ZFS Partition using zfs unmount and it > unmounted but Disk Utility still gives an error. > > Can I do this or am I looking at re-partitioning the whole drive and > starting from scratch with Leopard, BootCamp partition and an HFS+ partition > for my data? > > Or, is a new release coming out that might relieve me of some of these > issues? > > Thanks, > -Jason > > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] replace same sized disk fails with too small error
+1 On Thu, Jan 22, 2009 at 11:12 PM, Paul Schlie wrote: > It also wouldn't be a bad idea for ZFS to also verify drives designated as > hot spares in fact have sufficient capacity to be compatible replacements > for particular configurations, prior to actually being critically required > (as if drives otherwise appearing to have equivalent capacity may not, it > wouldn't be a nice thing to first discover upon attempted replacement of a > failed drive). > > > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS encryption?? - [Fwd: [osol-announce] SXCE Build 105 available
> Does anyone know specifically if b105 has ZFS encryption? IIRC it has been pushed back to b109. -mg signature.asc Description: OpenPGP digital signature ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] moving zfs root rpool between systems
On Tue, Jan 6, 2009 at 4:23 PM, John Arden wrote: > I have two 280R systems. System A has Solaris 10u6, and its (2) drives > are configured as a ZFS rpool, and are mirrored. I would like to pull > these drives, and move them to my other 280, system B, which is > currently hard drive-less. > > Although unsupported by Sun, I have done this before without issue using > UFS. > > I have also successfully moved zpool's from system to system, using > zpool export & zpool import. This is unique, at least to me as the root > file system and the OS are in rpool. > > Is this possible? > > John > Same basic concept as the UFS move. I can't think of anything that would prevent you from doing so. More to the point, if you pull the drives from the current 280, stick them into the new 280, and give it a shot, you aren't going to hurt anything doing so. If you're really paranoid, issue a snapshot before you pull the drives and power down the system. --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zfs read performance degrades over a short time
I appear to be seeing the performance of a local ZFS file system degrading over a short period of time. My system configuration: 32 bit Athlon 1800+ CPU 1 Gbyte of RAM Solaris 10 U6 SunOS filer 5.10 Generic_137138-09 i86pc i386 i86pc 2x250 GByte Western Digital WD2500JB IDE hard drives 1 zfs pool (striped with the two drives, 449 GBytes total) 1 hard drive has an 8 Gbyte UFS slice holding the OS root filesystem My test consists of reading a single 2.3 Gbyte file using 'dd'. Immediately after booting the system, I am seeing a read speed of nearly 100 Mbyte per sec. This is wonderful and shows that the ZFS striping is giving a performance boost as the maximum data rate from each drive is much lower than that - somewhere around 60 Mbytes/sec. Unfortunately if I rerun the test multiple times, the performance degrades. By the 18th time (about 10 minutes later) the test is run, the performance is down to 54 Mbyte/sec. It is interesting to note that the read performance of the UFS root filesystem has not degraded during the above testing.The large file read performance of the root UFS filesystem remains constant at 58 Mbytes. I'm running a fresh install of sol-10-u6-ga1-x86-dvd.iso. The only processes running on the system are the standard ones that happen with a fresh install. Does anyone know if this performance degradation is normal? If not, does anyone have hints on what I should do to track down the problem. Thanks. Ray --The log of my runs is shown below # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 23.8 user0.0 sys11.1 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 24.5 user0.0 sys 8.9 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 27.4 user0.0 sys 8.8 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 26.9 user0.0 sys 8.4 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 32.0 user0.0 sys 8.5 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 31.4 user0.0 sys 8.6 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 33.2 user0.0 sys 8.8 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 34.1 user0.0 sys 8.9 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 35.9 user0.0 sys 8.9 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 36.8 user0.0 sys 8.7 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 38.1 user0.0 sys 8.9 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 38.4 user0.0 sys 8.9 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 39.4 user0.0 sys 9.0 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 40.3 user0.0 sys 8.9 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 40.6 user0.0 sys 9.0 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 41.3 user0.0 sys 9.0 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 42.2 user0.0 sys 9.0 # time dd if=big2 of=/dev/null bs=1024k 2201+1 records in 2201+1 records out real 42.4 user0.0 sys 9.2 # -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Where does set the value to zio->io_offset?
Assume starting one disk write action, the vdev_disk_io_start will be called from zio_execute. static int vdev_disk_io_start(zio_t *zio) { .. bp->b_lblkno = lbtodb(zio->io_offset); .. } After scaning over the zfs source, I find the zio->io_offset is only set value in zio_create by the parameter offset. zio_write calls zio_create with the value 0 for the parameter offset. I can't find anywhere else the zio->io_offset being set. After the new block born, the correct offset has been filled in bp->blk_dva (see metaslab_alloc), when and where the correct value set to zio->io_offset? Who can tell me? thanks -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss