Re: [zfs-discuss] zpool export taking hours

2009-07-28 Thread fyleow
fyleow wrote: I have a raidz1 tank of 5x 640 GB hard drives on my newly installed OpenSolaris 2009.06 system. I did a zpool export tank and the process has been running for 3 hours now taking up 100% CPU usage. When I do a zfs list tank it's still shown as mounted. What's going on

Re: [zfs-discuss] Help with setting up ZFS

2009-07-28 Thread Brian
The two plugs that I indicated are multi-lane SAS ports, which /require/ using a breakout cable; don't worry - that the design for them. multi-lane means exactly that - several actual SAS connections in a single plug. The other 6 ports next to them (in black) are SATA ports

[zfs-discuss] ZFS Mirror : drive unexpectedly unplugged

2009-07-28 Thread Avérous Julien-Pierre
Hi ! I'm a Mac User, but I think that I will get more response here about this question than on a Mac forum. And first, sorry for my approximative English. I have a ZFS Pool named MyPool with two device (two external USB drive), configured as mirror : NAME STATE READ WRITE

Re: [zfs-discuss] ZFS Mirror : drive unexpectedly unplugged

2009-07-28 Thread Thomas Burgess
I don't have an answer to your question exactly because i'm a noob and i'm not using mac but i can say that on FreeBSD which i'm using atm there is a method to name devices ahead of time so if the drive letters change you avoid this exact problem. I'm sure opensolaris and mac have something

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-28 Thread Ross
I think people can understand the concept of missing flushes. The big conceptual problem is how this manages to hose an entire filesystem, which is assumed to have rather a lot of data which ZFS has already verified to be ok. Hardware ignoring flushes and loosing recent data is understandable,

Re: [zfs-discuss] ZFS Mirror : drive unexpectedly unplugged

2009-07-28 Thread Avérous Julien-Pierre
Thank you for you response, wonslung. I can export / import, yes, but for this I should unmount all filesystems depending of the pool, and it's not always possible (and it's sad to be forced to do that). For the same name device, I don't know how to do that. I will search for this. -- This

Re: [zfs-discuss] ZFS Mirror : drive unexpectedly unplugged

2009-07-28 Thread Avérous Julien-Pierre
There is a little mistake : If I do a attach of disk4s2 on disk2s2, it say to me that disk3s2 is busy (it's suspicious : the drive is not used) The good version is : If I do a attach of disk4s2 on disk2s2, it say to me that disk4s2 is busy (it's suspicious : the drive is not used) (disk3s2

Re: [zfs-discuss] ZFS Mirror : drive unexpectedly unplugged

2009-07-28 Thread Thomas Burgess
sometimes the disk will be busy just from being in the directory or if something is trying to connect to it. Again, i'm no expert so i'm going to refrain from commenting on your issue further. 2009/7/28 Avérous Julien-Pierre no-re...@opensolaris.org There is a little mistake : If I do a

[zfs-discuss] How to mirror an entire zfs pool to another pool

2009-07-28 Thread Thomas Walker
We are upgrading to new storage hardware. We currently have a zfs pool with the old storage volumes. I would like to create a new zfs pool, completely separate, with the new storage volumes. I do not want to just replace the old volumes with new volumes in the pool we are currently using. I

Re: [zfs-discuss] How to mirror an entire zfs pool to another pool

2009-07-28 Thread michael schuster
Thomas Walker wrote: We are upgrading to new storage hardware. We currently have a zfs pool with the old storage volumes. I would like to create a new zfs pool, completely separate, with the new storage volumes. I do not want to just replace the old volumes with new volumes in the pool we are

Re: [zfs-discuss] How to mirror an entire zfs pool to another pool

2009-07-28 Thread Darren J Moffat
Thomas Walker wrote: We are upgrading to new storage hardware. We currently have a zfs pool with the old storage volumes. I would like to create a new zfs pool, completely separate, with the new storage volumes. I do not want to just replace the old volumes with new volumes in the pool we

Re: [zfs-discuss] How to mirror an entire zfs pool to another pool

2009-07-28 Thread Thomas Walker
zpool create newpool zfs snapshot -r oldp...@sendit zfs send -R oldp...@sendit | zfs recv -vFd newpool I think this is probably something like what I want, the problem is I'm not really getting it yet. If you could explain just what is happening here in an example. Let's say I have

Re: [zfs-discuss] How to mirror an entire zfs pool to another pool

2009-07-28 Thread Darren J Moffat
I think this is probably something like what I want, the problem is I'm not really getting it yet. If you could explain just what is happening here in an example. Let's say I have this setup; oldpool = 10 x 500GB volumes, with two mounted filesystems; fs1 and fs2 I create newpool = 12 x 1TB

[zfs-discuss] USF drive on S10u7

2009-07-28 Thread dick hoogendijk
What is the best way to attach an USB harddisk to Solaris 10u7? I know some program is running to auto detect such a device (have forgotten the name, because I do almost all work on OSOL (hal). do I use that program or disable it an manualy attach the drive to the system? -- Dick Hoogendijk --

Re: [zfs-discuss] [indiana-discuss] zfs issues?

2009-07-28 Thread James Lever
Thanks for that Brian. I've logged a bug: CR 6865661 *HOT* Created, P1 opensolaris/triage-queue zfs scrub rpool causes zpool hang Just discovered after trying to create a further crash dump that it's failing and rebooting with the following error (just caught it prior to the reboot):

Re: [zfs-discuss] How to mirror an entire zfs pool to another pool

2009-07-28 Thread Thomas Walker
I think you've given me enough information to get started on a test of the procedure. Thanks very much. Thomas Walker -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org

Re: [zfs-discuss] When writing to SLOG at full speed all disk IO is blocked

2009-07-28 Thread Marcelo Leal
Ok Bob, but i think that is the problem about picket fencing... and so we are talking about commit the sync operations to disk. What i'm seeing is no read activity from disks when the slog is beeing written. The disks are zero (no read, no write). Thanks a lot for your reply. Leal [

Re: [zfs-discuss] USF drive on S10u7

2009-07-28 Thread Cindy . Swearingen
Hi Dick, The Solaris 10 volume management service is volfs. If you attach the USB hard disk and run volcheck, the disk should be mounted under the /rmdisk directory. If the auto-mounting doesn't occur, you can disable volfs and mount it manually. You can read more about this feature here:

Re: [zfs-discuss] [sam-qfs-discuss] sam-fs on zfs-pool

2009-07-28 Thread David Gwynne
On 27/07/2009, at 10:14 PM, Tobias Exner wrote: Hi list, I've did some tests and run into a very strange situation.. I created a zvol using zfs create -V and initialize an sam- filesystem on this zvol. After that I restored some testdata using a dump from another system. So far so good.

Re: [zfs-discuss] [sam-qfs-discuss] sam-fs on zfs-pool

2009-07-28 Thread Tomas Ögren
On 28 July, 2009 - David Gwynne sent me these 1,9K bytes: On 27/07/2009, at 10:14 PM, Tobias Exner wrote: Hi list, I've did some tests and run into a very strange situation.. I created a zvol using zfs create -V and initialize an sam- filesystem on this zvol. After that I restored

Re: [zfs-discuss] When writing to SLOG at full speed all disk IO is blocked

2009-07-28 Thread Bob Friesenhahn
On Tue, 28 Jul 2009, Marcelo Leal wrote: Ok Bob, but i think that is the problem about picket fencing... and so we are talking about commit the sync operations to disk. What i'm seeing is no read activity from disks when the slog is beeing written. The disks are zero (no read, no write).

Re: [zfs-discuss] How to mirror an entire zfs pool to another pool

2009-07-28 Thread Gaëtan Lehmann
Le 28 juil. 09 à 15:54, Darren J Moffat a écrit : How do I monitor the progress of the transfer? Once Unfortunately there is no easy way to do that just now. When the 'zfs recv' finishes is it is done. I've just found pv (pipe viewer) today (http://www.ivarch.com/programs/pv.shtml )

Re: [zfs-discuss] zpool is lain to burnination (bwahahahah!)

2009-07-28 Thread Graeme Clark
Hi Again, A bit more futzing around and I notice that output from a plain 'zdb' returns this: store version=14 name='store' state=0 txg=0 pool_guid=13934602390719084200 hostid=8462299 hostname='store' vdev_tree type='root' id=0

Re: [zfs-discuss] Help with setting up ZFS

2009-07-28 Thread Will Murnane
On Tue, Jul 28, 2009 at 03:04, Brianno-re...@opensolaris.org wrote: Just a quick question before I address everyone else. I bought this connector http://www.newegg.com/Product/Product.aspx?Item=N82E16812198020 However its pretty clear to me now (after Ive ordered it) that it won't at all

Re: [zfs-discuss] [sam-qfs-discuss] sam-fs on zfs-pool

2009-07-28 Thread Richard Elling
On Jul 28, 2009, at 8:53 AM, Tomas Ögren wrote: On 28 July, 2009 - David Gwynne sent me these 1,9K bytes: On 27/07/2009, at 10:14 PM, Tobias Exner wrote: Hi list, I've did some tests and run into a very strange situation.. I created a zvol using zfs create -V and initialize an sam-

Re: [zfs-discuss] USF drive on S10u7

2009-07-28 Thread dick hoogendijk
On Tue, 28 Jul 2009 09:03:14 -0600 cindy.swearin...@sun.com wrote: The Solaris 10 volume management service is volfs. #svcs -a | grep vol has told me that ;-) If the auto-mounting doesn't occur, you can disable volfs and mount it manually. I don't want the automounting to occur, so I diabled

Re: [zfs-discuss] When writing to SLOG at full speed all disk IO is blocked

2009-07-28 Thread Ross
My understanding is that there's never any need for a reader to wait for a write in progress. ZFS keeps all writes in memory until they're committed to disk - if you ever try to read something that's either waiting to be, or is being written to disk, ZFS will serve it straight from RAM. One

Re: [zfs-discuss] zfs destroy slow?

2009-07-28 Thread Brent Jones
On Mon, Jul 27, 2009 at 3:58 AM, Markus Koveromarkus.kov...@nebula.fi wrote: Oh well, whole system seems to be deadlocked. nice. Little too keen keeping data safe :-P Yours Markus Kovero From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf

Re: [zfs-discuss] zfs destroy slow?

2009-07-28 Thread Brent Jones
I submitted a bug, but I don't think its been assigned a case number yet. I see this exact same behavior on my X4540's. I create a lot of snapshots, and when I tidy up, zfs destroy can 'stall' any and all ZFS related commands for hours, or even days (in the case of nested snapshots). The

Re: [zfs-discuss] USF drive on S10u7

2009-07-28 Thread Bob Friesenhahn
On Tue, 28 Jul 2009, dick hoogendijk wrote: I don't want the automounting to occur, so I diabled volfs. I then did a rmformat to learn the device name, followed by a zpool create archive /dev/rdsk/devicename It is better to edit /etc/vold.conf since vold is used for other purposes as well

Re: [zfs-discuss] zfs destroy slow?

2009-07-28 Thread Eric Schrock
On 07/27/09 03:39, Markus Kovero wrote: Hi, how come zfs destroy being so slow, eg. destroying 6TB dataset renders zfs admin commands useless for time being, in this case for hours? (running osol 111b with latest patches.) I'm not sure what latest patches means w.r.t. ON build, but this is

Re: [zfs-discuss] zpool is lain to burnination (bwahahahah!)

2009-07-28 Thread Victor Latushkin
On 28.07.09 20:31, Graeme Clark wrote: Hi Again, A bit more futzing around and I notice that output from a plain 'zdb' returns this: store version=14 name='store' state=0 txg=0 pool_guid=13934602390719084200 hostid=8462299 hostname='store' vdev_tree

[zfs-discuss] Set New File/Folder ZFS ACLs Automatically through Samba?

2009-07-28 Thread Jeff Hulen
Do any of you know how to set the default ZFS ACLs for newly created files and folders when those files and folders are created through Samba? I want to have all new files and folders only inherit extended (non-trivial) ACLs that are set on the parent folders. But when a file is created through

Re: [zfs-discuss] Why is Solaris 10 ZFS performance so terrible?

2009-07-28 Thread Rich Morris
On Mon, Jul 20, 2009 at 7:52 PM, Bob Friesenhahn wrote: Sun has opened internal CR 6859997. It is now in Dispatched state at High priority. CR 6859997 has been accepted and is actively being worked on. The following info has been added to that CR: This is a problem with the ZFS file

[zfs-discuss] zfs send/recv syntax

2009-07-28 Thread Joseph L. Casale
Is it possible to send an entire pool (including all its zfs filesystems) to a zfs filesystem in a different pool on another host? Or must I send each zfs filesystem one at a time? Thanks! jlc ___ zfs-discuss mailing list zfs-discuss@opensolaris.org

[zfs-discuss] avail drops to 32.1T from 40.8T after create -o mountpoint

2009-07-28 Thread Glen Gunselman
This is my first ZFS pool. I'm using an X4500 with 48 TB drives. Solaris is 5/09. After the create zfs list shows 40.8T but after creating 4 filesystems/mountpoints the available drops 8.8TB to 32.1TB. What happened to the 8.8TB. Is this much overhead normal? zpool create -f zpool1 raidz

Re: [zfs-discuss] avail drops to 32.1T from 40.8T after create -o mountpoint

2009-07-28 Thread Mario Goebbels
This is my first ZFS pool. I'm using an X4500 with 48 TB drives. Solaris is 5/09. After the create zfs list shows 40.8T but after creating 4 filesystems/mountpoints the available drops 8.8TB to 32.1TB. What happened to the 8.8TB. Is this much overhead normal? IIRC zpool list includes the

Re: [zfs-discuss] zfs send/recv syntax

2009-07-28 Thread Ian Collins
On Wed 29/07/09 10:09 , Joseph L. Casale jcas...@activenetwerx.com sent: Is it possible to send an entire pool (including all its zfsfilesystems) to a zfs filesystem in a different pool on another host? Or must I send each zfs filesystem one at a time? Yes, use -R on the sending side and

Re: [zfs-discuss] avail drops to 32.1T from 40.8T after create -o mountpoint

2009-07-28 Thread Scott Lawson
Glen Gunselman wrote: This is my first ZFS pool. I'm using an X4500 with 48 TB drives. Solaris is 5/09. After the create zfs list shows 40.8T but after creating 4 filesystems/mountpoints the available drops 8.8TB to 32.1TB. What happened to the 8.8TB. Is this much overhead normal?

Re: [zfs-discuss] Why is Solaris 10 ZFS performance so terrible?

2009-07-28 Thread Bob Friesenhahn
On Tue, 28 Jul 2009, Rich Morris wrote: 6412053 is a related CR which mentions that the zfetch code may not be issuing I/O at a sufficient pace. This behavior is also seen on a Thumper running the test script in CR 6859997 since, even when prefetch is ramping up as expected, less than half

Re: [zfs-discuss] zfs send/recv syntax

2009-07-28 Thread Trevor Pretty
Try send/receive to the same host (ssh localhost). I used this when trying send/receive as it removes ssh between hosts "problems" The on disk format of ZFS has changed there is something about it in the man pages from memory so I don't think you can go S10 - OpenSolaris without doing an

Re: [zfs-discuss] Why is Solaris 10 ZFS performance so terrible?

2009-07-28 Thread Bob Friesenhahn
On Tue, 28 Jul 2009, Rich Morris wrote: The fix for this problem may be more feedback between the ARC and the zfetch code. Or it may make sense to restart the prefetch stream after some time has passed or perhaps whenever there's a miss on a block that was expected to have already been

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-28 Thread Rennie Allen
Can *someone* please name a single drive+firmware or RAID controller+firmware that ignores FLUSH CACHE / FLUSH CACHE EXT commands? Or worse, responds ok when the flush hasn't occurred? I think it would be a shorter list if one were to name the drives/controllers that actually implement a

Re: [zfs-discuss] zfs send/recv syntax

2009-07-28 Thread Ian Collins
On Wed 29/07/09 10:49 , Joseph L. Casale jcas...@activenetwerx.com sent: Yes, use -R on the sending side and -d on the receiving side. I tried that first, going from Solaris 10 to osol 0906: # zfs send -vR mypo...@snap|ssh j...@catania pfexec /usr/sbin/zfs recv -dF mypool/somename

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-28 Thread Rennie Allen
This is also (theoretically) why a drive purchased from Sun is more that expensive then a drive purchased from your neighbourhood computer shop: It's more significant than that. Drives aimed at the consumer market are at a competitive disadvantage if they do handle cache flush

Re: [zfs-discuss] zfs send/recv syntax

2009-07-28 Thread Cindy Swearingen
I apologize for replying in the middle of this thread, but I never saw the initial snapshot syntax of mypool2, which needs to be recursive (zfs snapshot -r mypo...@snap) to snapshot all the datasets in mypool2. Then, use zfs send -R to pick up and restore all the dataset properties. What was the

Re: [zfs-discuss] [n/zfs-discuss] Strange speeds with x4500, Solaris 10 10/08

2009-07-28 Thread Jorgen Lundman
This thread started over in nfs-discuss, as it appeared to be an nfs problem initially. Or at the very least, interaction between nfs and zil. Just summarising speeds we have found when untarring something. Always in a new/empty directory. Only looking at write speed. read is always very

Re: [zfs-discuss] [n/zfs-discuss] Strange speeds with x4500, Solaris 10 10/08

2009-07-28 Thread Bob Friesenhahn
On Wed, 29 Jul 2009, Jorgen Lundman wrote: For example, I know rsync and tar does not use fdsync (but dovecot does) on its close(), but does NFS make it fdsync anyway? NFS is required to do synchronous writes. This is what allows NFS clients to recover seamlessly if the server

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-28 Thread Eric D. Mudama
On Mon, Jul 27 at 13:50, Richard Elling wrote: On Jul 27, 2009, at 10:27 AM, Eric D. Mudama wrote: Can *someone* please name a single drive+firmware or RAID controller+firmware that ignores FLUSH CACHE / FLUSH CACHE EXT commands? Or worse, responds ok when the flush hasn't occurred? two

[zfs-discuss] Failing device in a replicated configuration....on a non replicated pool??

2009-07-28 Thread fyleow
I was greeted by this today. The Sun Message ID page says this should happen when there were errors in a replicated configuration. Clearly there's only one drive here. If there are unrecoverable errors how can my applications not be affected since there's no mirror or parity to recover from?

Re: [zfs-discuss] Failing device in a replicated configuration....on a non replicated pool??

2009-07-28 Thread Bob Friesenhahn
On Tue, 28 Jul 2009, fyleow wrote: I was greeted by this today. The Sun Message ID page says this should happen when there were errors in a replicated configuration. Clearly there's only one drive here. If there are unrecoverable errors how can my applications not be affected since there's no