Re: [zfs-discuss] How to recover -- LUNs go offline, now permanent errors?
Cindy, I gave your suggestion a try. I did the zpool clear and then did another zpool scrub and all is happy now. Thank you for your help. David -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] How to recover -- LUNs go offline, now permanent errors?
Cindy, Thanks for the reply. I'll get that a try and then send an update. Thanks, David -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] How to recover -- LUNs go offline, now permanent errors?
I recently had an issue with my LUNs from our storage unit going offline. This caused the zpool to get numerous errors on the luns. The pool is on-line, and I did a scrub, but one of the raid sets is degraded: raidz2-3 DEGRADED 0 0 0 c7t60001FF011C6F3103B00011D1BF1d0 DEGRADED 0 0 0 too many errors c7t60001FF011C6F3023900011D1BF1d0 DEGRADED 0 0 0 too many errors c7t60001FF011C6F2F53700011D1BF1d0 DEGRADED 0 0 0 too many errors c7t60001FF011C6F2E43500011D1BF1d0 DEGRADED 0 0 0 too many errors c7t60001FF011C6F2D23300011D1BF1d0 DEGRADED 0 0 0 too many errors c7t60001FF011C6F2A93100011D1BF1d0 DEGRADED 0 0 0 too many errors c7t60001FF011C6F29A2F00011D1BF1d0 DEGRADED 0 0 0 too many errors c7t60001FF011C6F2682D00011D1BF1d0 DEGRADED 0 0 0 too many errors c7t60001FF011C6F24C2B00011D1BF1d0 DEGRADED 0 0 0 too many errors c7t60001FF011C6F2192900011D1BF1d0 DEGRADED 0 0 0 too many errors Also I have the following: errors: Permanent errors have been detected in the following files: <0x3a>:<0x3b04> Originally, there was a file, and then a directory listed, but I removed them. Now I'm stuck with the hex codes above. How do I interpret them? Can this pool be recovered, or basically how do I proceed? The system is Solaris 10 U9 with all recent patches. Thanks, David -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Zpool metadata corruption from S10U9 to S11 express
An update: I had mirrored my boot drive when I installed Solaris 10U9 originally, so I went ahead and rebooted the system to this disk instead of my Solaris 11 install. After getting the system up, I imported the zpool, and everything worked normally. So I guess there is some sort of incompatibility between Solaris 10 and Solaris 11. I would have thought that Solaris 11 could import an older pool level. Any other insight on importing pools between these two versions of Solaris would be helpful. Thanks, David -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Zpool metadata corruption from S10U9 to S11 express
I was recently running Solaris 10 U9 and I decided that I would like to go to Solaris 11 Express so I exported my zpool, hoping that I would just do an import once I had the new system installed with Solaris 11. Now when I try to do an import I'm getting the following: # /home/dws# zpool import pool: tank id: 13155614069147461689 state: FAULTED status: The pool metadata is corrupted. action: The pool cannot be imported due to damaged devices or data. see: http://www.sun.com/msg/ZFS-8000-72 config: tank FAULTED corrupted data logs mirror-6 ONLINE c9t57d0 ONLINE c9t58d0 ONLINE mirror-7 ONLINE c9t59d0 ONLINE c9t60d0 ONLINE Is there something else I can do to see what is wrong. Original attempt when specifying the name resulted in: # /home/dws# zpool import tank cannot import 'tank': I/O error Destroy and re-create the pool from a backup source. I verified that I have all 60 of my luns. The controller numbers have changed, but I don't believe that should matter. Any suggestions about getting additional information about what is happening would be greatly appreciated. Thanks, David -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] How to map solaris disk devices to physical location for ZFS pool setup
Hi, I'm setting up a ZFS environment running on a Sun x4440 + J4400 arrays (similar to 7410 environment) and I was trying to figure out the best way to map a disk drive physical location (tray and slot) to the Solaris device c#t#d#. Do I need to install the CAM software to do this, or is there another way? I would like to understand the solaris device to physical drive location so that I can setup my ZFS pool mirrors/raid properly. I'm currently running Solaris Express build 119. Thanks, David -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs destroy is taking a long time...
A few more details: The system is a Sun x4600 running Solaris 10 Update 4. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zfs destroy is taking a long time...
I was wondering if anyone has any experience with how long a "zfs destroy" of about 40 TB should take? So far, it has been about an hour... Is there any good way to tell if it is working or if it is hung? Doing a "zfs list" just hangs. If you do a more specific zfs list, then it is okay... zfs list pool/another-fs Thanks, David -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Replacing a Lun (Raid 0) (Santricity)
Yes! That worked to get the spare back to an available state. Thanks! So that leaves me with the trying to put together a recommended procedure to replace a failed lun/disk from our Flexline 380. Does anyone have configuration in which they are using a RAID 0 lun, which they need to replace? Thanks, David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Replacing a Lun (Raid 0) (Santricity)
Addtional information: It looks like perhaps the original drive is in use, and the hot spare is assigned but not in use see below about zpool iostat: raidz22.76T 4.49T 0 0 29.0K 18.4K c10t600A0B80001139967CF945E80E95d0 - - 0 0 2.46K 1.33K c10t600A0B8000115EA2FEEB45E8145Ed0 - - 0 0 2.46K 1.33K spare - - 0 0 1.81K 1.33K c10t600A0B80001139967D2345E81075d0 - - 0 0 2.47K 1.33K c10t600A0B80001139967CE145E80D4Dd0 - - 0 0 0 1.33K c10t600A0B80001139967D3F45E81157d0 - - 0 0 2.47K 1.33K c10t600A0B8000115EA2FF1145E817BEd0 - - 0 0 2.46K 1.33K c10t600A0B80001139967D5D45E813EBd0 - - 0 0 2.46K 1.33K c10t600A0B8000115EA2FE7D45E80CDEd0 - - 0 0 2.47K 1.33K c10t600A0B80001139967C6145E808C7d0 - - 0 0 2.47K 1.33K c10t600A0B8000115EA2FE9945E80E6Ad0 - - 0 0 2.46K 1.33K c10t600A0B80001139967C8B45E80A59d0 - - 0 0 2.46K 1.33K c10t600A0B80001139967CA745E80B21d0 - - 0 0 2.47K 1.33K c10t600A0B8000115EA2FEB545E810D4d0 - - 0 0 2.47K 1.33K c10t600A0B80001139967CD145E80CD7d0 - - 0 0 2.47K 1.33K c10t600A0B8000115EA2FED145E8129Cd0 - - 0 0 2.47K 1.33K c10t600A0B80001139967CFB45E80EA5d0 - - 0 0 2.47K 1.33K c10t600A0B8000115EA2FEED45E8146Ed0 - - 0 0 2.46K 1.33K spares c10t600A0B80001139967CE145E80D4Dd0 INUSE currently in use So how do I get the hot spare out of the inuse state? This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Replacing a Lun (Raid 0) (Santricity)
I would like advise about how to replace a raid 0 lun. The lun is basically a raid 0 lun which is from a single disk volume group / volume from our Flexline 380 unit. So every disk in the unit is a volume group/volume/lun mapped to the host. We then let ZFS do the raid. We have a lun now which has been getting read errors and basically the underlying drive needs to be replaced. I've done this in the past but as I remember it was cumbersome and didn't go that smoothly so I would like some advise about how to go about it. As I recall if you just fail the disk via Santricity, the lun really doesn't go offline from a host point of via and hence ZFS still tries to write to the lun/disk. I believe I either unmapped the lun to the host at which point ZFS kicked in a hot spare, or I offlined the disk via ZFS first. Then at that point the drive can be replace on the storage and the volume re-initialized then remapped back to the host. So part of my question... does the above sound reasonable, or should I be doing this differently? Also I'm a little unsure about how to get the original lun back in operation and the spare back to a spare. For example I have the following situation now: raidz2 ONLINE 0 0 0 c10t600A0B80001139967CF945E80E95d0ONLINE 0 0 0 c10t600A0B8000115EA2FEEB45E8145Ed0ONLINE 0 0 0 spare ONLINE 0 0 0 c10t600A0B80001139967D2345E81075d0 ONLINE 0 0 0 c10t600A0B80001139967CE145E80D4Dd0 ONLINE 0 0 0 c10t600A0B80001139967D3F45E81157d0ONLINE 0 0 0 c10t600A0B8000115EA2FF1145E817BEd0ONLINE 0 0 0 c10t600A0B80001139967D5D45E813EBd0ONLINE 0 0 0 c10t600A0B8000115EA2FE7D45E80CDEd0ONLINE 0 0 0 c10t600A0B80001139967C6145E808C7d0ONLINE 0 0 0 c10t600A0B8000115EA2FE9945E80E6Ad0ONLINE 0 0 0 c10t600A0B80001139967C8B45E80A59d0ONLINE 0 0 0 c10t600A0B80001139967CA745E80B21d0ONLINE 0 0 0 c10t600A0B8000115EA2FEB545E810D4d0ONLINE 0 0 0 c10t600A0B80001139967CD145E80CD7d0ONLINE 0 0 0 c10t600A0B8000115EA2FED145E8129Cd0ONLINE 0 0 0 c10t600A0B80001139967CFB45E80EA5d0ONLINE 0 0 0 c10t600A0B8000115EA2FEED45E8146Ed0ONLINE 0 0 0 spares c10t600A0B80001139967CE145E80D4Dd0 INUSE currently in use c10t600A0B8000115EA2FEE145E81328d0 AVAIL c10t600A0B80001139967D0B45E80F21d0 AVAIL c10t600A0B8000115EA2FEFD45E81506d0 AVAIL c10t600A0B80001139967D3545E81107d0 AVAIL c10t600A0B80001139967D5345E81289d0 AVAIL c10t600A0B8000115EA2FF2345E81864d0 AVAIL c10t600A0B80001139967D6F45E8149Bd0 AVAIL I thought if I replace the original device with the same (but new) lun that should work, but I get the following: zpool replace tank c10t600A0B80001139967D2345E81075d0 c10t600A0B80001139967D2345E81075d0 invalid vdev specification use '-f' to override the following errors: /dev/dsk/c10t600A0B80001139967D2345E81075d0s0 is part of active ZFS pool tank. Please see zpool(1M). The above lun is the same as before, but the underlying disk on the storage was replaced. Do I need to do something to this lun to make ZFS think it is a new disk? Or should I be doing something different? I now have another disk which has gone back so I need to fix my above situation with the hot spare first, and then go through the process again for my second failure. The environment is S10U4, running on a x4600 with Flexline 380 storage units. Tia, David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zpool status (advanced listing)?
Under the GUI, there is an "advanced" option which shows vdev capacity, etc. I'm drawing a blank about how to get with the commands... Thanks, David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove snapshots
To list your snapshots: /usr/sbin/zfs list -H -t snapshot -o name Then you could use that in a for loop: for i in `/usr/sbin/zfs list -H -t snapshot -o name` ; do echo "Destroying snapshot: $i" /usr/sbin/zfs destroy $i done The above would destroy all your snapshots. You could put a grep on the end to select certain ones by a specific string... For example: /usr/sbin/zfs list -H -t snapshot -o name | grep "aug-16" David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] General recommendations on raidz groups of different sizes
What are your thoughts or recommendations on having a zpool made up of raidz groups of different sizes? Are there going to be performance issues? For example: pool: testpool1 state: ONLINE scrub: none requested config: NAMESTATE READ WRITE CKSUM testpool1 ONLINE 0 0 0 raidz1ONLINE 0 0 0 c12t600A0B800029E5EA07234685122Ad0 ONLINE 0 0 0 c12t600A0B800029E5EA07254685123Cd0 ONLINE 0 0 0 c12t600A0B800029E5EA072F46851256d0 ONLINE 0 0 0 c12t600A0B800029E5EA073146851266d0 ONLINE 0 0 0 c12t600A0B800029E5EA073746851278d0 ONLINE 0 0 0 c12t600A0B800029E5EA074146851292d0 ONLINE 0 0 0 c12t600A0B800029E5EA0747468512B6d0 ONLINE 0 0 0 c12t600A0B800029E5EA0749468512C2d0 ONLINE 0 0 0 raidz1ONLINE 0 0 0 c12t600A0B800029E5EA074F468512E0d0 ONLINE 0 0 0 c12t600A0B800029E5EA0751468512E8d0 ONLINE 0 0 0 c12t600A0B800029E5EA07574685130Cd0 ONLINE 0 0 0 c12t600A0B800029E5EA075946851318d0 ONLINE 0 0 0 c12t600A0B800029E5EA075F4685132Ed0 ONLINE 0 0 0 c12t600A0B800029E5EA076546851342d0 ONLINE 0 0 0 Thanks, David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs list hangs if zfs send is killed (leaving zfs receive process)
Well, the zfs receive process finally died, and now my zfs list works just fine. If there is a better way to capture what is going on, please let me know and I can duplicate the hang. David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Again ZFS with expanding LUNs!
I don't believe LUN expansion is quite yet possible under Solaris 10 (11/06). I believe this might make it into the next update but I'm not sure on that. Someone from Sun would need to comment on when this will make it into the production release of Solaris. I know this because I was working with a person from Sun testing a utility which will detect the larger LUN and update the disk label. Then at that point if you export the pool and import it again the added space will show up. The utility was just for testing however. I'm waiting as well for the automatic LUN expansion stuff. David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zfs list hangs if zfs send is killed (leaving zfs receive process)
I was in the process of doing a large zfs send | zfs receive when I decided that I wanted to terminate the the zfs send process. I killed it, but the zfs receive doesn't want to die... In the meantime my zfs list command just hangs. Here is the tail end of the truss output from a "truss zfs list": ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08043484) = 0 ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) Err#3 ESRCH ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0 ioctl(3, ZFS_IOC_OBJSET_STATS, 0x0804464C) = 0 ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08045788) Err#3 ESRCH ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) = 0 ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08043484) = 0 ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) Err#3 ESRCH ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0 brk(0x082CA000) = 0 ioctl(3, ZFS_IOC_OBJSET_STATS, 0x0804464C) = 0 ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08045788) Err#3 ESRCH ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) = 0 ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08043484) = 0 ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) Err#3 ESRCH ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0 ioctl(3, ZFS_IOC_OBJSET_STATS, 0x0804464C) = 0 ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08045788) Err#3 ESRCH ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) = 0 ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08043484) = 0 ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) Err#3 ESRCH ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0 ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) Err#3 ESRCH ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08046950) = 0 ioctl(3, ZFS_IOC_OBJSET_STATS, 0x0804464C) = 0 ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08046950) Err#3 ESRCH ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08045814) = 0 ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0 ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0 ioctl(3, ZFS_IOC_OBJSET_STATS, 0x0804464C) = 0 ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08045788) = 0 ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08043484) = 0 Let me know if something else is needed to figure out what is going on. The system is an x4600 running Solaris 10, with a fairly recent patch set. bash-3.00# uname -a SunOS gdo-node-1 5.10 Generic_125101-07 i86pc i386 i86pc bash-3.00# ps -ef | grep zfs root 17593 17438 0 Jul 10 pts/3 605:57 zfs receive gdo-node-2/pcmdi/ftp/pub-backup root 25972 18474 0 17:04:03 pts/4 0:00 truss zfs list root 25968 18474 0 17:03:12 pts/4 0:08 zfs list root 26086 17438 0 17:31:55 pts/3 0:00 grep zfs root 25926 23059 0 16:55:44 pts/5 0:12 zfs list root 25973 25972 0 17:04:03 pts/4 0:08 zfs list David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zpool status -v: machine readable format?
I was wondering if anyone had a script to parse the "zpool status -v" output into a more machine readable format? Thanks, David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: How to get new ZFS Solaris 10 U3 features going from Solaris 10 U2
Thank you to everyone that has replied. It sounds like I have a few options with regards to upgrading or just waiting and patching the current environment. David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] How to get new ZFS Solaris 10 U3 features going from Solaris 10 U2
We currently have a couple of servers at Solaris 10 U2, and we would like to get to Solaris 10 U3 for the new zfs features. Can this be accomplished via patching, or do you have to do an upgrade from S10U2 to S10U3? Also what about a system with Zones? What is the best practice for upgrading a system with zones? Thanks, David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Re: Corrupted LUN in RAIDZ group -- How to repair?
I have run zpool scrub again, and I now see checksum errors again. Wouldn't the checksum errors gotten fixed with the first zpool scrub? Can anyone recommend actions I should do at this point? Thanks, David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] How to NOT mount a ZFS storage pool/ZFS file system?
I currently have a system which has two ZFS storage pools. One of the pools is coming from a faulty piece of hardware. I would like to bring up our server mounting the storage pool which is okay and NOT mounting the one with from the hardware with problems. Is there a simple way to NOT mount one of my ZFS storage pools? The system is currently down due to the disk issues from one of the above pools. Thanks, David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Corrupted LUN in RAIDZ group -- How to repair?
James, Thanks for the reply. It looks like now the scrub has completed. Should I now clear these warnings? bash-3.00# zpool status -x pool: mypool state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: http://www.sun.com/msg/ZFS-8000-9P scrub: scrub completed with 0 errors on Sun Sep 10 07:44:36 2006 config: NAMESTATE READ WRITE CKSUM mypoolONLINE 0 0 0 raidz ONLINE 0 0 0 c10t600A0B800011730E66C544C5EBB8d0 ONLINE 0 0 0 c10t600A0B800011730E66CA44C5EBEAd0 ONLINE 0 0 0 c10t600A0B800011730E66CF44C5EC1Cd0 ONLINE 0 0 0 c10t600A0B800011730E66D444C5EC5Cd0 ONLINE 0 0 0 c10t600A0B800011730E66D944C5ECA0d0 ONLINE 0 0 0 c10t600A0B800011652EE5C144C5ECDFd0 ONLINE 0 0 0 c10t600A0B800011730E66E244C5ED2Cd0 ONLINE 0 0 0 c10t600A0B800011652EE5C644C5ED87d0 ONLINE 0 0 0 c10t600A0B800011730E66EB44C5EDD8d0 ONLINE 0 0 0 c10t600A0B800011652EE5CB44C5EE29d0 ONLINE 0 0 0 c10t600A0B800011730E66F444C5EE7Ed0 ONLINE 0 015 c10t600A0B800011652EE5D044C5EEC9d0 ONLINE 0 0 0 c10t600A0B800011730E66FD44C5EF1Ad0 ONLINE 50 0 0 c10t600A0B800011652EE5D544C5EF63d0 ONLINE 0 0 0 raidz ONLINE 0 0 0 c10t600A0B800011652EE5B844C5EBCBd0 ONLINE 0 0 0 c10t600A0B800011652EE5BA44C5EBF5d0 ONLINE 0 0 0 c10t600A0B800011652EE5BC44C5EC2Dd0 ONLINE 0 0 0 c10t600A0B800011652EE5BE44C5EC6Bd0 ONLINE 0 0 0 c10t600A0B800011730E66DB44C5ECB4d0 ONLINE 0 0 0 c10t600A0B800011652EE5C344C5ECF9d0 ONLINE 0 0 0 c10t600A0B800011730E66E444C5ED5Ad0 ONLINE 0 0 0 c10t600A0B800011652EE5C844C5EDA1d0 ONLINE 0 0 0 c10t600A0B800011730E66ED44C5EDFAd0 ONLINE 0 0 0 c10t600A0B800011652EE5CD44C5EE47d0 ONLINE 0 0 0 c10t600A0B800011730E66F644C5EE96d0 ONLINE 0 014 c10t600A0B800011652EE5D244C5EEE7d0 ONLINE 0 0 0 c10t600A0B800011730E66FF44C5EF32d0 ONLINE 70 0 0 c10t600A0B800011652EE5D744C5EF7Fd0 ONLINE 0 0 0 David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Corrupted LUN in RAIDZ group -- How to repair?
Background: We have a ZFS pool setup from LUNS which are from a SAN connected StorageTek/Engenio Flexline 380 storage system. Just this past Friday the storage environment went down causing the system to go down. After looking at the storage environment, we had several volume groups which needed to be carefully put back together to prevent corruption. Well, one of the volume groups and the volumes/LUNs coming from it got corrupted. Since our ZFS pools is setup to only have a LUN from each volume group we basically ended up with a single disk loss in our RAIDZ group. So I believe we should be able to recover from this. My question is how to replace this disk (LUN). Basically the LUN is again okay, but the data on the LUN is not. I have tried to do a zpool replace, but ZFS seems to know that the disk/lun is the same device. Using a -f (force) didn't work either. How does one replace a LUN with ZFS? I'm currently doing a "scrub", but don't know if that will help. I first just had read errors on a lun in the raidz group, but just tonight noticed that I now have a checksum error on another lun as well. (see zpool status command output below) Below is a zpool status -x output. Can anyone advise how to recover from this? # zpool status -x pool: mypool state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: http://www.sun.com/msg/ZFS-8000-9P scrub: scrub in progress, 66.00% done, 10h45m to go config: NAMESTATE READ WRITE CKSUM mypoolONLINE 0 0 0 raidz ONLINE 0 0 0 c10t600A0B800011730E66C544C5EBB8d0 ONLINE 0 0 0 c10t600A0B800011730E66CA44C5EBEAd0 ONLINE 0 0 0 c10t600A0B800011730E66CF44C5EC1Cd0 ONLINE 0 0 0 c10t600A0B800011730E66D444C5EC5Cd0 ONLINE 0 0 0 c10t600A0B800011730E66D944C5ECA0d0 ONLINE 0 0 0 c10t600A0B800011652EE5C144C5ECDFd0 ONLINE 0 0 0 c10t600A0B800011730E66E244C5ED2Cd0 ONLINE 0 0 0 c10t600A0B800011652EE5C644C5ED87d0 ONLINE 0 0 0 c10t600A0B800011730E66EB44C5EDD8d0 ONLINE 0 0 0 c10t600A0B800011652EE5CB44C5EE29d0 ONLINE 0 0 0 c10t600A0B800011730E66F444C5EE7Ed0 ONLINE 0 0 9 c10t600A0B800011652EE5D044C5EEC9d0 ONLINE 0 0 0 c10t600A0B800011730E66FD44C5EF1Ad0 ONLINE 50 0 0 c10t600A0B800011652EE5D544C5EF63d0 ONLINE 0 0 0 raidz ONLINE 0 0 0 c10t600A0B800011652EE5B844C5EBCBd0 ONLINE 0 0 0 c10t600A0B800011652EE5BA44C5EBF5d0 ONLINE 0 0 0 c10t600A0B800011652EE5BC44C5EC2Dd0 ONLINE 0 0 0 c10t600A0B800011652EE5BE44C5EC6Bd0 ONLINE 0 0 0 c10t600A0B800011730E66DB44C5ECB4d0 ONLINE 0 0 0 c10t600A0B800011652EE5C344C5ECF9d0 ONLINE 0 0 0 c10t600A0B800011730E66E444C5ED5Ad0 ONLINE 0 0 0 c10t600A0B800011652EE5C844C5EDA1d0 ONLINE 0 0 0 c10t600A0B800011730E66ED44C5EDFAd0 ONLINE 0 0 0 c10t600A0B800011652EE5CD44C5EE47d0 ONLINE 0 0 0 c10t600A0B800011730E66F644C5EE96d0 ONLINE 0 0 6 c10t600A0B800011652EE5D244C5EEE7d0 ONLINE 0 0 0 c10t600A0B800011730E66FF44C5EF32d0 ONLINE 70 0 0 c10t600A0B800011652EE5D744C5EF7Fd0 ONLINE 0 0 0 This system is at Solaris 10, U2. Thank you, David This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss