zpool import can be a little pessimistic about corrupted labels.
First, try physically removing the problem disk and try to import again.
If that doesn't work, then verify the labels on each disk using:
        zdb -l /dev/rdsk/c5d1s0
each disk should have 4 readable labels.
 -- richard

On Apr 15, 2010, at 2:01 AM, Kevin Denton wrote:

> After attempting unsuccessfully to replace a failed drive in a 10 drive 
> raidz2 array and reading as many forum entries as I could find I followed a 
> suggestion to export and import the pool.
> 
> In another attempt to import the pool I reinstalled the OS, but I have so far 
> been unable to import the pool.
> 
> Here is the output from format and zpool commands:
> 
> ke...@opensolaris:~# zpool status
>  pool: rpool
> state: ONLINE
> scrub: none requested
> config:
> 
>    NAME        STATE     READ WRITE CKSUM
>    rpool       ONLINE       0     0     0
>      c8d0s0    ONLINE       0     0     0
> 
> errors: No known data errors
> ke...@opensolaris:~# format
> Searching for disks...done
> 
> 
> AVAILABLE DISK SELECTIONS:
>       0. c4d0 <ST350083-         9QG0LW8-0001-465.76GB>
>          /p...@0,0/pci8086,2...@1e/pci-...@1/i...@0/c...@0,0
>       1. c4d1 <ST350063-         9QG1E50-0001-465.76GB>
>          /p...@0,0/pci8086,2...@1e/pci-...@1/i...@0/c...@1,0
>       2. c5d0 <ST350063-         9QG3AM7-0001-465.76GB>
>          /p...@0,0/pci8086,2...@1e/pci-...@1/i...@1/c...@0,0
>       3. c5d1 <ST350063-         9QG19MY-0001-465.76GB>
>          /p...@0,0/pci8086,2...@1e/pci-...@1/i...@1/c...@1,0
>       4. c6d0 <ST350063-         9QG19VY-0001-465.76GB>
>          /p...@0,0/pci8086,2...@1e/pci-...@2/i...@0/c...@0,0
>       5. c6d1 <ST350063-         5QG019W-0001-465.76GB>
>          /p...@0,0/pci8086,2...@1e/pci-...@2/i...@0/c...@1,0
>       6. c7d0 <ST350063-         9QG1DKF-0001-465.76GB>
>          /p...@0,0/pci8086,2...@1e/pci-...@2/i...@1/c...@0,0
>       7. c7d1 <ST350063-         5QG0B2Y-0001-465.76GB>
>          /p...@0,0/pci8086,2...@1e/pci-...@2/i...@1/c...@1,0
>       8. c8d0 <DEFAULT cyl 9961 alt 2 hd 255 sec 63>
>          /p...@0,0/pci-...@1f,1/i...@0/c...@0,0
>       9. c10d0 <ST350083-         9QG0LR5-0001-465.76GB>
>          /p...@0,0/pci-...@1f,2/i...@0/c...@0,0
>      10. c11d0 <ST350083-         9QG0LW6-0001-465.76GB>
>          /p...@0,0/pci-...@1f,2/i...@1/c...@0,0
> Specify disk (enter its number): ^C
> ke...@opensolaris:~# zpool import
>  pool: storage
>    id: 18058787158441119951
> state: UNAVAIL
> status: The pool was last accessed by another system.
> action: The pool cannot be imported due to damaged devices or data.
>   see: http://www.sun.com/msg/ZFS-8000-EY
> config:
> 
>    storage          UNAVAIL  insufficient replicas
>      raidz2-0       DEGRADED
>        c4d0         ONLINE
>        c4d1         ONLINE
>        c5d0         ONLINE
>        replacing-3  DEGRADED
>          c5d1       ONLINE
>          c5d1       FAULTED  corrupted data
>        c6d0         ONLINE
>        c6d1         ONLINE
>        c7d0         ONLINE
>        c7d1         ONLINE
>        c10d0        ONLINE
>        c11d0        ONLINE
> ke...@opensolaris:~# zpool import -f
>  pool: storage
>    id: 18058787158441119951
> state: UNAVAIL
> status: The pool was last accessed by another system.
> action: The pool cannot be imported due to damaged devices or data.
>   see: http://www.sun.com/msg/ZFS-8000-EY
> config:
> 
>    storage          UNAVAIL  insufficient replicas
>      raidz2-0       DEGRADED
>        c4d0         ONLINE
>        c4d1         ONLINE
>        c5d0         ONLINE
>        replacing-3  DEGRADED
>          c5d1       ONLINE
>          c5d1       FAULTED  corrupted data
>        c6d0         ONLINE
>        c6d1         ONLINE
>        c7d0         ONLINE
>        c7d1         ONLINE
>        c10d0        ONLINE
>        c11d0        ONLINE
> ke...@opensolaris:~# zpool import -f storage
> cannot import 'storage': one or more devices is currently unavailable
>    Destroy and re-create the pool from
>    a backup source.
> 
> 
> Prior to exporting the pool I was able to offline the failed drive.
> 
> Finally about a month ago I upgraded the zpool version to enable dedupe.
> 
> The suggestions I have read include "playing with" the metadata and this is 
> something I would need help with as I am just an "informed" user.
> 
> I am hoping that as only one drive failed and this is a dual parity raid that 
> there is someway to recover the pool.
> 
> Thanks in advance,
> Kevin
> -- 
> This message posted from opensolaris.org
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

ZFS storage and performance consulting at http://www.RichardElling.com
ZFS training on deduplication, NexentaStor, and NAS performance
Las Vegas, April 29-30, 2010 http://nexenta-vegas.eventbrite.com 





_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to