Sure enough Cindy, the eSATA cables had been crossed. I exported, powered off, 
reversed the cables, booted, imported, and the pool is currently resilvering 
with both c5t0d0 & c5t1d0 present in the mirror. :) Thank you!!

Alex



On May 24, 2011, at 9:58 AM, Cindy Swearingen wrote:

> Hi Alex,
> 
> If the hardware and cables were moved around then this is probably
> the root cause of your problem. You should see if you can move the
> devices/cabling back to what they were before the move.
> 
> The zpool history output provides the original device name, which
> isn't c5t1d0, either:
> 
> # zpool create tank c13t0d0
> 
> You might grep the zpool history output to find out which disk was
> eventually attached, like this:
> 
> # zpool history | grep attach
> 
> But its clear from the zdb -l output, that the devid for this
> particular device changed, which we've seen happen on some hardware. If
> the devid persists, ZFS can follow the devid of the device even if its
> physical path changes and is able to recover more gracefully.
> 
> If you continue to use this hardware for your storage pool, you should
> export the pool before making any kind of hardware change.
> 
> Thanks,
> 
> Cindy
> 
> 
> On 05/21/11 18:05, Alex Dolski wrote:
>> Hi Cindy,
>> Thanks for the advice. This is just a little old Gateway PC provisioned as 
>> an informal workgroup server. The main storage is two SATA drives in an 
>> external enclosure, connected to a Sil3132 PCIe eSATA controller. The OS is 
>> snv_134b, upgraded from snv_111a.
>> I can't identify a cause in particular. The box has been running for several 
>> months without much oversight. It's possible that the two eSATA cables got 
>> reconnected to different ports after a recent move.
>> The backup has been made and I will try the export & import, per your advice 
>> (if zpool command works - it does again at the moment, no reboot!). I will 
>> also try switching the eSATA cables to opposite ports.
>> Thanks,
>> Alex
>> Command output follows:
>> # format
>> Searching for disks...done
>> AVAILABLE DISK SELECTIONS:
>>       0. c5t1d0 <ATA-WDC WD5000AAKS-0-1D05-465.76GB>
>>          /pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0
>>       1. c8d0 <DEFAULT cyl 9726 alt 2 hd 255 sec 63>
>>          /pci@0,0/pci-ide@1f,2/ide@0/cmdk@0,0
>>       2. c9d0 <DEFAULT cyl 38910 alt 2 hd 255 sec 63>
>>          /pci@0,0/pci-ide@1f,2/ide@1/cmdk@0,0
>>       3. c11t0d0 <WD-Ext HDD 1021-2002-931.51GB>
>>          /pci@0,0/pci107b,5058@1a,7/storage@1/disk@0,0
>> # zpool history tank
>> History for 'tank':
>> 2010-06-18.15:14:16 zpool create tank c13t0d0
>> 2011-05-07.02:00:07 zpool scrub tank
>> 2011-05-14.02:00:08 zpool scrub tank
>> 2011-05-21.02:00:12 zpool scrub tank
>> <a million 'zfs snapshot' and 'zfs destroy' events from zfs-auto-snap 
>> omitted>
>> # zdb -l /dev/dsk/c5t1d0s0
>> --------------------------------------------
>> LABEL 0
>> --------------------------------------------
>>    version: 14
>>    name: 'tank'
>>    state: 0
>>    txg: 3374337
>>    pool_guid: 6242690959503408617
>>    hostid: 8697169
>>    hostname: 'wdssandbox'
>>    top_guid: 17982590661103377266
>>    guid: 1717308203478351258
>>    vdev_children: 1
>>    vdev_tree:
>>        type: 'mirror'
>>        id: 0
>>        guid: 17982590661103377266
>>        whole_disk: 0
>>        metaslab_array: 23
>>        metaslab_shift: 32
>>        ashift: 9
>>        asize: 500094468096
>>        is_log: 0
>>        children[0]:
>>            type: 'disk'
>>            id: 0
>>            guid: 1717308203478351258
>>            path: '/dev/dsk/c5t1d0s0'
>>            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1939879/a'
>>            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
>>            whole_disk: 1
>>            DTL: 27
>>        children[1]:
>>            type: 'disk'
>>            id: 1
>>            guid: 9267693216478869057
>>            path: '/dev/dsk/c5t1d0s0'
>>            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1769949/a'
>>            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
>>            whole_disk: 1
>>            DTL: 893
>> --------------------------------------------
>> LABEL 1
>> --------------------------------------------
>>    version: 14
>>    name: 'tank'
>>    state: 0
>>    txg: 3374337
>>    pool_guid: 6242690959503408617
>>    hostid: 8697169
>>    hostname: 'wdssandbox'
>>    top_guid: 17982590661103377266
>>    guid: 1717308203478351258
>>    vdev_children: 1
>>    vdev_tree:
>>        type: 'mirror'
>>        id: 0
>>        guid: 17982590661103377266
>>        whole_disk: 0
>>        metaslab_array: 23
>>        metaslab_shift: 32
>>        ashift: 9
>>        asize: 500094468096
>>        is_log: 0
>>        children[0]:
>>            type: 'disk'
>>            id: 0
>>            guid: 1717308203478351258
>>            path: '/dev/dsk/c5t1d0s0'
>>            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1939879/a'
>>            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
>>            whole_disk: 1
>>            DTL: 27
>>        children[1]:
>>            type: 'disk'
>>            id: 1
>>            guid: 9267693216478869057
>>            path: '/dev/dsk/c5t1d0s0'
>>            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1769949/a'
>>            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
>>            whole_disk: 1
>>            DTL: 893
>> --------------------------------------------
>> LABEL 2
>> --------------------------------------------
>>    version: 14
>>    name: 'tank'
>>    state: 0
>>    txg: 3374337
>>    pool_guid: 6242690959503408617
>>    hostid: 8697169
>>    hostname: 'wdssandbox'
>>    top_guid: 17982590661103377266
>>    guid: 1717308203478351258
>>    vdev_children: 1
>>    vdev_tree:
>>        type: 'mirror'
>>        id: 0
>>        guid: 17982590661103377266
>>        whole_disk: 0
>>        metaslab_array: 23
>>        metaslab_shift: 32
>>        ashift: 9
>>        asize: 500094468096
>>        is_log: 0
>>        children[0]:
>>            type: 'disk'
>>            id: 0
>>            guid: 1717308203478351258
>>            path: '/dev/dsk/c5t1d0s0'
>>            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1939879/a'
>>            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
>>            whole_disk: 1
>>            DTL: 27
>>        children[1]:
>>            type: 'disk'
>>            id: 1
>>            guid: 9267693216478869057
>>            path: '/dev/dsk/c5t1d0s0'
>>            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1769949/a'
>>            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
>>            whole_disk: 1
>>            DTL: 893
>> --------------------------------------------
>> LABEL 3
>> --------------------------------------------
>>    version: 14
>>    name: 'tank'
>>    state: 0
>>    txg: 3374337
>>    pool_guid: 6242690959503408617
>>    hostid: 8697169
>>    hostname: 'wdssandbox'
>>    top_guid: 17982590661103377266
>>    guid: 1717308203478351258
>>    vdev_children: 1
>>    vdev_tree:
>>        type: 'mirror'
>>        id: 0
>>        guid: 17982590661103377266
>>        whole_disk: 0
>>        metaslab_array: 23
>>        metaslab_shift: 32
>>        ashift: 9
>>        asize: 500094468096
>>        is_log: 0
>>        children[0]:
>>            type: 'disk'
>>            id: 0
>>            guid: 1717308203478351258
>>            path: '/dev/dsk/c5t1d0s0'
>>            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1939879/a'
>>            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
>>            whole_disk: 1
>>            DTL: 27
>>        children[1]:
>>            type: 'disk'
>>            id: 1
>>            guid: 9267693216478869057
>>            path: '/dev/dsk/c5t1d0s0'
>>            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1769949/a'
>>            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
>>            whole_disk: 1
>>            DTL: 893
>> On May 20, 2011, at 8:34 AM, Cindy Swearingen wrote:
>>> Hi Alex
>>> 
>>> More scary than interesting to me.
>>> 
>>> What kind of hardware and which Solaris release?
>>> 
>>> Do you know what steps lead up to this problem? Any recent hardware
>>> changes?
>>> 
>>> This output should tell you which disks were in this pool originally:
>>> 
>>> # zpool history tank
>>> 
>>> If the history identifies tank's actual disks, maybe you can determine
>>> which disk is masquerading as c5t1d0.
>>> 
>>> If that doesn't work, accessing the individual disk entries in format
>>> should tell which one is the problem, if its only one.
>>> 
>>> I would like to see the output of this command:
>>> 
>>> # zdb -l /dev/dsk/c5t1d0s0
>>> 
>>> Make sure you have a good backup of your data. If you need to pull a
>>> disk to check cabling, or rule out controller issues, you should
>>> probably export this pool first. Have a good backup.
>>> 
>>> Others have resolved minor device issues by exporting/importing the
>>> pool but with format/zpool commands hanging on your system, I'm not
>>> confident that this operation will work for you.
>>> 
>>> Thanks,
>>> 
>>> Cindy
>>> 
>>> On 05/19/11 12:17, Alex wrote:
>>>> I thought this was interesting - it looks like we have a failing drive in 
>>>> our mirror, but the two device nodes in the mirror are the same:
>>>> pool: tank
>>>> state: DEGRADED
>>>> status: One or more devices could not be used because the label is missing 
>>>> or
>>>>       invalid.  Sufficient replicas exist for the pool to continue
>>>>       functioning in a degraded state.
>>>> action: Replace the device using 'zpool replace'.
>>>>  see: http://www.sun.com/msg/ZFS-8000-4J
>>>> scrub: scrub completed after 1h9m with 0 errors on Sat May 14 03:09:45 2011
>>>> config:
>>>>       NAME        STATE     READ WRITE CKSUM
>>>>       tank        DEGRADED     0     0     0
>>>>         mirror-0  DEGRADED     0     0     0
>>>>           c5t1d0  ONLINE       0     0     0
>>>>           c5t1d0  FAULTED      0     0     0  corrupted data
>>>> c5t1d0 does indeed only appear once in the "format" list. I wonder how to 
>>>> go about correcting this if I can't uniquely identify the failing drive.
>>>> "format" takes forever to spill its guts, and the zpool commands all 
>>>> hang...... clearly there is hardware error here, probably causing that, 
>>>> but not sure how to identify which disk to pull.

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to