Hi Cindy,

Thanks for the advice. This is just a little old Gateway PC provisioned as an 
informal workgroup server. The main storage is two SATA drives in an external 
enclosure, connected to a Sil3132 PCIe eSATA controller. The OS is snv_134b, 
upgraded from snv_111a.

I can't identify a cause in particular. The box has been running for several 
months without much oversight. It's possible that the two eSATA cables got 
reconnected to different ports after a recent move.

The backup has been made and I will try the export & import, per your advice 
(if zpool command works - it does again at the moment, no reboot!). I will also 
try switching the eSATA cables to opposite ports.

Thanks,
Alex


Command output follows:

# format
Searching for disks...done

AVAILABLE DISK SELECTIONS:
       0. c5t1d0 <ATA-WDC WD5000AAKS-0-1D05-465.76GB>
          /pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0
       1. c8d0 <DEFAULT cyl 9726 alt 2 hd 255 sec 63>
          /pci@0,0/pci-ide@1f,2/ide@0/cmdk@0,0
       2. c9d0 <DEFAULT cyl 38910 alt 2 hd 255 sec 63>
          /pci@0,0/pci-ide@1f,2/ide@1/cmdk@0,0
       3. c11t0d0 <WD-Ext HDD 1021-2002-931.51GB>
          /pci@0,0/pci107b,5058@1a,7/storage@1/disk@0,0


# zpool history tank
History for 'tank':
2010-06-18.15:14:16 zpool create tank c13t0d0
2011-05-07.02:00:07 zpool scrub tank
2011-05-14.02:00:08 zpool scrub tank
2011-05-21.02:00:12 zpool scrub tank
<a million 'zfs snapshot' and 'zfs destroy' events from zfs-auto-snap omitted>


# zdb -l /dev/dsk/c5t1d0s0
--------------------------------------------
LABEL 0
--------------------------------------------
    version: 14
    name: 'tank'
    state: 0
    txg: 3374337
    pool_guid: 6242690959503408617
    hostid: 8697169
    hostname: 'wdssandbox'
    top_guid: 17982590661103377266
    guid: 1717308203478351258
    vdev_children: 1
    vdev_tree:
        type: 'mirror'
        id: 0
        guid: 17982590661103377266
        whole_disk: 0
        metaslab_array: 23
        metaslab_shift: 32
        ashift: 9
        asize: 500094468096
        is_log: 0
        children[0]:
            type: 'disk'
            id: 0
            guid: 1717308203478351258
            path: '/dev/dsk/c5t1d0s0'
            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1939879/a'
            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
            whole_disk: 1
            DTL: 27
        children[1]:
            type: 'disk'
            id: 1
            guid: 9267693216478869057
            path: '/dev/dsk/c5t1d0s0'
            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1769949/a'
            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
            whole_disk: 1
            DTL: 893
--------------------------------------------
LABEL 1
--------------------------------------------
    version: 14
    name: 'tank'
    state: 0
    txg: 3374337
    pool_guid: 6242690959503408617
    hostid: 8697169
    hostname: 'wdssandbox'
    top_guid: 17982590661103377266
    guid: 1717308203478351258
    vdev_children: 1
    vdev_tree:
        type: 'mirror'
        id: 0
        guid: 17982590661103377266
        whole_disk: 0
        metaslab_array: 23
        metaslab_shift: 32
        ashift: 9
        asize: 500094468096
        is_log: 0
        children[0]:
            type: 'disk'
            id: 0
            guid: 1717308203478351258
            path: '/dev/dsk/c5t1d0s0'
            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1939879/a'
            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
            whole_disk: 1
            DTL: 27
        children[1]:
            type: 'disk'
            id: 1
            guid: 9267693216478869057
            path: '/dev/dsk/c5t1d0s0'
            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1769949/a'
            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
            whole_disk: 1
            DTL: 893
--------------------------------------------
LABEL 2
--------------------------------------------
    version: 14
    name: 'tank'
    state: 0
    txg: 3374337
    pool_guid: 6242690959503408617
    hostid: 8697169
    hostname: 'wdssandbox'
    top_guid: 17982590661103377266
    guid: 1717308203478351258
    vdev_children: 1
    vdev_tree:
        type: 'mirror'
        id: 0
        guid: 17982590661103377266
        whole_disk: 0
        metaslab_array: 23
        metaslab_shift: 32
        ashift: 9
        asize: 500094468096
        is_log: 0
        children[0]:
            type: 'disk'
            id: 0
            guid: 1717308203478351258
            path: '/dev/dsk/c5t1d0s0'
            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1939879/a'
            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
            whole_disk: 1
            DTL: 27
        children[1]:
            type: 'disk'
            id: 1
            guid: 9267693216478869057
            path: '/dev/dsk/c5t1d0s0'
            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1769949/a'
            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
            whole_disk: 1
            DTL: 893
--------------------------------------------
LABEL 3
--------------------------------------------
    version: 14
    name: 'tank'
    state: 0
    txg: 3374337
    pool_guid: 6242690959503408617
    hostid: 8697169
    hostname: 'wdssandbox'
    top_guid: 17982590661103377266
    guid: 1717308203478351258
    vdev_children: 1
    vdev_tree:
        type: 'mirror'
        id: 0
        guid: 17982590661103377266
        whole_disk: 0
        metaslab_array: 23
        metaslab_shift: 32
        ashift: 9
        asize: 500094468096
        is_log: 0
        children[0]:
            type: 'disk'
            id: 0
            guid: 1717308203478351258
            path: '/dev/dsk/c5t1d0s0'
            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1939879/a'
            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
            whole_disk: 1
            DTL: 27
        children[1]:
            type: 'disk'
            id: 1
            guid: 9267693216478869057
            path: '/dev/dsk/c5t1d0s0'
            devid: 'id1,sd@SATA_____WDC_WD5000AAKS-0_____WD-WCAWF1769949/a'
            phys_path: '/pci@0,0/pci8086,2845@1c,3/pci1095,3132@0/disk@1,0:a'
            whole_disk: 1
            DTL: 893







On May 20, 2011, at 8:34 AM, Cindy Swearingen wrote:

> Hi Alex
> 
> More scary than interesting to me.
> 
> What kind of hardware and which Solaris release?
> 
> Do you know what steps lead up to this problem? Any recent hardware
> changes?
> 
> This output should tell you which disks were in this pool originally:
> 
> # zpool history tank
> 
> If the history identifies tank's actual disks, maybe you can determine
> which disk is masquerading as c5t1d0.
> 
> If that doesn't work, accessing the individual disk entries in format
> should tell which one is the problem, if its only one.
> 
> I would like to see the output of this command:
> 
> # zdb -l /dev/dsk/c5t1d0s0
> 
> Make sure you have a good backup of your data. If you need to pull a
> disk to check cabling, or rule out controller issues, you should
> probably export this pool first. Have a good backup.
> 
> Others have resolved minor device issues by exporting/importing the
> pool but with format/zpool commands hanging on your system, I'm not
> confident that this operation will work for you.
> 
> Thanks,
> 
> Cindy
> 
> On 05/19/11 12:17, Alex wrote:
>> I thought this was interesting - it looks like we have a failing drive in 
>> our mirror, but the two device nodes in the mirror are the same:
>>  pool: tank
>> state: DEGRADED
>> status: One or more devices could not be used because the label is missing or
>>        invalid.  Sufficient replicas exist for the pool to continue
>>        functioning in a degraded state.
>> action: Replace the device using 'zpool replace'.
>>   see: http://www.sun.com/msg/ZFS-8000-4J
>> scrub: scrub completed after 1h9m with 0 errors on Sat May 14 03:09:45 2011
>> config:
>>        NAME        STATE     READ WRITE CKSUM
>>        tank        DEGRADED     0     0     0
>>          mirror-0  DEGRADED     0     0     0
>>            c5t1d0  ONLINE       0     0     0
>>            c5t1d0  FAULTED      0     0     0  corrupted data
>> c5t1d0 does indeed only appear once in the "format" list. I wonder how to go 
>> about correcting this if I can't uniquely identify the failing drive.
>> "format" takes forever to spill its guts, and the zpool commands all 
>> hang...... clearly there is hardware error here, probably causing that, but 
>> not sure how to identify which disk to pull.

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to