Hi, 

great! Thank you.

With this command, I can see something like that:

/dev/dsk/c3t2d0
LABEL 2
and
LABEL 3 
--------------------------------------------
    version: 15
    name: 'tank'
    state: 0
    txg: 121888
    pool_guid: 4916680739993977334
    hostid: 3848478874
    hostname: 'xxx'
    top_guid: 11305277212649201582
    guid: 11245012245677467134
...

And for
/dev/dsk/c3t2d0s0

LABEL 0-3
--------------------------------------------
    version: 5000
    name: 'raid_10'
    state: 0
    txg: 7494567
    pool_guid: 1071744107269411238
    hostid: 134742016
    hostname: 'dnas'
    top_guid: 565884742309860805
    guid: 13206765472826992314
    vdev_children: 3

So it seems, that /dev/dsk/c3t2d0 have some bad (old) labels, and
/dev/dsk/c3t2d0s0 have correct one. How can I remove the bad unused labels
and how can I avoid this in future when I create new pool with formerly used
disks?

Thank you very much,
Filip


-----Original Message-----
From: Richard Elling [mailto:richard.ell...@richardelling.com] 
Sent: Wednesday, December 03, 2014 6:02 PM
To: Filip Marvan
Cc: Zach Malone; omnios-discuss@lists.omniti.com
Subject: Re: [OmniOS-discuss] zpool import strange informations

Try this instead:
        for i in /dev/dsk/c3t2d0*
        do
                echo $i
                zdb -l $i
        done

This will show you the ZFS labels on each and every slice/partition for the
disk.
Ideally, you'll only see one set of ZFS labels for each disk.
 -- richard

Attachment: smime.p7s
Description: S/MIME cryptographic signature

_______________________________________________
OmniOS-discuss mailing list
OmniOS-discuss@lists.omniti.com
http://lists.omniti.com/mailman/listinfo/omnios-discuss

Reply via email to