Hi, great! Thank you.
With this command, I can see something like that: /dev/dsk/c3t2d0 LABEL 2 and LABEL 3 -------------------------------------------- version: 15 name: 'tank' state: 0 txg: 121888 pool_guid: 4916680739993977334 hostid: 3848478874 hostname: 'xxx' top_guid: 11305277212649201582 guid: 11245012245677467134 ... And for /dev/dsk/c3t2d0s0 LABEL 0-3 -------------------------------------------- version: 5000 name: 'raid_10' state: 0 txg: 7494567 pool_guid: 1071744107269411238 hostid: 134742016 hostname: 'dnas' top_guid: 565884742309860805 guid: 13206765472826992314 vdev_children: 3 So it seems, that /dev/dsk/c3t2d0 have some bad (old) labels, and /dev/dsk/c3t2d0s0 have correct one. How can I remove the bad unused labels and how can I avoid this in future when I create new pool with formerly used disks? Thank you very much, Filip -----Original Message----- From: Richard Elling [mailto:richard.ell...@richardelling.com] Sent: Wednesday, December 03, 2014 6:02 PM To: Filip Marvan Cc: Zach Malone; omnios-discuss@lists.omniti.com Subject: Re: [OmniOS-discuss] zpool import strange informations Try this instead: for i in /dev/dsk/c3t2d0* do echo $i zdb -l $i done This will show you the ZFS labels on each and every slice/partition for the disk. Ideally, you'll only see one set of ZFS labels for each disk. -- richard
smime.p7s
Description: S/MIME cryptographic signature
_______________________________________________ OmniOS-discuss mailing list OmniOS-discuss@lists.omniti.com http://lists.omniti.com/mailman/listinfo/omnios-discuss