Ok I played around with the physical configuration and placed them on the original controller and zdb -l is now able to unpack LABEL 0,1,2,3 for all drives in the pool. I also changed the hostname in opensolaris to "freenas.local" as that is what was listed in the zdb -l(although I doubt this matters).
The new setup looks like this: @freenas:~/dskp0s# zpool import pool: Raidz id: 14119036174566039103 state: FAULTED status: The pool metadata is corrupted. action: The pool cannot be imported due to damaged devices or data. The pool may be active on another system, but can be imported using the '-f' flag. see: http://www.sun.com/msg/ZFS-8000-72 config: Raidz FAULTED corrupted data raidz1-0 FAULTED corrupted data c5t0d0p0 ONLINE c5t1d0p0 ONLINE c5t2d0s2 ONLINE c5t3d0p0 ONLINE c5t4d0p0 ONLINE c5t5d0p0 ONLINE @freenas:~/dskp0s# ls -l /dev/dsk/c5* lrwxrwxrwx 1 root root 62 Jul 23 08:05 /dev/dsk/c5t0d0p0 -> ../../devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@0,0:q lrwxrwxrwx 1 root root 62 Jul 23 08:05 /dev/dsk/c5t1d0p0 -> ../../devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@1,0:q lrwxrwxrwx 1 root root 62 Jul 23 08:05 /dev/dsk/c5t2d0s2 -> ../../devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@2,0:c lrwxrwxrwx 1 root root 62 Jul 23 08:05 /dev/dsk/c5t3d0p0 -> ../../devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@3,0:q lrwxrwxrwx 1 root root 62 Jul 23 08:05 /dev/dsk/c5t4d0p0 -> ../../devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@4,0:q lrwxrwxrwx 1 root root 62 Jul 23 08:05 /dev/dsk/c5t5d0p0 -> ../../devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@5,0:q The output of the above command was edited to show only the devices listed in the pool. I then made symlinks to the devices directly as follows in a dir call /dskp0s @freenas:~/dskp0s# ls -l total 17 lrwxrwxrwx 1 root root 57 Jul 23 08:40 aacdu0 -> /devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@0,0:q lrwxrwxrwx 1 root root 57 Jul 23 08:40 aacdu1 -> /devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@1,0:q lrwxrwxrwx 1 root root 57 Jul 23 08:40 aacdu2 -> /devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@2,0:q lrwxrwxrwx 1 root root 57 Jul 23 08:40 aacdu3 -> /devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@3,0:q lrwxrwxrwx 1 root root 57 Jul 23 08:41 aacdu4 -> /devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@4,0:q lrwxrwxrwx 1 root root 57 Jul 23 08:41 aacdu5 -> /devices/p...@0,0/pci8086,2...@1e/pci9005,2...@2/d...@5,0:q -rw-r--r-- 1 root root 1992 Jul 26 07:30 zpool.cache Note: aacdu2 symlink was linked to d...@2,0:q instead of d...@2,0:c because in FreeNAS the disks should be identical(maybe this is part of the problem)? zdb -l completes with either symlink. then I ran these commands from /dskp0s directory @freenas:~/dskp0s# zpool import Raidz cannot import 'Raidz': pool may be in use from other system use '-f' to import anyway @freenas:~/dskp0s# zpool import -d . Raidz cannot import 'Raidz': pool may be in use from other system use '-f' to import anyway @freenas:~/dskp0s# zpool import -f Raidz cannot import 'Raidz': one or more devices is currently unavailable Destroy and re-create the pool from a backup source. @freenas:~/dskp0s# zpool import -d . -f Raidz cannot import 'Raidz': one or more devices is currently unavailable Destroy and re-create the pool from a backup source. @freenas:~/dskp0s# zpool import -F Raidz cannot import 'Raidz': pool may be in use from other system use '-f' to import anyway @freenas:~/dskp0s# zpool import -d . -F Raidz cannot import 'Raidz': pool may be in use from other system use '-f' to import anyway @freenas:~/dskp0s# zdb -l aacdu0 -------------------------------------------- LABEL 0 -------------------------------------------- version: 6 name: 'Raidz' state: 0 txg: 11730350 pool_guid: 14119036174566039103 hostid: 0 hostname: 'freenas.local' top_guid: 16879648846521942561 guid: 6543046729241888600 vdev_tree: type: 'raidz' id: 0 guid: 16879648846521942561 nparity: 1 metaslab_array: 14 metaslab_shift: 32 ashift: 9 asize: 6000992059392 children[0]: type: 'disk' id: 0 guid: 6543046729241888600 path: '/dev/aacdu0' whole_disk: 0 children[1]: type: 'disk' id: 1 guid: 14313209149820231630 path: '/dev/aacdu1' whole_disk: 0 children[2]: type: 'disk' id: 2 guid: 5383435113781649515 path: '/dev/aacdu2' whole_disk: 0 children[3]: type: 'disk' id: 3 guid: 9586044621389086913 path: '/dev/aacdu3' whole_disk: 0 DTL: 1372 children[4]: type: 'disk' id: 4 guid: 10401318729908601665 path: '/dev/aacdu4' whole_disk: 0 children[5]: type: 'disk' id: 5 guid: 6282477769796963197 path: '/dev/aacdu5' whole_disk: 0 -------------------------------------------- LABEL 1 -------------------------------------------- version: 6 name: 'Raidz' state: 0 txg: 11730350 pool_guid: 14119036174566039103 hostid: 0 hostname: 'freenas.local' top_guid: 16879648846521942561 guid: 6543046729241888600 vdev_tree: type: 'raidz' id: 0 guid: 16879648846521942561 nparity: 1 metaslab_array: 14 metaslab_shift: 32 ashift: 9 asize: 6000992059392 children[0]: type: 'disk' id: 0 guid: 6543046729241888600 path: '/dev/aacdu0' whole_disk: 0 children[1]: type: 'disk' id: 1 guid: 14313209149820231630 path: '/dev/aacdu1' whole_disk: 0 children[2]: type: 'disk' id: 2 guid: 5383435113781649515 path: '/dev/aacdu2' whole_disk: 0 children[3]: type: 'disk' id: 3 guid: 9586044621389086913 path: '/dev/aacdu3' whole_disk: 0 DTL: 1372 children[4]: type: 'disk' id: 4 guid: 10401318729908601665 path: '/dev/aacdu4' whole_disk: 0 children[5]: type: 'disk' id: 5 guid: 6282477769796963197 path: '/dev/aacdu5' whole_disk: 0 -------------------------------------------- LABEL 2 -------------------------------------------- version: 6 name: 'Raidz' state: 0 txg: 11730350 pool_guid: 14119036174566039103 hostid: 0 hostname: 'freenas.local' top_guid: 16879648846521942561 guid: 6543046729241888600 vdev_tree: type: 'raidz' id: 0 guid: 16879648846521942561 nparity: 1 metaslab_array: 14 metaslab_shift: 32 ashift: 9 asize: 6000992059392 children[0]: type: 'disk' id: 0 guid: 6543046729241888600 path: '/dev/aacdu0' whole_disk: 0 children[1]: type: 'disk' id: 1 guid: 14313209149820231630 path: '/dev/aacdu1' whole_disk: 0 children[2]: type: 'disk' id: 2 guid: 5383435113781649515 path: '/dev/aacdu2' whole_disk: 0 children[3]: type: 'disk' id: 3 guid: 9586044621389086913 path: '/dev/aacdu3' whole_disk: 0 DTL: 1372 children[4]: type: 'disk' id: 4 guid: 10401318729908601665 path: '/dev/aacdu4' whole_disk: 0 children[5]: type: 'disk' id: 5 guid: 6282477769796963197 path: '/dev/aacdu5' whole_disk: 0 -------------------------------------------- LABEL 3 -------------------------------------------- version: 6 name: 'Raidz' state: 0 txg: 11730350 pool_guid: 14119036174566039103 hostid: 0 hostname: 'freenas.local' top_guid: 16879648846521942561 guid: 6543046729241888600 vdev_tree: type: 'raidz' id: 0 guid: 16879648846521942561 nparity: 1 metaslab_array: 14 metaslab_shift: 32 ashift: 9 asize: 6000992059392 children[0]: type: 'disk' id: 0 guid: 6543046729241888600 path: '/dev/aacdu0' whole_disk: 0 children[1]: type: 'disk' id: 1 guid: 14313209149820231630 path: '/dev/aacdu1' whole_disk: 0 children[2]: type: 'disk' id: 2 guid: 5383435113781649515 path: '/dev/aacdu2' whole_disk: 0 children[3]: type: 'disk' id: 3 guid: 9586044621389086913 path: '/dev/aacdu3' whole_disk: 0 DTL: 1372 children[4]: type: 'disk' id: 4 guid: 10401318729908601665 path: '/dev/aacdu4' whole_disk: 0 children[5]: type: 'disk' id: 5 guid: 6282477769796963197 path: '/dev/aacdu5' whole_disk: 0 The zpool.cache file is from the original machine, I copied it to try "zpool import -c" but that didn't work. Zpool import -f doesn't work even if I specify a directory and states that one or more devices is unavailable (am I symlinking incorrectly) Zpool import -F states that pool may be in use from other system use '-f' to import anyway (which doesn't work). (I tried to place the drives back on the original FreeNAS setup and export but nothing happens. "Zpool export Raidz" does not return any output I have not tried any other command as I am afraid of causing more damage. Any suggestions would be appreciated. -- This message posted from opensolaris.org _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss