Hi all,

I have a server that is build on top of an Asus board which is equipped with an Areca 1680 HBA. Since ZFS like raw disks, I changed its mode from RAID to JBOD in the firmware and rebootet the host.
Now, I do have 16 drives in the chassis and the line out like this:

root@vsm01:~# format
Searching for disks...done

c5t0d0: configured with capacity of 978.00MB


AVAILABLE DISK SELECTIONS:
       0. c3t1d0 <SEAGATE-ST3300655SS-R001 cyl 36469 alt 2 hd 255 sec 63>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@1,0
       1. c3t1d1 <Seagate-ST3500320NS-R001 cyl 60799 alt 2 hd 255 sec 63>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@1,1
       2. c3t1d2 <Seagate-ST3500320NS-R001-465.76GB>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@1,2
       3. c3t1d3 <Seagate-ST3500320NS-R001 cyl 60799 alt 2 hd 255 sec 63>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@1,3
       4. c3t1d4 <Seagate-ST3500320NS-R001 cyl 60799 alt 2 hd 255 sec 63>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@1,4
       5. c3t1d5 <Seagate-ST3500320NS-R001-465.76GB>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@1,5
       6. c3t1d6 <Seagate-ST3500320NS-R001-465.76GB>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@1,6
       7. c3t1d7 <Seagate-ST3500320NS-R001 cyl 60799 alt 2 hd 255 sec 63>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@1,7
       8. c3t2d0 <Seagate-ST3500320NS-R001-465.76GB>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@2,0
       9. c3t2d1 <Seagate-ST3500320NS-R001-465.76GB>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@2,1
10. c3t2d2 <Hitachi-HUA722010CLA330-R001 cyl 60782 alt 2 hd 255 sec 126>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@2,2
      11. c3t2d3 <Hitachi-HUA722010CLA330-R001-931.51GB>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@2,3
      12. c3t2d4 <Hitachi-HUA722010CLA330-R001-931.51GB>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@2,4
      13. c3t2d5 <Hitachi-HUA722010CLA330-R001-931.51GB>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@2,5
14. c3t2d6 <Hitachi-HUA722010CLA330-R001 cyl 60798 alt 2 hd 255 sec 126>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@2,6
      15. c3t2d7 <Hitachi-HUA722010CLA330-R001-931.51GB>
          /pci@0,0/pci8086,29f1@1/pci17d3,1680@0/disk@2,7
      16. c5t0d0 <Generic-USB EDC-1.00 cyl 978 alt 2 hd 64 sec 32>
          /pci@0,0/pci1043,819e@1d,7/storage@4/disk@0,0

Looked okay for me, so I went ahead an created a zpool containing two mirrors like this:

root@vsm01:~# zpool create vsm_pool1_1T mirror c3t2d1 c3t1d6 mirror c3t2d0 c3t1d5

This went just fine and the zpool was created like this:

root@vsm01:~# zpool status vsm_pool1_1T
  pool: vsm_pool1_1T
 state: ONLINE
  scan: none requested
config:

        NAME        STATE     READ WRITE CKSUM
        vsm_pool1_1T  ONLINE       0     0     0
          mirror-0  ONLINE       0     0     0
            c3t2d1  ONLINE       0     0     0
            c3t1d6  ONLINE       0     0     0
          mirror-1  ONLINE       0     0     0
            c3t2d0  ONLINE       0     0     0
            c3t1d5  ONLINE       0     0     0

errors: No known data errors

Now, creating another zpool from the remaining 500GB drives failed with this weird error:

root@vsm01:~# zpool create vsm_pool2_1T mirror c3t1d4 c3t1d3mirror c3t1d2 c3t1d1
invalid vdev specification
use '-f' to override the following errors:
/dev/dsk/c3t2d1s0 is part of active ZFS pool vsm_pool1_1T. Please see zpool(1M).

Anybody has an idea of what is going wrong here. It doesn't seem to matter which of the drives I want to use for the new zpool, I am always getting this error message - ven when trying only mirror c3t1d4 c3t1d3 or mirror c3t1d2 c3t1d1 alone.

Thanks,
budy






_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to