shawn@indianapolis:~$ pfexec format Searching for disks...done AVAILABLE DISK SELECTIONS: 0. c1d0 <WDC WD30- WD-WCAWZ084341-0001-2.73TB> /pci@0,0/pci-ide@11/ide@0/cmdk@0,0 1. c1d1 <WDC WD30- WD-WCAWZ087742-0001-2.73TB> /pci@0,0/pci-ide@11/ide@0/cmdk@1,0 2. c2d0 <Unknown-Unknown-0001 cyl 30397 alt 2 hd 255 sec 63> /pci@0,0/pci-ide@11/ide@1/cmdk@0,0 3. c2d1 <ST315003- 9VS25XB-0001-1.36TB> /pci@0,0/pci-ide@11/ide@1/cmdk@1,0 4. c4d0 <Unknown-Unknown-0001 cyl 30397 alt 2 hd 255 sec 63> /pci@0,0/pci-ide@14,1/ide@1/cmdk@0,0 Specify disk (enter its number): ^C shawn@indianapolis:~$ zpool status tank pool: tank state: ONLINE scan: scrub repaired 0 in 4h53m with 0 errors on Tue Dec 20 12:03:36 2011 config:
NAME STATE READ WRITE CKSUM tank ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 c1d0 ONLINE 0 0 0 c1d1 ONLINE 0 0 0 c2d1 ONLINE 0 0 0 errors: No known data errors On Tue, Jan 17, 2012 at 9:18 AM, Tom Evans <tevans...@googlemail.com> wrote: > On Tue, Jan 17, 2012 at 4:00 PM, Christer Solskogen > <christer.solsko...@gmail.com> wrote: >> A overhead of almost 300GB? That seems a bit to much, don't you think? >> The pool consist of one vdev with two 1,5TB disks and one 3TB in raidz1. >> > > Confused about your disks - can you show the output of zpool status. > > If you have a raidz of N disks with a minimum size of Y GB, you can > expect ``zpool list'' to show a size of N*Y and ``zfs list'' to show a > size of roughly (N-1)*Y. > > So, on my box with 2 x 6 x 1.5 TB drives in raidz, I see a zpool size > of 16.3 TB, and a zfs size of 13.3 TB. > > Cheers > > Tom _______________________________________________ freebsd-stable@freebsd.org mailing list http://lists.freebsd.org/mailman/listinfo/freebsd-stable To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"