Hi all,

We are running pretty large vdevs since the initial testing showed that
our setup was not too much off the optimum. However, under real world
load we do see quite some weird behaviour:

The system itself is a X4500 with 500 GB drives and right now the system
seems to be under heavy load, e.g. ls takes minutes to return on only a
few hundred entries, top shows 10% kernel, rest idle.

zpool ioststat -v atlashome 60 shows (not the first output):

              capacity     operations    bandwidth
pool         used  avail   read  write   read  write
----------  -----  -----  -----  -----  -----  -----
atlashome   2.11T  18.8T  2.29K     36  71.7M   138K
  raidz2     466G  6.36T    493     11  14.9M  34.1K
    c0t0d0      -      -     48      5  1.81M  3.52K
    c1t0d0      -      -     48      5  1.81M  3.46K
    c4t0d0      -      -     48      5  1.81M  3.27K
    c6t0d0      -      -     48      5  1.81M  3.40K
    c7t0d0      -      -     47      5  1.81M  3.40K
    c0t1d0      -      -     47      5  1.81M  3.20K
    c1t1d0      -      -     47      6  1.81M  3.59K
    c4t1d0      -      -     47      6  1.81M  3.53K
    c5t1d0      -      -     47      5  1.81M  3.33K
    c6t1d0      -      -     48      6  1.81M  3.67K
    c7t1d0      -      -     48      6  1.81M  3.66K
    c0t2d0      -      -     48      5  1.82M  3.42K
    c1t2d0      -      -     48      6  1.81M  3.56K
    c4t2d0      -      -     48      6  1.81M  3.54K
    c5t2d0      -      -     48      5  1.81M  3.41K
  raidz2     732G  6.10T    800     12  24.6M  52.3K
    c6t2d0      -      -    139      5  7.52M  4.54K
    c7t2d0      -      -    139      5  7.52M  4.81K
    c0t3d0      -      -    140      5  7.52M  4.98K
    c1t3d0      -      -    139      5  7.51M  4.47K
    c4t3d0      -      -    139      5  7.51M  4.82K
    c5t3d0      -      -    139      5  7.51M  4.99K
    c6t3d0      -      -    139      5  7.52M  4.44K
    c7t3d0      -      -    139      5  7.52M  4.78K
    c0t4d0      -      -    139      5  7.52M  4.97K
    c1t4d0      -      -    139      5  7.51M  4.60K
    c4t4d0      -      -    139      5  7.51M  4.86K
    c6t4d0      -      -    139      5  7.51M  4.99K
    c7t4d0      -      -    139      5  7.51M  4.52K
    c0t5d0      -      -    139      5  7.51M  4.78K
    c1t5d0      -      -    138      5  7.51M  4.94K
  raidz2     960G  6.31T  1.02K     12  32.2M  52.0K
    c4t5d0      -      -    178      5  9.29M  4.79K
    c5t5d0      -      -    178      5  9.28M  4.64K
    c6t5d0      -      -    179      5  9.29M  4.44K
    c7t5d0      -      -    178      4  9.26M  4.26K
    c0t6d0      -      -    178      5  9.28M  4.78K
    c1t6d0      -      -    178      5  9.20M  4.58K
    c4t6d0      -      -    178      5  9.26M  4.25K
    c5t6d0      -      -    177      4  9.21M  4.18K
    c6t6d0      -      -    178      5  9.29M  4.69K
    c7t6d0      -      -    177      5  9.26M  4.61K
    c0t7d0      -      -    177      5  9.29M  4.34K
    c1t7d0      -      -    177      5  9.24M  4.28K
    c4t7d0      -      -    177      5  9.29M  4.78K
    c5t7d0      -      -    177      5  9.27M  4.75K
    c6t7d0      -      -    177      5  9.29M  4.34K
    c7t7d0      -      -    177      5  9.27M  4.28K
----------  -----  -----  -----  -----  -----  -----

Questions:
(a) Why the first vdev does not get an equal share of the load
(b) Why is a large raidz2 so bad? When I use a standard Linux box with
hardware raid6 over 16 disks I usually get more bandwidth and at least
about the same small file performance
(c) Would the use of several smaller vdev would help much? And which
layout would be a good compromise for getting space as well as
performance and reliability? 46 disks have so few prime factors

Thanks a lot

Carsten
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to