iostat doesn't show any high service times and fsstat also shows low throughput. Occasionally I can generate enough load that you do see some very high asvc_t but when that occurs the pool is performing as expected. As a precaution I just added two extra drives to the zpool incase zfs was having difficulty finding a location to allocate new blocks and it has made no difference. You can still see still see block allocation cycling evenly across the available mirrors.
iostat -xnz 3 before, during, and after the blocking extended device statistics r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device 0.0 33.3 0.0 184.1 0.0 0.1 0.0 1.5 0 1 c12t0d0 0.0 15.3 0.0 98.7 0.0 0.0 0.0 1.0 0 0 c12t1d0 0.0 15.3 0.0 98.7 0.0 0.0 0.0 1.1 0 0 c12t2d0 0.0 14.3 0.0 98.7 0.0 0.0 0.0 1.0 0 0 c12t6d0 0.0 20.7 0.0 256.6 0.0 0.0 0.0 1.4 0 0 c11t5d0 0.0 48.0 0.0 273.4 0.0 0.1 0.0 1.3 0 1 c11t6d0 0.0 34.0 0.0 199.9 0.0 0.0 0.0 1.5 0 1 c11t7d0 0.0 20.3 0.0 256.6 0.0 0.0 0.0 1.4 0 0 c10t5d0 0.0 47.7 0.0 273.4 0.0 0.1 0.0 1.2 0 1 c10t6d0 0.0 34.0 0.0 199.9 0.0 0.0 0.0 1.5 0 1 c10t7d0 0.0 9.7 0.0 1237.9 0.0 0.0 0.0 4.7 0 1 c14t0d0 0.0 33.3 0.0 184.1 0.0 0.1 0.0 1.8 0 1 c13t0d0 0.0 15.7 0.0 98.7 0.0 0.0 0.0 1.2 0 0 c13t1d0 0.0 15.7 0.0 98.7 0.0 0.0 0.0 1.0 0 0 c13t2d0 0.0 13.7 0.0 98.7 0.0 0.0 0.0 1.1 0 0 c13t6d0 extended device statistics r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c8t1d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c8t2d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c8t3d0 0.3 0.3 1.7 0.0 0.0 0.0 0.0 6.0 0 0 c8t4d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c8t5d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c8t6d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c8t7d0 0.0 2.0 0.0 1.3 0.0 0.0 0.0 0.2 0 0 c12t0d0 0.0 15.7 0.0 235.9 0.0 0.0 0.0 0.8 0 0 c12t1d0 0.3 16.3 1.8 235.9 0.0 0.0 0.0 1.2 0 1 c12t2d0 0.3 15.0 2.8 130.6 0.0 0.0 0.0 1.0 0 1 c12t3d0 0.0 11.7 0.0 127.9 0.0 0.0 0.0 0.8 0 0 c12t4d0 0.3 0.3 0.2 0.0 0.0 0.0 0.0 9.0 0 1 c12t5d0 0.0 40.3 0.0 599.9 0.0 0.0 0.0 0.8 0 1 c12t6d0 0.3 0.3 2.8 0.0 0.0 0.0 0.0 3.2 0 0 c11t0d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c11t1d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c11t2d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c11t3d0 0.7 0.3 3.0 0.0 0.0 0.0 0.0 6.5 0 1 c11t4d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c11t5d0 0.3 2.0 1.5 1.3 0.0 0.0 0.0 2.8 0 1 c11t6d0 0.0 2.0 0.0 1.3 0.0 0.0 0.0 0.2 0 0 c11t7d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c10t0d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c10t1d0 0.3 0.3 0.2 0.0 0.0 0.0 0.0 6.7 0 0 c10t2d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c10t3d0 0.3 0.3 1.5 0.0 0.0 0.0 0.0 9.1 0 1 c10t4d0 0.3 0.3 1.5 0.0 0.0 0.0 0.0 7.7 0 1 c10t5d0 0.3 2.0 1.5 1.3 0.0 0.0 0.0 2.2 0 0 c10t6d0 0.0 2.0 0.0 1.3 0.0 0.0 0.0 0.2 0 0 c10t7d0 0.0 0.3 0.0 42.6 0.0 0.0 0.0 0.7 0 0 c14t1d0 0.3 0.3 1.5 0.0 0.0 0.0 0.0 5.9 0 0 c9t1d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c9t2d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c9t3d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c9t4d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c9t5d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c9t6d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c9t7d0 0.0 2.0 0.0 1.3 0.0 0.0 0.0 0.4 0 0 c13t0d0 0.0 15.7 0.0 235.9 0.0 0.0 0.0 0.8 0 0 c13t1d0 0.0 16.3 0.0 235.9 0.0 0.0 0.0 0.9 0 0 c13t2d0 0.0 14.7 0.0 130.6 0.0 0.0 0.0 0.7 0 0 c13t3d0 0.0 12.0 0.0 127.9 0.0 0.0 0.0 0.9 0 0 c13t4d0 0.0 0.3 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c13t5d0 0.0 39.6 0.0 599.9 0.0 0.0 0.0 0.8 0 1 c13t6d0 extended device statistics r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device 0.0 12.0 0.0 45.2 0.0 0.0 0.0 0.4 0 0 c8t1d0 0.0 32.3 0.0 175.7 0.0 0.1 0.0 2.4 0 1 c8t2d0 0.0 34.3 0.0 175.7 0.0 0.1 0.0 3.1 0 1 c8t3d0 0.0 32.3 0.0 175.7 0.0 0.1 0.0 2.6 0 1 c8t4d0 0.0 24.3 0.0 203.8 0.0 0.0 0.0 1.0 0 1 c12t4d0 0.0 23.3 0.0 142.0 0.0 0.0 0.0 0.8 0 0 c12t5d0 0.0 12.3 0.0 45.2 0.0 0.0 0.0 0.4 0 0 c9t1d0 0.0 31.3 0.0 175.7 0.0 0.1 0.0 2.2 0 1 c9t2d0 0.0 33.3 0.0 175.7 0.0 0.1 0.0 2.2 0 1 c9t3d0 0.0 31.7 0.0 175.7 0.0 0.1 0.0 2.5 0 1 c9t4d0 0.0 24.0 0.0 203.8 0.0 0.0 0.0 0.9 0 1 c13t4d0 0.0 23.0 0.0 142.0 0.0 0.0 0.0 0.8 0 0 c13t5d0 extended device statistics > > is 'iostat -en' error free? it is error free for all devices On May 4, 2011, at 12:28 PM, Michael Schuster wrote: > On Wed, May 4, 2011 at 21:21, Adam Serediuk <asered...@gmail.com> wrote: >> We have an X4540 running Solaris 11 Express snv_151a that has developed an >> issue where its write performance is absolutely abysmal. Even touching a >> file takes over five seconds both locally and remotely. >> >> /pool1/data# time touch foo >> >> real 0m5.305s >> user 0m0.001s >> sys 0m0.004s >> /pool1/data# time rm foo >> >> real 0m5.912s >> user 0m0.001s >> sys 0m0.005s >> >> The system exhibits this issue under the slightest load. We have >> sync=disabled set on all filesystems in this pool. The pool is at 75% >> capacity and is healthy. This issue started suddenly several days ago and >> persists after reboot. prstat shows zpool-pool1/150 taking 10% CPU >> constantly whereas other similar systems in our infrastructure under the >> same load do not. Even doing a 'zfs set' on a property takes up to 10 >> seconds and on other systems is instantaneous. Something appears to be >> blocking internally. >> >> Both iostat and zpool iostat show very little to zero load on the devices >> even while blocking. > > it'd be interesting to see a completet iostat run while this is > happening - I'd suggest 'iostat -xnz 3' or sth like that. look for > high svc_t values ... > > HTH > Michael > -- > regards/mit freundlichen GrĂ¼ssen > Michael Schuster _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss