What more info could you provide? Quite a lot more, actually, like: how many streams of SQL and copy are you running? how are the filesystems/zvols configured (recordsize, etc)? some CPU, VM and network stats would also be nice.
Based on the nexenta iostats you've provided (a tiny window on what's happening), it appears that you have an 8k recordsize for SQL. If you add up all the IOPS for the SQL, it's roughly 2000 reads at around 3ms each. Which might indicate at least 6 reads outstanding at any time. So how many queries do you have running in parallel? If you add more, I'd expect the service times to increase. 3ms isn't much for spinning rust, but isn't this why you are planning to use lots of L2ARC? Could be a similar story on writes. How many parallel streams? How many files? What's the average file size? What's the client filesysyem? How much does it sync to the server? Could it be that your client apps are always waiting for the spinning rust? Does an SSD log make any difference on this pool? Sent from my iPhone On 22 Oct 2010, at 19:57, Ian D <rewar...@hotmail.com> wrote: > Some numbers... > > zpool status > > pool: Pool_sas > state: ONLINE > scan: none requested > config: > > NAME STATE READ WRITE CKSUM > Pool_sas ONLINE 0 0 0 > c4t5000C5000006A6D3d0 ONLINE 0 0 0 > c4t5000C5000006A777d0 ONLINE 0 0 0 > c4t5000C5000006AA43d0 ONLINE 0 0 0 > c4t5000C5000006AC4Fd0 ONLINE 0 0 0 > c4t5000C5000006AEF7d0 ONLINE 0 0 0 > c4t5000C5000006B27Fd0 ONLINE 0 0 0 > c4t5000C5000006B28Bd0 ONLINE 0 0 0 > c4t5000C5000006B46Bd0 ONLINE 0 0 0 > c4t5000C5000006B563d0 ONLINE 0 0 0 > c4t5000C5000006B643d0 ONLINE 0 0 0 > c4t5000C5000006B6D3d0 ONLINE 0 0 0 > c4t5000C5000006BBE7d0 ONLINE 0 0 0 > c4t5000C5000006C407d0 ONLINE 0 0 0 > c4t5000C5000006C657d0 ONLINE 0 0 0 > > errors: No known data errors > > pool: Pool_test > state: ONLINE > scan: none requested > config: > > NAME STATE READ WRITE CKSUM > Pool_test ONLINE 0 0 0 > c4t5000C5002103F093d0 ONLINE 0 0 0 > c4t5000C50021101683d0 ONLINE 0 0 0 > c4t5000C50021102AA7d0 ONLINE 0 0 0 > c4t5000C500211034D3d0 ONLINE 0 0 0 > c4t5000C500211035DFd0 ONLINE 0 0 0 > c4t5000C5002110480Fd0 ONLINE 0 0 0 > c4t5000C50021104F0Fd0 ONLINE 0 0 0 > c4t5000C50021119A43d0 ONLINE 0 0 0 > c4t5000C5002112392Fd0 ONLINE 0 0 0 > > errors: No known data errors > > pool: syspool > state: ONLINE > scan: none requested > config: > > NAME STATE READ WRITE CKSUM > syspool ONLINE 0 0 0 > c0t0d0s0 ONLINE 0 0 0 > > errors: No known data errors > ========================================================= > > Pool_sas is made of 14x 146G 15K SAS Drives in a big stripe. For this test > there is no log device or cache. Connected to it is a RedHat box using iSCSI > through an Intel X520 10GbE NIC. It runs several large MySQL queries at once- > each taking minutes to compute. > > Pool_test is a stripe of 2TB SATA drives and a terrabyte of files is being > copied to it for another box during this test. > > Here's the pastebin of "iostat -xdn 10" on the Linux box: > http://pastebin.com/431ESYaz > > Here's the pastebin of "iostat -xdn 10" on the Nexenta box: > http://pastebin.com/9g7KD3Ku > > Here's the pastebin "zpool iostat -v 10" on the Nexenta box: > http://pastebin.com/05fJL5sw > > From these numbers it looks like the Linux box is waiting for data all the > time while the Nexenta box isn't pulling nearly as much throughput and IOPS > as it could. Where is the bottleneck? > > One thing suspicious is that we notice a slow down of one pool when the other > is under load. How can that be? > > Ian > -- > This message posted from opensolaris.org > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss