On Jul 25, 2012, at 7:34 AM, Matt Breitbach wrote: > NFS – iSCSI and FC/FCoE to come once I get it into the proper lab.
ok, so NFS for these tests. I'm not convinced a single ESXi box can drive the load to saturate 10GbE. Also, depending on how you are configuring the system, the I/O that you think is 4KB might look very different coming out of ESXi. Use nfssvrtop or one of the many dtrace one-liners for observing NFS traffic to see what is really on the wire. And I'm very interested to know if you see 16KB reads during the "write-only" workload. more below... > From: Richard Elling [mailto:richard.ell...@gmail.com] > Sent: Tuesday, July 24, 2012 11:36 PM > To: matth...@flash.shanje.com > Cc: zfs-discuss@opensolaris.org > Subject: Re: [zfs-discuss] IO load questions > > Important question, what is the interconnect? iSCSI? FC? NFS? > -- richard > > On Jul 24, 2012, at 9:44 AM, matth...@flash.shanje.com wrote: > > > Working on a POC for high IO workloads, and I’m running in to a bottleneck > that I’m not sure I can solve. Testbed looks like this : > > SuperMicro 6026-6RFT+ barebones w/ dual 5506 CPU’s, 72GB RAM, and ESXi > VM – 4GB RAM, 1vCPU > Connectivity dual 10Gbit Ethernet to Cisco Nexus 5010 > > Target Nexenta system : > > Intel barebones, Dual Xeon 5620 CPU’s, 192GB RAM, Nexenta 3.1.3 Enterprise > Intel x520 dual port 10Gbit Ethernet – LACP Active VPC to Nexus 5010 switches. > 2x LSI 9201-16E HBA’s, 1x LSI 9200-8e HBA > 5 DAE’s (3 in use for this test) > 1 DAE – connected (multipathed) to LSI 9200-8e. Loaded w/ 6x Stec ZeusRAM > SSD’s – striped for ZIL, and 6x OCZ Talos C 230GB drives for L2ARC. > 2 DAE’s connected (multipathed) to one LSI 9201-16E – 24x 600GB 15k Seagate > Cheetah drives > Obviously data integrity is not guaranteed > > Testing using IOMeter from windows guest, 10GB test file, queue depth of 64 > I have a share set up with 4k recordsizes, compression disabled, access time > disabled, and am seeing performance as follows : > > ~50,000 IOPS 4k random read. 200MB/sec, 30% CPU utilization on Nexenta, ~90% > utilization on guest OS. I’m guessing guest OS is bottlenecking. Going to > try physical hardware next week > ~25,000 IOPS 4k random write. 100MB/sec, ~70% CPU utilization on Nexenta, > ~45% CPU utilization on guest OS. Feels like Nexenta CPU is bottleneck. Load > average of 2.5 For cases where you are not bandwidth limited, larger recordsizes can be more efficient. There is no good rule-of-thumb for this, and larger recordsizes will, at some point, hit the bandwidth bottlenecks. I've had good luck with 8KB and 32KB recordsize for ESXi+Windows over NFS. I've never bothered to test 16KB, due to lack of time. > A quick test with 128k recordsizes and 128k IO looked to be 400MB/sec > performance, can’t remember CPU utilization on either side. Will retest and > report those numbers. It would not surprise me to see a CPU bottleneck on the ESXi side at these levels. -- richard > > It feels like something is adding more overhead here than I would expect on > the 4k recordsizes/IO workloads. Any thoughts where I should start on this? > I’d really like to see closer to 10Gbit performance here, but it seems like > the hardware isn’t able to cope with it? > > Theoretical peak performance for a single 10GbE wire is near 300k IOPS @ 4KB, > unidirectional. > This workload is extraordinarily difficult to achieve with a single client > using any of the popular > storage protocols. > -- richard > > -- > ZFS Performance and Training > richard.ell...@richardelling.com > +1-760-896-4422 > > > > > > > -- ZFS Performance and Training richard.ell...@richardelling.com +1-760-896-4422
_______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss