This system is for serving VM images through iSCSI to roughly 30 xenserver hosts. I would like to know what type of performance I can expect in the coming months as we grow this system out. We currently have 2 intel ssds mirrored for the zil and 2 intel ssds for the l2arc in a stripe. I am interested more in max throughput of the local storage at this point and time.
On Wed, Aug 10, 2011 at 12:01 PM, Roy Sigurd Karlsbakk <r...@karlsbakk.net> wrote: > What sort of load will this server be serving? sync or async writes? what > sort of reads? random i/o or sequential? if sequential, how many > streams/concurrent users? those are factors you need to evaluate before > running a test. A local test will usually be using async i/o and a dd with > only 4k blocksize is bound to be slow, probably because of cpu overhead. > > roy > > ----- Original Message ----- >> Hello All, >> Sorry for the lack of information. Here is some answers to some >> questions: >> 1) createPool.sh: >> essentially can take 2 params, one is number of disks in pool, the >> second is either blank or mirrored, blank means number of disks in the >> pool i.e. raid 0, mirrored makes 2 disk mirrors. >> >> #!/bin/sh >> disks=( `cat diskList | grep Hitachi | awk '{print $2}' | tr '\n' ' '` >> ) >> #echo ${disks[1]} >> #$useDisks=" " >> for (( i = 0; i < $1; i++ )) >> do >> #echo "Thus far: "$useDisks >> if [ "$2" = "mirrored" ] >> then >> if [ $(($i % 2)) -eq 0 ] >> then >> useDisks="$useDisks mirror ${disks[i]}" >> else >> useDisks=$useDisks" "${disks[i]} >> fi >> else >> useDisks=$useDisks" "${disks[i]} >> fi >> >> if [ $(($i - $1)) -le 2 ] >> then >> echo "spares are: ${disks[i]}" >> fi >> done >> >> #echo $useDisks >> zpool create -f fooPool0 $useDisks >> >> >> >> 2) hardware: >> Each server attached to each storage array is a dell r710 with 32 GB >> memory each. To test for issues with another platform the below info, >> is from a dell 1950 server with 8GB memory. However, I see similar >> results from the r710s as well. >> >> >> 3) In order to deal with caching, I am writing larger amounts of data >> to the disk then I have memory for. >> >> 4) I have tested with bonnie++ as well and here are the results, i >> have read that it is best to test with 4x the amount of memory: >> /usr/local/sbin/bonnie++ -s 32000 -d /fooPool0/test -u gdurham >> Using uid:101, gid:10. >> Writing with putc()...done >> Writing intelligently...done >> Rewriting...done >> Reading with getc()...done >> Reading intelligently...done >> start 'em...done...done...done... >> Create files in sequential order...done. >> Stat files in sequential order...done. >> Delete files in sequential order...done. >> Create files in random order...done. >> Stat files in random order...done. >> Delete files in random order...done. >> Version 1.03d ------Sequential Output------ --Sequential Input- >> --Random- >> -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- >> Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec >> %CP >> cm-srfe03 32000M 230482 97 477644 76 223687 44 209868 91 541182 >> 41 1900 5 >> ------Sequential Create------ --------Random Create-------- >> -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- >> files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP >> 16 29126 100 +++++ +++ +++++ +++ 24761 100 +++++ +++ +++++ +++ >> cm-srfe03,32000M,230482,97,477644,76,223687,44,209868,91,541182,41,1899.7,5,16,29126,100,+++++,+++,+++++,+++,24761,100,+++++,+++,+++++,+++ >> >> >> I will run these with the r710 server as well and will report the >> results. >> >> Thanks for the help! >> >> -Greg >> >> >> >> On Wed, Aug 10, 2011 at 9:16 AM, phil.har...@gmail.com >> <phil.har...@gmail.com> wrote: >> > I would generally agree that dd is not a great benchmarking tool, >> > but you >> > could use multiple instances to multiple files, and larger block >> > sizes are >> > more efficient. And it's always good to check iostat and mpstat for >> > io and >> > cpu bottlenecks. Also note that an initial run that creates files >> > may be >> > quicker because it just allocates blocks, whereas subsequent >> > rewrites >> > require copy-on-write. >> > >> > ----- Reply message ----- >> > From: "Peter Tribble" <peter.trib...@gmail.com> >> > To: "Gregory Durham" <gregory.dur...@gmail.com> >> > Cc: <zfs-discuss@opensolaris.org> >> > Subject: [zfs-discuss] Issues with supermicro >> > Date: Wed, Aug 10, 2011 10:56 >> > >> > >> > On Wed, Aug 10, 2011 at 1:45 AM, Gregory Durham >> > <gregory.dur...@gmail.com> wrote: >> >> Hello, >> >> We just purchased two of the sc847e26-rjbod1 units to be used in a >> >> storage environment running Solaris 11 express. >> >> >> >> We are using Hitachi HUA723020ALA640 6 gb/s drives with an LSI SAS >> >> 9200-8e hba. We are not using failover/redundancy. Meaning that one >> >> port of the hba goes to the primary front backplane interface, and >> >> the >> >> other goes to the primary rear backplane interface. >> >> >> >> For testing, we have done the following: >> >> Installed 12 disks in the front, 0 in the back. >> >> Created a stripe of different numbers of disks. After each test, I >> >> destroy the underlying storage volume and create a new one. As you >> >> can >> >> see by the results, adding more disks, makes no difference to the >> >> performance. This should make a large difference from 4 disks to 8 >> >> disks, however no difference is shown. >> >> >> >> Any help would be greatly appreciated! >> >> >> >> This is the result: >> >> >> >> root@cm-srfe03:/home/gdurham~# time dd if=/dev/zero >> >> of=/fooPool0/86gb.tst bs=4096 count=20971520 >> >> ^C3503681+0 records in >> >> 3503681+0 records out >> >> 14351077376 bytes (14 GB) copied, 39.3747 s, 364 MB/s >> > >> > So, the problem here is that you're not testing the storage at all. >> > You're basically measuring dd. >> > >> > To get meaningful results, you need to do two things: >> > >> > First, run it for long enough so you eliminate any write cache >> > effects. Writes go to memory and only get sent to disk in the >> > background. >> > >> > Second, use a proper benchmark suite, and one that isn't itself >> > a bottleneck. Something like vdbench, although there are others. >> > >> > -- >> > -Peter Tribble >> > http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/ >> > _______________________________________________ >> > zfs-discuss mailing list >> > zfs-discuss@opensolaris.org >> > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss >> > >> _______________________________________________ >> zfs-discuss mailing list >> zfs-discuss@opensolaris.org >> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > -- > Vennlige hilsener / Best regards > > roy > -- > Roy Sigurd Karlsbakk > (+47) 97542685 > r...@karlsbakk.net > http://blogg.karlsbakk.net/ > -- > I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det > er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av > idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og > relevante synonymer på norsk. > _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss