Hello All, Sorry for the lack of information. Here is some answers to some questions: 1) createPool.sh: essentially can take 2 params, one is number of disks in pool, the second is either blank or mirrored, blank means number of disks in the pool i.e. raid 0, mirrored makes 2 disk mirrors.
#!/bin/sh disks=( `cat diskList | grep Hitachi | awk '{print $2}' | tr '\n' ' '` ) #echo ${disks[1]} #$useDisks=" " for (( i = 0; i < $1; i++ )) do #echo "Thus far: "$useDisks if [ "$2" = "mirrored" ] then if [ $(($i % 2)) -eq 0 ] then useDisks="$useDisks mirror ${disks[i]}" else useDisks=$useDisks" "${disks[i]} fi else useDisks=$useDisks" "${disks[i]} fi if [ $(($i - $1)) -le 2 ] then echo "spares are: ${disks[i]}" fi done #echo $useDisks zpool create -f fooPool0 $useDisks 2) hardware: Each server attached to each storage array is a dell r710 with 32 GB memory each. To test for issues with another platform the below info, is from a dell 1950 server with 8GB memory. However, I see similar results from the r710s as well. 3) In order to deal with caching, I am writing larger amounts of data to the disk then I have memory for. 4) I have tested with bonnie++ as well and here are the results, i have read that it is best to test with 4x the amount of memory: /usr/local/sbin/bonnie++ -s 32000 -d /fooPool0/test -u gdurham Using uid:101, gid:10. Writing with putc()...done Writing intelligently...done Rewriting...done Reading with getc()...done Reading intelligently...done start 'em...done...done...done... Create files in sequential order...done. Stat files in sequential order...done. Delete files in sequential order...done. Create files in random order...done. Stat files in random order...done. Delete files in random order...done. Version 1.03d ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP cm-srfe03 32000M 230482 97 477644 76 223687 44 209868 91 541182 41 1900 5 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 29126 100 +++++ +++ +++++ +++ 24761 100 +++++ +++ +++++ +++ cm-srfe03,32000M,230482,97,477644,76,223687,44,209868,91,541182,41,1899.7,5,16,29126,100,+++++,+++,+++++,+++,24761,100,+++++,+++,+++++,+++ I will run these with the r710 server as well and will report the results. Thanks for the help! -Greg On Wed, Aug 10, 2011 at 9:16 AM, phil.har...@gmail.com <phil.har...@gmail.com> wrote: > I would generally agree that dd is not a great benchmarking tool, but you > could use multiple instances to multiple files, and larger block sizes are > more efficient. And it's always good to check iostat and mpstat for io and > cpu bottlenecks. Also note that an initial run that creates files may be > quicker because it just allocates blocks, whereas subsequent rewrites > require copy-on-write. > > ----- Reply message ----- > From: "Peter Tribble" <peter.trib...@gmail.com> > To: "Gregory Durham" <gregory.dur...@gmail.com> > Cc: <zfs-discuss@opensolaris.org> > Subject: [zfs-discuss] Issues with supermicro > Date: Wed, Aug 10, 2011 10:56 > > > On Wed, Aug 10, 2011 at 1:45 AM, Gregory Durham > <gregory.dur...@gmail.com> wrote: >> Hello, >> We just purchased two of the sc847e26-rjbod1 units to be used in a >> storage environment running Solaris 11 express. >> >> We are using Hitachi HUA723020ALA640 6 gb/s drives with an LSI SAS >> 9200-8e hba. We are not using failover/redundancy. Meaning that one >> port of the hba goes to the primary front backplane interface, and the >> other goes to the primary rear backplane interface. >> >> For testing, we have done the following: >> Installed 12 disks in the front, 0 in the back. >> Created a stripe of different numbers of disks. After each test, I >> destroy the underlying storage volume and create a new one. As you can >> see by the results, adding more disks, makes no difference to the >> performance. This should make a large difference from 4 disks to 8 >> disks, however no difference is shown. >> >> Any help would be greatly appreciated! >> >> This is the result: >> >> root@cm-srfe03:/home/gdurham~# time dd if=/dev/zero >> of=/fooPool0/86gb.tst bs=4096 count=20971520 >> ^C3503681+0 records in >> 3503681+0 records out >> 14351077376 bytes (14 GB) copied, 39.3747 s, 364 MB/s > > So, the problem here is that you're not testing the storage at all. > You're basically measuring dd. > > To get meaningful results, you need to do two things: > > First, run it for long enough so you eliminate any write cache > effects. Writes go to memory and only get sent to disk in the > background. > > Second, use a proper benchmark suite, and one that isn't itself > a bottleneck. Something like vdbench, although there are others. > > -- > -Peter Tribble > http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/ > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss