What sort of load will this server be serving? sync or async writes? what sort of reads? random i/o or sequential? if sequential, how many streams/concurrent users? those are factors you need to evaluate before running a test. A local test will usually be using async i/o and a dd with only 4k blocksize is bound to be slow, probably because of cpu overhead.
roy ----- Original Message ----- > Hello All, > Sorry for the lack of information. Here is some answers to some > questions: > 1) createPool.sh: > essentially can take 2 params, one is number of disks in pool, the > second is either blank or mirrored, blank means number of disks in the > pool i.e. raid 0, mirrored makes 2 disk mirrors. > > #!/bin/sh > disks=( `cat diskList | grep Hitachi | awk '{print $2}' | tr '\n' ' '` > ) > #echo ${disks[1]} > #$useDisks=" " > for (( i = 0; i < $1; i++ )) > do > #echo "Thus far: "$useDisks > if [ "$2" = "mirrored" ] > then > if [ $(($i % 2)) -eq 0 ] > then > useDisks="$useDisks mirror ${disks[i]}" > else > useDisks=$useDisks" "${disks[i]} > fi > else > useDisks=$useDisks" "${disks[i]} > fi > > if [ $(($i - $1)) -le 2 ] > then > echo "spares are: ${disks[i]}" > fi > done > > #echo $useDisks > zpool create -f fooPool0 $useDisks > > > > 2) hardware: > Each server attached to each storage array is a dell r710 with 32 GB > memory each. To test for issues with another platform the below info, > is from a dell 1950 server with 8GB memory. However, I see similar > results from the r710s as well. > > > 3) In order to deal with caching, I am writing larger amounts of data > to the disk then I have memory for. > > 4) I have tested with bonnie++ as well and here are the results, i > have read that it is best to test with 4x the amount of memory: > /usr/local/sbin/bonnie++ -s 32000 -d /fooPool0/test -u gdurham > Using uid:101, gid:10. > Writing with putc()...done > Writing intelligently...done > Rewriting...done > Reading with getc()...done > Reading intelligently...done > start 'em...done...done...done... > Create files in sequential order...done. > Stat files in sequential order...done. > Delete files in sequential order...done. > Create files in random order...done. > Stat files in random order...done. > Delete files in random order...done. > Version 1.03d ------Sequential Output------ --Sequential Input- > --Random- > -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- > Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec > %CP > cm-srfe03 32000M 230482 97 477644 76 223687 44 209868 91 541182 > 41 1900 5 > ------Sequential Create------ --------Random Create-------- > -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- > files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP > 16 29126 100 +++++ +++ +++++ +++ 24761 100 +++++ +++ +++++ +++ > cm-srfe03,32000M,230482,97,477644,76,223687,44,209868,91,541182,41,1899.7,5,16,29126,100,+++++,+++,+++++,+++,24761,100,+++++,+++,+++++,+++ > > > I will run these with the r710 server as well and will report the > results. > > Thanks for the help! > > -Greg > > > > On Wed, Aug 10, 2011 at 9:16 AM, phil.har...@gmail.com > <phil.har...@gmail.com> wrote: > > I would generally agree that dd is not a great benchmarking tool, > > but you > > could use multiple instances to multiple files, and larger block > > sizes are > > more efficient. And it's always good to check iostat and mpstat for > > io and > > cpu bottlenecks. Also note that an initial run that creates files > > may be > > quicker because it just allocates blocks, whereas subsequent > > rewrites > > require copy-on-write. > > > > ----- Reply message ----- > > From: "Peter Tribble" <peter.trib...@gmail.com> > > To: "Gregory Durham" <gregory.dur...@gmail.com> > > Cc: <zfs-discuss@opensolaris.org> > > Subject: [zfs-discuss] Issues with supermicro > > Date: Wed, Aug 10, 2011 10:56 > > > > > > On Wed, Aug 10, 2011 at 1:45 AM, Gregory Durham > > <gregory.dur...@gmail.com> wrote: > >> Hello, > >> We just purchased two of the sc847e26-rjbod1 units to be used in a > >> storage environment running Solaris 11 express. > >> > >> We are using Hitachi HUA723020ALA640 6 gb/s drives with an LSI SAS > >> 9200-8e hba. We are not using failover/redundancy. Meaning that one > >> port of the hba goes to the primary front backplane interface, and > >> the > >> other goes to the primary rear backplane interface. > >> > >> For testing, we have done the following: > >> Installed 12 disks in the front, 0 in the back. > >> Created a stripe of different numbers of disks. After each test, I > >> destroy the underlying storage volume and create a new one. As you > >> can > >> see by the results, adding more disks, makes no difference to the > >> performance. This should make a large difference from 4 disks to 8 > >> disks, however no difference is shown. > >> > >> Any help would be greatly appreciated! > >> > >> This is the result: > >> > >> root@cm-srfe03:/home/gdurham~# time dd if=/dev/zero > >> of=/fooPool0/86gb.tst bs=4096 count=20971520 > >> ^C3503681+0 records in > >> 3503681+0 records out > >> 14351077376 bytes (14 GB) copied, 39.3747 s, 364 MB/s > > > > So, the problem here is that you're not testing the storage at all. > > You're basically measuring dd. > > > > To get meaningful results, you need to do two things: > > > > First, run it for long enough so you eliminate any write cache > > effects. Writes go to memory and only get sent to disk in the > > background. > > > > Second, use a proper benchmark suite, and one that isn't itself > > a bottleneck. Something like vdbench, although there are others. > > > > -- > > -Peter Tribble > > http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/ > > _______________________________________________ > > zfs-discuss mailing list > > zfs-discuss@opensolaris.org > > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss