What sort of load will this server be serving? sync or async writes? what sort 
of reads? random i/o or sequential? if sequential, how many streams/concurrent 
users? those are factors you need to evaluate before running a test. A local 
test will usually be using async i/o and a dd with only 4k blocksize is bound 
to be slow, probably because of cpu overhead.

roy

----- Original Message -----
> Hello All,
> Sorry for the lack of information. Here is some answers to some
> questions:
> 1) createPool.sh:
> essentially can take 2 params, one is number of disks in pool, the
> second is either blank or mirrored, blank means number of disks in the
> pool i.e. raid 0, mirrored makes 2 disk mirrors.
> 
> #!/bin/sh
> disks=( `cat diskList | grep Hitachi | awk '{print $2}' | tr '\n' ' '`
> )
> #echo ${disks[1]}
> #$useDisks=" "
> for (( i = 0; i < $1; i++ ))
> do
> #echo "Thus far: "$useDisks
> if [ "$2" = "mirrored" ]
> then
> if [ $(($i % 2)) -eq 0 ]
> then
> useDisks="$useDisks mirror ${disks[i]}"
> else
> useDisks=$useDisks" "${disks[i]}
> fi
> else
> useDisks=$useDisks" "${disks[i]}
> fi
> 
> if [ $(($i - $1)) -le 2 ]
> then
> echo "spares are: ${disks[i]}"
> fi
> done
> 
> #echo $useDisks
> zpool create -f fooPool0 $useDisks
> 
> 
> 
> 2) hardware:
> Each server attached to each storage array is a dell r710 with 32 GB
> memory each. To test for issues with another platform the below info,
> is from a dell 1950 server with 8GB memory. However, I see similar
> results from the r710s as well.
> 
> 
> 3) In order to deal with caching, I am writing larger amounts of data
> to the disk then I have memory for.
> 
> 4) I have tested with bonnie++ as well and here are the results, i
> have read that it is best to test with 4x the amount of memory:
> /usr/local/sbin/bonnie++ -s 32000 -d /fooPool0/test -u gdurham
> Using uid:101, gid:10.
> Writing with putc()...done
> Writing intelligently...done
> Rewriting...done
> Reading with getc()...done
> Reading intelligently...done
> start 'em...done...done...done...
> Create files in sequential order...done.
> Stat files in sequential order...done.
> Delete files in sequential order...done.
> Create files in random order...done.
> Stat files in random order...done.
> Delete files in random order...done.
> Version 1.03d ------Sequential Output------ --Sequential Input-
> --Random-
> -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks--
> Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec
> %CP
> cm-srfe03 32000M 230482 97 477644 76 223687 44 209868 91 541182
> 41 1900 5
> ------Sequential Create------ --------Random Create--------
> -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete--
> files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP
> 16 29126 100 +++++ +++ +++++ +++ 24761 100 +++++ +++ +++++ +++
> cm-srfe03,32000M,230482,97,477644,76,223687,44,209868,91,541182,41,1899.7,5,16,29126,100,+++++,+++,+++++,+++,24761,100,+++++,+++,+++++,+++
> 
> 
> I will run these with the r710 server as well and will report the
> results.
> 
> Thanks for the help!
> 
> -Greg
> 
> 
> 
> On Wed, Aug 10, 2011 at 9:16 AM, phil.har...@gmail.com
> <phil.har...@gmail.com> wrote:
> > I would generally agree that dd is not a great benchmarking tool,
> > but you
> > could use multiple instances to multiple files, and larger block
> > sizes are
> > more efficient. And it's always good to check iostat and mpstat for
> > io and
> > cpu bottlenecks. Also note that an initial run that creates files
> > may be
> > quicker because it just allocates blocks, whereas subsequent
> > rewrites
> > require copy-on-write.
> >
> > ----- Reply message -----
> > From: "Peter Tribble" <peter.trib...@gmail.com>
> > To: "Gregory Durham" <gregory.dur...@gmail.com>
> > Cc: <zfs-discuss@opensolaris.org>
> > Subject: [zfs-discuss] Issues with supermicro
> > Date: Wed, Aug 10, 2011 10:56
> >
> >
> > On Wed, Aug 10, 2011 at 1:45 AM, Gregory Durham
> > <gregory.dur...@gmail.com> wrote:
> >> Hello,
> >> We just purchased two of the sc847e26-rjbod1 units to be used in a
> >> storage environment running Solaris 11 express.
> >>
> >> We are using Hitachi HUA723020ALA640 6 gb/s drives with an LSI SAS
> >> 9200-8e hba. We are not using failover/redundancy. Meaning that one
> >> port of the hba goes to the primary front backplane interface, and
> >> the
> >> other goes to the primary rear backplane interface.
> >>
> >> For testing, we have done the following:
> >> Installed 12 disks in the front, 0 in the back.
> >> Created a stripe of different numbers of disks. After each test, I
> >> destroy the underlying storage volume and create a new one. As you
> >> can
> >> see by the results, adding more disks, makes no difference to the
> >> performance. This should make a large difference from 4 disks to 8
> >> disks, however no difference is shown.
> >>
> >> Any help would be greatly appreciated!
> >>
> >> This is the result:
> >>
> >> root@cm-srfe03:/home/gdurham~# time dd if=/dev/zero
> >> of=/fooPool0/86gb.tst bs=4096 count=20971520
> >> ^C3503681+0 records in
> >> 3503681+0 records out
> >> 14351077376 bytes (14 GB) copied, 39.3747 s, 364 MB/s
> >
> > So, the problem here is that you're not testing the storage at all.
> > You're basically measuring dd.
> >
> > To get meaningful results, you need to do two things:
> >
> > First, run it for long enough so you eliminate any write cache
> > effects. Writes go to memory and only get sent to disk in the
> > background.
> >
> > Second, use a proper benchmark suite, and one that isn't itself
> > a bottleneck. Something like vdbench, although there are others.
> >
> > --
> > -Peter Tribble
> > http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/
> > _______________________________________________
> > zfs-discuss mailing list
> > zfs-discuss@opensolaris.org
> > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> >
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

-- 
Vennlige hilsener / Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to