On Wed, Jul 1, 2009 at 10:33 AM, Jim Dunham<[email protected]> wrote:
> Ross,
>
>> I don't know if it's my network setup or what at this point, but during my
>> benchmarks I am unable to get more then 20-23MB/s doing 4k sequential reads
>> with 4 outstanding I/Os. This is constant whether going to a ZVOL or a soft
>> partition.
>>
>> My setup:
>>
>> Server: Dell PE 2950, 2 quad xeons, 4GB memory, 2 on board bnx (management
>> network only), intel quad igb running LAG to a layer 3 switch, flow control
>> and jumbo frames enabled, Solaris 10 update 7 with all the latest patches
>> installed. Storage is 14 SAS 15k in a MD1000 enclosure hooked up to a Dell
>> PERC 6e w/ 512MB of BBU write-back cache, disks setup as individual RAID0
>> disks, then put in a zpool as 2 raidz2 sets of 7 drives each. I have the
>> write-back enabled on each disk for ZIL performance since I don't have an
>> SSD disk for ZIL logging.
>>
>> Client ESX windows 2003r2 guest running MS ini 2.08.
>>
>> I believe the switch and the ESX are setup properly tests against a Linux
>> host running IET and HW RAID6 show it's capable of around 30-40MB/s.
>>
>> Any feedback on this would be appreciated as I am getting a little burnt
>> out trying to diagnose this.
>
> Getting 20-23MB/s doing 4k sequential reads is between 5K to 6K IOPS. These
> being constant going between ZVOL or soft partition, I would have to surmise
> that CPU performance is likely one of your limiting factors. What is the CPU
> cost of the iSCSI Target Daemon (iscsitgtd) while driving this I/O load?

The IOPS estimate is dead on. I didn't run top during the benchmark,
so I can't say definitively what it was, the load avg was about 1.5,
but there was NFS operations occuring on a separate zpool at the time,
so that isn't reliable.

I'll re-run the test and look at the usage.

> If you are measuring 20-23MB/s performance on the iSCSI Client side, what is
> the iSCSI Target's I/O performance in accessing the ZVOLs or soft
> partitions? Specific to comparing iSCSI client and target performance, does
> the work load performance change, as the I/O size changes?

On the client side I was seeing an 18-20% cpu usage, on a quad cpu esx
guest, the load was single cpu bound, but changed cpus during the run.

The benchmark runs workloads of my choosing, the random workload gives
me pretty much the disks' capabilities, it's the sequential that is
causing a problem. Sequential is sensitive to network latency so that
makes me think it's network related, but I'll shoot at anything at
this point.

Is iscsitgt multi threaded in the backing store routines?

Can it efficiently handle multiple outstanding I/Os?

Because if the backing store I/O is single threaded that would also
show the same performance degradation between it and the Linux target
that uses 8 I/O threads per target for backing store operations.

-Ross
_______________________________________________
storage-discuss mailing list
[email protected]
http://mail.opensolaris.org/mailman/listinfo/storage-discuss

Reply via email to