On Sat, Jan 25, 2014 at 7:01 AM, Steve Dainard <[email protected]>wrote:

> Not sure what a good method to bench this would be, but:
>
> An NFS mount point on virt host:
> [root@ovirt001 iso-store]# dd if=/dev/zero of=test1 bs=4k count=100000
> 100000+0 records in
> 100000+0 records out
> 409600000 bytes (410 MB) copied, 3.95399 s, 104 MB/s
>
> Raw brick performance on gluster server (yes, I know I shouldn't write
> directly to the brick):
> [root@gluster1 iso-store]# dd if=/dev/zero of=test bs=4k count=100000
> 100000+0 records in
> 100000+0 records out
> 409600000 bytes (410 MB) copied, 3.06743 s, 134 MB/s
>
> Gluster mount point on gluster server:
> [root@gluster1 iso-store]#
> dd if=/dev/zero of=test bs=4k count=100000
> 100000+0 records in
> 100000+0 records out
> 409600000 bytes (410 MB) copied, 19.5766 s, 20.9 MB/s
>
> The storage servers are a bit older, but are both dual socket quad core
> opterons with 4x 7200rpm drives.
>
> I'm in the process of setting up a share from my desktop and I'll see if I
> can bench between the two systems. Not sure if my ssd will impact the
> tests, I've heard there isn't an advantage using ssd storage for glusterfs.
>
> Does anyone have a hardware reference design for glusterfs as a backend
> for virt? Or is there a benchmark utility?
>

Check this thread out,
http://raobharata.wordpress.com/2012/10/29/qemu-glusterfs-native-integration/it's
quite dated but I remember seeing similar figures.

In fact when I used FIO on a libgfapi mounted VM I got slightly faster
read/write speeds than on the physical box itself (I assume because of some
level of caching). On NFS it was close to half.. You'll probably get a
little more interesting results using FIO opposed to dd



>
> *Steve Dainard *
> IT Infrastructure Manager
> Miovision <http://miovision.com/> | *Rethink Traffic*
> 519-513-2407 ex.250
> 877-646-8476 (toll-free)
>
> *Blog <http://miovision.com/blog>  |  **LinkedIn
> <https://www.linkedin.com/company/miovision-technologies>  |  Twitter
> <https://twitter.com/miovision>  |  Facebook
> <https://www.facebook.com/miovision>*
> ------------------------------
>  Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener,
> ON, Canada | N2C 1L3
> This e-mail may contain information that is privileged or confidential. If
> you are not the intended recipient, please delete the e-mail and any
> attachments and notify us immediately.
>
>
> On Thu, Jan 23, 2014 at 7:18 PM, Andrew Cathrow <[email protected]>wrote:
>
>> Are we sure that the issue is the guest I/O - what's the raw performance
>> on the host accessing the gluster storage?
>>
>> ------------------------------
>>
>> *From: *"Steve Dainard" <[email protected]>
>> *To: *"Itamar Heim" <[email protected]>
>> *Cc: *"Ronen Hod" <[email protected]>, "users" <[email protected]>, "Sanjay
>> Rao" <[email protected]>
>> *Sent: *Thursday, January 23, 2014 4:56:58 PM
>> *Subject: *Re: [Users] Extremely poor disk access speeds in Windows guest
>>
>>
>> I have two options, virtio and virtio-scsi.
>>
>> I was using virtio, and have also attempted virtio-scsi on another
>> Windows guest with the same results.
>>
>> Using the newest drivers, virtio-win-0.1-74.iso.
>>
>> *Steve Dainard *
>> IT Infrastructure Manager
>> Miovision <http://miovision.com/> | *Rethink Traffic*
>> 519-513-2407 ex.250
>> 877-646-8476 (toll-free)
>>
>> *Blog <http://miovision.com/blog>  |  **LinkedIn
>> <https://www.linkedin.com/company/miovision-technologies>  |  Twitter
>> <https://twitter.com/miovision>  |  Facebook
>> <https://www.facebook.com/miovision>*
>> ------------------------------
>>  Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener,
>> ON, Canada | N2C 1L3
>> This e-mail may contain information that is privileged or confidential.
>> If you are not the intended recipient, please delete the e-mail and any
>> attachments and notify us immediately.
>>
>>
>> On Thu, Jan 23, 2014 at 4:24 PM, Itamar Heim <[email protected]> wrote:
>>
>>> On 01/23/2014 07:46 PM, Steve Dainard wrote:
>>>
>>>> Backing Storage: Gluster Replica
>>>> Storage Domain: NFS
>>>> Ovirt Hosts: CentOS 6.5
>>>> Ovirt version: 3.3.2
>>>> Network: GigE
>>>> # of VM's: 3 - two Linux guests are idle, one Windows guest is
>>>> installing updates.
>>>>
>>>> I've installed a Windows 2008 R2 guest with virtio disk, and all the
>>>> drivers from the latest virtio iso. I've also installed the spice agent
>>>> drivers.
>>>>
>>>> Guest disk access is horribly slow, Resource monitor during Windows
>>>> updates shows Disk peaking at 1MB/sec (scale never increases) and Disk
>>>> Queue Length Peaking at 5 and looks to be sitting at that level 99% of
>>>> the time. 113 updates in Windows has been running solidly for about 2.5
>>>> hours and is at 89/113 updates complete.
>>>>
>>>
>>> virtio-block or virtio-scsi?
>>> which windows guest driver version for that?
>>>
>>>
>>>> I can't say my Linux guests are blisteringly fast, but updating a guest
>>>> from RHEL 6.3 fresh install to 6.5 took about 25 minutes.
>>>>
>>>> If anyone has any ideas, please let me know - I haven't found any tuning
>>>> docs for Windows guests that could explain this issue.
>>>>
>>>> Thanks,
>>>>
>>>>
>>>> *Steve Dainard *
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list
>>>> [email protected]
>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>>
>>>
>>
>> _______________________________________________
>> Users mailing list
>> [email protected]
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>>
>
> _______________________________________________
> Users mailing list
> [email protected]
> http://lists.ovirt.org/mailman/listinfo/users
>
>
_______________________________________________
Users mailing list
[email protected]
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to