Skickat från min iPhone

> 8 jan 2014 kl. 18:47 skrev "Darrell Budic" <[email protected]>:
> 
> Grégoire-
> 
> I think this is expected behavior. Well, at least the high glusterfsd CPU use 
> during disk creation, anyway. I tried creating a 10 G disk on my test 
> environment and observed similar high CPU usage by glusterfsd. Did the 
> creation on the i5 system, it showed 95%-105% cpu for glusterfsd during 
> creation, with the core2 system running ~35-65% glusterfsd utilization during 
> the creation. Minor disk wait was observed on both systems, < 10% peak and 
> generally < 5%. I imagine my ZFS cached backends helped a lot here. Took 
> about 3 minutes, roughly what I’d expect for the i5’s disk system. Network 
> usage was about 45% of the 1G link. No errors or messages logged to 
> /var/log/messages.
> 
> Depending on what your test setup looks like, I’d check my network for packet 
> loss or errors first. Then look at my storage setup and test pure throughput 
> on the disks to see what you’ve got, maybe see what else is running. Did you 
> use an NFS cluster or a PosixFS cluster for this?
> 
> My test setup, running a version of the nightly self-hosted setup w/ gluster 
> distributed/replicated disks as shared storage, in a NFS cluster:
> 
> Core i5 3570K @ 3.4Ghz, 16G Ram
> Boot disks: 2x 32G SATA SSDs in raid-1
> Storage system: 4x500G Seagate RE3s in a ZFS raid-10 w/ 1GB ZIL & ~22G L2ARC 
> caching from boot drives
> 1 1G ethernet
> 2 VMs running
> 
> Core2 Duo E8500 @ 3.16GHz, 8G Ram
> Boot disks: 2x 32G SATA SSDS in raid-1
> Storage system: 2x1500G WD Green drives in a  ZFS Raid w/ 1GB ZIL & ~22G 
> L2ARC cache from boot drives
> 1 1G ethernet
> 
> They are connected through a Netgear Prosafe+ workgroup style switch, not 
> much going on between them.
> 
>  -Darrell

Just curious, are you doing ZFS in Linux?

/K

> 
>> On Jan 8, 2014, at 7:49 AM, [email protected] wrote:
>> 
>> Hello,
>> 
>> Do you need more informations about this issue ? Do you think this problem 
>> is likely to show up in other cases ? I mean, is that an expected behaviour 
>> with my environment, or is it unexpected ?
>> 
>> Is there a way to limit the bandwidth usage for creation of pre-allocated 
>> disk so that it doesn't impact production ?
>> 
>> Thank you,
>> Regards,
>> Grégoire
>> 
>> Le 2014-01-02 17:42, Vijay Bellur a écrit :
>>> Adding gluster-users.
>>> On 01/02/2014 08:50 PM, [email protected] wrote:
>>>> Hello,
>>>> I have a Gluster volume in distributed/replicated mode. I have 2 hosts.
>>>> When I try to create a VM with a preallocated disk, it uses 100% of the
>>>> available CPU and bandwidth (I have 1 Gigabit network card).
>>>> The result is I can't even create a preallocated disk because the engine
>>>> detects a network failure.
>>>> I get that kind of messages in /var/log/messages :
>>>> "
>>>> Jan  2 14:13:54 localhost sanlock[3811]: 2014-01-02 14:13:54+0100 167737
>>>> [3811]: s4 kill 21114 sig 15 count 1
>>>> Jan  2 14:13:54 localhost wdmd[3800]: test failed rem 51 now 167737 ping
>>>> 167718 close 167728 renewal 167657 expire 167737 client 3811
>>>> sanlock_ef4978d6-5711-4e01-a0ec-7ffbd9     cdbe5d:1
>>>> "
>>>> And that in the Ovirt Gui :
>>>> "
>>>> 2014-janv.-02, 15:35 Operation Add-Disk failed to complete.
>>>> 2014-janv.-02, 15:35 Storage Pool Manager runs on Host HOST2 (Address:
>>>> X.X.X.X).
>>>> 2014-janv.-02, 15:35 Invalid status on Data Center GlusterSewan. Setting
>>>> Data Center status to Non Responsive (On host HOST2, Error: done).
>>>> 2014-janv.-02, 15:35 State was set to Up for host HOST2.
>>>> 2014-janv.-02, 15:33 Used Network resources of host HOST2 [98%] exceeded
>>>> defined threshold [95%].
>>>> 2014-janv.-02, 15:33 Add-Disk operation of test_Disk1 was initiated on
>>>> VM test by admin@internal.
>>>> I understand that the creation of a 10 Go disk image generates a lot of
>>>> traffic, but is there a way to limit it so that it doesn't have an
>>>> impact on the production ? Furthermore, Why does it use so much CPU
>>>> ressources ? I can see on my monitoring graph a big peak of CPU usage
>>>> when I launched the operation (probably until 100%).
>>> Do you happen to notice what is consuming CPU? Since the same cluster
>>> does both virtualization and storage, a GigE network might get
>>> saturated very quickly. Is it possible to separate out the management
>>> and data/gluster traffic in this setup?
>>> Regards,
>>> Vijay
>> _______________________________________________
>> Users mailing list
>> [email protected]
>> http://lists.ovirt.org/mailman/listinfo/users
> 
> _______________________________________________
> Users mailing list
> [email protected]
> http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________
Users mailing list
[email protected]
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to