Hi,
Is the performance.strict-o-direct=on a mandatory option to avoid data
inconsistency, although it has a pretty big impact on volume iops
performance?
Thank you !




On Fri, Sep 14, 2018, 13:03 Paolo Margara <[email protected]> wrote:

> Hi,
>
> but performance.strict-o-direct is not one of the option enabled by
> gdeploy during installation because it's supposed to give some sort of
> benefit?
>
>
> Paolo
>
> Il 14/09/2018 11:34, Leo David ha scritto:
>
> performance.strict-o-direct:  on
> This was the bloody option that created the botleneck ! It was ON.
> So now i get an average of 17k random writes,  which is not bad at all.
> Below,  the volume options that worked for me:
>
> performance.strict-write-ordering: off
> performance.strict-o-direct: off
> server.event-threads: 4
> client.event-threads: 4
> performance.read-ahead: off
> network.ping-timeout: 30
> performance.quick-read: off
> cluster.eager-lock: enable
> performance.stat-prefetch: on
> performance.low-prio-threads: 32
> network.remote-dio: off
> user.cifs: off
> performance.io-cache: off
> server.allow-insecure: on
> features.shard: on
> transport.address-family: inet
> storage.owner-uid: 36
> storage.owner-gid: 36
> nfs.disable: on
>
> If any other tweaks can be done,  please let me know.
> Thank you !
>
> Leo
>
>
> On Fri, Sep 14, 2018 at 12:01 PM, Leo David <[email protected]> wrote:
>
>> Hi Everyone,
>> So i have decided to take out all of the gluster volume custom options,
>> and add them one by one while activating/deactivating the storage domain &
>> rebooting one vm after each  added option :(
>>
>> The default options that giving bad iops ( ~1-2k) performance are :
>>
>> performance.stat-prefetch on
>> cluster.eager-lock enable
>> performance.io-cache off
>> performance.read-ahead off
>> performance.quick-read off
>> user.cifs off
>> network.ping-timeout 30
>> network.remote-dio off
>> performance.strict-o-direct on
>> performance.low-prio-threads 32
>>
>> After adding only:
>>
>>
>> server.allow-insecure on
>> features.shard on
>> storage.owner-gid 36
>> storage.owner-uid 36
>> transport.address-family inet
>> nfs.disable on
>> The performance increased to 7k-10k iops.
>>
>> The problem is that i don't know if that's sufficient ( maybe it can be
>> more improved ) , or even worse than this there might be chances to into
>> different volume issues by taking out some volume really needed options...
>>
>> If would have handy the default options that are applied to volumes as
>> optimization in a 3way replica, I think that might help..
>>
>> Any thoughts ?
>>
>> Thank you very much !
>>
>>
>> Leo
>>
>>
>>
>>
>>
>> On Fri, Sep 14, 2018 at 8:54 AM, Leo David <[email protected]> wrote:
>>
>>> Any thoughs on these ? Is that UI optimization only a gluster volume
>>> custom configuration ? If so, i guess it can be done from cli, but I am not
>>> aware of the corect optimized parameters of the volume....
>>>
>>>
>>> On Thu, Sep 13, 2018, 18:25 Leo David <[email protected]> wrote:
>>>
>>>> Thank you Jayme. I am trying to do this, but I am getting an error,
>>>> since the volume is replica 1 distribute, and it seems that oVirt expects a
>>>> replica 3 volume.
>>>> Would it be another way to optimize the volume in this situation ?
>>>>
>>>>
>>>> On Thu, Sep 13, 2018, 17:49 Jayme <[email protected]> wrote:
>>>>
>>>>> I had similar problems until I clicked "optimize volume for vmstore"
>>>>> in the admin GUI for each data volume.  I'm not sure if this is what is
>>>>> causing your problem here but I'd recommend trying that first.  It is
>>>>> suppose to be optimized by default but for some reason my ovirt 4.2 
>>>>> cockpit
>>>>> deploy did not apply those settings automatically.
>>>>>
>>>>> On Thu, Sep 13, 2018 at 10:21 AM Leo David <[email protected]> wrote:
>>>>>
>>>>>> Hi Everyone,
>>>>>> I am encountering the following issue on a single instance
>>>>>> hyper-converged 4.2 setup.
>>>>>> The following fio test was done:
>>>>>>
>>>>>> fio --randrepeat=1 --ioengine=libaio --direct=1 --gtod_reduce=1
>>>>>> --name=test --filename=test --bs=4k --iodepth=64 --size=4G
>>>>>> --readwrite=randwrite
>>>>>> The results are very poor doing the test inside of a vm with a
>>>>>> prealocated disk on the ssd store:  ~2k IOPS
>>>>>> Same test done on the oVirt node directly on the mounted ssd_lvm:
>>>>>> ~30k IOPS
>>>>>> Same test done, this time on the gluster mount path: ~20K IOPS
>>>>>>
>>>>>> What could be the issue that the vms have this slow hdd performance (
>>>>>> 2k on ssd !! )?
>>>>>> Thank you very much !
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best regards, Leo David
>>>>>> _______________________________________________
>>>>>> Users mailing list -- [email protected]
>>>>>> To unsubscribe send an email to [email protected]
>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>> oVirt Code of Conduct:
>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>> List Archives:
>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/FCCIZFRWINWWLQSYWRPF6HNKPQMZB2XP/
>>>>>>
>>>>>
>>
>>
>> --
>> Best regards, Leo David
>>
>
>
>
> --
> Best regards, Leo David
>
>
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives: 
> https://lists.ovirt.org/archives/list/[email protected]/message/5HKQQKX3BJLZ3HQ5SCHPLPON24OGMGSS/
>
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/LEQL77GT3PZBEIJ6H3H67H7B7AKWM75I/

Reply via email to