On Wed, Feb 27, 2019 at 11:21 AM Leo David <[email protected]> wrote:
>
> Thank you Sahina, I'm in that conversation too :).
> On the other hand...
> In this case, setting this option on, would only make sense in multi-node 
> setups, and not in single instance ones, where we only have one hypervisor 
> accsessing the volume.
> Please correct me if this is wrong.
> Have a nice day,

In single instance deployments too, the option ensures all writes
(with o-direct flag) are flushed to disk and not cached.
>
> Leo
>
>
> On Tue, Feb 26, 2019, 08:24 Sahina Bose <[email protected]> wrote:
>>
>>
>>
>>
>> On Fri, Sep 14, 2018 at 3:35 PM Paolo Margara <[email protected]> 
>> wrote:
>>>
>>> Hi,
>>>
>>> but performance.strict-o-direct is not one of the option enabled by gdeploy 
>>> during installation because it's supposed to give some sort of benefit?
>>
>>
>> See 
>> https://lists.ovirt.org/archives/list/[email protected]/message/VS764WDBR2PLGGDZVRGBEM6OJCAFEM3R/
>>  on why the option is set.
>>
>>>
>>> Paolo
>>>
>>>
>>> Il 14/09/2018 11:34, Leo David ha scritto:
>>>
>>> performance.strict-o-direct:  on
>>> This was the bloody option that created the botleneck ! It was ON.
>>> So now i get an average of 17k random writes,  which is not bad at all. 
>>> Below,  the volume options that worked for me:
>>>
>>> performance.strict-write-ordering: off
>>> performance.strict-o-direct: off
>>> server.event-threads: 4
>>> client.event-threads: 4
>>> performance.read-ahead: off
>>> network.ping-timeout: 30
>>> performance.quick-read: off
>>> cluster.eager-lock: enable
>>> performance.stat-prefetch: on
>>> performance.low-prio-threads: 32
>>> network.remote-dio: off
>>> user.cifs: off
>>> performance.io-cache: off
>>> server.allow-insecure: on
>>> features.shard: on
>>> transport.address-family: inet
>>> storage.owner-uid: 36
>>> storage.owner-gid: 36
>>> nfs.disable: on
>>>
>>> If any other tweaks can be done,  please let me know.
>>> Thank you !
>>>
>>> Leo
>>>
>>>
>>> On Fri, Sep 14, 2018 at 12:01 PM, Leo David <[email protected]> wrote:
>>>>
>>>> Hi Everyone,
>>>> So i have decided to take out all of the gluster volume custom options,  
>>>> and add them one by one while activating/deactivating the storage domain & 
>>>> rebooting one vm after each  added option :(
>>>>
>>>> The default options that giving bad iops ( ~1-2k) performance are :
>>>>
>>>> performance.stat-prefetch on
>>>> cluster.eager-lock enable
>>>> performance.io-cache off
>>>> performance.read-ahead off
>>>> performance.quick-read off
>>>> user.cifs off
>>>> network.ping-timeout 30
>>>> network.remote-dio off
>>>> performance.strict-o-direct on
>>>> performance.low-prio-threads 32
>>>>
>>>> After adding only:
>>>>
>>>>
>>>> server.allow-insecure on
>>>> features.shard on
>>>> storage.owner-gid 36
>>>> storage.owner-uid 36
>>>> transport.address-family inet
>>>> nfs.disable on
>>>>
>>>> The performance increased to 7k-10k iops.
>>>>
>>>> The problem is that i don't know if that's sufficient ( maybe it can be 
>>>> more improved ) , or even worse than this there might be chances to into 
>>>> different volume issues by taking out some volume really needed options...
>>>>
>>>> If would have handy the default options that are applied to volumes as 
>>>> optimization in a 3way replica, I think that might help..
>>>>
>>>> Any thoughts ?
>>>>
>>>> Thank you very much !
>>>>
>>>>
>>>> Leo
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On Fri, Sep 14, 2018 at 8:54 AM, Leo David <[email protected]> wrote:
>>>>>
>>>>> Any thoughs on these ? Is that UI optimization only a gluster volume 
>>>>> custom configuration ? If so, i guess it can be done from cli, but I am 
>>>>> not aware of the corect optimized parameters of the volume....
>>>>>
>>>>>
>>>>> On Thu, Sep 13, 2018, 18:25 Leo David <[email protected]> wrote:
>>>>>>
>>>>>> Thank you Jayme. I am trying to do this, but I am getting an error, 
>>>>>> since the volume is replica 1 distribute, and it seems that oVirt 
>>>>>> expects a replica 3 volume.
>>>>>> Would it be another way to optimize the volume in this situation ?
>>>>>>
>>>>>>
>>>>>> On Thu, Sep 13, 2018, 17:49 Jayme <[email protected]> wrote:
>>>>>>>
>>>>>>> I had similar problems until I clicked "optimize volume for vmstore" in 
>>>>>>> the admin GUI for each data volume.  I'm not sure if this is what is 
>>>>>>> causing your problem here but I'd recommend trying that first.  It is 
>>>>>>> suppose to be optimized by default but for some reason my ovirt 4.2 
>>>>>>> cockpit deploy did not apply those settings automatically.
>>>>>>>
>>>>>>> On Thu, Sep 13, 2018 at 10:21 AM Leo David <[email protected]> wrote:
>>>>>>>>
>>>>>>>> Hi Everyone,
>>>>>>>> I am encountering the following issue on a single instance 
>>>>>>>> hyper-converged 4.2 setup.
>>>>>>>> The following fio test was done:
>>>>>>>>
>>>>>>>> fio --randrepeat=1 --ioengine=libaio --direct=1 --gtod_reduce=1 
>>>>>>>> --name=test --filename=test --bs=4k --iodepth=64 --size=4G 
>>>>>>>> --readwrite=randwrite
>>>>>>>> The results are very poor doing the test inside of a vm with a 
>>>>>>>> prealocated disk on the ssd store:  ~2k IOPS
>>>>>>>> Same test done on the oVirt node directly on the mounted ssd_lvm: ~30k 
>>>>>>>> IOPS
>>>>>>>> Same test done, this time on the gluster mount path: ~20K IOPS
>>>>>>>>
>>>>>>>> What could be the issue that the vms have this slow hdd performance ( 
>>>>>>>> 2k on ssd !! )?
>>>>>>>> Thank you very much !
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Best regards, Leo David
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list -- [email protected]
>>>>>>>> To unsubscribe send an email to [email protected]
>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>>>> oVirt Code of Conduct: 
>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>> List Archives: 
>>>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/FCCIZFRWINWWLQSYWRPF6HNKPQMZB2XP/
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Best regards, Leo David
>>>
>>>
>>>
>>>
>>> --
>>> Best regards, Leo David
>>>
>>>
>>> _______________________________________________
>>> Users mailing list -- [email protected]
>>> To unsubscribe send an email to [email protected]
>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>> oVirt Code of Conduct: 
>>> https://www.ovirt.org/community/about/community-guidelines/
>>> List Archives: 
>>> https://lists.ovirt.org/archives/list/[email protected]/message/5HKQQKX3BJLZ3HQ5SCHPLPON24OGMGSS/
>>>
>>> _______________________________________________
>>> Users mailing list -- [email protected]
>>> To unsubscribe send an email to [email protected]
>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>> oVirt Code of Conduct: 
>>> https://www.ovirt.org/community/about/community-guidelines/
>>> List Archives: 
>>> https://lists.ovirt.org/archives/list/[email protected]/message/E2KCDRMU2KESRNC35YXQUB365B4BF3BU/
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/K6VVFN36N5NJP3Y5FJKVHO3EVTLPG3HT/

Reply via email to