Thank you Sahina, I'm in that conversation too :). On the other hand... In this case, setting this option on, would only make sense in multi-node setups, and not in single instance ones, where we only have one hypervisor accsessing the volume. Please correct me if this is wrong. Have a nice day,
Leo On Tue, Feb 26, 2019, 08:24 Sahina Bose <[email protected]> wrote: > > > > On Fri, Sep 14, 2018 at 3:35 PM Paolo Margara <[email protected]> > wrote: > >> Hi, >> >> but performance.strict-o-direct is not one of the option enabled by >> gdeploy during installation because it's supposed to give some sort of >> benefit? >> > > See > https://lists.ovirt.org/archives/list/[email protected]/message/VS764WDBR2PLGGDZVRGBEM6OJCAFEM3R/ > on why the option is set. > > >> Paolo >> >> Il 14/09/2018 11:34, Leo David ha scritto: >> >> performance.strict-o-direct: on >> This was the bloody option that created the botleneck ! It was ON. >> So now i get an average of 17k random writes, which is not bad at all. >> Below, the volume options that worked for me: >> >> performance.strict-write-ordering: off >> performance.strict-o-direct: off >> server.event-threads: 4 >> client.event-threads: 4 >> performance.read-ahead: off >> network.ping-timeout: 30 >> performance.quick-read: off >> cluster.eager-lock: enable >> performance.stat-prefetch: on >> performance.low-prio-threads: 32 >> network.remote-dio: off >> user.cifs: off >> performance.io-cache: off >> server.allow-insecure: on >> features.shard: on >> transport.address-family: inet >> storage.owner-uid: 36 >> storage.owner-gid: 36 >> nfs.disable: on >> >> If any other tweaks can be done, please let me know. >> Thank you ! >> >> Leo >> >> >> On Fri, Sep 14, 2018 at 12:01 PM, Leo David <[email protected]> wrote: >> >>> Hi Everyone, >>> So i have decided to take out all of the gluster volume custom options, >>> and add them one by one while activating/deactivating the storage domain & >>> rebooting one vm after each added option :( >>> >>> The default options that giving bad iops ( ~1-2k) performance are : >>> >>> performance.stat-prefetch on >>> cluster.eager-lock enable >>> performance.io-cache off >>> performance.read-ahead off >>> performance.quick-read off >>> user.cifs off >>> network.ping-timeout 30 >>> network.remote-dio off >>> performance.strict-o-direct on >>> performance.low-prio-threads 32 >>> >>> After adding only: >>> >>> >>> server.allow-insecure on >>> features.shard on >>> storage.owner-gid 36 >>> storage.owner-uid 36 >>> transport.address-family inet >>> nfs.disable on >>> The performance increased to 7k-10k iops. >>> >>> The problem is that i don't know if that's sufficient ( maybe it can be >>> more improved ) , or even worse than this there might be chances to into >>> different volume issues by taking out some volume really needed options... >>> >>> If would have handy the default options that are applied to volumes as >>> optimization in a 3way replica, I think that might help.. >>> >>> Any thoughts ? >>> >>> Thank you very much ! >>> >>> >>> Leo >>> >>> >>> >>> >>> >>> On Fri, Sep 14, 2018 at 8:54 AM, Leo David <[email protected]> wrote: >>> >>>> Any thoughs on these ? Is that UI optimization only a gluster volume >>>> custom configuration ? If so, i guess it can be done from cli, but I am not >>>> aware of the corect optimized parameters of the volume.... >>>> >>>> >>>> On Thu, Sep 13, 2018, 18:25 Leo David <[email protected]> wrote: >>>> >>>>> Thank you Jayme. I am trying to do this, but I am getting an error, >>>>> since the volume is replica 1 distribute, and it seems that oVirt expects >>>>> a >>>>> replica 3 volume. >>>>> Would it be another way to optimize the volume in this situation ? >>>>> >>>>> >>>>> On Thu, Sep 13, 2018, 17:49 Jayme <[email protected]> wrote: >>>>> >>>>>> I had similar problems until I clicked "optimize volume for vmstore" >>>>>> in the admin GUI for each data volume. I'm not sure if this is what is >>>>>> causing your problem here but I'd recommend trying that first. It is >>>>>> suppose to be optimized by default but for some reason my ovirt 4.2 >>>>>> cockpit >>>>>> deploy did not apply those settings automatically. >>>>>> >>>>>> On Thu, Sep 13, 2018 at 10:21 AM Leo David <[email protected]> wrote: >>>>>> >>>>>>> Hi Everyone, >>>>>>> I am encountering the following issue on a single instance >>>>>>> hyper-converged 4.2 setup. >>>>>>> The following fio test was done: >>>>>>> >>>>>>> fio --randrepeat=1 --ioengine=libaio --direct=1 --gtod_reduce=1 >>>>>>> --name=test --filename=test --bs=4k --iodepth=64 --size=4G >>>>>>> --readwrite=randwrite >>>>>>> The results are very poor doing the test inside of a vm with a >>>>>>> prealocated disk on the ssd store: ~2k IOPS >>>>>>> Same test done on the oVirt node directly on the mounted ssd_lvm: >>>>>>> ~30k IOPS >>>>>>> Same test done, this time on the gluster mount path: ~20K IOPS >>>>>>> >>>>>>> What could be the issue that the vms have this slow hdd performance >>>>>>> ( 2k on ssd !! )? >>>>>>> Thank you very much ! >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Best regards, Leo David >>>>>>> _______________________________________________ >>>>>>> Users mailing list -- [email protected] >>>>>>> To unsubscribe send an email to [email protected] >>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>> oVirt Code of Conduct: >>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>> List Archives: >>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/FCCIZFRWINWWLQSYWRPF6HNKPQMZB2XP/ >>>>>>> >>>>>> >>> >>> >>> -- >>> Best regards, Leo David >>> >> >> >> >> -- >> Best regards, Leo David >> >> >> _______________________________________________ >> Users mailing list -- [email protected] >> To unsubscribe send an email to [email protected] >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> List Archives: >> https://lists.ovirt.org/archives/list/[email protected]/message/5HKQQKX3BJLZ3HQ5SCHPLPON24OGMGSS/ >> >> _______________________________________________ >> Users mailing list -- [email protected] >> To unsubscribe send an email to [email protected] >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> List Archives: >> https://lists.ovirt.org/archives/list/[email protected]/message/E2KCDRMU2KESRNC35YXQUB365B4BF3BU/ >> >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/L57QZ4NJO3H4XB2YLQ3EN537AHKOYKQO/

