Yes, in the end that makes perfectly sense. Thank you very much Sahina !
On Fri, Mar 1, 2019, 07:45 Sahina Bose <[email protected]> wrote: > On Wed, Feb 27, 2019 at 11:21 AM Leo David <[email protected]> wrote: > > > > Thank you Sahina, I'm in that conversation too :). > > On the other hand... > > In this case, setting this option on, would only make sense in > multi-node setups, and not in single instance ones, where we only have one > hypervisor accsessing the volume. > > Please correct me if this is wrong. > > Have a nice day, > > In single instance deployments too, the option ensures all writes > (with o-direct flag) are flushed to disk and not cached. > > > > Leo > > > > > > On Tue, Feb 26, 2019, 08:24 Sahina Bose <[email protected]> wrote: > >> > >> > >> > >> > >> On Fri, Sep 14, 2018 at 3:35 PM Paolo Margara <[email protected]> > wrote: > >>> > >>> Hi, > >>> > >>> but performance.strict-o-direct is not one of the option enabled by > gdeploy during installation because it's supposed to give some sort of > benefit? > >> > >> > >> See > https://lists.ovirt.org/archives/list/[email protected]/message/VS764WDBR2PLGGDZVRGBEM6OJCAFEM3R/ > on why the option is set. > >> > >>> > >>> Paolo > >>> > >>> > >>> Il 14/09/2018 11:34, Leo David ha scritto: > >>> > >>> performance.strict-o-direct: on > >>> This was the bloody option that created the botleneck ! It was ON. > >>> So now i get an average of 17k random writes, which is not bad at > all. Below, the volume options that worked for me: > >>> > >>> performance.strict-write-ordering: off > >>> performance.strict-o-direct: off > >>> server.event-threads: 4 > >>> client.event-threads: 4 > >>> performance.read-ahead: off > >>> network.ping-timeout: 30 > >>> performance.quick-read: off > >>> cluster.eager-lock: enable > >>> performance.stat-prefetch: on > >>> performance.low-prio-threads: 32 > >>> network.remote-dio: off > >>> user.cifs: off > >>> performance.io-cache: off > >>> server.allow-insecure: on > >>> features.shard: on > >>> transport.address-family: inet > >>> storage.owner-uid: 36 > >>> storage.owner-gid: 36 > >>> nfs.disable: on > >>> > >>> If any other tweaks can be done, please let me know. > >>> Thank you ! > >>> > >>> Leo > >>> > >>> > >>> On Fri, Sep 14, 2018 at 12:01 PM, Leo David <[email protected]> wrote: > >>>> > >>>> Hi Everyone, > >>>> So i have decided to take out all of the gluster volume custom > options, and add them one by one while activating/deactivating the storage > domain & rebooting one vm after each added option :( > >>>> > >>>> The default options that giving bad iops ( ~1-2k) performance are : > >>>> > >>>> performance.stat-prefetch on > >>>> cluster.eager-lock enable > >>>> performance.io-cache off > >>>> performance.read-ahead off > >>>> performance.quick-read off > >>>> user.cifs off > >>>> network.ping-timeout 30 > >>>> network.remote-dio off > >>>> performance.strict-o-direct on > >>>> performance.low-prio-threads 32 > >>>> > >>>> After adding only: > >>>> > >>>> > >>>> server.allow-insecure on > >>>> features.shard on > >>>> storage.owner-gid 36 > >>>> storage.owner-uid 36 > >>>> transport.address-family inet > >>>> nfs.disable on > >>>> > >>>> The performance increased to 7k-10k iops. > >>>> > >>>> The problem is that i don't know if that's sufficient ( maybe it can > be more improved ) , or even worse than this there might be chances to into > different volume issues by taking out some volume really needed options... > >>>> > >>>> If would have handy the default options that are applied to volumes > as optimization in a 3way replica, I think that might help.. > >>>> > >>>> Any thoughts ? > >>>> > >>>> Thank you very much ! > >>>> > >>>> > >>>> Leo > >>>> > >>>> > >>>> > >>>> > >>>> > >>>> > >>>> On Fri, Sep 14, 2018 at 8:54 AM, Leo David <[email protected]> wrote: > >>>>> > >>>>> Any thoughs on these ? Is that UI optimization only a gluster volume > custom configuration ? If so, i guess it can be done from cli, but I am not > aware of the corect optimized parameters of the volume.... > >>>>> > >>>>> > >>>>> On Thu, Sep 13, 2018, 18:25 Leo David <[email protected]> wrote: > >>>>>> > >>>>>> Thank you Jayme. I am trying to do this, but I am getting an error, > since the volume is replica 1 distribute, and it seems that oVirt expects a > replica 3 volume. > >>>>>> Would it be another way to optimize the volume in this situation ? > >>>>>> > >>>>>> > >>>>>> On Thu, Sep 13, 2018, 17:49 Jayme <[email protected]> wrote: > >>>>>>> > >>>>>>> I had similar problems until I clicked "optimize volume for > vmstore" in the admin GUI for each data volume. I'm not sure if this is > what is causing your problem here but I'd recommend trying that first. It > is suppose to be optimized by default but for some reason my ovirt 4.2 > cockpit deploy did not apply those settings automatically. > >>>>>>> > >>>>>>> On Thu, Sep 13, 2018 at 10:21 AM Leo David <[email protected]> > wrote: > >>>>>>>> > >>>>>>>> Hi Everyone, > >>>>>>>> I am encountering the following issue on a single instance > hyper-converged 4.2 setup. > >>>>>>>> The following fio test was done: > >>>>>>>> > >>>>>>>> fio --randrepeat=1 --ioengine=libaio --direct=1 --gtod_reduce=1 > --name=test --filename=test --bs=4k --iodepth=64 --size=4G > --readwrite=randwrite > >>>>>>>> The results are very poor doing the test inside of a vm with a > prealocated disk on the ssd store: ~2k IOPS > >>>>>>>> Same test done on the oVirt node directly on the mounted ssd_lvm: > ~30k IOPS > >>>>>>>> Same test done, this time on the gluster mount path: ~20K IOPS > >>>>>>>> > >>>>>>>> What could be the issue that the vms have this slow hdd > performance ( 2k on ssd !! )? > >>>>>>>> Thank you very much ! > >>>>>>>> > >>>>>>>> > >>>>>>>> > >>>>>>>> > >>>>>>>> -- > >>>>>>>> Best regards, Leo David > >>>>>>>> _______________________________________________ > >>>>>>>> Users mailing list -- [email protected] > >>>>>>>> To unsubscribe send an email to [email protected] > >>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > >>>>>>>> oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > >>>>>>>> List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/FCCIZFRWINWWLQSYWRPF6HNKPQMZB2XP/ > >>>> > >>>> > >>>> > >>>> > >>>> -- > >>>> Best regards, Leo David > >>> > >>> > >>> > >>> > >>> -- > >>> Best regards, Leo David > >>> > >>> > >>> _______________________________________________ > >>> Users mailing list -- [email protected] > >>> To unsubscribe send an email to [email protected] > >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > >>> oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > >>> List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/5HKQQKX3BJLZ3HQ5SCHPLPON24OGMGSS/ > >>> > >>> _______________________________________________ > >>> Users mailing list -- [email protected] > >>> To unsubscribe send an email to [email protected] > >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > >>> oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > >>> List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/E2KCDRMU2KESRNC35YXQUB365B4BF3BU/ >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/C4ALLDPXMSQDK5D5RIBT4D3IGQCQR5AS/

