Yes, in the end that makes perfectly sense.
Thank you very much Sahina !

On Fri, Mar 1, 2019, 07:45 Sahina Bose <[email protected]> wrote:

> On Wed, Feb 27, 2019 at 11:21 AM Leo David <[email protected]> wrote:
> >
> > Thank you Sahina, I'm in that conversation too :).
> > On the other hand...
> > In this case, setting this option on, would only make sense in
> multi-node setups, and not in single instance ones, where we only have one
> hypervisor accsessing the volume.
> > Please correct me if this is wrong.
> > Have a nice day,
>
> In single instance deployments too, the option ensures all writes
> (with o-direct flag) are flushed to disk and not cached.
> >
> > Leo
> >
> >
> > On Tue, Feb 26, 2019, 08:24 Sahina Bose <[email protected]> wrote:
> >>
> >>
> >>
> >>
> >> On Fri, Sep 14, 2018 at 3:35 PM Paolo Margara <[email protected]>
> wrote:
> >>>
> >>> Hi,
> >>>
> >>> but performance.strict-o-direct is not one of the option enabled by
> gdeploy during installation because it's supposed to give some sort of
> benefit?
> >>
> >>
> >> See
> https://lists.ovirt.org/archives/list/[email protected]/message/VS764WDBR2PLGGDZVRGBEM6OJCAFEM3R/
> on why the option is set.
> >>
> >>>
> >>> Paolo
> >>>
> >>>
> >>> Il 14/09/2018 11:34, Leo David ha scritto:
> >>>
> >>> performance.strict-o-direct:  on
> >>> This was the bloody option that created the botleneck ! It was ON.
> >>> So now i get an average of 17k random writes,  which is not bad at
> all. Below,  the volume options that worked for me:
> >>>
> >>> performance.strict-write-ordering: off
> >>> performance.strict-o-direct: off
> >>> server.event-threads: 4
> >>> client.event-threads: 4
> >>> performance.read-ahead: off
> >>> network.ping-timeout: 30
> >>> performance.quick-read: off
> >>> cluster.eager-lock: enable
> >>> performance.stat-prefetch: on
> >>> performance.low-prio-threads: 32
> >>> network.remote-dio: off
> >>> user.cifs: off
> >>> performance.io-cache: off
> >>> server.allow-insecure: on
> >>> features.shard: on
> >>> transport.address-family: inet
> >>> storage.owner-uid: 36
> >>> storage.owner-gid: 36
> >>> nfs.disable: on
> >>>
> >>> If any other tweaks can be done,  please let me know.
> >>> Thank you !
> >>>
> >>> Leo
> >>>
> >>>
> >>> On Fri, Sep 14, 2018 at 12:01 PM, Leo David <[email protected]> wrote:
> >>>>
> >>>> Hi Everyone,
> >>>> So i have decided to take out all of the gluster volume custom
> options,  and add them one by one while activating/deactivating the storage
> domain & rebooting one vm after each  added option :(
> >>>>
> >>>> The default options that giving bad iops ( ~1-2k) performance are :
> >>>>
> >>>> performance.stat-prefetch on
> >>>> cluster.eager-lock enable
> >>>> performance.io-cache off
> >>>> performance.read-ahead off
> >>>> performance.quick-read off
> >>>> user.cifs off
> >>>> network.ping-timeout 30
> >>>> network.remote-dio off
> >>>> performance.strict-o-direct on
> >>>> performance.low-prio-threads 32
> >>>>
> >>>> After adding only:
> >>>>
> >>>>
> >>>> server.allow-insecure on
> >>>> features.shard on
> >>>> storage.owner-gid 36
> >>>> storage.owner-uid 36
> >>>> transport.address-family inet
> >>>> nfs.disable on
> >>>>
> >>>> The performance increased to 7k-10k iops.
> >>>>
> >>>> The problem is that i don't know if that's sufficient ( maybe it can
> be more improved ) , or even worse than this there might be chances to into
> different volume issues by taking out some volume really needed options...
> >>>>
> >>>> If would have handy the default options that are applied to volumes
> as optimization in a 3way replica, I think that might help..
> >>>>
> >>>> Any thoughts ?
> >>>>
> >>>> Thank you very much !
> >>>>
> >>>>
> >>>> Leo
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>> On Fri, Sep 14, 2018 at 8:54 AM, Leo David <[email protected]> wrote:
> >>>>>
> >>>>> Any thoughs on these ? Is that UI optimization only a gluster volume
> custom configuration ? If so, i guess it can be done from cli, but I am not
> aware of the corect optimized parameters of the volume....
> >>>>>
> >>>>>
> >>>>> On Thu, Sep 13, 2018, 18:25 Leo David <[email protected]> wrote:
> >>>>>>
> >>>>>> Thank you Jayme. I am trying to do this, but I am getting an error,
> since the volume is replica 1 distribute, and it seems that oVirt expects a
> replica 3 volume.
> >>>>>> Would it be another way to optimize the volume in this situation ?
> >>>>>>
> >>>>>>
> >>>>>> On Thu, Sep 13, 2018, 17:49 Jayme <[email protected]> wrote:
> >>>>>>>
> >>>>>>> I had similar problems until I clicked "optimize volume for
> vmstore" in the admin GUI for each data volume.  I'm not sure if this is
> what is causing your problem here but I'd recommend trying that first.  It
> is suppose to be optimized by default but for some reason my ovirt 4.2
> cockpit deploy did not apply those settings automatically.
> >>>>>>>
> >>>>>>> On Thu, Sep 13, 2018 at 10:21 AM Leo David <[email protected]>
> wrote:
> >>>>>>>>
> >>>>>>>> Hi Everyone,
> >>>>>>>> I am encountering the following issue on a single instance
> hyper-converged 4.2 setup.
> >>>>>>>> The following fio test was done:
> >>>>>>>>
> >>>>>>>> fio --randrepeat=1 --ioengine=libaio --direct=1 --gtod_reduce=1
> --name=test --filename=test --bs=4k --iodepth=64 --size=4G
> --readwrite=randwrite
> >>>>>>>> The results are very poor doing the test inside of a vm with a
> prealocated disk on the ssd store:  ~2k IOPS
> >>>>>>>> Same test done on the oVirt node directly on the mounted ssd_lvm:
> ~30k IOPS
> >>>>>>>> Same test done, this time on the gluster mount path: ~20K IOPS
> >>>>>>>>
> >>>>>>>> What could be the issue that the vms have this slow hdd
> performance ( 2k on ssd !! )?
> >>>>>>>> Thank you very much !
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> --
> >>>>>>>> Best regards, Leo David
> >>>>>>>> _______________________________________________
> >>>>>>>> Users mailing list -- [email protected]
> >>>>>>>> To unsubscribe send an email to [email protected]
> >>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> >>>>>>>> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> >>>>>>>> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/FCCIZFRWINWWLQSYWRPF6HNKPQMZB2XP/
> >>>>
> >>>>
> >>>>
> >>>>
> >>>> --
> >>>> Best regards, Leo David
> >>>
> >>>
> >>>
> >>>
> >>> --
> >>> Best regards, Leo David
> >>>
> >>>
> >>> _______________________________________________
> >>> Users mailing list -- [email protected]
> >>> To unsubscribe send an email to [email protected]
> >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> >>> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> >>> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/5HKQQKX3BJLZ3HQ5SCHPLPON24OGMGSS/
> >>>
> >>> _______________________________________________
> >>> Users mailing list -- [email protected]
> >>> To unsubscribe send an email to [email protected]
> >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> >>> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> >>> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/E2KCDRMU2KESRNC35YXQUB365B4BF3BU/
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/C4ALLDPXMSQDK5D5RIBT4D3IGQCQR5AS/

Reply via email to