On Tue, Mar 15, 2022 at 7:43 AM Jonas Liechti <[email protected]> wrote: > Thank you for the confirmation Strahil! > > As our current environment is more or less the same (except the Hardware > RAID, which is not possible with NVMe disks), we planned to use Gluster. I > guess we will proceed as we originally planned as we are satisfied with the > performance. > Sorry for being late to the party.
Two items to consider: A. If you have more than two drives, consider using MDRAID 10 on NVME drives We have been using it for a while now and getting *very* good performance (more or less in-line with a single NVME drive). As an added bonus, we managed to survive a single drive failure without fault. B .If possible use a 40GbE network between the nodes instead of using 10GbE. It makes a huge impact. (SSDs or even a large HDD array can easily saturate a single 10GbE link). - Gilboa > On 3/11/22 07:08, Strahil Nikolov via Users wrote: > > Red Hat Gluster Storage is discontinued, but the Gluster (upstream) is > pretty active and as Sandro Bonazzola (RH) confirmed -> there are no plans > to remove support for Gluster. > I think it's still a good choice, especially if you don't have SAN/ > Higly-Available NFS. > > Also, storage migration is transparent for the VMs, so you can add SAN on > a later stage and move all VMs from Gluster to SAN without disruption* . > > Keep in mind that Gluster is a tier2 storage and if you really need a lot > of IOPS, CEPH might be suitable. > > > Best Regards, > Strahil Nikolov > > *: Note that this is valid if the FUSE client is used. Other oVirt users > report huge performance increase when using libgfapi interface, but it has > drawbacks like storage migration can happen only when you switch off > libgfapi, power off the VM (on a scheduled basis), power on the VM, live > migrate the VM to other storage type, enable libgfapi again for the rest of > the VMs. > > > > Thanks to Nikolov and Strahil for the valuable input! I was off for a few > weeks, so I would like to apologize if I'm potentially reviving a zombie > thread. > > I am a bit confused about where to go with this environment after the > discontinuation of the hyperconverged setup. What alternative options are > there for us? Or do you think going the Gluster way would still be > advisable, even though it seems as it is being discontinued over time? > > Thanks for any input on this! > > Best regards, > Jonas > On 1/22/22 14:31, Strahil Nikolov via Users wrote: > > Using the wizzard is utilizing the Gluster Andible roles. > I would highly recommend using it, unless you know what you are doing (for > example storage alignment when using Hardware raid). > > Keep in mind that the DHT xlator (the logic in distributed volumes) is > shard aware, so your shards are spread between subvolumes and additional > performance can be gained.So using replicated-distributed volumes have > their benefits. > > If you decide to avoid the software raid, use only replica3 volumes as > with SSDs/NVMEs usually the failures are not physical, but logical (maximum > writes reached -> predictive failure -> total failure). > > Also, consider mounting via noatime/relatime and > context="system_u:object_r:glusterd_brick_t:s0" for your gluster bricks. > > Best Regards, > Strahil Nikolov > > On Fri, Jan 21, 2022 at 11:00, Gilboa Davara > <[email protected]> <[email protected]> wrote: > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > Privacy Statement: https://www.ovirt.org/privacy-policy.html > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > > https://lists.ovirt.org/archives/list/[email protected]/message/U2ZEWLRF5D6FENQEI5QXL77CMWB7XF32/ > > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > Privacy Statement: https://www.ovirt.org/privacy-policy.html > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/26AHNDSOJSIVTGYOEUFOY444YYBZCAIW/ > > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > Privacy Statement: https://www.ovirt.org/privacy-policy.html > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/JFFS2T25TIIHFEQMS2Y3BU4DARSIDE3U/ >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/DWLJDJ5VO2CV6DNYLMVOWSUXKCQQLPUW/

