[ovirt-users] Re: Beginning oVirt / Hyperconverged
On Tue, Nov 12, 2019 at 3:27 AM wrote: > I guess this is a little late now... but: > > I wanted to do the same, especially because the additional servers (beyond > 3) would lower the relative storage cost overhead when using erasure > coding, but it's 'off the trodden path' and I cannot recommend it, after > seeing just how fragile the whole house of cards is already. > Right, we only support replica 3 or replica 2 + arbiter, not erasure coded gluster volumes or replica 4. > The ability to extend a HCI cluster with additional storage nodes is very > much in the genes of Gluster, but oVirt is a little more than just Gluster > and while this capability is on the roadmap according to the RHEV roadmap > RH Summit video, it's not there today. > > But here is what you can already do: > * Save your pennies until you have enough for an extra two servers: HCI is > supposed to do great with 6 or 9 boxes (das rutschte einfach so > raus...sorry!) > > * You can always add any additional host as a compute host. I guess you > shouldn't add it as a host for the hosted-engine VM (extra tick mark), not > because it's impossible, but because it doesn't help making things more > resilient. The extra compute nodes are well managed with all capabilities > of oVirt (HA, live-migration etc.), because that's actually very close to > the original usage model. Of course, you won't benefit from the local > storage capacity. There is a facility for adding local storage to hosts, > but I didn't find any documentation on usage or recommendations and would > represent an obstacle to things like live-migration. > > * You can turn the forth host into a 'single-node HCI' for disaster > recovery testing (that's what I did, except that I didn't get very far > yet). Single node-HCI obviously has zero redundancy, but it also loses > benefits like GUI based upgrades (a chicken and egg issue). But the ability > to play with an additional DC and replication may be well worth allocating > the fourth machine to that purpose. Once you have lost a three-way HCI to > operator error or because of a software issue (as I have), that backup > seems mightily attractive, even if there doesn't seem to be any easy way to > extend that into a triple (wouldn't it be nice if you could do that without > even an interruption?). But a manual rebuild of the primary triple and > back-migration from the surviving single node HCI could save your job and > work. > > * You can always add nodes to the existing Gluster, but perhaps best not > extend the pre-configured Gluster volumes across the extra nodes. While > that sounds easy and attractive, I seriously doubt it is supported by the > current code base, making the management engine and the VDSM agent > essentially blind to the extra nodes or incapable of properly handling > faults and failures. But extra volumes to be used as application level > container storage or similar should be fine, even if they won't be properly > managed by oVirt. > You can expand the existing gluster volumes across the extra nodes. The configuration of the volume is still a replica 3 or replica 2 + arbiter depending on how you initially set it up - and that would mean that it can only tolerate 1 node failure in the replica set subvolume. > In my lab installations I moved the disks around on the four hosts so that > the two primary nodes and the fourth (DR) node had a full complement, while > the arbiter node had to manage with a subset of the boot SSD being used to > host the arbitration brick. In theory you should be able to manage better > in a three node HCI by spreading the three "normal" volumes "clockwise" > around the nodes. I used that setup for my initial Gluster based tests (not > using the HCI wizard), but that has oVirt treat the Gluster like a SAN. > > The HCI wizard doesn't support the clockwise or round-robin allocation, > but if you feel adventurous you should be able to turn a standard > installation into that, purely operating with brick > additions/changes/removals but again, given the fragility of the stack and > the extend of the documentation, I'd recommend against it. > You can now also edit your ansible playbook to specify the placement of the bricks across nodes during deployment - see https://github.com/gluster/gluster-ansible-features/pull/34, but does require some tweaking. > A full three-way replication (instead of 2+1 arbitration) can be > beneficial if your cluster runs on standard sized "Lego" left-overs and you > can get an extra box easily. But it will actually add write amplification > as a bit of a CPU overhead, but perhaps more importantly, at the network > level. If you got 10Gbit or better and slower storage, that may not be an > issue, if you got SSDs and Gbit, you can easily loose 33% write bandwidth, > while read-bandwith should be unaffected. > ___ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > P
[ovirt-users] Re: Beginning oVirt / Hyperconverged
I guess this is a little late now... but: I wanted to do the same, especially because the additional servers (beyond 3) would lower the relative storage cost overhead when using erasure coding, but it's 'off the trodden path' and I cannot recommend it, after seeing just how fragile the whole house of cards is already. The ability to extend a HCI cluster with additional storage nodes is very much in the genes of Gluster, but oVirt is a little more than just Gluster and while this capability is on the roadmap according to the RHEV roadmap RH Summit video, it's not there today. But here is what you can already do: * Save your pennies until you have enough for an extra two servers: HCI is supposed to do great with 6 or 9 boxes (das rutschte einfach so raus...sorry!) * You can always add any additional host as a compute host. I guess you shouldn't add it as a host for the hosted-engine VM (extra tick mark), not because it's impossible, but because it doesn't help making things more resilient. The extra compute nodes are well managed with all capabilities of oVirt (HA, live-migration etc.), because that's actually very close to the original usage model. Of course, you won't benefit from the local storage capacity. There is a facility for adding local storage to hosts, but I didn't find any documentation on usage or recommendations and would represent an obstacle to things like live-migration. * You can turn the forth host into a 'single-node HCI' for disaster recovery testing (that's what I did, except that I didn't get very far yet). Single node-HCI obviously has zero redundancy, but it also loses benefits like GUI based upgrades (a chicken and egg issue). But the ability to play with an additional DC and replication may be well worth allocating the fourth machine to that purpose. Once you have lost a three-way HCI to operator error or because of a software issue (as I have), that backup seems mightily attractive, even if there doesn't seem to be any easy way to extend that into a triple (wouldn't it be nice if you could do that without even an interruption?). But a manual rebuild of the primary triple and back-migration from the surviving single node HCI could save your job and work. * You can always add nodes to the existing Gluster, but perhaps best not extend the pre-configured Gluster volumes across the extra nodes. While that sounds easy and attractive, I seriously doubt it is supported by the current code base, making the management engine and the VDSM agent essentially blind to the extra nodes or incapable of properly handling faults and failures. But extra volumes to be used as application level container storage or similar should be fine, even if they won't be properly managed by oVirt. In my lab installations I moved the disks around on the four hosts so that the two primary nodes and the fourth (DR) node had a full complement, while the arbiter node had to manage with a subset of the boot SSD being used to host the arbitration brick. In theory you should be able to manage better in a three node HCI by spreading the three "normal" volumes "clockwise" around the nodes. I used that setup for my initial Gluster based tests (not using the HCI wizard), but that has oVirt treat the Gluster like a SAN. The HCI wizard doesn't support the clockwise or round-robin allocation, but if you feel adventurous you should be able to turn a standard installation into that, purely operating with brick additions/changes/removals but again, given the fragility of the stack and the extend of the documentation, I'd recommend against it. A full three-way replication (instead of 2+1 arbitration) can be beneficial if your cluster runs on standard sized "Lego" left-overs and you can get an extra box easily. But it will actually add write amplification as a bit of a CPU overhead, but perhaps more importantly, at the network level. If you got 10Gbit or better and slower storage, that may not be an issue, if you got SSDs and Gbit, you can easily loose 33% write bandwidth, while read-bandwith should be unaffected. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/4K7ZZ2WL57GWVR6CL3TM6L6XKJI5JVI2/
[ovirt-users] Re: Beginning oVirt / Hyperconverged
Iirc you need to have multiples of three severs for hyperconverged. You could have a 4 server cluster but only three of them would be participating in storage and one would be compute only. Someone correct me if I’m wrong. On Thu, Oct 31, 2019 at 6:41 AM wrote: > Hey folks! > > Quick question, really: I have 4 servers of identical hardware. The > documentation says "you need 3", not "you need 3 or more"; is it possible > to run hyperconverged with 4 servers (in same rack, let's neglect the > possibility of 2 failing servers (split brain))? > > Also: I have a nice seperate server (low power footprint, big on cpu (8) > and has 32gb ram)-- would love to use this as engine server only. Can I > initiate a hyperconverged system with 4 working host starting from a > stand-alone engine? Or does "hyperconverged" a must for HA engine on the > cluster? > > One last one: Installing ovirt nodes, obviously. This will by default > create a lvm across the entire RAID volume; can I simply create folders > (with correct permissions) and use those as gluster bricks? Or do I need to > partition the RAID in a special way? > > Thanks for your pointers, > Chris. > ___ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/UL3KCGA64RKITPZXULRYU6IM6CHZVVOR/ > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/KVMC2UYQSBUYYN43PYSEB6AP2MZFDJAG/