On Tue, Nov 12, 2019 at 3:27 AM <[email protected]> wrote:

> I guess this is a little late now... but:
>
> I wanted to do the same, especially because the additional servers (beyond
> 3) would lower the relative storage cost overhead when using erasure
> coding, but it's 'off the trodden path' and I cannot recommend it, after
> seeing just how fragile the whole house of cards is already.
>

Right, we only support replica 3 or replica 2 + arbiter, not erasure coded
gluster volumes or replica 4.


> The ability to extend a HCI cluster with additional storage nodes is very
> much in the genes of Gluster, but oVirt is a little more than just Gluster
> and while this capability is on the roadmap according to the RHEV roadmap
> RH Summit video, it's not there today.
>
> But here is what you can already do:
> * Save your pennies until you have enough for an extra two servers: HCI is
> supposed to do great with 6 or 9 boxes (das rutschte einfach so
> raus...sorry!)
>
> * You can always add any additional host as a compute host. I guess you
> shouldn't add it as a host for the hosted-engine VM (extra tick mark), not
> because it's impossible, but because it doesn't help making things more
> resilient. The extra compute nodes are well managed with all capabilities
> of oVirt (HA, live-migration etc.), because that's actually very close to
> the original usage model. Of course, you won't benefit from the local
> storage capacity. There is a facility for adding local storage to hosts,
> but I didn't find any documentation on usage or recommendations and would
> represent an obstacle to things like live-migration.
>
> * You can turn the forth host into a 'single-node HCI' for disaster
> recovery testing (that's what I did, except that I didn't get very far
> yet). Single node-HCI obviously has zero redundancy, but it also loses
> benefits like GUI based upgrades (a chicken and egg issue). But the ability
> to play with an additional DC and replication may be well worth allocating
> the fourth machine to that purpose. Once you have lost a three-way HCI to
> operator error or because of a software issue (as I have), that backup
> seems mightily attractive, even if there doesn't seem to be any easy way to
> extend that into a triple (wouldn't it be nice if you could do that without
> even an interruption?). But a manual rebuild of the primary triple and
> back-migration from the surviving single node HCI could save your job and
> work.
>
> * You can always add nodes to the existing Gluster, but perhaps best not
> extend the pre-configured Gluster volumes across the extra nodes. While
> that sounds easy and attractive, I seriously doubt it is supported by the
> current code base, making the management engine and the VDSM agent
> essentially blind to the extra nodes or incapable of properly handling
> faults and failures. But extra volumes to be used as application level
> container storage or similar should be fine, even if they won't be properly
> managed by oVirt.
>

You can expand the existing gluster volumes across the extra nodes. The
configuration of the volume is still a replica 3 or replica 2 + arbiter
depending on how you initially set it up - and that would mean that it can
only tolerate 1 node failure in the replica set subvolume.


> In my lab installations I moved the disks around on the four hosts so that
> the two primary nodes and the fourth (DR) node had a full complement, while
> the arbiter node had to manage with a subset of the boot SSD being used to
> host the arbitration brick. In theory you should be able to manage better
> in a three node HCI by spreading the three "normal" volumes "clockwise"
> around the nodes. I used that setup for my initial Gluster based tests (not
> using the HCI wizard), but that has oVirt treat the Gluster like a SAN.
>
> The HCI wizard doesn't support the clockwise or round-robin allocation,
> but if you feel adventurous you should be able to turn a standard
> installation into that, purely operating with brick
> additions/changes/removals but again, given the fragility of the stack and
> the extend of the documentation, I'd recommend against it.
>

You can now also edit your ansible playbook to specify the placement of the
bricks across nodes during deployment - see
https://github.com/gluster/gluster-ansible-features/pull/34, but does
require some tweaking.


> A full three-way replication (instead of 2+1 arbitration) can be
> beneficial if your cluster runs on standard sized "Lego" left-overs and you
> can get an extra box easily. But it will actually add write amplification
> as a bit of a CPU overhead, but perhaps more importantly, at the network
> level. If you got 10Gbit or better and slower storage, that may not be an
> issue, if you got SSDs and Gbit, you can easily loose 33% write bandwidth,
> while read-bandwith should be unaffected.
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/4K7ZZ2WL57GWVR6CL3TM6L6XKJI5JVI2/
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/Y24CL2ZKJN5OVOHYE26CJIVQYG65PEJP/

Reply via email to