For the migration of the HCI to another set of NICs, it's possible. WARNING: Ensure that all hosts can reach themselves on the new IPs. ssh is a good test (unless it has been hardened).
There are 2 paths to take: - Add the new ip as a peer and replace-brick to use the new ip/hostname/fqdn gluster peer probe new_ip gluster volume replace-brick VOL old_ip:/path_to_brick new_ip:/path_to_brick Note: You might need 'force' in order to do it. This one is tricky to cleanup the old IP , you have to tinker inside the gluster configs . If you are ok to see both IPs for the gluster node , then this one is the easier - Reduce the replica count with remove brick , gluster peer detach old_ip, and then gluster peer probe and add brick: gluster volume remove-brick replica 2 VOL old_ip:/path_to_brick Most probably it won't work with 'commit' and you might need to use 'force' gluster peer detach old_ip gluster peer probe new_ip umount /path_to_brick mkfs.xfs -f -i size=512 /block/device/hosting/the/brick mount /path_to_brick gluster volume add-brick VOL replica 3 new_ip:/path_to_brick and wait for the heal to end Once done, repeat for all nodes. As you got a test gluster cluster, you can test it there. Best Regards, Strahil Nikolov На 12 август 2020 г. 6:27:56 GMT+03:00, [email protected] написа: >Thanks for putting in the effort! > >I learned a lot of new things. >I also learned that I need to learn a few more now. >The table could use some alternating background or a grid: Too easy to >get lost. > >Environments change over time, e.g. you find you really should have >split the management, storage, migration and north-south networks, but >any documentation I find says "do it right from the start". > >So I wonder if there is any easy enough way to split out the storage >network to a newly added second set of NICs in a HCI environment, or if >a re-install is really the only reasonable thing to do. > >They have created such a nice looking GUI around the whole network >configuration stuff, but from what I have experienced, just hitting >buttons is very dangerous there, while many don't have a description of >help item. > >Since you're at it: I have been able to make nested virtualization work >to a degree. >3-node HCI (physical) is hosting another 3-node HCI (virtual), virtual >oVirt Cockpit deployed a working gluster, launched and prepared the >HostedEngine the nested level, managed to move it on the Gluster and if >I boot the virtual nodes, their VDSM will launch the nested hosted >engine, but that machine can't see the network any more. I can connect >to it via hosted-engine --console, it has an Ethernet interface, but no >traffic gets through either way: Ideas? What do I look for? > >I am wildy guessing that ovn has nesting issues, so would going the >Linux bridge approach help there? How do I chose between the two? Must >I leave my beloved Cockpit wizard and use the script installer? >_______________________________________________ >Users mailing list -- [email protected] >To unsubscribe send an email to [email protected] >Privacy Statement: https://www.ovirt.org/privacy-policy.html >oVirt Code of Conduct: >https://www.ovirt.org/community/about/community-guidelines/ >List Archives: >https://lists.ovirt.org/archives/list/[email protected]/message/IWR5B5EUYF5JBNWSHATJQ3RH7SIXWHBH/ _______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/JLQF6OLF2J2RGU2ID7AWJUHCSVRS2TIA/

