[ovirt-users] Re: oVirt 2.5.1.2 HCI Gluster Recovery Options

2022-12-12 Thread Strahil Nikolov via Users
I asked if the OS survived, as then you would need to:- Recreate the LVM 
structure (it's the same on all 3 nodes)- Create the filesystem with isize=512- 
Permanently mount the file system- Fix the volume in order to heal
I think that you might have fixed the first 3 and the simplest way from that 
point is to use 'remove-brick' to switch into replica 2 mode:
#Get the syntaxgluster volume remove-brick help
#The failed brick won't have a pidgluster volume status 
#Actual replace:#Tripple Check the removed brick name!!!gluster volume 
remove-brick replica 2  commit force
#Now add the new brick (which might be the same name as the old one):gluster 
volume add-brick  replica 3  
#Last but still very importantgluster volume heal  full
#Verifygluster volume listgluster volume status
You should be able to do everything from the UI (I personally never did it from 
there).
If you need to make the bricks manually, take a look at: 
https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.5/html/administration_guide/brick_configuration
 .Disk alignment is very important on all layers.
For the mount options you can 
use:noatime,context="system_u:object_r:glusterd_brick_t:s0"
Best Regards,Strahil Nikolov 
 
 
  On Sat, Dec 10, 2022 at 21:49, Clint Boggio wrote:   
Good Day Strahil;

The OS did survive. The OS was on a RAID1 array and the gluster bricks are on a 
RAID5 array. Since the time I requested guidance I did a whole lot of reading 
and learning. I found in /etc/ansible several yaml files that contained the 
configuration of the gluster bricks and the lvm volumes contained within them. 
I also found that lvm configurations can be restored from /etc/lvm/backup and 
/etc/lvm/archive. All very helpfull I must say.

I re-created the RAID array and did a light initialization to wipe the array. 
Then I attempted to restore the backups from /etc/lvm/backup and 
/etc/lvm/array. I was unable to inspire the thinpool to enable with the "manual 
repair required" error each time. As there was not enough unused space on the 
system to cycle out the metadata lvm I was unable to proceed.

I then deleted the LVM configuration and re-created the LVM volumes from the 
/etc/ansible yaml files I found. I made them identical to the other two nodes 
that were functioning in the cluster. I have them succesfully mounted via fstab 
with the new UUID that were generated when I recreated the volumes. 

Based on the documents and articles I have read and studied these past few 
weeks, at this point I believe that I need to re-apply the appropriate 
gluster-related attributes to the correct level of the gluster mounts and 
attempt to start the gluster service. I have not done this yet because I don't 
want to cause another outage and I want to be sure that what I am doing will 
result in the afflicted gluster volumes healing and sychronising with the 
healthy nodes.

1. What attributes and at what level should I apply them to the refurbished 
node ?
2. What problems could I encounter when I try to start the gluster service 
after making the changes ?
3. Should I remove the new node from the gluster cluster and then re-add it or 
will it heal in it's current refurbished state ?

I appreciate you taking time to check on the status of this situation and thank 
you for any help or insight you can provide.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UEIP7GKIG3IMT2UMPTLYV475XMVRULOE/
  
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/F5Q7AIJNWHUSSGG7ECNFNCFZKEGEUEAZ/


[ovirt-users] Re: oVirt 2.5.1.2 HCI Gluster Recovery Options

2022-12-10 Thread Clint Boggio
Good Day Strahil;

The OS did survive. The OS was on a RAID1 array and the gluster bricks are on a 
RAID5 array. Since the time I requested guidance I did a whole lot of reading 
and learning. I found in /etc/ansible several yaml files that contained the 
configuration of the gluster bricks and the lvm volumes contained within them. 
I also found that lvm configurations can be restored from /etc/lvm/backup and 
/etc/lvm/archive. All very helpfull I must say.

I re-created the RAID array and did a light initialization to wipe the array. 
Then I attempted to restore the backups from /etc/lvm/backup and 
/etc/lvm/array. I was unable to inspire the thinpool to enable with the "manual 
repair required" error each time. As there was not enough unused space on the 
system to cycle out the metadata lvm I was unable to proceed.

I then deleted the LVM configuration and re-created the LVM volumes from the 
/etc/ansible yaml files I found. I made them identical to the other two nodes 
that were functioning in the cluster. I have them succesfully mounted via fstab 
with the new UUID that were generated when I recreated the volumes. 

Based on the documents and articles I have read and studied these past few 
weeks, at this point I believe that I need to re-apply the appropriate 
gluster-related attributes to the correct level of the gluster mounts and 
attempt to start the gluster service. I have not done this yet because I don't 
want to cause another outage and I want to be sure that what I am doing will 
result in the afflicted gluster volumes healing and sychronising with the 
healthy nodes.

1. What attributes and at what level should I apply them to the refurbished 
node ?
2. What problems could I encounter when I try to start the gluster service 
after making the changes ?
3. Should I remove the new node from the gluster cluster and then re-add it or 
will it heal in it's current refurbished state ?

I appreciate you taking time to check on the status of this situation and thank 
you for any help or insight you can provide.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UEIP7GKIG3IMT2UMPTLYV475XMVRULOE/


[ovirt-users] Re: oVirt 2.5.1.2 HCI Gluster Recovery Options

2022-12-09 Thread Strahil Nikolov via Users
Did the OS of the node survived ?
Best Regards,Strahil Nikolov 
 
 
  On Tue, Dec 6, 2022 at 0:01, Clint Boggio wrote:   
Greetings oVirt community;

I've got a three-node HCI cluster with hosted engine and on one of the three 
nodes I lost the underlying RAID array under the Gluster bricks. It looks like 
I'm going to have to rebuild the RAID array and thus likely lose the Gluster 
configuration that lives there. The host OS (oVirt Node) is fine and it's just 
the Gluster/RAID piece that is affected. 

Can one of you please point me to the best documentation link to the process of 
re-configuring that failed node to be identical to the rest of the nodes please 
? It's configured for replica-three and I don't want to have a problem with 
nodes 2 and 3 before I get node 1 back up and healthy.

Any help will be greatly appreciated.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QSWBLY3QJOSOTGIHYPJYQWP3XRQLEV7P/
  
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/2C6RWMK5EEQ5LYGZOTRFKLKSJUHICLME/