On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni <[email protected]> wrote:
> Hello Team, > > We are running a setup of 3-way replica HC gluster setup configured during > the initial deployment from the cockpit console using ansible. > > NODE1 > - /dev/sda (OS) > - /dev/sdb ( Gluster Bricks ) > * /gluster_bricks/engine/engine/ > * /gluster_bricks/data/data/ > * /gluster_bricks/vmstore/vmstore/ > > NODE2 and NODE3 with a similar setup. > > There is a mishap that /dev/sdb on NODE2 totally got crashed and now there > is nothing inside. However, I have created similar directories after > mounting it back i.e., > > * /gluster_bricks/engine/engine/ > * /gluster_bricks/data/data/ > * /gluster_bricks/vmstore/vmstore/ > but it is not yet recovered. > > ===================================================== > [root@node2 ~]# gluster volume status > Status of volume: data > Gluster process TCP Port RDMA Port Online > Pid > > ------------------------------------------------------------------------------ > Brick *.*.*.1:/gluster_bricks/data/data 49152 0 Y 11111 > Brick *.*.*.2:/gluster_bricks/data/data N/A N/A N N/A > Brick *.*.*.3:/gluster_bricks/data/data 49152 0 Y 4303 > Self-heal Daemon on localhost N/A N/A Y > 23976 > Self-heal Daemon on *.*.*.1 N/A N/A Y 27838 > Self-heal Daemon on *.*.*.3 N/A N/A Y 27424 > > Task Status of Volume data > > ------------------------------------------------------------------------------ > There are no active volume tasks > > Status of volume: engine > Gluster process TCP Port RDMA Port Online > Pid > > ------------------------------------------------------------------------------ > Brick *.*.*.1:/gluster_bricks/engine/eng > ine 49153 0 Y > 11117 > Brick *.*.*.2:/gluster_bricks/engine/eng > ine N/A N/A N > N/A > Brick *.*.*.3:/gluster_bricks/engine/eng > ine 49153 0 Y > 4314 > Self-heal Daemon on localhost N/A N/A Y > 23976 > Self-heal Daemon on *.*.*.3 N/A N/A Y 27424 > Self-heal Daemon on *.*.*.1 N/A N/A Y 27838 > > Task Status of Volume engine > > ------------------------------------------------------------------------------ > There are no active volume tasks > > Status of volume: vmstore > Gluster process TCP Port RDMA Port Online > Pid > > ------------------------------------------------------------------------------ > Brick *.*.*.1:/gluster_bricks/vmstore/vm > store 49154 0 Y > 21603 > Brick *.*.*.2:/gluster_bricks/vmstore/vm > store N/A N/A N > N/A > Brick *.*.*.3:/gluster_bricks/vmstore/vm > store 49154 0 Y > 26845 > Self-heal Daemon on localhost N/A N/A Y > 23976 > Self-heal Daemon on *.*.*.3 N/A N/A Y 27424 > Self-heal Daemon on *.*.*.1 N/A N/A Y 27838 > > Task Status of Volume vmstore > > ------------------------------------------------------------------------------ > There are no active volume tasks > ============================================================= > > > Can someone please suggest the steps to recover the setup? > > I have tried the below workaround but it doesn't help. > > https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html > You can reset the brick - if you're on oVirt 4.2.x, there's a UI option in the bricks subtab to do this. > > -- > > ABHISHEK SAHNI > Mob : +91-990-701-5143 > > > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/ >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/SOFSDQNT4OIOM6LQLBCNZ4PYEEHBTXFU/

