Click on volume for which you want to reset the brick-> under bricks tab select the brick you wan to reset -> once you do you will see the 'Reset Brick' option is active. Attached is a screenshot -> https://i.imgur.com/QUMSrzt.png
On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni <[email protected]> wrote: > Thanks Sahina for your response, I am not able to find it on UI, please > help me to navigate? and yes I am using oVirt 4.2.6.4-1. > > On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <[email protected]> wrote: > >> >> >> On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni < >> [email protected]> wrote: >> >>> Hello Team, >>> >>> We are running a setup of 3-way replica HC gluster setup configured >>> during the initial deployment from the cockpit console using ansible. >>> >>> NODE1 >>> - /dev/sda (OS) >>> - /dev/sdb ( Gluster Bricks ) >>> * /gluster_bricks/engine/engine/ >>> * /gluster_bricks/data/data/ >>> * /gluster_bricks/vmstore/vmstore/ >>> >>> NODE2 and NODE3 with a similar setup. >>> >>> There is a mishap that /dev/sdb on NODE2 totally got crashed and now >>> there is nothing inside. However, I have created similar directories after >>> mounting it back i.e., >>> >>> * /gluster_bricks/engine/engine/ >>> * /gluster_bricks/data/data/ >>> * /gluster_bricks/vmstore/vmstore/ >>> but it is not yet recovered. >>> >>> ===================================================== >>> [root@node2 ~]# gluster volume status >>> Status of volume: data >>> Gluster process TCP Port RDMA Port Online >>> Pid >>> >>> ------------------------------------------------------------------------------ >>> Brick *.*.*.1:/gluster_bricks/data/data 49152 0 Y >>> 11111 >>> Brick *.*.*.2:/gluster_bricks/data/data N/A N/A N >>> N/A >>> Brick *.*.*.3:/gluster_bricks/data/data 49152 0 Y >>> 4303 >>> Self-heal Daemon on localhost N/A N/A Y >>> 23976 >>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>> 27838 >>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>> 27424 >>> >>> Task Status of Volume data >>> >>> ------------------------------------------------------------------------------ >>> There are no active volume tasks >>> >>> Status of volume: engine >>> Gluster process TCP Port RDMA Port Online >>> Pid >>> >>> ------------------------------------------------------------------------------ >>> Brick *.*.*.1:/gluster_bricks/engine/eng >>> ine 49153 0 Y >>> 11117 >>> Brick *.*.*.2:/gluster_bricks/engine/eng >>> ine N/A N/A N >>> N/A >>> Brick *.*.*.3:/gluster_bricks/engine/eng >>> ine 49153 0 Y >>> 4314 >>> Self-heal Daemon on localhost N/A N/A Y >>> 23976 >>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>> 27424 >>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>> 27838 >>> >>> Task Status of Volume engine >>> >>> ------------------------------------------------------------------------------ >>> There are no active volume tasks >>> >>> Status of volume: vmstore >>> Gluster process TCP Port RDMA Port Online >>> Pid >>> >>> ------------------------------------------------------------------------------ >>> Brick *.*.*.1:/gluster_bricks/vmstore/vm >>> store 49154 0 Y >>> 21603 >>> Brick *.*.*.2:/gluster_bricks/vmstore/vm >>> store N/A N/A N >>> N/A >>> Brick *.*.*.3:/gluster_bricks/vmstore/vm >>> store 49154 0 Y >>> 26845 >>> Self-heal Daemon on localhost N/A N/A Y >>> 23976 >>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>> 27424 >>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>> 27838 >>> >>> Task Status of Volume vmstore >>> >>> ------------------------------------------------------------------------------ >>> There are no active volume tasks >>> ============================================================= >>> >>> >>> Can someone please suggest the steps to recover the setup? >>> >>> I have tried the below workaround but it doesn't help. >>> >>> >>> https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html >>> >> >> You can reset the brick - if you're on oVirt 4.2.x, there's a UI option >> in the bricks subtab to do this. >> >> >>> >>> -- >>> >>> ABHISHEK SAHNI >>> Mob : +91-990-701-5143 >>> >>> >>> _______________________________________________ >>> Users mailing list -- [email protected] >>> To unsubscribe send an email to [email protected] >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>> oVirt Code of Conduct: >>> https://www.ovirt.org/community/about/community-guidelines/ >>> List Archives: >>> https://lists.ovirt.org/archives/list/[email protected]/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/ >>> >> > > -- > > ABHISHEK SAHNI > Mob : +91-990-701-5143 > > > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/GEZZIGBWA6HQ55JV2PSP7VSSYPZQ2JAC/ > Thanks, Kaustav
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/R3Y3YMUD5JIRPTZ4JVJTNZI7MXR4SQAW/

