Hello Kaustav, That's weird, I never saw any volumes under the storage tab since installation. I am using HC setup deployed using cockpit console.
https://imgur.com/a/nH9rzK8 Did I miss something? On Tue, Nov 27, 2018 at 2:50 PM Kaustav Majumder <kmaju...@redhat.com> wrote: > Click on volume for which you want to reset the brick-> under bricks tab > select the brick you wan to reset -> once you do you will see the 'Reset > Brick' option is active. > Attached is a screenshot -> https://i.imgur.com/QUMSrzt.png > > On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni < > abhishek.sahni1...@gmail.com> wrote: > >> Thanks Sahina for your response, I am not able to find it on UI, please >> help me to navigate? and yes I am using oVirt 4.2.6.4-1. >> >> On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <sab...@redhat.com> wrote: >> >>> >>> >>> On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni < >>> abhishek.sahni1...@gmail.com> wrote: >>> >>>> Hello Team, >>>> >>>> We are running a setup of 3-way replica HC gluster setup configured >>>> during the initial deployment from the cockpit console using ansible. >>>> >>>> NODE1 >>>> - /dev/sda (OS) >>>> - /dev/sdb ( Gluster Bricks ) >>>> * /gluster_bricks/engine/engine/ >>>> * /gluster_bricks/data/data/ >>>> * /gluster_bricks/vmstore/vmstore/ >>>> >>>> NODE2 and NODE3 with a similar setup. >>>> >>>> There is a mishap that /dev/sdb on NODE2 totally got crashed and now >>>> there is nothing inside. However, I have created similar directories after >>>> mounting it back i.e., >>>> >>>> * /gluster_bricks/engine/engine/ >>>> * /gluster_bricks/data/data/ >>>> * /gluster_bricks/vmstore/vmstore/ >>>> but it is not yet recovered. >>>> >>>> ===================================================== >>>> [root@node2 ~]# gluster volume status >>>> Status of volume: data >>>> Gluster process TCP Port RDMA Port >>>> Online Pid >>>> >>>> ------------------------------------------------------------------------------ >>>> Brick *.*.*.1:/gluster_bricks/data/data 49152 0 Y >>>> 11111 >>>> Brick *.*.*.2:/gluster_bricks/data/data N/A N/A N >>>> N/A >>>> Brick *.*.*.3:/gluster_bricks/data/data 49152 0 Y >>>> 4303 >>>> Self-heal Daemon on localhost N/A N/A Y >>>> 23976 >>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>> 27838 >>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>> 27424 >>>> >>>> Task Status of Volume data >>>> >>>> ------------------------------------------------------------------------------ >>>> There are no active volume tasks >>>> >>>> Status of volume: engine >>>> Gluster process TCP Port RDMA Port >>>> Online Pid >>>> >>>> ------------------------------------------------------------------------------ >>>> Brick *.*.*.1:/gluster_bricks/engine/eng >>>> ine 49153 0 Y >>>> 11117 >>>> Brick *.*.*.2:/gluster_bricks/engine/eng >>>> ine N/A N/A N >>>> N/A >>>> Brick *.*.*.3:/gluster_bricks/engine/eng >>>> ine 49153 0 Y >>>> 4314 >>>> Self-heal Daemon on localhost N/A N/A Y >>>> 23976 >>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>> 27424 >>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>> 27838 >>>> >>>> Task Status of Volume engine >>>> >>>> ------------------------------------------------------------------------------ >>>> There are no active volume tasks >>>> >>>> Status of volume: vmstore >>>> Gluster process TCP Port RDMA Port >>>> Online Pid >>>> >>>> ------------------------------------------------------------------------------ >>>> Brick *.*.*.1:/gluster_bricks/vmstore/vm >>>> store 49154 0 Y >>>> 21603 >>>> Brick *.*.*.2:/gluster_bricks/vmstore/vm >>>> store N/A N/A N >>>> N/A >>>> Brick *.*.*.3:/gluster_bricks/vmstore/vm >>>> store 49154 0 Y >>>> 26845 >>>> Self-heal Daemon on localhost N/A N/A Y >>>> 23976 >>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>> 27424 >>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>> 27838 >>>> >>>> Task Status of Volume vmstore >>>> >>>> ------------------------------------------------------------------------------ >>>> There are no active volume tasks >>>> ============================================================= >>>> >>>> >>>> Can someone please suggest the steps to recover the setup? >>>> >>>> I have tried the below workaround but it doesn't help. >>>> >>>> >>>> https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html >>>> >>> >>> You can reset the brick - if you're on oVirt 4.2.x, there's a UI option >>> in the bricks subtab to do this. >>> >>> >>>> >>>> -- >>>> >>>> ABHISHEK SAHNI >>>> Mob : +91-990-701-5143 >>>> >>>> >>>> _______________________________________________ >>>> Users mailing list -- users@ovirt.org >>>> To unsubscribe send an email to users-le...@ovirt.org >>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>> oVirt Code of Conduct: >>>> https://www.ovirt.org/community/about/community-guidelines/ >>>> List Archives: >>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/ >>>> >>> >> >> -- >> >> ABHISHEK SAHNI >> Mob : +91-990-701-5143 >> >> >> _______________________________________________ >> Users mailing list -- users@ovirt.org >> To unsubscribe send an email to users-le...@ovirt.org >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> List Archives: >> https://lists.ovirt.org/archives/list/users@ovirt.org/message/GEZZIGBWA6HQ55JV2PSP7VSSYPZQ2JAC/ >> > > > Thanks, > Kaustav > -- ABHISHEK SAHNI Mob : +91-990-701-5143
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/SY3HGC7ZTOB7S5YNGFAW7BXC6TOOGCVO/