You are welcome. On Tue, Nov 27, 2018 at 4:24 PM Abhishek Sahni <[email protected]> wrote:
> Thanks a lot. :-) > > On Tue, Nov 27, 2018 at 4:22 PM Kaustav Majumder <[email protected]> > wrote: > >> >> >> On Tue, Nov 27, 2018 at 4:05 PM Abhishek Sahni <[email protected]> >> wrote: >> >>> That is amazing, resetting bricks resolved the issue. >>> >>> Thanks much Sahina and Kaustav. >>> >>> However, Do we have manual steps to recover those bricks. >>> >> https://gluster.readthedocs.io/en/latest/release-notes/3.9.0/ >> >>> >>> On Tue, Nov 27, 2018 at 3:57 PM Abhishek Sahni <[email protected]> >>> wrote: >>> >>>> I just enabled it on default cluster and now the volumes are visible. >>>> It seems like gluster service was disabled by default on cluster. >>>> >>>> On Tue, Nov 27, 2018 at 3:51 PM Sahina Bose <[email protected]> wrote: >>>> >>>>> >>>>> >>>>> On Tue, Nov 27, 2018 at 3:45 PM Kaustav Majumder <[email protected]> >>>>> wrote: >>>>> >>>>>> I am not sure why ovirt is not showing any volume. >>>>>> Sahina, is this a bug? >>>>>> >>>>> >>>>> Check if gluster service is enabled on the cluster. >>>>> The volumes are managed only if this is true >>>>> >>>>> >>>>>> On Tue, Nov 27, 2018 at 3:10 PM Abhishek Sahni < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> Hello Kaustav, >>>>>>> >>>>>>> That's weird, I never saw any volumes under the storage tab since >>>>>>> installation. I am using HC setup deployed using cockpit console. >>>>>>> >>>>>>> https://imgur.com/a/nH9rzK8 >>>>>>> >>>>>>> Did I miss something? >>>>>>> >>>>>>> >>>>>>> On Tue, Nov 27, 2018 at 2:50 PM Kaustav Majumder < >>>>>>> [email protected]> wrote: >>>>>>> >>>>>>>> Click on volume for which you want to reset the brick-> under >>>>>>>> bricks tab select the brick you wan to reset -> once you do you will >>>>>>>> see >>>>>>>> the 'Reset Brick' option is active. >>>>>>>> Attached is a screenshot -> https://i.imgur.com/QUMSrzt.png >>>>>>>> >>>>>>>> On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni < >>>>>>>> [email protected]> wrote: >>>>>>>> >>>>>>>>> Thanks Sahina for your response, I am not able to find it on UI, >>>>>>>>> please help me to navigate? and yes I am using oVirt 4.2.6.4-1. >>>>>>>>> >>>>>>>>> On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <[email protected]> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni < >>>>>>>>>> [email protected]> wrote: >>>>>>>>>> >>>>>>>>>>> Hello Team, >>>>>>>>>>> >>>>>>>>>>> We are running a setup of 3-way replica HC gluster setup >>>>>>>>>>> configured during the initial deployment from the cockpit console >>>>>>>>>>> using >>>>>>>>>>> ansible. >>>>>>>>>>> >>>>>>>>>>> NODE1 >>>>>>>>>>> - /dev/sda (OS) >>>>>>>>>>> - /dev/sdb ( Gluster Bricks ) >>>>>>>>>>> * /gluster_bricks/engine/engine/ >>>>>>>>>>> * /gluster_bricks/data/data/ >>>>>>>>>>> * /gluster_bricks/vmstore/vmstore/ >>>>>>>>>>> >>>>>>>>>>> NODE2 and NODE3 with a similar setup. >>>>>>>>>>> >>>>>>>>>>> There is a mishap that /dev/sdb on NODE2 totally got crashed and >>>>>>>>>>> now there is nothing inside. However, I have created similar >>>>>>>>>>> directories >>>>>>>>>>> after mounting it back i.e., >>>>>>>>>>> >>>>>>>>>>> * /gluster_bricks/engine/engine/ >>>>>>>>>>> * /gluster_bricks/data/data/ >>>>>>>>>>> * /gluster_bricks/vmstore/vmstore/ >>>>>>>>>>> but it is not yet recovered. >>>>>>>>>>> >>>>>>>>>>> ===================================================== >>>>>>>>>>> [root@node2 ~]# gluster volume status >>>>>>>>>>> Status of volume: data >>>>>>>>>>> Gluster process TCP Port RDMA Port >>>>>>>>>>> Online Pid >>>>>>>>>>> >>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>> Brick *.*.*.1:/gluster_bricks/data/data 49152 0 Y >>>>>>>>>>> 11111 >>>>>>>>>>> Brick *.*.*.2:/gluster_bricks/data/data N/A N/A N >>>>>>>>>>> N/A >>>>>>>>>>> Brick *.*.*.3:/gluster_bricks/data/data 49152 0 Y >>>>>>>>>>> 4303 >>>>>>>>>>> Self-heal Daemon on localhost N/A N/A >>>>>>>>>>> Y 23976 >>>>>>>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>>>>>>> 27838 >>>>>>>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>>>>>>> 27424 >>>>>>>>>>> >>>>>>>>>>> Task Status of Volume data >>>>>>>>>>> >>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>> There are no active volume tasks >>>>>>>>>>> >>>>>>>>>>> Status of volume: engine >>>>>>>>>>> Gluster process TCP Port RDMA Port >>>>>>>>>>> Online Pid >>>>>>>>>>> >>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>> Brick *.*.*.1:/gluster_bricks/engine/eng >>>>>>>>>>> ine 49153 0 >>>>>>>>>>> Y 11117 >>>>>>>>>>> Brick *.*.*.2:/gluster_bricks/engine/eng >>>>>>>>>>> ine N/A N/A >>>>>>>>>>> N N/A >>>>>>>>>>> Brick *.*.*.3:/gluster_bricks/engine/eng >>>>>>>>>>> ine 49153 0 >>>>>>>>>>> Y 4314 >>>>>>>>>>> Self-heal Daemon on localhost N/A N/A >>>>>>>>>>> Y 23976 >>>>>>>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>>>>>>> 27424 >>>>>>>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>>>>>>> 27838 >>>>>>>>>>> >>>>>>>>>>> Task Status of Volume engine >>>>>>>>>>> >>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>> There are no active volume tasks >>>>>>>>>>> >>>>>>>>>>> Status of volume: vmstore >>>>>>>>>>> Gluster process TCP Port RDMA Port >>>>>>>>>>> Online Pid >>>>>>>>>>> >>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>> Brick *.*.*.1:/gluster_bricks/vmstore/vm >>>>>>>>>>> store 49154 0 >>>>>>>>>>> Y 21603 >>>>>>>>>>> Brick *.*.*.2:/gluster_bricks/vmstore/vm >>>>>>>>>>> store N/A N/A >>>>>>>>>>> N N/A >>>>>>>>>>> Brick *.*.*.3:/gluster_bricks/vmstore/vm >>>>>>>>>>> store 49154 0 >>>>>>>>>>> Y 26845 >>>>>>>>>>> Self-heal Daemon on localhost N/A N/A >>>>>>>>>>> Y 23976 >>>>>>>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>>>>>>> 27424 >>>>>>>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>>>>>>> 27838 >>>>>>>>>>> >>>>>>>>>>> Task Status of Volume vmstore >>>>>>>>>>> >>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>> There are no active volume tasks >>>>>>>>>>> ============================================================= >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Can someone please suggest the steps to recover the setup? >>>>>>>>>>> >>>>>>>>>>> I have tried the below workaround but it doesn't help. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> You can reset the brick - if you're on oVirt 4.2.x, there's a UI >>>>>>>>>> option in the bricks subtab to do this. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> >>>>>>>>>>> ABHISHEK SAHNI >>>>>>>>>>> Mob : +91-990-701-5143 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Users mailing list -- [email protected] >>>>>>>>>>> To unsubscribe send an email to [email protected] >>>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>>>>>> oVirt Code of Conduct: >>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>>>>> List Archives: >>>>>>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/ >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> >>>>>>>>> ABHISHEK SAHNI >>>>>>>>> Mob : +91-990-701-5143 >>>>>>>>> >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> Users mailing list -- [email protected] >>>>>>>>> To unsubscribe send an email to [email protected] >>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>>>> oVirt Code of Conduct: >>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>>> List Archives: >>>>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/GEZZIGBWA6HQ55JV2PSP7VSSYPZQ2JAC/ >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> Thanks, >>>>>>>> Kaustav >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> ABHISHEK SAHNI >>>>>>> Mob : +91-990-701-5143 >>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> KAUSTAV MAJUMDER >>>>>> >>>>>> ASSOCIATE SOFTWARE ENGINEER >>>>>> >>>>>> Red Hat India PVT LTD. <https://www.redhat.com/> >>>>>> >>>>>> [email protected] M: 08981884037 IM: IRC: kmajumder >>>>>> <https://red.ht/sig> >>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>> @redhatway <https://twitter.com/redhatway> @redhatinc >>>>>> <https://instagram.com/redhatinc> @redhatsnaps >>>>>> <https://snapchat.com/add/redhatsnaps> >>>>>> >>>>> >>>> >>>> -- >>>> Thanks, >>>> >>>> Abhishek Sahni >>>> Computer Centre >>>> IISER Bhopal >>>> >>> >>> >>> -- >>> Thanks, >>> >>> Abhishek Sahni >>> Computer Centre >>> IISER Bhopal >>> >> >> Thanks, >> Kaustav >> > > > -- > > ABHISHEK SAHNI > Mob : +91-990-701-5143 > > >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/IZLW425VXW5ZJAW7APV7WFWCZOXPYTTS/

