Thanks a lot. :-) On Tue, Nov 27, 2018 at 4:22 PM Kaustav Majumder <kmaju...@redhat.com> wrote:
> > > On Tue, Nov 27, 2018 at 4:05 PM Abhishek Sahni <abhis...@iiserb.ac.in> > wrote: > >> That is amazing, resetting bricks resolved the issue. >> >> Thanks much Sahina and Kaustav. >> >> However, Do we have manual steps to recover those bricks. >> > https://gluster.readthedocs.io/en/latest/release-notes/3.9.0/ > >> >> On Tue, Nov 27, 2018 at 3:57 PM Abhishek Sahni <abhis...@iiserb.ac.in> >> wrote: >> >>> I just enabled it on default cluster and now the volumes are visible. It >>> seems like gluster service was disabled by default on cluster. >>> >>> On Tue, Nov 27, 2018 at 3:51 PM Sahina Bose <sab...@redhat.com> wrote: >>> >>>> >>>> >>>> On Tue, Nov 27, 2018 at 3:45 PM Kaustav Majumder <kmaju...@redhat.com> >>>> wrote: >>>> >>>>> I am not sure why ovirt is not showing any volume. >>>>> Sahina, is this a bug? >>>>> >>>> >>>> Check if gluster service is enabled on the cluster. >>>> The volumes are managed only if this is true >>>> >>>> >>>>> On Tue, Nov 27, 2018 at 3:10 PM Abhishek Sahni < >>>>> abhishek.sahni1...@gmail.com> wrote: >>>>> >>>>>> Hello Kaustav, >>>>>> >>>>>> That's weird, I never saw any volumes under the storage tab since >>>>>> installation. I am using HC setup deployed using cockpit console. >>>>>> >>>>>> https://imgur.com/a/nH9rzK8 >>>>>> >>>>>> Did I miss something? >>>>>> >>>>>> >>>>>> On Tue, Nov 27, 2018 at 2:50 PM Kaustav Majumder <kmaju...@redhat.com> >>>>>> wrote: >>>>>> >>>>>>> Click on volume for which you want to reset the brick-> under bricks >>>>>>> tab select the brick you wan to reset -> once you do you will see the >>>>>>> 'Reset Brick' option is active. >>>>>>> Attached is a screenshot -> https://i.imgur.com/QUMSrzt.png >>>>>>> >>>>>>> On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni < >>>>>>> abhishek.sahni1...@gmail.com> wrote: >>>>>>> >>>>>>>> Thanks Sahina for your response, I am not able to find it on UI, >>>>>>>> please help me to navigate? and yes I am using oVirt 4.2.6.4-1. >>>>>>>> >>>>>>>> On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <sab...@redhat.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni < >>>>>>>>> abhishek.sahni1...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Hello Team, >>>>>>>>>> >>>>>>>>>> We are running a setup of 3-way replica HC gluster setup >>>>>>>>>> configured during the initial deployment from the cockpit console >>>>>>>>>> using >>>>>>>>>> ansible. >>>>>>>>>> >>>>>>>>>> NODE1 >>>>>>>>>> - /dev/sda (OS) >>>>>>>>>> - /dev/sdb ( Gluster Bricks ) >>>>>>>>>> * /gluster_bricks/engine/engine/ >>>>>>>>>> * /gluster_bricks/data/data/ >>>>>>>>>> * /gluster_bricks/vmstore/vmstore/ >>>>>>>>>> >>>>>>>>>> NODE2 and NODE3 with a similar setup. >>>>>>>>>> >>>>>>>>>> There is a mishap that /dev/sdb on NODE2 totally got crashed and >>>>>>>>>> now there is nothing inside. However, I have created similar >>>>>>>>>> directories >>>>>>>>>> after mounting it back i.e., >>>>>>>>>> >>>>>>>>>> * /gluster_bricks/engine/engine/ >>>>>>>>>> * /gluster_bricks/data/data/ >>>>>>>>>> * /gluster_bricks/vmstore/vmstore/ >>>>>>>>>> but it is not yet recovered. >>>>>>>>>> >>>>>>>>>> ===================================================== >>>>>>>>>> [root@node2 ~]# gluster volume status >>>>>>>>>> Status of volume: data >>>>>>>>>> Gluster process TCP Port RDMA Port >>>>>>>>>> Online Pid >>>>>>>>>> >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> Brick *.*.*.1:/gluster_bricks/data/data 49152 0 Y >>>>>>>>>> 11111 >>>>>>>>>> Brick *.*.*.2:/gluster_bricks/data/data N/A N/A N >>>>>>>>>> N/A >>>>>>>>>> Brick *.*.*.3:/gluster_bricks/data/data 49152 0 Y >>>>>>>>>> 4303 >>>>>>>>>> Self-heal Daemon on localhost N/A N/A >>>>>>>>>> Y 23976 >>>>>>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>>>>>> 27838 >>>>>>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>>>>>> 27424 >>>>>>>>>> >>>>>>>>>> Task Status of Volume data >>>>>>>>>> >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> There are no active volume tasks >>>>>>>>>> >>>>>>>>>> Status of volume: engine >>>>>>>>>> Gluster process TCP Port RDMA Port >>>>>>>>>> Online Pid >>>>>>>>>> >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> Brick *.*.*.1:/gluster_bricks/engine/eng >>>>>>>>>> ine 49153 0 >>>>>>>>>> Y 11117 >>>>>>>>>> Brick *.*.*.2:/gluster_bricks/engine/eng >>>>>>>>>> ine N/A N/A >>>>>>>>>> N N/A >>>>>>>>>> Brick *.*.*.3:/gluster_bricks/engine/eng >>>>>>>>>> ine 49153 0 >>>>>>>>>> Y 4314 >>>>>>>>>> Self-heal Daemon on localhost N/A N/A >>>>>>>>>> Y 23976 >>>>>>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>>>>>> 27424 >>>>>>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>>>>>> 27838 >>>>>>>>>> >>>>>>>>>> Task Status of Volume engine >>>>>>>>>> >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> There are no active volume tasks >>>>>>>>>> >>>>>>>>>> Status of volume: vmstore >>>>>>>>>> Gluster process TCP Port RDMA Port >>>>>>>>>> Online Pid >>>>>>>>>> >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> Brick *.*.*.1:/gluster_bricks/vmstore/vm >>>>>>>>>> store 49154 0 >>>>>>>>>> Y 21603 >>>>>>>>>> Brick *.*.*.2:/gluster_bricks/vmstore/vm >>>>>>>>>> store N/A N/A >>>>>>>>>> N N/A >>>>>>>>>> Brick *.*.*.3:/gluster_bricks/vmstore/vm >>>>>>>>>> store 49154 0 >>>>>>>>>> Y 26845 >>>>>>>>>> Self-heal Daemon on localhost N/A N/A >>>>>>>>>> Y 23976 >>>>>>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>>>>>> 27424 >>>>>>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>>>>>> 27838 >>>>>>>>>> >>>>>>>>>> Task Status of Volume vmstore >>>>>>>>>> >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> There are no active volume tasks >>>>>>>>>> ============================================================= >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Can someone please suggest the steps to recover the setup? >>>>>>>>>> >>>>>>>>>> I have tried the below workaround but it doesn't help. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html >>>>>>>>>> >>>>>>>>> >>>>>>>>> You can reset the brick - if you're on oVirt 4.2.x, there's a UI >>>>>>>>> option in the bricks subtab to do this. >>>>>>>>> >>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> >>>>>>>>>> ABHISHEK SAHNI >>>>>>>>>> Mob : +91-990-701-5143 >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>>>>> oVirt Code of Conduct: >>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>>>> List Archives: >>>>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/ >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> >>>>>>>> ABHISHEK SAHNI >>>>>>>> Mob : +91-990-701-5143 >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>>> oVirt Code of Conduct: >>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>> List Archives: >>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/GEZZIGBWA6HQ55JV2PSP7VSSYPZQ2JAC/ >>>>>>>> >>>>>>> >>>>>>> >>>>>>> Thanks, >>>>>>> Kaustav >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> ABHISHEK SAHNI >>>>>> Mob : +91-990-701-5143 >>>>>> >>>>>> >>>>>> >>>>> >>>>> -- >>>>> >>>>> KAUSTAV MAJUMDER >>>>> >>>>> ASSOCIATE SOFTWARE ENGINEER >>>>> >>>>> Red Hat India PVT LTD. <https://www.redhat.com/> >>>>> >>>>> kmajum...@redhat.com M: 08981884037 IM: IRC: kmajumder >>>>> <https://red.ht/sig> >>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>> @redhatway <https://twitter.com/redhatway> @redhatinc >>>>> <https://instagram.com/redhatinc> @redhatsnaps >>>>> <https://snapchat.com/add/redhatsnaps> >>>>> >>>> >>> >>> -- >>> Thanks, >>> >>> Abhishek Sahni >>> Computer Centre >>> IISER Bhopal >>> >> >> >> -- >> Thanks, >> >> Abhishek Sahni >> Computer Centre >> IISER Bhopal >> > > Thanks, > Kaustav > -- ABHISHEK SAHNI Mob : +91-990-701-5143
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/UGXJ5LYVNVMGLOQ3JAE2QPU5E7TH67FH/