On Tue, Nov 27, 2018 at 3:45 PM Kaustav Majumder <[email protected]> wrote:
> I am not sure why ovirt is not showing any volume. > Sahina, is this a bug? > Check if gluster service is enabled on the cluster. The volumes are managed only if this is true > On Tue, Nov 27, 2018 at 3:10 PM Abhishek Sahni < > [email protected]> wrote: > >> Hello Kaustav, >> >> That's weird, I never saw any volumes under the storage tab since >> installation. I am using HC setup deployed using cockpit console. >> >> https://imgur.com/a/nH9rzK8 >> >> Did I miss something? >> >> >> On Tue, Nov 27, 2018 at 2:50 PM Kaustav Majumder <[email protected]> >> wrote: >> >>> Click on volume for which you want to reset the brick-> under bricks tab >>> select the brick you wan to reset -> once you do you will see the 'Reset >>> Brick' option is active. >>> Attached is a screenshot -> https://i.imgur.com/QUMSrzt.png >>> >>> On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni < >>> [email protected]> wrote: >>> >>>> Thanks Sahina for your response, I am not able to find it on UI, please >>>> help me to navigate? and yes I am using oVirt 4.2.6.4-1. >>>> >>>> On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <[email protected]> wrote: >>>> >>>>> >>>>> >>>>> On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni < >>>>> [email protected]> wrote: >>>>> >>>>>> Hello Team, >>>>>> >>>>>> We are running a setup of 3-way replica HC gluster setup configured >>>>>> during the initial deployment from the cockpit console using ansible. >>>>>> >>>>>> NODE1 >>>>>> - /dev/sda (OS) >>>>>> - /dev/sdb ( Gluster Bricks ) >>>>>> * /gluster_bricks/engine/engine/ >>>>>> * /gluster_bricks/data/data/ >>>>>> * /gluster_bricks/vmstore/vmstore/ >>>>>> >>>>>> NODE2 and NODE3 with a similar setup. >>>>>> >>>>>> There is a mishap that /dev/sdb on NODE2 totally got crashed and now >>>>>> there is nothing inside. However, I have created similar directories >>>>>> after >>>>>> mounting it back i.e., >>>>>> >>>>>> * /gluster_bricks/engine/engine/ >>>>>> * /gluster_bricks/data/data/ >>>>>> * /gluster_bricks/vmstore/vmstore/ >>>>>> but it is not yet recovered. >>>>>> >>>>>> ===================================================== >>>>>> [root@node2 ~]# gluster volume status >>>>>> Status of volume: data >>>>>> Gluster process TCP Port RDMA Port >>>>>> Online Pid >>>>>> >>>>>> ------------------------------------------------------------------------------ >>>>>> Brick *.*.*.1:/gluster_bricks/data/data 49152 0 Y >>>>>> 11111 >>>>>> Brick *.*.*.2:/gluster_bricks/data/data N/A N/A N >>>>>> N/A >>>>>> Brick *.*.*.3:/gluster_bricks/data/data 49152 0 Y >>>>>> 4303 >>>>>> Self-heal Daemon on localhost N/A N/A Y >>>>>> 23976 >>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>> 27838 >>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>> 27424 >>>>>> >>>>>> Task Status of Volume data >>>>>> >>>>>> ------------------------------------------------------------------------------ >>>>>> There are no active volume tasks >>>>>> >>>>>> Status of volume: engine >>>>>> Gluster process TCP Port RDMA Port >>>>>> Online Pid >>>>>> >>>>>> ------------------------------------------------------------------------------ >>>>>> Brick *.*.*.1:/gluster_bricks/engine/eng >>>>>> ine 49153 0 Y >>>>>> 11117 >>>>>> Brick *.*.*.2:/gluster_bricks/engine/eng >>>>>> ine N/A N/A N >>>>>> N/A >>>>>> Brick *.*.*.3:/gluster_bricks/engine/eng >>>>>> ine 49153 0 Y >>>>>> 4314 >>>>>> Self-heal Daemon on localhost N/A N/A Y >>>>>> 23976 >>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>> 27424 >>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>> 27838 >>>>>> >>>>>> Task Status of Volume engine >>>>>> >>>>>> ------------------------------------------------------------------------------ >>>>>> There are no active volume tasks >>>>>> >>>>>> Status of volume: vmstore >>>>>> Gluster process TCP Port RDMA Port >>>>>> Online Pid >>>>>> >>>>>> ------------------------------------------------------------------------------ >>>>>> Brick *.*.*.1:/gluster_bricks/vmstore/vm >>>>>> store 49154 0 Y >>>>>> 21603 >>>>>> Brick *.*.*.2:/gluster_bricks/vmstore/vm >>>>>> store N/A N/A N >>>>>> N/A >>>>>> Brick *.*.*.3:/gluster_bricks/vmstore/vm >>>>>> store 49154 0 Y >>>>>> 26845 >>>>>> Self-heal Daemon on localhost N/A N/A Y >>>>>> 23976 >>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>> 27424 >>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>> 27838 >>>>>> >>>>>> Task Status of Volume vmstore >>>>>> >>>>>> ------------------------------------------------------------------------------ >>>>>> There are no active volume tasks >>>>>> ============================================================= >>>>>> >>>>>> >>>>>> Can someone please suggest the steps to recover the setup? >>>>>> >>>>>> I have tried the below workaround but it doesn't help. >>>>>> >>>>>> >>>>>> https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html >>>>>> >>>>> >>>>> You can reset the brick - if you're on oVirt 4.2.x, there's a UI >>>>> option in the bricks subtab to do this. >>>>> >>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> ABHISHEK SAHNI >>>>>> Mob : +91-990-701-5143 >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Users mailing list -- [email protected] >>>>>> To unsubscribe send an email to [email protected] >>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>> oVirt Code of Conduct: >>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>> List Archives: >>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/ >>>>>> >>>>> >>>> >>>> -- >>>> >>>> ABHISHEK SAHNI >>>> Mob : +91-990-701-5143 >>>> >>>> >>>> _______________________________________________ >>>> Users mailing list -- [email protected] >>>> To unsubscribe send an email to [email protected] >>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>> oVirt Code of Conduct: >>>> https://www.ovirt.org/community/about/community-guidelines/ >>>> List Archives: >>>> https://lists.ovirt.org/archives/list/[email protected]/message/GEZZIGBWA6HQ55JV2PSP7VSSYPZQ2JAC/ >>>> >>> >>> >>> Thanks, >>> Kaustav >>> >> >> >> -- >> >> ABHISHEK SAHNI >> Mob : +91-990-701-5143 >> >> >> > > -- > > KAUSTAV MAJUMDER > > ASSOCIATE SOFTWARE ENGINEER > > Red Hat India PVT LTD. <https://www.redhat.com/> > > [email protected] M: 08981884037 IM: IRC: kmajumder > <https://red.ht/sig> > TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> > @redhatway <https://twitter.com/redhatway> @redhatinc > <https://instagram.com/redhatinc> @redhatsnaps > <https://snapchat.com/add/redhatsnaps> >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/QFSCKVUB52TQWKXP44C76AC5CY5SUT6H/

