I just enabled it on default cluster and now the volumes are visible. It seems like gluster service was disabled by default on cluster.
On Tue, Nov 27, 2018 at 3:51 PM Sahina Bose <[email protected]> wrote: > > > On Tue, Nov 27, 2018 at 3:45 PM Kaustav Majumder <[email protected]> > wrote: > >> I am not sure why ovirt is not showing any volume. >> Sahina, is this a bug? >> > > Check if gluster service is enabled on the cluster. > The volumes are managed only if this is true > > >> On Tue, Nov 27, 2018 at 3:10 PM Abhishek Sahni < >> [email protected]> wrote: >> >>> Hello Kaustav, >>> >>> That's weird, I never saw any volumes under the storage tab since >>> installation. I am using HC setup deployed using cockpit console. >>> >>> https://imgur.com/a/nH9rzK8 >>> >>> Did I miss something? >>> >>> >>> On Tue, Nov 27, 2018 at 2:50 PM Kaustav Majumder <[email protected]> >>> wrote: >>> >>>> Click on volume for which you want to reset the brick-> under bricks >>>> tab select the brick you wan to reset -> once you do you will see the >>>> 'Reset Brick' option is active. >>>> Attached is a screenshot -> https://i.imgur.com/QUMSrzt.png >>>> >>>> On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni < >>>> [email protected]> wrote: >>>> >>>>> Thanks Sahina for your response, I am not able to find it on UI, >>>>> please help me to navigate? and yes I am using oVirt 4.2.6.4-1. >>>>> >>>>> On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <[email protected]> >>>>> wrote: >>>>> >>>>>> >>>>>> >>>>>> On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> Hello Team, >>>>>>> >>>>>>> We are running a setup of 3-way replica HC gluster setup configured >>>>>>> during the initial deployment from the cockpit console using ansible. >>>>>>> >>>>>>> NODE1 >>>>>>> - /dev/sda (OS) >>>>>>> - /dev/sdb ( Gluster Bricks ) >>>>>>> * /gluster_bricks/engine/engine/ >>>>>>> * /gluster_bricks/data/data/ >>>>>>> * /gluster_bricks/vmstore/vmstore/ >>>>>>> >>>>>>> NODE2 and NODE3 with a similar setup. >>>>>>> >>>>>>> There is a mishap that /dev/sdb on NODE2 totally got crashed and now >>>>>>> there is nothing inside. However, I have created similar directories >>>>>>> after >>>>>>> mounting it back i.e., >>>>>>> >>>>>>> * /gluster_bricks/engine/engine/ >>>>>>> * /gluster_bricks/data/data/ >>>>>>> * /gluster_bricks/vmstore/vmstore/ >>>>>>> but it is not yet recovered. >>>>>>> >>>>>>> ===================================================== >>>>>>> [root@node2 ~]# gluster volume status >>>>>>> Status of volume: data >>>>>>> Gluster process TCP Port RDMA Port >>>>>>> Online Pid >>>>>>> >>>>>>> ------------------------------------------------------------------------------ >>>>>>> Brick *.*.*.1:/gluster_bricks/data/data 49152 0 Y >>>>>>> 11111 >>>>>>> Brick *.*.*.2:/gluster_bricks/data/data N/A N/A N >>>>>>> N/A >>>>>>> Brick *.*.*.3:/gluster_bricks/data/data 49152 0 Y >>>>>>> 4303 >>>>>>> Self-heal Daemon on localhost N/A N/A Y >>>>>>> 23976 >>>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>>> 27838 >>>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>>> 27424 >>>>>>> >>>>>>> Task Status of Volume data >>>>>>> >>>>>>> ------------------------------------------------------------------------------ >>>>>>> There are no active volume tasks >>>>>>> >>>>>>> Status of volume: engine >>>>>>> Gluster process TCP Port RDMA Port >>>>>>> Online Pid >>>>>>> >>>>>>> ------------------------------------------------------------------------------ >>>>>>> Brick *.*.*.1:/gluster_bricks/engine/eng >>>>>>> ine 49153 0 Y >>>>>>> 11117 >>>>>>> Brick *.*.*.2:/gluster_bricks/engine/eng >>>>>>> ine N/A N/A N >>>>>>> N/A >>>>>>> Brick *.*.*.3:/gluster_bricks/engine/eng >>>>>>> ine 49153 0 Y >>>>>>> 4314 >>>>>>> Self-heal Daemon on localhost N/A N/A Y >>>>>>> 23976 >>>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>>> 27424 >>>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>>> 27838 >>>>>>> >>>>>>> Task Status of Volume engine >>>>>>> >>>>>>> ------------------------------------------------------------------------------ >>>>>>> There are no active volume tasks >>>>>>> >>>>>>> Status of volume: vmstore >>>>>>> Gluster process TCP Port RDMA Port >>>>>>> Online Pid >>>>>>> >>>>>>> ------------------------------------------------------------------------------ >>>>>>> Brick *.*.*.1:/gluster_bricks/vmstore/vm >>>>>>> store 49154 0 Y >>>>>>> 21603 >>>>>>> Brick *.*.*.2:/gluster_bricks/vmstore/vm >>>>>>> store N/A N/A N >>>>>>> N/A >>>>>>> Brick *.*.*.3:/gluster_bricks/vmstore/vm >>>>>>> store 49154 0 Y >>>>>>> 26845 >>>>>>> Self-heal Daemon on localhost N/A N/A Y >>>>>>> 23976 >>>>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y >>>>>>> 27424 >>>>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y >>>>>>> 27838 >>>>>>> >>>>>>> Task Status of Volume vmstore >>>>>>> >>>>>>> ------------------------------------------------------------------------------ >>>>>>> There are no active volume tasks >>>>>>> ============================================================= >>>>>>> >>>>>>> >>>>>>> Can someone please suggest the steps to recover the setup? >>>>>>> >>>>>>> I have tried the below workaround but it doesn't help. >>>>>>> >>>>>>> >>>>>>> https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html >>>>>>> >>>>>> >>>>>> You can reset the brick - if you're on oVirt 4.2.x, there's a UI >>>>>> option in the bricks subtab to do this. >>>>>> >>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> ABHISHEK SAHNI >>>>>>> Mob : +91-990-701-5143 >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Users mailing list -- [email protected] >>>>>>> To unsubscribe send an email to [email protected] >>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>> oVirt Code of Conduct: >>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>> List Archives: >>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/ >>>>>>> >>>>>> >>>>> >>>>> -- >>>>> >>>>> ABHISHEK SAHNI >>>>> Mob : +91-990-701-5143 >>>>> >>>>> >>>>> _______________________________________________ >>>>> Users mailing list -- [email protected] >>>>> To unsubscribe send an email to [email protected] >>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>> oVirt Code of Conduct: >>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>> List Archives: >>>>> https://lists.ovirt.org/archives/list/[email protected]/message/GEZZIGBWA6HQ55JV2PSP7VSSYPZQ2JAC/ >>>>> >>>> >>>> >>>> Thanks, >>>> Kaustav >>>> >>> >>> >>> -- >>> >>> ABHISHEK SAHNI >>> Mob : +91-990-701-5143 >>> >>> >>> >> >> -- >> >> KAUSTAV MAJUMDER >> >> ASSOCIATE SOFTWARE ENGINEER >> >> Red Hat India PVT LTD. <https://www.redhat.com/> >> >> [email protected] M: 08981884037 IM: IRC: kmajumder >> <https://red.ht/sig> >> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >> @redhatway <https://twitter.com/redhatway> @redhatinc >> <https://instagram.com/redhatinc> @redhatsnaps >> <https://snapchat.com/add/redhatsnaps> >> > -- Thanks, Abhishek Sahni Computer Centre IISER Bhopal
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/EOW22SMUYO5O2LJ2IMDSC4X6NAUJPXVX/

