I just enabled it on default cluster and now the volumes are visible. It
seems like gluster service was disabled by default on cluster.

On Tue, Nov 27, 2018 at 3:51 PM Sahina Bose <[email protected]> wrote:

>
>
> On Tue, Nov 27, 2018 at 3:45 PM Kaustav Majumder <[email protected]>
> wrote:
>
>> I am not sure why ovirt is not showing any volume.
>> Sahina, is this a bug?
>>
>
> Check if gluster service is enabled on the cluster.
> The volumes are managed only if this is true
>
>
>> On Tue, Nov 27, 2018 at 3:10 PM Abhishek Sahni <
>> [email protected]> wrote:
>>
>>> Hello Kaustav,
>>>
>>> That's weird, I never saw any volumes under the storage tab since
>>> installation. I am using HC setup deployed using cockpit console.
>>>
>>> https://imgur.com/a/nH9rzK8
>>>
>>> Did I miss something?
>>>
>>>
>>> On Tue, Nov 27, 2018 at 2:50 PM Kaustav Majumder <[email protected]>
>>> wrote:
>>>
>>>> Click on volume for which you want to reset the brick-> under bricks
>>>> tab select the brick you wan to reset -> once you do you will see the
>>>> 'Reset Brick' option is active.
>>>> Attached is a screenshot -> https://i.imgur.com/QUMSrzt.png
>>>>
>>>> On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni <
>>>> [email protected]> wrote:
>>>>
>>>>> Thanks Sahina for your response, I am not able to find it on UI,
>>>>> please help me to navigate? and yes I am using oVirt 4.2.6.4-1.
>>>>>
>>>>> On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <[email protected]>
>>>>> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni <
>>>>>> [email protected]> wrote:
>>>>>>
>>>>>>> Hello Team,
>>>>>>>
>>>>>>> We are running a setup of 3-way replica HC gluster setup configured
>>>>>>> during the initial deployment from the cockpit console using ansible.
>>>>>>>
>>>>>>> NODE1
>>>>>>>   - /dev/sda   (OS)
>>>>>>>   - /dev/sdb   ( Gluster Bricks )
>>>>>>>        * /gluster_bricks/engine/engine/
>>>>>>>        * /gluster_bricks/data/data/
>>>>>>>        * /gluster_bricks/vmstore/vmstore/
>>>>>>>
>>>>>>> NODE2 and NODE3 with a similar setup.
>>>>>>>
>>>>>>> There is a mishap that /dev/sdb on NODE2 totally got crashed and now
>>>>>>> there is nothing inside. However, I have created similar directories 
>>>>>>> after
>>>>>>> mounting it back i.e.,
>>>>>>>
>>>>>>>        * /gluster_bricks/engine/engine/
>>>>>>>        * /gluster_bricks/data/data/
>>>>>>>        * /gluster_bricks/vmstore/vmstore/
>>>>>>> but it is not yet recovered.
>>>>>>>
>>>>>>> =====================================================
>>>>>>> [root@node2 ~]# gluster volume status
>>>>>>> Status of volume: data
>>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>>> Online  Pid
>>>>>>>
>>>>>>> ------------------------------------------------------------------------------
>>>>>>> Brick *.*.*.1:/gluster_bricks/data/data  49152     0          Y
>>>>>>>  11111
>>>>>>> Brick *.*.*.2:/gluster_bricks/data/data  N/A       N/A        N
>>>>>>>  N/A
>>>>>>> Brick *.*.*.3:/gluster_bricks/data/data  49152     0          Y
>>>>>>>  4303
>>>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>>>      23976
>>>>>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>>>>>  27838
>>>>>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>>>>>  27424
>>>>>>>
>>>>>>> Task Status of Volume data
>>>>>>>
>>>>>>> ------------------------------------------------------------------------------
>>>>>>> There are no active volume tasks
>>>>>>>
>>>>>>> Status of volume: engine
>>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>>> Online  Pid
>>>>>>>
>>>>>>> ------------------------------------------------------------------------------
>>>>>>> Brick *.*.*.1:/gluster_bricks/engine/eng
>>>>>>> ine                                         49153     0          Y
>>>>>>>      11117
>>>>>>> Brick *.*.*.2:/gluster_bricks/engine/eng
>>>>>>> ine                                         N/A       N/A        N
>>>>>>>      N/A
>>>>>>> Brick *.*.*.3:/gluster_bricks/engine/eng
>>>>>>> ine                                         49153     0          Y
>>>>>>>      4314
>>>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>>>      23976
>>>>>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>>>>>  27424
>>>>>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>>>>>  27838
>>>>>>>
>>>>>>> Task Status of Volume engine
>>>>>>>
>>>>>>> ------------------------------------------------------------------------------
>>>>>>> There are no active volume tasks
>>>>>>>
>>>>>>> Status of volume: vmstore
>>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>>> Online  Pid
>>>>>>>
>>>>>>> ------------------------------------------------------------------------------
>>>>>>> Brick *.*.*.1:/gluster_bricks/vmstore/vm
>>>>>>> store                                       49154     0          Y
>>>>>>>      21603
>>>>>>> Brick *.*.*.2:/gluster_bricks/vmstore/vm
>>>>>>> store                                       N/A       N/A        N
>>>>>>>      N/A
>>>>>>> Brick *.*.*.3:/gluster_bricks/vmstore/vm
>>>>>>> store                                       49154     0          Y
>>>>>>>      26845
>>>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>>>      23976
>>>>>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>>>>>  27424
>>>>>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>>>>>  27838
>>>>>>>
>>>>>>> Task Status of Volume vmstore
>>>>>>>
>>>>>>> ------------------------------------------------------------------------------
>>>>>>> There are no active volume tasks
>>>>>>> =============================================================
>>>>>>>
>>>>>>>
>>>>>>> Can someone please suggest the steps to recover the setup?
>>>>>>>
>>>>>>> I have tried the below workaround but it doesn't help.
>>>>>>>
>>>>>>>
>>>>>>> https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html
>>>>>>>
>>>>>>
>>>>>>  You can reset the brick - if you're on oVirt 4.2.x, there's a UI
>>>>>> option in the bricks subtab to do this.
>>>>>>
>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>>
>>>>>>> ABHISHEK SAHNI
>>>>>>> Mob : +91-990-701-5143
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Users mailing list -- [email protected]
>>>>>>> To unsubscribe send an email to [email protected]
>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>>> oVirt Code of Conduct:
>>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>>> List Archives:
>>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/
>>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>>
>>>>> ABHISHEK SAHNI
>>>>> Mob : +91-990-701-5143
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list -- [email protected]
>>>>> To unsubscribe send an email to [email protected]
>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>> oVirt Code of Conduct:
>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>> List Archives:
>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/GEZZIGBWA6HQ55JV2PSP7VSSYPZQ2JAC/
>>>>>
>>>>
>>>>
>>>> Thanks,
>>>> Kaustav
>>>>
>>>
>>>
>>> --
>>>
>>> ABHISHEK SAHNI
>>> Mob : +91-990-701-5143
>>>
>>>
>>>
>>
>> --
>>
>> KAUSTAV MAJUMDER
>>
>> ASSOCIATE SOFTWARE ENGINEER
>>
>> Red Hat India PVT LTD. <https://www.redhat.com/>
>>
>> [email protected]    M: 08981884037     IM: IRC: kmajumder
>> <https://red.ht/sig>
>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
>> @redhatway <https://twitter.com/redhatway>   @redhatinc
>> <https://instagram.com/redhatinc>   @redhatsnaps
>> <https://snapchat.com/add/redhatsnaps>
>>
>

-- 
Thanks,

Abhishek Sahni
Computer Centre
IISER Bhopal
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/EOW22SMUYO5O2LJ2IMDSC4X6NAUJPXVX/

Reply via email to