On Tue, Nov 27, 2018 at 3:45 PM Kaustav Majumder <[email protected]>
wrote:

> I am not sure why ovirt is not showing any volume.
> Sahina, is this a bug?
>

Check if gluster service is enabled on the cluster.
The volumes are managed only if this is true


> On Tue, Nov 27, 2018 at 3:10 PM Abhishek Sahni <
> [email protected]> wrote:
>
>> Hello Kaustav,
>>
>> That's weird, I never saw any volumes under the storage tab since
>> installation. I am using HC setup deployed using cockpit console.
>>
>> https://imgur.com/a/nH9rzK8
>>
>> Did I miss something?
>>
>>
>> On Tue, Nov 27, 2018 at 2:50 PM Kaustav Majumder <[email protected]>
>> wrote:
>>
>>> Click on volume for which you want to reset the brick-> under bricks tab
>>> select the brick you wan to reset -> once you do you will see the 'Reset
>>> Brick' option is active.
>>> Attached is a screenshot -> https://i.imgur.com/QUMSrzt.png
>>>
>>> On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni <
>>> [email protected]> wrote:
>>>
>>>> Thanks Sahina for your response, I am not able to find it on UI, please
>>>> help me to navigate? and yes I am using oVirt 4.2.6.4-1.
>>>>
>>>> On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <[email protected]> wrote:
>>>>
>>>>>
>>>>>
>>>>> On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni <
>>>>> [email protected]> wrote:
>>>>>
>>>>>> Hello Team,
>>>>>>
>>>>>> We are running a setup of 3-way replica HC gluster setup configured
>>>>>> during the initial deployment from the cockpit console using ansible.
>>>>>>
>>>>>> NODE1
>>>>>>   - /dev/sda   (OS)
>>>>>>   - /dev/sdb   ( Gluster Bricks )
>>>>>>        * /gluster_bricks/engine/engine/
>>>>>>        * /gluster_bricks/data/data/
>>>>>>        * /gluster_bricks/vmstore/vmstore/
>>>>>>
>>>>>> NODE2 and NODE3 with a similar setup.
>>>>>>
>>>>>> There is a mishap that /dev/sdb on NODE2 totally got crashed and now
>>>>>> there is nothing inside. However, I have created similar directories 
>>>>>> after
>>>>>> mounting it back i.e.,
>>>>>>
>>>>>>        * /gluster_bricks/engine/engine/
>>>>>>        * /gluster_bricks/data/data/
>>>>>>        * /gluster_bricks/vmstore/vmstore/
>>>>>> but it is not yet recovered.
>>>>>>
>>>>>> =====================================================
>>>>>> [root@node2 ~]# gluster volume status
>>>>>> Status of volume: data
>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>> Online  Pid
>>>>>>
>>>>>> ------------------------------------------------------------------------------
>>>>>> Brick *.*.*.1:/gluster_bricks/data/data  49152     0          Y
>>>>>>  11111
>>>>>> Brick *.*.*.2:/gluster_bricks/data/data  N/A       N/A        N
>>>>>>  N/A
>>>>>> Brick *.*.*.3:/gluster_bricks/data/data  49152     0          Y
>>>>>>  4303
>>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>>    23976
>>>>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>>>>  27838
>>>>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>>>>  27424
>>>>>>
>>>>>> Task Status of Volume data
>>>>>>
>>>>>> ------------------------------------------------------------------------------
>>>>>> There are no active volume tasks
>>>>>>
>>>>>> Status of volume: engine
>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>> Online  Pid
>>>>>>
>>>>>> ------------------------------------------------------------------------------
>>>>>> Brick *.*.*.1:/gluster_bricks/engine/eng
>>>>>> ine                                         49153     0          Y
>>>>>>    11117
>>>>>> Brick *.*.*.2:/gluster_bricks/engine/eng
>>>>>> ine                                         N/A       N/A        N
>>>>>>    N/A
>>>>>> Brick *.*.*.3:/gluster_bricks/engine/eng
>>>>>> ine                                         49153     0          Y
>>>>>>    4314
>>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>>    23976
>>>>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>>>>  27424
>>>>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>>>>  27838
>>>>>>
>>>>>> Task Status of Volume engine
>>>>>>
>>>>>> ------------------------------------------------------------------------------
>>>>>> There are no active volume tasks
>>>>>>
>>>>>> Status of volume: vmstore
>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>> Online  Pid
>>>>>>
>>>>>> ------------------------------------------------------------------------------
>>>>>> Brick *.*.*.1:/gluster_bricks/vmstore/vm
>>>>>> store                                       49154     0          Y
>>>>>>    21603
>>>>>> Brick *.*.*.2:/gluster_bricks/vmstore/vm
>>>>>> store                                       N/A       N/A        N
>>>>>>    N/A
>>>>>> Brick *.*.*.3:/gluster_bricks/vmstore/vm
>>>>>> store                                       49154     0          Y
>>>>>>    26845
>>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>>    23976
>>>>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>>>>  27424
>>>>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>>>>  27838
>>>>>>
>>>>>> Task Status of Volume vmstore
>>>>>>
>>>>>> ------------------------------------------------------------------------------
>>>>>> There are no active volume tasks
>>>>>> =============================================================
>>>>>>
>>>>>>
>>>>>> Can someone please suggest the steps to recover the setup?
>>>>>>
>>>>>> I have tried the below workaround but it doesn't help.
>>>>>>
>>>>>>
>>>>>> https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html
>>>>>>
>>>>>
>>>>>  You can reset the brick - if you're on oVirt 4.2.x, there's a UI
>>>>> option in the bricks subtab to do this.
>>>>>
>>>>>
>>>>>>
>>>>>> --
>>>>>>
>>>>>> ABHISHEK SAHNI
>>>>>> Mob : +91-990-701-5143
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> Users mailing list -- [email protected]
>>>>>> To unsubscribe send an email to [email protected]
>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>> oVirt Code of Conduct:
>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>> List Archives:
>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/
>>>>>>
>>>>>
>>>>
>>>> --
>>>>
>>>> ABHISHEK SAHNI
>>>> Mob : +91-990-701-5143
>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list -- [email protected]
>>>> To unsubscribe send an email to [email protected]
>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>> oVirt Code of Conduct:
>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>> List Archives:
>>>> https://lists.ovirt.org/archives/list/[email protected]/message/GEZZIGBWA6HQ55JV2PSP7VSSYPZQ2JAC/
>>>>
>>>
>>>
>>> Thanks,
>>> Kaustav
>>>
>>
>>
>> --
>>
>> ABHISHEK SAHNI
>> Mob : +91-990-701-5143
>>
>>
>>
>
> --
>
> KAUSTAV MAJUMDER
>
> ASSOCIATE SOFTWARE ENGINEER
>
> Red Hat India PVT LTD. <https://www.redhat.com/>
>
> [email protected]    M: 08981884037     IM: IRC: kmajumder
> <https://red.ht/sig>
> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
> @redhatway <https://twitter.com/redhatway>   @redhatinc
> <https://instagram.com/redhatinc>   @redhatsnaps
> <https://snapchat.com/add/redhatsnaps>
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/QFSCKVUB52TQWKXP44C76AC5CY5SUT6H/

Reply via email to