You are welcome.

On Tue, Nov 27, 2018 at 4:24 PM Abhishek Sahni <[email protected]>
wrote:

> Thanks a lot. :-)
>
> On Tue, Nov 27, 2018 at 4:22 PM Kaustav Majumder <[email protected]>
> wrote:
>
>>
>>
>> On Tue, Nov 27, 2018 at 4:05 PM Abhishek Sahni <[email protected]>
>> wrote:
>>
>>> That is amazing, resetting bricks resolved the issue.
>>>
>>> Thanks much Sahina and Kaustav.
>>>
>>> However, Do we have manual steps to recover those bricks.
>>>
>> https://gluster.readthedocs.io/en/latest/release-notes/3.9.0/
>>
>>>
>>> On Tue, Nov 27, 2018 at 3:57 PM Abhishek Sahni <[email protected]>
>>> wrote:
>>>
>>>> I just enabled it on default cluster and now the volumes are visible.
>>>> It seems like gluster service was disabled by default on cluster.
>>>>
>>>> On Tue, Nov 27, 2018 at 3:51 PM Sahina Bose <[email protected]> wrote:
>>>>
>>>>>
>>>>>
>>>>> On Tue, Nov 27, 2018 at 3:45 PM Kaustav Majumder <[email protected]>
>>>>> wrote:
>>>>>
>>>>>> I am not sure why ovirt is not showing any volume.
>>>>>> Sahina, is this a bug?
>>>>>>
>>>>>
>>>>> Check if gluster service is enabled on the cluster.
>>>>> The volumes are managed only if this is true
>>>>>
>>>>>
>>>>>> On Tue, Nov 27, 2018 at 3:10 PM Abhishek Sahni <
>>>>>> [email protected]> wrote:
>>>>>>
>>>>>>> Hello Kaustav,
>>>>>>>
>>>>>>> That's weird, I never saw any volumes under the storage tab since
>>>>>>> installation. I am using HC setup deployed using cockpit console.
>>>>>>>
>>>>>>> https://imgur.com/a/nH9rzK8
>>>>>>>
>>>>>>> Did I miss something?
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Nov 27, 2018 at 2:50 PM Kaustav Majumder <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>> Click on volume for which you want to reset the brick-> under
>>>>>>>> bricks tab select the brick you wan to reset -> once you do you will 
>>>>>>>> see
>>>>>>>> the 'Reset Brick' option is active.
>>>>>>>> Attached is a screenshot -> https://i.imgur.com/QUMSrzt.png
>>>>>>>>
>>>>>>>> On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni <
>>>>>>>> [email protected]> wrote:
>>>>>>>>
>>>>>>>>> Thanks Sahina for your response, I am not able to find it on UI,
>>>>>>>>> please help me to navigate? and yes I am using oVirt 4.2.6.4-1.
>>>>>>>>>
>>>>>>>>> On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <[email protected]>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni <
>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hello Team,
>>>>>>>>>>>
>>>>>>>>>>> We are running a setup of 3-way replica HC gluster setup
>>>>>>>>>>> configured during the initial deployment from the cockpit console 
>>>>>>>>>>> using
>>>>>>>>>>> ansible.
>>>>>>>>>>>
>>>>>>>>>>> NODE1
>>>>>>>>>>>   - /dev/sda   (OS)
>>>>>>>>>>>   - /dev/sdb   ( Gluster Bricks )
>>>>>>>>>>>        * /gluster_bricks/engine/engine/
>>>>>>>>>>>        * /gluster_bricks/data/data/
>>>>>>>>>>>        * /gluster_bricks/vmstore/vmstore/
>>>>>>>>>>>
>>>>>>>>>>> NODE2 and NODE3 with a similar setup.
>>>>>>>>>>>
>>>>>>>>>>> There is a mishap that /dev/sdb on NODE2 totally got crashed and
>>>>>>>>>>> now there is nothing inside. However, I have created similar 
>>>>>>>>>>> directories
>>>>>>>>>>> after mounting it back i.e.,
>>>>>>>>>>>
>>>>>>>>>>>        * /gluster_bricks/engine/engine/
>>>>>>>>>>>        * /gluster_bricks/data/data/
>>>>>>>>>>>        * /gluster_bricks/vmstore/vmstore/
>>>>>>>>>>> but it is not yet recovered.
>>>>>>>>>>>
>>>>>>>>>>> =====================================================
>>>>>>>>>>> [root@node2 ~]# gluster volume status
>>>>>>>>>>> Status of volume: data
>>>>>>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>>>>>>> Online  Pid
>>>>>>>>>>>
>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>> Brick *.*.*.1:/gluster_bricks/data/data  49152     0          Y
>>>>>>>>>>>      11111
>>>>>>>>>>> Brick *.*.*.2:/gluster_bricks/data/data  N/A       N/A        N
>>>>>>>>>>>      N/A
>>>>>>>>>>> Brick *.*.*.3:/gluster_bricks/data/data  49152     0          Y
>>>>>>>>>>>      4303
>>>>>>>>>>> Self-heal Daemon on localhost               N/A       N/A
>>>>>>>>>>> Y       23976
>>>>>>>>>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>>>>>>>>>      27838
>>>>>>>>>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>>>>>>>>>      27424
>>>>>>>>>>>
>>>>>>>>>>> Task Status of Volume data
>>>>>>>>>>>
>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>> There are no active volume tasks
>>>>>>>>>>>
>>>>>>>>>>> Status of volume: engine
>>>>>>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>>>>>>> Online  Pid
>>>>>>>>>>>
>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>> Brick *.*.*.1:/gluster_bricks/engine/eng
>>>>>>>>>>> ine                                         49153     0
>>>>>>>>>>> Y       11117
>>>>>>>>>>> Brick *.*.*.2:/gluster_bricks/engine/eng
>>>>>>>>>>> ine                                         N/A       N/A
>>>>>>>>>>> N       N/A
>>>>>>>>>>> Brick *.*.*.3:/gluster_bricks/engine/eng
>>>>>>>>>>> ine                                         49153     0
>>>>>>>>>>> Y       4314
>>>>>>>>>>> Self-heal Daemon on localhost               N/A       N/A
>>>>>>>>>>> Y       23976
>>>>>>>>>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>>>>>>>>>      27424
>>>>>>>>>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>>>>>>>>>      27838
>>>>>>>>>>>
>>>>>>>>>>> Task Status of Volume engine
>>>>>>>>>>>
>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>> There are no active volume tasks
>>>>>>>>>>>
>>>>>>>>>>> Status of volume: vmstore
>>>>>>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>>>>>>> Online  Pid
>>>>>>>>>>>
>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>> Brick *.*.*.1:/gluster_bricks/vmstore/vm
>>>>>>>>>>> store                                       49154     0
>>>>>>>>>>> Y       21603
>>>>>>>>>>> Brick *.*.*.2:/gluster_bricks/vmstore/vm
>>>>>>>>>>> store                                       N/A       N/A
>>>>>>>>>>> N       N/A
>>>>>>>>>>> Brick *.*.*.3:/gluster_bricks/vmstore/vm
>>>>>>>>>>> store                                       49154     0
>>>>>>>>>>> Y       26845
>>>>>>>>>>> Self-heal Daemon on localhost               N/A       N/A
>>>>>>>>>>> Y       23976
>>>>>>>>>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>>>>>>>>>      27424
>>>>>>>>>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>>>>>>>>>      27838
>>>>>>>>>>>
>>>>>>>>>>> Task Status of Volume vmstore
>>>>>>>>>>>
>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>> There are no active volume tasks
>>>>>>>>>>> =============================================================
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Can someone please suggest the steps to recover the setup?
>>>>>>>>>>>
>>>>>>>>>>> I have tried the below workaround but it doesn't help.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>  You can reset the brick - if you're on oVirt 4.2.x, there's a UI
>>>>>>>>>> option in the bricks subtab to do this.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>>
>>>>>>>>>>> ABHISHEK SAHNI
>>>>>>>>>>> Mob : +91-990-701-5143
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> _______________________________________________
>>>>>>>>>>> Users mailing list -- [email protected]
>>>>>>>>>>> To unsubscribe send an email to [email protected]
>>>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>>>>>>> oVirt Code of Conduct:
>>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>>>>> List Archives:
>>>>>>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>>
>>>>>>>>> ABHISHEK SAHNI
>>>>>>>>> Mob : +91-990-701-5143
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> Users mailing list -- [email protected]
>>>>>>>>> To unsubscribe send an email to [email protected]
>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>>>>> oVirt Code of Conduct:
>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>>> List Archives:
>>>>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/GEZZIGBWA6HQ55JV2PSP7VSSYPZQ2JAC/
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Kaustav
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>>
>>>>>>> ABHISHEK SAHNI
>>>>>>> Mob : +91-990-701-5143
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>> --
>>>>>>
>>>>>> KAUSTAV MAJUMDER
>>>>>>
>>>>>> ASSOCIATE SOFTWARE ENGINEER
>>>>>>
>>>>>> Red Hat India PVT LTD. <https://www.redhat.com/>
>>>>>>
>>>>>> [email protected]    M: 08981884037     IM: IRC: kmajumder
>>>>>> <https://red.ht/sig>
>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
>>>>>> @redhatway <https://twitter.com/redhatway>   @redhatinc
>>>>>> <https://instagram.com/redhatinc>   @redhatsnaps
>>>>>> <https://snapchat.com/add/redhatsnaps>
>>>>>>
>>>>>
>>>>
>>>> --
>>>> Thanks,
>>>>
>>>> Abhishek Sahni
>>>> Computer Centre
>>>> IISER Bhopal
>>>>
>>>
>>>
>>> --
>>> Thanks,
>>>
>>> Abhishek Sahni
>>> Computer Centre
>>> IISER Bhopal
>>>
>>
>> Thanks,
>> Kaustav
>>
>
>
> --
>
> ABHISHEK SAHNI
> Mob : +91-990-701-5143
>
>
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/IZLW425VXW5ZJAW7APV7WFWCZOXPYTTS/

Reply via email to