Click on volume for which you want to reset the brick-> under bricks tab
select the brick you wan to reset -> once you do you will see the 'Reset
Brick' option is active.
Attached is a screenshot -> https://i.imgur.com/QUMSrzt.png

On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni <[email protected]>
wrote:

> Thanks Sahina for your response, I am not able to find it on UI, please
> help me to navigate? and yes I am using oVirt 4.2.6.4-1.
>
> On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <[email protected]> wrote:
>
>>
>>
>> On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni <
>> [email protected]> wrote:
>>
>>> Hello Team,
>>>
>>> We are running a setup of 3-way replica HC gluster setup configured
>>> during the initial deployment from the cockpit console using ansible.
>>>
>>> NODE1
>>>   - /dev/sda   (OS)
>>>   - /dev/sdb   ( Gluster Bricks )
>>>        * /gluster_bricks/engine/engine/
>>>        * /gluster_bricks/data/data/
>>>        * /gluster_bricks/vmstore/vmstore/
>>>
>>> NODE2 and NODE3 with a similar setup.
>>>
>>> There is a mishap that /dev/sdb on NODE2 totally got crashed and now
>>> there is nothing inside. However, I have created similar directories after
>>> mounting it back i.e.,
>>>
>>>        * /gluster_bricks/engine/engine/
>>>        * /gluster_bricks/data/data/
>>>        * /gluster_bricks/vmstore/vmstore/
>>> but it is not yet recovered.
>>>
>>> =====================================================
>>> [root@node2 ~]# gluster volume status
>>> Status of volume: data
>>> Gluster process                             TCP Port  RDMA Port  Online
>>> Pid
>>>
>>> ------------------------------------------------------------------------------
>>> Brick *.*.*.1:/gluster_bricks/data/data  49152     0          Y
>>>  11111
>>> Brick *.*.*.2:/gluster_bricks/data/data  N/A       N/A        N
>>>  N/A
>>> Brick *.*.*.3:/gluster_bricks/data/data  49152     0          Y
>>>  4303
>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>  23976
>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>  27838
>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>  27424
>>>
>>> Task Status of Volume data
>>>
>>> ------------------------------------------------------------------------------
>>> There are no active volume tasks
>>>
>>> Status of volume: engine
>>> Gluster process                             TCP Port  RDMA Port  Online
>>> Pid
>>>
>>> ------------------------------------------------------------------------------
>>> Brick *.*.*.1:/gluster_bricks/engine/eng
>>> ine                                         49153     0          Y
>>>  11117
>>> Brick *.*.*.2:/gluster_bricks/engine/eng
>>> ine                                         N/A       N/A        N
>>>  N/A
>>> Brick *.*.*.3:/gluster_bricks/engine/eng
>>> ine                                         49153     0          Y
>>>  4314
>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>  23976
>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>  27424
>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>  27838
>>>
>>> Task Status of Volume engine
>>>
>>> ------------------------------------------------------------------------------
>>> There are no active volume tasks
>>>
>>> Status of volume: vmstore
>>> Gluster process                             TCP Port  RDMA Port  Online
>>> Pid
>>>
>>> ------------------------------------------------------------------------------
>>> Brick *.*.*.1:/gluster_bricks/vmstore/vm
>>> store                                       49154     0          Y
>>>  21603
>>> Brick *.*.*.2:/gluster_bricks/vmstore/vm
>>> store                                       N/A       N/A        N
>>>  N/A
>>> Brick *.*.*.3:/gluster_bricks/vmstore/vm
>>> store                                       49154     0          Y
>>>  26845
>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>  23976
>>> Self-heal Daemon on *.*.*.3              N/A       N/A        Y
>>>  27424
>>> Self-heal Daemon on *.*.*.1              N/A       N/A        Y
>>>  27838
>>>
>>> Task Status of Volume vmstore
>>>
>>> ------------------------------------------------------------------------------
>>> There are no active volume tasks
>>> =============================================================
>>>
>>>
>>> Can someone please suggest the steps to recover the setup?
>>>
>>> I have tried the below workaround but it doesn't help.
>>>
>>>
>>> https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html
>>>
>>
>>  You can reset the brick - if you're on oVirt 4.2.x, there's a UI option
>> in the bricks subtab to do this.
>>
>>
>>>
>>> --
>>>
>>> ABHISHEK SAHNI
>>> Mob : +91-990-701-5143
>>>
>>>
>>> _______________________________________________
>>> Users mailing list -- [email protected]
>>> To unsubscribe send an email to [email protected]
>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>> oVirt Code of Conduct:
>>> https://www.ovirt.org/community/about/community-guidelines/
>>> List Archives:
>>> https://lists.ovirt.org/archives/list/[email protected]/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/
>>>
>>
>
> --
>
> ABHISHEK SAHNI
> Mob : +91-990-701-5143
>
>
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/GEZZIGBWA6HQ55JV2PSP7VSSYPZQ2JAC/
>


Thanks,
Kaustav
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/R3Y3YMUD5JIRPTZ4JVJTNZI7MXR4SQAW/

Reply via email to