On 12/26/2016 03:16 PM, Service Mail wrote:
Hello,
I have a cluster with 3 nodes with some replica volumes. I have
restart one of the brick while still writing to the others so that I
can test the healing (self healing is on). After few hrs the healing
seems to be stopped but if I check the status, after many hours, I
still can see the following:
[root@brick-a ~]# gluster volume heal backup info
Brick brick-a:/pool/backup
<gfid:670563bf-76da-41bc-a4b8-4bf5e05d7a4e>
<gfid:562106de-2768-4c66-9d9a-3e5da19ec3f6>
<gfid:2084e791-b3f8-4234-989b-6f3dfa395e6b>
<gfid:981b9503-efc0-4617-864a-7ce0768524e5>
<gfid:394201a8-d551-491b-81b9-89044c11076f>
<gfid:863335e0-79ec-46d4-8704-40e33e248653>
<gfid:a8342c60-76d9-4dc9-8e3e-a5c97a8f843a>
<gfid:9517abe3-a466-404c-8987-164fbd0c01ad>
<gfid:c075be3f-4f1b-4235-9172-1ad34bd273d1>
Status: Connected
Number of entries: 9
Can you check if the .glusterfs hardlinks are present for these entries
in brick-b? i.e `stat
brick-b:/pool/backup/.glusterfs/67/05/670563bf-76da-41bc-a4b8-4bf5e05d7a4e`
etc.
-Ravi
Brick brick-b:/pool/backup
Status: Connected
Far as I understand, this means some data on brick b (the one I have
restarted) is not healthy
I have tried "#gluster volume heal backup full" however the info
command still list those files as corrupted.
But the most concerning thing to me is that checking the directory
structure on all nodes they look different:
[root@brick-a backup]# du * -sh
301Gdata
1.9Tdata2
1.9Tdata3
[root@brick-b backup]# du * -sh 304G data 1.4T data2 1.7T data3
Could be that the 9 entries listed with the "info" commands are big
files being "synced" to the others bricks? how can I verified this?
Can you help?
Thanks
_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users