‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
On Thursday, November 8, 2018 11:05 AM, Ravishankar N <ravishan...@redhat.com> 
wrote:

> It is not a split-brain. Nodes 1 and 3 have xattrs indicating a pending
> entry heal on node2 , so heal must have happened ideally. Can you check
> a few things?

> -   Is there any disconnects between each of the shds and the brick
>     processes (check via statedump or look for disconnect messages in
>     glustershd.log). Does restarting shd via a `volume start force` solve
>     the problem?

Yes there is one disconnect at 14:21 (UTC 13:21) because node2 ran out of 
memory (although it has 32 GB of RAM) and I had to reboot it. Here are the 
relevant log entries taken from glustershd.log on node1:

[2018-11-05 13:21:16.284239] C 
[rpc-clnt-ping.c:166:rpc_clnt_ping_timer_expired] 0-myvol-pro-client-1: server 
192.168.10.33:49154 has not responded in the last 42 seconds, disconnecting.
[2018-11-05 13:21:16.284385] I [MSGID: 114018] 
[client.c:2254:client_rpc_notify] 0-myvol-pro-client-1: disconnected from 
myvol-pro-client-1. Client process will keep trying to connect to glusterd 
until brick's port is available
[2018-11-05 13:21:16.284889] W [rpc-clnt-ping.c:222:rpc_clnt_ping_cbk] 
0-myvol-pro-client-1: socket disconnected

I also just ran a "volume start force" and saw that the glustershd processes 
got restarted on all 3 nodes but that did not trigger any healing. There are 
still the same amount of files/dirs pending heal...

> -   Is the symlink pointing to oc_dir present inside .glusterfs/25/e2 in
>     all 3 bricks?

They are yes for node1 and node3 but node2 there is no such symlink...

I hope that helps to debug the issue further, else please let me know if you 
need more info
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to