+ Pranith In the meantime, can you please provide the logs of all the gluster server machines and the client machines?
Logs can be found in /var/log/glusterfs directory. Regards, Raghavendra On Mon, Feb 8, 2016 at 9:20 AM, Dominique Roux <[email protected]> wrote: > Hi guys, > > I faced a problem a week ago. > In our environment we have three servers in a quorum. The gluster volume > is spreaded over two bricks and has the type replicated. > > We now, for simulating a fail of one brick, isolated one of the two > bricks with iptables, so that communication to the other two peers > wasn't possible anymore. > After that VMs (opennebula) which had I/O in this time crashed. > We stopped the glusterfsd hard (kill -9) and restarted it, what made > things work again (Certainly we also had to restart the failed VMs). But > I think this shouldn't happen. Since quorum was not reached (2/3 hosts > were still up and connected). > > Here some infos of our system: > OS: CentOS Linux release 7.1.1503 > Glusterfs version: glusterfs 3.7.3 > > gluster volume info: > > Volume Name: cluster1 > Type: Replicate > Volume ID: > Status: Started > Number of Bricks: 1 x 2 = 2 > Transport-type: tcp > Bricks: > Brick1: srv01:/home/gluster > Brick2: srv02:/home/gluster > Options Reconfigured: > cluster.self-heal-daemon: enable > cluster.server-quorum-type: server > network.remote-dio: enable > cluster.eager-lock: enable > performance.stat-prefetch: on > performance.io-cache: off > performance.read-ahead: off > performance.quick-read: off > server.allow-insecure: on > nfs.disable: 1 > > Hope you can help us. > > Thanks a lot. > > Best regards > Dominique > _______________________________________________ > Gluster-users mailing list > [email protected] > http://www.gluster.org/mailman/listinfo/gluster-users >
_______________________________________________ Gluster-users mailing list [email protected] http://www.gluster.org/mailman/listinfo/gluster-users
