Next time when this happens, could you collect statedump of the brick
processes where this activity is going on at intervals of 10 seconds?

You can refer about how to take statedump at:
https://gluster.readthedocs.io/en/latest/Troubleshooting/statedump/

On Tue, May 16, 2017 at 7:43 PM, Jan Wrona <[email protected]> wrote:

> Hi,
>
> I have three servers in the linked list topology [1], GlusterFS 3.8.10,
> CentOS 7. Each server has two bricks, both on the same XFS filesystem. The
> XFS is constructed over the whole MD RAID device:
> md5 : active raid5 sdj1[6] sdh1[8] sde1[2] sdg1[9] sdd1[1] sdi1[5] sdf1[3]
> sdc1[0]
>       6836411904 blocks super 1.2 level 5, 512k chunk, algorithm 2 [8/8]
> [UUUUUUUU]
>       bitmap: 2/8 pages [8KB], 65536KB chunk
>
> Everything works fine until one of the RAID devices starts its regular
> check. During the check, the client's mount sometimes completely stops
> responding. I'm mounting using the Pacemaker's Filesystem OCF RA [2] with 
> OCF_CHECK_LEVEL=20,
> which basically tries to write a small status file to the filesystem every
> 2 minutes to see if its OK. But even this small write operation sometimes
> times out (2 minutes) during the check. Pacemaker then remounts the Gluster
> and everything goes back to normal.
>
> I understand that the RAID check is draining a lot of I/O performance, but
> the underlying XFS remains responsive (of course it is slower, but by far
> not as much as Gluster). The check intervals on the servers are not
> overlapping. I've even decreased the /proc/sys/dev/raid/speed_limit_max
> from the default 200 MB/s to the 50 MB/s, but it helped only a little, the
> mount still tends to freeze for a few seconds during the check.
>
> What are your suggestions to solve this issue?
>
> Regards,
> Jan Wrona
>
> [1] https://joejulian.name/blog/how-to-expand-glusterfs-
> replicated-clusters-by-one-server/
> [2] https://github.com/ClusterLabs/resource-agents/blob/master/heartbeat/
> Filesystem
>
> _______________________________________________
> Gluster-users mailing list
> [email protected]
> http://lists.gluster.org/mailman/listinfo/gluster-users
>



-- 
Pranith
_______________________________________________
Gluster-users mailing list
[email protected]
http://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to