Hello,

we are using glusterfs 3.2.5 and have a distributed volume with over 10M directories. We recently added new node and initiated rebalance. After several days glusterfsd consumed all the memory and was killed by the kernel. At that stage it was still doing the layout rebalance and got over 9M directories.

Rebalance reports failed status:
# gluster volume rebalance vol status
rebalance failed

In the glusterd log however I found:
I [glusterd-rebalance.c:473:glusterd_defrag_start] 0-rebalance: rebalance on /etc/glusterd/mount/vol complete

I stumbled upon the patch on gluster-devel which mentions memory leak related rebalance: http://dev.gluster.com/pipermail/glusterfs/2011-June/003369.html. But can't figure out if this was included in the 3.2.5 release.

Another question is - is it safe to initiate rebalance of the data?

Thanks in advance
--
Emir Imamagic
Sektor za racunalne sustave
Sveuciliste u Zagrebu, Sveucilisni racunski centar (Srce), www.srce.unizg.hr
emir.imama...@srce.hr, tel: +385 1 616 5809, fax: +385 1 616 5559
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

Reply via email to