Re: [Gluster-devel] Spurious failures

2015-09-24 Thread Michael Scherer
Le jeudi 24 septembre 2015 à 06:50 -0400, Kotresh Hiremath Ravishankar a écrit : > >>> Ok, this definitely requires some tests and toughts. It only use ipv4 > >>> too ? > >>> (I guess yes, since ipv6 is removed from the rackspace build slaves) > > Yes! > > Could we

Re: [Gluster-devel] Spurious failures

2015-09-24 Thread Kotresh Hiremath Ravishankar
>>> Ok, this definitely requires some tests and toughts. It only use ipv4 >>> too ? >>> (I guess yes, since ipv6 is removed from the rackspace build slaves) Yes! Could we know when can these settings be done on all linux slave machines? If it takes sometime, we

Re: [Gluster-devel] Spurious failures

2015-09-24 Thread Kotresh Hiremath Ravishankar
Thank you:) and also please check the script I had given passes in all machines Thanks and Regards, Kotresh H R - Original Message - > From: "Michael Scherer" > To: "Kotresh Hiremath Ravishankar" > Cc: "Krutika Dhananjay" ,

Re: [Gluster-devel] Spurious failures

2015-09-24 Thread Kotresh Hiremath Ravishankar
Hi, >>>So, it is ok if I restrict that to be used only on 127.0.0.1 ? I think no, testcases use 'H0' to create volumes H0=${H0:=`hostname`}; Geo-rep expects passwordLess SSH to 'H0' Thanks and Regards, Kotresh H R - Original Message - >

[Gluster-devel] Memory leak in GlusterFS FUSE client

2015-09-24 Thread Oleksandr Natalenko
In our GlusterFS deployment we've encountered something like memory leak in GlusterFS FUSE client. We use replicated (×2) GlusterFS volume to store mail (exim+dovecot, maildir format). Here is inode stats for both bricks and mountpoint: === Brick 1 (Server 1): Filesystem

Re: [Gluster-devel] [Gluster-users] Memory leak in GlusterFS FUSE client

2015-09-24 Thread Oleksandr Natalenko
I've checked statedump of volume in question and haven't found lots of iobuf as mentioned in that bugreport. However, I've noticed that there are lots of LRU records like this: === [conn.1.bound_xl./bricks/r6sdLV07_vd0_mail/mail.lru.1] gfid=c4b29310-a19d-451b-8dd1-b3ac2d86b595 nlookup=1

Re: [Gluster-devel] [Gluster-users] Memory leak in GlusterFS FUSE client

2015-09-24 Thread Oleksandr Natalenko
We use bare GlusterFS installation with no oVirt involved. 24.09.2015 10:29, Gabi C wrote: google vdsm memory leak..it's been discussed on list last year and earlier this one... ___ Gluster-devel mailing list Gluster-devel@gluster.org

Re: [Gluster-devel] Spurious failures

2015-09-24 Thread Michael Scherer
Le jeudi 24 septembre 2015 à 02:24 -0400, Kotresh Hiremath Ravishankar a écrit : > Hi, > > >>>So, it is ok if I restrict that to be used only on 127.0.0.1 ? > I think no, testcases use 'H0' to create volumes > H0=${H0:=`hostname`}; > Geo-rep expects

[Gluster-devel] glusterfs 3.6.6 released

2015-09-24 Thread Raghavendra Bhat
Hi, glusterfs-3.6.6 has been released and the packages for RHEL/Fedora/Centos can be found here. http://download.gluster.org/pub/gluster/glusterfs/3.6/LATEST/ Requesting people running 3.6.x to please try it out and let us know if there are any issues. This release supposedly fixes the bugs

Re: [Gluster-devel] Spurious failures

2015-09-24 Thread Michael Scherer
Le jeudi 24 septembre 2015 à 07:59 -0400, Kotresh Hiremath Ravishankar a écrit : > Thank you:) and also please check the script I had given passes in all > machines So it worked everywhere, but on slave0 and slave1. Not sure what is wrong, or if they are used, I will check later. -- Michael