[Gluster-devel] GlusterFS Office Hours at FOSDEM
Hi, Lalatendu Mohanty, Niels de Vos, and myself will be holding GlusterFS Office Hours at FOSDEM. Look for us at the CentOS booth, from 16h00 to 17h00 on Saturday, 31 January. FOSDEM is taking place this weekend, 31 Jan and 1 Feb, at ULB Solbosch Campus, Brussels. FOSDEM is a free event, no registration is necessary. For more info see https://fosdem.org/2015/ Hope to see you there. -- Kaleb ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] GSOC-2015 heads up
Hi all, As it happens every year, the processes for GSOC begin in February, starting with the organization registration, which runs this year from February 9th to 20th. Rest of the timeline can be viewed at [1]. Last year we participated in GSOC under the Fedora organization, we hadn't registered as an organization in time. The Fedora organization graciously gave us one of their slots, and allowed us to become a part of GSOC for the first time. Vipul Nayyar implemented glusterfsiostat [2], under Krishnan P's mentorship. He created a tool to obtain and visualize I/O stats on GlusterFS clients, and also contributed improvements to GlusterFS code, specifically to the io-stats xlator. He successfully completed this project, and along the way became more involved with our community and is currently helping maintain the Gluster community infrastructure. This year we need to aim higher, and have multiple contributions coming in. This first and foremost requires our community to register itself as an organization, as we cannot continue to ask the Fedora community to lend us their valuable slots. We can discuss the details regarding the registration and further steps during the upcoming community meeting. Following this single success from last year, let's aim for multiple successes this year and grow our community. Cheers! ~kaushal [1]: https://www.google-melange.com/gsoc/homepage/google/gsoc2015 [2]: https://forge.gluster.org/glusterfsiostat ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] GlusterFS 4.0 updates needed
data classification has been updated to better reflect cache tiering's current status. - Original Message - From: Jeff Darcy jda...@redhat.com To: Gluster Devel gluster-devel@gluster.org Sent: Monday, January 26, 2015 7:53:37 AM Subject: [Gluster-devel] GlusterFS 4.0 updates needed Ahead of Wednesday's community meeting, I'd like to get as much 4.0 status together as possible. Developers, please check out the sub-projects on this page: http://www.gluster.org/community/documentation/index.php/Planning40 If you're involved in any of those, please update the status section of the corresponding feature page (near the bottom) as appropriate. Links to design notes, email threads, code review requests, etc. are most welcome. Yes, even things that have been sent to this list before. I'll be setting up a survey to choose a time for an online (Hangout/IRC) summit in the first week of February. If we can separate the items that are truly being worked on vs. those that just seemed cool at the time but haven't received one minute of anyone's attention since (e.g. sharding) then we can keep that meeting focused and productive. Thanks! ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] Bug in EXPECT family of macros in test framework
On Tue, Jan 27, 2015 at 12:35:35PM +0100, Xavier Hernandez wrote: Hi, I've just discovered that EXPECT, EXPECT_WITHIN and others have a bug that can hide some failures. The problem is that the test is made using a regular expression comparison. For example, EXPECT checks this: if ! [[ $a =~ $e ]]; then There are some tests that use a regular expression as the first argument to EXPECT, however there are many other tests that simply use a number or a text. This can cause problems. For example if the test is: EXPECT 0 test And the test returns 10, the regular expression 0 *will* match with 10, so the test will be considered successful when it shouldn't. It will also report success for test cases that check for an empty result: EXPECT test This will always succeed, even if the test returns something. What would be the best approach to solve this ? Oh, yuck! I think we should test like this, untested: [[ $a =~ ^$e\$ ]] It might break some regular expressions, the expressions are now forced to be very strict and should match the whole line. Thoughts? Niels pgpXr35MHc8Sw.pgp Description: PGP signature ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] Current state of multi-thread epoll (and review request)
Hi, Here is the current state of multi-thread epoll patch, 1) Patches: epoll patch: http://review.gluster.org/#/c/3842/ epoll configuration patch: http://review.gluster.org/#/c/9488/ 2) Failures: - epoll patch - In 2 runs the following 2 test cases failed in separate instances, - ./tests/bugs/quota/bug-1038598.t - Under investigation - ./tests/bugs/glusterd/bug-1104642.t - This is possibly a TC issue, as we check peer count and then immediately check for the configuration update. A peer count does not guarantee that the peer has also updated configuration. I will be sending a patch (today) to address this test case. - epoll configuration - Overall this patch passed, could be a one off as it depends on the above patch, or could be due to the fact that this makes glusterd and the CLI single threaded, and hence we do not hit the above failures. - NetBSD - NetBSD fails badly with the above patch - Under investigation 3) Requests - Request that people review the above patches, so that we can get concerns addressed quicker - The new patch on the configuration of epoll threads, needs some attention in terms of correctness of layers where this is allowed to be configured (among others) etc. Request reviewers to take a look at this. Thanks, Shyam ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] [Gluster-users] lockd: server not responding, timed out
Hi Niels, I see no kernel NFS service running on the gluster node. Here are the output root@glusterprod001:~# rpcinfo program version netid addressserviceowner 104tcp6 ::.0.111 portmapper superuser 103tcp6 ::.0.111 portmapper superuser 104udp6 ::.0.111 portmapper superuser 103udp6 ::.0.111 portmapper superuser 104tcp 0.0.0.0.0.111 portmapper superuser 103tcp 0.0.0.0.0.111 portmapper superuser 102tcp 0.0.0.0.0.111 portmapper superuser 104udp 0.0.0.0.0.111 portmapper superuser 103udp 0.0.0.0.0.111 portmapper superuser 102udp 0.0.0.0.0.111 portmapper superuser 104local /run/rpcbind.sock portmapper superuser 103local /run/rpcbind.sock portmapper superuser 153tcp 0.0.0.0.150.65 mountd superuser 151tcp 0.0.0.0.150.66 mountd superuser 133tcp 0.0.0.0.8.1nfssuperuser 1000214tcp 0.0.0.0.150.68 nlockmgr superuser 1002273tcp 0.0.0.0.8.1- superuser 1000211udp 0.0.0.0.2.215 nlockmgr superuser 1000211tcp 0.0.0.0.2.217 nlockmgr superuser 1000241udp 0.0.0.0.136.211status 105 1000241tcp 0.0.0.0.170.60 status 105 1000241udp6 ::.182.65 status 105 1000241tcp6 ::.172.250 status 105 root@glusterprod001:~# ss State Recv-Q Send-QLocal Address:PortPeer Address:Port ESTAB 0 0 10.101.165.61:1015 10.101.165.63:24007 ESTAB 0 0 10.101.165.61:93610.101.165.61:49156 ESTAB 0 0 10.101.165.61:1012 10.101.165.66:24007 ESTAB 0 0 10.101.165.61:49157 10.101.165.64:987 ESTAB 0 0 10.101.165.61:99910.101.165.62:49153 ESTAB 0 0 10.101.165.61:49157 10.101.165.66:988 ESTAB 0 0 10.101.165.61:49155 10.101.165.62:834 ESTAB 0 0 10.101.165.61:49156 10.101.165.63:998 ESTAB 0 0 10.101.165.61:91210.101.165.65:49153 ESTAB 0 0 127.0.0.1:982127.0.0.1:24007 ESTAB 0 0 10.101.165.61:49156 10.101.165.65:997 ESTAB 0 0 10.101.165.61:49155 10.101.165.61:850 ESTAB 0 0 10.101.165.61:92210.101.165.62:49154 ESTAB 0 0 10.101.165.61:89610.101.165.65:49154 ESTAB 0 0 10.101.165.61:1010 10.101.165.61:24007 ESTAB 0 0 10.101.165.61:imaps 10.101.165.61:49156 ESTAB 0 0 10.101.165.61:49155 10.101.165.63:981 ESTAB 0 0 10.101.165.61:93010.101.165.64:49155 ESTAB 0 0 10.101.165.61:4379 10.101.165.65:44899 ESTAB 0 0 10.101.165.61:98310.101.165.61:49157 ESTAB 0 0
Re: [Gluster-devel] Ability to turn off 'glusterfs' protocol
On Tue, Jan 27, 2015 at 02:10:17AM +0100, Csaba Henk wrote: Hi Niels, On Mon, Jan 26, 2015 at 10:07 AM, Niels de Vos nde...@redhat.com wrote: On Sun, Jan 25, 2015 at 10:08:20PM +0530, Ramana Raja wrote: 1) How would the above suggestion impact gfapi access? gfapi uses the GlusterFS protocol. When access through the protocol is denied, only clients from within the trusted storage pool can use it. gfapi is just a client, similar to the FUSE mount. 2) Would the list of trusted clients be configured via gluster volume set? There are the 'auth.allow' and 'auth.reject' volume options. I think they would be used as an alternative to a turn off glusterfs protocol feature. Does it mean that the implementation of feature would essentially boil down to an auth ruleset calculated by glusterfs? I guess that depends on the goal of the feature. Where does the need arrise to turn off the glusterfs protocol? Should nothing outside of the trusted storage pool be able to connect to the bricks? This would effectively only allow NFS/Samba/... when the service is located on a system that is part of the trusted storage pool. If you need a more fine-grained permission scheme, auth.allow and auth.reject might be more appropriate. Or, use SSL authentication? HTH, Niels pgpPNrRk9vj_K.pgp Description: PGP signature ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] GlusterFS 3.7 updates needed
Hi All, As we approach feature freeze for 3.7 (28th Feb), it would great to have a status update for features being worked on for the release. Action(s) needed for status updates: 1. If you are the owner for one of the features listed in the planning page [1], can you please update status of the implementation and also include any relevant mailing list, IRC discussions and design decisions that can help in providing more clarity on the feature? 2. If you are already working on a feature for 3.7 and have not yet opened a new feature page as yet or linked the feature page in the planning page, can you please ensure that we have one listed by Feb 10th (i.e. 2 weeks from now)? It might be hard to accommodate new feature proposals and reviews for such features if we don't have an appropriate feature page linked in the release planning page by Feb 10th. 3. If you have planned a feature, are blocked and need more assistance to complete it by the feature freeze date, please let us know. Thanks, Vijay [1] http://www.gluster.org/community/documentation/index.php/Planning37 ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] GlusterFS 3.7 updates needed
On 2015-01-27 11:49, Vijay Bellur wrote: Hi All, As we approach feature freeze for 3.7 (28th Feb), it would great to have a status update for features being worked on for the release. Not exactly a new feature, but something I would like to see happen: [Gluster-devel] Glusterd daemon management code refactoring http://www.gluster.org/pipermail/gluster-devel/2014-December/043180.html /Anders -- Anders Blomdell Email: anders.blomd...@control.lth.se Department of Automatic Control Lund University Phone:+46 46 222 4625 P.O. Box 118 Fax: +46 46 138118 SE-221 00 Lund, Sweden ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] GlusterFS 3.7 updates needed
Glusterd daemon management code refactoring is being worked on. See http://review.gluster.org/9428 (and dependent patches) for current status. I have added you to the list of reviewers. Hope that is OK with you. - Original Message - On 2015-01-27 11:49, Vijay Bellur wrote: Hi All, As we approach feature freeze for 3.7 (28th Feb), it would great to have a status update for features being worked on for the release. Not exactly a new feature, but something I would like to see happen: [Gluster-devel] Glusterd daemon management code refactoring http://www.gluster.org/pipermail/gluster-devel/2014-December/043180.html /Anders -- Anders Blomdell Email: anders.blomd...@control.lth.se Department of Automatic Control Lund University Phone:+46 46 222 4625 P.O. Box 118 Fax: +46 46 138118 SE-221 00 Lund, Sweden ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] Bug in EXPECT family of macros in test framework
Hi, I've just discovered that EXPECT, EXPECT_WITHIN and others have a bug that can hide some failures. The problem is that the test is made using a regular expression comparison. For example, EXPECT checks this: if ! [[ $a =~ $e ]]; then There are some tests that use a regular expression as the first argument to EXPECT, however there are many other tests that simply use a number or a text. This can cause problems. For example if the test is: EXPECT 0 test And the test returns 10, the regular expression 0 *will* match with 10, so the test will be considered successful when it shouldn't. It will also report success for test cases that check for an empty result: EXPECT test This will always succeed, even if the test returns something. What would be the best approach to solve this ? Xavi ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] Quota with hard-links
Hi All, This is regarding quota accounting for hard-links. Currently, accounting is done only once for the links created within the same directory and accounting is done separately when links are created in separate directory. With this approach account may go wrong when rename is performed on the hardlink files across directories. We are implementing one of the below mentioned policy for hard-links when quota is enabled. 1) Allow creating hard-links only within same directory. (We can hit the same problem if quota is enabled on the pre-existing data which contains hard-links) 2) Allow creating hard-links only within the same branch where the limit is set (We can hit the same problem if quota is enabled on the pre-existing data which contains hard-links and also when quota-limit is set/unset) 3) Account for all the hard-links. (Marker treats all hard-links as a new file) Please provide your feedback. Thanks, Vijay ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] quota and snapshot testcase failure (zfs on CentOS 6.6)
Hi Kiran, Testcase './tests/basic/quota-anon-fd-nfs.t' is removed from the test suite.There are some issues with this testcase, we are working on it. We will add this test-case soon once the issue is fixed. Thanks, Vijay On Tuesday 27 January 2015 06:11 PM, Vijaikumar M wrote: Hi Kiran, quota.t failure issue has been fixed with patch http://review.gluster.org/#/c/9410/. Can you please re-try the test with this patch and see if it works? Thanks, Vijay On Wednesday 19 November 2014 10:32 AM, Pranith Kumar Karampuri wrote: On 11/19/2014 10:30 AM, Atin Mukherjee wrote: On 11/18/2014 10:35 PM, Pranith Kumar Karampuri wrote: On 11/12/2014 04:52 PM, Kiran Patil wrote: I have create zpool with name d and mnt and they appear in filesystem as follows. d on /d type zfs (rw,xattr) mnt on /mnt type zfs (rw,xattr) Debug enabled output of quota.t testcase is at http://ur1.ca/irbt1. CC vijaikumar quota-anon-fd-nfs.t spurious failure fix is addressed by http://review.gluster.org/#/c/9108/ This is just quota.t in tests/basic, not the anon-fd one Pranith ~Atin On Wed, Nov 12, 2014 at 3:22 PM, Kiran Patil ki...@fractalio.com mailto:ki...@fractalio.com wrote: Hi, Gluster suite report, Gluster version: glusterfs 3.6.1 On disk filesystem: Zfs 0.6.3-1.1 Operating system: CentOS release 6.6 (Final) We are seeing quota and snapshot testcase failures. We are not sure why quota is failing since quotas worked fine on gluster 3.4. Test Summary Report --- ./tests/basic/quota-anon-fd-nfs.t (Wstat: 0 Tests: 16 Failed: 1) Failed test: 16 ./tests/basic/quota.t (Wstat: 0 Tests: 73 Failed: 4) Failed tests: 24, 28, 32, 65 ./tests/basic/uss.t (Wstat: 0 Tests: 147 Failed: 78) Failed tests: 8-11, 16-25, 28-29, 31-32, 39-40, 45-47 49-57, 60-61, 63-64, 71-72, 78-87, 90-91 93-94, 101-102, 107-115, 118-119, 121-122 129-130, 134, 136-137, 139-140, 142-143 145-146 ./tests/basic/volume-snapshot.t (Wstat: 0 Tests: 30 Failed: 12) Failed tests: 11-18, 21-24 ./tests/basic/volume-status.t (Wstat: 0 Tests: 14 Failed: 1) Failed test: 14 ./tests/bugs/bug-1023974.t (Wstat: 0 Tests: 15 Failed: 1) Failed test: 12 ./tests/bugs/bug-1038598.t (Wstat: 0 Tests: 28 Failed: 6) Failed tests: 17, 21-22, 26-28 ./tests/bugs/bug-1045333.t (Wstat: 0 Tests: 16 Failed: 9) Failed tests: 7-15 ./tests/bugs/bug-1049834.t (Wstat: 0 Tests: 18 Failed: 7) Failed tests: 11-14, 16-18 ./tests/bugs/bug-1087203.t (Wstat: 0 Tests: 43 Failed: 2) Failed tests: 31, 41 ./tests/bugs/bug-1090042.t (Wstat: 0 Tests: 12 Failed: 3) Failed tests: 9-11 ./tests/bugs/bug-1109770.t (Wstat: 0 Tests: 19 Failed: 4) Failed tests: 8-11 ./tests/bugs/bug-1109889.t (Wstat: 0 Tests: 20 Failed: 4) Failed tests: 8-11 ./tests/bugs/bug-1112559.t (Wstat: 0 Tests: 11 Failed: 3) Failed tests: 8-9, 11 ./tests/bugs/bug-1112613.t (Wstat: 0 Tests: 22 Failed: 5) Failed tests: 12-14, 17-18 ./tests/bugs/bug-1113975.t (Wstat: 0 Tests: 13 Failed: 4) Failed tests: 8-9, 11-12 ./tests/bugs/bug-847622.t (Wstat: 0 Tests: 10 Failed: 1) Failed test: 8 ./tests/bugs/bug-861542.t (Wstat: 0 Tests: 13 Failed: 7) Failed tests: 7-13 ./tests/features/ssl-authz.t (Wstat: 0 Tests: 18 Failed: 1) Failed test: 18 Files=277, Tests=7908, 8147 wallclock secs ( 4.56 usr 0.78 sys + 774.74 cusr 666.97 csys = 1447.05 CPU) Result: FAIL Thanks, Kiran. ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] quota and snapshot testcase failure (zfs on CentOS 6.6)
Hi Kiran, quota.t failure issue has been fixed with patch http://review.gluster.org/#/c/9410/. Can you please re-try the test with this patch and see if it works? Thanks, Vijay On Wednesday 19 November 2014 10:32 AM, Pranith Kumar Karampuri wrote: On 11/19/2014 10:30 AM, Atin Mukherjee wrote: On 11/18/2014 10:35 PM, Pranith Kumar Karampuri wrote: On 11/12/2014 04:52 PM, Kiran Patil wrote: I have create zpool with name d and mnt and they appear in filesystem as follows. d on /d type zfs (rw,xattr) mnt on /mnt type zfs (rw,xattr) Debug enabled output of quota.t testcase is at http://ur1.ca/irbt1. CC vijaikumar quota-anon-fd-nfs.t spurious failure fix is addressed by http://review.gluster.org/#/c/9108/ This is just quota.t in tests/basic, not the anon-fd one Pranith ~Atin On Wed, Nov 12, 2014 at 3:22 PM, Kiran Patil ki...@fractalio.com mailto:ki...@fractalio.com wrote: Hi, Gluster suite report, Gluster version: glusterfs 3.6.1 On disk filesystem: Zfs 0.6.3-1.1 Operating system: CentOS release 6.6 (Final) We are seeing quota and snapshot testcase failures. We are not sure why quota is failing since quotas worked fine on gluster 3.4. Test Summary Report --- ./tests/basic/quota-anon-fd-nfs.t (Wstat: 0 Tests: 16 Failed: 1) Failed test: 16 ./tests/basic/quota.t (Wstat: 0 Tests: 73 Failed: 4) Failed tests: 24, 28, 32, 65 ./tests/basic/uss.t (Wstat: 0 Tests: 147 Failed: 78) Failed tests: 8-11, 16-25, 28-29, 31-32, 39-40, 45-47 49-57, 60-61, 63-64, 71-72, 78-87, 90-91 93-94, 101-102, 107-115, 118-119, 121-122 129-130, 134, 136-137, 139-140, 142-143 145-146 ./tests/basic/volume-snapshot.t (Wstat: 0 Tests: 30 Failed: 12) Failed tests: 11-18, 21-24 ./tests/basic/volume-status.t (Wstat: 0 Tests: 14 Failed: 1) Failed test: 14 ./tests/bugs/bug-1023974.t (Wstat: 0 Tests: 15 Failed: 1) Failed test: 12 ./tests/bugs/bug-1038598.t (Wstat: 0 Tests: 28 Failed: 6) Failed tests: 17, 21-22, 26-28 ./tests/bugs/bug-1045333.t (Wstat: 0 Tests: 16 Failed: 9) Failed tests: 7-15 ./tests/bugs/bug-1049834.t (Wstat: 0 Tests: 18 Failed: 7) Failed tests: 11-14, 16-18 ./tests/bugs/bug-1087203.t (Wstat: 0 Tests: 43 Failed: 2) Failed tests: 31, 41 ./tests/bugs/bug-1090042.t (Wstat: 0 Tests: 12 Failed: 3) Failed tests: 9-11 ./tests/bugs/bug-1109770.t (Wstat: 0 Tests: 19 Failed: 4) Failed tests: 8-11 ./tests/bugs/bug-1109889.t (Wstat: 0 Tests: 20 Failed: 4) Failed tests: 8-11 ./tests/bugs/bug-1112559.t (Wstat: 0 Tests: 11 Failed: 3) Failed tests: 8-9, 11 ./tests/bugs/bug-1112613.t (Wstat: 0 Tests: 22 Failed: 5) Failed tests: 12-14, 17-18 ./tests/bugs/bug-1113975.t (Wstat: 0 Tests: 13 Failed: 4) Failed tests: 8-9, 11-12 ./tests/bugs/bug-847622.t (Wstat: 0 Tests: 10 Failed: 1) Failed test: 8 ./tests/bugs/bug-861542.t (Wstat: 0 Tests: 13 Failed: 7) Failed tests: 7-13 ./tests/features/ssl-authz.t (Wstat: 0 Tests: 18 Failed: 1) Failed test: 18 Files=277, Tests=7908, 8147 wallclock secs ( 4.56 usr 0.78 sys + 774.74 cusr 666.97 csys = 1447.05 CPU) Result: FAIL Thanks, Kiran. ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] glusterfs-3.6.2 GA
glusterfs-3.6.2 has been released. The release source tar file and packages for Fedora {20,21,rawhide}, RHEL/CentOS {5,6,7}, Debian {wheezy,jessie}, Pidora2014, and Raspbian wheezy are available at http://download.gluster.org/pub/gluster/glusterfs/3.6/3.6.2/ (Ubuntu packages will be available soon.) This release fixes the following bugs. Thanks to all who submitted bugs and patches and reviewed the changes. 1184191 - Cluster/DHT : Fixed crash due to null deref 1180404 - nfs server restarts when a snapshot is deactivated 1180411 - CIFS:[USS]: glusterfsd OOM killed when 255 snapshots were browsed at CIFS mount and Control+C is issued 1180070 - [AFR] getfattr on fuse mount gives error : Software caused connection abort 1175753 - [readdir-ahead]: indicate EOF for readdirp 1175752 - [USS]: On a successful lookup, snapd logs are filled with Warnings dict OR key (entry-point) is NULL 1175749 - glusterfs client crashed while migrating the fds 1179658 - Add brick fails if parent dir of new brick and existing brick is same and volume was accessed using libgfapi and smb. 1146524 - glusterfs.spec.in - synch minor diffs with fedora dist-git glusterfs.spec 1175744 - [USS]: Unable to access .snaps after snapshot restore after directories were deleted and recreated 1175742 - [USS]: browsing .snaps directory with CIFS fails with Invalid argument 1175739 - [USS]: Non root user who has no access to a directory, from NFS mount, is able to access the files under .snaps under that directory 1175758 - [USS] : Rebalance process tries to connect to snapd and in case when snapd crashes it might affect rebalance process 1175765 - USS]: When snapd is crashed gluster volume stop/delete operation fails making the cluster in inconsistent state 1173528 - Change in volume heal info command output 1166515 - [Tracker] RDMA support in glusterfs 1166505 - mount fails for nfs protocol in rdma volumes 1138385 - [DHT:REBALANCE]: Rebalance failures are seen with error message remote operation failed: File exists 1177418 - entry self-heal in 3.5 and 3.6 are not compatible 1170954 - Fix mutex problems reported by coverity scan 1177899 - nfs: ls shows Permission denied with root-squash 1175738 - [USS]: data unavailability for a period of time when USS is enabled/disabled 1175736 - [USS]:After deactivating a snapshot trying to access the remaining activated snapshots from NFS mount gives 'Invalid argument' error 1175735 - [USS]: snapd process is not killed once the glusterd comes back 1175733 - [USS]: If the snap name is same as snap-directory than cd to virtual snap directory fails 1175756 - [USS] : Snapd crashed while trying to access the snapshots under .snaps directory 1175755 - SNAPSHOT[USS]:gluster volume set for uss doesnot check any boundaries 1175732 - [SNAPSHOT]: nouuid is appended for every snapshoted brick which causes duplication if the original brick has already nouuid 1175730 - [USS]: creating file/directories under .snaps shows wrong error message 1175754 - [SNAPSHOT]: before the snap is marked to be deleted if the node goes down than the snaps are propagated on other nodes and glusterd hungs 1159484 - ls -alR can not heal the disperse volume 1138897 - NetBSD port 1175728 - [USS]: All uss related logs are reported under /var/log/glusterfs, it makes sense to move it into subfolder 1170548 - [USS] : don't display the snapshots which are not activated 1170921 - [SNAPSHOT]: snapshot should be deactivated by default when created 1175694 - [SNAPSHOT]: snapshoted volume is read only but it shows rw attributes in mount 1161885 - Possible file corruption on dispersed volumes 1170959 - EC_MAX_NODES is defined incorrectly 1175645 - [USS]: Typo error in the description for USS under gluster volume set help 1171259 - mount.glusterfs does not understand -n option Regards, Kaleb, on behalf of Raghavendra Bhat, who did all the work. ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] Minutes of todays Gluster Community Bug Triage meeting
On Tue, Jan 27, 2015 at 12:07:35PM +0100, Niels de Vos wrote: Hi all, This meeting is scheduled for anyone that is interested in learning more about, or assisting with the Bug Triage. Meeting details: - location: #gluster-meeting on Freenode IRC - date: every Tuesday - time: 12:00 UTC, 13:00 CET (run: date -d 12:00 UTC) - agenda: https://public.pad.fsfe.org/p/gluster-bug-triage Currently the following items are listed: * Roll Call * Status of last weeks action items * Group Triage * Open Floor The last two topics have space for additions. If you have a suitable bug or topic to discuss, please add it to the agenda. Minutes: http://meetbot.fedoraproject.org/gluster-meeting/2015-01-27/gluster-meeting.2015-01-27-12.03.html Minutes (text): http://meetbot.fedoraproject.org/gluster-meeting/2015-01-27/gluster-meeting.2015-01-27-12.03.txt Log: http://meetbot.fedoraproject.org/gluster-meeting/2015-01-27/gluster-meeting.2015-01-27-12.03.log.html Meeting started by ndevos at 12:03:35 UTC. Meeting summary --- * Roll Call (ndevos, 12:04:07) * Action items from last week (ndevos, 12:06:26) * hagarth will look for somebody that can act like a bug assigner manager kind of person (ndevos, 12:06:31) * hagarth should update the MAINTAINERS file, add current maintainers and new components like Snapshot (ndevos, 12:08:12) * lalatenduM will send a reminder to the users- and devel- ML about (and how to) fixing Coverity defects (ndevos, 12:09:09) * lalatenduM initiate a discussion with the RH Gluster team to triage their own bugs when they report them (ndevos, 12:09:53) * ndevos adds NetBSD queries to the bug triage wiki page (ndevos, 12:10:49) * hagarth start/sync email on regular (nightly) automated tests (ndevos, 12:11:50) * lalatenduM will look into using nightly builds for automated testing, and will report issues/success to the mailinglist (ndevos, 12:12:42) * Group Triage (ndevos, 12:16:20) * Open Floor (ndevos, 12:31:24) Meeting ended at 12:41:57 UTC. People Present (lines said) --- * ndevos (45) * hagarth (18) * kkeithley_ (13) * hchiramm (10) * gothos (7) * zodbot (2) pgpFpSJW82D3j.pgp Description: PGP signature ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] [Gluster-users] v3.6.2
After shutting down all NFS and gluster processes, there was still an NFS process. [root@gfs01bkp ~]# rpcinfo -p program vers proto port service 104 tcp111 portmapper 103 tcp111 portmapper 102 tcp111 portmapper 104 udp111 portmapper 103 udp111 portmapper 102 udp111 portmapper 153 tcp 38465 mountd 151 tcp 38466 mountd 133 tcp 2049 nfs 1000241 udp 34738 status 1000241 tcp 37269 status [root@gfs01bkp ~]# netstat -anp | grep 2049 [root@gfs01bkp ~]# netstat -anp | grep 38465 [root@gfs01bkp ~]# netstat -anp | grep 38466 I killed off the processes using rpcinfo -d [root@gfs01bkp ~]# rpcinfo -p program vers proto port service 104 tcp111 portmapper 103 tcp111 portmapper 102 tcp111 portmapper 104 udp111 portmapper 103 udp111 portmapper 102 udp111 portmapper 1000241 udp 34738 status 1000241 tcp 37269 status Then I restarted the glusterd and did a 'mount -a'. Worked perfectly. And the errors that were showing up in the logs every 3-seconds stopped. Thanks for your help. Greatly appreciated. David -- Original Message -- From: Xavier Hernandez xhernan...@datalab.es To: David F. Robinson david.robin...@corvidtec.com; Kaushal M kshlms...@gmail.com Cc: Gluster Users gluster-us...@gluster.org; Gluster Devel gluster-devel@gluster.org Sent: 1/27/2015 10:02:31 AM Subject: Re: [Gluster-devel] [Gluster-users] v3.6.2 Hi, I had a similar problem once. It happened after doing some unrelated tests with NFS. I thought it was a problem I generated doing weird things, so I didn't investigate the cause further. To see if this is the same case, try this: * Unmount all NFS mounts and stop all gluster volumes * Check that there are no gluster processes running (ps ax | grep gluster), specially any glusterfs. glusterd is ok. * Check that there are no NFS processes running (ps ax | grep nfs) * Check with 'rpcinfo -p' that there's no nfs service registered The output should be similar to this: program vers proto port service 10 4 tcp 111 portmapper 10 3 tcp 111 portmapper 10 2 tcp 111 portmapper 10 4 udp 111 portmapper 10 3 udp 111 portmapper 10 2 udp 111 portmapper 100024 1 udp 33482 status 100024 1 tcp 37034 status If there are more services registered, you can directly delete them or check if they correspond to an active process. For example, if the output is this: program vers proto port service 10 4 tcp 111 portmapper 10 3 tcp 111 portmapper 10 2 tcp 111 portmapper 10 4 udp 111 portmapper 10 3 udp 111 portmapper 10 2 udp 111 portmapper 100021 3 udp 39618 nlockmgr 100021 3 tcp 41067 nlockmgr 100024 1 udp 33482 status 100024 1 tcp 37034 status You can do a netstat -anp | grep 39618 to see if there is some process really listening at the nlockmgr port. You can repeat this for port 41067. If there is some process, you should stop it. If there is no process listening on that port, you should remove it with a command like this: rpcinfo -d 100021 3 You must execute this command for all stale ports for any services other than portmapper and status. Once done you should get the output shown before. After that, you can try to start your volume and see if everything is registered (rpcinfo -p) and if gluster has started the nfs server (gluster volume status). If everything is ok, you should be able to mount the volume using NFS. Xavi On 01/27/2015 03:18 PM, David F. Robinson wrote: Turning off nfslock did not help. Also, still getting these messages every 3-seconds: [2015-01-27 14:16:12.921880] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:15.922431] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:18.923080] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:21.923748] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:24.924472] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:27.925192] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:30.925895] W [socket.c:611:__socket_rwv] 0-management: readv on
Re: [Gluster-devel] [Gluster-users] v3.6.2
I rebooted the machine to see if the problem would return and it does. Same issue after a reboot. Any suggestions? One other thing I tested was to comment out the NFS mounts in /etc/fstab: # gfsib01bkp.corvidtec.com:/homegfs_bkp /backup_nfs/homegfs nfs vers=3,intr,bg,rsize=32768,wsize=32768 0 0 After the machine comes back up, I remove the comment and do a 'mount -a'. The mount works fine. It looks like it is a timing during startup issue. Is it trying to do the NFS mount while glusterd is still starting up? David -- Original Message -- From: Xavier Hernandez xhernan...@datalab.es To: David F. Robinson david.robin...@corvidtec.com; Kaushal M kshlms...@gmail.com Cc: Gluster Users gluster-us...@gluster.org; Gluster Devel gluster-devel@gluster.org Sent: 1/27/2015 10:02:31 AM Subject: Re: [Gluster-devel] [Gluster-users] v3.6.2 Hi, I had a similar problem once. It happened after doing some unrelated tests with NFS. I thought it was a problem I generated doing weird things, so I didn't investigate the cause further. To see if this is the same case, try this: * Unmount all NFS mounts and stop all gluster volumes * Check that there are no gluster processes running (ps ax | grep gluster), specially any glusterfs. glusterd is ok. * Check that there are no NFS processes running (ps ax | grep nfs) * Check with 'rpcinfo -p' that there's no nfs service registered The output should be similar to this: program vers proto port service 10 4 tcp 111 portmapper 10 3 tcp 111 portmapper 10 2 tcp 111 portmapper 10 4 udp 111 portmapper 10 3 udp 111 portmapper 10 2 udp 111 portmapper 100024 1 udp 33482 status 100024 1 tcp 37034 status If there are more services registered, you can directly delete them or check if they correspond to an active process. For example, if the output is this: program vers proto port service 10 4 tcp 111 portmapper 10 3 tcp 111 portmapper 10 2 tcp 111 portmapper 10 4 udp 111 portmapper 10 3 udp 111 portmapper 10 2 udp 111 portmapper 100021 3 udp 39618 nlockmgr 100021 3 tcp 41067 nlockmgr 100024 1 udp 33482 status 100024 1 tcp 37034 status You can do a netstat -anp | grep 39618 to see if there is some process really listening at the nlockmgr port. You can repeat this for port 41067. If there is some process, you should stop it. If there is no process listening on that port, you should remove it with a command like this: rpcinfo -d 100021 3 You must execute this command for all stale ports for any services other than portmapper and status. Once done you should get the output shown before. After that, you can try to start your volume and see if everything is registered (rpcinfo -p) and if gluster has started the nfs server (gluster volume status). If everything is ok, you should be able to mount the volume using NFS. Xavi On 01/27/2015 03:18 PM, David F. Robinson wrote: Turning off nfslock did not help. Also, still getting these messages every 3-seconds: [2015-01-27 14:16:12.921880] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:15.922431] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:18.923080] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:21.923748] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:24.924472] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:27.925192] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:30.925895] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:33.926563] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:36.927248] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) -- Original Message -- From: Kaushal M kshlms...@gmail.com mailto:kshlms...@gmail.com To: David F. Robinson david.robin...@corvidtec.com mailto:david.robin...@corvidtec.com Cc: Joe Julian j...@julianfamily.org mailto:j...@julianfamily.org; Gluster Users gluster-us...@gluster.org mailto:gluster-us...@gluster.org; Gluster Devel gluster-devel@gluster.org mailto:gluster-devel@gluster.org Sent: 1/27/2015 1:49:56 AM Subject: Re: Re[2]: [Gluster-devel]
Re: [Gluster-devel] [Gluster-users] v3.6.2
In my /etc/fstab, I have the following: gfsib01bkp.corvidtec.com:/homegfs_bkp /backup/homegfs glusterfs transport=tcp,_netdev 0 0 gfsib01bkp.corvidtec.com:/Software_bkp /backup/Software glusterfs transport=tcp,_netdev 0 0 gfsib01bkp.corvidtec.com:/Source_bkp /backup/Source glusterfs transport=tcp,_netdev 0 0 #... Setup NFS mounts as well gfsib01bkp.corvidtec.com:/homegfs_bkp /backup_nfs/homegfs nfs vers=3,intr,bg,rsize=32768,wsize=32768 0 0 It looks like it is trying to start the nfs mount before gluster has finished coming up and that this is hanging the nfs ports. I have _netdev in the glusterfs mount point to make sure the network has come up (including infiniband) prior to starting gluster. Shouldn't the gluster init scripts check for gluster startup prior to starting the nfs mount? It doesn't look like this is working properly. David -- Original Message -- From: Xavier Hernandez xhernan...@datalab.es To: David F. Robinson david.robin...@corvidtec.com; Kaushal M kshlms...@gmail.com Cc: Gluster Users gluster-us...@gluster.org; Gluster Devel gluster-devel@gluster.org Sent: 1/27/2015 10:02:31 AM Subject: Re: [Gluster-devel] [Gluster-users] v3.6.2 Hi, I had a similar problem once. It happened after doing some unrelated tests with NFS. I thought it was a problem I generated doing weird things, so I didn't investigate the cause further. To see if this is the same case, try this: * Unmount all NFS mounts and stop all gluster volumes * Check that there are no gluster processes running (ps ax | grep gluster), specially any glusterfs. glusterd is ok. * Check that there are no NFS processes running (ps ax | grep nfs) * Check with 'rpcinfo -p' that there's no nfs service registered The output should be similar to this: program vers proto port service 10 4 tcp 111 portmapper 10 3 tcp 111 portmapper 10 2 tcp 111 portmapper 10 4 udp 111 portmapper 10 3 udp 111 portmapper 10 2 udp 111 portmapper 100024 1 udp 33482 status 100024 1 tcp 37034 status If there are more services registered, you can directly delete them or check if they correspond to an active process. For example, if the output is this: program vers proto port service 10 4 tcp 111 portmapper 10 3 tcp 111 portmapper 10 2 tcp 111 portmapper 10 4 udp 111 portmapper 10 3 udp 111 portmapper 10 2 udp 111 portmapper 100021 3 udp 39618 nlockmgr 100021 3 tcp 41067 nlockmgr 100024 1 udp 33482 status 100024 1 tcp 37034 status You can do a netstat -anp | grep 39618 to see if there is some process really listening at the nlockmgr port. You can repeat this for port 41067. If there is some process, you should stop it. If there is no process listening on that port, you should remove it with a command like this: rpcinfo -d 100021 3 You must execute this command for all stale ports for any services other than portmapper and status. Once done you should get the output shown before. After that, you can try to start your volume and see if everything is registered (rpcinfo -p) and if gluster has started the nfs server (gluster volume status). If everything is ok, you should be able to mount the volume using NFS. Xavi On 01/27/2015 03:18 PM, David F. Robinson wrote: Turning off nfslock did not help. Also, still getting these messages every 3-seconds: [2015-01-27 14:16:12.921880] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:15.922431] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:18.923080] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:21.923748] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:24.924472] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:27.925192] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:30.925895] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:33.926563] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:36.927248] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) -- Original Message -- From: Kaushal M kshlms...@gmail.com
Re: [Gluster-devel] [Gluster-users] v3.6.2
Not elegant, but here is my short-term fix to prevent the issue after a reboot: Added 'noauto' to the mount in /etc/rc.local: /etc/fstab: #... Note: Used the 'noauto' for the NFS mounts and put the mount in /etc/rc.local to ensure that #... glsuter has been started before attempting to mount using NFS. Otherwise, hangs ports #... during startup. gfsib01bkp.corvidtec.com:/homegfs_bkp /backup_nfs/homegfs nfs vers=3,intr,bg,rsize=32768,wsize=32768,noauto 0 0 gfsib01a.corvidtec.com:/homegfs /homegfs_nfs nfs vers=3,intr,bg,rsize=32768,wsize=32768,noauto 0 0 /etc/rc.local: /etc/init.d/glusterd restart (sleep 20; mount -a; mount /backup_nfs/homegfs) -- Original Message -- From: Xavier Hernandez xhernan...@datalab.es To: David F. Robinson david.robin...@corvidtec.com; Kaushal M kshlms...@gmail.com Cc: Gluster Users gluster-us...@gluster.org; Gluster Devel gluster-devel@gluster.org Sent: 1/27/2015 10:02:31 AM Subject: Re: [Gluster-devel] [Gluster-users] v3.6.2 Hi, I had a similar problem once. It happened after doing some unrelated tests with NFS. I thought it was a problem I generated doing weird things, so I didn't investigate the cause further. To see if this is the same case, try this: * Unmount all NFS mounts and stop all gluster volumes * Check that there are no gluster processes running (ps ax | grep gluster), specially any glusterfs. glusterd is ok. * Check that there are no NFS processes running (ps ax | grep nfs) * Check with 'rpcinfo -p' that there's no nfs service registered The output should be similar to this: program vers proto port service 10 4 tcp 111 portmapper 10 3 tcp 111 portmapper 10 2 tcp 111 portmapper 10 4 udp 111 portmapper 10 3 udp 111 portmapper 10 2 udp 111 portmapper 100024 1 udp 33482 status 100024 1 tcp 37034 status If there are more services registered, you can directly delete them or check if they correspond to an active process. For example, if the output is this: program vers proto port service 10 4 tcp 111 portmapper 10 3 tcp 111 portmapper 10 2 tcp 111 portmapper 10 4 udp 111 portmapper 10 3 udp 111 portmapper 10 2 udp 111 portmapper 100021 3 udp 39618 nlockmgr 100021 3 tcp 41067 nlockmgr 100024 1 udp 33482 status 100024 1 tcp 37034 status You can do a netstat -anp | grep 39618 to see if there is some process really listening at the nlockmgr port. You can repeat this for port 41067. If there is some process, you should stop it. If there is no process listening on that port, you should remove it with a command like this: rpcinfo -d 100021 3 You must execute this command for all stale ports for any services other than portmapper and status. Once done you should get the output shown before. After that, you can try to start your volume and see if everything is registered (rpcinfo -p) and if gluster has started the nfs server (gluster volume status). If everything is ok, you should be able to mount the volume using NFS. Xavi On 01/27/2015 03:18 PM, David F. Robinson wrote: Turning off nfslock did not help. Also, still getting these messages every 3-seconds: [2015-01-27 14:16:12.921880] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:15.922431] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:18.923080] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:21.923748] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:24.924472] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:27.925192] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:30.925895] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:33.926563] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) [2015-01-27 14:16:36.927248] W [socket.c:611:__socket_rwv] 0-management: readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed (Invalid argument) -- Original Message -- From: Kaushal M kshlms...@gmail.com mailto:kshlms...@gmail.com To: David F. Robinson david.robin...@corvidtec.com mailto:david.robin...@corvidtec.com Cc: Joe Julian j...@julianfamily.org mailto:j...@julianfamily.org; Gluster Users gluster-us...@gluster.org mailto:gluster-us...@gluster.org; Gluster Devel