[Gluster-devel] GlusterFS Office Hours at FOSDEM

2015-01-27 Thread Kaleb KEITHLEY

Hi,

Lalatendu Mohanty, Niels de Vos, and myself will be holding GlusterFS 
Office Hours at FOSDEM.


Look for us at the CentOS booth, from 16h00 to 17h00 on Saturday, 31 
January.


FOSDEM is taking place this weekend, 31 Jan and 1 Feb, at ULB Solbosch 
Campus, Brussels. FOSDEM is a free event, no registration is necessary. 
For more info see https://fosdem.org/2015/


Hope to see you there.

--

Kaleb
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] GSOC-2015 heads up

2015-01-27 Thread Kaushal M
Hi all,
As it happens every year, the processes for GSOC begin in February,
starting with the organization registration, which runs this year from
February 9th to 20th. Rest of the timeline can be viewed at [1].

Last year we participated in GSOC under the Fedora organization, we hadn't
registered as an organization in time. The Fedora organization graciously
gave us one of their slots, and allowed us to become a part of GSOC for the
first time. Vipul Nayyar implemented glusterfsiostat [2], under Krishnan
P's mentorship. He created a tool to obtain and visualize I/O stats on
GlusterFS clients, and also contributed improvements to GlusterFS code,
specifically to the io-stats xlator. He successfully completed this
project, and along the way became more involved with our community and is
currently helping maintain the Gluster community infrastructure.

This year we need to aim higher, and have multiple contributions coming in.
This first and foremost requires our community to register itself as an
organization, as we cannot continue to ask the Fedora community to lend us
their valuable slots. We can discuss the details regarding the registration
and further steps during the upcoming community meeting.

Following this single success from last year, let's aim for multiple
successes this year and grow our community.

Cheers!

~kaushal

[1]: https://www.google-melange.com/gsoc/homepage/google/gsoc2015
[2]: https://forge.gluster.org/glusterfsiostat
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] GlusterFS 4.0 updates needed

2015-01-27 Thread Dan Lambright
data classification has been updated to better reflect cache tiering's current 
status. 

- Original Message -
 From: Jeff Darcy jda...@redhat.com
 To: Gluster Devel gluster-devel@gluster.org
 Sent: Monday, January 26, 2015 7:53:37 AM
 Subject: [Gluster-devel] GlusterFS 4.0 updates needed
 
 Ahead of Wednesday's community meeting, I'd like to get as much 4.0
 status together as possible.  Developers, please check out the
 sub-projects on this page:
 
 http://www.gluster.org/community/documentation/index.php/Planning40
 
 If you're involved in any of those, please update the status section
 of the corresponding feature page (near the bottom) as appropriate.
 Links to design notes, email threads, code review requests, etc. are
 most welcome.  Yes, even things that have been sent to this list before.
 I'll be setting up a survey to choose a time for an online (Hangout/IRC)
 summit in the first week of February.  If we can separate the items
 that are truly being worked on vs. those that just seemed cool at the
 time but haven't received one minute of anyone's attention since (e.g.
 sharding) then we can keep that meeting focused and productive.  Thanks!
 ___
 Gluster-devel mailing list
 Gluster-devel@gluster.org
 http://www.gluster.org/mailman/listinfo/gluster-devel
 
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Bug in EXPECT family of macros in test framework

2015-01-27 Thread Niels de Vos
On Tue, Jan 27, 2015 at 12:35:35PM +0100, Xavier Hernandez wrote:
 Hi,
 
 I've just discovered that EXPECT, EXPECT_WITHIN and others have a bug that
 can hide some failures.
 
 The problem is that the test is made using a regular expression comparison.
 For example, EXPECT checks this:
 
 if ! [[ $a =~ $e ]]; then
 
 There are some tests that use a regular expression as the first argument to
 EXPECT, however there are many other tests that simply use a number or a
 text.
 
 This can cause problems. For example if the test is:
 
 EXPECT 0 test
 
 And the test returns 10, the regular expression 0 *will* match with 10,
 so the test will be considered successful when it shouldn't.
 
 It will also report success for test cases that check for an empty result:
 
 EXPECT  test
 
 This will always succeed, even if the test returns something.
 
 What would be the best approach to solve this ?

Oh, yuck!

I think we should test like this, untested: [[ $a =~ ^$e\$ ]]

It might break some regular expressions, the expressions are now forced
to be very strict and should match the whole line.

Thoughts?
Niels


pgpXr35MHc8Sw.pgp
Description: PGP signature
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] Current state of multi-thread epoll (and review request)

2015-01-27 Thread Shyam

Hi,

Here is the current state of multi-thread epoll patch,

1) Patches:
epoll patch: http://review.gluster.org/#/c/3842/
epoll configuration patch: http://review.gluster.org/#/c/9488/

2) Failures:
- epoll patch
  - In 2 runs the following 2 test cases failed in separate instances,
- ./tests/bugs/quota/bug-1038598.t
  - Under investigation
- ./tests/bugs/glusterd/bug-1104642.t
  - This is possibly a TC issue, as we check peer count and then 
immediately check for the configuration update. A peer count does not 
guarantee that the peer has also updated configuration. I will be 
sending a patch (today) to address this test case.


- epoll configuration
 - Overall this patch passed, could be a one off as it depends on the 
above patch, or could be due to the fact that this makes glusterd and 
the CLI single threaded, and hence we do not hit the above failures.


- NetBSD
  - NetBSD fails badly with the above patch
- Under investigation

3) Requests
- Request that people review the above patches, so that we can get 
concerns addressed quicker


- The new patch on the configuration of epoll threads, needs some 
attention in terms of correctness of layers where this is allowed to be 
configured (among others) etc. Request reviewers to take a look at this.


Thanks,
Shyam
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] [Gluster-users] lockd: server not responding, timed out

2015-01-27 Thread Peter Auyeung
Hi Niels,

I see no kernel NFS service running on the gluster node.

Here are the output

root@glusterprod001:~# rpcinfo
   program version netid addressserviceowner
104tcp6  ::.0.111   portmapper superuser
103tcp6  ::.0.111   portmapper superuser
104udp6  ::.0.111   portmapper superuser
103udp6  ::.0.111   portmapper superuser
104tcp   0.0.0.0.0.111  portmapper superuser
103tcp   0.0.0.0.0.111  portmapper superuser
102tcp   0.0.0.0.0.111  portmapper superuser
104udp   0.0.0.0.0.111  portmapper superuser
103udp   0.0.0.0.0.111  portmapper superuser
102udp   0.0.0.0.0.111  portmapper superuser
104local /run/rpcbind.sock  portmapper superuser
103local /run/rpcbind.sock  portmapper superuser
153tcp   0.0.0.0.150.65 mountd superuser
151tcp   0.0.0.0.150.66 mountd superuser
133tcp   0.0.0.0.8.1nfssuperuser
1000214tcp   0.0.0.0.150.68 nlockmgr   superuser
1002273tcp   0.0.0.0.8.1-  superuser
1000211udp   0.0.0.0.2.215  nlockmgr   superuser
1000211tcp   0.0.0.0.2.217  nlockmgr   superuser
1000241udp   0.0.0.0.136.211status 105
1000241tcp   0.0.0.0.170.60 status 105
1000241udp6  ::.182.65  status 105
1000241tcp6  ::.172.250 status 105

root@glusterprod001:~# ss
State   Recv-Q Send-QLocal 
Address:PortPeer Address:Port
ESTAB   0  0 
10.101.165.61:1015   10.101.165.63:24007
ESTAB   0  0 
10.101.165.61:93610.101.165.61:49156
ESTAB   0  0 
10.101.165.61:1012   10.101.165.66:24007
ESTAB   0  0 
10.101.165.61:49157  10.101.165.64:987
ESTAB   0  0 
10.101.165.61:99910.101.165.62:49153
ESTAB   0  0 
10.101.165.61:49157  10.101.165.66:988
ESTAB   0  0 
10.101.165.61:49155  10.101.165.62:834
ESTAB   0  0 
10.101.165.61:49156  10.101.165.63:998
ESTAB   0  0 
10.101.165.61:91210.101.165.65:49153
ESTAB   0  0 
127.0.0.1:982127.0.0.1:24007
ESTAB   0  0 
10.101.165.61:49156  10.101.165.65:997
ESTAB   0  0 
10.101.165.61:49155  10.101.165.61:850
ESTAB   0  0 
10.101.165.61:92210.101.165.62:49154
ESTAB   0  0 
10.101.165.61:89610.101.165.65:49154
ESTAB   0  0 
10.101.165.61:1010   10.101.165.61:24007
ESTAB   0  0 
10.101.165.61:imaps  10.101.165.61:49156
ESTAB   0  0 
10.101.165.61:49155  10.101.165.63:981
ESTAB   0  0 
10.101.165.61:93010.101.165.64:49155
ESTAB   0  0 
10.101.165.61:4379   10.101.165.65:44899
ESTAB   0  0 
10.101.165.61:98310.101.165.61:49157
ESTAB   0  0   

Re: [Gluster-devel] Ability to turn off 'glusterfs' protocol

2015-01-27 Thread Niels de Vos
On Tue, Jan 27, 2015 at 02:10:17AM +0100, Csaba Henk wrote:
 Hi Niels,
 
 On Mon, Jan 26, 2015 at 10:07 AM, Niels de Vos nde...@redhat.com wrote:
  On Sun, Jan 25, 2015 at 10:08:20PM +0530, Ramana Raja wrote:
  1) How would the above suggestion impact gfapi access?
 
  gfapi uses the GlusterFS protocol. When access through the protocol is
  denied, only clients from within the trusted storage pool can use it.
  gfapi is just a client, similar to the FUSE mount.
 
  2) Would the list of trusted clients be configured via gluster volume 
  set?
 
  There are the 'auth.allow' and 'auth.reject' volume options. I think
  they would be used as an alternative to a turn off glusterfs protocol
  feature.
 
 Does it mean that the implementation of feature would essentially boil down
 to an auth ruleset calculated by glusterfs?

I guess that depends on the goal of the feature. Where does the need
arrise to turn off the glusterfs protocol? Should nothing outside of
the trusted storage pool be able to connect to the bricks? This would
effectively only allow NFS/Samba/... when the service is located on a
system that is part of the trusted storage pool.

If you need a more fine-grained permission scheme, auth.allow and
auth.reject might be more appropriate. Or, use SSL authentication?

HTH,
Niels


pgpPNrRk9vj_K.pgp
Description: PGP signature
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] GlusterFS 3.7 updates needed

2015-01-27 Thread Vijay Bellur

Hi All,

As we approach feature freeze for 3.7 (28th Feb), it would great to have 
a status update for features being worked on for the release.


Action(s) needed for status updates:

1. If you are the owner for one of the features listed in the planning 
page [1], can you please update status of the implementation and also 
include any relevant mailing list, IRC discussions and design decisions 
that can help in providing more clarity on the feature?


2. If you are already working on a feature for 3.7 and have not yet 
opened a new feature page as yet or linked the feature page in the 
planning page, can you please ensure that we have one listed by Feb 10th 
(i.e. 2 weeks from now)? It might be hard to accommodate new feature 
proposals and reviews for such features if we don't have an appropriate 
feature page linked in the release planning page by Feb 10th.


3. If you have planned a feature, are blocked and need more assistance 
to complete it by the feature freeze date, please let us know.


Thanks,
Vijay

[1] http://www.gluster.org/community/documentation/index.php/Planning37


___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] GlusterFS 3.7 updates needed

2015-01-27 Thread Anders Blomdell
On 2015-01-27 11:49, Vijay Bellur wrote:
 Hi All,
 
 As we approach feature freeze for 3.7 (28th Feb), it would great to
 have a status update for features being worked on for the release.
Not exactly a new feature, but something I would like to see happen:

[Gluster-devel] Glusterd daemon management code refactoring
http://www.gluster.org/pipermail/gluster-devel/2014-December/043180.html

/Anders
-- 
Anders Blomdell  Email: anders.blomd...@control.lth.se
Department of Automatic Control
Lund University  Phone:+46 46 222 4625
P.O. Box 118 Fax:  +46 46 138118
SE-221 00 Lund, Sweden

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] GlusterFS 3.7 updates needed

2015-01-27 Thread Krishnan Parthasarathi
Glusterd daemon management code refactoring is being worked on.
See http://review.gluster.org/9428 (and dependent patches) for
current status. I have added you to the list of reviewers. Hope that
is OK with you.

- Original Message -
 On 2015-01-27 11:49, Vijay Bellur wrote:
  Hi All,
  
  As we approach feature freeze for 3.7 (28th Feb), it would great to
  have a status update for features being worked on for the release.
 Not exactly a new feature, but something I would like to see happen:
 
 [Gluster-devel] Glusterd daemon management code refactoring
 http://www.gluster.org/pipermail/gluster-devel/2014-December/043180.html
 
 /Anders
 --
 Anders Blomdell  Email: anders.blomd...@control.lth.se
 Department of Automatic Control
 Lund University  Phone:+46 46 222 4625
 P.O. Box 118 Fax:  +46 46 138118
 SE-221 00 Lund, Sweden
 
 
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] Bug in EXPECT family of macros in test framework

2015-01-27 Thread Xavier Hernandez

Hi,

I've just discovered that EXPECT, EXPECT_WITHIN and others have a bug 
that can hide some failures.


The problem is that the test is made using a regular expression 
comparison. For example, EXPECT checks this:


if ! [[ $a =~ $e ]]; then

There are some tests that use a regular expression as the first argument 
to EXPECT, however there are many other tests that simply use a number 
or a text.


This can cause problems. For example if the test is:

EXPECT 0 test

And the test returns 10, the regular expression 0 *will* match with 
10, so the test will be considered successful when it shouldn't.


It will also report success for test cases that check for an empty result:

EXPECT  test

This will always succeed, even if the test returns something.

What would be the best approach to solve this ?

Xavi
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] Quota with hard-links

2015-01-27 Thread Vijaikumar M

Hi All,

This is regarding quota accounting for hard-links. Currently, accounting 
is done only once for the links created within the same directory and 
accounting is done separately when links are created in separate directory.
With this approach account may go wrong when rename is performed on the 
hardlink files across directories.


We are implementing one of the below mentioned policy for hard-links 
when quota is enabled.


1) Allow creating hard-links only within same directory.
(We can hit the same problem if quota is enabled on the 
pre-existing data which contains hard-links)


2) Allow creating hard-links only within the same branch where the limit 
is set
(We can hit the same problem if quota is enabled on the 
pre-existing data which contains hard-links and also when quota-limit is 
set/unset)


3) Account for all the hard-links.
(Marker treats all hard-links as a new file)

Please provide your feedback.

Thanks,
Vijay

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] quota and snapshot testcase failure (zfs on CentOS 6.6)

2015-01-27 Thread Vijaikumar M

Hi Kiran,

Testcase './tests/basic/quota-anon-fd-nfs.t' is removed from the test 
suite.There are some issues with this testcase, we are working on it. We 
will add this test-case soon once the issue is fixed.


Thanks,
Vijay


On Tuesday 27 January 2015 06:11 PM, Vijaikumar M wrote:

Hi Kiran,

quota.t failure issue has been fixed with patch 
http://review.gluster.org/#/c/9410/. Can you please re-try the test 
with this patch and see if it works?


Thanks,
Vijay


On Wednesday 19 November 2014 10:32 AM, Pranith Kumar Karampuri wrote:


On 11/19/2014 10:30 AM, Atin Mukherjee wrote:


On 11/18/2014 10:35 PM, Pranith Kumar Karampuri wrote:

On 11/12/2014 04:52 PM, Kiran Patil wrote:

I have create zpool with name d and mnt and they appear in filesystem
as follows.

d on /d type zfs (rw,xattr)
mnt on /mnt type zfs (rw,xattr)

Debug enabled output of quota.t testcase is at http://ur1.ca/irbt1.

CC vijaikumar

quota-anon-fd-nfs.t spurious failure fix is addressed by
http://review.gluster.org/#/c/9108/

This is just quota.t in tests/basic, not the anon-fd one

Pranith

~Atin

On Wed, Nov 12, 2014 at 3:22 PM, Kiran Patil ki...@fractalio.com
mailto:ki...@fractalio.com wrote:

 Hi,

 Gluster suite report,

 Gluster version: glusterfs 3.6.1

 On disk filesystem: Zfs 0.6.3-1.1

 Operating system: CentOS release 6.6 (Final)

 We are seeing quota and snapshot testcase failures.

 We are not sure why quota is failing since quotas worked fine on
 gluster 3.4.

 Test Summary Report
 ---
 ./tests/basic/quota-anon-fd-nfs.t  (Wstat: 0 Tests: 16 
Failed: 1)

   Failed test:  16
 ./tests/basic/quota.t  (Wstat: 0 Tests: 73 Failed: 4)
   Failed tests:  24, 28, 32, 65
 ./tests/basic/uss.t  (Wstat: 0 Tests: 147 Failed: 78)
   Failed tests:  8-11, 16-25, 28-29, 31-32, 39-40, 45-47
 49-57, 60-61, 63-64, 71-72, 78-87, 90-91
 93-94, 101-102, 107-115, 118-119, 121-122
 129-130, 134, 136-137, 139-140, 142-143
 145-146
 ./tests/basic/volume-snapshot.t  (Wstat: 0 Tests: 30 Failed: 12)
   Failed tests:  11-18, 21-24
 ./tests/basic/volume-status.t  (Wstat: 0 Tests: 14 Failed: 1)
   Failed test:  14
 ./tests/bugs/bug-1023974.t   (Wstat: 0 Tests: 15 Failed: 1)
   Failed test:  12
 ./tests/bugs/bug-1038598.t   (Wstat: 0 Tests: 28 Failed: 6)
   Failed tests:  17, 21-22, 26-28
 ./tests/bugs/bug-1045333.t   (Wstat: 0 Tests: 16 Failed: 9)
   Failed tests:  7-15
 ./tests/bugs/bug-1049834.t   (Wstat: 0 Tests: 18 Failed: 7)
   Failed tests:  11-14, 16-18
 ./tests/bugs/bug-1087203.t   (Wstat: 0 Tests: 43 Failed: 2)
   Failed tests:  31, 41
 ./tests/bugs/bug-1090042.t   (Wstat: 0 Tests: 12 Failed: 3)
   Failed tests:  9-11
 ./tests/bugs/bug-1109770.t   (Wstat: 0 Tests: 19 Failed: 4)
   Failed tests:  8-11
 ./tests/bugs/bug-1109889.t   (Wstat: 0 Tests: 20 Failed: 4)
   Failed tests:  8-11
 ./tests/bugs/bug-1112559.t   (Wstat: 0 Tests: 11 Failed: 3)
   Failed tests:  8-9, 11
 ./tests/bugs/bug-1112613.t   (Wstat: 0 Tests: 22 Failed: 5)
   Failed tests:  12-14, 17-18
 ./tests/bugs/bug-1113975.t   (Wstat: 0 Tests: 13 Failed: 4)
   Failed tests:  8-9, 11-12
 ./tests/bugs/bug-847622.t  (Wstat: 0 Tests: 10 Failed: 1)
   Failed test:  8
 ./tests/bugs/bug-861542.t  (Wstat: 0 Tests: 13 Failed: 7)
   Failed tests:  7-13
 ./tests/features/ssl-authz.t   (Wstat: 0 Tests: 18 Failed: 1)
   Failed test:  18
 Files=277, Tests=7908, 8147 wallclock secs ( 4.56 usr 0.78 sys +
 774.74 cusr 666.97 csys = 1447.05 CPU)
 Result: FAIL

 Thanks,
 Kiran.




___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel




___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel







___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] quota and snapshot testcase failure (zfs on CentOS 6.6)

2015-01-27 Thread Vijaikumar M

Hi Kiran,

quota.t failure issue has been fixed with patch 
http://review.gluster.org/#/c/9410/. Can you please re-try the test with 
this patch and see if it works?


Thanks,
Vijay


On Wednesday 19 November 2014 10:32 AM, Pranith Kumar Karampuri wrote:


On 11/19/2014 10:30 AM, Atin Mukherjee wrote:


On 11/18/2014 10:35 PM, Pranith Kumar Karampuri wrote:

On 11/12/2014 04:52 PM, Kiran Patil wrote:

I have create zpool with name d and mnt and they appear in filesystem
as follows.

d on /d type zfs (rw,xattr)
mnt on /mnt type zfs (rw,xattr)

Debug enabled output of quota.t testcase is at http://ur1.ca/irbt1.

CC vijaikumar

quota-anon-fd-nfs.t spurious failure fix is addressed by
http://review.gluster.org/#/c/9108/

This is just quota.t in tests/basic, not the anon-fd one

Pranith

~Atin

On Wed, Nov 12, 2014 at 3:22 PM, Kiran Patil ki...@fractalio.com
mailto:ki...@fractalio.com wrote:

 Hi,

 Gluster suite report,

 Gluster version: glusterfs 3.6.1

 On disk filesystem: Zfs 0.6.3-1.1

 Operating system: CentOS release 6.6 (Final)

 We are seeing quota and snapshot testcase failures.

 We are not sure why quota is failing since quotas worked fine on
 gluster 3.4.

 Test Summary Report
 ---
 ./tests/basic/quota-anon-fd-nfs.t  (Wstat: 0 Tests: 16 Failed: 1)
   Failed test:  16
 ./tests/basic/quota.t  (Wstat: 0 Tests: 73 Failed: 4)
   Failed tests:  24, 28, 32, 65
 ./tests/basic/uss.t  (Wstat: 0 Tests: 147 Failed: 78)
   Failed tests:  8-11, 16-25, 28-29, 31-32, 39-40, 45-47
 49-57, 60-61, 63-64, 71-72, 78-87, 90-91
 93-94, 101-102, 107-115, 118-119, 121-122
 129-130, 134, 136-137, 139-140, 142-143
 145-146
 ./tests/basic/volume-snapshot.t  (Wstat: 0 Tests: 30 Failed: 12)
   Failed tests:  11-18, 21-24
 ./tests/basic/volume-status.t  (Wstat: 0 Tests: 14 Failed: 1)
   Failed test:  14
 ./tests/bugs/bug-1023974.t   (Wstat: 0 Tests: 15 Failed: 1)
   Failed test:  12
 ./tests/bugs/bug-1038598.t   (Wstat: 0 Tests: 28 Failed: 6)
   Failed tests:  17, 21-22, 26-28
 ./tests/bugs/bug-1045333.t   (Wstat: 0 Tests: 16 Failed: 9)
   Failed tests:  7-15
 ./tests/bugs/bug-1049834.t   (Wstat: 0 Tests: 18 Failed: 7)
   Failed tests:  11-14, 16-18
 ./tests/bugs/bug-1087203.t   (Wstat: 0 Tests: 43 Failed: 2)
   Failed tests:  31, 41
 ./tests/bugs/bug-1090042.t   (Wstat: 0 Tests: 12 Failed: 3)
   Failed tests:  9-11
 ./tests/bugs/bug-1109770.t   (Wstat: 0 Tests: 19 Failed: 4)
   Failed tests:  8-11
 ./tests/bugs/bug-1109889.t   (Wstat: 0 Tests: 20 Failed: 4)
   Failed tests:  8-11
 ./tests/bugs/bug-1112559.t   (Wstat: 0 Tests: 11 Failed: 3)
   Failed tests:  8-9, 11
 ./tests/bugs/bug-1112613.t   (Wstat: 0 Tests: 22 Failed: 5)
   Failed tests:  12-14, 17-18
 ./tests/bugs/bug-1113975.t   (Wstat: 0 Tests: 13 Failed: 4)
   Failed tests:  8-9, 11-12
 ./tests/bugs/bug-847622.t  (Wstat: 0 Tests: 10 Failed: 1)
   Failed test:  8
 ./tests/bugs/bug-861542.t  (Wstat: 0 Tests: 13 Failed: 7)
   Failed tests:  7-13
 ./tests/features/ssl-authz.t   (Wstat: 0 Tests: 18 Failed: 1)
   Failed test:  18
 Files=277, Tests=7908, 8147 wallclock secs ( 4.56 usr 0.78 sys +
 774.74 cusr 666.97 csys = 1447.05 CPU)
 Result: FAIL

 Thanks,
 Kiran.




___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel




___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel





___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] glusterfs-3.6.2 GA

2015-01-27 Thread Kaleb KEITHLEY


glusterfs-3.6.2 has been released.

The release source tar file and packages for Fedora {20,21,rawhide}, 
RHEL/CentOS {5,6,7}, Debian {wheezy,jessie}, Pidora2014, and Raspbian 
wheezy are available at 
http://download.gluster.org/pub/gluster/glusterfs/3.6/3.6.2/


(Ubuntu packages will be available soon.)

This release fixes the following bugs. Thanks to all who submitted bugs 
and patches and reviewed the changes.



 1184191 - Cluster/DHT : Fixed crash due to null deref
 1180404 - nfs server restarts when a snapshot is deactivated
 1180411 - CIFS:[USS]: glusterfsd OOM killed when 255 snapshots were
browsed at CIFS mount and Control+C is issued
 1180070 - [AFR] getfattr on fuse mount gives error : Software caused
connection abort
 1175753 - [readdir-ahead]: indicate EOF for readdirp
 1175752 - [USS]: On a successful lookup, snapd logs are filled with
Warnings dict OR key (entry-point) is NULL
 1175749 - glusterfs client crashed while migrating the fds
 1179658 - Add brick fails if parent dir of new brick and existing
brick is same and volume was accessed using libgfapi and smb.
 1146524 - glusterfs.spec.in - synch minor diffs with fedora dist-git
glusterfs.spec
 1175744 - [USS]: Unable to access .snaps after snapshot restore after
directories were deleted and recreated
 1175742 - [USS]: browsing .snaps directory with CIFS fails with
Invalid argument
 1175739 - [USS]: Non root user who has no access to a directory, from
NFS mount, is able to access the files under .snaps under that directory
 1175758 - [USS] : Rebalance process tries to connect to snapd and in
case when snapd crashes it might affect rebalance process
 1175765 - USS]: When snapd is crashed gluster volume stop/delete
operation fails making the cluster in inconsistent state
 1173528 - Change in volume heal info command output
 1166515 - [Tracker] RDMA support in glusterfs
 1166505 - mount fails for nfs protocol in rdma volumes
 1138385 - [DHT:REBALANCE]: Rebalance failures are seen with error
message  remote operation failed: File exists
 1177418 - entry self-heal in 3.5 and 3.6 are not compatible
 1170954 - Fix mutex problems reported by coverity scan
 1177899 - nfs: ls shows Permission denied with root-squash
 1175738 - [USS]: data unavailability for a period of time when USS is
enabled/disabled
 1175736 - [USS]:After deactivating a snapshot trying to access the
remaining activated snapshots from NFS mount gives 'Invalid argument' error
 1175735 - [USS]: snapd process is not killed once the glusterd comes back
 1175733 - [USS]: If the snap name is same as snap-directory than cd to
virtual snap directory fails
 1175756 - [USS] : Snapd crashed while trying to access the snapshots
under .snaps directory
 1175755 - SNAPSHOT[USS]:gluster volume set for uss doesnot check any
boundaries
 1175732 - [SNAPSHOT]: nouuid is appended for every snapshoted brick
which causes duplication if the original brick has already nouuid
 1175730 - [USS]: creating file/directories under .snaps shows wrong
error message
 1175754 - [SNAPSHOT]: before the snap is marked to be deleted if the
node goes down than the snaps are propagated on other nodes and glusterd
hungs
 1159484 - ls -alR can not heal the disperse volume
 1138897 - NetBSD port
 1175728 - [USS]: All uss related logs are reported under
/var/log/glusterfs, it makes sense to move it into subfolder
 1170548 - [USS] : don't display the snapshots which are not activated
 1170921 - [SNAPSHOT]: snapshot should be deactivated by default when
created
 1175694 - [SNAPSHOT]: snapshoted volume is read only but it shows rw
attributes in mount
 1161885 - Possible file corruption on dispersed volumes
 1170959 - EC_MAX_NODES is defined incorrectly
 1175645 - [USS]: Typo error in the description for USS under gluster
volume set help
 1171259 - mount.glusterfs does not understand -n option

Regards,
Kaleb, on behalf of Raghavendra Bhat, who did all the work.



___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] Minutes of todays Gluster Community Bug Triage meeting

2015-01-27 Thread Niels de Vos
On Tue, Jan 27, 2015 at 12:07:35PM +0100, Niels de Vos wrote:
 Hi all,
 
 This meeting is scheduled for anyone that is interested in learning more
 about, or assisting with the Bug Triage.
 
 Meeting details:
 - location: #gluster-meeting on Freenode IRC
 - date: every Tuesday
 - time: 12:00 UTC, 13:00 CET (run: date -d 12:00 UTC)
 - agenda: https://public.pad.fsfe.org/p/gluster-bug-triage
 
 Currently the following items are listed:
 * Roll Call
 * Status of last weeks action items
 * Group Triage
 * Open Floor
 
 The last two topics have space for additions. If you have a suitable bug
 or topic to discuss, please add it to the agenda.

Minutes: 
http://meetbot.fedoraproject.org/gluster-meeting/2015-01-27/gluster-meeting.2015-01-27-12.03.html
Minutes (text): 
http://meetbot.fedoraproject.org/gluster-meeting/2015-01-27/gluster-meeting.2015-01-27-12.03.txt
Log: 
http://meetbot.fedoraproject.org/gluster-meeting/2015-01-27/gluster-meeting.2015-01-27-12.03.log.html

Meeting started by ndevos at 12:03:35 UTC.

Meeting summary
---
* Roll Call  (ndevos, 12:04:07)

* Action items from last week  (ndevos, 12:06:26)

* hagarth will look for somebody that can act like a bug assigner
  manager kind of person  (ndevos, 12:06:31)

* hagarth should update the MAINTAINERS file, add current maintainers
  and new components like Snapshot  (ndevos, 12:08:12)

* lalatenduM will send a reminder to the users- and devel- ML about (and
  how to) fixing Coverity defects  (ndevos, 12:09:09)

* lalatenduM initiate a discussion with the RH Gluster team to triage
  their own bugs when they report them  (ndevos, 12:09:53)

* ndevos adds NetBSD queries to the bug triage wiki page  (ndevos,
  12:10:49)

* hagarth start/sync email on regular (nightly) automated tests
  (ndevos, 12:11:50)

* lalatenduM will look into using nightly builds for automated testing,
  and will report issues/success to the mailinglist  (ndevos, 12:12:42)

* Group Triage  (ndevos, 12:16:20)

* Open Floor  (ndevos, 12:31:24)

Meeting ended at 12:41:57 UTC.


People Present (lines said)
---
* ndevos (45)
* hagarth (18)
* kkeithley_ (13)
* hchiramm (10)
* gothos (7)
* zodbot (2)


pgpFpSJW82D3j.pgp
Description: PGP signature
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] [Gluster-users] v3.6.2

2015-01-27 Thread David F. Robinson
After shutting down all NFS and gluster processes, there was still an 
NFS process.


[root@gfs01bkp ~]# rpcinfo -p
   program vers proto   port  service
104   tcp111  portmapper
103   tcp111  portmapper
102   tcp111  portmapper
104   udp111  portmapper
103   udp111  portmapper
102   udp111  portmapper
153   tcp  38465  mountd
151   tcp  38466  mountd
133   tcp   2049  nfs
1000241   udp  34738  status
1000241   tcp  37269  status
[root@gfs01bkp ~]# netstat -anp | grep 2049
[root@gfs01bkp ~]# netstat -anp | grep 38465
[root@gfs01bkp ~]# netstat -anp | grep 38466

I killed off the processes using rpcinfo -d

[root@gfs01bkp ~]# rpcinfo -p
   program vers proto   port  service
104   tcp111  portmapper
103   tcp111  portmapper
102   tcp111  portmapper
104   udp111  portmapper
103   udp111  portmapper
102   udp111  portmapper
1000241   udp  34738  status
1000241   tcp  37269  status

Then I restarted the glusterd and did a 'mount -a'.  Worked perfectly.  
And the errors that were showing up in the logs every 3-seconds stopped.


Thanks for your help.  Greatly appreciated.

David




-- Original Message --
From: Xavier Hernandez xhernan...@datalab.es
To: David F. Robinson david.robin...@corvidtec.com; Kaushal M 
kshlms...@gmail.com
Cc: Gluster Users gluster-us...@gluster.org; Gluster Devel 
gluster-devel@gluster.org

Sent: 1/27/2015 10:02:31 AM
Subject: Re: [Gluster-devel] [Gluster-users] v3.6.2


Hi,

I had a similar problem once. It happened after doing some unrelated 
tests with NFS. I thought it was a problem I generated doing weird 
things, so I didn't investigate the cause further.


To see if this is the same case, try this:

* Unmount all NFS mounts and stop all gluster volumes
* Check that there are no gluster processes running (ps ax | grep 
gluster), specially any glusterfs. glusterd is ok.

* Check that there are no NFS processes running (ps ax | grep nfs)
* Check with 'rpcinfo -p' that there's no nfs service registered

The output should be similar to this:

   program vers proto port service
10 4 tcp 111 portmapper
10 3 tcp 111 portmapper
10 2 tcp 111 portmapper
10 4 udp 111 portmapper
10 3 udp 111 portmapper
10 2 udp 111 portmapper
100024 1 udp 33482 status
100024 1 tcp 37034 status

If there are more services registered, you can directly delete them or 
check if they correspond to an active process. For example, if the 
output is this:


   program vers proto port service
10 4 tcp 111 portmapper
10 3 tcp 111 portmapper
10 2 tcp 111 portmapper
10 4 udp 111 portmapper
10 3 udp 111 portmapper
10 2 udp 111 portmapper
100021 3 udp 39618 nlockmgr
100021 3 tcp 41067 nlockmgr
100024 1 udp 33482 status
100024 1 tcp 37034 status

You can do a netstat -anp | grep 39618 to see if there is some 
process really listening at the nlockmgr port. You can repeat this for 
port 41067. If there is some process, you should stop it. If there is 
no process listening on that port, you should remove it with a command 
like this:


rpcinfo -d 100021 3

You must execute this command for all stale ports for any services 
other than portmapper and status. Once done you should get the output 
shown before.


After that, you can try to start your volume and see if everything is 
registered (rpcinfo -p) and if gluster has started the nfs server 
(gluster volume status).


If everything is ok, you should be able to mount the volume using NFS.

Xavi

On 01/27/2015 03:18 PM, David F. Robinson wrote:

Turning off nfslock did not help. Also, still getting these messages
every 3-seconds:

[2015-01-27 14:16:12.921880] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:15.922431] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:18.923080] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:21.923748] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:24.924472] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:27.925192] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:30.925895] W [socket.c:611:__socket_rwv] 
0-management:

readv on 

Re: [Gluster-devel] [Gluster-users] v3.6.2

2015-01-27 Thread David F. Robinson
I rebooted the machine to see if the problem would return and it does.  
Same issue after a reboot.

Any suggestions?

One other thing I tested was to comment out the NFS mounts in 
/etc/fstab:
# gfsib01bkp.corvidtec.com:/homegfs_bkp /backup_nfs/homegfs nfs 
vers=3,intr,bg,rsize=32768,wsize=32768 0 0
After the machine comes back up, I remove the comment and do a 'mount 
-a'.  The mount works fine.


It looks like it is a timing during startup issue.  Is it trying to do 
the NFS mount while glusterd is still starting up?


David


-- Original Message --
From: Xavier Hernandez xhernan...@datalab.es
To: David F. Robinson david.robin...@corvidtec.com; Kaushal M 
kshlms...@gmail.com
Cc: Gluster Users gluster-us...@gluster.org; Gluster Devel 
gluster-devel@gluster.org

Sent: 1/27/2015 10:02:31 AM
Subject: Re: [Gluster-devel] [Gluster-users] v3.6.2


Hi,

I had a similar problem once. It happened after doing some unrelated 
tests with NFS. I thought it was a problem I generated doing weird 
things, so I didn't investigate the cause further.


To see if this is the same case, try this:

* Unmount all NFS mounts and stop all gluster volumes
* Check that there are no gluster processes running (ps ax | grep 
gluster), specially any glusterfs. glusterd is ok.

* Check that there are no NFS processes running (ps ax | grep nfs)
* Check with 'rpcinfo -p' that there's no nfs service registered

The output should be similar to this:

   program vers proto port service
10 4 tcp 111 portmapper
10 3 tcp 111 portmapper
10 2 tcp 111 portmapper
10 4 udp 111 portmapper
10 3 udp 111 portmapper
10 2 udp 111 portmapper
100024 1 udp 33482 status
100024 1 tcp 37034 status

If there are more services registered, you can directly delete them or 
check if they correspond to an active process. For example, if the 
output is this:


   program vers proto port service
10 4 tcp 111 portmapper
10 3 tcp 111 portmapper
10 2 tcp 111 portmapper
10 4 udp 111 portmapper
10 3 udp 111 portmapper
10 2 udp 111 portmapper
100021 3 udp 39618 nlockmgr
100021 3 tcp 41067 nlockmgr
100024 1 udp 33482 status
100024 1 tcp 37034 status

You can do a netstat -anp | grep 39618 to see if there is some 
process really listening at the nlockmgr port. You can repeat this for 
port 41067. If there is some process, you should stop it. If there is 
no process listening on that port, you should remove it with a command 
like this:


rpcinfo -d 100021 3

You must execute this command for all stale ports for any services 
other than portmapper and status. Once done you should get the output 
shown before.


After that, you can try to start your volume and see if everything is 
registered (rpcinfo -p) and if gluster has started the nfs server 
(gluster volume status).


If everything is ok, you should be able to mount the volume using NFS.

Xavi

On 01/27/2015 03:18 PM, David F. Robinson wrote:

Turning off nfslock did not help. Also, still getting these messages
every 3-seconds:

[2015-01-27 14:16:12.921880] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:15.922431] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:18.923080] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:21.923748] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:24.924472] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:27.925192] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:30.925895] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:33.926563] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:36.927248] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
-- Original Message --
From: Kaushal M kshlms...@gmail.com mailto:kshlms...@gmail.com
To: David F. Robinson david.robin...@corvidtec.com
mailto:david.robin...@corvidtec.com
Cc: Joe Julian j...@julianfamily.org mailto:j...@julianfamily.org;
Gluster Users gluster-us...@gluster.org
mailto:gluster-us...@gluster.org; Gluster Devel
gluster-devel@gluster.org mailto:gluster-devel@gluster.org
Sent: 1/27/2015 1:49:56 AM
Subject: Re: Re[2]: [Gluster-devel] 

Re: [Gluster-devel] [Gluster-users] v3.6.2

2015-01-27 Thread David F. Robinson

In my /etc/fstab, I have the following:

  gfsib01bkp.corvidtec.com:/homegfs_bkp  /backup/homegfs
glusterfs   transport=tcp,_netdev 0 0
  gfsib01bkp.corvidtec.com:/Software_bkp /backup/Software   
glusterfs   transport=tcp,_netdev 0 0
  gfsib01bkp.corvidtec.com:/Source_bkp   /backup/Source 
glusterfs   transport=tcp,_netdev 0 0


  #... Setup NFS mounts as well
  gfsib01bkp.corvidtec.com:/homegfs_bkp /backup_nfs/homegfs nfs 
vers=3,intr,bg,rsize=32768,wsize=32768 0 0



It looks like it is trying to start the nfs mount before gluster has 
finished coming up and that this is hanging the nfs ports.  I have 
_netdev in the glusterfs mount point to make sure the network has come 
up (including infiniband) prior to starting gluster.  Shouldn't the 
gluster init scripts check for gluster startup prior to starting the nfs 
mount?  It doesn't look like this is working properly.


David



-- Original Message --
From: Xavier Hernandez xhernan...@datalab.es
To: David F. Robinson david.robin...@corvidtec.com; Kaushal M 
kshlms...@gmail.com
Cc: Gluster Users gluster-us...@gluster.org; Gluster Devel 
gluster-devel@gluster.org

Sent: 1/27/2015 10:02:31 AM
Subject: Re: [Gluster-devel] [Gluster-users] v3.6.2


Hi,

I had a similar problem once. It happened after doing some unrelated 
tests with NFS. I thought it was a problem I generated doing weird 
things, so I didn't investigate the cause further.


To see if this is the same case, try this:

* Unmount all NFS mounts and stop all gluster volumes
* Check that there are no gluster processes running (ps ax | grep 
gluster), specially any glusterfs. glusterd is ok.

* Check that there are no NFS processes running (ps ax | grep nfs)
* Check with 'rpcinfo -p' that there's no nfs service registered

The output should be similar to this:

   program vers proto port service
10 4 tcp 111 portmapper
10 3 tcp 111 portmapper
10 2 tcp 111 portmapper
10 4 udp 111 portmapper
10 3 udp 111 portmapper
10 2 udp 111 portmapper
100024 1 udp 33482 status
100024 1 tcp 37034 status

If there are more services registered, you can directly delete them or 
check if they correspond to an active process. For example, if the 
output is this:


   program vers proto port service
10 4 tcp 111 portmapper
10 3 tcp 111 portmapper
10 2 tcp 111 portmapper
10 4 udp 111 portmapper
10 3 udp 111 portmapper
10 2 udp 111 portmapper
100021 3 udp 39618 nlockmgr
100021 3 tcp 41067 nlockmgr
100024 1 udp 33482 status
100024 1 tcp 37034 status

You can do a netstat -anp | grep 39618 to see if there is some 
process really listening at the nlockmgr port. You can repeat this for 
port 41067. If there is some process, you should stop it. If there is 
no process listening on that port, you should remove it with a command 
like this:


rpcinfo -d 100021 3

You must execute this command for all stale ports for any services 
other than portmapper and status. Once done you should get the output 
shown before.


After that, you can try to start your volume and see if everything is 
registered (rpcinfo -p) and if gluster has started the nfs server 
(gluster volume status).


If everything is ok, you should be able to mount the volume using NFS.

Xavi

On 01/27/2015 03:18 PM, David F. Robinson wrote:

Turning off nfslock did not help. Also, still getting these messages
every 3-seconds:

[2015-01-27 14:16:12.921880] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:15.922431] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:18.923080] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:21.923748] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:24.924472] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:27.925192] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:30.925895] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:33.926563] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:36.927248] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
-- Original Message --
From: Kaushal M kshlms...@gmail.com 

Re: [Gluster-devel] [Gluster-users] v3.6.2

2015-01-27 Thread David F. Robinson
Not elegant, but here is my short-term fix to prevent the issue after a 
reboot:


Added 'noauto' to the mount in /etc/rc.local:

/etc/fstab:
#... Note: Used the 'noauto' for the NFS mounts and put the mount in 
/etc/rc.local to ensure that
#... glsuter has been started before attempting to mount using NFS. 
Otherwise, hangs ports

#... during startup.
gfsib01bkp.corvidtec.com:/homegfs_bkp /backup_nfs/homegfs nfs 
vers=3,intr,bg,rsize=32768,wsize=32768,noauto 0 0
gfsib01a.corvidtec.com:/homegfs /homegfs_nfs nfs 
vers=3,intr,bg,rsize=32768,wsize=32768,noauto 0 0



/etc/rc.local:
/etc/init.d/glusterd restart
(sleep 20; mount -a; mount /backup_nfs/homegfs)



-- Original Message --
From: Xavier Hernandez xhernan...@datalab.es
To: David F. Robinson david.robin...@corvidtec.com; Kaushal M 
kshlms...@gmail.com
Cc: Gluster Users gluster-us...@gluster.org; Gluster Devel 
gluster-devel@gluster.org

Sent: 1/27/2015 10:02:31 AM
Subject: Re: [Gluster-devel] [Gluster-users] v3.6.2


Hi,

I had a similar problem once. It happened after doing some unrelated 
tests with NFS. I thought it was a problem I generated doing weird 
things, so I didn't investigate the cause further.


To see if this is the same case, try this:

* Unmount all NFS mounts and stop all gluster volumes
* Check that there are no gluster processes running (ps ax | grep 
gluster), specially any glusterfs. glusterd is ok.

* Check that there are no NFS processes running (ps ax | grep nfs)
* Check with 'rpcinfo -p' that there's no nfs service registered

The output should be similar to this:

   program vers proto port service
10 4 tcp 111 portmapper
10 3 tcp 111 portmapper
10 2 tcp 111 portmapper
10 4 udp 111 portmapper
10 3 udp 111 portmapper
10 2 udp 111 portmapper
100024 1 udp 33482 status
100024 1 tcp 37034 status

If there are more services registered, you can directly delete them or 
check if they correspond to an active process. For example, if the 
output is this:


   program vers proto port service
10 4 tcp 111 portmapper
10 3 tcp 111 portmapper
10 2 tcp 111 portmapper
10 4 udp 111 portmapper
10 3 udp 111 portmapper
10 2 udp 111 portmapper
100021 3 udp 39618 nlockmgr
100021 3 tcp 41067 nlockmgr
100024 1 udp 33482 status
100024 1 tcp 37034 status

You can do a netstat -anp | grep 39618 to see if there is some 
process really listening at the nlockmgr port. You can repeat this for 
port 41067. If there is some process, you should stop it. If there is 
no process listening on that port, you should remove it with a command 
like this:


rpcinfo -d 100021 3

You must execute this command for all stale ports for any services 
other than portmapper and status. Once done you should get the output 
shown before.


After that, you can try to start your volume and see if everything is 
registered (rpcinfo -p) and if gluster has started the nfs server 
(gluster volume status).


If everything is ok, you should be able to mount the volume using NFS.

Xavi

On 01/27/2015 03:18 PM, David F. Robinson wrote:

Turning off nfslock did not help. Also, still getting these messages
every 3-seconds:

[2015-01-27 14:16:12.921880] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:15.922431] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:18.923080] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:21.923748] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:24.924472] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:27.925192] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:30.925895] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:33.926563] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
[2015-01-27 14:16:36.927248] W [socket.c:611:__socket_rwv] 
0-management:

readv on /var/run/1f0cee5a2d074e39b32ee5a81c70e68c.socket failed
(Invalid argument)
-- Original Message --
From: Kaushal M kshlms...@gmail.com mailto:kshlms...@gmail.com
To: David F. Robinson david.robin...@corvidtec.com
mailto:david.robin...@corvidtec.com
Cc: Joe Julian j...@julianfamily.org mailto:j...@julianfamily.org;
Gluster Users gluster-us...@gluster.org
mailto:gluster-us...@gluster.org; Gluster Devel