Hello Carlos,

This is just one of the volumes that has the problem, but there are several :

[root@lpr-nas01 ~]# gluster vol info caviar_data11
Volume Name: caviar_data11
Type: Distribute
Volume ID: 4440ef2e-5d60-41a9-bac8-cf4751fc9be2
Status: Started
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: lpr-nas01:/brick-xiv/caviar_data11

[root@lpr-nas01 ~]# gluster vol status caviar_data11 detail
Status of volume: caviar_data11
------------------------------------------------------------------------------
Brick                : Brick lpr-nas01:/brick-xiv/caviar_data11
Port                 : 49186
Online               : Y
Pid                  : 4592
File System          : ext4
Device               : /dev/sdb
Mount Options        : rw
Inode Size           : 256
Disk Space Free      : 7.2TB
Total Disk Space     : 22.7TB
Inode Count          : 381786112
Free Inodes          : 115286880


Mounting with glusterfs works fine and I have no problems whatsoever.
Nfs service is switched off
No auto-heal is running
Name resolution shouldn't be a problem as the bricks are mounted on the same 
host running the gluster daemon. I also test the mount from that server.

[root@lpr-nas01 ~]# ping lpr-nas01
PING lpr-nas01.onprvp.fgov.be (192.168.151.21) 56(84) bytes of data.
64 bytes from lpr-nas01.onprvp.fgov.be (192.168.151.21): icmp_seq=1 ttl=64 
time=0.023 ms
[root@lpr-nas01 ~]# nslookup 192.168.151.21
Server:                 192.168.147.31
Address:             192.168.147.31#53

21.151.168.192.in-addr.arpa      name = lpr-nas01.onprvp.fgov.be.


The files were created on a linux machine with those uid/gid, the windows 
machine need only read acces, so it always falls into the "others" category.
Selinux/iptables are switched off.


Thanks for helping out!


From: Carlos Capriotti [mailto:[email protected]]
Sent: maandag 31 maart 2014 18:03
To: VAN CAUSBROECK Wannes
Cc: [email protected])
Subject: Re: [Gluster-users] nfs acces denied

maybe it would be nice to see your volume info for affected volumes.

Also, on the server side, what happens if you mount the share using glusterfs 
instead of nfs ?

any change the native nfs server is running on your server ?

Are there any auto-heal processes running ?

There are a few name resolution messages on your logs, that seem to refer to 
the nodes themselves. Any DNS conflicts ? Maybe add the names of servers to the 
hosts file ?

You MS client seems to be having issues with user/group translation. It seems 
to create files with gid 1003. (I could be wrong).

Again, is SElinux/ACLs/iptables disabled ?

All is very inconclusive os far.

On Mon, Mar 31, 2014 at 5:26 PM, VAN CAUSBROECK Wannes 
<[email protected]<mailto:[email protected]>>
 wrote:
Well, with 'client' i do actually mean the server itself.
i've tried forcing linux and windows to nfs V3 and tcp, and on windows i played 
around with the uid and gid, but the result is always the same


On 31 Mar 2014, at 17:22, "Carlos Capriotti" 
<[email protected]<mailto:[email protected]>> wrote:
Well, saying your client-side is "linux" does not help much. Distro, flavor, 
etc helps a lot, but I'll take a wild guess here.

First, force your NFS mount (client) to use nfs version 3.

The same for Microsoft. (It is fair to say I have no idea if the MS client 
supports v4 or not).

Additionally, check that firewalls are disabled on both sides, just for 
testing. The same goes for SElinux.

Windows and ACL, and user mapping is something that might be in your way too. 
There is a Technet document that describes how to handle this mapping if I am 
not wrong.
Just for testing, mount your nfs share you your own server, using 
localhost:/nfs_share and see how it goes.

It is a good start.

Kr,

Carlos

On Mon, Mar 31, 2014 at 3:58 PM, VAN CAUSBROECK Wannes 
<[email protected]<mailto:[email protected]>>
 wrote:
Hello all,

I've already tried to post this, but i'm unsure it arrived to the mailing list.

I have some issues regarding my nfs mounts. My setup is as follows:
Rhel 6.4, gluster 3.4.2-1 running on a vm (4 cores, 8GB ram) attached to a san. 
I have one disk on which are all the bricks (formatted ext4 in 64 bit mode) of 
25TB.
On the gluster side of things, everything works without issues. The trouble 
starts when I mount a volume as an nfs mount.
Lots of volumes work without issues, but others behave strangely. The volumes 
that act weird generally contain many files (can be accidental?).
The volumes in question mount without issues, but when I try to go into any 
subdirectory sometimes it works, sometimes I get errors.

On windows with nfs client: access denied

In nfslog:
[2014-03-31 13:57:58.771241] I [dht-layout.c:638:dht_layout_normalize] 
0-caviar_data11-dht: found anomalies in 
<gfid:c8d94120-6851-46ea-9f28-c629a44b1015>. holes=1 overlaps=0
[2014-03-31 13:57:58.771348] E 
[nfs3-helpers.c:3595:nfs3_fh_resolve_inode_lookup_cbk] 0-nfs-nfsv3: Lookup 
failed: <gfid:c8d94120-6851-46ea-9f28-c629a44b1015>: Invalid argument
[2014-03-31 13:57:58.771380] E [nfs3.c:1380:nfs3_lookup_resume] 0-nfs-nfsv3: 
Unable to resolve FH: (192.168.148.46:984<http://192.168.148.46:984>) 
caviar_data11 : c8d94120-6851-46ea-9f28-c629a44b1015
[2014-03-31 13:57:58.771819] W [nfs3-helpers.c:3380:nfs3_log_common_res] 
0-nfs-nfsv3: XID: 1ec28530, LOOKUP: NFS: 22(Invalid argument for operation), 
POSIX: 14(Bad address)
[2014-03-31 13:57:58.798967] I [dht-layout.c:638:dht_layout_normalize] 
0-caviar_data11-dht: found anomalies in 
<gfid:14972193-1039-4d7a-aed5-0d7e7eccf57b>. holes=1 overlaps=0
[2014-03-31 13:57:58.799039] E 
[nfs3-helpers.c:3595:nfs3_fh_resolve_inode_lookup_cbk] 0-nfs-nfsv3: Lookup 
failed: <gfid:14972193-1039-4d7a-aed5-0d7e7eccf57b>: Invalid argument
[2014-03-31 13:57:58.799056] E [nfs3.c:1380:nfs3_lookup_resume] 0-nfs-nfsv3: 
Unable to resolve FH: (192.168.148.46:984<http://192.168.148.46:984>) 
caviar_data11 : 14972193-1039-4d7a-aed5-0d7e7eccf57b
[2014-03-31 13:57:58.799088] W [nfs3-helpers.c:3380:nfs3_log_common_res] 
0-nfs-nfsv3: XID: 1ec28531, LOOKUP: NFS: 22(Invalid argument for operation), 
POSIX: 14(Bad address)
....


On linux:
[root@lpr-nas01 brick-xiv2]# ll /media/2011/201105/20110530/
ls: /media/2011/201105/20110530/37: No such file or directory
total 332
...
drwxrwsr-x 2 nfsnobody 1003 4096 Jun  6  2011 32
drwxrwsr-x 2 nfsnobody 1003 4096 Jun  6  2011 34
drwxrwsr-x 2 nfsnobody 1003 4096 Jun  6  2011 35
drwxrwsr-x 2 nfsnobody 1003 4096 Jun  6  2011 36
drwxrwsr-x 2 nfsnobody 1003 4096 Jun  6  2011 37
...

[root@lpr-nas01 brick-xiv2]# ll /media/2011/201105/20110530/37
ls: /media/2011/201105/20110530/37/NN.0000073824357.00001.tif: No such file or 
directory
ls: /media/2011/201105/20110530/37/NN.0000073824357.00003.tif: No such file or 
directory
total 54
-rwxrwxr-x 0 nfsnobody 1003  9340 Jun  6  2011 NN.0000073824357.00001.tif
-rwxrwxr-x 1 nfsnobody 1003 35312 Jun  6  2011 NN.0000073824357.00002.tif
-rwxrwxr-x 0 nfsnobody 1003  9340 Jun  6  2011 NN.0000073824357.00003.tif


I see in the nfslog:
...
[2014-03-31 12:44:18.941083] I [dht-layout.c:638:dht_layout_normalize] 
0-caviar_data11-dht: found anomalies in /2011/201107/20110716/55. holes=1 
overlaps=0
[2014-03-31 12:44:18.958078] I [dht-layout.c:638:dht_layout_normalize] 
0-caviar_data11-dht: found anomalies in /2011/201107/20110716/30. holes=1 
overlaps=0
[2014-03-31 12:44:18.959980] I [dht-layout.c:638:dht_layout_normalize] 
0-caviar_data11-dht: found anomalies in /2011/201107/20110716/90. holes=1 
overlaps=0
[2014-03-31 12:44:18.961094] E [dht-helper.c:429:dht_subvol_get_hashed] 
(-->/usr/lib64/glusterfs/3.4.2/xlator/debug/io-stats.so(io_stats_lookup+0x157) 
[0x7fd6a61282e7] (-->/usr/lib64/libglusterfs.so.0(default_lookup+0x6d) 
[0x3dfe01c03d] 
(-->/usr/lib64/glusterfs/3.4.2/xlator/cluster/distribute.so(dht_lookup+0xa7e) 
[0x7fd6a656af2e]))) 0-caviar_data11-dht: invalid argument: loc->parent
[2014-03-31 12:44:18.961283] W [client-rpc-fops.c:2624:client3_3_lookup_cbk] 
0-caviar_data11-client-0: remote operation failed: Invalid argument. Path: 
<gfid:00000000-0000-0000-0000-000000000000> 
(00000000-0000-0000-0000-000000000000)
[2014-03-31 12:44:18.961319] E [acl3.c:334:acl3_getacl_resume] 0-nfs-ACL: 
Unable to resolve FH: (192.168.151.21:740<http://192.168.151.21:740>) 
caviar_data11 : 00000000-0000-0000-0000-000000000000
[2014-03-31 12:44:18.961338] E [acl3.c:342:acl3_getacl_resume] 0-nfs-ACL: 
unable to open_and_resume
...

The weirdest thing is it changes from time to time which files and directories 
work and which don't
Any ideas?

Thanks!

_______________________________________________
Gluster-users mailing list
[email protected]<mailto:[email protected]>
http://supercolony.gluster.org/mailman/listinfo/gluster-users


_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to