maybe it would be nice to see your volume info for affected volumes.

Also, on the server side, what happens if you mount the share using
glusterfs instead of nfs ?

any change the native nfs server is running on your server ?

Are there any auto-heal processes running ?

There are a few name resolution messages on your logs, that seem to refer
to the nodes themselves. Any DNS conflicts ? Maybe add the names of servers
to the hosts file ?

You MS client seems to be having issues with user/group translation. It
seems to create files with gid 1003. (I could be wrong).

Again, is SElinux/ACLs/iptables disabled ?

All is very inconclusive os far.


On Mon, Mar 31, 2014 at 5:26 PM, VAN CAUSBROECK Wannes <
[email protected]> wrote:

>  Well, with 'client' i do actually mean the server itself.
> i've tried forcing linux and windows to nfs V3 and tcp, and on windows i
> played around with the uid and gid, but the result is always the same
>
>
>
> On 31 Mar 2014, at 17:22, "Carlos Capriotti" <[email protected]>
> wrote:
>
>   Well, saying your client-side is "linux" does not help much. Distro,
> flavor, etc helps a lot, but I'll take a wild guess here.
>
>  First, force your NFS mount (client) to use nfs version 3.
>
>  The same for Microsoft. (It is fair to say I have no idea if the MS
> client supports v4 or not).
>
>  Additionally, check that firewalls are disabled on both sides, just for
> testing. The same goes for SElinux.
>
>  Windows and ACL, and user mapping is something that might be in your way
> too. There is a Technet document that describes how to handle this mapping
> if I am not wrong.
>
>  Just for testing, mount your nfs share you your own server, using
> localhost:/nfs_share and see how it goes.
>
>  It is a good start.
>
>  Kr,
>
>  Carlos
>
>
> On Mon, Mar 31, 2014 at 3:58 PM, VAN CAUSBROECK Wannes <
> [email protected]> wrote:
>
>>  Hello all,
>>
>>
>>
>> I've already tried to post this, but i'm unsure it arrived to the mailing
>> list.
>>
>>
>>
>> I have some issues regarding my nfs mounts. My setup is as follows:
>>
>> Rhel 6.4, gluster 3.4.2-1 running on a vm (4 cores, 8GB ram) attached to
>> a san. I have one disk on which are all the bricks (formatted ext4 in 64
>> bit mode) of 25TB.
>>
>> On the gluster side of things, everything works without issues. The
>> trouble starts when I mount a volume as an nfs mount.
>>
>> Lots of volumes work without issues, but others behave strangely. The
>> volumes that act weird generally contain many files (can be accidental?).
>>
>> The volumes in question mount without issues, but when I try to go into
>> any subdirectory sometimes it works, sometimes I get errors.
>>
>>
>>
>> On windows with nfs client: access denied
>>
>>
>>
>> In nfslog:
>>
>> [2014-03-31 13:57:58.771241] I [dht-layout.c:638:dht_layout_normalize]
>> 0-caviar_data11-dht: found anomalies in
>> <gfid:c8d94120-6851-46ea-9f28-c629a44b1015>. holes=1 overlaps=0
>>
>> [2014-03-31 13:57:58.771348] E
>> [nfs3-helpers.c:3595:nfs3_fh_resolve_inode_lookup_cbk] 0-nfs-nfsv3: Lookup
>> failed: <gfid:c8d94120-6851-46ea-9f28-c629a44b1015>: Invalid argument
>>
>> [2014-03-31 13:57:58.771380] E [nfs3.c:1380:nfs3_lookup_resume]
>> 0-nfs-nfsv3: Unable to resolve FH: (192.168.148.46:984) caviar_data11 :
>> c8d94120-6851-46ea-9f28-c629a44b1015
>>
>> [2014-03-31 13:57:58.771819] W [nfs3-helpers.c:3380:nfs3_log_common_res]
>> 0-nfs-nfsv3: XID: 1ec28530, LOOKUP: NFS: 22(Invalid argument for
>> operation), POSIX: 14(Bad address)
>>
>> [2014-03-31 13:57:58.798967] I [dht-layout.c:638:dht_layout_normalize]
>> 0-caviar_data11-dht: found anomalies in
>> <gfid:14972193-1039-4d7a-aed5-0d7e7eccf57b>. holes=1 overlaps=0
>>
>> [2014-03-31 13:57:58.799039] E
>> [nfs3-helpers.c:3595:nfs3_fh_resolve_inode_lookup_cbk] 0-nfs-nfsv3: Lookup
>> failed: <gfid:14972193-1039-4d7a-aed5-0d7e7eccf57b>: Invalid argument
>>
>> [2014-03-31 13:57:58.799056] E [nfs3.c:1380:nfs3_lookup_resume]
>> 0-nfs-nfsv3: Unable to resolve FH: (192.168.148.46:984) caviar_data11 :
>> 14972193-1039-4d7a-aed5-0d7e7eccf57b
>>
>> [2014-03-31 13:57:58.799088] W [nfs3-helpers.c:3380:nfs3_log_common_res]
>> 0-nfs-nfsv3: XID: 1ec28531, LOOKUP: NFS: 22(Invalid argument for
>> operation), POSIX: 14(Bad address)
>>
>> ....
>>
>>
>>
>>
>>
>> On linux:
>>
>> [root@lpr-nas01 brick-xiv2]# ll /media/2011/201105/20110530/
>>
>> ls: /media/2011/201105/20110530/37: No such file or directory
>>
>> total 332
>>
>> ...
>>
>> drwxrwsr-x 2 nfsnobody 1003 4096 Jun  6  2011 32
>>
>> drwxrwsr-x 2 nfsnobody 1003 4096 Jun  6  2011 34
>>
>> drwxrwsr-x 2 nfsnobody 1003 4096 Jun  6  2011 35
>>
>> drwxrwsr-x 2 nfsnobody 1003 4096 Jun  6  2011 36
>>
>> drwxrwsr-x 2 nfsnobody 1003 4096 Jun  6  2011 37
>>
>> ...
>>
>>
>>
>> [root@lpr-nas01 brick-xiv2]# ll /media/2011/201105/20110530/37
>>
>> ls: /media/2011/201105/20110530/37/NN.0000073824357.00001.tif: No such
>> file or directory
>>
>> ls: /media/2011/201105/20110530/37/NN.0000073824357.00003.tif: No such
>> file or directory
>>
>> total 54
>>
>> -rwxrwxr-x 0 nfsnobody 1003  9340 Jun  6  2011 NN.0000073824357.00001.tif
>>
>> -rwxrwxr-x 1 nfsnobody 1003 35312 Jun  6  2011 NN.0000073824357.00002.tif
>>
>> -rwxrwxr-x 0 nfsnobody 1003  9340 Jun  6  2011 NN.0000073824357.00003.tif
>>
>>
>>
>>
>>
>> I see in the nfslog:
>>
>> ...
>>
>> [2014-03-31 12:44:18.941083] I [dht-layout.c:638:dht_layout_normalize]
>> 0-caviar_data11-dht: found anomalies in /2011/201107/20110716/55. holes=1
>> overlaps=0
>>
>> [2014-03-31 12:44:18.958078] I [dht-layout.c:638:dht_layout_normalize]
>> 0-caviar_data11-dht: found anomalies in /2011/201107/20110716/30. holes=1
>> overlaps=0
>>
>> [2014-03-31 12:44:18.959980] I [dht-layout.c:638:dht_layout_normalize]
>> 0-caviar_data11-dht: found anomalies in /2011/201107/20110716/90. holes=1
>> overlaps=0
>>
>> [2014-03-31 12:44:18.961094] E [dht-helper.c:429:dht_subvol_get_hashed]
>> (-->/usr/lib64/glusterfs/3.4.2/xlator/debug/io-stats.so(io_stats_lookup+0x157)
>> [0x7fd6a61282e7] (-->/usr/lib64/libglusterfs.so.0(default_lookup+0x6d)
>> [0x3dfe01c03d]
>> (-->/usr/lib64/glusterfs/3.4.2/xlator/cluster/distribute.so(dht_lookup+0xa7e)
>> [0x7fd6a656af2e]))) 0-caviar_data11-dht: invalid argument: loc->parent
>>
>> [2014-03-31 12:44:18.961283] W
>> [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-caviar_data11-client-0:
>> remote operation failed: Invalid argument. Path:
>> <gfid:00000000-0000-0000-0000-000000000000>
>> (00000000-0000-0000-0000-000000000000)
>>
>> [2014-03-31 12:44:18.961319] E [acl3.c:334:acl3_getacl_resume] 0-nfs-ACL:
>> Unable to resolve FH: (192.168.151.21:740) caviar_data11 :
>> 00000000-0000-0000-0000-000000000000
>>
>> [2014-03-31 12:44:18.961338] E [acl3.c:342:acl3_getacl_resume] 0-nfs-ACL:
>> unable to open_and_resume
>>
>> ...
>>
>>
>>
>> The weirdest thing is it changes from time to time which files and
>> directories work and which don't
>>
>> Any ideas?
>>
>>
>>
>> Thanks!
>>
>> _______________________________________________
>> Gluster-users mailing list
>> [email protected]
>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>
>
>
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to