Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-11 Thread Rick Macklem
NAGY Andreas wrote:
>Thanks! Please keep me updated if you find put more or when a updated version 
>is available.
Will try to remember to do so.

>As I now know it is working, I will start tomorrow to build up a testsystem 
>with 3 NFS servers >(two of them in a ha with CARP and HAST) and several ESXi 
>hosts which will all access there >NFS datastores over 4 uplinks with NICs on 
>different subnets.
>It should always be possible to do there some testing.
Have fun. That's way beyond anything I do for NFS testing.

Btw, unless you don't want me to, I will list you as "Tested by:" on the 
commits.
(If you don't want me to do this, just email.)

Good luck with the testing, rick
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-10 Thread NAGY Andreas
Thanks! Please keep me updated if you find put more or when a updated version 
is available.

As I now know it is working, I will start tomorrow to build up a testsystem 
with 3 NFS servers (two of them in a ha with CARP and HAST) and several ESXi 
hosts which will all access there NFS datastores over 4 uplinks with NICs on 
different subnets.
It should always be possible to do there some testing.

andi



Von: Rick Macklem <rmack...@uoguelph.ca>
Gesendet: 10.03.2018 11:20 nachm.
An: NAGY Andreas; 'freebsd-stable@freebsd.org'
Betreff: Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

NAGY Andreas wrote:
>Thanks, the not issuing delegation warnings disappeared with this patch.
>
>But now there are some new warnings I haven't seen so far:
>2018-03-10T13:01:39.441Z cpu8:68046)WARNING: NFS41: NFS41FSOpGetObject:2148: 
>Failed to >get object 0x43910e71b386 [36 c6b10167 9b157f95 5aa100fb 8ffcf2c1 c 
>2 9f22ad6d 0 0 0 0 0]: >Stale file handle
I doubt these would be related to the patch. A stale FH means that the client 
tried to
access a file via its FH after it was removed. (Normally this is a client bug, 
but hopefully
not one that will cause grief.)
>These only appear several times after a the NFS share is mounted or remounted 
>after a >connection loss.
>Everything works fine, but haven't seen them till I applied the last patch.
>
>andi
Ok. Thanks for testing all of these patches. I will probably get cleaned up 
versions of
them committed in April.

The main outstanding issue is the Readdir one about directory changing too much.
Hopefully I can find out something about it via email.

Have fun with it, rick
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-10 Thread Rick Macklem
NAGY Andreas wrote:
>Thanks, the not issuing delegation warnings disappeared with this patch.
>
>But now there are some new warnings I haven't seen so far:
>2018-03-10T13:01:39.441Z cpu8:68046)WARNING: NFS41: NFS41FSOpGetObject:2148: 
>Failed to >get object 0x43910e71b386 [36 c6b10167 9b157f95 5aa100fb 8ffcf2c1 c 
>2 9f22ad6d 0 0 0 0 0]: >Stale file handle
I doubt these would be related to the patch. A stale FH means that the client 
tried to
access a file via its FH after it was removed. (Normally this is a client bug, 
but hopefully
not one that will cause grief.)
>These only appear several times after a the NFS share is mounted or remounted 
>after a >connection loss.
>Everything works fine, but haven't seen them till I applied the last patch.
>
>andi
Ok. Thanks for testing all of these patches. I will probably get cleaned up 
versions of
them committed in April.

The main outstanding issue is the Readdir one about directory changing too much.
Hopefully I can find out something about it via email.

Have fun with it, rick
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-10 Thread NAGY Andreas
Thanks, the not issuing delegation warnings disappeared with this patch.

But now there are some new warnings I haven't seen so far:
2018-03-10T13:01:39.441Z cpu8:68046)WARNING: NFS41: NFS41FSOpGetObject:2148: 
Failed to get object 0x43910e71b386 [36 c6b10167 9b157f95 5aa100fb 8ffcf2c1 c 2 
9f22ad6d 0 0 0 0 0]: Stale file handle

These only appear several times after a the NFS share is mounted or remounted 
after a connection loss. 
Everything works fine, but haven't seen them till I applied the last patch.

andi

___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-09 Thread Rick Macklem
NAGY Andreas wrote:
>>I'll take another look and see if I can guess why it doesn't like "2" as a 
>>reason for >not issuing a delegation. (As noted before, I don't think this is 
>>serious, but???)
>
>This warnings are still there, but don't seem to have any impact. It looks 
>like they >only appeare when files are created or modified on the datastore 
>from the >datastore browser or from shell, have not seen this warnings when 
>working in a >VM on a virtual disk that is stored on the nfs datastore.
I've attached one more little patch. If this one is applied after 
wantdeleg.patch
and wantdeleg2.patch it might get rid of these. (There was another place
that needed to be changed that got missed by wantdeleg.patch.)

Have fun with it, rick



wantdeleg3.patch
Description: wantdeleg3.patch
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-09 Thread NAGY Andreas
>The attached patch changes BindConnectiontoSession to reply 
>NFS4ERR_BAD_SESSION when the session doesn't exist. This might trigger 
>recovery after a server reboot.
>This patch must be applied after bindconn.patch.

Works perfect! ESXi host reconnects to the datastore as soon as the nfsserver 
is available.

>I took a quick look at your packet trace and it appears that this client does 
>all write FILESYNC. As such, setting vfs.nfsd.async=1 won't have any affect.
>If you apply the attached patch, it should change the FILESYNC->UNSTABLE so 
>that vfs.nfsd.async=1 will make a difference. Again, doing this does put data 
>at risk when the server crashes.

Yes, ESXi writes everything sync. With the patch + vfs.nfsd.async=1 (only for 
testing)  writes are a bit faster. 
If have not tuned anything on the fs settings so far (just formatted it as 
standard UFS), compared to multipath iSCSI with VMFS reads are a little bit 
faster, but writes are a bit slower.
That's just what I see from a simple ATTO benchmark from within a Windows VM, 
have not done any details IOP benchmark,...
The RAID controller on this machines does not support IT mode, but I think I 
will still use ZFS, but only as filesystem on a single hw raid disk. Must check 
what are the best setting for this on the hw raid + zfs for nfs.

>This one is a mystery to me. It seemed to be upset that the directory is 
>changing (I assume either the Change or ModifyTime attributes). However, if 
>entries are being deleted, the directory is changing and, as far as I know, 
>the Change and ModifyTime attributes are supposed to change.
>I might try posting on nf...@ietf.org in case somebody involved with this 
>client reads that list and can explain what this is?

Maybe it is really just a bug in the VMware integrated host client browser. 
Deleting folders on the mounted datastore in the ESXi shell is no problem and 
does also not generate any warnings in the vmkernel.log.

>I'll take another look and see if I can guess why it doesn't like "2" as a 
>reason for not issuing a delegation. (As noted before, I don't think this is 
>serious, but???)

This warnings are still there, but don't seem to have any impact. It looks like 
they only appeare when files are created or modified on the datastore from the 
datastore browser or from shell, have not seen this warnings when working in a 
VM on a virtual disk that is stored on the nfs datastore.

>Yes, before I posted that I didn't understand why multiple TCP links would be 
>faster.
>I didn't notice at the time that you mentioned using different subnets and, as 
>such, links couldn't be trunked below TCP. In your case trunking above TCP 
>makes sense.

Yes, actually I am working on a lab environment/testsys I often get some 
servers as leftovers from projects and there are also plenty of Cisco 1GB 
switches, but 10GBs are rare. I already did some tests with iSCSI multipathing 
but I prefer NFS and now with this I get also the same speed. 
As I have never seen a working setup with multiple paths for NFS, so I am 
really happy that you got this working in such a short time.

andi

___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-08 Thread Rick Macklem
NAGY Andreas wrote:
>Actually I have only made some quick benchmarks with ATTO in a Windows VM 
>>which has a vmdk on the NFS41 datastore which is mounted over two 1GB links 
>in >different subnets.
>Read is nearly the double of just a single connection and write is just a bit 
>faster. >Don't know if write speed could be improved, actually the share is 
>UFS on a HW >raid controller which has local write speeds about 500MB/s.
I took a quick look at your packet trace and it appears that this client
does all write FILESYNC. As such, setting vfs.nfsd.async=1 won't have any 
affect.

If you apply the attached patch, it should change the FILESYNC->UNSTABLE so
that vfs.nfsd.async=1 will make a difference. Again, doing this does put data
at risk when the server crashes.

rick


writeasync.patch
Description: writeasync.patch
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-08 Thread Rick Macklem
NAGY Andreas wrote:
>- after a reboot of the FreeBSD machine the ESXi does not restore the NFS 
>>datastore again with following warning (just disconnecting the links is fine)
>2018-03-08T12:39:44.602Z cpu23:66484)WARNING: NFS41: 
> >NFS41_Bug:2361: BUG - Invalid BIND_CONN_TO_SESSION error: >NFS4ERR_NOTSUPP
The attached patch changes BindConnectiontoSession to reply
NFS4ERR_BAD_SESSION when the session doesn't exist. This might trigger
recovery after a server reboot.
This patch must be applied after bindconn.patch.

rick


bindconn2.patch
Description: bindconn2.patch
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-08 Thread Rick Macklem
NAGY Andreas wrote:
>Thanks you, really great how fast you adapt the source/make patches for this. 
>Saw so many >posts were people did not get NFS41 working with ESXi and FreeBSD 
>and now I have it already >running with your changes.
>
>I have now compiled the kernel with all 4 patches, and it works now.
Ok. Sounds like we are making progress. It also takes someone willing to test 
patches, so
thanks for doing so.
>Some problems are still left:
>
>- the "Server returned improper reason for no delegation: 2" warnings are 
>still in the >vmkernel.log.
>2018-03-08T11:41:20.290Z cpu0:68011 opID=488969b0)WARNING: 
> NFS41: >NFS41ValidateDelegation:608: Server returned improper reason for no 
> delegation: 2
I'll take another look and see if I can guess why it doesn't like "2" as a 
reason for not
issuing a delegation. (As noted before, I don't think this is serious, but???)

>- can't delete a folder with the VMware host client datastore browser:
 >   2018-03-08T11:34:00.349Z cpu1:67981 opID=f5159ce3)WARNING: 
 > NFS41: >NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
 > 0x43046e4cb158: Transient >file system condition, suggest retry
[more of these snipped]
>2018-03-08T11:34:00.352Z cpu1:67981 opID=f5159ce3)WARNING: 
> UserFile: 2155: >hostd-worker: Directory changing too often to perform 
> readdir operation (11 retries), >returning busy
This one is a mystery to me. It seemed to be upset that the directory is 
changing (I
assume either the Change or ModifyTime attributes). However, if entries are 
being
deleted, the directory is changing and, as far as I know, the Change and 
ModifyTime
attributes are supposed to change.
I might try posting on nf...@ietf.org in case somebody involved with this 
client reads
that list and can explain what this is?

>- after a reboot of the FreeBSD machine the ESXi does not restore the NFS 
>datastore again >with following warning (just disconnecting the links is fine)
>2018-03-08T12:39:44.602Z cpu23:66484)WARNING: NFS41: 
> NFS41_Bug:2361: BUG - >Invalid BIND_CONN_TO_SESSION error: NFS4ERR_NOTSUPP
Hmm. Normally after a server reboot, the clients will try some RPC that starts 
with a
Sequence (the session op) and the server will reply NFS4ERR_BAD_SESSION.
This triggers recovery in the client.
The BindConnectiontoSession operation is done in an RPC by itself, so there is 
no
Sequence op to trigger NFS4ERR_BAD_SESSION.
Maybe this client expects to see NFS4ERR_BAD_SESSION for the 
BindConnectiontoSession.
I'll post a patch that modifies the BindConnectiontoSession to do that.

>Actually I have only made some quick benchmarks with ATTO in a Windows VM 
>which has a >vmdk on the NFS41 datastore which is mounted over two 1GB links 
>in different subnets.
>Read is nearly the double of just a single connection and write is just a bit 
>faster. Don't know if >write speed could be improved, actually the share is 
>UFS on a HW raid controller which has >local write speeds about 500MB/s.
Yes, before I posted that I didn't understand why multiple TCP links would be 
faster.
I didn't notice at the time that you mentioned using different subnets and, as 
such,
links couldn't be trunked below TCP. In your case trunking above TCP makes 
sense.

Getting slower write rates than read rates from NFS is normal.
Did you try "sysctl vfs.nfsd.async=1"?
The other thing that might help for UFS is increasing the size of the buffer 
cache.
(If this server is mainly an NFS server you could probably make the buffer cache
 greater than half of the machine's ram.
 Note to others, since ZFS doesn't use the buffer cache, the opposite is true 
for
 ZFS.)

rick

___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-08 Thread NAGY Andreas
Thanks you, really great how fast you adapt the source/make patches for this. 
Saw so many posts were people did not get NFS41 working with ESXi and FreeBSD 
and now I have it already running with your changes.

I have now compiled the kernel with all 4 patches, and it works now.

Some problems are still left:

- the "Server returned improper reason for no delegation: 2" warnings are still 
in the vmkernel.log.
2018-03-08T11:41:20.290Z cpu0:68011 opID=488969b0)WARNING: 
NFS41: NFS41ValidateDelegation:608: Server returned improper reason for no 
delegation: 2

- can't delete a folder with the VMware host client datastore browser:
2018-03-08T11:34:00.349Z cpu1:67981 opID=f5159ce3)WARNING: 
NFS41: NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
0x43046e4cb158: Transient file system condition, suggest retry
2018-03-08T11:34:00.349Z cpu1:67981 opID=f5159ce3)WARNING: 
NFS41: NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
0x43046e4cb158: Transient file system condition, suggest retry
2018-03-08T11:34:00.349Z cpu1:67981 opID=f5159ce3)WARNING: 
NFS41: NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
0x43046e4cb158: Transient file system condition, suggest retry
2018-03-08T11:34:00.350Z cpu1:67981 opID=f5159ce3)WARNING: 
NFS41: NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
0x43046e4cb158: Transient file system condition, suggest retry
2018-03-08T11:34:00.350Z cpu1:67981 opID=f5159ce3)WARNING: 
NFS41: NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
0x43046e4cb158: Transient file system condition, suggest retry
2018-03-08T11:34:00.350Z cpu1:67981 opID=f5159ce3)WARNING: 
NFS41: NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
0x43046e4cb158: Transient file system condition, suggest retry
2018-03-08T11:34:00.351Z cpu1:67981 opID=f5159ce3)WARNING: 
NFS41: NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
0x43046e4cb158: Transient file system condition, suggest retry
2018-03-08T11:34:00.351Z cpu1:67981 opID=f5159ce3)WARNING: 
NFS41: NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
0x43046e4cb158: Transient file system condition, suggest retry
2018-03-08T11:34:00.351Z cpu1:67981 opID=f5159ce3)WARNING: 
NFS41: NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
0x43046e4cb158: Transient file system condition, suggest retry
2018-03-08T11:34:00.351Z cpu1:67981 opID=f5159ce3)WARNING: 
NFS41: NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
0x43046e4cb158: Transient file system condition, suggest retry
2018-03-08T11:34:00.352Z cpu1:67981 opID=f5159ce3)WARNING: 
NFS41: NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 
0x43046e4cb158: Transient file system condition, suggest retry
2018-03-08T11:34:00.352Z cpu1:67981 opID=f5159ce3)WARNING: 
UserFile: 2155: hostd-worker: Directory changing too often to perform readdir 
operation (11 retries), returning busy

- after a reboot of the FreeBSD machine the ESXi does not restore the NFS 
datastore again with following warning (just disconnecting the links is fine)
2018-03-08T12:39:44.602Z cpu23:66484)WARNING: NFS41: 
NFS41_Bug:2361: BUG - Invalid BIND_CONN_TO_SESSION error: NFS4ERR_NOTSUPP

Actually I have only made some quick benchmarks with ATTO in a Windows VM which 
has a vmdk on the NFS41 datastore which is mounted over two 1GB links in 
different subnets.
Read is nearly the double of just a single connection and write is just a bit 
faster. Don't know if write speed could be improved, actually the share is UFS 
on a HW raid controller which has local write speeds about 500MB/s.

At following link is the vmkernel.log from mouning the NFS share, attaching a 
vmdk from the share to a Win VM, running ATTO benchmark on it, 
disconnecting/reconnecting network and also the problem with the 
BIND_CONN_TO_SESSION error: NFS4ERR_NOTSUPP after reboot.
Till the reboot I have also made a trace on one of the two links. 
(nfs41_trace_before_reboot.pcap and nfs41_trace_after_reboot.pcap)

https://files.fm/u/wvybmdmc

andi

-Original Message-
From: Rick Macklem [mailto:rmack...@uoguelph.ca] 
Sent: Donnerstag, 8. März 2018 03:48
To: NAGY Andreas <andreas.n...@frequentis.com>; 'freebsd-stable@freebsd.org' 
<freebsd-stable@freebsd.org>
Subject: Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

NAGY Andreas wrote:
>attached the trace. If I see it correct it uses FORE_OR_BOTH. 
>(bctsa_dir: >CDFC4_FORE_OR_BOTH (0x0003))
Yes. The scary part is the ExchangeID before the BindConnectiontoSession.
(Normally that is only done at the beginning of a new mount to get a ClientID,  
followed immediately by a CreateSessio

Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-07 Thread Rick Macklem
NAGY Andreas wrote:
>attached the trace. If I see it correct it uses FORE_OR_BOTH. (bctsa_dir: 
>>CDFC4_FORE_OR_BOTH (0x0003))
Yes. The scary part is the ExchangeID before the BindConnectiontoSession.
(Normally that is only done at the beginning of a new mount to get a ClientID,
 followed immediately by a CreateSession. I don't know why it would do this?)

The attached patch might get BindConnectiontoSession to work. I have no way
to test it beyond seeing it compile. Hopefully it will apply cleanly.

>The trace is only with the first patch, have not compiled the wantdeleg 
>patches so >far.
That's fine. I don't think that matters much.

>I think this is related to the BIND_CONN_TO_SESSION; after a disconnect the 
>ESXi >cannot connect to the NFS also with this warning:
>2018-03-07T16:55:11.227Z cpu21:66484)WARNING: NFS41: NFS41_Bug:2361: >BUG - 
>Invalid BIND_CONN_TO_SESSION error: NFS4ERR_NOTSUPP
If the attached patch works, you'll find out what it fixes.

>Another thing I noticed today is that it is not possible to delete a folder 
>with the >ESXi datastorebrowser on the NFS mount. Maybe it is a VMWare bug, 
>but with >NFS3 it works.
>
>Here the vmkernel.log with only one connection contains mounting, trying to 
>>delete a folder and disconnect:
>
>2018-03-07T16:46:04.543Z cpu12:68008 opID=55bea165)World: 12235: VC opID 
>>c55dbe59 maps to vmkernel opID 55bea165
>2018-03-07T16:46:04.543Z cpu12:68008 opID=55bea165)NFS41: 
>>NFS41_VSIMountSet:423: Mount server: 10.0.0.225, port: 2049, path: /, label: 
>>nfsds1, security: 1 user: , options: 
>2018-03-07T16:46:04.543Z cpu12:68008 opID=55bea165)StorageApdHandler: >977: 
>APD Handle  Created with lock[StorageApd-0x43046e4c6d70]
>2018-03-07T16:46:04.544Z cpu11:66486)NFS41: 
>>NFS41ProcessClusterProbeResult:3873: Reclaiming state, cluster 0x43046e4c7ee0 
>>[7]
>2018-03-07T16:46:04.545Z cpu12:68008 opID=55bea165)NFS41: 
>>NFS41FSCompleteMount:3791: Lease time: 120
>2018-03-07T16:46:04.545Z cpu12:68008 opID=55bea165)NFS41: 
>>NFS41FSCompleteMount:3792: Max read xfer size: 0x2
>2018-03-07T16:46:04.545Z cpu12:68008 opID=55bea165)NFS41: 
>>NFS41FSCompleteMount:3793: Max write xfer size: 0x2
>2018-03-07T16:46:04.545Z cpu12:68008 opID=55bea165)NFS41: 
>>NFS41FSCompleteMount:3794: Max file size: 0x8000
>2018-03-07T16:46:04.545Z cpu12:68008 opID=55bea165)NFS41: 
>>NFS41FSCompleteMount:3795: Max file name: 255
>2018-03-07T16:46:04.545Z cpu12:68008 opID=55bea165)WARNING: NFS41: 
>>NFS41FSCompleteMount:3800: The max file name size (255) of file system is 
>>larger than that of FSS (128)
>2018-03-07T16:46:04.546Z cpu12:68008 opID=55bea165)NFS41: 
>>NFS41FSAPDNotify:5960: Restored connection to the server 10.0.0.225 mount 
>>point nfsds1, mounted as 1a7893c8-eec764a7-- ("/")
>2018-03-07T16:46:04.546Z cpu12:68008 opID=55bea165)NFS41: 
>>NFS41_VSIMountSet:435: nfsds1 mounted successfully
>2018-03-07T16:47:19.869Z cpu21:67981 opID=e47706ec)World: 12235: VC opID 
>>c55dbe91 maps to vmkernel opID e47706ec
>2018-03-07T16:47:19.869Z cpu21:67981 opID=e47706ec)WARNING: NFS41: 
>>NFS41FileOpReaddir:4728: Failed to process READDIR result for fh 0x43046e4c6
I have no idea if getting BindConnectiontoSession working will fix this or not?

rick



bindconn.patch
Description: bindconn.patch
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-07 Thread NAGY Andreas
eas.n...@frequentis.com>; 'freebsd-stable@freebsd.org' 
<freebsd-stable@freebsd.org>
Subject: Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

NAGY Andreas wrote:
>Okay, that was the main reason for using NFS 4.1.
>Is it planned to implement it, or is the focus on pNFS?
I took a quick look and implementing this for some cases will be pretty easy. 
Binding a FORE channel is implied, so for that case all the server does is 
reply OK to the BIND_CONN_TO_SESSION.

To know if the ESXi client case is a simple one, I need to see what the 
BIND_CONN_TO_SESSION arguments look like.
If you can capture packets for when this second connection is done and email it 
to me as an attachment, I can look at what the BIND_CONN_TO_SESSION args are.
# tcpdump -s 0 -w  host  run on the 
FreeBSD server should get the  I need.

Alternately, if you have wireshark handy, you can just use it to look for the 
BIND_CONN_TO_SESSION request and see if it specifies (FORE, BACK, FORE_OR_BOTH 
or BACK_OR_BOTH) in it.
FORE or FORE_OR_BOTH means it is easy to do and I can probably have a patch for 
testing in a day or two.

rick


nfs41_trace.pcap
Description: nfs41_trace.pcap
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-07 Thread Rick Macklem
NAGY Andreas wrote:
>Okay, that was the main reason for using NFS 4.1.
>Is it planned to implement it, or is the focus on pNFS?
I took a quick look and implementing this for some cases will be pretty
easy. Binding a FORE channel is implied, so for that case all the server
does is reply OK to the BIND_CONN_TO_SESSION.

To know if the ESXi client case is a simple one, I need to see what the
BIND_CONN_TO_SESSION arguments look like.
If you can capture packets for when this second connection is done and
email it to me as an attachment, I can look at what the BIND_CONN_TO_SESSION 
args are.
# tcpdump -s 0 -w  host 
run on the FreeBSD server should get the  I need.

Alternately, if you have wireshark handy, you can just use it to look
for the BIND_CONN_TO_SESSION request and see if it specifies
(FORE, BACK, FORE_OR_BOTH or BACK_OR_BOTH) in it.
FORE or FORE_OR_BOTH means it is easy to do and I can probably have
a patch for testing in a day or two.

rick
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-06 Thread Rick Macklem
NAGY Andreas wrote:
>Compiling with the last patch also failed:
>
>error: use of undeclared identifier 'NFSV4OPEN_WDSUPPFTYPE

If you apply the attached patch along with wantdeleg.patch, it should
build. At most, this will get rid of the warnings about invalid reason for
not issuing a delegation, so the wantdeleg*.patches probably aren't very
interesting.

rick



wantdeleg2.patch
Description: wantdeleg2.patch
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-06 Thread Rick Macklem
NAGY Andreas wrote:
>Okay, that was the main reason for using NFS 4.1.
>Is it planned to implement it, or is the focus on pNFS?
Do the VMware people claim that this improves performance?
(I know nothing about the world of VMs, but for real hardware
 I can't see any advantage of having more than one TCP connection?
 As far as I know, the Linux client never tries to acquire a second TCP
 connection. I would have assumed trunking would be handled below
 TCP.)

This is the first client that I am aware of (and just yesterday when you pointed
it out) that uses BIND_CONN_TO_SESSION for an additional TCP connection.
(Up until now I was only aware of it being used for RDMA setups and I have no
 hardware to play with such things.)

If the VMware folk claim it does improve performance, I might get around to
it someday, although you are correct that I am working on pNFS support for the
server right now.

rick
[stuff snipped]

___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-06 Thread NAGY Andreas
Okay, that was the main reason for using NFS 4.1.
Is it planned to implement it, or is the focus on pNFS?

Thanks,
Andi



Von: Rick Macklem <rmack...@uoguelph.ca>
Gesendet: 05.03.2018 11:49 nachm.
An: NAGY Andreas; 'freebsd-stable@freebsd.org'
Betreff: Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

Nope, that isn't supported, rick
(Hope no one is too upset by a top post.)


From: NAGY Andreas <andreas.n...@frequentis.com>
Sent: Monday, March 5, 2018 8:22:10 AM
To: Rick Macklem; 'freebsd-stable@freebsd.org'
Subject: RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

Thanks, I am actually compiling with both patches.

I try now to get NFS 4.1 multipathing working. So I have now two connection on 
different subnets between the ESXi host and the FreeBSD host with exports for 
the same mountpoint on both subnets.
Now I get the following errors in the vmkernel.log:
2018-03-05T13:06:07.488Z cpu10:66503)WARNING: NFS41: NFS41_Bug:2361: BUG - 
Invalid BIND_CONN_TO_SESSION error: NFS4ERR_NOTSUPP

Is there session trunking available in the FreeBSD NFS41 implementation?

Br,
andi

-Original Message-
From: Rick Macklem [mailto:rmack...@uoguelph.ca]
Sent: Montag, 5. März 2018 02:16
To: NAGY Andreas <andreas.n...@frequentis.com>; 'freebsd-stable@freebsd.org' 
<freebsd-stable@freebsd.org>
Subject: Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

NAGY Andreas wrote:
[stuff snipped]
>In the source I saw nfs_async = 0; is it right that NFS will work in async 
>mode if I >compile the kernel with nfs_async = 1?
>
>I know the risk of running it async, but is it not the same risk having the 
>datastore >connected via iSCSI which standard is also not sync?
If you want to use it, you can just set it by setting the sysctl 
vfs.nfsd.async=1.
- If the server crashes/reboots you can lose data. Also, after the reboot, the
  client will only see an temporarily unresponsive server and will not have any
  indication of data loss.
  (I am not familiar with iSCSI, so I can't comment on how safe that is.)
- If you are using ZFS, there is also a ZFS config (sync=disabled). I'm not a 
ZFS
  guy, so I don't know anything more, but I'm sure others reading this list can
  tell you how to set it.
[more stuff snipped]
>So far I did not see any issue with the mount. Only in the vmkernel.log there 
>are >often following entrees:
>WARNING: NFS41: NFS41ValidateDelegation:608: Server returned improper
>>reason for no delegation: 2
The attached patch *might* get rid of these, although I don't think it matters 
much, since it is just complaining about the "reason" the server returns for 
not issuing a delegation (issuing delegations is entirely at the discretion of 
the server and is disabled by default).
[more stuff snipped]
Good luck with it, rick
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-05 Thread Rick Macklem
Nope, that isn't supported, rick
(Hope no one is too upset by a top post.)


From: NAGY Andreas <andreas.n...@frequentis.com>
Sent: Monday, March 5, 2018 8:22:10 AM
To: Rick Macklem; 'freebsd-stable@freebsd.org'
Subject: RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

Thanks, I am actually compiling with both patches.

I try now to get NFS 4.1 multipathing working. So I have now two connection on 
different subnets between the ESXi host and the FreeBSD host with exports for 
the same mountpoint on both subnets.
Now I get the following errors in the vmkernel.log:
2018-03-05T13:06:07.488Z cpu10:66503)WARNING: NFS41: NFS41_Bug:2361: BUG - 
Invalid BIND_CONN_TO_SESSION error: NFS4ERR_NOTSUPP

Is there session trunking available in the FreeBSD NFS41 implementation?

Br,
andi

-Original Message-
From: Rick Macklem [mailto:rmack...@uoguelph.ca]
Sent: Montag, 5. März 2018 02:16
To: NAGY Andreas <andreas.n...@frequentis.com>; 'freebsd-stable@freebsd.org' 
<freebsd-stable@freebsd.org>
Subject: Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

NAGY Andreas wrote:
[stuff snipped]
>In the source I saw nfs_async = 0; is it right that NFS will work in async 
>mode if I >compile the kernel with nfs_async = 1?
>
>I know the risk of running it async, but is it not the same risk having the 
>datastore >connected via iSCSI which standard is also not sync?
If you want to use it, you can just set it by setting the sysctl 
vfs.nfsd.async=1.
- If the server crashes/reboots you can lose data. Also, after the reboot, the
  client will only see an temporarily unresponsive server and will not have any
  indication of data loss.
  (I am not familiar with iSCSI, so I can't comment on how safe that is.)
- If you are using ZFS, there is also a ZFS config (sync=disabled). I'm not a 
ZFS
  guy, so I don't know anything more, but I'm sure others reading this list can
  tell you how to set it.
[more stuff snipped]
>So far I did not see any issue with the mount. Only in the vmkernel.log there 
>are >often following entrees:
>WARNING: NFS41: NFS41ValidateDelegation:608: Server returned improper
>>reason for no delegation: 2
The attached patch *might* get rid of these, although I don't think it matters 
much, since it is just complaining about the "reason" the server returns for 
not issuing a delegation (issuing delegations is entirely at the discretion of 
the server and is disabled by default).
[more stuff snipped]
Good luck with it, rick
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-05 Thread NAGY Andreas
Compiling with the last patch also failed:

error: use of undeclared identifier 'NFSV4OPEN_WDSUPPFTYPE'



-Original Message-
From: NAGY Andreas 
Sent: Montag, 5. März 2018 14:22
To: Rick Macklem <rmack...@uoguelph.ca>; 'freebsd-stable@freebsd.org' 
<freebsd-stable@freebsd.org>
Subject: RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

Thanks, I am actually compiling with both patches.

I try now to get NFS 4.1 multipathing working. So I have now two connection on 
different subnets between the ESXi host and the FreeBSD host with exports for 
the same mountpoint on both subnets.
Now I get the following errors in the vmkernel.log:
2018-03-05T13:06:07.488Z cpu10:66503)WARNING: NFS41: NFS41_Bug:2361: BUG - 
Invalid BIND_CONN_TO_SESSION error: NFS4ERR_NOTSUPP

Is there session trunking available in the FreeBSD NFS41 implementation?

Br,
andi

-Original Message-
From: Rick Macklem [mailto:rmack...@uoguelph.ca] 
Sent: Montag, 5. März 2018 02:16
To: NAGY Andreas <andreas.n...@frequentis.com>; 'freebsd-stable@freebsd.org' 
<freebsd-stable@freebsd.org>
Subject: Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

NAGY Andreas wrote:
[stuff snipped]
>In the source I saw nfs_async = 0; is it right that NFS will work in async 
>mode if I >compile the kernel with nfs_async = 1?
>
>I know the risk of running it async, but is it not the same risk having the 
>datastore >connected via iSCSI which standard is also not sync?
If you want to use it, you can just set it by setting the sysctl 
vfs.nfsd.async=1.
- If the server crashes/reboots you can lose data. Also, after the reboot, the
  client will only see an temporarily unresponsive server and will not have any
  indication of data loss.
  (I am not familiar with iSCSI, so I can't comment on how safe that is.)
- If you are using ZFS, there is also a ZFS config (sync=disabled). I'm not a 
ZFS
  guy, so I don't know anything more, but I'm sure others reading this list can
  tell you how to set it.
[more stuff snipped]
>So far I did not see any issue with the mount. Only in the vmkernel.log there 
>are >often following entrees:
>WARNING: NFS41: NFS41ValidateDelegation:608: Server returned improper 
>>reason for no delegation: 2
The attached patch *might* get rid of these, although I don't think it matters 
much, since it is just complaining about the "reason" the server returns for 
not issuing a delegation (issuing delegations is entirely at the discretion of 
the server and is disabled by default).
[more stuff snipped]
Good luck with it, rick
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-05 Thread NAGY Andreas
Thanks, I am actually compiling with both patches.

I try now to get NFS 4.1 multipathing working. So I have now two connection on 
different subnets between the ESXi host and the FreeBSD host with exports for 
the same mountpoint on both subnets.
Now I get the following errors in the vmkernel.log:
2018-03-05T13:06:07.488Z cpu10:66503)WARNING: NFS41: NFS41_Bug:2361: BUG - 
Invalid BIND_CONN_TO_SESSION error: NFS4ERR_NOTSUPP

Is there session trunking available in the FreeBSD NFS41 implementation?

Br,
andi

-Original Message-
From: Rick Macklem [mailto:rmack...@uoguelph.ca] 
Sent: Montag, 5. März 2018 02:16
To: NAGY Andreas <andreas.n...@frequentis.com>; 'freebsd-stable@freebsd.org' 
<freebsd-stable@freebsd.org>
Subject: Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

NAGY Andreas wrote:
[stuff snipped]
>In the source I saw nfs_async = 0; is it right that NFS will work in async 
>mode if I >compile the kernel with nfs_async = 1?
>
>I know the risk of running it async, but is it not the same risk having the 
>datastore >connected via iSCSI which standard is also not sync?
If you want to use it, you can just set it by setting the sysctl 
vfs.nfsd.async=1.
- If the server crashes/reboots you can lose data. Also, after the reboot, the
  client will only see an temporarily unresponsive server and will not have any
  indication of data loss.
  (I am not familiar with iSCSI, so I can't comment on how safe that is.)
- If you are using ZFS, there is also a ZFS config (sync=disabled). I'm not a 
ZFS
  guy, so I don't know anything more, but I'm sure others reading this list can
  tell you how to set it.
[more stuff snipped]
>So far I did not see any issue with the mount. Only in the vmkernel.log there 
>are >often following entrees:
>WARNING: NFS41: NFS41ValidateDelegation:608: Server returned improper 
>>reason for no delegation: 2
The attached patch *might* get rid of these, although I don't think it matters 
much, since it is just complaining about the "reason" the server returns for 
not issuing a delegation (issuing delegations is entirely at the discretion of 
the server and is disabled by default).
[more stuff snipped]
Good luck with it, rick
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-04 Thread Rick Macklem
NAGY Andreas wrote:
[stuff snipped]
>In the source I saw nfs_async = 0; is it right that NFS will work in async 
>mode if I >compile the kernel with nfs_async = 1?
>
>I know the risk of running it async, but is it not the same risk having the 
>datastore >connected via iSCSI which standard is also not sync?
If you want to use it, you can just set it by setting the sysctl 
vfs.nfsd.async=1.
- If the server crashes/reboots you can lose data. Also, after the reboot, the
  client will only see an temporarily unresponsive server and will not have any
  indication of data loss.
  (I am not familiar with iSCSI, so I can't comment on how safe that is.)
- If you are using ZFS, there is also a ZFS config (sync=disabled). I'm not a 
ZFS
  guy, so I don't know anything more, but I'm sure others reading this list can
  tell you how to set it.
[more stuff snipped]
>So far I did not see any issue with the mount. Only in the vmkernel.log there 
>are >often following entrees:
>WARNING: NFS41: NFS41ValidateDelegation:608: Server returned improper >reason 
>for no delegation: 2
The attached patch *might* get rid of these, although I don't think it matters
much, since it is just complaining about the "reason" the server returns for
not issuing a delegation (issuing delegations is entirely at the discretion of
the server and is disabled by default).
[more stuff snipped]
Good luck with it, rick


wantdeleg.patch
Description: wantdeleg.patch
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-04 Thread NAGY Andreas
Okay, the slow write was not a NFS problem, it was the hw raid controller which 
switched to write through because of a broken battery.

In the source I saw nfs_async = 0; is it right that NFS will work in async mode 
if I compile the kernel with nfs_async = 1?

I know the risk of running it async, but is it not the same risk having the 
datastore connected via iSCSI which standard is also not sync?

The last weeks I tested the following setup:
Two FreeBSD hosts with a more or less good hw RAID controller in a HAST cluster 
providing a datastore to two ESXi hosts via iSCSI.
This setup worked quiet well, but I now want to switch to NFS, and hope to get 
equivalent speeds.

Thanks so far,
andi
 

-Original Message-
From: NAGY Andreas 
Sent: Sonntag, 4. März 2018 14:26
To: Rick Macklem <rmack...@uoguelph.ca>; freebsd-stable@freebsd.org
Subject: RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

Thanks, got it working with your patch.

So far I did not see any issue with the mount. Only in the vmkernel.log there 
are often following entrees:
WARNING: NFS41: NFS41ValidateDelegation:608: Server returned improper reason 
for no delegation: 2

Actually I have only a single link between the ESXi host and the FreeBSD host, 
but as soon as I figure out what Is the right way to configure multiple paths 
for NFS I will do more testing. 

I need also to check out what can be tuned. I expected that writes to the NFS 
datastore will be slower than iSCSI but not as slow as it is now.

andi 


-Original Message-
From: Rick Macklem [mailto:rmack...@uoguelph.ca]
Sent: Sonntag, 4. März 2018 06:48
To: NAGY Andreas <andreas.n...@frequentis.com>; freebsd-stable@freebsd.org
Subject: Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

NAGY Andreas wrote:
>Hi and thanks!
>
>First time using/needing a patch could you give me a short advise how to use 
>it >and for which version?
The only difference with kernel versions will be the line#s.
>So far I have made a fresh FreeBSD 11.1 RELEASE install as a VM on a 
>ESXi host >updated the system and did a svn checkout 
>http://svn.freebsd.org/base/release/11.1.0/
>
>Then tried to apply the patch in /usr/src/sys via patch < 
>/tmp/reclaimcom2.patch
>
>Output was:
>Hmm...  Looks like a unified diff to me...
>The text leading up to this was:
>--
>|--- fs/nfsserver/nfs_nfsdserv.c.savrecl2018-02-10 20:34:31.166445000 
>-0500
>|+++ fs/nfsserver/nfs_nfsdserv.c2018-02-10 20:36:07.94749 -0500
>--
>Patching file fs/nfsserver/nfs_nfsdserv.c using Plan A...
>No such line 4225 in input file, ignoring Hunk #1 succeeded at 4019 
>(offset -207 lines).
>done
Since it says "Hunk #1 succeeded...", I think it patched ok.
However, you can check by looking at nfsrvd_reclaimcomplete() in 
sys/fs/nfsserver/nfs_nfsdserv.c.
Before the patch it would look like:
if (*tl == newnfs_true)
nd->nd_repstat = NFSERR_NOTSUPP;
   else
nd->nd_repstat = nfsrv_checkreclaimcomplete(nd); 
whereas after being patched, it will look like:
  nd->nd_repstat = nfsrv_checkreclaimcomplete(nd);
   if (*tl == newnfs_true)
   nd->nd_repstat = 0;

rick
[stuff snipped]
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-04 Thread NAGY Andreas
Thanks, got it working with your patch.

So far I did not see any issue with the mount. Only in the vmkernel.log there 
are often following entrees:
WARNING: NFS41: NFS41ValidateDelegation:608: Server returned improper reason 
for no delegation: 2

Actually I have only a single link between the ESXi host and the FreeBSD host, 
but as soon as I figure out what Is the right way to configure multiple paths 
for NFS I will do more testing. 

I need also to check out what can be tuned. I expected that writes to the NFS 
datastore will be slower than iSCSI but not as slow as it is now.

andi 


-Original Message-
From: Rick Macklem [mailto:rmack...@uoguelph.ca] 
Sent: Sonntag, 4. März 2018 06:48
To: NAGY Andreas <andreas.n...@frequentis.com>; freebsd-stable@freebsd.org
Subject: Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

NAGY Andreas wrote:
>Hi and thanks!
>
>First time using/needing a patch could you give me a short advise how to use 
>it >and for which version?
The only difference with kernel versions will be the line#s.
>So far I have made a fresh FreeBSD 11.1 RELEASE install as a VM on a 
>ESXi host >updated the system and did a svn checkout 
>http://svn.freebsd.org/base/release/11.1.0/
>
>Then tried to apply the patch in /usr/src/sys via patch < 
>/tmp/reclaimcom2.patch
>
>Output was:
>Hmm...  Looks like a unified diff to me...
>The text leading up to this was:
>--
>|--- fs/nfsserver/nfs_nfsdserv.c.savrecl2018-02-10 20:34:31.166445000 
>-0500
>|+++ fs/nfsserver/nfs_nfsdserv.c2018-02-10 20:36:07.94749 -0500
>--
>Patching file fs/nfsserver/nfs_nfsdserv.c using Plan A...
>No such line 4225 in input file, ignoring Hunk #1 succeeded at 4019 
>(offset -207 lines).
>done
Since it says "Hunk #1 succeeded...", I think it patched ok.
However, you can check by looking at nfsrvd_reclaimcomplete() in 
sys/fs/nfsserver/nfs_nfsdserv.c.
Before the patch it would look like:
if (*tl == newnfs_true)
nd->nd_repstat = NFSERR_NOTSUPP;
   else
nd->nd_repstat = nfsrv_checkreclaimcomplete(nd); 
whereas after being patched, it will look like:
  nd->nd_repstat = nfsrv_checkreclaimcomplete(nd);
   if (*tl == newnfs_true)
   nd->nd_repstat = 0;

rick
[stuff snipped]
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-03 Thread Rick Macklem
NAGY Andreas wrote:
>Hi and thanks!
>
>First time using/needing a patch could you give me a short advise how to use 
>it >and for which version?
The only difference with kernel versions will be the line#s.
>So far I have made a fresh FreeBSD 11.1 RELEASE install as a VM on a ESXi host 
>>updated the system and did a svn checkout 
>http://svn.freebsd.org/base/release/11.1.0/
>
>Then tried to apply the patch in /usr/src/sys via patch < 
>/tmp/reclaimcom2.patch
>
>Output was:
>Hmm...  Looks like a unified diff to me...
>The text leading up to this was:
>--
>|--- fs/nfsserver/nfs_nfsdserv.c.savrecl2018-02-10 20:34:31.166445000 
>-0500
>|+++ fs/nfsserver/nfs_nfsdserv.c2018-02-10 20:36:07.94749 -0500
>--
>Patching file fs/nfsserver/nfs_nfsdserv.c using Plan A...
>No such line 4225 in input file, ignoring
>Hunk #1 succeeded at 4019 (offset -207 lines).
>done
Since it says "Hunk #1 succeeded...", I think it patched ok.
However, you can check by looking at nfsrvd_reclaimcomplete() in
sys/fs/nfsserver/nfs_nfsdserv.c.
Before the patch it would look like:
if (*tl == newnfs_true)
nd->nd_repstat = NFSERR_NOTSUPP;
   else
nd->nd_repstat = nfsrv_checkreclaimcomplete(nd);
whereas after being patched, it will look like:
  nd->nd_repstat = nfsrv_checkreclaimcomplete(nd);
   if (*tl == newnfs_true)
   nd->nd_repstat = 0;

rick
[stuff snipped]
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


RE: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-03 Thread NAGY Andreas
Hi and thanks!

First time using/needing a patch could you give me a short advise how to use it 
and for which version?

So far I have made a fresh FreeBSD 11.1 RELEASE install as a VM on a ESXi host 
updated the system and did a svn checkout 
http://svn.freebsd.org/base/release/11.1.0/

Then tried to apply the patch in /usr/src/sys via patch < /tmp/reclaimcom2.patch

Output was: 
Hmm...  Looks like a unified diff to me...
The text leading up to this was:
--
|--- fs/nfsserver/nfs_nfsdserv.c.savrecl2018-02-10 20:34:31.166445000 
-0500
|+++ fs/nfsserver/nfs_nfsdserv.c2018-02-10 20:36:07.94749 -0500
--
Patching file fs/nfsserver/nfs_nfsdserv.c using Plan A...
No such line 4225 in input file, ignoring
Hunk #1 succeeded at 4019 (offset -207 lines).
done

So I think this was not correct, as I also noticed that nfs_nfsdserv.c 4102 
lines.

andi



-Original Message-
From: Rick Macklem [mailto:rmack...@uoguelph.ca] 
Sent: Samstag, 3. März 2018 03:01
To: NAGY Andreas <andreas.n...@frequentis.com>; freebsd-stable@freebsd.org
Subject: Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi 
client

NAGY Andreas wrote:
>I am trying to get a FreeBSD NFS 4.1 export working with VMware Esxi 6.5u1, 
>but >it is always mounted as read only.
>
>After some research, I found out that this is a known problem, and there are 
>>threads about this from 2015 also in the mailinglist archive.
>
>As it seems VMware will not change the bahvior of there NFS 4.1 client I 
>wanted >to ask here if there is a patch or workaround for this available.
I believe the attached small patch deals with the ReclaimComplete issue.
However, someone else who tested this had additional issues with the mount:
- The client logged a couple of things (that sounded weird to me;-)
  - Something about Readdir seeing directories change too much..
  - Something about "wrong reason for not issuing a delegation"...
 (I don't what either of these are caused by or whether they result in serious
  breakage of the mount.)
They also ran into a hang when transferring a large file. It sounded to me like 
something that might be a network interface device driver issue and I suggested 
they disable TSO, LRO and jumbo frames, but I never heard back from them, so I 
don't know more about this.

So, feel free to test with the attached patch and if you run into problems with 
the mount, email w.r.t. what they are. If we persevere we might get it going ok.

rick
[stuff snipped]
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: NFS 4.1 RECLAIM_COMPLETE FS failed error in combination with ESXi client

2018-03-02 Thread Rick Macklem
NAGY Andreas wrote:
>I am trying to get a FreeBSD NFS 4.1 export working with VMware Esxi 6.5u1, 
>but >it is always mounted as read only.
>
>After some research, I found out that this is a known problem, and there are 
>>threads about this from 2015 also in the mailinglist archive.
>
>As it seems VMware will not change the bahvior of there NFS 4.1 client I 
>wanted >to ask here if there is a patch or workaround for this available.
I believe the attached small patch deals with the ReclaimComplete issue.
However, someone else who tested this had additional issues with the mount:
- The client logged a couple of things (that sounded weird to me;-)
  - Something about Readdir seeing directories change too much..
  - Something about "wrong reason for not issuing a delegation"...
 (I don't what either of these are caused by or whether they result in serious
  breakage of the mount.)
They also ran into a hang when transferring a large file. It sounded to me like
something that might be a network interface device driver issue and I suggested
they disable TSO, LRO and jumbo frames, but I never heard back from them,
so I don't know more about this.

So, feel free to test with the attached patch and if you run into problems
with the mount, email w.r.t. what they are. If we persevere we might get it
going ok.

rick
[stuff snipped]

reclaimcom2.patch
Description: reclaimcom2.patch
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"