Doug Rabson wrote:

On 10 Apr 2008, at 11:27, Kris Kennaway wrote:

Doug Rabson wrote:
On 10 Apr 2008, at 10:55, Kris Kennaway wrote:
Doug Rabson wrote:
dfr         2008-04-10 09:39:29 UTC
FreeBSD src repository
Modified files:        (Branch: RELENG_7)
  lib/libc/gen         lockf.c

Hmm, I am concerned by this quick merge because rpc.lockd is still not usable for me on HEAD. I had to disable it entirely on all of my client machines because it will usually hang and become unkillable when locking files over NFS (when interoperating with a 7.0 server). I haven't had time to get the tcpdump trace you requested because of focusing on other problems. Because of the client problems I haven't tested it at all on the server side yet.

I really don't think this has received sufficient testing yet.
I'm pretty sure I fixed that problem with my change to src/lib/libc/rpc/svc_vc.c revision 1.27.

OK, I will retry, but that would just mean I can start testing. Let's hope the rest of the code is bug-free :)

Thanks. If problems show up with RELENG_7 users, I can disable use of the kernel NLM without necessarily having to back out the whole thing. Lets see what shows up - I'll hold off committing to RELENG_6 for now.




After cvs update and rebuilding kernel, includes, libc and rpc.lockd, I still get:

hydra2# rpc.lockd
Can't find or load kernel support for rpc.lockd - using non-kernel implementation
hydra2# ps aux | grep lockd
root 1217 0.0 0.0 3212 1188 ?? Ss 12:21PM 0:00.00 rpc.lockd: server (rpc.lockd) daemon 1218 0.0 0.0 3212 1360 ?? S 12:21PM 0:00.00 rpc.lockd: client (rpc.lockd) root 1219 0.0 0.0 3212 1188 ?? D 12:21PM 0:00.00 rpc.lockd: server (rpc.lockd)
root    1221  0.0  0.0   552   352  p0  D+   12:21PM   0:00.00 grep lockd
hydra2# vi /etc/rc.conf
load: 0.95  cmd: vi 1222 [lockd] 0.00u 0.00s 0% 1420k

/etc is on a NFS-mounted /

Kris
_______________________________________________
[email protected] mailing list
http://lists.freebsd.org/mailman/listinfo/cvs-all
To unsubscribe, send any mail to "[EMAIL PROTECTED]"

Reply via email to