On Tue, 2008-07-01 at 07:49 +0200, Carsten Aulbert wrote:
> Hi again,
> 
> Ian Kent wrote:
> > But then it's been included in RHEL-5 from initial release and it's
> > holding up fine.
> 
> We'll try autofs5 on at leat one node possibly later today, either Jan,
> Steffen or I should succeed in backporting it (and getting around the
> LDAP problem).

btw, I discovered a mistake in one of the patches included in the
previous patch. I'll post an updated patch later today. Sorry for the
trouble.

> 
> However I still have v4 related question.
> 
> We're merciless and run this on one of the nodes:
> 
> $ cat test_mount
> #!/bin/sh
> 
> n_node=1000
> 
> for i in `seq 1 $n_node`;do
>         n=`echo $RANDOM%1342+10001 | bc| sed -e "s/1/n/"`
>         $HOME/bin/mount.sh $n&
>         echo -n .
> done
> 
> $ cat mount.sh
> #!/bin/sh
> 
> dir="/distributed/spray/data/EatH/S5R1"
> 
> ping -c1 -w1 $1 > /dev/null&& file="/atlas/node/$1$dir/"`ls -f
> /atlas/node/$1$dir/|head -n 50 | tail -n 1`
> md5sum ${file}
> 
> 
> Running this gives this in syslog:
> Jul  1 07:37:19 n1312 rpc.idmapd[2309]: nfsopen:
> open(/var/lib/nfs/rpc_pipefs/nfs/clntaa58/idmap): Too many open files
> Jul  1 07:37:19 n1312 rpc.idmapd[2309]: nfsopen:
> open(/var/lib/nfs/rpc_pipefs/nfs/clntaa58/idmap): Too many open files
> Jul  1 07:37:19 n1312 rpc.idmapd[2309]: nfsopen:
> open(/var/lib/nfs/rpc_pipefs/nfs/clntaa5e/idmap): Too many open files
> Jul  1 07:37:19 n1312 rpc.idmapd[2309]: nfsopen:
> open(/var/lib/nfs/rpc_pipefs/nfs/clntaa5e/idmap): Too many open files
> Jul  1 07:37:19 n1312 rpc.idmapd[2309]: nfsopen:
> open(/var/lib/nfs/rpc_pipefs/nfs/clntaa9c/idmap): Too many open files
> 
> Which is not surprising to me. However, there are a few things I'm
> wondering about.
> 
> (1) Shall I try the nfs list at sourceforge or is that list only full of
> spam?

Yep, for sure, the NFS maintainer in present on that list, hopefully he
will be able to help.

> (2) All our mounts use nfsvers=3 why is rpc.idmapd involved at all?

Not sure, I really must find time to get up to speed on that stuff.

> (3) Why is this daemon growing so extremely large?
> # ps aux|grep rpc.idmapd
> root      2309  0.1 16.2 2037152 1326944 ?     Ss   Jun30   1:24
> /usr/sbin/rpc.idmapd

Ditto.

> (4) The script maxes out at about 340 concurrent mounts, any idea how to
> increase this number?

Complicated question.

We can go into that further in a separate thread if you see bind to
reserved port fail messages in the log otherwise I'm not sure so we
would need to investigate.
 
> (5) Finally autofs related again: After running the script, /proc/mounts
> has these leftovers:
> 
> n0765:/local /atlas/node/n0765 nfs
> rw,vers=3,rsize=32768,wsize=32768,soft,intr,proto=tcp,timeo=600,retrans=2,sec=sys,addr=10.10.7.65
> 0 0
> n1058:/local /atlas/node/n1058 nfs
> rw,vers=3,rsize=32768,wsize=32768,soft,intr,proto=tcp,timeo=600,retrans=2,sec=sys,addr=10.10.10.58
> 0 0
> n0232:/local /atlas/node/n0232 nfs
> rw,vers=3,rsize=32768,wsize=32768,soft,intr,proto=tcp,timeo=600,retrans=2,sec=sys,addr=10.10.2.32
> 0 0
> n0409:/local /atlas/node/n0409 nfs
> rw,vers=3,rsize=32768,wsize=32768,soft,intr,proto=tcp,timeo=600,retrans=2,sec=sys,addr=10.10.4.9
> 0 0
> n0022:/local /atlas/node/n0022 nfs
> rw,vers=3,rsize=32768,wsize=32768,soft,intr,proto=tcp,timeo=600,retrans=2,sec=sys,addr=10.10.0.22
> 0 0
> n0549:/local /atlas/node/n0549 nfs
> rw,vers=3,rsize=32768,wsize=32768,soft,intr,proto=tcp,timeo=600,retrans=2,sec=sys,addr=10.10.5.49
> 0 0
> n0016:/local /atlas/node/n0016 nfs
> rw,vers=3,rsize=32768,wsize=32768,soft,intr,proto=tcp,timeo=600,retrans=2,sec=sys,addr=10.10.0.16
> 0 0
> n0975:/local /atlas/node/n0975 nfs
> rw,vers=3,rsize=32768,wsize=32768,soft,intr,proto=tcp,timeo=600,retrans=2,sec=sys,addr=10.10.9.75
> 0 0
> 
> 
> Which I need to umount manually now, remove the empty directories under
> /atlas/node before I could restart autofs.

Check if /etc/mtab is out of sync with /proc/mounts when you see this.

If so then your mount(8) mtab locking is broken otherwise it's something
else and, rather than try and dig up v4 patches, I'd recommend v5. I
haven't been able to completely resolve this in v5 yet but it is much
better so you will need to see how it goes.

Ian


_______________________________________________
autofs mailing list
[email protected]
http://linux.kernel.org/mailman/listinfo/autofs

Reply via email to