On Sep 26, 2012, at 11:34 AM, Jérémie Dubois-Lacoste wrote:

> We have typically ~80 clients running.
> In our case the locks were eating memory on the oss, not the mds, but
> yes it might
> still have the same cause.

Hmmm.  With only 80 clients I wouldn't expect that locks would use up too much 
memory.  But the number of cached locks is controlled on a per target basis, so 
if the OSS node has quite a few OSTs, the aggregate number of locks could get 
large.

> I'm not sure how to find the lifespan and number of locks per node. I
> think we're using the default settings. Do you know how to check this?

Take a look at /proc/fs/lustre/ldlm/namespaces/*/{lru_max_age,lru_size}. The 
lru_max_age file shows the amount of time a lock will be cached before it ages 
out.  The lru_size file shows the max number of locks that will be cached.

-- 
Rick Mohr
Senior HPC System Administrator
National Institute for Computational Sciences
http://www.nics.tennessee.edu


_______________________________________________
Lustre-discuss mailing list
[email protected]
http://lists.lustre.org/mailman/listinfo/lustre-discuss

Reply via email to