Re: [Lustre-discuss] lru_size very small

2008-08-23 Thread Andreas Dilger
On Aug 22, 2008  15:39 -0400, Brock Palen wrote:
 It looks like lru_size is not a static parameter.  While on most of  
 our hosts it starts as zero.  Once the file system is accessed some  
 the values start to rise.  The values get highest for the MDS.
 
 cat nobackup-MDT-mdc-01022c433800/lru_size
   3877

Yes, in 1.6.5 instead of having a static LRU size it is dynamic based
on load.  This optimizes the number of locks available to nodes that
have very different workloads than others (e.g. login/build nodes vs.
compute nodes vs. backup nodes).

 So in 1.6.5.1  are lock dynamically adjusted based on ram available  
 on the MDS/OSS's?  Notice how the value above is _much_ higher than  
 the default '100' in the manual.

The total number of locks available are now a function of the RAM
on the server.  I think the maximum is 50 locks/MB, but this is
hooked into the kernel VM so that in case of too much memory pressure
then the LRU size is shrunk.

 I should point out this value was 0  till I did a   'find . | wc -l'   
 in a directory.  The same is for regular access.  users on nodes that  
 access lustre have locks.  Nodes that have not had lustre access yet  
 are still 0  (by access I mean an application that uses our lustre  
 mount vs our NFS mount.)
 
 Any feedback on the nature of locks and lru_size?
 We are looking to do what the manual says about upping the number on  
 the login nodes.

Yes, the manual needs an update.

Cheers, Andreas
--
Andreas Dilger
Sr. Staff Engineer, Lustre Group
Sun Microsystems of Canada, Inc.

___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] lru_size very small

2008-08-23 Thread Brock Palen
Great!

So I read this as being lru_size no-longer needs to be manually  
adjusted.  Thats great!
Thanks!

Brock Palen
www.umich.edu/~brockp
Center for Advanced Computing
[EMAIL PROTECTED]
(734)936-1985



On Aug 23, 2008, at 7:22 AM, Andreas Dilger wrote:
 On Aug 22, 2008  15:39 -0400, Brock Palen wrote:
 It looks like lru_size is not a static parameter.  While on most of
 our hosts it starts as zero.  Once the file system is accessed some
 the values start to rise.  The values get highest for the MDS.

 cat nobackup-MDT-mdc-01022c433800/lru_size
   3877

 Yes, in 1.6.5 instead of having a static LRU size it is dynamic based
 on load.  This optimizes the number of locks available to nodes that
 have very different workloads than others (e.g. login/build nodes vs.
 compute nodes vs. backup nodes).

 So in 1.6.5.1  are lock dynamically adjusted based on ram available
 on the MDS/OSS's?  Notice how the value above is _much_ higher than
 the default '100' in the manual.

 The total number of locks available are now a function of the RAM
 on the server.  I think the maximum is 50 locks/MB, but this is
 hooked into the kernel VM so that in case of too much memory pressure
 then the LRU size is shrunk.

 I should point out this value was 0  till I did a   'find . | wc -l'
 in a directory.  The same is for regular access.  users on nodes that
 access lustre have locks.  Nodes that have not had lustre access yet
 are still 0  (by access I mean an application that uses our lustre
 mount vs our NFS mount.)

 Any feedback on the nature of locks and lru_size?
 We are looking to do what the manual says about upping the number on
 the login nodes.

 Yes, the manual needs an update.

 Cheers, Andreas
 --
 Andreas Dilger
 Sr. Staff Engineer, Lustre Group
 Sun Microsystems of Canada, Inc.




___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] lru_size very small

2008-08-21 Thread Brock Palen
Sorry for throwing up so many quick questions on the list in a short  
time.

Looking at the manual about locking,  the manual states

The default value of LRU size is 100

I looked on our login nodes to increase its value, currently lustre  
set lru_size to 32 for the MDS and 1 for 9 of the OST's, 3 for 1 OST,  
4 for 1 OST and 0 for 3 OST's.

I should note though that all 14 OST's are spread across two OSS,  
both with 16GB of ram (x4500's).

Compared to what the manual says this sounds really small.
Would this be a sign that we don't have enough memory in our OSS/ 
MDS's  for our number of clients?

I looked on a few of our clients, many only have 1 lru_size for the  
MDS and 0 for all the OST's.

Am I reading something wrong?  Or do we have to set this at start up,  
not let lustre figure it out from clients/ram  as stated in the manual.

This state worries me because it gives me the felling the cache will  
not function at all because of the lack of available locks.  I don't  
want to end up on the wrong end of can speed up Lustre dramatically.

Thanks.

633 clients,
16 GB MDS/MGS
2x16GB OSS's.


Brock Palen
www.umich.edu/~brockp
Center for Advanced Computing
[EMAIL PROTECTED]
(734)936-1985



___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss