On Sat, Sep 27, 2008 at 8:43 PM, Mag Gam <[EMAIL PROTECTED]> wrote:
> Hello:
>
> We have implemented RHEL 5.2.1 at our university. We use the systems
> which are 8 cores with 64GB of memory. We are on a 10g network. These
> servers are mostly for computational purposes. That essentially means
> read data process the data and continue. We don't want to keep the
> data in memory since we read files which are about 250kb but there are
> over 1 million per directory.
>
> Are there any tuning parameter we can set for this type of setup?
>
> I can elaborate if you have any more questions.

There's some pretty good docs on the interwebs.  I believe that some
guys from RH did a pres @ the RH Summit this year - you might stumble
on a copy of that as well....

I believe that one param that will help is "swappiness" - default is
60, tweaking it determines how the kernel deals with swapping pages in
and out of memory.  You'd prolly want it set at 100 whereas
application guys want it at 0 so stuff never leaves memory.  Not sure
how it might apply to a million tiny files tho...

_______________________________________________
rhelv5-list mailing list
[email protected]
https://www.redhat.com/mailman/listinfo/rhelv5-list

Reply via email to