Prakash Velayutham wrote:
when a user requests that software, he should be given both the CPUs in
the node regardless of what he requests, as the software takes up a lot of RAM 
(almost 6 - 7
GB). How can I make this condition work?

Two things should handle it:

1. In torque, set a default amount of memory requested for a normal job. It doesn't have to be realistic, but it can't be tiny, either. 1-2 GB, perhaps?

2. On your memory hog jobs, enter '#PBS -l mem=7168mb' or similar into his job files, replacing 7168mb with slightly less memory than 'pbsnodes' reports free when the node is idle.

Combining those two should allow two normal jobs to run without problems, even if they get substantially larger than 1gb each. But your hog jobs will request nearly all the memory on the system, and jobs that default to requesting 1 GB of memory will get pushed to other systems.

Granted, if someone had a tiny memory job, they could also explicitly request the 50 MB or whatever it might require and sneak onto the hog's system. But if they're estimating their needs accurately, and your hog job is really only using one CPU, there's no harm there.

--
Mike Renfro  / R&D Engineer, Center for Manufacturing Research,
931 372-3601 / Tennessee Technological University -- [EMAIL PROTECTED]
_______________________________________________
mauiusers mailing list
[email protected]
http://www.supercluster.org/mailman/listinfo/mauiusers

Reply via email to