Hi,

Am 07.12.2012 um 11:16 schrieb Arnau Bria:

> I've configured our cluster in the way that slots

slots are consumable by default in `qconf -sc`.


> /memory are consumable
> resources. Our nodes have their limits and there are some default
> resources requirements at job submission. All this conf should avoid
> memory/processor oversubscription (at least, from what I've read).
> something like http://jeetworks.org/node/93 ... is this the recommended
> way for avoiding over-subscription?

Yes. OTOH you can use an RQS.

But instead of the intermediate files the author provided I would prefer the 
command similar to:

$ for node in node{01..24}; do qconf -mattr exechost complex_values slots=8 
$node; done

(You need Bash 4 for leading zeros here.)

NB: default requests for consumables I would put in the complex definition.


> I've also configured core-binding, and the default for each job is 1
> slot. 
> 
> But with this conf I have some questions:
> 
> 1.-) when submitting a job specifying more than 1 job slot (-l slots=2
> -binding linear:2), OGS fails and suggest to use parallel environment.

Yep.


> I've read
> somewhere that this is OGS design, so I need a pe. I've not found a
> clear doc about pe (yes, how to create and manage, but not a complete
> definition of each parameter and its implications) so, anyone could
> share some doc about it?

The definitions of the entries are in the man page. What is unclear therein?


> what is the minimun conf I need for allowing
> slots requirement al job submission? somethign like:
> 
> $ qconf -sp smp
> pe_name            smp
> slots              1024
> user_lists         NONE
> xuser_lists        NONE
> start_proc_args    NONE
> stop_proc_args     NONE
> allocation_rule    $pe_slots

With this setting you are limited to one host, but this is correct for smp.


> control_slaves     FALSE

For tightly integrated jobs the above needs to be set to TRUE in case you want 
to issue a local startup of a second process (or going to a second node). For 
forks and threads instead it doesn't matter.


> job_is_first_task  TRUE
> urgency_slots      min
> accounting_summary FALSE
> 
> 
> is enough? (this is what I'm using and works fine)

Yes.


> 2.-) I've still not configured share based priorities, but when done,
> if a user request more slots/memory than the default, but it does not
> use them, is this request taken into account for share calculation?
> I mean, user A request 8GB and 1 cpu and uses 8GB and 3600 sec of cpu,

h_cpu or h_rt?


> and user B requests 16GB and 2 cpus, but uses 8GB and 3600 of cpu. Are
> both users priorities recalculated by resource usage or by resource
> requests?

Depends of the setting: ACCT_RESERVED_USAGE and SHARETREE_RESERVED_USAGE in 
SGE's configuration (`man sge_conf`).

-- Reuti


> TIA,
> Arnau
> _______________________________________________
> users mailing list
> [email protected]
> https://gridengine.org/mailman/listinfo/users


_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users

Reply via email to