Am 29.01.2016 um 15:31 schrieb Dennis Mungai:
> Add SHARE=FORCE to your partition settings for each partition entry in
> the configuration file.

https://computing.llnl.gov/linux/slurm/cons_res_share.html

selection setting was:
SelectType=select/cons_res
SelectTypeParameters=CR_Core_Memory

Shared=FORCE as you recommend leads to:
"
Cores are allocated to jobs. A core may run more than one job.
"

What does that have to do with the problem?
Can you elaborate on that?

/Benjamin


> On Jan 29, 2016 5:08 PM, David Roman <[email protected]> wrote:
> Hello,
> 
> I'm a newbies with SLURM. Perhaps could you help me to understand my
> mistake.
> 
> I have 2 nodes (2 sockets with 4 core per socket = 8 CPUs per node) I
> created 3 partitions
> 
> DEV with node2
> OP    with node1
> LOW with node1 and node2
> 
> I created 2 jobs
> Job_A uses 8 CPUS in partion DEV
> Job_B uses 16 CPUS in partion LOW
> 
> If I start Job_A before Job_B, all is ok. Job_A is in RUNNING state and
> Job_B is in PENDING state
> 
> BUT, If I start Job_B before Job_A. The both jobs are in RUNNING state.
> 
> Thanks for your help,
> 
> David.
> 
> 
> Here my slurm.conf without comments
> 
> ClusterName=Noveltits
> ControlMachine=slurm
> SlurmUser=slurm
> SlurmctldPort=6817
> SlurmdPort=6818
> AuthType=auth/munge
> StateSaveLocation=/tmp
> SlurmdSpoolDir=/tmp/slurmd
> SwitchType=switch/none
> MpiDefault=none
> SlurmctldPidFile=/var/run/slurmctld.pid
> SlurmdPidFile=/var/run/slurmd.pid
> ProctrackType=proctrack/pgid
> CacheGroups=0
> ReturnToService=0
> SlurmctldTimeout=300
> SlurmdTimeout=300
> InactiveLimit=0
> MinJobAge=300
> KillWait=30
> Waittime=0
> SchedulerType=sched/backfill
> SelectType=select/cons_res
> SelectTypeParameters=CR_CORE_Memory
> FastSchedule=0
> SlurmctldDebug=3
> SlurmdDebug=3
> JobCompType=jobcomp/none
> 
> PreemptMode=SUSPEND,GANG
> PreemptType=preempt/partition_prio
> 
> 
> NodeName=slurm_node[1-2] CPUs=8 SocketsPerBoard=2 CoresPerSocket=4
> ThreadsPerCore=1
> PartitionName=op  Nodes=slurm_node1     Priority=100 Default=No 
> MaxTime=INFINITE State=UP
> PartitionName=dev Nodes=slurm_node2     Priority=1   Default=yes
> MaxTime=INFINITE State=UP PreemptMode=OFF
> PartitionName=low Nodes=slurm_node[1-2] Priority=1   Default=No 
> MaxTime=INFINITE State=UP
> 
> 
> ______________________________________________________________________
> 
> This e-mail contains information which is confidential. It is intended
> only for the use of the named recipient. If you have received this
> e-mail in error, please let us know by replying to the sender, and
> immediately delete it from your system. Please note, that in these
> circumstances, the use, disclosure, distribution or copying of this
> information is strictly prohibited. KEMRI-Wellcome Trust Programme
> cannot accept any responsibility for the accuracy or completeness of
> this message as it has been transmitted over a public network. Although
> the Programme has taken reasonable precautions to ensure no viruses are
> present in emails, it cannot accept responsibility for any loss or
> damage arising from the use of the email or attachments. Any views
> expressed in this message are those of the individual sender, except
> where the sender specifically states them to be the views of
> KEMRI-Wellcome Trust Programme.
> ______________________________________________________________________

-- 
FSU Jena | JULIELab.de/Staff/Benjamin+Redling.html
vox: +49 3641 9 44323 | fax: +49 3641 9 44321

Reply via email to