Can you change your consumable resources from CR_Core_Memory to CR_CPU_Memory?

On Jan 29, 2016 5:42 PM, Benjamin Redling <[email protected]> wrote:

Am 29.01.2016 um 15:31 schrieb Dennis Mungai:
> Add SHARE=FORCE to your partition settings for each partition entry in
> the configuration file.

https://computing.llnl.gov/linux/slurm/cons_res_share.html

selection setting was:
SelectType=select/cons_res
SelectTypeParameters=CR_Core_Memory

Shared=FORCE as you recommend leads to:
"
Cores are allocated to jobs. A core may run more than one job.
"

What does that have to do with the problem?
Can you elaborate on that?

/Benjamin


> On Jan 29, 2016 5:08 PM, David Roman <[email protected]> wrote:
> Hello,
>
> I'm a newbies with SLURM. Perhaps could you help me to understand my
> mistake.
>
> I have 2 nodes (2 sockets with 4 core per socket = 8 CPUs per node) I
> created 3 partitions
>
> DEV with node2
> OP    with node1
> LOW with node1 and node2
>
> I created 2 jobs
> Job_A uses 8 CPUS in partion DEV
> Job_B uses 16 CPUS in partion LOW
>
> If I start Job_A before Job_B, all is ok. Job_A is in RUNNING state and
> Job_B is in PENDING state
>
> BUT, If I start Job_B before Job_A. The both jobs are in RUNNING state.
>
> Thanks for your help,
>
> David.
>
>
> Here my slurm.conf without comments
>
> ClusterName=Noveltits
> ControlMachine=slurm
> SlurmUser=slurm
> SlurmctldPort=6817
> SlurmdPort=6818
> AuthType=auth/munge
> StateSaveLocation=/tmp
> SlurmdSpoolDir=/tmp/slurmd
> SwitchType=switch/none
> MpiDefault=none
> SlurmctldPidFile=/var/run/slurmctld.pid
> SlurmdPidFile=/var/run/slurmd.pid
> ProctrackType=proctrack/pgid
> CacheGroups=0
> ReturnToService=0
> SlurmctldTimeout=300
> SlurmdTimeout=300
> InactiveLimit=0
> MinJobAge=300
> KillWait=30
> Waittime=0
> SchedulerType=sched/backfill
> SelectType=select/cons_res
> SelectTypeParameters=CR_CORE_Memory
> FastSchedule=0
> SlurmctldDebug=3
> SlurmdDebug=3
> JobCompType=jobcomp/none
>
> PreemptMode=SUSPEND,GANG
> PreemptType=preempt/partition_prio
>
>
> NodeName=slurm_node[1-2] CPUs=8 SocketsPerBoard=2 CoresPerSocket=4
> ThreadsPerCore=1
> PartitionName=op  Nodes=slurm_node1     Priority=100 Default=No
> MaxTime=INFINITE State=UP
> PartitionName=dev Nodes=slurm_node2     Priority=1   Default=yes
> MaxTime=INFINITE State=UP PreemptMode=OFF
> PartitionName=low Nodes=slurm_node[1-2] Priority=1   Default=No
> MaxTime=INFINITE State=UP
>
>
> ______________________________________________________________________
>
> This e-mail contains information which is confidential. It is intended
> only for the use of the named recipient. If you have received this
> e-mail in error, please let us know by replying to the sender, and
> immediately delete it from your system. Please note, that in these
> circumstances, the use, disclosure, distribution or copying of this
> information is strictly prohibited. KEMRI-Wellcome Trust Programme
> cannot accept any responsibility for the accuracy or completeness of
> this message as it has been transmitted over a public network. Although
> the Programme has taken reasonable precautions to ensure no viruses are
> present in emails, it cannot accept responsibility for any loss or
> damage arising from the use of the email or attachments. Any views
> expressed in this message are those of the individual sender, except
> where the sender specifically states them to be the views of
> KEMRI-Wellcome Trust Programme.
> ______________________________________________________________________

______________________________________________________________________

This e-mail contains information which is confidential. It is intended only for 
the use of the named recipient. If you have received this e-mail in error, 
please let us know by replying to the sender, and immediately delete it from 
your system.  Please note, that in these circumstances, the use, disclosure, 
distribution or copying of this information is strictly prohibited. 
KEMRI-Wellcome Trust Programme cannot accept any responsibility for the  
accuracy or completeness of this message as it has been transmitted over a 
public network. Although the Programme has taken reasonable precautions to 
ensure no viruses are present in emails, it cannot accept responsibility for 
any loss or damage arising from the use of the email or attachments. Any views 
expressed in this message are those of the individual sender, except where the 
sender specifically states them to be the views of KEMRI-Wellcome Trust 
Programme.
______________________________________________________________________
--
FSU Jena | JULIELab.de/Staff/Benjamin+Redling.html
vox: +49 3641 9 44323 | fax: +49 3641 9 44321

Reply via email to