On 2016-01-29 16:08, Bruce Roberts wrote:
> Not really related to the question, but the documentation you are referencing 
> is years old.  You should probably reference the current documentation at 
> SchedMD 
> 
> http://slurm.schedmd.com
> 
> In this case 
> 
> http://slurm.schedmd.com/cons_res_share.html

Full ACK. Was leaving the office and just used the highest page ranking
to make a point.


> On January 29, 2016 6:42:24 AM PST, Benjamin Redling 
> <[email protected]> wrote:
>>
>> Am 29.01.2016 um 15:31 schrieb Dennis Mungai:
>>> Add SHARE=FORCE to your partition settings for each partition entry
>> in
>>> the configuration file.
>>
>> https://computing.llnl.gov/linux/slurm/cons_res_share.html
>>
>> selection setting was:
>> SelectType=select/cons_res
>> SelectTypeParameters=CR_Core_Memory
>>
>> Shared=FORCE as you recommend leads to:
>> "
>> Cores are allocated to jobs. A core may run more than one job.
>> "
>>
>> What does that have to do with the problem?
>> Can you elaborate on that?
>>
>> /Benjamin
>>
>>
>>> On Jan 29, 2016 5:08 PM, David Roman <[email protected]> wrote:
>>> Hello,
>>>
>>> I'm a newbies with SLURM. Perhaps could you help me to understand my
>>> mistake.
>>>
>>> I have 2 nodes (2 sockets with 4 core per socket = 8 CPUs per node) I
>>> created 3 partitions
>>>
>>> DEV with node2
>>> OP    with node1
>>> LOW with node1 and node2
>>>
>>> I created 2 jobs
>>> Job_A uses 8 CPUS in partion DEV
>>> Job_B uses 16 CPUS in partion LOW
>>>
>>> If I start Job_A before Job_B, all is ok. Job_A is in RUNNING state
>> and
>>> Job_B is in PENDING state
>>>
>>> BUT, If I start Job_B before Job_A. The both jobs are in RUNNING
>> state.
>>>
>>> Thanks for your help,
>>>
>>> David.
>>>
>>>
>>> Here my slurm.conf without comments
>>>
>>> ClusterName=Noveltits
>>> ControlMachine=slurm
>>> SlurmUser=slurm
>>> SlurmctldPort=6817
>>> SlurmdPort=6818
>>> AuthType=auth/munge
>>> StateSaveLocation=/tmp
>>> SlurmdSpoolDir=/tmp/slurmd
>>> SwitchType=switch/none
>>> MpiDefault=none
>>> SlurmctldPidFile=/var/run/slurmctld.pid
>>> SlurmdPidFile=/var/run/slurmd.pid
>>> ProctrackType=proctrack/pgid
>>> CacheGroups=0
>>> ReturnToService=0
>>> SlurmctldTimeout=300
>>> SlurmdTimeout=300
>>> InactiveLimit=0
>>> MinJobAge=300
>>> KillWait=30
>>> Waittime=0
>>> SchedulerType=sched/backfill
>>> SelectType=select/cons_res
>>> SelectTypeParameters=CR_CORE_Memory
>>> FastSchedule=0
>>> SlurmctldDebug=3
>>> SlurmdDebug=3
>>> JobCompType=jobcomp/none
>>>
>>> PreemptMode=SUSPEND,GANG
>>> PreemptType=preempt/partition_prio
>>>
>>>
>>> NodeName=slurm_node[1-2] CPUs=8 SocketsPerBoard=2 CoresPerSocket=4
>>> ThreadsPerCore=1
>>> PartitionName=op  Nodes=slurm_node1     Priority=100 Default=No 
>>> MaxTime=INFINITE State=UP
>>> PartitionName=dev Nodes=slurm_node2     Priority=1   Default=yes
>>> MaxTime=INFINITE State=UP PreemptMode=OFF
>>> PartitionName=low Nodes=slurm_node[1-2] Priority=1   Default=No 
>>> MaxTime=INFINITE State=UP
>>>
>>>
>>>
>> ______________________________________________________________________
>>>
>>> This e-mail contains information which is confidential. It is
>> intended
>>> only for the use of the named recipient. If you have received this
>>> e-mail in error, please let us know by replying to the sender, and
>>> immediately delete it from your system. Please note, that in these
>>> circumstances, the use, disclosure, distribution or copying of this
>>> information is strictly prohibited. KEMRI-Wellcome Trust Programme
>>> cannot accept any responsibility for the accuracy or completeness of
>>> this message as it has been transmitted over a public network.
>> Although
>>> the Programme has taken reasonable precautions to ensure no viruses
>> are
>>> present in emails, it cannot accept responsibility for any loss or
>>> damage arising from the use of the email or attachments. Any views
>>> expressed in this message are those of the individual sender, except
>>> where the sender specifically states them to be the views of
>>> KEMRI-Wellcome Trust Programme.
>>>
>> ______________________________________________________________________
>>
>> -- 
>> FSU Jena | JULIELab.de/Staff/Benjamin+Redling.html
>> vox: +49 3641 9 44323 | fax: +49 3641 9 44321
> 


-- 
FSU Jena | JULIELab.de/Staff/Benjamin+Redling.html
vox: +49 3641 9 44323 | fax: +49 3641 9 44321

Reply via email to