No this doesn't work. When I try:

-map-by core:pe=2 -n 32
on 4 nodes
#PBS -l nodes=4:ppn=16,pmem=2gb
giving a total of 64 cores
I get
A request for multiple cpus-per-proc was given, but a directive
was also give to map to an object level that cannot support that
directive.

Please specify a mapping level that has more than one cpu, or
else let us define a default mapping that will allow multiple
cpus-per-proc.
--------------------------------------------------------------------------
[n001.cluster.com:28479] [[28163,0],0] ORTE_ERROR_LOG: Not initialized
in file util/session_dir.c at line 579
H2O-64_REC.log (END)
---
Ron Cohen
recoh...@gmail.com
skypename: ronaldcohen
twitter: @recohen3


On Fri, Mar 25, 2016 at 2:11 PM, Ronald Cohen <recoh...@gmail.com> wrote:
> or is it mpirun -map-by core:pe=8 -n 16 ?
>
> ---
> Ron Cohen
> recoh...@gmail.com
> skypename: ronaldcohen
> twitter: @recohen3
>
>
> On Fri, Mar 25, 2016 at 2:10 PM, Ronald Cohen <recoh...@gmail.com> wrote:
>> Thank you--I looked on the man page and it is not clear to me what
>> pe=2 does. Is that the number of threads? So if I want 16 mpi procs
>> with 2 threads is it for 32 cores (two nodes)
>>
>> mpirun -map-by core:pe=2 -n 16
>>
>> ?
>>
>> Sorry if I mangled this.
>>
>>
>> Ron
>>
>> ---
>> Ron Cohen
>> recoh...@gmail.com
>> skypename: ronaldcohen
>> twitter: @recohen3
>>
>>
>> On Fri, Mar 25, 2016 at 2:03 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>> Okay, what I would suggest is that you use the following cmd line:
>>>
>>> mpirun -map-by core:pe=2 (or 8 or whatever number you want)
>>>
>>> This should give you the best performance as it will tight-pack the procs 
>>> and assign them to the correct number of cores. See if that helps
>>>
>>>> On Mar 25, 2016, at 10:38 AM, Ronald Cohen <recoh...@gmail.com> wrote:
>>>>
>>>> 1.10.2
>>>>
>>>> Ron
>>>>
>>>> ---
>>>> Ron Cohen
>>>> recoh...@gmail.com
>>>> skypename: ronaldcohen
>>>> twitter: @recohen3
>>>>
>>>>
>>>> On Fri, Mar 25, 2016 at 1:30 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>>> Hmmm…what version of OMPI are you using?
>>>>>
>>>>>
>>>>> On Mar 25, 2016, at 10:27 AM, Ronald Cohen <recoh...@gmail.com> wrote:
>>>>>
>>>>> --report-bindings didn't report anything
>>>>> ---
>>>>> Ron Cohen
>>>>> recoh...@gmail.com
>>>>> skypename: ronaldcohen
>>>>> twitter: @recohen3
>>>>>
>>>>>
>>>>> On Fri, Mar 25, 2016 at 1:24 PM, Ronald Cohen <recoh...@gmail.com> wrote:
>>>>>
>>>>> —display-allocation an
>>>>> didn't seem to give useful information:
>>>>>
>>>>> ======================   ALLOCATED NODES   ======================
>>>>>       n005: slots=16 max_slots=0 slots_inuse=0 state=UP
>>>>>       n008.cluster.com: slots=16 max_slots=0 slots_inuse=0 state=UP
>>>>>       n007.cluster.com: slots=16 max_slots=0 slots_inuse=0 state=UP
>>>>>       n006.cluster.com: slots=16 max_slots=0 slots_inuse=0 state=UP
>>>>> =================================================================
>>>>>
>>>>> for
>>>>> mpirun -display-allocation  --map-by ppr:8:node -n 32
>>>>>
>>>>> Ron
>>>>>
>>>>> ---
>>>>> Ron Cohen
>>>>> recoh...@gmail.com
>>>>> skypename: ronaldcohen
>>>>> twitter: @recohen3
>>>>>
>>>>>
>>>>> On Fri, Mar 25, 2016 at 1:17 PM, Ronald Cohen <recoh...@gmail.com> wrote:
>>>>>
>>>>> Actually there was the same number of procs per node in each case. I
>>>>> verified this by logging into the nodes while they were running--in
>>>>> both cases 4 per node .
>>>>>
>>>>> Ron
>>>>>
>>>>> ---
>>>>> Ron Cohen
>>>>> recoh...@gmail.com
>>>>> skypename: ronaldcohen
>>>>> twitter: @recohen3
>>>>>
>>>>>
>>>>> On Fri, Mar 25, 2016 at 1:14 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>>>
>>>>>
>>>>> On Mar 25, 2016, at 9:59 AM, Ronald Cohen <recoh...@gmail.com> wrote:
>>>>>
>>>>> It is very strange but my program runs slower with any of these
>>>>> choices than if IO simply use:
>>>>>
>>>>> mpirun  -n 16
>>>>> with
>>>>> #PBS -l
>>>>> nodes=n013.cluster.com:ppn=4+n014.cluster.com:ppn=4+n015.cluster.com:ppn=4+n016.cluster.com:ppn=4
>>>>> for example.
>>>>>
>>>>>
>>>>> This command will tightly pack as many procs as possible on a node - note
>>>>> that we may well not see the PBS directives regarding number of ppn. Add
>>>>> —display-allocation and let’s see how many slots we think were assigned on
>>>>> each node
>>>>>
>>>>>
>>>>> The timing for the latter is 165 seconds, and for
>>>>> #PBS -l nodes=4:ppn=16,pmem=1gb
>>>>> mpirun  --map-by ppr:4:node -n 16
>>>>> it is 368 seconds.
>>>>>
>>>>>
>>>>> It will typically be faster if you pack more procs/node as they can use
>>>>> shared memory for communication.
>>>>>
>>>>>
>>>>> Ron
>>>>>
>>>>> ---
>>>>> Ron Cohen
>>>>> recoh...@gmail.com
>>>>> skypename: ronaldcohen
>>>>> twitter: @recohen3
>>>>>
>>>>>
>>>>> On Fri, Mar 25, 2016 at 12:43 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>>>
>>>>>
>>>>> On Mar 25, 2016, at 9:40 AM, Ronald Cohen <recoh...@gmail.com> wrote:
>>>>>
>>>>> Thank you! I will try it!
>>>>>
>>>>>
>>>>> What would
>>>>> -cpus-per-proc  4 -n 16
>>>>> do?
>>>>>
>>>>>
>>>>> This would bind each process to 4 cores, filling each node with procs 
>>>>> until
>>>>> the cores on that node were exhausted, to a total of 16 processes within 
>>>>> the
>>>>> allocation.
>>>>>
>>>>>
>>>>> Ron
>>>>> ---
>>>>> Ron Cohen
>>>>> recoh...@gmail.com
>>>>> skypename: ronaldcohen
>>>>> twitter: @recohen3
>>>>>
>>>>>
>>>>> On Fri, Mar 25, 2016 at 12:38 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>>>
>>>>> Add -rank-by node to your cmd line. You’ll still get 4 procs/node, but 
>>>>> they
>>>>> will be ranked by node instead of consecutively within a node.
>>>>>
>>>>>
>>>>>
>>>>> On Mar 25, 2016, at 9:30 AM, Ronald Cohen <recoh...@gmail.com> wrote:
>>>>>
>>>>> I am using
>>>>>
>>>>> mpirun  --map-by ppr:4:node -n 16
>>>>>
>>>>> and this loads the processes in round robin fashion. This seems to be
>>>>> twice as slow for my code as loading them node by node, 4 processes
>>>>> per node.
>>>>>
>>>>> How can I not load them round robin, but node by node?
>>>>>
>>>>> Thanks!
>>>>>
>>>>> Ron
>>>>>
>>>>>
>>>>> ---
>>>>> Ron Cohen
>>>>> recoh...@gmail.com
>>>>> skypename: ronaldcohen
>>>>> twitter: @recohen3
>>>>>
>>>>> ---
>>>>> Ronald Cohen
>>>>> Geophysical Laboratory
>>>>> Carnegie Institution
>>>>> 5251 Broad Branch Rd., N.W.
>>>>> Washington, D.C. 20015
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post:
>>>>> http://www.open-mpi.org/community/lists/users/2016/03/28828.php
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post:
>>>>> http://www.open-mpi.org/community/lists/users/2016/03/28829.php
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post:
>>>>> http://www.open-mpi.org/community/lists/users/2016/03/28830.php
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post:
>>>>> http://www.open-mpi.org/community/lists/users/2016/03/28831.php
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post:
>>>>> http://www.open-mpi.org/community/lists/users/2016/03/28832.php
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post:
>>>>> http://www.open-mpi.org/community/lists/users/2016/03/28833.php
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post:
>>>>> http://www.open-mpi.org/community/lists/users/2016/03/28837.php
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post:
>>>>> http://www.open-mpi.org/community/lists/users/2016/03/28840.php
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> Link to this post: 
>>>> http://www.open-mpi.org/community/lists/users/2016/03/28843.php
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/users/2016/03/28844.php

Reply via email to