Hmmm…add —report-bindings to your cmd line. The only thing I can think would be 
that the binding is different

> On Mar 25, 2016, at 10:17 AM, Ronald Cohen <recoh...@gmail.com> wrote:
> 
> Actually there was the same number of procs per node in each case. I
> verified this by logging into the nodes while they were running--in
> both cases 4 per node .
> 
> Ron
> 
> ---
> Ron Cohen
> recoh...@gmail.com
> skypename: ronaldcohen
> twitter: @recohen3
> 
> 
> On Fri, Mar 25, 2016 at 1:14 PM, Ralph Castain <r...@open-mpi.org> wrote:
>> 
>>> On Mar 25, 2016, at 9:59 AM, Ronald Cohen <recoh...@gmail.com> wrote:
>>> 
>>> It is very strange but my program runs slower with any of these
>>> choices than if IO simply use:
>>> 
>>> mpirun  -n 16
>>> with
>>> #PBS -l 
>>> nodes=n013.cluster.com:ppn=4+n014.cluster.com:ppn=4+n015.cluster.com:ppn=4+n016.cluster.com:ppn=4
>>> for example.
>> 
>> This command will tightly pack as many procs as possible on a node - note 
>> that we may well not see the PBS directives regarding number of ppn. Add 
>> —display-allocation and let’s see how many slots we think were assigned on 
>> each node
>> 
>>> 
>>> The timing for the latter is 165 seconds, and for
>>> #PBS -l nodes=4:ppn=16,pmem=1gb
>>> mpirun  --map-by ppr:4:node -n 16
>>> it is 368 seconds.
>> 
>> It will typically be faster if you pack more procs/node as they can use 
>> shared memory for communication.
>> 
>>> 
>>> Ron
>>> 
>>> ---
>>> Ron Cohen
>>> recoh...@gmail.com
>>> skypename: ronaldcohen
>>> twitter: @recohen3
>>> 
>>> 
>>> On Fri, Mar 25, 2016 at 12:43 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>> 
>>>>> On Mar 25, 2016, at 9:40 AM, Ronald Cohen <recoh...@gmail.com> wrote:
>>>>> 
>>>>> Thank you! I will try it!
>>>>> 
>>>>> 
>>>>> What would
>>>>> -cpus-per-proc  4 -n 16
>>>>> do?
>>>> 
>>>> This would bind each process to 4 cores, filling each node with procs 
>>>> until the cores on that node were exhausted, to a total of 16 processes 
>>>> within the allocation.
>>>> 
>>>>> 
>>>>> Ron
>>>>> ---
>>>>> Ron Cohen
>>>>> recoh...@gmail.com
>>>>> skypename: ronaldcohen
>>>>> twitter: @recohen3
>>>>> 
>>>>> 
>>>>> On Fri, Mar 25, 2016 at 12:38 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>>>> Add -rank-by node to your cmd line. You’ll still get 4 procs/node, but 
>>>>>> they will be ranked by node instead of consecutively within a node.
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>>> On Mar 25, 2016, at 9:30 AM, Ronald Cohen <recoh...@gmail.com> wrote:
>>>>>>> 
>>>>>>> I am using
>>>>>>> 
>>>>>>> mpirun  --map-by ppr:4:node -n 16
>>>>>>> 
>>>>>>> and this loads the processes in round robin fashion. This seems to be
>>>>>>> twice as slow for my code as loading them node by node, 4 processes
>>>>>>> per node.
>>>>>>> 
>>>>>>> How can I not load them round robin, but node by node?
>>>>>>> 
>>>>>>> Thanks!
>>>>>>> 
>>>>>>> Ron
>>>>>>> 
>>>>>>> 
>>>>>>> ---
>>>>>>> Ron Cohen
>>>>>>> recoh...@gmail.com
>>>>>>> skypename: ronaldcohen
>>>>>>> twitter: @recohen3
>>>>>>> 
>>>>>>> ---
>>>>>>> Ronald Cohen
>>>>>>> Geophysical Laboratory
>>>>>>> Carnegie Institution
>>>>>>> 5251 Broad Branch Rd., N.W.
>>>>>>> Washington, D.C. 20015
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>> Link to this post: 
>>>>>>> http://www.open-mpi.org/community/lists/users/2016/03/28828.php
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> Link to this post: 
>>>>>> http://www.open-mpi.org/community/lists/users/2016/03/28829.php
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post: 
>>>>> http://www.open-mpi.org/community/lists/users/2016/03/28830.php
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> Link to this post: 
>>>> http://www.open-mpi.org/community/lists/users/2016/03/28831.php
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/users/2016/03/28832.php
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/users/2016/03/28833.php
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2016/03/28834.php

Reply via email to