One thing you might want to try: add this to your mpirun cmd line:

--display-allocation

This will tell you how many slots we think we've been given on your cluster.

On Aug 21, 2014, at 12:50 PM, Ralph Castain <r...@open-mpi.org> wrote:

> Starting early in the 1.7 series, we began to bind procs by default to cores 
> when -np <= 2, and to sockets if np > 2. Is it possible this is what you are 
> seeing?
> 
> 
> On Aug 21, 2014, at 12:45 PM, Andrej Prsa <aprs...@gmail.com> wrote:
> 
>> Dear devels,
>> 
>> I have been trying out 1.8.2rcs recently and found a show-stopping
>> problem on our cluster. Running any job with any number of processors
>> larger than 32 will always employ only 32 cores per node (our nodes
>> have 48 cores). We are seeing identical behavior with 1.8.2rc4,
>> 1.8.2rc2, and 1.8.1. Running identical programs shows no such issues
>> with version 1.6.5, where all 48 cores per node are working. While our
>> system is running torque/maui, the problem is evident by running mpirun
>> directly.
>> 
>> I am attaching hwloc topology in case that helps -- I am aware of a
>> buggy bios code that trips hwloc, but I don't know if that might be an
>> issue or not. I am happy to help debugging if you can provide me with
>> guidance.
>> 
>> Thanks,
>> Andrej
>> <cluster.output><cluster.tar.bz2>_______________________________________________
>> devel mailing list
>> de...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/devel/2014/08/15676.php
> 

Reply via email to