One thing you might want to try: add this to your mpirun cmd line: --display-allocation
This will tell you how many slots we think we've been given on your cluster. On Aug 21, 2014, at 12:50 PM, Ralph Castain <r...@open-mpi.org> wrote: > Starting early in the 1.7 series, we began to bind procs by default to cores > when -np <= 2, and to sockets if np > 2. Is it possible this is what you are > seeing? > > > On Aug 21, 2014, at 12:45 PM, Andrej Prsa <aprs...@gmail.com> wrote: > >> Dear devels, >> >> I have been trying out 1.8.2rcs recently and found a show-stopping >> problem on our cluster. Running any job with any number of processors >> larger than 32 will always employ only 32 cores per node (our nodes >> have 48 cores). We are seeing identical behavior with 1.8.2rc4, >> 1.8.2rc2, and 1.8.1. Running identical programs shows no such issues >> with version 1.6.5, where all 48 cores per node are working. While our >> system is running torque/maui, the problem is evident by running mpirun >> directly. >> >> I am attaching hwloc topology in case that helps -- I am aware of a >> buggy bios code that trips hwloc, but I don't know if that might be an >> issue or not. I am happy to help debugging if you can provide me with >> guidance. >> >> Thanks, >> Andrej >> <cluster.output><cluster.tar.bz2>_______________________________________________ >> devel mailing list >> de...@open-mpi.org >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >> Link to this post: >> http://www.open-mpi.org/community/lists/devel/2014/08/15676.php >