In fixing a problem for Mellanox, I noted that we had somehow lost our default binding policy intentions - i.e., we were no longer automatically binding by default. I fixed that, and so you’d get this if the numactl and numactl-devel libs are missing.
I suspect we should eliminate those warnings if we bind by default? On the other matter: that has been the agreed-upon behavior for some time now. If you don’t specify anything, we launch the number of procs equal to the number of slots, with the slots auto-detected using hwloc and equating slots to number of discovered cores. However, if you use -host, then we assume you are telling us the number of slots=1 for every time you mention the host name. This overrides any auto-discovery. HTH Ralph > On Mar 20, 2015, at 8:16 AM, Rolf vandeVaart <rvandeva...@nvidia.com> wrote: > > Greetings: > > I am now seeing the following message for all my calls to mpirun on ompi > master. This started with last night’s MTT run. Is this intentional? > > [rvandevaart@ivy0 ~]$ mpirun -np 1 hostname > -------------------------------------------------------------------------- > WARNING: a request was made to bind a process. While the system > supports binding the process itself, at least one node does NOT > support binding memory to the process location. > > Node: ivy0 > > This usually is due to not having the required NUMA support installed > on the node. In some Linux distributions, the required support is > contained in the libnumactl and libnumactl-devel packages. > This is a warning only; your job will continue, though performance may be > degraded. > -------------------------------------------------------------------------- > ivy0.nvidia.com <http://ivy0.nvidia.com/> > > > On another note, I noticed on both 1.8 and master that we get different > number of nodes if we specify the hostname. This is not too big a deal, but > surprised me. > > [rvandevaart@ivy0 ~]$ /opt/openmpi/v1.8.4/bin/mpirun hostname > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > ivy0.nvidia.com <http://ivy0.nvidia.com/> > [rvandevaart@ivy0 ~]$ /opt/openmpi/v1.8.4/bin/mpirun -host ivy0 hostname > ivy0.nvidia.com <http://ivy0.nvidia.com/> > [rvandevaart@ivy0 ~]$ > > This email message is for the sole use of the intended recipient(s) and may > contain confidential information. Any unauthorized review, use, disclosure > or distribution is prohibited. If you are not the intended recipient, please > contact the sender by reply email and destroy all copies of the original > message. > _______________________________________________ > devel mailing list > de...@open-mpi.org <mailto:de...@open-mpi.org> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel > <http://www.open-mpi.org/mailman/listinfo.cgi/devel> > Link to this post: > http://www.open-mpi.org/community/lists/devel/2015/03/17140.php > <http://www.open-mpi.org/community/lists/devel/2015/03/17140.php>