?? Doesn't that tell pbs to allocate 1 node with 2 slots on it? I don't see where you get 4
Sent from my iPad > On Jul 31, 2017, at 10:00 AM, Mahmood Naderan <mahmood...@gmail.com> wrote: > > OK. The next question is how touse it with torque (PBS)? currently we write > this directive > > Nodes=1:ppn=2 > > which means 4 threads. Then we omit -np and -hostfile in the mpirun command. > >> On 31 Jul 2017 20:24, "Elken, Tom" <tom.el...@intel.com> wrote: >> Hi Mahmood, >> >> >> >> With the -hostfile case, Open MPI is trying to helpfully run things faster >> by keeping both processes on one host. Ways to avoid this… >> >> >> >> On the mpirun command line add: >> >> >> >> -pernode (runs 1 process per node), oe >> >> -npernode 1 , but these two has been deprecated in favor of the wonderful >> syntax: >> >> --map-by ppr:1:node >> >> >> >> Or you could change your hostfile to: >> >> cluster slots=1 >> compute-0-0 slots=1 >> >> >> >> >> -Tom >> >> >> >> From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Mahmood >> Naderan >> Sent: Monday, July 31, 2017 6:47 AM >> To: Open MPI Users <users@lists.open-mpi.org> >> Subject: [OMPI users] -host vs -hostfile >> >> >> >> Hi, >> >> I have stuck at a problem which I don't remember that on previous versions. >> when I run a test program with -host, it works. I mean, the process spans to >> the hosts I specified. However, when I specify -hostfile, it doesn't work!! >> >> mahmood@cluster:mpitest$ /share/apps/computer/openmpi-2.0.1/bin/mpirun -host >> compute-0-0,cluster -np 2 a.out >> **************************************************************************** >> * hwloc 1.11.2 has encountered what looks like an error from the operating >> system. >> * >> * Package (P#1 cpuset 0xffff0000) intersects with NUMANode (P#1 cpuset >> 0xff00ffff) without inclusion! >> * Error occurred in topology.c line 1048 >> * >> * The following FAQ entry in the hwloc documentation may help: >> * What should I do when hwloc reports "operating system" warnings? >> * Otherwise please report this error message to the hwloc user's mailing >> list, >> * along with the output+tarball generated by the hwloc-gather-topology >> script. >> **************************************************************************** >> Hello world from processor cluster.hpc.org, rank 1 out of 2 processors >> Hello world from processor compute-0-0.local, rank 0 out of 2 processors >> mahmood@cluster:mpitest$ cat hosts >> cluster >> compute-0-0 >> >> mahmood@cluster:mpitest$ /share/apps/computer/openmpi-2.0.1/bin/mpirun >> -hostfile hosts -np 2 a.out >> **************************************************************************** >> * hwloc 1.11.2 has encountered what looks like an error from the operating >> system. >> * >> * Package (P#1 cpuset 0xffff0000) intersects with NUMANode (P#1 cpuset >> 0xff00ffff) without inclusion! >> * Error occurred in topology.c line 1048 >> * >> * The following FAQ entry in the hwloc documentation may help: >> * What should I do when hwloc reports "operating system" warnings? >> * Otherwise please report this error message to the hwloc user's mailing >> list, >> * along with the output+tarball generated by the hwloc-gather-topology >> script. >> **************************************************************************** >> Hello world from processor cluster.hpc.org, rank 0 out of 2 processors >> Hello world from processor cluster.hpc.org, rank 1 out of 2 processors >> >> >> how can I resolve that? >> Regards, >> Mahmood >> >> >> >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org >> https://rfd.newmexicoconsortium.org/mailman/listinfo/users > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users
_______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users