Re: [OMPI users] Order of ranks in mpirun

2019-05-15 Thread Ralph Castain via users
> On May 15, 2019, at 7:18 PM, Adam Sylvester via users > wrote: > > Up to this point, I've been running a single MPI rank per physical host > (using multithreading within my application to use all available cores). I > use this command: > mpirun -N 1 --bind-to none --hostfile hosts.txt >

Re: [OMPI users] Errors using contexts with OSHMEM 4.1.0

2019-05-17 Thread Ralph Castain via users
Hi Lee Ann I fear so - and assign it to @hoopoepg , @brminich and @yosefe Ralph > On May 17, 2019, at 11:14 AM, Riesen, Lee Ann via users > wrote: > > I haven't received a reply to this. Should I submit a bug report? Lee Ann > > - > Lee Ann Riesen, Enterprise and Government Group,

Re: [OMPI users] process mapping

2019-06-21 Thread Ralph Castain via users
On Jun 21, 2019, at 1:52 PM, Noam Bernstein mailto:noam.bernst...@nrl.navy.mil> > wrote: On Jun 21, 2019, at 4:45 PM, Ralph Castain mailto:r...@open-mpi.org> > wrote: Hilarious - I wrote that code and I have no idea who added that option or what it is supposed to do. I can assure, however,

Re: [OMPI users] process mapping

2019-06-21 Thread Ralph Castain via users
, at 1:43 PM, Noam Bernstein mailto:noam.bernst...@nrl.navy.mil> > wrote: On Jun 21, 2019, at 4:04 PM, Ralph Castain via users mailto:users@lists.open-mpi.org> > wrote: I’m unaware of any “map-to cartofile” option, nor do I find it in mpirun’s help or man page. Are you seeing it somew

Re: [OMPI users] process mapping

2019-06-21 Thread Ralph Castain via users
I’m unaware of any “map-to cartofile” option, nor do I find it in mpirun’s help or man page. Are you seeing it somewhere? On Jun 21, 2019, at 12:43 PM, Noam Bernstein via users mailto:users@lists.open-mpi.org> > wrote: Hi - are there any examples of the cartofile format?  Or is there some

Re: [OMPI users] How is the rank determined (Open MPI and Podman)

2019-07-24 Thread Ralph Castain via users
rent MODEX keys > are used. It seems like MODEX can not fetch messages in another order > than it was sent. Is that so? > > Not sure how to tell the other processes to not use CMA, while some > processes are still transmitting their user namespace ID to PROC 0. > >

Re: [OMPI users] How is the rank determined (Open MPI and Podman)

2019-07-22 Thread Ralph Castain via users
If that works, then it might be possible to include the namespace ID in the job-info provided by PMIx at startup - would have to investigate, so please confirm that the modex option works first. > On Jul 22, 2019, at 1:22 AM, Gilles Gouaillardet via users > wrote: > > Adrian, > > > An

Re: [OMPI users] TMPDIR for running openMPI job under grid

2019-07-26 Thread Ralph Castain via users
Upgrade to OMPI v4 or at least something in the v3 series. If you continue to have a problem, then set PMIX_MCA_ptl=tcp in your environment. On Jul 26, 2019, at 12:12 PM, Kulshrestha, Vipul via users mailto:users@lists.open-mpi.org> > wrote: Hi,  I am trying to setup my open-mpi application

Re: [OMPI users] MPI_Comm_Spawn failure: All nodes already filled

2019-08-07 Thread Ralph Castain via users
Yeah, we do currently require that to be true. Process mapping is distributed across the daemons - i.e., the daemon on each node independently computes the map. We have talked about picking up the hostfile on the head node and sending out the contents, but haven't implemented that yet. On Aug

[OMPI users] SGE Users/Dev: Request

2019-07-26 Thread Ralph Castain via users
I just wanted to address a question to the SGE users and/or developers on this list. As you may know, we have been developing PMIx for the last few years and have now integrated it into various RMs. This allows the RMs to directly launch application processes without going through mpirun and

Re: [OMPI users] MPI_Comm_Spawn failure: All nodes already filled

2019-08-06 Thread Ralph Castain via users
I'm afraid I cannot replicate this problem on OMPI master, so it could be something different about OMPI 4.0.1 or your environment. Can you download and test one of the nightly tarballs from the "master" branch and see if it works for you? https://www.open-mpi.org/nightly/master/ Ralph On

Re: [OMPI users] OMPI was not built with SLURM's PMI support

2019-08-08 Thread Ralph Castain via users
Did you configure Slurm to use PMIx? If so, then you simply need to set the "--mpi=pmix" or "--mpi=pmix_v2" (depending on which version of PMIx you used) flag on your srun cmd line so it knows to use it. If not (and you can't fix it), then you have to explicitly configure OMPI to use Slurm's

Re: [OMPI users] OMPI was not built with SLURM's PMI support

2019-08-09 Thread Ralph Castain via users
Artem - do you have any suggestions? On Aug 8, 2019, at 12:06 PM, Jing Gong mailto:gongj...@kth.se> > wrote: Hi Ralph, $ Did you remember to add "--mpi=pmix" to your srun cmd line? On the cluster, $ srun  --mpi=list srun: MPI types are... srun: none srun: openmpi srun: pmi2 srun: pmix srun:

Re: [OMPI users] OMPI was not built with SLURM's PMI support

2019-08-09 Thread Ralph Castain via users
ailed   --> Returned value Not found (-13) instead of ORTE_SUCCESS -- What is the issue? Thanks a lot. /Jing From: users mailto:users-boun...@lists.open-mpi.org> > on behalf of Ra

Re: [OMPI users] Unable to run a python code on cluster with mpirun in parallel

2019-09-09 Thread Ralph Castain via users
Take a look at "man orte_hosts" for a full explanation of how to use hostfile - /etc/hosts is not a properly formatted hostfile. You really just want a file that lists the names of the hosts, one per line, as that is the simplest hostfile. > On Sep 7, 2019, at 4:23 AM, Sepinoud Azimi via users