Re: [OMPI users] Undefined ompi_mpi_info_null issue

2015-06-11 Thread Gilles Gouaillardet
Ray, this symbol is defined in libmpi.so. can you run ldd /N/dc2/projects/ray/quarry/h5py/h5py-2.5.0/build/lib.linux-x86_64-2.7/h5py/_errors.so and make sure this is linked with openmpi 1.8.4 ? Cheers, Gilles On 6/12/2015 1:29 AM, Ray Sheppard wrote: Hi List, I know I saw this issue year

Re: [OMPI users] Default value of btl_openib_memalign_threshold

2015-06-11 Thread Steve Wise
Hey Jeff, what did you run to generate the memory corruption? Can you run the same test with --mca btl_openib_memalign_threshold 12288 and see if you get the same corruption? I'm not hitting any corruption over iw_cxgb4 with a simple test. On 6/10/2015 2:39 PM, Jeff Squyres (jsquyres) wrote

Re: [OMPI users] Default value of btl_openib_memalign_threshold

2015-06-11 Thread Steve Wise
FYI: I opened: https://github.com/open-mpi/ompi/issues/638 to track this. Steve. On 6/10/2015 4:07 PM, Ralph Castain wrote: Done On Jun 10, 2015, at 1:55 PM, Steve Wise > wrote: If you're trying to release 1.8.6, I recommend you revert the change that

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Ralph Castain
Yeah, I’d put my money on a race condition under that scenario. I don’t have anything that large I can test on, but I’ll see what I can do > On Jun 11, 2015, at 1:17 PM, Leiter, Kenneth W CIV USARMY ARL (US) > wrote: > > Yes, each parent launches ten children and no other parents participate i

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Leiter, Kenneth W CIV USARMY ARL (US)
Yes, each parent launches ten children and no other parents participate in that spawn (i.e. the spawn uses MPI_COMM_SELF as the communicator). No threading. I am using the example from: https://github.com/bfroehle/mpi4py/tree/master/demo/spawning In lieu of my actual application which has a lo

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Ralph Castain
So to be clear: each parent launches 10 children, and no other parents participate in that spawn? And there is no threading in the app, yes? > On Jun 11, 2015, at 12:53 PM, Leiter, Kenneth W CIV USARMY ARL (US) > wrote: > > Howard, > > I do not run into a problem when I have one parent spaw

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Ralph Castain
Hmmm…from what this shows, it should have complained about being “overloaded” instead of “oversubscribed” - yes? > On Jun 11, 2015, at 12:41 PM, Howard Pritchard wrote: > > Hi Ralph, > > Attached is the map and reservaion output (I was adjusting the number of > spawned ranks using an env. v

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Leiter, Kenneth W CIV USARMY ARL (US)
Howard, I do not run into a problem when I have one parent spawning many children (tested up to 100 children ranks), but am seeing the problem when I have, for example, 8 parents launching 10 children each. - Ken From: users [users-boun...@open-mpi.org] on behal

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Howard Pritchard
Hi Ralph, Attached is the map and reservaion output (I was adjusting the number of spawned ranks using an env. variable. I had one master which spawned 23 children. Howard 2015-06-11 12:39 GMT-06:00 Ralph Castain : > Howard: could you add —display-devel-map —display-allocation and send the >

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Leiter, Kenneth W CIV USARMY ARL (US)
I am requesting a large interactive job, but not using all the nodes when executing the test. - Ken From: users [users-boun...@open-mpi.org] on behalf of Douglas L Reeder [d...@centurylink.net] Sent: Thursday, June 11, 2015 3:11 PM To: Open MPI Users Sub

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Douglas L Reeder
Ken, Why are the qsub ncpus and mpiirun -np different values. Doug On Jun 11, 2015, at 12:08 PM, Leiter, Kenneth W CIV USARMY ARL (US) wrote: > Hi Howard, > > My qsub command is: > qsub -l select=10:ncpus=32:mpiprocs=32 -q debug -l walltime=01:00:00 -I > > I have also tried using ccm mode wi

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Leiter, Kenneth W CIV USARMY ARL (US)
The ompi_info is attached. I can try with a vanilla configure, here is what I configured with: ./configure --with-alps --with-ugni --without-verbs --with-cray-xpmem --with-cray-pmi --with-udreg --without-tm --enable-debug I am using PrgEnv-intel/5.2.40 From:

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Ralph Castain
Howard: could you add —display-devel-map —display-allocation and send the output along? I’d like to see why it things you are oversubscribed. Thanks > On Jun 11, 2015, at 11:36 AM, Howard Pritchard wrote: > > Hi Ken, > > Could you post the output of your ompi_info? > > I have PrgEnv-gnu/5.2

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Howard Pritchard
Hi Ken, Could you post the output of your ompi_info? I have PrgEnv-gnu/5.2.56 and gcc/4.9.2 loaded in my env on nersc system. Following configure line: ./configure --enable-mpi-java --prefix=my_favorite_install_location The general rule of thumb on cray's with master (not with older versions th

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Leiter, Kenneth W CIV USARMY ARL (US)
I will try on a non-cray machine as well. - Ken -Original Message- From: users [mailto:users-boun...@open-mpi.org] On Behalf Of Howard Pritchard Sent: Thursday, June 11, 2015 12:21 PM To: Open MPI Users Subject: Re: [OMPI users] orted segmentation fault in pmix on master Hello Ken, Coul

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Leiter, Kenneth W CIV USARMY ARL (US)
Hi Howard, My qsub command is: qsub -l select=10:ncpus=32:mpiprocs=32 -q debug -l walltime=01:00:00 -I I have also tried using ccm mode with no change in outcome. My mpirun command is: mpirun -np 9 -debug-daemons ./parent child I have also attached the debug output for the particular daemon th

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Nathan Hjelm
Won't help him. aprun does not support dynamics. -Nathan On Thu, Jun 11, 2015 at 12:43:38PM -0400, Joshua Ladd wrote: >Ken, > >Could you try to launch the job with aprun instead of mpirun? > >Thanks, > >Josh >On Thu, Jun 11, 2015 at 12:21 PM, Howard Pritchard >wrote: >

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Joshua Ladd
Ken, Could you try to launch the job with aprun instead of mpirun? Thanks, Josh On Thu, Jun 11, 2015 at 12:21 PM, Howard Pritchard wrote: > Hello Ken, > > Could you give the details of the allocation request (qsub args) > as well as the mpirun command line args? I'm trying to reproduce > on t

[OMPI users] Undefined ompi_mpi_info_null issue

2015-06-11 Thread Ray Sheppard
Hi List, I know I saw this issue years ago but have forgotten the details. I looked through old posts but only found about half a dozen pertaining to WinDoze. I am trying to build a Python (2.7.3) extension (h5py) that calls HDF5 (1.8.14). I built both the OpenMPI (1.8.4) and the HDF5 modu

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Howard Pritchard
Hello Ken, Could you give the details of the allocation request (qsub args) as well as the mpirun command line args? I'm trying to reproduce on the nersc system. It would be interesting if you have access to a similar size non-cray cluster if you get the same problems. Howard 2015-06-11 9:13 G

Re: [OMPI users] Missing file "openmpi/ompi/mpi/f77/constants.h"

2015-06-11 Thread Dave Love
Filippo Spiga writes: > Dear OpenMPI experts, > > I am rebuilding IPM (https://github.com/nerscadmin/ipm) based on OpenMPI > 1.8.5. However, despite OMPI is compiled with the "--with-devel-headers" > option, IPM build fails because the file "openmpi/ompi/mpi/f77/constants.h" > is missing Whic

Re: [OMPI users] Looking for LAM-MPI sources to create a mirror

2015-06-11 Thread Dave Love
"Jeff Squyres (jsquyres)" writes: > Sadly, I have minimal experience with .debs... if someone would contribute > the necessary packaging, we could talk about hosting a source deb on the main > Open MPI site. What's wrong with the Debian packages (if you really want LAM)? $ apt-cache show la

Re: [OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Ralph Castain
I don’t have a Cray, but let me see if I can reproduce this on something else > On Jun 11, 2015, at 7:26 AM, Leiter, Kenneth W CIV USARMY ARL (US) > wrote: > > Hello, > > I am attempting to use the openmpi development master for a code that uses > dynamic process management (i.e. MPI_Comm_spaw

[OMPI users] orted segmentation fault in pmix on master

2015-06-11 Thread Leiter, Kenneth W CIV USARMY ARL (US)
Hello, I am attempting to use the openmpi development master for a code that uses dynamic process management (i.e. MPI_Comm_spawn) on our Cray XC40 at the Army Research Laboratory. After reading through the mailing list I came to the conclusion that the master branch is the only hope for getting t

Re: [OMPI users] building openmpi-v1.10-dev-59-g4e9cea6 still breaks

2015-06-11 Thread Siegmar Gross
Hi, > Just fixed it - thanks! > > > On Jun 10, 2015, at 10:31 AM, Nathan Hjelm wrote: > > > > Looks like a bug in the CM optimization that went into 1.10 yesterday. I > > will let the author know to fix this one. Today I could build openmpi-v1.8.5-50-g793344d, openmpi-v1.10-dev-79-g841ec20

Re: [OMPI users] Missing file "openmpi/ompi/mpi/f77/constants.h"

2015-06-11 Thread Jeff Squyres (jsquyres)
Can you send all the information listed here: http://www.open-mpi.org/community/help/ > On Jun 11, 2015, at 2:17 AM, Filippo Spiga wrote: > > Dear OpenMPI experts, > > I am rebuilding IPM (https://github.com/nerscadmin/ipm) based on OpenMPI > 1.8.5. However, despite OMPI is compiled wit

[OMPI users] Missing file "openmpi/ompi/mpi/f77/constants.h"

2015-06-11 Thread Filippo Spiga
Dear OpenMPI experts, I am rebuilding IPM (https://github.com/nerscadmin/ipm) based on OpenMPI 1.8.5. However, despite OMPI is compiled with the "--with-devel-headers" option, IPM build fails because the file "openmpi/ompi/mpi/f77/constants.h" is missing Shall I hack IPM because this is a depre