Re: [OMPI users] ERROR: C_FUNLOC function

2014-12-17 Thread Jeff Squyres (jsquyres)
Siegmar -- This fix was just pushed to the OMPI master. A new master tarball should be available shortly (probably within an hour or so -- look for a tarball dated Dec 17 at http://www.open-mpi.org/nightly/master/). I anticipate that this fix will also make it in for the v1.8.4 release (see

Re: [OMPI users] OMPI users] OMPI users] OMPI users] MPI inside MPI (still)

2014-12-17 Thread Alex A. Schmidt
Ralph, I am afraid I will have to insist on i/o redirection matter for the spawnee process. I have a "child" mpi code that do just 2 things: read the 3 parameters passed to it and print them, and then read data from stdin and show it. So, if "stdin_file" is a text file with two lines, say: 10

Re: [OMPI users] OMPI users] OMPI users] OMPI users] MPI inside MPI (still)

2014-12-17 Thread Ralph Castain
Have you tried putting the "<" as a separate parameter? In other words, since you are specifying the argv, you have to specify each of them separately. So it looks more like: "mpirun", "-n", "1", "myapp", "<", "stdinfile" Does that work? Ralph On Wed, Dec 17, 2014 at 8:07 AM, Alex A. Schmidt

Re: [OMPI users] ERROR: C_FUNLOC function

2014-12-17 Thread Siegmar Gross
Hi Jeff, > This fix was just pushed to the OMPI master. A new master tarball > should be available shortly (probably within an hour or so -- look > for a tarball dated Dec 17 at http://www.open-mpi.org/nightly/master/). Yes, I could build it now. Thank you very much to everybody who helped to

Re: [OMPI users] OMPI users] OMPI users] OMPI users] MPI inside MPI (still)

2014-12-17 Thread Alex A. Schmidt
Ralph, Sorry, "<" as an element of argv to mpi_comm_spawn is interpreted just the same, as another parameter by the spawnee process. But I am confused: wouldn't it be redundant to put "mpirun" "-n" "1" "myapp" as elements of argv, considering role of the other parameters of mpi_comm_spawn like

Re: [OMPI users] OMPI users] OMPI users] OMPI users] MPI inside MPI (still)

2014-12-17 Thread George Bosilca
I don't think this has any chance of working. The redirection is something interpreted by the shell, and when Open MPI "fork-exec" a process it does not behave as the shell. Thus a potentially non-portable solution would be to instead of launching the mpirun directly to launch it through a shell.

Re: [OMPI users] ERROR: C_FUNLOC function

2014-12-17 Thread Jeff Squyres (jsquyres)
Siegmar -- I filed https://github.com/open-mpi/ompi/issues/317 and https://github.com/open-mpi/ompi/issues/318. On Dec 17, 2014, at 3:33 PM, Siegmar Gross wrote: > Hi Jeff, > >> This fix was just pushed to the OMPI master. A new master tarball >>

Re: [OMPI users] OMPI users] OMPI users] OMPI users] MPI inside MPI (still)

2014-12-17 Thread Alex A. Schmidt
Putting "/bin/sh" in command with info key "ompi_non_mpi" set to ".true." (if command is empty, mpi_comm_spawn tries to execute ' ') of mpi_comm_spawn and "-c" "mpirun -n 1 myapp" in args results in this message: /usr/bin/sh: -c: option requires an argument Putting a single string in args as

Re: [OMPI users] OMPI users] OMPI users] OMPI users] MPI inside MPI (still)

2014-12-17 Thread Alex A. Schmidt
Let me rephrase the previous message: Putting "/bin/sh" in command with info key "ompi_non_mpi" set to ".true." (if command is empty, mpi_comm_spawn tries to execute ' ') of mpi_comm_spawn and "-c" "mpirun -n 1 myapp" in args results in this message:

[OMPI users] Deadlock in OpenMPI 1.8.3 and PETSc 3.4.5

2014-12-17 Thread Ben Menadue
Hi PETSc and OpenMPI teams, I'm running into a deadlock in PETSc 3.4.5 with OpenMPI 1.8.3: 1. PetscCommDestroy calls MPI_Attr_delete 2. MPI_Attr_delete acquires a lock 3. MPI_Attr_delete calls Petsc_DelComm_Outer (through a callback) 4. Petsc_DelComm_Outer calls MPI_Attr_get 5. MPI_Attr_get

Re: [OMPI users] [EXTERNAL] Re: How to find MPI ranks located in remote nodes?

2014-12-17 Thread Jeff Squyres (jsquyres)
Returning to a super-old thread that was never finished... On Dec 2, 2014, at 6:49 PM, George Bosilca wrote: > That's not enough. They will have to check for the right version of Open MPI > and then for the availability of the OMPI_ functions. That looks > as having the

Re: [OMPI users] Operators for MPI handles not correctly overloaded with Intel Fortran

2014-12-17 Thread Jeff Squyres (jsquyres)
Jorg -- I'm sorry for the giant delay in replying; the US holiday and the MPI Forum meeting last week made a disaster out of my already-out-of-control INBOX. :-( Hmm. This almost sounds like a bug in the intel compiler. Do you have the latest version of their compiler, perchance? On Dec

Re: [OMPI users] OMPI users] OMPI users] OMPI users] OMPI users] MPI inside MPI (still)

2014-12-17 Thread Gilles Gouaillardet
Alex, You do not want to spawn mpirun. Or if this is really what you want, then just use system("env -i ...") I think what you need is spawn a shell that do the redirection and then invoke your app. This is something like MPI_Comm_spawn("/bin/sh", "-c", "siesta < infile") That being said, i

Re: [OMPI users] OpenMPI 1.8.4rc3, 1.6.5 and 1.6.3: segmentation violation in mca_io_romio_dist_MPI_File_close

2014-12-17 Thread Eric Chamberland
Hi! Here is a "poor man's fix" that works for me (the idea is not from me, thanks to Thomas H.): #1- char* lCwd = getcwd(0,0); #2- chdir(lPathToFile); #3- MPI_File_open(...,lFileNameWithoutTooLongPath,...); #4- chdir(lCwd); #5- ... I think there are some limitations but it works very well

Re: [OMPI users] [EXTERNAL] Re: How to find MPI ranks located in remote nodes?

2014-12-17 Thread George Bosilca
On Wed, Dec 17, 2014 at 7:29 PM, Jeff Squyres (jsquyres) wrote: > Returning to a super-old thread that was never finished... > > > On Dec 2, 2014, at 6:49 PM, George Bosilca wrote: > > > That's not enough. They will have to check for the right version

Re: [OMPI users] Deadlock in OpenMPI 1.8.3 and PETSc 3.4.5

2014-12-17 Thread Howard Pritchard
Hi Ben, Would you mind checking if you still observe this deadlock condition if you use the 1.8.4 rc4 candidate? openmpi-1.8.4rc4.tar.gz I realize the behavior will likely be the same, but this is just to double

Re: [OMPI users] OMPI users] OMPI users] OMPI users] OMPI users] MPI inside MPI (still)

2014-12-17 Thread Alex A. Schmidt
The option system("env -i ...") has been tested earlier by me and it does work. There is doubt though it would work along with a job scheduler. I will reserve this as a last resort solution. mpi_comm_spawn("/bin/sh","-c","siesta < infile",..) definitely does not work. Patching siesta to start

Re: [OMPI users] OMPI users] OMPI users] OMPI users] OMPI users] MPI inside MPI (still)

2014-12-17 Thread Ralph Castain
We can certainly add an MPI_Info key to redirect stdin, stdout, and stderr. However, that won't happen in the immediate future, nor would it come into the 1.8 series. Meantime, I suspect wrapping these codes in scripts sounds like the way to go. You would call mpirun to start the job in the

Re: [OMPI users] Deadlock in OpenMPI 1.8.3 and PETSc 3.4.5

2014-12-17 Thread George Bosilca
Ben, I can't find anything in the MPI standard suggesting that a recursive behavior of the attribute deletion is enforced/supported by the MPI standard. Thus, the current behavior of Open MPI (a single lock for all attributes), while maybe a little strict, is standard compliant (and thus

[hwloc-users] wrong os_index on AIX -> please test

2014-12-17 Thread Brice Goglin
Hello I am seeing assert failures on AIX 6.1 because our PU os_index is off by one. They go from -1 to 62 instead of 0 to 63. We have a comment saying /* It seems logical processors are numbered from 1 here, while the * bindprocessor functions numbers them from 0... */ This

Re: [hwloc-users] wrong os_index on AIX -> please test

2014-12-17 Thread Erik Schnetter
Brice bash-3.2$ uname -a AIX pandora1 1 6 00F63F144C00 bash-3.2$ ./lstopo-no-graphics --version lstopo-no-graphics 1.7.2 bash-3.2$ ./lstopo-no-graphics - | grep "PU L#0" PU L#0 (P#0) -erik On Wed, Dec 17, 2014 at 9:34 AM, Brice Goglin wrote: > Hello > > I am