[hwloc-devel] Create success (hwloc git dev-1179-gf061101)

2016-07-13 Thread MPI Team
Creating nightly hwloc snapshot git tarball was a success. Snapshot: hwloc dev-1179-gf061101 Start time: Wed Jul 13 21:01:02 EDT 2016 End time: Wed Jul 13 21:03:02 EDT 2016 Your friendly daemon, Cyrador

Re: [OMPI devel] SHMEM, "mpp/shmem.fh", CMake and infinite loops

2016-07-13 Thread Gilles Gouaillardet
Paul, thanks for testing the workaround /* i was on a trip and could not do it myself */ At first glance, i agree with Jeff and the root cause seems to be a CMake bug. /* i cannot find any rationale for automatically including some directories that were not requested by the user */

Re: [OMPI devel] 2.0.0rc4 Crash in MPI_File_write_all_end

2016-07-13 Thread Gilles Gouaillardet
Eric, OpenMPI 2.0.0 has been released, so the fix should land into the v2.x branch shortly. If i understand correctly, you script download/compile OpenMPI and then download/compile PETSc. In this is correct, and for the time being, feel free to patch Open MPI v2.x before compiling it,

Re: [OMPI devel] 2.0.0rc4 Crash in MPI_File_write_all_end

2016-07-13 Thread Eric Chamberland
Hi, FYI: I've tested the SHA e28951e From git clone launched around 01h19: http://www.giref.ulaval.ca/~cmpgiref/dernier_ompi/2016.07.13.01h19m30s_config.log Eric On 13/07/16 04:01 PM, Pritchard Jr., Howard wrote: Jeff, I think this was fixed in PR 1227 on v2.x Howard

Re: [OMPI devel] 2.0.0rc4 Crash in MPI_File_write_all_end

2016-07-13 Thread Pritchard Jr., Howard
Jeff, I think this was fixed in PR 1227 on v2.x Howard -- Howard Pritchard HPC-DES Los Alamos National Laboratory On 7/13/16, 1:47 PM, "devel on behalf of Jeff Squyres (jsquyres)" wrote: >I literally just noticed that this

Re: [OMPI devel] 2.0.0rc4 Crash in MPI_File_write_all_end

2016-07-13 Thread Jeff Squyres (jsquyres)
I literally just noticed that this morning (that singleton was broken on master), but hadn't gotten to bisecting / reporting it yet... I also haven't tested 2.0.0. I really hope singletons aren't broken then... /me goes to test 2.0.0... Whew -- 2.0.0 singletons are fine. :-) > On Jul 13,

Re: [OMPI devel] 2.0.0rc4 Crash in MPI_File_write_all_end

2016-07-13 Thread Ralph Castain
Hmmm…I see where the singleton on master might be broken - will check later today > On Jul 13, 2016, at 11:37 AM, Eric Chamberland > wrote: > > Hi Howard, > > ok, I will wait for 2.0.1rcX... ;) > > I've put in place a script to download/compile

Re: [OMPI devel] 2.0.0rc4 Crash in MPI_File_write_all_end

2016-07-13 Thread Eric Chamberland
Hi Howard, ok, I will wait for 2.0.1rcX... ;) I've put in place a script to download/compile OpenMPI+PETSc(3.7.2) and our code from the git repos. Now I am in a somewhat uncomfortable situation where neither the ompi-release.git or ompi.git repos are working for me. The first gives me the

[OMPI devel] Option to switch from shmem to kmem module

2016-07-13 Thread Abhishek Joshi
Hi, Is there any option in mpirun which enables us to switch dynamically from shmem mode to knem/xpmem mode beyond a specifiable message size? This is because, according to my tests, knem performs better than shmem only at large message sizes. -- Abhishek

Re: [OMPI devel] SHMEM, "mpp/shmem.fh", CMake and infinite loops

2016-07-13 Thread Jeff Squyres (jsquyres)
Thanks for the report. I don't know much about OSHMEM, but I'm guessing the files were laid out that way for a reason (e.g., maybe the OSHMEM spec calls for both of those files to exist?). I've filed an issue here to track it: https://github.com/open-mpi/ompi/issues/1868 Additionally,

Re: [OMPI devel] MCA_SPML_CALL call in compiled objects

2016-07-13 Thread Jeff Squyres (jsquyres)
Thanks Ben. Rainer Keller just filed a PR for this -- we'll get it in v2.0.1: https://github.com/open-mpi/ompi/pull/1867 > On Jul 12, 2016, at 12:08 AM, Ben Menadue wrote: > > Hi, > > Looks like there's a #include missing from >

Re: [OMPI devel] 2.0.0rc4 Crash in MPI_File_write_all_end

2016-07-13 Thread Howard Pritchard
Hi Eric, Thanks very much for finding this problem. We decided in order to have a reasonably timely release, that we'd triage issues and turn around a new RC if something drastic appeared. We want to fix this issue (and it will be fixed), but we've decided to defer the fix for this issue to a

Re: [OMPI devel] SHMEM, "mpp/shmem.fh", CMake and infinite loops

2016-07-13 Thread Paul Kapinos
Hi Gilles, On 07/13/16 01:10, Gilles Gouaillardet wrote: Paul, The two header files in include/mpp simply include the file with the same name in the upper directory. Yessir! (and CMake do not care about the upper directory and build infinite loop) A simple workaround is to replace these