Jeff, I think this was fixed in PR 1227 on v2.x
Howard -- Howard Pritchard HPC-DES Los Alamos National Laboratory On 7/13/16, 1:47 PM, "devel on behalf of Jeff Squyres (jsquyres)" <devel-boun...@open-mpi.org on behalf of jsquy...@cisco.com> wrote: >I literally just noticed that this morning (that singleton was broken on >master), but hadn't gotten to bisecting / reporting it yet... > >I also haven't tested 2.0.0. I really hope singletons aren't broken >then... > >/me goes to test 2.0.0... > >Whew -- 2.0.0 singletons are fine. :-) > > >> On Jul 13, 2016, at 3:01 PM, Ralph Castain <r...@open-mpi.org> wrote: >> >> HmmmÅ I see where the singleton on master might be broken - will check >>later today >> >>> On Jul 13, 2016, at 11:37 AM, Eric Chamberland >>><eric.chamberl...@giref.ulaval.ca> wrote: >>> >>> Hi Howard, >>> >>> ok, I will wait for 2.0.1rcX... ;) >>> >>> I've put in place a script to download/compile OpenMPI+PETSc(3.7.2) >>>and our code from the git repos. >>> >>> Now I am in a somewhat uncomfortable situation where neither the >>>ompi-release.git or ompi.git repos are working for me. >>> >>> The first gives me the errors with MPI_File_write_all_end I reported, >>>but the former gives me errors like these: >>> >>> [lorien:106919] [[INVALID],INVALID] ORTE_ERROR_LOG: Bad parameter in >>>file ess_singleton_module.c at line 167 >>> *** An error occurred in MPI_Init_thread >>> *** on a NULL communicator >>> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now >>>abort, >>> *** and potentially your MPI job) >>> [lorien:106919] Local abort before MPI_INIT completed completed >>>successfully, but am not able to aggregate error messages, and not able >>>to guarantee that all other processes were killed! >>> >>> So, for my continuous integration of OpenMPI I am in a no man's >>>land... :( >>> >>> Thanks anyway for the follow-up! >>> >>> Eric >>> >>> On 13/07/16 07:49 AM, Howard Pritchard wrote: >>>> Hi Eric, >>>> >>>> Thanks very much for finding this problem. We decided in order to >>>>have >>>> a reasonably timely >>>> release, that we'd triage issues and turn around a new RC if something >>>> drastic >>>> appeared. We want to fix this issue (and it will be fixed), but we've >>>> decided to >>>> defer the fix for this issue to a 2.0.1 bug fix release. >>>> >>>> Howard >>>> >>>> >>>> >>>> 2016-07-12 13:51 GMT-06:00 Eric Chamberland >>>> <eric.chamberl...@giref.ulaval.ca >>>> <mailto:eric.chamberl...@giref.ulaval.ca>>: >>>> >>>> Hi Edgard, >>>> >>>> I just saw that your patch got into ompi/master... any chances it >>>> goes into ompi-release/v2.x before rc5? >>>> >>>> thanks, >>>> >>>> Eric >>>> >>>> >>>> On 08/07/16 03:14 PM, Edgar Gabriel wrote: >>>> >>>> I think I found the problem, I filed a pr towards master, and >>>>if >>>> that >>>> passes I will file a pr for the 2.x branch. >>>> >>>> Thanks! >>>> Edgar >>>> >>>> >>>> On 7/8/2016 1:14 PM, Eric Chamberland wrote: >>>> >>>> >>>> On 08/07/16 01:44 PM, Edgar Gabriel wrote: >>>> >>>> ok, but just to be able to construct a test case, >>>> basically what you are >>>> doing is >>>> >>>> MPI_File_write_all_begin (fh, NULL, 0, some datatype); >>>> >>>> MPI_File_write_all_end (fh, NULL, &status), >>>> >>>> is this correct? >>>> >>>> Yes, but with 2 processes: >>>> >>>> rank 0 writes something, but not rank 1... >>>> >>>> other info: rank 0 didn't wait for rank1 after >>>> MPI_File_write_all_end so >>>> it continued to the next MPI_File_write_all_begin with a >>>> different >>>> datatype but on the same file... >>>> >>>> thanks! >>>> >>>> Eric >>>> _______________________________________________ >>>> devel mailing list >>>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>>> Subscription: >>>> https://www.open-mpi.org/mailman/listinfo.cgi/devel >>>> Link to this post: >>>> >>>>http://www.open-mpi.org/community/lists/devel/2016/07/19173.php >>>> >>>> >>>> _______________________________________________ >>>> devel mailing list >>>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>>> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel >>>> Link to this post: >>>> http://www.open-mpi.org/community/lists/devel/2016/07/19192.php >>>> >>>> >>> _______________________________________________ >>> devel mailing list >>> de...@open-mpi.org >>> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel >>> Link to this post: >>>http://www.open-mpi.org/community/lists/devel/2016/07/19201.php >> >> _______________________________________________ >> devel mailing list >> de...@open-mpi.org >> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel >> Link to this post: >>http://www.open-mpi.org/community/lists/devel/2016/07/19202.php > > >-- >Jeff Squyres >jsquy...@cisco.com >For corporate legal information go to: >http://www.cisco.com/web/about/doing_business/legal/cri/ > >_______________________________________________ >devel mailing list >de...@open-mpi.org >Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel >Link to this post: >http://www.open-mpi.org/community/lists/devel/2016/07/19203.php