Thanks Ralph, I will use the PR patch meanwhile. George On Feb 29, 2016 12:44 PM, "Ralph Castain" <r...@open-mpi.org> wrote:
> The fix is waiting in a PR - Howard said he will do the final review this > evening (need verification that Cray isn’t broken) > > > On Feb 29, 2016, at 10:35 AM, George Bosilca <bosi...@icl.utk.edu> > wrote: > > > > Singletons are broken with version (e5d6b97db4fa1) compiled in both > debug and optimized builds. I got this on my OS X laptop, I can try on a > linux machine if necessary. > > > > The error message is attached below. My application is calling > MPI_Init_thread, but this is not the root of the problem as a simple > MPI_Init/MPI_Finalize fails with a similar message. > > > > Thanks, > > George. > > > > > > > > > -------------------------------------------------------------------------- > > It looks like orte_init failed for some reason; your parallel process is > > likely to abort. There are many reasons that a parallel process can > > fail during orte_init; some of which are due to configuration or > > environment problems. This failure appears to be an internal failure; > > here's some additional information (which may only be relevant to an > > Open MPI developer): > > > > init pmix failed > > --> Returned value Bad parameter (-5) instead of ORTE_SUCCESS > > > -------------------------------------------------------------------------- > > > -------------------------------------------------------------------------- > > It looks like orte_init failed for some reason; your parallel process is > > likely to abort. There are many reasons that a parallel process can > > fail during orte_init; some of which are due to configuration or > > environment problems. This failure appears to be an internal failure; > > here's some additional information (which may only be relevant to an > > Open MPI developer): > > > > orte_ess_init failed > > --> Returned value Bad parameter (-5) instead of ORTE_SUCCESS > > > -------------------------------------------------------------------------- > > > -------------------------------------------------------------------------- > > It looks like MPI_INIT failed for some reason; your parallel process is > > likely to abort. There are many reasons that a parallel process can > > fail during MPI_INIT; some of which are due to configuration or > environment > > problems. This failure appears to be an internal failure; here's some > > additional information (which may only be relevant to an Open MPI > > developer): > > > > ompi_mpi_init: ompi_rte_init failed > > --> Returned "Bad parameter" (-5) instead of "Success" (0) > > > -------------------------------------------------------------------------- > > *** An error occurred in MPI_Init_thread > > *** on a NULL communicator > > *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort, > > *** and potentially your MPI job) > > [Babuska:42464] Local abort before MPI_INIT completed completed > successfully, but am not able to aggregate error messages, and not able to > guarantee that all other processes were killed! > > > > > > _______________________________________________ > > devel mailing list > > de...@open-mpi.org > > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel > > Link to this post: > http://www.open-mpi.org/community/lists/devel/2016/02/18642.php > > _______________________________________________ > devel mailing list > de...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel > Link to this post: > http://www.open-mpi.org/community/lists/devel/2016/02/18643.php