Thanks Ralph,  I will use the PR  patch meanwhile.

George
On Feb 29, 2016 12:44 PM, "Ralph Castain" <r...@open-mpi.org> wrote:

> The fix is waiting in a PR - Howard said he will do the final review this
> evening (need verification that Cray isn’t broken)
>
> > On Feb 29, 2016, at 10:35 AM, George Bosilca <bosi...@icl.utk.edu>
> wrote:
> >
> > Singletons are broken with version (e5d6b97db4fa1) compiled in both
> debug and optimized builds. I got this on my OS X laptop, I can try on a
> linux machine if necessary.
> >
> > The error message is attached below. My application is calling
> MPI_Init_thread, but this is not the root of the problem as a simple
> MPI_Init/MPI_Finalize fails with a similar message.
> >
> >   Thanks,
> >     George.
> >
> >
> >
> >
> --------------------------------------------------------------------------
> > It looks like orte_init failed for some reason; your parallel process is
> > likely to abort.  There are many reasons that a parallel process can
> > fail during orte_init; some of which are due to configuration or
> > environment problems.  This failure appears to be an internal failure;
> > here's some additional information (which may only be relevant to an
> > Open MPI developer):
> >
> >   init pmix failed
> >   --> Returned value Bad parameter (-5) instead of ORTE_SUCCESS
> >
> --------------------------------------------------------------------------
> >
> --------------------------------------------------------------------------
> > It looks like orte_init failed for some reason; your parallel process is
> > likely to abort.  There are many reasons that a parallel process can
> > fail during orte_init; some of which are due to configuration or
> > environment problems.  This failure appears to be an internal failure;
> > here's some additional information (which may only be relevant to an
> > Open MPI developer):
> >
> >   orte_ess_init failed
> >   --> Returned value Bad parameter (-5) instead of ORTE_SUCCESS
> >
> --------------------------------------------------------------------------
> >
> --------------------------------------------------------------------------
> > It looks like MPI_INIT failed for some reason; your parallel process is
> > likely to abort.  There are many reasons that a parallel process can
> > fail during MPI_INIT; some of which are due to configuration or
> environment
> > problems.  This failure appears to be an internal failure; here's some
> > additional information (which may only be relevant to an Open MPI
> > developer):
> >
> >   ompi_mpi_init: ompi_rte_init failed
> >   --> Returned "Bad parameter" (-5) instead of "Success" (0)
> >
> --------------------------------------------------------------------------
> > *** An error occurred in MPI_Init_thread
> > *** on a NULL communicator
> > *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> > ***    and potentially your MPI job)
> > [Babuska:42464] Local abort before MPI_INIT completed completed
> successfully, but am not able to aggregate error messages, and not able to
> guarantee that all other processes were killed!
> >
> >
> > _______________________________________________
> > devel mailing list
> > de...@open-mpi.org
> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> > Link to this post:
> http://www.open-mpi.org/community/lists/devel/2016/02/18642.php
>
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post:
> http://www.open-mpi.org/community/lists/devel/2016/02/18643.php

Reply via email to