When you do, that PR has already been committed, so you can just pull the next nightly 2.x tarball and test from there
> On Aug 24, 2016, at 10:39 AM, Paul Hargrove <phhargr...@lbl.gov> wrote: > > I am afraid it might take a day or two before I can get to testing that patch. > > -Paul > > On Tue, Aug 23, 2016 at 10:16 PM, Gilles Gouaillardet <gil...@rist.or.jp > <mailto:gil...@rist.or.jp>> wrote: > Paul, > > > you can download a patch at > https://patch-diff.githubusercontent.com/raw/open-mpi/ompi-release/pull/1336.patch > > <https://patch-diff.githubusercontent.com/raw/open-mpi/ompi-release/pull/1336.patch> > (note you need recent autotools in order to use it) > > > Cheers, > > > Gilles > > On 8/23/2016 10:40 PM, r...@open-mpi.org <mailto:r...@open-mpi.org> wrote: >> Looks like Solaris has a “getupeercred” - can you take a look at it, Gilles? >> We’d have to add that to our AC_CHECK_FUNCS and update the native sec >> component. >> >> >>> On Aug 23, 2016, at 6:32 AM, r...@open-mpi.org <mailto:r...@open-mpi.org> >>> wrote: >>> >>> I took a quick glance at this one, and the only way I can see to get that >>> error is from this block of code: >>> >>> #if defined(HAVE_STRUCT_UCRED_UID) >>> euid = ucred.uid; >>> gid = ucred.gid; >>> #else >>> euid = ucred.cr_uid; >>> gid = ucred.cr_gid; >>> #endif >>> >>> #elif defined(HAVE_GETPEEREID) >>> pmix_output_verbose(2, pmix_globals.debug_output, >>> "sec:native checking getpeereid for peer >>> credentials"); >>> if (0 != getpeereid(peer->sd, &euid, &gid)) { >>> pmix_output_verbose(2, pmix_globals.debug_output, >>> "sec: getsockopt getpeereid failed: %s", >>> strerror (pmix_socket_errno)); >>> return PMIX_ERR_INVALID_CRED; >>> } >>> #else >>> return PMIX_ERR_NOT_SUPPORTED; >>> #endif >>> >>> >>> I can only surmise, therefore, that Solaris doesn’t pass either of the two >>> #if define’d tests. Is there a Solaris alternative? >>> >>> >>>> On Aug 23, 2016, at 5:55 AM, r...@open-mpi.org <mailto:r...@open-mpi.org> >>>> wrote: >>>> >>>> Thanks Gilles! >>>> >>>>> On Aug 23, 2016, at 3:42 AM, Gilles Gouaillardet >>>>> <gilles.gouaillar...@gmail.com <mailto:gilles.gouaillar...@gmail.com>> >>>>> wrote: >>>>> >>>>> Thanks Paul, >>>>> >>>>> at first glance, something is going wrong in the sec module under solaris. >>>>> I will keep digging tomorrow >>>>> >>>>> Cheers, >>>>> >>>>> Gilles >>>>> >>>>> On Tuesday, August 23, 2016, Paul Hargrove <phhargr...@lbl.gov >>>>> <mailto:phhargr...@lbl.gov>> wrote: >>>>> On Solaris 11.3 on x86-64: >>>>> >>>>> $ mpirun -mca btl sm,self,openib -np 2 -host pcp-d-3,pcp-d-4 >>>>> examples/ring_c' >>>>> [pcp-d-4:25075] PMIX ERROR: NOT-SUPPORTED in file >>>>> /shared/OMPI/openmpi-2.0.1rc1-solaris11-x86-ib-gcc/openmpi-2.0.1rc1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_listener.c >>>>> at line 529 >>>>> [pcp-d-4:25078] PMIX ERROR: UNREACHABLE in file >>>>> /shared/OMPI/openmpi-2.0.1rc1-solaris11-x86-ib-gcc/openmpi-2.0.1rc1/opal/mca/pmix/pmix112/pmix/src/client/pmix_client.c >>>>> at line 983 >>>>> [pcp-d-4:25078] PMIX ERROR: UNREACHABLE in file >>>>> /shared/OMPI/openmpi-2.0.1rc1-solaris11-x86-ib-gcc/openmpi-2.0.1rc1/opal/mca/pmix/pmix112/pmix/src/client/pmix_client.c >>>>> at line 199 >>>>> -------------------------------------------------------------------------- >>>>> It looks like MPI_INIT failed for some reason; your parallel process is >>>>> likely to abort. There are many reasons that a parallel process can >>>>> fail during MPI_INIT; some of which are due to configuration or >>>>> environment >>>>> problems. This failure appears to be an internal failure; here's some >>>>> additional information (which may only be relevant to an Open MPI >>>>> developer): >>>>> >>>>> ompi_mpi_init: ompi_rte_init failed >>>>> --> Returned "(null)" (-43) instead of "Success" (0) >>>>> -------------------------------------------------------------------------- >>>>> *** An error occurred in MPI_Init >>>>> *** on a NULL communicator >>>>> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort, >>>>> *** and potentially your MPI job) >>>>> [pcp-d-4:25078] Local abort before MPI_INIT completed completed >>>>> successfully, but am not able to aggregate error messages, and not able >>>>> to guarantee that all other processes were killed! >>>>> ------------------------------------------------------- >>>>> Primary job terminated normally, but 1 process returned >>>>> a non-zero exit code.. Per user-direction, the job has been aborted. >>>>> ------------------------------------------------------- >>>>> -------------------------------------------------------------------------- >>>>> mpirun detected that one or more processes exited with non-zero status, >>>>> thus causing >>>>> the job to be terminated. The first process to do so was: >>>>> >>>>> Process name: [[25599,1],1] >>>>> Exit code: 1 >>>>> -------------------------------------------------------------------------- >>>>> >>>>> -Paul >>>>> >>>>> -- >>>>> Paul H. Hargrove phhargr...@lbl.gov <> >>>>> Computer Languages & Systems Software (CLaSS) Group >>>>> Computer Science Department Tel: +1-510-495-2352 >>>>> <tel:%2B1-510-495-2352> >>>>> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900 >>>>> <tel:%2B1-510-486-6900>_______________________________________________ >>>>> devel mailing list >>>>> devel@lists.open-mpi.org <mailto:devel@lists.open-mpi.org> >>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel >>>>> <https://rfd.newmexicoconsortium.org/mailman/listinfo/devel> >>>> _______________________________________________ >>>> devel mailing list >>>> devel@lists.open-mpi.org <mailto:devel@lists.open-mpi.org> >>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel >>>> <https://rfd.newmexicoconsortium.org/mailman/listinfo/devel> >> >> >> >> _______________________________________________ >> devel mailing list >> devel@lists.open-mpi.org <mailto:devel@lists.open-mpi.org> >> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel >> <https://rfd.newmexicoconsortium.org/mailman/listinfo/devel> > > _______________________________________________ > devel mailing list > devel@lists.open-mpi.org <mailto:devel@lists.open-mpi.org> > https://rfd.newmexicoconsortium.org/mailman/listinfo/devel > <https://rfd.newmexicoconsortium.org/mailman/listinfo/devel> > > > > -- > Paul H. Hargrove phhargr...@lbl.gov > <mailto:phhargr...@lbl.gov> > Computer Languages & Systems Software (CLaSS) Group > Computer Science Department Tel: +1-510-495-2352 > Lawrence Berkeley National Laboratory Fax: +1-510-486-6900 > _______________________________________________ > devel mailing list > devel@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
_______________________________________________ devel mailing list devel@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/devel