Hmmm...well, my bad. There does indeed appear to be something funny going on with Leopard. No idea what - it used to work fine. I haven't tested it in awhile though - I've been test building regularly on Leopard, but running on Tiger (I misspoke earlier).
For now, I'm afraid you can't run on Leopard. Have to figure it out later when I have more time. Ralph > ------ Forwarded Message >> From: Aurélien Bouteiller <boute...@eecs.utk.edu> >> Reply-To: Open MPI Developers <de...@open-mpi.org> >> Date: Thu, 31 Jan 2008 02:18:27 -0500 >> To: Open MPI Developers <de...@open-mpi.org> >> Subject: Re: [OMPI devel] orte_ns_base_select failed: returned value -1 >> instead of ORTE_SUCCESS >> >> I tried using a fresh trunk, same problem have occured. Here is the >> complete configure line. I am using libtool 1.5.22 from fink. >> Otherwise everything is standard OS 10.5. >> >> $ ../trunk/configure --prefix=/Users/bouteill/ompi/build --enable- >> mpirun-prefix-by-default --disable-io-romio --enable-debug --enable- >> picky --enable-mem-debug --enable-mem-profile --enable-visibility -- >> disable-dlopen --disable-shared --enable-static >> >> The error message generated by abort contains garbage (line numbers do >> not match anything in .c files and according to gdb the failure does >> not occur during ns initialization). This looks like a heap corruption >> or something as bad. >> >> orterun (argc=4, argv=0xbffff81c) at ../../../../trunk/orte/tools/ >> orterun/orterun.c:529 >> 529 cb_states = ORTE_PROC_STATE_TERMINATED | >> ORTE_PROC_STATE_AT_STG1; >> (gdb) n >> 530 rc = orte_rmgr.spawn_job(apps, num_apps, &jobid, 0, NULL, >> job_state_callback, cb_states, &attributes); >> (gdb) n >> 531 while (NULL != (item = opal_list_remove_first(&attributes))) >> OBJ_RELEASE(item); >> (gdb) n >> ** Stepping over inlined function code. ** >> 532 OBJ_DESTRUCT(&attributes); >> (gdb) n >> 534 if (orterun_globals.do_not_launch) { >> (gdb) n >> 539 OPAL_THREAD_LOCK(&orterun_globals.lock); >> (gdb) n >> 541 if (ORTE_SUCCESS == rc) { >> (gdb) n >> 542 while (!orterun_globals.exit) { >> (gdb) n >> 543 opal_condition_wait(&orterun_globals.cond, >> (gdb) n >> [grosse-pomme.local:77335] [NO-NAME] ORTE_ERROR_LOG: Bad parameter in >> file /SourceCache/openmpi/openmpi-5/openmpi/orte/mca/oob/base/ >> oob_base_init.c at line 74 >> >> Aurelien >> >> >> Le 30 janv. 08 à 17:18, Ralph Castain a écrit : >> >>> Are you running on the trunk, or an earlier release? >>> >>> If the trunk, then I suspect you have a stale library hanging >>> around. I >>> build and run statically on Leopard regularly. >>> >>> >>> On 1/30/08 2:54 PM, "Aurélien Bouteiller" <boute...@eecs.utk.edu> >>> wrote: >>> >>>> I get a runtime error in static build on Mac OS 10.5 (automake 1.10, >>>> autoconf 2.60, gcc-apple-darwin 4.01, libtool 1.5.22). >>>> >>>> The error does not occur in dso builds, and everything seems to work >>>> fine on Linux. >>>> >>>> Here is the error log. >>>> >>>> ~/ompi$ mpirun -np 2 NetPIPE_3.6/NPmpi >>>> [grosse-pomme.local:34247] [NO-NAME] ORTE_ERROR_LOG: Bad parameter in >>>> file /SourceCache/openmpi/openmpi-5/openmpi/orte/mca/oob/base/ >>>> oob_base_init.c at line 74 >>>> [grosse-pomme.local:34247] [NO-NAME] ORTE_ERROR_LOG: Bad parameter in >>>> file /SourceCache/openmpi/openmpi-5/openmpi/orte/mca/ns/proxy/ >>>> ns_proxy_component.c at line 222 >>>> [grosse-pomme.local:34247] [NO-NAME] ORTE_ERROR_LOG: Error in file / >>>> SourceCache/openmpi/openmpi-5/openmpi/orte/runtime/orte_init_stage1.c >>>> at line 230 >>>> -------------------------------------------------------------------------- >>>> It looks like orte_init failed for some reason; your parallel >>>> process is >>>> likely to abort. There are many reasons that a parallel process can >>>> fail during orte_init; some of which are due to configuration or >>>> environment problems. This failure appears to be an internal >>>> failure; >>>> here's some additional information (which may only be relevant to an >>>> Open MPI developer): >>>> >>>> orte_ns_base_select failed >>>> --> Returned value -1 instead of ORTE_SUCCESS >>>> >>>> -------------------------------------------------------------------------- >>>> -------------------------------------------------------------------------- >>>> It looks like MPI_INIT failed for some reason; your parallel >>>> process is >>>> likely to abort. There are many reasons that a parallel process can >>>> fail during MPI_INIT; some of which are due to configuration or >>>> environment >>>> problems. This failure appears to be an internal failure; here's >>>> some >>>> additional information (which may only be relevant to an Open MPI >>>> developer): >>>> >>>> ompi_mpi_init: orte_init_stage1 failed >>>> --> Returned "Error" (-1) instead of "Success" (0) >>>> -------------------------------------------------------------------------- >>>> *** An error occurred in MPI_Init >>>> *** before MPI was initialized >>>> *** MPI_ERRORS_ARE_FATAL (goodbye) >>>> >>>> >>>> >>>> -- >>>> Dr. Aurélien Bouteiller >>>> Sr. Research Associate - Innovative Computing Laboratory >>>> Suite 350, 1122 Volunteer Boulevard >>>> Knoxville, TN 37996 >>>> 865 974 6321 >>>> >>>> >>>> >>>> >>>> >>>> _______________________________________________ >>>> devel mailing list >>>> de...@open-mpi.org >>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>> >>> >>> >>> _______________________________________________ >>> devel mailing list >>> de...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >> >> >> _______________________________________________ >> devel mailing list >> de...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/devel > > ------ End of Forwarded Message > > > > ------ End of Forwarded Message > >