I can't seem to replicate this first problem - it runs fine for me even if the rankfile contains only one entry.
What you could do is build 1.4.1 with --enable-debug and then run with --debug-devel -mca rmaps_base_verbose 5 to get more info on what is happening. On Feb 27, 2010, at 11:26 AM, Ralph Castain wrote: > I'm looking at the first problem - will get back to you on it. > > As to the second issue: it was r21728, and no - it does not appear to have > been moved to the 1.4 series (rankfile is untested on a regular basis). I > will do so now. > > Thanks! > Ralph > > On Feb 15, 2010, at 10:39 AM, Bogdan Costescu wrote: > >> Hi! >> >> With version 1.4.1 I get a rather strange crash in mpirun whenever I >> try to run a job using (I think) a rankfile which doesn't contain the >> specified number of ranks. F.e. I ask for 4 ranks ('-np 4'), but the >> rankfile contains only one entry: >> >> rank 0=mbm-01-24 slot=1:* >> >> and the following comes out: >> >> [mbm-01-24:20985] *** Process received signal *** >> [mbm-01-24:20985] Signal: Segmentation fault (11) >> [mbm-01-24:20985] Signal code: Address not mapped (1) >> [mbm-01-24:20985] Failing at address: 0x50 >> [mbm-01-24:20985] [ 0] /lib64/libpthread.so.0 [0x2b9de894f7c0] >> [mbm-01-24:20985] [ 1] >> /sw/openmpi/1.4.1/gcc/4.4.3/lib/libopen-rte.so.0(orte_util_encode_pidmap+0xbb) >> [0x2b9de79b9f7b] >> [mbm-01-24:20985] [ 2] >> /sw/openmpi/1.4.1/gcc/4.4.3/lib/libopen-rte.so.0(orte_odls_base_default_get_add_procs_data+0x2d0) >> [0x2b9de79d49c0] >> [mbm-01-24:20985] [ 3] >> /sw/openmpi/1.4.1/gcc/4.4.3/lib/libopen-rte.so.0(orte_plm_base_launch_apps+0xbc) >> [0x2b9de79e1fcc] >> [mbm-01-24:20985] [ 4] >> /sw/openmpi/1.4.1/gcc/4.4.3/lib/libopen-rte.so.0 [0x2b9de79e6251] >> [mbm-01-24:20985] [ 5] mpirun [0x403782] >> [mbm-01-24:20985] [ 6] mpirun [0x402cb4] >> [mbm-01-24:20985] [ 7] /lib64/libc.so.6(__libc_start_main+0xf4) >> [0x2b9de8b79994] >> [mbm-01-24:20985] [ 8] mpirun [0x402bd9] >> [mbm-01-24:20985] *** End of error message *** >> Segmentation fault >> >> However if the rankfile contains a second entry, like: >> >> rank 0=mbm-01-24 slot=1:* >> rank 1=mbm-01-24 slot=1:* >> >> I get an error, but no segmentation fault. I guess that the >> segmentation fault is unintended... Is this known ? If not, how could >> I debug this ? >> Now to the second problem: the exact same error keeps coming even if I >> specify 4 ranks, the messages are: >> >> -------------------------------------------------------------------------- >> mpirun was unable to start the specified application as it encountered an >> error: >> >> Error name: Error >> Node: mbm-01-24 >> >> when attempting to start process rank 0. >> -------------------------------------------------------------------------- >> [mbm-01-24:21011] Rank 0: PAFFINITY cannot get physical core id for >> logical core 4 in physical socket 1 (1) >> -------------------------------------------------------------------------- >> We were unable to successfully process/set the requested processor >> affinity settings: >> >> Specified slot list: 1:* >> Error: Error >> >> This could mean that a non-existent processor was specified, or >> that the specification had improper syntax. >> -------------------------------------------------------------------------- >> >> The node has 2 slots, each with 4 cores, so what I'm trying to achieve >> is using the 4 cores of the second slot. When searching the archives, >> I stumbled on an e-mail from not too long ago which seemingly dealt >> with the same error: >> >> http://www.open-mpi.org/community/lists/devel/2009/07/6513.php >> >> which suggests that a fix was found, but no commit was specified, so I >> can't track down whether this was actually also applied to the stable >> series. Could someone more knowledgeable in this area shed some light >> ? >> >> Thanks in advance! >> Bogdan >> _______________________________________________ >> devel mailing list >> de...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/devel >