I would have to look at the code, but I suspect it doesn't handle "*". Could be upgraded to do so, but that would depend on the relevant developer to do so :-)
On Jun 9, 2010, at 10:16 AM, Grzegorz Maj wrote: > Thanks a lot, it works fine for me. > But going back to my problems - is it some bug in open-mpi or I should > use "slot=*" option in some other way? > > 2010/6/9 Ralph Castain <r...@open-mpi.org>: >> I would recommend using the sequential mapper instead: >> >> mpirun -mca rmaps seq >> >> You can then just list your hosts in your hostfile, and we will put the >> ranks sequentially on those hosts. So you get something like this >> >> host01 <= rank0 >> host01 <= rank1 >> host02 <= rank2 >> host03 <= rank3 >> host01 <= rank4 >> >> Ralph >> >> On Jun 9, 2010, at 4:39 AM, Grzegorz Maj wrote: >> >>> In my previous mail I said that slot=0-3 would be a solution. >>> Unfortunately it gives me exactly the same segfault as in case with >>> *:* >>> >>> 2010/6/9 Grzegorz Maj <ma...@wp.pl>: >>>> Hi, >>>> I'd like mpirun to run tasks with specific ranks on specific hosts, >>>> but I don't want to provide any particular sockets/slots/cores. >>>> The following example uses just one host, but generally I'll use more. >>>> In my hostfile I just have: >>>> >>>> root@host01 slots=4 >>>> >>>> I was playing with my rankfile to achieve what I've mentioned, but I >>>> always get some problems. >>>> >>>> 1) With rankfile like: >>>> rank 0=host01 slot=* >>>> rank 1=host01 slot=* >>>> rank 2=host01 slot=* >>>> rank 3=host01 slot=* >>>> >>>> I get: >>>> >>>> -------------------------------------------------------------------------- >>>> We were unable to successfully process/set the requested processor >>>> affinity settings: >>>> >>>> Specified slot list: * >>>> Error: Error >>>> >>>> This could mean that a non-existent processor was specified, or >>>> that the specification had improper syntax. >>>> -------------------------------------------------------------------------- >>>> -------------------------------------------------------------------------- >>>> mpirun was unable to start the specified application as it encountered an >>>> error: >>>> >>>> Error name: Error >>>> Node: host01 >>>> >>>> when attempting to start process rank 0. >>>> -------------------------------------------------------------------------- >>>> [host01:13715] Rank 0: PAFFINITY cannot get physical processor id for >>>> logical processor 4 >>>> >>>> >>>> I think it tries to find processor #4, bug there are only 0-3 >>>> >>>> 2) With rankfile like: >>>> rank 0=host01 slot=*:* >>>> rank 1=host01 slot=*:* >>>> rank 2=host01 slot=*:* >>>> rank 3=host01 slot=*:* >>>> >>>> Everything looks well, i.e. my programs are spread across 4 processors. >>>> But when running MPI program as follows: >>>> >>>> MPI::Init(argc, argv); >>>> fprintf(stderr, "after init %d\n", MPI::Is_initialized()); >>>> nprocs_mpi = MPI::COMM_WORLD.Get_size(); >>>> fprintf(stderr, "won't get here\n"); >>>> >>>> I get: >>>> >>>> after init 1 >>>> [host01:14348] *** Process received signal *** >>>> [host01:14348] Signal: Segmentation fault (11) >>>> [host01:14348] Signal code: Address not mapped (1) >>>> [host01:14348] Failing at address: 0x8 >>>> [host01:14348] [ 0] [0xffffe410] >>>> [host01:14348] [ 1] p(_ZNK3MPI4Comm8Get_sizeEv+0x19) [0x8051299] >>>> [host01:14348] [ 2] p(main+0x86) [0x804ee4e] >>>> [host01:14348] [ 3] /lib/libc.so.6(__libc_start_main+0xe5) [0x4180b5c5] >>>> [host01:14348] [ 4] p(__gxx_personality_v0+0x125) [0x804ecc1] >>>> [host01:14348] *** End of error message *** >>>> >>>> I'm using OPEN MPI v. 1.4.2 (downloaded yesterday). >>>> In my rankfile I really want to write something like slot=*. I know >>>> slot=0-3 would be a solution, but when generating rankfile I may not >>>> be sure how many processors are there available on a particular host. >>>> >>>> Any help would be appreciated. >>>> >>>> Regards, >>>> Grzegorz Maj >>>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users