Okay, thanks. It's the same problem as the other person encountered. Basically, 
it looks to OMPI as if you are launching > 128 independent app contexts, and 
our arrays were limited to 128.

He has provided a patch that I'll review (couple of things I'd rather change) 
and then apply to our developer's trunk. I would expect it to migrate over to 
the 1.4 release series at some point (can't guarantee which one).


On Feb 27, 2010, at 6:47 AM, Federico Golfrè Andreasi wrote:

> Hi,
> 
> the program is executed as one application on 129 cpus defined by the 
> hostfile. 
> Than rank 0, inside the code, execute another program with 129 cpus, with a 
> one-to-one relation, rank0 of the spawined process runs on the same host of 
> rank0 of the spawning one and so on...
> Excuting the spawning program does not give any problem,
> but in the moment of spawning (with more than 128 cpus) it holds. 
> 
> Thank you!
> 
> Federico
> 
> 
> 
> 
> 2010/2/27 Ralph Castain <r...@open-mpi.org>
> Since another user was doing something that caused a similar problem, perhaps 
> we are missing a key piece of info here. Are you launching one app_context 
> across 128 nodes? Or are you launching 128 app_contexts, each on a separate 
> node?
> 
> 
> On Feb 26, 2010, at 10:23 AM, Federico Golfrè Andreasi wrote:
> 
>> I'm doing some tests and it seems that is not able to do a spawn multiple 
>> with more than 128 nodes.
>> 
>> It just hold, with no error message.
>> 
>> What do you think? What can I try to understand the problem.
>> 
>> Thanks,
>> 
>> Federico
>> 
>> 
>> 
>> 
>> 2010/2/26 Ralph Castain <r...@open-mpi.org>
>> No known limitations of which we are aware...the variables are all set to 
>> int32_t, so INT32_MAX would be the only limit I can imagine. In which case, 
>> you'll run out of memory long before you hit it.
>> 
>> 
>> 2010/2/26 Federico Golfrè Andreasi <federico.gol...@gmail.com>
>> HI !
>> 
>> have you ever did some analysis to understand if there is a limitation in 
>> the number of nodes usable with OpenMPI-v1.4 ?
>> Using also the functions MPI_Comm_spawn o MPI_Comm_spawn_multiple.
>> 
>> Thanks,
>>    Federico
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to