On Jul 12, 2010, at 3:23 PM, Brian Budge wrote:

> Hi Ralph -
> 
> So you can just start this daemon on all of the nodes when the
> machines are booted, for example, and then these connections can be
> made programmatically?

Ummm...not exactly. You have to start only one ompi-server, but it must be on a 
node that is reachable from everywhere. Then you need to pass an mca param to 
each singleton telling it how to connect to that ompi-server. The man page on 
ompi-server gives the info on how to do that.


> 
> Sounds great.  I look forward to that functionality.
> 
>  Brian
> 
> On Mon, Jul 12, 2010 at 12:38 PM, Ralph Castain <r...@open-mpi.org> wrote:
>> 
>> On Jul 12, 2010, at 11:12 AM, Brian Budge wrote:
>> 
>>> HI Ralph -
>>> 
>>> Thanks for the reply.  I think this patch sounds great!  The idea in
>>> our software is that it won't be known until after the program is
>>> running whether or not MPI is needed, so it would be best if the
>>> communication initialization could be done programmatically instead of
>>> through an external program like mpirun.  Is there a plan for this to
>>> enter the mainline?
>> 
>> Just to clarify: while the revised OMPI will allow singleton connect/accept, 
>> it will still only do so in the presence of a running ompi-server. There is 
>> no way to wireup singletons without that external agent.
>> 
>> However, ompi-server is not specific to any job, so it can be allowed to run 
>> in the background as a daemon, used only when required.
>> 
>>> 
>>>  Brian
>>> 
>>> On Mon, Jul 12, 2010 at 8:29 AM, Ralph Castain <r...@open-mpi.org> wrote:
>>>> 
>>>> On Jul 12, 2010, at 9:07 AM, Brian Budge wrote:
>>>> 
>>>>> Hi Jody -
>>>>> 
>>>>> Thanks for the reply.  is there a way of "fusing" intercommunicators?
>>>>> Let's say I have a higher level node scheduler, and it makes a new
>>>>> node available to a COMM that is already running.  So the master
>>>>> spawns another process for that node.  How can the new process
>>>>> communicate with the other already started processes?
>>>> 
>>>> They can connect/accept via ompi-server - checkout "man ompi-server". You 
>>>> can also have them all rendezvous at a common mpirun if you prefer by 
>>>> using the appropriate mca param to give the required contact info.
>>>> 
>>>>> 
>>>>> Also, how can you specify with MPI_Comm_spawn/multiple() how do you
>>>>> specify IP addresses on which to start the processes?
>>>> 
>>>> Look at "man MPI_Comm_spawn"
>>>> 
>>>>> 
>>>>> If my higher level node scheduler needs to take away a process from my
>>>>> COMM, is it good/bad for that node to call MPI_Finalize as it exits?
>>>> 
>>>> The entire job will automatically abort if it fails to call Finalize as 
>>>> this is considered an abnormal termination event.
>>>> 
>>>>> 
>>>>> I would prefer not to use any of the MPI command-line utilities
>>>>> (mpirun/mpiexec) if that's possible.
>>>> 
>>>> Not possible at the moment. I have provided a patch to another user on 
>>>> this list that made it possible to use ompi-server as a rendezvous point 
>>>> for singletons, but that isn't in any formal release yet.
>>>> 
>>>>> 
>>>>> Thanks,
>>>>>  Brian
>>>>> 
>>>>> On Sat, Jul 10, 2010 at 11:53 PM, jody <jody....@gmail.com> wrote:
>>>>>> Hi Brian
>>>>>> When you spawn processes with MPI_Comm_spawn(), one of the arguments
>>>>>> will be set to an intercommunicator of thes spawner and the spawnees.
>>>>>> You can use this intercommunicator as the communicator argument
>>>>>> in the MPI_functions.
>>>>>> 
>>>>>> Jody
>>>>>> On Fri, Jul 9, 2010 at 5:56 PM, Brian Budge <brian.bu...@gmail.com> 
>>>>>> wrote:
>>>>>>> Hi all -
>>>>>>> 
>>>>>>> I've been looking at the dynamic process features of mpi-2.  I have 
>>>>>>> managed
>>>>>>> to actually launch processes using spawn, but haven't seen examples for
>>>>>>> actually communicating once these processes are launched.  I am 
>>>>>>> additionally
>>>>>>> interested in how processes created through multiple spawn calls can
>>>>>>> communicate.
>>>>>>> 
>>>>>>> Does anyone know of resources that describe these topics?  My google-fu 
>>>>>>> must
>>>>>>> not be up to par :)
>>>>>>> 
>>>>>>> Thanks,
>>>>>>>   Brian
>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>> 
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


Reply via email to