Done!

On May 1, 2014, at 11:22 AM, George Bosilca <bosi...@icl.utk.edu> wrote:

> Apparently we are good today at 2PM EST. Fire-up the webex ;)
> 
>  George.
> 
> On May 1, 2014, at 10:35 , Jeff Squyres (jsquyres) <jsquy...@cisco.com> wrote:
> 
>> http://doodle.com/hhm4yyr76ipcxgk2
>> 
>> 
>> On May 1, 2014, at 10:25 AM, Ralph Castain <r...@open-mpi.org>
>> wrote:
>> 
>>> sure - might be faster that way :-)
>>> 
>>> On May 1, 2014, at 6:59 AM, Jeff Squyres (jsquyres) <jsquy...@cisco.com> 
>>> wrote:
>>> 
>>>> Want to have a phone call/webex to discuss?
>>>> 
>>>> 
>>>> On May 1, 2014, at 9:43 AM, Ralph Castain <r...@open-mpi.org> wrote:
>>>> 
>>>>> The problem we'll have with BTLs in opal is going to revolve around that 
>>>>> ompi_process_name_t and will occur in a number of places. I've been 
>>>>> trying to grok George's statement about accessors and can't figure out a 
>>>>> clean way to make that work IF every RTE gets to define the process name 
>>>>> a different way.
>>>>> 
>>>>> For example, suppose I define ompi_process_name_t to be a string. I can 
>>>>> hash the string down to an opal_identifier_t, but that is a structureless 
>>>>> 64-bit value - there is no concept of a jobid or vpid in it. So if you 
>>>>> now want to extract a jobid for that identifier, the only way you can do 
>>>>> it is to "up-call" back to the RTE to parse it.
>>>>> 
>>>>> This means that every RTE would have to initialize OPAL with a 
>>>>> registration of its opal_identifier parser function(s), which seems like 
>>>>> a really ugly solution.
>>>>> 
>>>>> Maybe it is time to shift the process identifier down to the opal layer? 
>>>>> If we define opal_identifier_t to include the required jobid/vpid, 
>>>>> perhaps adding a void* so someone can put whatever they want in it?
>>>>> 
>>>>> Note that I'm not wild about extending the identifier size beyond 64-bits 
>>>>> as the memory footprint issue is growing in concern, and I still haven't 
>>>>> seen any real use-case proposed for extending it.
>>>>> 
>>>>> 
>>>>> On May 1, 2014, at 3:41 AM, Jeff Squyres (jsquyres) <jsquy...@cisco.com> 
>>>>> wrote:
>>>>> 
>>>>>> On Apr 30, 2014, at 10:01 PM, George Bosilca <bosi...@icl.utk.edu> wrote:
>>>>>> 
>>>>>>> Why do you need the ompi_process_name_t? Isn’t the opal_identifier_t 
>>>>>>> enough to dig for the info of the peer into the opal_db?
>>>>>> 
>>>>>> 
>>>>>> At the moment, I use the ompi_process_name_t for RML sends/receives in 
>>>>>> the usnic BTL.  I know this will have to change when the BTLs move down 
>>>>>> to OPAL (when is that going to happen, BTW?).  So my future use case may 
>>>>>> be somewhat moot.
>>>>>> 
>>>>>> More detail
>>>>>> ===========
>>>>>> 
>>>>>> "Why does the usnic BTL use RML sends/receives?", you ask.
>>>>>> 
>>>>>> The reason is rooted in the fact that the usnic BTL uses an unreliable, 
>>>>>> connectionless transport under the covert.  We had some customers have 
>>>>>> network misconfigurations that resulted in usnic traffic not flowing 
>>>>>> properly (e.g., MTU mismatches in the network).  But since we don't have 
>>>>>> a connection-oriented underlying API that will eventually timeout/fail 
>>>>>> to connect/etc. when there's a problem with the network configuration, 
>>>>>> we added a "connection validation" service in the usnic BTL that fires 
>>>>>> up in a thread in the local rank 0 on each server.  This thread provides 
>>>>>> service to all the MPI processes on its server.
>>>>>> 
>>>>>> In short: the service thread sends UDP pings and ACKs to peer service 
>>>>>> threads on other servers (upon demand/upon first send between servers) 
>>>>>> to verify network connectivity.  If the pings eventually fail/timeout 
>>>>>> (i.e., don't get ACKs back), the service thread does a show_help and 
>>>>>> kills the job. 
>>>>>> 
>>>>>> There's more details, but that's the gist of it.
>>>>>> 
>>>>>> This basically gives us the ability to highlight problems in the network 
>>>>>> and kill the MPI job rather than spin infinitely while trying to deliver 
>>>>>> MPI/BTL messages to a peer that will never get there.
>>>>>> 
>>>>>> Since this is really a server-to-server network connectivity issue (vs. 
>>>>>> an MPI peer-to-peer connectivity issue), we only need to have one 
>>>>>> service thread for a whole server.  The other MPI procs on the server 
>>>>>> use RML to talk to it.  E.g., "Please ping the server where MPI proc X 
>>>>>> lives," and so on.  This seemed better than having a service thread in 
>>>>>> each MPI process.
>>>>>> 
>>>>>> We've thought a bit about what to do when the BTLs move down to OPAL 
>>>>>> (since they won't be able to use RML any more), but don't have a final 
>>>>>> solution yet...  We do still want to be able to utilize this capability 
>>>>>> even after the BTL move.
>>>>>> 
>>>>>> -- 
>>>>>> Jeff Squyres
>>>>>> jsquy...@cisco.com
>>>>>> For corporate legal information go to: 
>>>>>> http://www.cisco.com/web/about/doing_business/legal/cri/
>>>>>> 
>>>>>> _______________________________________________
>>>>>> devel mailing list
>>>>>> de...@open-mpi.org
>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>>>>> Link to this post: 
>>>>>> http://www.open-mpi.org/community/lists/devel/2014/05/14673.php
>>>>> 
>>>>> _______________________________________________
>>>>> devel mailing list
>>>>> de...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>>>> Link to this post: 
>>>>> http://www.open-mpi.org/community/lists/devel/2014/05/14674.php
>>>> 
>>>> 
>>>> -- 
>>>> Jeff Squyres
>>>> jsquy...@cisco.com
>>>> For corporate legal information go to: 
>>>> http://www.cisco.com/web/about/doing_business/legal/cri/
>>>> 
>>>> _______________________________________________
>>>> devel mailing list
>>>> de...@open-mpi.org
>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>>> Link to this post: 
>>>> http://www.open-mpi.org/community/lists/devel/2014/05/14675.php
>>> 
>>> _______________________________________________
>>> devel mailing list
>>> de...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/devel/2014/05/14676.php
>> 
>> 
>> -- 
>> Jeff Squyres
>> jsquy...@cisco.com
>> For corporate legal information go to: 
>> http://www.cisco.com/web/about/doing_business/legal/cri/
>> 
>> _______________________________________________
>> devel mailing list
>> de...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/devel/2014/05/14677.php
> 
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post: 
> http://www.open-mpi.org/community/lists/devel/2014/05/14678.php


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/

Reply via email to