On Wed, Feb/06/2008 10:54:05AM, Josh Hursey wrote:
> 
> On Feb 6, 2008, at 10:25 AM, Jeff Squyres wrote:
> 
> > On Jan 31, 2008, at 5:07 PM, Josh Hursey wrote:
> >
> >> For the visualization it would be really nice to see how well  
> >> tested a
> >> particular interconnect, resource manager, and/or 'feature' is when
> >> ramping up to a release. However these peices of information are hard
> >> to obtain, and in some cases quantify (e.g., what do we mean by
> >> testing a 'feature'?).
> >>
> >> Thinking about this it occurred to me that what we really need is for
> >> OMPI to tell MTT what it is doing for some of these cases.
> >> Two examples, MTT cannot tell:
> >> - which set of compile time options are enabled/disabled
> >> automatically
> >>   e.g. [ "./configure --with-foo" vs "./configure"]
> >
> > Yes, this could be done.
> >
> >
> >> - which BTL(s) or MTL are used to run a test
> >>   e.g. [ "mpirun -mca btl tcp,self foo" vs. "mpirun foo"]
> >
> > Don't we offer this in a limited way right now with the "network"
> > field in the MPI details section?  I think we hesitated to put OMPI-
> > specific semantics on that field -- e.g., whether you're using the MX
> > BTL or MTL is an OMPI issue; you're still using the MX protocol/ 
> > network.
> >
> > I suppose we could agument those strings in the OMPI case: mx:mtl and
> > mx:btl, for example.
> >
> > So to be clear: does the network field not give you what you need?
> 
> The network field gives us exactly what we want. The problem is that  
> it is not filled in when we run "mpirun foo" since we do not specify  
> the BTLs on the command line (unless the INI explicitly specifies it).  
> The problems becomes further complicated when you run something like  
> "mpirun -mca btl openib,tcp,self" where the 'tcp' BTL is not going to  
> be used due to exclusivity (at least that is what I'm told), so we  
> miss report the BTLs used in this case.
> 
> >
> >
> >> For the configure options we *could* parse the config.log to extract
> >> this data. The question is, if we did this, what do we want to look?
> >> And is this something we want to do? Is there another way?
> >
> > I think having a network-like field for the MPI install section might
> > be good, and possibly have an OMPI:: funclet to automatically do the
> > parsing.  But we need to be mindful of MPIs that won't have a
> > configure script, so what information goes there might be dubious (or
> > just empty?).
> 
> Yeah I think an Open MPI specific function should do the parsing since  
> the configure options we want to grab will be specific to Open MPI. I  
> think in the case of no configure script it would just be empty.
> 

The info we are looking for in config.log is not available
in ompi_info? Parsing config.log throws a monkey wrench into
an AlreadyInstalled testing scenario.

-Ethan

> >
> >
> >>
> >> For the BTL(s)/MTL this is a much more subtle question since this
> >> depends on the connectivity of the interfaces on a machine, and the
> >> runtime selection logic. If we added a parameter to mpirun (e.g. "--
> >> showme connectivity") that displayed connectivity information to
> >> stdout (or a file) would this be useful? What should it look like?
> >
> > Ya, this is on my to-do list.  IB CM stuff in the openib BTL has been
> > consuming my time recently (much more complicated than I originally
> > thought); I swear I'll be getting to the connectivity map issue before
> > v1.3...
> 
> Is there a bug about this somewhere? There is a slim chance that I  
> (maybe Tim P) could help with this effort as well in the near term  
> (next month). For the simple case we could just dump the connectivity  
> information from Rank 0, then the more complex case will be full  
> mapping.
> 
> -- Josh
> 
> >
> >
> >>
> >> We have talked about some of this in the past, but I could not find a
> >> Bug talking about it in MTT.
> >>
> >> What do you think about this?
> >>
> >> Cheers,
> >> Josh
> >> _______________________________________________
> >> mtt-devel mailing list
> >> mtt-de...@open-mpi.org
> >> http://www.open-mpi.org/mailman/listinfo.cgi/mtt-devel
> >
> >
> > -- 
> > Jeff Squyres
> > Cisco Systems
> >
> > _______________________________________________
> > mtt-devel mailing list
> > mtt-de...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/mtt-devel
> 
> _______________________________________________
> mtt-devel mailing list
> mtt-de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/mtt-devel

Reply via email to