Hi!!! Thanks for help!!! Right now I am just trying to install the normal openmpi(without using all development header files). But it is still giving me some error. I have downloaded the developer version from the openmpi.org site. Then I gave ./configure --prefix=/net/hc293/pooja/dev_openmpi (lots of out put) make all install (lots of output ) and error :ld returned 1 exit status make[2]: *** [libopen-pal.la] Error 1 make[2]: Leaving directory `/net/hc293/pooja/openmpi-1.2.1a0r14362-dev/opal' make[1]: *** [all-recursive] Error 1 make[1]: Leaving directory `/net/hc293/pooja/openmpi-1.2.1a0r14362-dev/opal' make: *** [all-recursive] Error 1
Also the dev_openmpi folder is empty. SO I am not able to complie normal ring_c.c example also. Please help Thanks and Regards Pooja > Configure with the --with-devel-headers switch. This will install > all the developer headers. > > If you care, check out "./configure --help" -- that shows all the > options available to the configure script (including --with-devel- > headers). > > > On Apr 13, 2007, at 7:36 PM, po...@cc.gatech.edu wrote: > >> Hi >> >> I have downloaded the developer version of source code by >> downloading a >> nightly Subversion snapshot tarball.And have installed the openmpi. >> Using >> >> ./configure --prefix=/usr/local >> make all install. >> >> But I want to install with all the development headers.So that I >> can write >> an application that can use Ompi internal headers. >> >> >> Thanks and Regards >> Pooja >> >> >> >> >> >>> On Apr 1, 2007, at 3:12 PM, Ralph Castain wrote: >>> >>>> I can't help you with the BTL question. On the others: >>> >>> Yes, you can "sorta" call BTL's directly from application programs >>> (are you trying to use MPI alongside other communication libraries, >>> and using the BTL components as a sample?), but there are issues >>> involved with this. >>> >>> First, you need to install Open MPI with all the development >>> headers. Open MPI normally only installs "mpi.h" and a small number >>> of other heads; installing *all* the headers will allow you to write >>> applications that use OMPI's internal headers (such as btl.h) while >>> developing outside of the Open MPI source tree. >>> >>> Second, you probably won't want to access the BTL's directly. To >>> make this make sense, here's how the code is organized (even if the >>> specific call sequence is not exactly this layered for performance/ >>> optimization reasons): >>> >>> MPI layer (e.g., MPI_SEND) >>> -> PML >>> -> BML >>> -> BTL >>> >>> You have two choices: >>> >>> 1. Go through the PML instead (this is what we do in the MPI >>> collectives, for example) -- but this imposes MPI semantics on >>> sending and receiving, which assumedly you are trying to avoid. >>> Check out ompi/mca/pml/pml.h. >>> >>> 2. Go through the BML instead -- the BTL Management Layer. This is >>> essentially a multiplexor for all the BTLs that have been >>> instantiated. I'm guessing that this is what you want to do >>> (remember that OMPI has true multi-device support; using the BML and >>> multiple BTLs is one of the ways that we do this). Have a look at >>> ompi/mca/bml/bml.h for the interface. >>> >>> There is also currently no mechanism to get the BML and BTL pointers >>> that were instantiated by the PML. However, if you're just doing >>> proof-of-concept code, you can extract these directly from the MPI >>> layer's global variables to see how this stuff works. >>> >>> To have full interoperability of the underlying BTLs and between >>> multiple upper-layer communication libraries (e.g., between OMPI and >>> something else) is something that we have talked about a little, but >>> have not done much work on. >>> >>> To see the BTL interface (just for completeness), see ompi/mca/btl/ >>> btl.h. >>> >>> You can probably see the pattern here... In all of Open MPI's >>> frameworks, the public interface is in <level>/mca/<framework>/ >>> <framework>.h, where <level> is one of opal, orte, or ompi, and >>> <framework> is the name of the framework. >>> >>>> 1. states are reported via the orte/mca/smr framework. You will see >>>> the >>>> states listed in orte/mca/smr/smr_types.h. We track both process >>>> and job >>>> states. Hopefully, the state names will be somewhat self- >>>> explanatory and >>>> indicative of the order in which they are traversed. The job states >>>> are set >>>> when *all* of the processes in the job reach the corresponding >>>> state. >>> >>> Note that these are very coarse-grained process-level states (e.g., >>> is a given process running or not?). It's not clear what kind of >>> states you were asking about -- the Open MPI code base has many >>> internal state machines for various message passing and other >>> mechanisms. >>> >>> What information are you looking for, specifically? >>> >>>> 2. I'm not sure what you mean by mapping MPI processes to "physical" >>>> processes, but I assume you mean how do we assign MPI ranks to >>>> processes on >>>> specific nodes. You will find that done in the orte/mca/rmaps >>>> framework. We >>>> currently only have one component in that framework - the round- >>>> robin >>>> implementation - that maps either by slot or by node, as indicated >>>> by the >>>> user. That code is fairly heavily commented, so you hopefully can >>>> understand >>>> what it is doing. >>>> >>>> Hope that helps! >>>> Ralph >>>> >>>> >>>> On 4/1/07 1:32 PM, "po...@cc.gatech.edu" <po...@cc.gatech.edu> >>>> wrote: >>>> >>>>> Hi >>>>> I am Pooja and I am working on a course project which requires me >>>>> -> to track the internal state changes of MPI and need me to >>>>> figure out >>>>> how does ORTE maps MPi Process to actual physical processes >>>>> ->Also I need to find way to get BTL transports work directly with >>>>> MPI >>>>> level calls. >>>>> I just want to know is this posible and if yes what procedure I >>>>> should >>>>> follow or I should look into which files (for change). >>>>> >>>>> >>>>> Please Help >>>>> >>>>> Thanks and Regards >>>>> Pooja >>>>> >>>>> _______________________________________________ >>>>> devel mailing list >>>>> de...@open-mpi.org >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>> >>>> >>>> _______________________________________________ >>>> devel mailing list >>>> de...@open-mpi.org >>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>> >>> >>> -- >>> Jeff Squyres >>> Cisco Systems >>> >>> _______________________________________________ >>> devel mailing list >>> de...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>> >> >> _______________________________________________ >> devel mailing list >> de...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/devel > > > -- > Jeff Squyres > Cisco Systems > > _______________________________________________ > devel mailing list > de...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/devel >