Sorry for the very late reply. Everything works now! Thanks a lot!! On Sep 25, 2013, at 7:00 PM, Ralph Castain wrote:
> I've committed a fix to the trunk (r29245) and scheduled it for v1.7.3 - > thanks for the debug info! > > Ralph > > On Sep 25, 2013, at 5:00 AM, Suraj Prabhakaran <suraj.prabhaka...@gmail.com> > wrote: > >> Dear Ralph, >> >> I am sorry but I think I missed adding plm verbosity to 5 last time. Here is >> the output of the complete program with and without -novm to the following >> mpiexec. >> >> mpiexec -mca state_base_verbose 10 -mca errmgr_base_verbose 10 -mca >> plm_base_verbose 5 -mca btl tcp,sm,self -np 2 ./addhosttest >> mpiexec -mca state_base_verbose 10 -mca errmgr_base_verbose 10 -mca >> plm_base_verbose 5 -mca btl tcp,sm,self -novm -np 2 ./addhosttest >> >> Here you can see that although I spawn only one process on grsacc18, >> something is also done with grsacc19. >> >> Sorry and thanks! >> Suraj >> <output.rtf><output-novm.rtf> >> >> >> On Sep 24, 2013, at 8:24 PM, Ralph Castain wrote: >> >>> What I find puzzling is that I don't see any output indicating that you >>> went thru the Torque launcher to launch the daemons - not a peep of debug >>> output. This makes me suspicious that something else is going on. Are you >>> sure you sent me all the output? >>> >>> Try adding -novm to your mpirun cmd line and let's see if that mode works >>> >>> On Sep 24, 2013, at 9:06 AM, Suraj Prabhakaran >>> <suraj.prabhaka...@gmail.com> wrote: >>> >>>> Hi Ralph, >>>> >>>> So here is what I do. I spawn just a "single" process on a new node which >>>> is basically not in the $PBS_NODEFILE list. >>>> My $PBS_NODEFILE list contains >>>> grsacc20 >>>> grsacc19 >>>> >>>> I then start the app with just 2 processes. So one host gets one process >>>> and they are successfully spawned through the torque (through tm_spawn()). >>>> MPI would have stored grsacc20 and grsacc19 to its list of hosts with >>>> launchid 0 and 1 correspondingly. >>>> I then use the add-host info and spawn ONE new process on a new host >>>> "grsacc18" through MPI_Comm_spawn. From what I saw in the code, the >>>> launchid of this new host is -1 since openmpi does not know about this and >>>> it is not available in the $PBS_NODEFILE. Since withouth the launchid, >>>> torque would not know where to spawn, I just retrieve the correct launchid >>>> of this host from a file just before tm_spawn() and use this launchid. >>>> This is the only modification that I made to openmpi. >>>> So, the host "grsacc18" will have a new launchid = 2 and will be used to >>>> spawn the process through torque. This worked perfectly until 1.6.5. >>>> >>>> As we see here from the outputs, although I spawn only a single process on >>>> grsacc18, I too have no clue why openmpi tries to spawn something on >>>> grsacc19. Of course, without pbs/torque involved, everything works fine. >>>> I have attached the simple test code. Please modify hostnames and >>>> executable path before use. >>>> >>>> Best, >>>> Suraj >>>> >>>> <addhosttest.c> >>>> >>>> >>>> On Sep 24, 2013, at 4:59 PM, Ralph Castain wrote: >>>> >>>>> I'm going to need a little help here. The problem is that you launch two >>>>> new daemons, and one of them exits immediately because it thinks it lost >>>>> the connection back to mpirun - before it even gets a chance to create it. >>>>> >>>>> Can you give me a little more info as to exactly what you are doing? >>>>> Perhaps send me your test code? >>>>> >>>>> On Sep 24, 2013, at 7:48 AM, Suraj Prabhakaran >>>>> <suraj.prabhaka...@gmail.com> wrote: >>>>> >>>>>> Hi Ralph, >>>>>> >>>>>> Output attached in a file. >>>>>> Thanks a lot! >>>>>> >>>>>> Best, >>>>>> Suraj >>>>>> >>>>>> <output.rtf> >>>>>> >>>>>> On Sep 24, 2013, at 4:11 PM, Ralph Castain wrote: >>>>>> >>>>>>> Afraid I don't see the problem offhand - can you add the following to >>>>>>> your cmd line? >>>>>>> >>>>>>> -mca state_base_verbose 10 -mca errmgr_base_verbose 10 >>>>>>> >>>>>>> Thanks >>>>>>> Ralph >>>>>>> >>>>>>> On Sep 24, 2013, at 6:35 AM, Suraj Prabhakaran >>>>>>> <suraj.prabhaka...@gmail.com> wrote: >>>>>>> >>>>>>>> Hi Ralph, >>>>>>>> >>>>>>>> I always got this output from any MPI job that ran on our nodes. There >>>>>>>> seems to be a problem somewhere but it never stopped the applications >>>>>>>> from running. But anyway, I ran it again now with only tcp and >>>>>>>> excluded the infiniband and I get the same output again. Except that >>>>>>>> this time, the error related to this openib is not there anymore. >>>>>>>> Printing out the log again. >>>>>>>> >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive processing msg >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive job launch command from >>>>>>>> [[6160,1],0] >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive adding hosts >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive calling spawn >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive done processing commands >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:setup_job >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:setup_vm >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:setup_vm add new daemon >>>>>>>> [[6160,0],2] >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:setup_vm assigning new daemon >>>>>>>> [[6160,0],2] to node grsacc18 >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:tm: launching vm >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:tm: final top-level argv: >>>>>>>> orted -mca ess tm -mca orte_ess_jobid 403701760 -mca >>>>>>>> orte_ess_vpid <template> -mca orte_ess_num_procs 3 -mca orte_hnp_uri >>>>>>>> "403701760.0;tcp://192.168.222.20:35163" -mca plm_base_verbose 5 -mca >>>>>>>> btl tcp,sm,self >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:tm: launching on node grsacc19 >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:tm: executing: >>>>>>>> orted -mca ess tm -mca orte_ess_jobid 403701760 -mca >>>>>>>> orte_ess_vpid 1 -mca orte_ess_num_procs 3 -mca orte_hnp_uri >>>>>>>> "403701760.0;tcp://192.168.222.20:35163" -mca plm_base_verbose 5 -mca >>>>>>>> btl tcp,sm,self >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:tm: launching on node grsacc18 >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:tm: executing: >>>>>>>> orted -mca ess tm -mca orte_ess_jobid 403701760 -mca >>>>>>>> orte_ess_vpid 2 -mca orte_ess_num_procs 3 -mca orte_hnp_uri >>>>>>>> "403701760.0;tcp://192.168.222.20:35163" -mca plm_base_verbose 5 -mca >>>>>>>> btl tcp,sm,self >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:tm:launch: finished spawning orteds >>>>>>>> [grsacc19:28821] mca:base:select:( plm) Querying component [rsh] >>>>>>>> [grsacc19:28821] [[6160,0],1] plm:rsh_lookup on agent ssh : rsh path >>>>>>>> NULL >>>>>>>> [grsacc19:28821] mca:base:select:( plm) Query of component [rsh] set >>>>>>>> priority to 10 >>>>>>>> [grsacc19:28821] mca:base:select:( plm) Selected component [rsh] >>>>>>>> [grsacc19:28821] [[6160,0],1] plm:rsh_setup on agent ssh : rsh path >>>>>>>> NULL >>>>>>>> [grsacc19:28821] [[6160,0],1] plm:base:receive start comm >>>>>>>> [grsacc19:28821] [[6160,0],1] plm:base:receive stop comm >>>>>>>> [grsacc18:16717] mca:base:select:( plm) Querying component [rsh] >>>>>>>> [grsacc18:16717] [[6160,0],2] plm:rsh_lookup on agent ssh : rsh path >>>>>>>> NULL >>>>>>>> [grsacc18:16717] mca:base:select:( plm) Query of component [rsh] set >>>>>>>> priority to 10 >>>>>>>> [grsacc18:16717] mca:base:select:( plm) Selected component [rsh] >>>>>>>> [grsacc18:16717] [[6160,0],2] plm:rsh_setup on agent ssh : rsh path >>>>>>>> NULL >>>>>>>> [grsacc18:16717] [[6160,0],2] plm:base:receive start comm >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:orted_report_launch from daemon >>>>>>>> [[6160,0],2] >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:orted_report_launch from daemon >>>>>>>> [[6160,0],2] on node grsacc18 >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:orted_report_launch completed >>>>>>>> for daemon [[6160,0],2] at contact >>>>>>>> 403701760.2;tcp://192.168.222.18:44229 >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:launch_apps for job [6160,2] >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive processing msg >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive update proc state >>>>>>>> command from [[6160,0],2] >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive got update_proc_state >>>>>>>> for job [6160,2] >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive got update_proc_state >>>>>>>> for vpid 0 state RUNNING exit_code 0 >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive done processing commands >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:launch wiring up iof for job >>>>>>>> [6160,2] >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive processing msg >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive done processing commands >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:launch registered event >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:launch sending dyn release of >>>>>>>> job [6160,2] to [[6160,1],0] >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:orted_cmd sending orted_exit >>>>>>>> commands >>>>>>>> [grsacc19:28815] [[6160,0],1] plm:base:receive stop comm >>>>>>>> [grsacc20:04578] [[6160,0],0] plm:base:receive stop comm >>>>>>>> -bash-4.1$ [grsacc18:16717] [[6160,0],2] plm:base:receive stop comm >>>>>>>> >>>>>>>> Best, >>>>>>>> Suraj >>>>>>>> On Sep 24, 2013, at 3:24 PM, Ralph Castain wrote: >>>>>>>> >>>>>>>>> Your output shows that it launched your apps, but they exited. The >>>>>>>>> error is reported here, though it appears we aren't flushing the >>>>>>>>> message out before exiting due to a race condition: >>>>>>>>> >>>>>>>>>> [grsacc20:04511] 1 more process has sent help message >>>>>>>>>> help-mpi-btl-openib.txt / no active ports found >>>>>>>>> >>>>>>>>> Here is the full text: >>>>>>>>> [no active ports found] >>>>>>>>> WARNING: There is at least non-excluded one OpenFabrics device found, >>>>>>>>> but there are no active ports detected (or Open MPI was unable to use >>>>>>>>> them). This is most certainly not what you wanted. Check your >>>>>>>>> cables, subnet manager configuration, etc. The openib BTL will be >>>>>>>>> ignored for this job. >>>>>>>>> >>>>>>>>> Local host: %s >>>>>>>>> >>>>>>>>> Looks like at least one node being used doesn't have an active >>>>>>>>> Infiniband port on it? >>>>>>>>> >>>>>>>>> >>>>>>>>> On Sep 24, 2013, at 6:11 AM, Suraj Prabhakaran >>>>>>>>> <suraj.prabhaka...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Hi Ralph, >>>>>>>>>> >>>>>>>>>> I tested it with the trunk r29228. I still have the following >>>>>>>>>> problem. Now, it even spawns the daemon on the new node through >>>>>>>>>> torque but then suddently quits. The following is the output. Can >>>>>>>>>> you please have a look? >>>>>>>>>> >>>>>>>>>> Thanks >>>>>>>>>> Suraj >>>>>>>>>> >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive processing msg >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive job launch command >>>>>>>>>> from [[6253,1],0] >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive adding hosts >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive calling spawn >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive done processing >>>>>>>>>> commands >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:setup_job >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:setup_vm >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:setup_vm add new daemon >>>>>>>>>> [[6253,0],2] >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:setup_vm assigning new daemon >>>>>>>>>> [[6253,0],2] to node grsacc18 >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:tm: launching vm >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:tm: final top-level argv: >>>>>>>>>> orted -mca ess tm -mca orte_ess_jobid 409796608 -mca >>>>>>>>>> orte_ess_vpid <template> -mca orte_ess_num_procs 3 -mca orte_hnp_uri >>>>>>>>>> "409796608.0;tcp://192.168.222.20:53097" -mca plm_base_verbose 6 >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:tm: launching on node grsacc19 >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:tm: executing: >>>>>>>>>> orted -mca ess tm -mca orte_ess_jobid 409796608 -mca >>>>>>>>>> orte_ess_vpid 1 -mca orte_ess_num_procs 3 -mca orte_hnp_uri >>>>>>>>>> "409796608.0;tcp://192.168.222.20:53097" -mca plm_base_verbose 6 >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:tm: launching on node grsacc18 >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:tm: executing: >>>>>>>>>> orted -mca ess tm -mca orte_ess_jobid 409796608 -mca >>>>>>>>>> orte_ess_vpid 2 -mca orte_ess_num_procs 3 -mca orte_hnp_uri >>>>>>>>>> "409796608.0;tcp://192.168.222.20:53097" -mca plm_base_verbose 6 >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:tm:launch: finished spawning orteds >>>>>>>>>> [grsacc19:28754] mca:base:select:( plm) Querying component [rsh] >>>>>>>>>> [grsacc19:28754] [[6253,0],1] plm:rsh_lookup on agent ssh : rsh path >>>>>>>>>> NULL >>>>>>>>>> [grsacc19:28754] mca:base:select:( plm) Query of component [rsh] >>>>>>>>>> set priority to 10 >>>>>>>>>> [grsacc19:28754] mca:base:select:( plm) Selected component [rsh] >>>>>>>>>> [grsacc19:28754] [[6253,0],1] plm:rsh_setup on agent ssh : rsh path >>>>>>>>>> NULL >>>>>>>>>> [grsacc19:28754] [[6253,0],1] plm:base:receive start comm >>>>>>>>>> [grsacc19:28754] [[6253,0],1] plm:base:receive stop comm >>>>>>>>>> [grsacc18:16648] mca:base:select:( plm) Querying component [rsh] >>>>>>>>>> [grsacc18:16648] [[6253,0],2] plm:rsh_lookup on agent ssh : rsh path >>>>>>>>>> NULL >>>>>>>>>> [grsacc18:16648] mca:base:select:( plm) Query of component [rsh] >>>>>>>>>> set priority to 10 >>>>>>>>>> [grsacc18:16648] mca:base:select:( plm) Selected component [rsh] >>>>>>>>>> [grsacc18:16648] [[6253,0],2] plm:rsh_setup on agent ssh : rsh path >>>>>>>>>> NULL >>>>>>>>>> [grsacc18:16648] [[6253,0],2] plm:base:receive start comm >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:orted_report_launch from >>>>>>>>>> daemon [[6253,0],2] >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:orted_report_launch from >>>>>>>>>> daemon [[6253,0],2] on node grsacc18 >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:orted_report_launch completed >>>>>>>>>> for daemon [[6253,0],2] at contact >>>>>>>>>> 409796608.2;tcp://192.168.222.18:47974 >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:launch_apps for job [6253,2] >>>>>>>>>> [grsacc20:04511] 1 more process has sent help message >>>>>>>>>> help-mpi-btl-openib.txt / no active ports found >>>>>>>>>> [grsacc20:04511] Set MCA parameter "orte_base_help_aggregate" to 0 >>>>>>>>>> to see all help / error messages >>>>>>>>>> [grsacc20:04511] 1 more process has sent help message >>>>>>>>>> help-mpi-btl-base.txt / btl:no-nics >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive processing msg >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive update proc state >>>>>>>>>> command from [[6253,0],2] >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive got update_proc_state >>>>>>>>>> for job [6253,2] >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive got update_proc_state >>>>>>>>>> for vpid 0 state RUNNING exit_code 0 >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive done processing >>>>>>>>>> commands >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:launch wiring up iof for job >>>>>>>>>> [6253,2] >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive processing msg >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive done processing >>>>>>>>>> commands >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:launch registered event >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:launch sending dyn release of >>>>>>>>>> job [6253,2] to [[6253,1],0] >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:orted_cmd sending orted_exit >>>>>>>>>> commands >>>>>>>>>> [grsacc19:28747] [[6253,0],1] plm:base:receive stop comm >>>>>>>>>> [grsacc20:04511] [[6253,0],0] plm:base:receive stop comm >>>>>>>>>> -bash-4.1$ [grsacc18:16648] [[6253,0],2] plm:base:receive stop comm >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Sep 23, 2013, at 1:55 AM, Ralph Castain wrote: >>>>>>>>>> >>>>>>>>>>> Found a bug in the Torque support - we were trying to connect to >>>>>>>>>>> the MOM again, which would hang (I imagine). I pushed a fix to the >>>>>>>>>>> trunk (r29227) and scheduled it to come to 1.7.3 if you want to try >>>>>>>>>>> it again. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Sep 22, 2013, at 4:21 PM, Suraj Prabhakaran >>>>>>>>>>> <suraj.prabhaka...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Dear Ralph, >>>>>>>>>>>> >>>>>>>>>>>> This is the output I get when I execute with the verbose option. >>>>>>>>>>>> >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:receive processing msg >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:receive job launch command >>>>>>>>>>>> from [[23526,1],0] >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:receive adding hosts >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:receive calling spawn >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:receive done processing >>>>>>>>>>>> commands >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:setup_job >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:setup_vm >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:setup_vm add new daemon >>>>>>>>>>>> [[23526,0],2] >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:setup_vm assigning new >>>>>>>>>>>> daemon [[23526,0],2] to node grsacc17/1-4 >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:setup_vm add new daemon >>>>>>>>>>>> [[23526,0],3] >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:setup_vm assigning new >>>>>>>>>>>> daemon [[23526,0],3] to node grsacc17/0-5 >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:tm: launching vm >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:tm: final top-level argv: >>>>>>>>>>>> orted -mca ess tm -mca orte_ess_jobid 1541799936 -mca >>>>>>>>>>>> orte_ess_vpid <template> -mca orte_ess_num_procs 4 -mca >>>>>>>>>>>> orte_hnp_uri "1541799936.0;tcp://192.168.222.20:49049" -mca >>>>>>>>>>>> plm_base_verbose 5 >>>>>>>>>>>> [warn] opal_libevent2021_event_base_loop: reentrant invocation. >>>>>>>>>>>> Only one event_base_loop can run on each event_base at once. >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:orted_cmd sending >>>>>>>>>>>> orted_exit commands >>>>>>>>>>>> [grsacc20:21012] [[23526,0],0] plm:base:receive stop comm >>>>>>>>>>>> >>>>>>>>>>>> Says something? >>>>>>>>>>>> >>>>>>>>>>>> Best, >>>>>>>>>>>> Suraj >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Sep 22, 2013, at 9:45 PM, Ralph Castain wrote: >>>>>>>>>>>> >>>>>>>>>>>>> I'll still need to look at the intercomm_create issue, but I just >>>>>>>>>>>>> tested both the trunk and current 1.7.3 branch for "add-host" and >>>>>>>>>>>>> both worked just fine. This was on my little test cluster which >>>>>>>>>>>>> only has rsh available - no Torque. >>>>>>>>>>>>> >>>>>>>>>>>>> You might add "-mca plm_base_verbose 5" to your cmd line to get >>>>>>>>>>>>> some debug output as to the problem. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Sep 21, 2013, at 5:48 PM, Ralph Castain <r...@open-mpi.org> >>>>>>>>>>>>> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Sep 21, 2013, at 4:54 PM, Suraj Prabhakaran >>>>>>>>>>>>>> <suraj.prabhaka...@gmail.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Dear all, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Really thanks a lot for your efforts. I too downloaded the >>>>>>>>>>>>>>> trunk to check if it works for my case and as of revision >>>>>>>>>>>>>>> 29215, it works for the original case I reported. Although it >>>>>>>>>>>>>>> works, I still see the following in the output. Does it mean >>>>>>>>>>>>>>> anything? >>>>>>>>>>>>>>> [grsacc17][[13611,1],0][btl_openib_proc.c:157:mca_btl_openib_proc_create] >>>>>>>>>>>>>>> [btl_openib_proc.c:157] ompi_modex_recv failed for peer >>>>>>>>>>>>>>> [[13611,2],0] >>>>>>>>>>>>>> >>>>>>>>>>>>>> Yes - it means we don't quite have this right yet :-( >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> However, on another topic relevant to my use case, I have >>>>>>>>>>>>>>> another problem to report. I am having problems using the >>>>>>>>>>>>>>> "add-host" info to the MPI_Comm_spawn() when MPI is compiled >>>>>>>>>>>>>>> with support for Torque resource manager. This problem is >>>>>>>>>>>>>>> totally new in the 1.7 series and it worked perfectly until >>>>>>>>>>>>>>> 1.6.5 >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Basically, I am working on implementing dynamic resource >>>>>>>>>>>>>>> management facilities in the Torque/Maui batch system. Through >>>>>>>>>>>>>>> a new tm call, an application can get new resources for a job. >>>>>>>>>>>>>> >>>>>>>>>>>>>> FWIW: you'll find that we added an API to the orte RAS framework >>>>>>>>>>>>>> to support precisely that operation. It allows an application to >>>>>>>>>>>>>> request that we dynamically obtain additional resources during >>>>>>>>>>>>>> execution (e.g., as part of a Comm_spawn call via an info_key). >>>>>>>>>>>>>> We originally implemented this with Slurm, but you could add the >>>>>>>>>>>>>> calls into the Torque component as well if you like. >>>>>>>>>>>>>> >>>>>>>>>>>>>> This is in the trunk now - will come over to 1.7.4 >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> I want to use MPI_Comm_spawn() to spawn new processes in the >>>>>>>>>>>>>>> new hosts. With my extended torque/maui batch system, I was >>>>>>>>>>>>>>> able to perfectly use the "add-host" info argument to >>>>>>>>>>>>>>> MPI_Comm_spawn() to spawn new processes on these hosts. Since >>>>>>>>>>>>>>> MPI and Torque refer to the hosts through the nodeids, I made >>>>>>>>>>>>>>> sure that OpenMPI uses the correct nodeid's for these new >>>>>>>>>>>>>>> hosts. >>>>>>>>>>>>>>> Until 1.6.5, this worked perfectly fine, except that due to the >>>>>>>>>>>>>>> Intercomm_merge problem, I could not really run a real >>>>>>>>>>>>>>> application to its completion. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> While this is now fixed in the trunk, I found that, however, >>>>>>>>>>>>>>> when using the "add-host" info argument, everything collapses >>>>>>>>>>>>>>> after printing out the following error. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> [warn] opal_libevent2021_event_base_loop: reentrant invocation. >>>>>>>>>>>>>>> Only one event_base_loop can run on each event_base at once. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I'll take a look - probably some stale code that hasn't been >>>>>>>>>>>>>> updated yet for async ORTE operations >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> And due to this, I am still not really able to run my >>>>>>>>>>>>>>> application! I also compiled the MPI without any Torque/PBS >>>>>>>>>>>>>>> support and just used the "add-host" argument normally. Again, >>>>>>>>>>>>>>> this worked perfectly in 1.6.5. But in the 1.7 series, it works >>>>>>>>>>>>>>> but after printing out the following error. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> [grsacc17][[13731,1],0][btl_openib_proc.c:157:mca_btl_openib_proc_create] >>>>>>>>>>>>>>> [btl_openib_proc.c:157] ompi_modex_recv failed for peer >>>>>>>>>>>>>>> [[13731,2],0] >>>>>>>>>>>>>>> [grsacc17][[13731,1],1][btl_openib_proc.c:157:mca_btl_openib_proc_create] >>>>>>>>>>>>>>> [btl_openib_proc.c:157] ompi_modex_recv failed for peer >>>>>>>>>>>>>>> [[13731,2],0] >>>>>>>>>>>>>> >>>>>>>>>>>>>> Yeah, the 1.7 series doesn't have the reentrant test in it - so >>>>>>>>>>>>>> we "illegally" re-enter libevent. The error again means we don't >>>>>>>>>>>>>> have Intercomm_create correct just yet. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I'll see what I can do about this and get back to you >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> In short, with pbs/torque support, it fails and without >>>>>>>>>>>>>>> pbs/torque support, it runs after spitting the above lines. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I would really appreciate some help on this, since I need these >>>>>>>>>>>>>>> features to actually test my case and (at least in my short >>>>>>>>>>>>>>> experience) no other MPI implementation seem friendly to such >>>>>>>>>>>>>>> dynamic scenarios. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks a lot! >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Best, >>>>>>>>>>>>>>> Suraj >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Sep 20, 2013, at 4:58 PM, Jeff Squyres (jsquyres) wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Just to close my end of this loop: as of trunk r29213, it all >>>>>>>>>>>>>>>> works for me. Thanks! >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Sep 18, 2013, at 12:52 PM, Ralph Castain >>>>>>>>>>>>>>>> <r...@open-mpi.org> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Thanks George - much appreciated >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Sep 18, 2013, at 9:49 AM, George Bosilca >>>>>>>>>>>>>>>>> <bosi...@icl.utk.edu> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> The test case was broken. I just pushed a fix. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> George. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On Sep 18, 2013, at 16:49 , Ralph Castain >>>>>>>>>>>>>>>>>> <r...@open-mpi.org> wrote: >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Hangs with any np > 1 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> However, I'm not sure if that's an issue with the test vs >>>>>>>>>>>>>>>>>>> the underlying implementation >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> On Sep 18, 2013, at 7:40 AM, "Jeff Squyres (jsquyres)" >>>>>>>>>>>>>>>>>>> <jsquy...@cisco.com> wrote: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Does it hang when you run with -np 4? >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Sent from my phone. No type good. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> On Sep 18, 2013, at 4:10 PM, "Ralph Castain" >>>>>>>>>>>>>>>>>>>> <r...@open-mpi.org> wrote: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Strange - it works fine for me on my Mac. However, I see >>>>>>>>>>>>>>>>>>>>> one difference - I only run it with np=1 >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> On Sep 18, 2013, at 2:22 AM, Jeff Squyres (jsquyres) >>>>>>>>>>>>>>>>>>>>> <jsquy...@cisco.com> wrote: >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> On Sep 18, 2013, at 9:33 AM, George Bosilca >>>>>>>>>>>>>>>>>>>>>> <bosi...@icl.utk.edu> wrote: >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 1. sm doesn't work between spawned processes. So you >>>>>>>>>>>>>>>>>>>>>>> must have another network enabled. >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> I know :-). I have tcp available as well (OMPI will >>>>>>>>>>>>>>>>>>>>>> abort if you only run with sm,self because the >>>>>>>>>>>>>>>>>>>>>> comm_spawn will fail with unreachable errors -- I just >>>>>>>>>>>>>>>>>>>>>> tested/proved this to myself). >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2. Don't use the test case attached to my email, I left >>>>>>>>>>>>>>>>>>>>>>> an xterm based spawn and the debugging. It can't work >>>>>>>>>>>>>>>>>>>>>>> without xterm support. Instead try using the test case >>>>>>>>>>>>>>>>>>>>>>> from the trunk, the one committed by Ralph. >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> I didn't see any "xterm" strings in there, but ok. :-) >>>>>>>>>>>>>>>>>>>>>> I ran with orte/test/mpi/intercomm_create.c, and that >>>>>>>>>>>>>>>>>>>>>> hangs for me as well: >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> ----- >>>>>>>>>>>>>>>>>>>>>> ❯❯❯ mpicc intercomm_create.c -o intercomm_create >>>>>>>>>>>>>>>>>>>>>> ❯❯❯ mpirun -np 4 intercomm_create >>>>>>>>>>>>>>>>>>>>>> b: MPI_Intercomm_create( intra, 0, intra, MPI_COMM_NULL, >>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 4] >>>>>>>>>>>>>>>>>>>>>> b: MPI_Intercomm_create( intra, 0, intra, MPI_COMM_NULL, >>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 5] >>>>>>>>>>>>>>>>>>>>>> b: MPI_Intercomm_create( intra, 0, intra, MPI_COMM_NULL, >>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 6] >>>>>>>>>>>>>>>>>>>>>> b: MPI_Intercomm_create( intra, 0, intra, MPI_COMM_NULL, >>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 7] >>>>>>>>>>>>>>>>>>>>>> c: MPI_Intercomm_create( MPI_COMM_WORLD, 0, intra, 0, >>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 4] >>>>>>>>>>>>>>>>>>>>>> c: MPI_Intercomm_create( MPI_COMM_WORLD, 0, intra, 0, >>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 5] >>>>>>>>>>>>>>>>>>>>>> c: MPI_Intercomm_create( MPI_COMM_WORLD, 0, intra, 0, >>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 6] >>>>>>>>>>>>>>>>>>>>>> c: MPI_Intercomm_create( MPI_COMM_WORLD, 0, intra, 0, >>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 7] >>>>>>>>>>>>>>>>>>>>>> a: MPI_Intercomm_create( ab_intra, 0, ac_intra, 0, 201, >>>>>>>>>>>>>>>>>>>>>> &inter) (0) >>>>>>>>>>>>>>>>>>>>>> a: MPI_Intercomm_create( ab_intra, 0, ac_intra, 0, 201, >>>>>>>>>>>>>>>>>>>>>> &inter) (0) >>>>>>>>>>>>>>>>>>>>>> a: MPI_Intercomm_create( ab_intra, 0, ac_intra, 0, 201, >>>>>>>>>>>>>>>>>>>>>> &inter) (0) >>>>>>>>>>>>>>>>>>>>>> a: MPI_Intercomm_create( ab_intra, 0, ac_intra, 0, 201, >>>>>>>>>>>>>>>>>>>>>> &inter) (0) >>>>>>>>>>>>>>>>>>>>>> [hang] >>>>>>>>>>>>>>>>>>>>>> ----- >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Similarly, on my Mac, it hangs with no output: >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> ----- >>>>>>>>>>>>>>>>>>>>>> ❯❯❯ mpicc intercomm_create.c -o intercomm_create >>>>>>>>>>>>>>>>>>>>>> ❯❯❯ mpirun -np 4 intercomm_create >>>>>>>>>>>>>>>>>>>>>> [hang] >>>>>>>>>>>>>>>>>>>>>> ----- >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> George. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> On Sep 18, 2013, at 07:53 , "Jeff Squyres (jsquyres)" >>>>>>>>>>>>>>>>>>>>>>> <jsquy...@cisco.com> wrote: >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> George -- >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> When I build the SVN trunk (r29201) on 64 bit linux, >>>>>>>>>>>>>>>>>>>>>>>> your attached test case hangs: >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> ----- >>>>>>>>>>>>>>>>>>>>>>>> ❯❯❯ mpicc intercomm_create.c -o intercomm_create >>>>>>>>>>>>>>>>>>>>>>>> ❯❯❯ mpirun -np 4 intercomm_create >>>>>>>>>>>>>>>>>>>>>>>> b: MPI_Intercomm_create( intra, 0, intra, >>>>>>>>>>>>>>>>>>>>>>>> MPI_COMM_NULL, 201, &inter) [rank 4] >>>>>>>>>>>>>>>>>>>>>>>> b: MPI_Intercomm_create( intra, 0, intra, >>>>>>>>>>>>>>>>>>>>>>>> MPI_COMM_NULL, 201, &inter) [rank 5] >>>>>>>>>>>>>>>>>>>>>>>> b: MPI_Intercomm_create( intra, 0, intra, >>>>>>>>>>>>>>>>>>>>>>>> MPI_COMM_NULL, 201, &inter) [rank 6] >>>>>>>>>>>>>>>>>>>>>>>> b: MPI_Intercomm_create( intra, 0, intra, >>>>>>>>>>>>>>>>>>>>>>>> MPI_COMM_NULL, 201, &inter) [rank 7] >>>>>>>>>>>>>>>>>>>>>>>> a: MPI_Intercomm_create( ab_intra, 0, ac_intra, 0, >>>>>>>>>>>>>>>>>>>>>>>> 201, &inter) (0) >>>>>>>>>>>>>>>>>>>>>>>> a: MPI_Intercomm_create( ab_intra, 0, ac_intra, 0, >>>>>>>>>>>>>>>>>>>>>>>> 201, &inter) (0) >>>>>>>>>>>>>>>>>>>>>>>> a: MPI_Intercomm_create( ab_intra, 0, ac_intra, 0, >>>>>>>>>>>>>>>>>>>>>>>> 201, &inter) (0) >>>>>>>>>>>>>>>>>>>>>>>> a: MPI_Intercomm_create( ab_intra, 0, ac_intra, 0, >>>>>>>>>>>>>>>>>>>>>>>> 201, &inter) (0) >>>>>>>>>>>>>>>>>>>>>>>> c: MPI_Intercomm_create( MPI_COMM_WORLD, 0, intra, 0, >>>>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 4] >>>>>>>>>>>>>>>>>>>>>>>> c: MPI_Intercomm_create( MPI_COMM_WORLD, 0, intra, 0, >>>>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 5] >>>>>>>>>>>>>>>>>>>>>>>> c: MPI_Intercomm_create( MPI_COMM_WORLD, 0, intra, 0, >>>>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 6] >>>>>>>>>>>>>>>>>>>>>>>> c: MPI_Intercomm_create( MPI_COMM_WORLD, 0, intra, 0, >>>>>>>>>>>>>>>>>>>>>>>> 201, &inter) [rank 7] >>>>>>>>>>>>>>>>>>>>>>>> [hang] >>>>>>>>>>>>>>>>>>>>>>>> ----- >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> On my Mac, it hangs without printing anything: >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> ----- >>>>>>>>>>>>>>>>>>>>>>>> ❯❯❯ mpicc intercomm_create.c -o intercomm_create >>>>>>>>>>>>>>>>>>>>>>>> ❯❯❯ mpirun -np 4 intercomm_create >>>>>>>>>>>>>>>>>>>>>>>> [hang] >>>>>>>>>>>>>>>>>>>>>>>> ----- >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> On Sep 18, 2013, at 1:48 AM, George Bosilca >>>>>>>>>>>>>>>>>>>>>>>> <bosi...@icl.utk.edu> wrote: >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> Here is a quick (and definitively not the cleanest) >>>>>>>>>>>>>>>>>>>>>>>>> patch that addresses the MPI_Intercomm issue at the >>>>>>>>>>>>>>>>>>>>>>>>> MPI level. It should be applied after removal of >>>>>>>>>>>>>>>>>>>>>>>>> 29166. >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> I also added the corrected test case stressing the >>>>>>>>>>>>>>>>>>>>>>>>> corner cases by doing barriers at every inter-comm >>>>>>>>>>>>>>>>>>>>>>>>> creation and doing a clean disconnect. >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>>>> Jeff Squyres >>>>>>>>>>>>>>>>>>>>>>>> jsquy...@cisco.com >>>>>>>>>>>>>>>>>>>>>>>> For corporate legal information go to: >>>>>>>>>>>>>>>>>>>>>>>> http://www.cisco.com/web/about/doing_business/legal/cri/ >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>>>>>>> devel mailing list >>>>>>>>>>>>>>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>>>>>> devel mailing list >>>>>>>>>>>>>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>> Jeff Squyres >>>>>>>>>>>>>>>>>>>>>> jsquy...@cisco.com >>>>>>>>>>>>>>>>>>>>>> For corporate legal information go to: >>>>>>>>>>>>>>>>>>>>>> http://www.cisco.com/web/about/doing_business/legal/cri/ >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>>>>> devel mailing list >>>>>>>>>>>>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>>>> devel mailing list >>>>>>>>>>>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>>> devel mailing list >>>>>>>>>>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>> devel mailing list >>>>>>>>>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>> devel mailing list >>>>>>>>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>> devel mailing list >>>>>>>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> Jeff Squyres >>>>>>>>>>>>>>>> jsquy...@cisco.com >>>>>>>>>>>>>>>> For corporate legal information go to: >>>>>>>>>>>>>>>> http://www.cisco.com/web/about/doing_business/legal/cri/ >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>> devel mailing list >>>>>>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>> devel mailing list >>>>>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> devel mailing list >>>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> devel mailing list >>>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> devel mailing list >>>>>>>>>>> de...@open-mpi.org >>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> devel mailing list >>>>>>>>>> de...@open-mpi.org >>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> devel mailing list >>>>>>>>> de...@open-mpi.org >>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> devel mailing list >>>>>>>> de...@open-mpi.org >>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>>> >>>>>>> _______________________________________________ >>>>>>> devel mailing list >>>>>>> de...@open-mpi.org >>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>> >>>>>> _______________________________________________ >>>>>> devel mailing list >>>>>> de...@open-mpi.org >>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>> >>>>> _______________________________________________ >>>>> devel mailing list >>>>> de...@open-mpi.org >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>> >>>> _______________________________________________ >>>> devel mailing list >>>> de...@open-mpi.org >>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >>> >>> _______________________________________________ >>> devel mailing list >>> de...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/devel >> >> _______________________________________________ >> devel mailing list >> de...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/devel > > _______________________________________________ > devel mailing list > de...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/devel