I don't know - I didn't write the app file code, and I've never seen anything 
defining its behavior. So I guess you could say it is intended - or not! :-/


On Mar 1, 2012, at 2:53 PM, Jeffrey Squyres wrote:

> Actually, I should say that I discovered that if you put --prefix on each 
> line of the app context file, then the first case (running the app context 
> file) works fine; it adheres to the --prefix behavior.
> 
> Ralph: is this intended behavior?  (I don't know if I have an opinion either 
> way)
> 
> 
> On Mar 1, 2012, at 4:51 PM, Jeffrey Squyres wrote:
> 
>> I see the problem.
>> 
>> It looks like the use of the app context file is triggering different 
>> behavior, and that behavior is erasing the use of --prefix.  If I replace 
>> the app context file with a complete command line, it works and the --prefix 
>> behavior is observed.
>> 
>> Specifically:
>> 
>> $mpirunfile $mcaparams --app addmpw-hostname
>> 
>> ^^ This one seems to ignore --prefix behavior.
>> 
>> $mpirunfile $mcaparams --host svbu-mpi,svbu-mpi001 -np 2 hostname
>> $mpirunfile $mcaparams --host svbu-mpi -np 1 hostname : --host svbu-mpi001 
>> -np 1 hostname
>> 
>> ^^ These two seem to adhere to the proper --prefix behavior.
>> 
>> Ralph -- can you have a look?
>> 
>> 
>> 
>> 
>> On Mar 1, 2012, at 2:59 PM, Yiguang Yan wrote:
>> 
>>> Hi Ralph,
>>> 
>>> Thanks, here is what I did as suggested by Jeff:
>>> 
>>>> What did this command line look like? Can you provide the configure line 
>>>> as well? 
>>> 
>>> As in my previous post, the script as following:
>>> 
>>> (1) debug messages:
>>>>>> 
>>> yiguang@gulftown testdmp]$ ./test.bash
>>> [gulftown:28340] mca: base: components_open: Looking for plm components
>>> [gulftown:28340] mca: base: components_open: opening plm components
>>> [gulftown:28340] mca: base: components_open: found loaded component rsh
>>> [gulftown:28340] mca: base: components_open: component rsh has no register 
>>> function
>>> [gulftown:28340] mca: base: components_open: component rsh open function 
>>> successful
>>> [gulftown:28340] mca: base: components_open: found loaded component slurm
>>> [gulftown:28340] mca: base: components_open: component slurm has no 
>>> register function
>>> [gulftown:28340] mca: base: components_open: component slurm open function 
>>> successful
>>> [gulftown:28340] mca: base: components_open: found loaded component tm
>>> [gulftown:28340] mca: base: components_open: component tm has no register 
>>> function
>>> [gulftown:28340] mca: base: components_open: component tm open function 
>>> successful
>>> [gulftown:28340] mca:base:select: Auto-selecting plm components
>>> [gulftown:28340] mca:base:select:(  plm) Querying component [rsh]
>>> [gulftown:28340] mca:base:select:(  plm) Query of component [rsh] set 
>>> priority to 10
>>> [gulftown:28340] mca:base:select:(  plm) Querying component [slurm]
>>> [gulftown:28340] mca:base:select:(  plm) Skipping component [slurm]. Query 
>>> failed to return a module
>>> [gulftown:28340] mca:base:select:(  plm) Querying component [tm]
>>> [gulftown:28340] mca:base:select:(  plm) Skipping component [tm]. Query 
>>> failed to return a module
>>> [gulftown:28340] mca:base:select:(  plm) Selected component [rsh]
>>> [gulftown:28340] mca: base: close: component slurm closed
>>> [gulftown:28340] mca: base: close: unloading component slurm
>>> [gulftown:28340] mca: base: close: component tm closed
>>> [gulftown:28340] mca: base: close: unloading component tm
>>> [gulftown:28340] plm:base:set_hnp_name: initial bias 28340 nodename hash 
>>> 3546479048
>>> [gulftown:28340] plm:base:set_hnp_name: final jobfam 17438
>>> [gulftown:28340] [[17438,0],0] plm:base:receive start comm
>>> [gulftown:28340] [[17438,0],0] plm:rsh: setting up job [17438,1]
>>> [gulftown:28340] [[17438,0],0] plm:base:setup_job for job [17438,1]
>>> [gulftown:28340] [[17438,0],0] plm:rsh: local shell: 0 (bash)
>>> [gulftown:28340] [[17438,0],0] plm:rsh: assuming same remote shell as local 
>>> shell
>>> [gulftown:28340] [[17438,0],0] plm:rsh: remote shell: 0 (bash)
>>> [gulftown:28340] [[17438,0],0] plm:rsh: final template argv:
>>>      /usr/bin/rsh <template>  orted --daemonize -mca ess env -mca 
>>> orte_ess_jobid 1142816768 -mca 
>>> orte_ess_vpid <template> -mca orte_ess_num_procs 4 --hnp-uri 
>>> "1142816768.0;tcp://198.177.146.70:43159;tcp://10.10.10.4:43159;tcp://172.23.10.1:43159;tcp://172.33.10.1:43159"
>>>  -
>>> -mca plm_rsh_agent rsh:ssh --mca btl_openib_warn_default_gid_prefix 0 --mca 
>>> btl openib,sm,self --mca 
>>> orte_tmpdir_base /tmp --mca plm_base_verbose 100
>>> [gulftown:28340] [[17438,0],0] plm:rsh:launch daemon already exists on node 
>>> gulftown
>>> [gulftown:28340] [[17438,0],0] plm:rsh: launching on node ibnode001
>>> [gulftown:28340] [[17438,0],0] plm:rsh: recording launch of daemon 
>>> [[17438,0],1]
>>> [gulftown:28340] [[17438,0],0] plm:rsh: executing: (//usr/bin/rsh) 
>>> [/usr/bin/rsh ibnode001  orted --daemonize -mca 
>>> ess env -mca orte_ess_jobid 1142816768 -mca orte_ess_vpid 1 -mca 
>>> orte_ess_num_procs 4 --hnp-uri 
>>> "1142816768.0;tcp://198.177.146.70:43159;tcp://10.10.10.4:43159;tcp://172.23.10.1:43159;tcp://172.33.10.1:43159"
>>>  -
>>> -mca plm_rsh_agent rsh:ssh --mca btl_openib_warn_default_gid_prefix 0 --mca 
>>> btl openib,sm,self --mca 
>>> orte_tmpdir_base /tmp --mca plm_base_verbose 100]
>>> bash: orted: command not found
>>> [gulftown:28340] [[17438,0],0] plm:rsh: launching on node ibnode002
>>> [gulftown:28340] [[17438,0],0] plm:rsh: recording launch of daemon 
>>> [[17438,0],2]
>>> [gulftown:28340] [[17438,0],0] plm:rsh: executing: (//usr/bin/rsh) 
>>> [/usr/bin/rsh ibnode002  orted --daemonize -mca 
>>> ess env -mca orte_ess_jobid 1142816768 -mca orte_ess_vpid 2 -mca 
>>> orte_ess_num_procs 4 --hnp-uri 
>>> "1142816768.0;tcp://198.177.146.70:43159;tcp://10.10.10.4:43159;tcp://172.23.10.1:43159;tcp://172.33.10.1:43159"
>>>  -
>>> -mca plm_rsh_agent rsh:ssh --mca btl_openib_warn_default_gid_prefix 0 --mca 
>>> btl openib,sm,self --mca 
>>> orte_tmpdir_base /tmp --mca plm_base_verbose 100]
>>> bash: orted: command not found
>>> [gulftown:28340] [[17438,0],0] plm:rsh: launching on node ibnode003
>>> [gulftown:28340] [[17438,0],0] plm:rsh: executing: (//usr/bin/rsh) 
>>> [/usr/bin/rsh ibnode003  orted --daemonize -mca 
>>> ess env -mca orte_ess_jobid 1142816768 -mca orte_ess_vpid 3 -mca 
>>> orte_ess_num_procs 4 --hnp-uri 
>>> "1142816768.0;tcp://198.177.146.70:43159;tcp://10.10.10.4:43159;tcp://172.23.10.1:43159;tcp://172.33.10.1:43159"
>>>  -
>>> -mca plm_rsh_agent rsh:ssh --mca btl_openib_warn_default_gid_prefix 0 --mca 
>>> btl openib,sm,self --mca 
>>> orte_tmpdir_base /tmp --mca plm_base_verbose 100]
>>> [gulftown:28340] [[17438,0],0] plm:rsh: recording launch of daemon 
>>> [[17438,0],3]
>>> bash: orted: command not found
>>> [gulftown:28340] [[17438,0],0] plm:base:daemon_callback
>>> <<<
>>> 
>>> (2) test.bash script:
>>>>>> 
>>> #!/bin/sh -f
>>> #nohup
>>> #
>>> # 
>>> >-------------------------------------------------------------------------------------------<
>>> adinahome=/usr/adina/system8.8dmp
>>> mpirunfile=$adinahome/bin/mpirun
>>> #
>>> # Set envars for mpirun and orted
>>> #
>>> export PATH=$adinahome/bin:$adinahome/tools:$PATH
>>> export LD_LIBRARY_PATH=$adinahome/lib:$LD_LIBRARY_PATH
>>> #
>>> #
>>> # run DMP problem
>>> #
>>> mcaprefix="--prefix $adinahome"
>>> mcarshagent="--mca plm_rsh_agent rsh:ssh"
>>> mcatmpdir="--mca orte_tmpdir_base /tmp"
>>> mcaopenibmsg="--mca btl_openib_warn_default_gid_prefix 0"
>>> mcaenvars="-x PATH -x LD_LIBRARY_PATH"
>>> mcabtlconn="--mca btl openib,sm,self"
>>> mcaplmbase="--mca plm_base_verbose 100"
>>> 
>>> mcaparams="$mcaprefix $mcaenvars $mcarshagent $mcaopenibmsg $mcabtlconn 
>>> $mcatmpdir $mcaplmbase"
>>> 
>>> $mpirunfile $mcaparams --app addmpw-hostname
>>> <<<
>>> 
>>> (3) the contend of app file addmpw-hostname:
>>>>>> 
>>> -n 1 -host gulftown hostname
>>> -n 1 -host ibnode001 hostname
>>> -n 1 -host ibnode002 hostname
>>> -n 1 -host ibnode003 thostname
>>> <<<
>>> 
>>> Any comments?
>>> 
>>> Thanks,
>>> Yiguang
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> -- 
>> Jeff Squyres
>> jsquy...@cisco.com
>> For corporate legal information go to: 
>> http://www.cisco.com/web/about/doing_business/legal/cri/
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> -- 
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to: 
> http://www.cisco.com/web/about/doing_business/legal/cri/
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


Reply via email to