I am trying to switch to OpenMPI, and I ran into a problem : my home directory must exist on all the nodes, or orted will crash.

I have a "master" machine where I initiate the mpirun command.
Then I have a bunch of slave machines, which will also execute the MPI job. My user exists on all the machines, but the home directory is not mounted on the slaves, so it's only visible on the master node. I can log on a slave node, but don't have a home there. Of course the binary I'm running exists on all the machines (not in my home !). And the problem can be reproduced by running a shell command too, to make things simpler.

We have thousands of slave nodes and we don't want to mount the user's homedirs on all the slaves, so a fix would be really really nice.

Example :

I have 3 hosts, master, slave1, slave2. My home directory exists only on master.

If I log on master and run "mpirun -host master,slave1uname -a" I get a segfault. If I log on slave1 and run "mpirun -host slave1,slave2 uname -a", it runs fine. My home directory does not exist on either slave1 or slave2. If I log on master and run "mpirun -host master uname -a" it runs fine. I can run across several master nodes, it's fine too.

So it runs fine if my home directory exists everywhere, or if it does not exist at all. If it exists only on some nodes and not others, orted crashes. I thought it could be related to my environment but I created a new user with an empty home and it does the same thing. As soon as I create the homedir on slave1 and slave2 it works fine.




I'm using OpenMPI 1.2.2, here is the error message and the result of ompi_info.

Short version (rnd04 is the master, r137n001 is a slave node).

-bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -host rnd04,r137n001 uname -a Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64 x86_64 x86_64 GNU/Linux
[r137n001:31533] *** Process received signal ***
[r137n001:31533] Signal: Segmentation fault (11)
[r137n001:31533] Signal code: Address not mapped (1)
[r137n001:31533] Failing at address: 0x1
[r137n001:31533] [ 0] [0xffffe600]
[r137n001:31533] [ 1] /lib/tls/libc.so.6 [0xbf3bfc]
[r137n001:31533] [ 2] /lib/tls/libc.so.6(_IO_vfprintf+0xcb) [0xbf3e3b]
[r137n001:31533] [ 3] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 (opal_show_help+0x263) [0xf7f78de3] [r137n001:31533] [ 4] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 (orte_rmgr_base_check_context_cwd+0xff) [0xf7fea7ef] [r137n001:31533] [ 5] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_odls_default.so(orte_odls_default_launch_local_procs+0xe7f) [0xf7ea041f]
[r137n001:31533] [ 6] /usr/local/openmpi-1.2.2/bin/orted [0x804a1ea]
[r137n001:31533] [ 7] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_gpr_proxy.so(orte_gpr_proxy_deliver_notify_msg+0x136) [0xf7ef65c6] [r137n001:31533] [ 8] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_gpr_proxy.so(orte_gpr_proxy_notify_recv+0x108) [0xf7ef4f68] [r137n001:31533] [ 9] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 [0xf7fd9a18] [r137n001:31533] [10] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_oob_tcp.so(mca_oob_tcp_msg_recv_complete+0x24c) [0xf7f05fdc] [r137n001:31533] [11] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_oob_tcp.so [0xf7f07f61] [r137n001:31533] [12] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 (opal_event_base_loop+0x388) [0xf7f67dd8] [r137n001:31533] [13] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 (opal_event_loop+0x29) [0xf7f67fb9] [r137n001:31533] [14] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_oob_tcp.so(mca_oob_tcp_msg_wait+0x37) [0xf7f053c7] [r137n001:31533] [15] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_oob_tcp.so(mca_oob_tcp_recv+0x374) [0xf7f09a04] [r137n001:31533] [16] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 (mca_oob_recv_packed+0x4d) [0xf7fd980d] [r137n001:31533] [17] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_gpr_proxy.so(orte_gpr_proxy_exec_compound_cmd+0x137) [0xf7ef55e7] [r137n001:31533] [18] /usr/local/openmpi-1.2.2/bin/orted(main+0x99d) [0x8049d0d] [r137n001:31533] [19] /lib/tls/libc.so.6(__libc_start_main+0xd3) [0xbcee23]
[r137n001:31533] [20] /usr/local/openmpi-1.2.2/bin/orted [0x80492e1]
[r137n001:31533] *** End of error message ***
mpirun noticed that job rank 1 with PID 31533 on node r137n001 exited on signal 11 (Segmentation fault).



If I create /home/toto on r137n001, it works fine :
(as root on r137n001: "mkdir /home/toto && chown toto:users /home/toto")

-bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -host rnd04,r137n001 uname -a Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64 x86_64 x86_64 GNU/Linux Linux r137n001 2.6.9-34.ELsmp #1 SMP Fri Feb 24 16:56:28 EST 2006 x86_64 x86_64 x86_64 GNU/Linux


I tried to use ssh instead of rsh, it crashes too.

If anyone knows a way to run OpenMPI jobs in this configuration where the home directory does not exist on all the nodes, it woud really help !

Or is there a way to fix orted so that it won't crash ?


Here is the crash with -d option :

-bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -d -host rnd04,r137n001 uname -a
[rnd04:10736] connect_uni: connection not allowed
[rnd04:10736] [0,0,0] setting up session dir with
[rnd04:10736]   universe default-universe-10736
[rnd04:10736]   user toto
[rnd04:10736]   host rnd04
[rnd04:10736]   jobid 0
[rnd04:10736]   procid 0
[rnd04:10736] procdir: /tmp/openmpi-sessions-toto@rnd04_0/default- universe-10736/0/0 [rnd04:10736] jobdir: /tmp/openmpi-sessions-toto@rnd04_0/default- universe-10736/0 [rnd04:10736] unidir: /tmp/openmpi-sessions-toto@rnd04_0/default- universe-10736
[rnd04:10736] top: openmpi-sessions-toto@rnd04_0
[rnd04:10736] tmp: /tmp
[rnd04:10736] [0,0,0] contact_file /tmp/openmpi-sessions-toto@rnd04_0/ default-universe-10736/universe-setup.txt
[rnd04:10736] [0,0,0] wrote setup file
[rnd04:10736] pls:rsh: local csh: 0, local sh: 1
[rnd04:10736] pls:rsh: assuming same remote shell as local shell
[rnd04:10736] pls:rsh: remote csh: 0, remote sh: 1
[rnd04:10736] pls:rsh: final template argv:
[rnd04:10736] pls:rsh: /usr/bin/rsh <template> orted --debug -- bootproxy 1 --name <template> --num_procs 3 --vpid_start 0 --nodename <template> --universe toto@rnd04:default-universe-10736 --nsreplica "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029" -- gprreplica "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029"
[rnd04:10736] pls:rsh: launching on node rnd04
[rnd04:10736] pls:rsh: rnd04 is a LOCAL node
[rnd04:10736] pls:rsh: reset PATH: /usr/local/openmpi-1.2.2/bin:/usr/ local/bin:/bin:/usr/bin:/usr/X11R6/bin:/usr/kerberos/bin [rnd04:10736] pls:rsh: reset LD_LIBRARY_PATH: /usr/local/ openmpi-1.2.2/lib:/usr/local/openmpi-1.2.2/lib:/usr/local/ openmpi-1.2.2/lib64
[rnd04:10736] pls:rsh: changing to directory /home/toto
[rnd04:10736] pls:rsh: executing: (/usr/local/openmpi-1.2.2/bin/ orted) orted --debug --bootproxy 1 --name 0.0.1 --num_procs 3 -- vpid_start 0 --nodename rnd04 --universe toto@rnd04:default- universe-10736 --nsreplica "0.0.0;tcp://172.28.20.143:33029;tcp:// 10.3.254.105:33029" --gprreplica "0.0.0;tcp:// 172.28.20.143:33029;tcp://10.3.254.105:33029" --set-sid [HOSTNAME=rnd04 SHELL=/bin/bash TERM=xterm-color HISTSIZE=1000 USER=toto LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/lib:/usr/local/ openmpi-1.2.2/lib:/usr/local/openmpi-1.2.2/lib64 LS_COLORS=no=00:fi=00:di=01;34:ln=01;36:pi=40;33:so=01;35:bd=40;33;01:cd =40;33;01:or=01;05;37;41:mi=01;05;37;41:ex=01;32:*.cmd=01;32:*.exe=01;32 :*.com=01;32:*.btm=01;32:*.bat=01;32:*.sh=01;32:*.csh=01;32:*.tar=01;31: *.tgz=01;31:*.arj=01;31:*.taz=01;31:*.lzh=01;31:*.zip=01;31:*.z=01;31:*. Z=01;31:*.gz=01;31:*.bz2=01;31:*.bz=01;31:*.tz=01;31:*.rpm=01;31:*.cpio= 01;31:*.jpg=01;35:*.gif=01;35:*.bmp=01;35:*.xbm=01;35:*.xpm=01;35:*.png= 01;35:*.tif=01;35: KDEDIR=/usr MAIL=/var/spool/mail/toto PATH=/usr/ local/openmpi-1.2.2/bin:/usr/local/bin:/bin:/usr/bin:/usr/X11R6/bin:/ usr/kerberos/bin INPUTRC=/etc/inputrc PWD=/home/toto LANG=en_US.UTF-8 SSH_ASKPASS=/usr/libexec/openssh/gnome-ssh-askpass SHLVL=1 HOME=/home/ toto LOGNAME=toto LESSOPEN=|/usr/bin/lesspipe.sh %s G_BROKEN_FILENAMES=1 _=/usr/local/openmpi-1.2.2/bin/mpirun OMPI_MCA_orte_debug=1 OMPI_MCA_seed=0]
[rnd04:10736] pls:rsh: launching on node r137n001
[rnd04:10736] pls:rsh: r137n001 is a REMOTE node
[rnd04:10736] pls:rsh: executing: (//usr/bin/rsh) /usr/bin/rsh r137n001 PATH=/usr/local/openmpi-1.2.2/bin:$PATH ; export PATH ; LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/lib:$LD_LIBRARY_PATH ; export LD_LIBRARY_PATH ; /usr/local/openmpi-1.2.2/bin/orted --debug -- bootproxy 1 --name 0.0.2 --num_procs 3 --vpid_start 0 --nodename r137n001 --universe toto@rnd04:default-universe-10736 --nsreplica "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029" -- gprreplica "0.0.0;tcp://172.28.20.143:33029;tcp:// 10.3.254.105:33029" [HOSTNAME=rnd04 SHELL=/bin/bash TERM=xterm-color HISTSIZE=1000 USER=toto LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/lib:/ usr/local/openmpi-1.2.2/lib64 LS_COLORS=no=00:fi=00:di=01;34:ln=01;36:pi=40;33:so=01;35:bd=40;33;01:cd =40;33;01:or=01;05;37;41:mi=01;05;37;41:ex=01;32:*.cmd=01;32:*.exe=01;32 :*.com=01;32:*.btm=01;32:*.bat=01;32:*.sh=01;32:*.csh=01;32:*.tar=01;31: *.tgz=01;31:*.arj=01;31:*.taz=01;31:*.lzh=01;31:*.zip=01;31:*.z=01;31:*. Z=01;31:*.gz=01;31:*.bz2=01;31:*.bz=01;31:*.tz=01;31:*.rpm=01;31:*.cpio= 01;31:*.jpg=01;35:*.gif=01;35:*.bmp=01;35:*.xbm=01;35:*.xpm=01;35:*.png= 01;35:*.tif=01;35: KDEDIR=/usr MAIL=/var/spool/mail/toto PATH=/usr/ local/bin:/bin:/usr/bin:/usr/X11R6/bin:/usr/kerberos/bin INPUTRC=/etc/ inputrc PWD=/home/toto LANG=en_US.UTF-8 SSH_ASKPASS=/usr/libexec/ openssh/gnome-ssh-askpass SHLVL=1 HOME=/home/toto LOGNAME=toto LESSOPEN=|/usr/bin/lesspipe.sh %s G_BROKEN_FILENAMES=1 _=/usr/local/ openmpi-1.2.2/bin/mpirun OMPI_MCA_orte_debug=1 OMPI_MCA_seed=0]
[rnd04:10737] [0,0,1] setting up session dir with
[rnd04:10737]   universe default-universe-10736
[rnd04:10737]   user toto
[rnd04:10737]   host rnd04
[rnd04:10737]   jobid 0
[rnd04:10737]   procid 1
[rnd04:10737] procdir: /tmp/openmpi-sessions-toto@rnd04_0/default- universe-10736/0/1 [rnd04:10737] jobdir: /tmp/openmpi-sessions-toto@rnd04_0/default- universe-10736/0 [rnd04:10737] unidir: /tmp/openmpi-sessions-toto@rnd04_0/default- universe-10736
[rnd04:10737] top: openmpi-sessions-toto@rnd04_0
[rnd04:10737] tmp: /tmp
[r137n001:31527] [0,0,2] setting up session dir with
[r137n001:31527]        universe default-universe-10736
[r137n001:31527]        user toto
[r137n001:31527]        host r137n001
[r137n001:31527]        jobid 0
[r137n001:31527]        procid 2
[r137n001:31527] procdir: /tmp/openmpi-sessions-toto@r137n001_0/ default-universe-10736/0/2 [r137n001:31527] jobdir: /tmp/openmpi-sessions-toto@r137n001_0/ default-universe-10736/0 [r137n001:31527] unidir: /tmp/openmpi-sessions-toto@r137n001_0/ default-universe-10736
[r137n001:31527] top: openmpi-sessions-toto@r137n001_0
[r137n001:31527] tmp: /tmp
Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64 x86_64 x86_64 GNU/Linux
[rnd04:10737] sess_dir_finalize: proc session dir not empty - leaving
[r137n001:31528] *** Process received signal ***
[r137n001:31528] Signal: Segmentation fault (11)
[r137n001:31528] Signal code: Address not mapped (1)
[r137n001:31528] Failing at address: 0x1
[r137n001:31528] [ 0] [0xffffe600]
[r137n001:31528] [ 1] /lib/tls/libc.so.6 [0xbf3bfc]
[r137n001:31528] [ 2] /lib/tls/libc.so.6(_IO_vfprintf+0xcb) [0xbf3e3b]
[r137n001:31528] [ 3] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 (opal_show_help+0x263) [0xf7f78de3] [r137n001:31528] [ 4] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 (orte_rmgr_base_check_context_cwd+0xff) [0xf7fea7ef] [r137n001:31528] [ 5] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_odls_default.so(orte_odls_default_launch_local_procs+0xe7f) [0xf7ea041f]
[r137n001:31528] [ 6] /usr/local/openmpi-1.2.2/bin/orted [0x804a1ea]
[r137n001:31528] [ 7] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_gpr_proxy.so(orte_gpr_proxy_deliver_notify_msg+0x136) [0xf7ef65c6] [r137n001:31528] [ 8] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_gpr_proxy.so(orte_gpr_proxy_notify_recv+0x108) [0xf7ef4f68] [r137n001:31528] [ 9] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 [0xf7fd9a18] [r137n001:31528] [10] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_oob_tcp.so(mca_oob_tcp_msg_recv_complete+0x24c) [0xf7f05fdc] [r137n001:31528] [11] /usr/local/openmpi-1.2.2/lib/openmpi/ mca_oob_tcp.so [0xf7f07f61] [r137n001:31528] [12] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 (opal_event_base_loop+0x388) [0xf7f67dd8] [r137n001:31528] [13] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 (opal_event_loop+0x29) [0xf7f67fb9] [r137n001:31528] [14] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 (opal_progress+0xbe) [0xf7f6123e] [r137n001:31528] [15] /usr/local/openmpi-1.2.2/bin/orted(main+0xd74) [0x804a0e4] [r137n001:31528] [16] /lib/tls/libc.so.6(__libc_start_main+0xd3) [0xbcee23]
[r137n001:31528] [17] /usr/local/openmpi-1.2.2/bin/orted [0x80492e1]
[r137n001:31528] *** End of error message ***
[r137n001:31527] sess_dir_finalize: proc session dir not empty - leaving
[rnd04:10736] spawn: in job_state_callback(jobid = 1, state = 0x80)
mpirun noticed that job rank 1 with PID 31528 on node r137n001 exited on signal 11 (Segmentation fault).
[rnd04:10737] sess_dir_finalize: job session dir not empty - leaving
[r137n001:31527] sess_dir_finalize: job session dir not empty - leaving
[rnd04:10737] sess_dir_finalize: proc session dir not empty - leaving
[rnd04:10736] sess_dir_finalize: proc session dir not empty - leaving
-bash-3.00$ [r137n001:31527] sess_dir_finalize: proc session dir not empty - leaving




-bash-3.00$ /usr/local/openmpi-1.2.2/bin/ompi_info --all
                Open MPI: 1.2.2
   Open MPI SVN revision: r14613
                Open RTE: 1.2.2
   Open RTE SVN revision: r14613
                    OPAL: 1.2.2
       OPAL SVN revision: r14613
MCA backtrace: execinfo (MCA v1.0, API v1.0, Component v1.2.2) MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component v1.2.2)
           MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.2.2)
MCA maffinity: first_use (MCA v1.0, API v1.0, Component v1.2.2)
               MCA timer: linux (MCA v1.0, API v1.0, Component v1.2.2)
         MCA installdirs: env (MCA v1.0, API v1.0, Component v1.2.2)
         MCA installdirs: config (MCA v1.0, API v1.0, Component v1.2.2)
           MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
           MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
                MCA coll: basic (MCA v1.0, API v1.0, Component v1.2.2)
                MCA coll: self (MCA v1.0, API v1.0, Component v1.2.2)
                MCA coll: sm (MCA v1.0, API v1.0, Component v1.2.2)
                MCA coll: tuned (MCA v1.0, API v1.0, Component v1.2.2)
                  MCA io: romio (MCA v1.0, API v1.0, Component v1.2.2)
               MCA mpool: rdma (MCA v1.0, API v1.0, Component v1.2.2)
               MCA mpool: sm (MCA v1.0, API v1.0, Component v1.2.2)
                 MCA pml: cm (MCA v1.0, API v1.0, Component v1.2.2)
                 MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.2.2)
                 MCA bml: r2 (MCA v1.0, API v1.0, Component v1.2.2)
              MCA rcache: vma (MCA v1.0, API v1.0, Component v1.2.2)
                 MCA btl: self (MCA v1.0, API v1.0.1, Component v1.2.2)
                 MCA btl: sm (MCA v1.0, API v1.0.1, Component v1.2.2)
                 MCA btl: tcp (MCA v1.0, API v1.0.1, Component v1.0)
                MCA topo: unity (MCA v1.0, API v1.0, Component v1.2.2)
                 MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.2.2)
              MCA errmgr: hnp (MCA v1.0, API v1.3, Component v1.2.2)
              MCA errmgr: orted (MCA v1.0, API v1.3, Component v1.2.2)
              MCA errmgr: proxy (MCA v1.0, API v1.3, Component v1.2.2)
                 MCA gpr: null (MCA v1.0, API v1.0, Component v1.2.2)
                 MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.2.2)
MCA gpr: replica (MCA v1.0, API v1.0, Component v1.2.2)
                 MCA iof: proxy (MCA v1.0, API v1.0, Component v1.2.2)
                 MCA iof: svc (MCA v1.0, API v1.0, Component v1.2.2)
                  MCA ns: proxy (MCA v1.0, API v2.0, Component v1.2.2)
MCA ns: replica (MCA v1.0, API v2.0, Component v1.2.2)
                 MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
MCA ras: dash_host (MCA v1.0, API v1.3, Component v1.2.2) MCA ras: gridengine (MCA v1.0, API v1.3, Component v1.2.2) MCA ras: localhost (MCA v1.0, API v1.3, Component v1.2.2)
                 MCA ras: slurm (MCA v1.0, API v1.3, Component v1.2.2)
MCA rds: hostfile (MCA v1.0, API v1.3, Component v1.2.2)
                 MCA rds: proxy (MCA v1.0, API v1.3, Component v1.2.2)
MCA rds: resfile (MCA v1.0, API v1.3, Component v1.2.2) MCA rmaps: round_robin (MCA v1.0, API v1.3, Component v1.2.2)
                MCA rmgr: proxy (MCA v1.0, API v2.0, Component v1.2.2)
                MCA rmgr: urm (MCA v1.0, API v2.0, Component v1.2.2)
                 MCA rml: oob (MCA v1.0, API v1.0, Component v1.2.2)
MCA pls: gridengine (MCA v1.0, API v1.3, Component v1.2.2)
                 MCA pls: proxy (MCA v1.0, API v1.3, Component v1.2.2)
                 MCA pls: rsh (MCA v1.0, API v1.3, Component v1.2.2)
                 MCA pls: slurm (MCA v1.0, API v1.3, Component v1.2.2)
                 MCA sds: env (MCA v1.0, API v1.0, Component v1.2.2)
                 MCA sds: pipe (MCA v1.0, API v1.0, Component v1.2.2)
                 MCA sds: seed (MCA v1.0, API v1.0, Component v1.2.2)
MCA sds: singleton (MCA v1.0, API v1.0, Component v1.2.2)
                 MCA sds: slurm (MCA v1.0, API v1.0, Component v1.2.2)
                  Prefix: /usr/local/openmpi-1.2.2
                  Bindir: /usr/local/openmpi-1.2.2/bin
                  Libdir: /usr/local/openmpi-1.2.2/lib
                  Incdir: /usr/local/openmpi-1.2.2/include
               Pkglibdir: /usr/local/openmpi-1.2.2/lib/openmpi
              Sysconfdir: /usr/local/openmpi-1.2.2/etc
Configured architecture: x86_64-unknown-linux-gnu
           Configured by: root
           Configured on: Tue Jun  5 14:32:20 CDT 2007
          Configure host: qpac171
                Built by: root
                Built on: Tue Jun  5 14:39:38 CDT 2007
              Built host: qpac171
              C bindings: yes
            C++ bindings: yes
      Fortran77 bindings: yes (all)
      Fortran90 bindings: no
Fortran90 bindings size: na
              C compiler: gcc
     C compiler absolute: /usr/bin/gcc
             C char size: 1
             C bool size: 1
            C short size: 2
              C int size: 4
             C long size: 4
            C float size: 4
           C double size: 8
          C pointer size: 4
            C char align: 1
            C bool align: 1
             C int align: 4
           C float align: 4
          C double align: 4
            C++ compiler: g++
   C++ compiler absolute: /usr/bin/g++
      Fortran77 compiler: g77
  Fortran77 compiler abs: /usr/bin/g77
      Fortran90 compiler: none
  Fortran90 compiler abs: none
       Fort integer size: 4
       Fort logical size: 4
Fort logical value true: 1
      Fort have integer1: yes
      Fort have integer2: yes
      Fort have integer4: yes
      Fort have integer8: yes
     Fort have integer16: no
         Fort have real4: yes
         Fort have real8: yes
        Fort have real16: no
      Fort have complex8: yes
     Fort have complex16: yes
     Fort have complex32: no
      Fort integer1 size: 1
      Fort integer2 size: 2
      Fort integer4 size: 4
      Fort integer8 size: 8
     Fort integer16 size: -1
          Fort real size: 4
         Fort real4 size: 4
         Fort real8 size: 8
        Fort real16 size: -1
      Fort dbl prec size: 4
          Fort cplx size: 4
      Fort dbl cplx size: 4
         Fort cplx8 size: 8
        Fort cplx16 size: 16
        Fort cplx32 size: -1
      Fort integer align: 4
     Fort integer1 align: 1
     Fort integer2 align: 2
     Fort integer4 align: 4
     Fort integer8 align: 8
    Fort integer16 align: -1
         Fort real align: 4
        Fort real4 align: 4
        Fort real8 align: 8
       Fort real16 align: -1
     Fort dbl prec align: 4
         Fort cplx align: 4
     Fort dbl cplx align: 4
        Fort cplx8 align: 4
       Fort cplx16 align: 8
       Fort cplx32 align: -1
             C profiling: yes
           C++ profiling: yes
     Fortran77 profiling: yes
     Fortran90 profiling: no
          C++ exceptions: no
          Thread support: posix (mpi: no, progress: no)
Build CFLAGS: -O3 -DNDEBUG -m32 -finline-functions -fno- strict-aliasing -pthread
          Build CXXFLAGS: -O3 -DNDEBUG -m32 -finline-functions -pthread
            Build FFLAGS: -m32
           Build FCFLAGS: -m32
           Build LDFLAGS: -export-dynamic
              Build LIBS: -lnsl -lutil  -lm
    Wrapper extra CFLAGS: -pthread -m32
  Wrapper extra CXXFLAGS: -pthread -m32
    Wrapper extra FFLAGS: -pthread -m32
   Wrapper extra FCFLAGS: -pthread -m32
   Wrapper extra LDFLAGS:
Wrapper extra LIBS: -ldl -Wl,--export-dynamic -lnsl -lutil -lm -ldl
  Internal debug support: no
     MPI parameter check: runtime
Memory profiling support: no
Memory debugging support: no
         libltdl support: yes
   Heterogeneous support: yes
mpirun default --prefix: no
MCA mca: parameter "mca_param_files" (current value: "/home/toto/.openmpi/mca-params.conf:/usr/local/openmpi-1.2.2/ etc/openmpi-mca-params.conf") Path for MCA configuration files containing default parameter values MCA mca: parameter "mca_component_path" (current value: "/usr/local/openmpi-1.2.2/lib/openmpi:/home/toto/.openmpi/ components") Path where to look for Open MPI and ORTE components MCA mca: parameter "mca_verbose" (current value: <none>)
                          Top-level verbosity parameter
MCA mca: parameter "mca_component_show_load_errors" (current value: "1") Whether to show errors for components that failed to load or not MCA mca: parameter "mca_component_disable_dlopen" (current value: "0") Whether to attempt to disable opening dynamic components or not MCA mpi: parameter "mpi_param_check" (current value: "1") Whether you want MPI API parameters checked at run-time or not. Possible values are 0 (no checking) and 1 (perform checking at run-time) MCA mpi: parameter "mpi_yield_when_idle" (current value: "0") Yield the processor when waiting for MPI communication (for MPI processes, will default to 1 when oversubscribing nodes) MCA mpi: parameter "mpi_event_tick_rate" (current value: "-1") How often to progress TCP communications (0 = never, otherwise specified in microseconds) MCA mpi: parameter "mpi_show_handle_leaks" (current value: "0") Whether MPI_FINALIZE shows all MPI handles that were not freed or not MCA mpi: parameter "mpi_no_free_handles" (current value: "0") Whether to actually free MPI objects when their handles are freed MCA mpi: parameter "mpi_show_mca_params" (current value: "0") Whether to show all MCA parameter value during MPI_INIT or not (good for reproducability of MPI jobs) MCA mpi: parameter "mpi_show_mca_params_file" (current value: <none>) If mpi_show_mca_params is true, setting this string to a valid filename tells Open MPI to dump all the MCA parameter values into a file suitable for reading via the mca_param_files parameter (good for reproducability of MPI jobs) MCA mpi: parameter "mpi_paffinity_alone" (current value: "0") If nonzero, assume that this job is the only (set of) process(es) running on each node and bind processes to processors, starting with processor ID 0 MCA mpi: parameter "mpi_keep_peer_hostnames" (current value: "1") If nonzero, save the string hostnames of all MPI peer processes (mostly for error / debugging output messages). This can add quite a bit of memory usage to each MPI process. MCA mpi: parameter "mpi_abort_delay" (current value: "0") If nonzero, print out an identifying message when MPI_ABORT is invoked (hostname, PID of the process that called MPI_ABORT) and delay for that many seconds before exiting (a negative delay value means to never abort). This allows attaching of a debugger before quitting the job. MCA mpi: parameter "mpi_abort_print_stack" (current value: "0") If nonzero, print out a stack trace when MPI_ABORT is invoked MCA mpi: parameter "mpi_preconnect_all" (current value: "0") Whether to force MPI processes to create connections / warmup with *all* peers during MPI_INIT (vs. making connections lazily -- upon the first MPI traffic between each process peer pair) MCA mpi: parameter "mpi_preconnect_oob" (current value: "0") Whether to force MPI processes to fully wire-up the OOB system between MPI processes. MCA mpi: parameter "mpi_leave_pinned" (current value: "0") Whether to use the "leave pinned" protocol or not. Enabling this setting can help bandwidth performance when repeatedly sending and receiving large messages with the same buffers over RDMA-based networks. MCA mpi: parameter "mpi_leave_pinned_pipeline" (current value: "0") Whether to use the "leave pinned pipeline" protocol or not.
                MCA orte: parameter "orte_debug" (current value: "0")
                          Top-level ORTE debug switch
MCA orte: parameter "orte_no_daemonize" (current value: "0") Whether to properly daemonize the ORTE daemons or not MCA orte: parameter "orte_base_user_debugger" (current value: "totalview @mpirun@ -a @mpirun_args@ : fxp @mpirun@ -a @mpirun_args@") Sequence of user-level debuggers to search for in orterun MCA orte: parameter "orte_abort_timeout" (current value: "10") Time to wait [in seconds] before giving up on aborting an ORTE operation
                MCA orte: parameter "orte_timing" (current value: "0")
Request that critical timing loops be measured MCA opal: parameter "opal_signal" (current value: "6,7,8,11") If a signal is received, display the stack trace frame
           MCA backtrace: parameter "backtrace" (current value: <none>)
Default selection set of components for the backtrace framework (<none> means "use all components that can be found") MCA backtrace: parameter "backtrace_base_verbose" (current value: "0") Verbosity level for the backtrace framework (0 = no verbosity) MCA backtrace: parameter "backtrace_execinfo_priority" (current value: "0")
              MCA memory: parameter "memory" (current value: <none>)
Default selection set of components for the memory framework (<none> means "use all components that can be found") MCA memory: parameter "memory_base_verbose" (current value: "0") Verbosity level for the memory framework (0 = no verbosity) MCA memory: parameter "memory_ptmalloc2_priority" (current value: "0")
           MCA paffinity: parameter "paffinity" (current value: <none>)
Default selection set of components for the paffinity framework (<none> means "use all components that can be found") MCA paffinity: parameter "paffinity_linux_priority" (current value: "10")
                          Priority of the linux paffinity component
MCA paffinity: information "paffinity_linux_have_cpu_set_t" (value: "1") Whether this component was compiled on a system with the type cpu_set_t or not (1 = yes, 0 = no) MCA paffinity: information "paffinity_linux_CPU_ZERO_ok" (value: "1") Whether this component was compiled on a system where CPU_ZERO() is functional or broken (1 = functional, 0 = broken/not available) MCA paffinity: information "paffinity_linux_sched_setaffinity_num_params" (value: "3") The number of parameters that sched_set_affinity() takes on the machine where this component was compiled
           MCA maffinity: parameter "maffinity" (current value: <none>)
Default selection set of components for the maffinity framework (<none> means "use all components that can be found") MCA maffinity: parameter "maffinity_first_use_priority" (current value: "10")
                          Priority of the first_use maffinity component
               MCA timer: parameter "timer" (current value: <none>)
Default selection set of components for the timer framework (<none> means "use all components that can be found") MCA timer: parameter "timer_base_verbose" (current value: "0") Verbosity level for the timer framework (0 = no verbosity) MCA timer: parameter "timer_linux_priority" (current value: "0")
           MCA allocator: parameter "allocator" (current value: <none>)
Default selection set of components for the allocator framework (<none> means "use all components that can be found") MCA allocator: parameter "allocator_base_verbose" (current value: "0") Verbosity level for the allocator framework (0 = no verbosity) MCA allocator: parameter "allocator_basic_priority" (current value: "0") MCA allocator: parameter "allocator_bucket_num_buckets" (current value: "30") MCA allocator: parameter "allocator_bucket_priority" (current value: "0")
                MCA coll: parameter "coll" (current value: <none>)
Default selection set of components for the coll framework (<none> means "use all components that can be found") MCA coll: parameter "coll_base_verbose" (current value: "0") Verbosity level for the coll framework (0 = no verbosity) MCA coll: parameter "coll_basic_priority" (current value: "10")
                          Priority of the basic coll component
MCA coll: parameter "coll_basic_crossover" (current value: "4") Minimum number of processes in a communicator before using the logarithmic algorithms MCA coll: parameter "coll_self_priority" (current value: "75") MCA coll: parameter "coll_sm_priority" (current value: "0")
                          Priority of the sm coll component
MCA coll: parameter "coll_sm_control_size" (current value: "4096") Length of the control data -- should usually be either the length of a cache line on most SMPs, or the size of a page on machines that support direct memory affinity page placement (in bytes) MCA coll: parameter "coll_sm_bootstrap_filename" (current value: "shared_mem_sm_bootstrap") Filename (in the Open MPI session directory) of the coll sm component bootstrap rendezvous mmap file MCA coll: parameter "coll_sm_bootstrap_num_segments" (current value: "8")
                          Number of segments in the bootstrap file
MCA coll: parameter "coll_sm_fragment_size" (current value: "8192") Fragment size (in bytes) used for passing data through shared memory (will be rounded up to the nearest control_size size) MCA coll: parameter "coll_sm_mpool" (current value: "sm")
                          Name of the mpool component to use
MCA coll: parameter "coll_sm_comm_in_use_flags" (current value: "2") Number of "in use" flags, used to mark a message passing area segment as currently being used or not (must be >= 2 and <= comm_num_segments) MCA coll: parameter "coll_sm_comm_num_segments" (current value: "8") Number of segments in each communicator's shared memory message passing area (must be >= 2, and must be a multiple of comm_in_use_flags) MCA coll: parameter "coll_sm_tree_degree" (current value: "4") Degree of the tree for tree-based operations (must be => 1 and <= min(control_size, 255)) MCA coll: information "coll_sm_shared_mem_used_bootstrap" (value: "160") Amount of shared memory used in the shared memory bootstrap area (in bytes) MCA coll: parameter "coll_sm_info_num_procs" (current value: "4") Number of processes to use for the calculation of the shared_mem_size MCA information parameter (must be => 2) MCA coll: information "coll_sm_shared_mem_used_data" (value: "548864") Amount of shared memory used in the shared memory data area for info_num_procs processes (in bytes) MCA coll: parameter "coll_tuned_priority" (current value: "30")
                          Priority of the tuned coll component
MCA coll: parameter "coll_tuned_pre_allocate_memory_comm_size_limit" (current value: "32768") Size of communicator were we stop pre- allocating memory for the fixed internal buffer used for message requests etc that is hung off the communicator data segment. I.e. if you have a 100'000 nodes you might not want to pre-allocate 200'000 request handle slots per communicator instance! MCA coll: parameter "coll_tuned_init_tree_fanout" (current value: "4") Inital fanout used in the tree topologies for each communicator. This is only an initial guess, if a tuned collective needs a different fanout for an operation, it build it dynamically. This parameter is only for the first guess and might
                          save a little time
MCA coll: parameter "coll_tuned_init_chain_fanout" (current value: "4") Inital fanout used in the chain (fanout followed by pipeline) topologies for each communicator. This is only an initial guess, if a tuned collective needs a different fanout for an operation, it build it dynamically. This parameter is only for the first guess and might save a little time MCA coll: parameter "coll_tuned_use_dynamic_rules" (current value: "0") Switch used to decide if we use static (compiled/if statements) or dynamic (built at runtime) decision function rules MCA io: parameter "io_base_freelist_initial_size" (current value: "16")
                          Initial MPI-2 IO request freelist size
MCA io: parameter "io_base_freelist_max_size" (current value: "64")
                          Max size of the MPI-2 IO request freelist
MCA io: parameter "io_base_freelist_increment" (current value: "16") Increment size of the MPI-2 IO request freelist
                  MCA io: parameter "io" (current value: <none>)
Default selection set of components for the io framework (<none> means "use all components that can be found") MCA io: parameter "io_base_verbose" (current value: "0") Verbosity level for the io framework (0 = no verbosity) MCA io: parameter "io_romio_priority" (current value: "10")
                          Priority of the io romio component
MCA io: parameter "io_romio_delete_priority" (current value: "10")
                          Delete priority of the io romio component
MCA io: parameter "io_romio_enable_parallel_optimizations" (current value: "0") Enable set of Open MPI-added options to improve collective file i/o performance
               MCA mpool: parameter "mpool" (current value: <none>)
Default selection set of components for the mpool framework (<none> means "use all components that can be found") MCA mpool: parameter "mpool_base_verbose" (current value: "0") Verbosity level for the mpool framework (0 = no verbosity) MCA mpool: parameter "mpool_rdma_rcache_name" (current value: "vma") The name of the registration cache the mpool should use MCA mpool: parameter "mpool_rdma_rcache_size_limit" (current value: "0") the maximum size of registration cache in bytes. 0 is unlimited (default 0) MCA mpool: parameter "mpool_rdma_print_stats" (current value: "0") print pool usage statistics at the end of the run MCA mpool: parameter "mpool_rdma_priority" (current value: "0") MCA mpool: parameter "mpool_sm_allocator" (current value: "bucket") Name of allocator component to use with sm mpool MCA mpool: parameter "mpool_sm_max_size" (current value: "536870912") Maximum size of the sm mpool shared memory file MCA mpool: parameter "mpool_sm_min_size" (current value: "134217728") Minimum size of the sm mpool shared memory file MCA mpool: parameter "mpool_sm_per_peer_size" (current value: "33554432") Size (in bytes) to allocate per local peer in the sm mpool shared memory file, bounded by min_size and max_size MCA mpool: parameter "mpool_sm_priority" (current value: "0") MCA mpool: parameter "mpool_base_use_mem_hooks" (current value: "0") use memory hooks for deregistering freed memory MCA mpool: parameter "mpool_use_mem_hooks" (current value: "0")
                          (deprecated, use mpool_base_use_mem_hooks)
MCA mpool: parameter "mpool_base_disable_sbrk" (current value: "0") use mallopt to override calling sbrk (doesn't return memory to OS!) MCA mpool: parameter "mpool_disable_sbrk" (current value: "0")
                          (deprecated, use mca_mpool_base_disable_sbrk)
                 MCA pml: parameter "pml" (current value: <none>)
Default selection set of components for the pml framework (<none> means "use all components that can be found") MCA pml: parameter "pml_base_verbose" (current value: "0") Verbosity level for the pml framework (0 = no verbosity) MCA pml: parameter "pml_cm_free_list_num" (current value: "4")
                          Initial size of request free lists
MCA pml: parameter "pml_cm_free_list_max" (current value: "-1")
                          Maximum size of request free lists
MCA pml: parameter "pml_cm_free_list_inc" (current value: "64") Number of elements to add when growing request free lists MCA pml: parameter "pml_cm_priority" (current value: "30")
                          CM PML selection priority
MCA pml: parameter "pml_ob1_free_list_num" (current value: "4") MCA pml: parameter "pml_ob1_free_list_max" (current value: "-1") MCA pml: parameter "pml_ob1_free_list_inc" (current value: "64") MCA pml: parameter "pml_ob1_priority" (current value: "20") MCA pml: parameter "pml_ob1_eager_limit" (current value: "131072") MCA pml: parameter "pml_ob1_send_pipeline_depth" (current value: "3") MCA pml: parameter "pml_ob1_recv_pipeline_depth" (current value: "4")
                 MCA bml: parameter "bml" (current value: <none>)
Default selection set of components for the bml framework (<none> means "use all components that can be found") MCA bml: parameter "bml_base_verbose" (current value: "0") Verbosity level for the bml framework (0 = no verbosity) MCA bml: parameter "bml_r2_show_unreach_errors" (current value: "1")
                          Show error message when procs are unreachable
MCA bml: parameter "bml_r2_priority" (current value: "0")
              MCA rcache: parameter "rcache" (current value: <none>)
Default selection set of components for the rcache framework (<none> means "use all components that can be found") MCA rcache: parameter "rcache_base_verbose" (current value: "0") Verbosity level for the rcache framework (0 = no verbosity) MCA rcache: parameter "rcache_vma_priority" (current value: "0") MCA btl: parameter "btl_base_debug" (current value: "0") If btl_base_debug is 1 standard debug is output, if > 1 verbose debug is output
                 MCA btl: parameter "btl" (current value: <none>)
Default selection set of components for the btl framework (<none> means "use all components that can be found") MCA btl: parameter "btl_base_verbose" (current value: "0") Verbosity level for the btl framework (0 = no verbosity) MCA btl: parameter "btl_self_free_list_num" (current value: "0")
                          Number of fragments by default
MCA btl: parameter "btl_self_free_list_max" (current value: "-1")
                          Maximum number of fragments
MCA btl: parameter "btl_self_free_list_inc" (current value: "32")
                          Increment by this number of fragments
MCA btl: parameter "btl_self_eager_limit" (current value: "131072") Eager size fragmeng (before the rendez- vous ptotocol) MCA btl: parameter "btl_self_min_send_size" (current value: "262144")
                          Minimum fragment size after the rendez-vous
MCA btl: parameter "btl_self_max_send_size" (current value: "262144")
                          Maximum fragment size after the rendez-vous
MCA btl: parameter "btl_self_min_rdma_size" (current value: "2147483647")
                          Maximum fragment size for the RDMA transfer
MCA btl: parameter "btl_self_max_rdma_size" (current value: "2147483647")
                          Maximum fragment size for the RDMA transfer
MCA btl: parameter "btl_self_exclusivity" (current value: "65536")
                          Device exclusivity
MCA btl: parameter "btl_self_flags" (current value: "10")
                          Active behavior flags
MCA btl: parameter "btl_self_priority" (current value: "0") MCA btl: parameter "btl_sm_free_list_num" (current value: "8") MCA btl: parameter "btl_sm_free_list_max" (current value: "-1") MCA btl: parameter "btl_sm_free_list_inc" (current value: "64") MCA btl: parameter "btl_sm_exclusivity" (current value: "65535") MCA btl: parameter "btl_sm_latency" (current value: "100") MCA btl: parameter "btl_sm_max_procs" (current value: "-1") MCA btl: parameter "btl_sm_sm_extra_procs" (current value: "2") MCA btl: parameter "btl_sm_mpool" (current value: "sm") MCA btl: parameter "btl_sm_eager_limit" (current value: "4096") MCA btl: parameter "btl_sm_max_frag_size" (current value: "32768") MCA btl: parameter "btl_sm_size_of_cb_queue" (current value: "128") MCA btl: parameter "btl_sm_cb_lazy_free_freq" (current value: "120") MCA btl: parameter "btl_sm_priority" (current value: "0") MCA btl: parameter "btl_tcp_if_include" (current value: <none>) MCA btl: parameter "btl_tcp_if_exclude" (current value: "lo") MCA btl: parameter "btl_tcp_free_list_num" (current value: "8") MCA btl: parameter "btl_tcp_free_list_max" (current value: "-1") MCA btl: parameter "btl_tcp_free_list_inc" (current value: "32") MCA btl: parameter "btl_tcp_sndbuf" (current value: "131072") MCA btl: parameter "btl_tcp_rcvbuf" (current value: "131072") MCA btl: parameter "btl_tcp_endpoint_cache" (current value: "30720") MCA btl: parameter "btl_tcp_exclusivity" (current value: "0") MCA btl: parameter "btl_tcp_eager_limit" (current value: "65536") MCA btl: parameter "btl_tcp_min_send_size" (current value: "65536") MCA btl: parameter "btl_tcp_max_send_size" (current value: "131072") MCA btl: parameter "btl_tcp_min_rdma_size" (current value: "131072") MCA btl: parameter "btl_tcp_max_rdma_size" (current value: "2147483647") MCA btl: parameter "btl_tcp_flags" (current value: "122") MCA btl: parameter "btl_tcp_priority" (current value: "0") MCA btl: parameter "btl_base_include" (current value: <none>) MCA btl: parameter "btl_base_exclude" (current value: <none>) MCA btl: parameter "btl_base_warn_component_unused" (current value: "1") This parameter is used to turn on warning messages when certain NICs are not used
                 MCA mtl: parameter "mtl" (current value: <none>)
Default selection set of components for the mtl framework (<none> means "use all components that can be found") MCA mtl: parameter "mtl_base_verbose" (current value: "0") Verbosity level for the mtl framework (0 = no verbosity)
                MCA topo: parameter "topo" (current value: <none>)
Default selection set of components for the topo framework (<none> means "use all components that can be found") MCA topo: parameter "topo_base_verbose" (current value: "0") Verbosity level for the topo framework (0 = no verbosity)
                 MCA osc: parameter "osc" (current value: <none>)
Default selection set of components for the osc framework (<none> means "use all components that can be found") MCA osc: parameter "osc_base_verbose" (current value: "0") Verbosity level for the osc framework (0 = no verbosity) MCA osc: parameter "osc_pt2pt_no_locks" (current value: "0") Enable optimizations available only if MPI_LOCK is not used. MCA osc: parameter "osc_pt2pt_eager_limit" (current value: "16384")
                          Max size of eagerly sent data
MCA osc: parameter "osc_pt2pt_priority" (current value: "0")
              MCA errmgr: parameter "errmgr" (current value: <none>)
Default selection set of components for the errmgr framework (<none> means "use all components that can be found") MCA errmgr: parameter "errmgr_hnp_debug" (current value: "0") MCA errmgr: parameter "errmgr_hnp_priority" (current value: "0") MCA errmgr: parameter "errmgr_orted_debug" (current value: "0") MCA errmgr: parameter "errmgr_orted_priority" (current value: "0") MCA errmgr: parameter "errmgr_proxy_debug" (current value: "0") MCA errmgr: parameter "errmgr_proxy_priority" (current value: "0") MCA gpr: parameter "gpr_base_maxsize" (current value: "2147483647") MCA gpr: parameter "gpr_base_blocksize" (current value: "512")
                 MCA gpr: parameter "gpr" (current value: <none>)
Default selection set of components for the gpr framework (<none> means "use all components that can be found") MCA gpr: parameter "gpr_null_priority" (current value: "0") MCA gpr: parameter "gpr_proxy_debug" (current value: "0") MCA gpr: parameter "gpr_proxy_priority" (current value: "0") MCA gpr: parameter "gpr_replica_debug" (current value: "0") MCA gpr: parameter "gpr_replica_isolate" (current value: "0") MCA gpr: parameter "gpr_replica_priority" (current value: "0") MCA iof: parameter "iof_base_window_size" (current value: "4096") MCA iof: parameter "iof_base_service" (current value: "0.0.0")
                 MCA iof: parameter "iof" (current value: <none>)
Default selection set of components for the iof framework (<none> means "use all components that can be found") MCA iof: parameter "iof_proxy_debug" (current value: "1") MCA iof: parameter "iof_proxy_priority" (current value: "0") MCA iof: parameter "iof_svc_debug" (current value: "1") MCA iof: parameter "iof_svc_priority" (current value: "0")
                  MCA ns: parameter "ns" (current value: <none>)
Default selection set of components for the ns framework (<none> means "use all components that can be found") MCA ns: parameter "ns_proxy_debug" (current value: "0") MCA ns: parameter "ns_proxy_maxsize" (current value: "2147483647") MCA ns: parameter "ns_proxy_blocksize" (current value: "512") MCA ns: parameter "ns_proxy_priority" (current value: "0") MCA ns: parameter "ns_replica_debug" (current value: "0") MCA ns: parameter "ns_replica_isolate" (current value: "0") MCA ns: parameter "ns_replica_maxsize" (current value: "2147483647") MCA ns: parameter "ns_replica_blocksize" (current value: "512") MCA ns: parameter "ns_replica_priority" (current value: "0")
                 MCA oob: parameter "oob" (current value: <none>)
Default selection set of components for the oob framework (<none> means "use all components that can be found") MCA oob: parameter "oob_base_verbose" (current value: "0") Verbosity level for the oob framework (0 = no verbosity) MCA oob: parameter "oob_tcp_peer_limit" (current value: "-1") MCA oob: parameter "oob_tcp_peer_retries" (current value: "60") MCA oob: parameter "oob_tcp_debug" (current value: "0") MCA oob: parameter "oob_tcp_include" (current value: <none>) MCA oob: parameter "oob_tcp_exclude" (current value: <none>) MCA oob: parameter "oob_tcp_sndbuf" (current value: "131072") MCA oob: parameter "oob_tcp_rcvbuf" (current value: "131072") MCA oob: parameter "oob_tcp_connect_sleep" (current value: "1") Enable (1) /Disable (0) random sleep for connection wireup MCA oob: parameter "oob_tcp_listen_mode" (current value: "event") Mode for HNP to accept incoming connections: event, listen_thread MCA oob: parameter "oob_tcp_listen_thread_max_queue" (current value: "10") High water mark for queued accepted socket list size MCA oob: parameter "oob_tcp_listen_thread_max_time" (current value: "10") Maximum amount of time (in milliseconds) to wait between processing accepted socket list MCA oob: parameter "oob_tcp_accept_spin_count" (current value: "10") Number of times to let accept return EWOULDBLOCK before updating accepted socket list MCA oob: parameter "oob_tcp_priority" (current value: "0")
                 MCA ras: parameter "ras" (current value: <none>)
MCA ras: parameter "ras_dash_host_priority" (current value: "5") Selection priority for the dash_host RAS component MCA ras: parameter "ras_gridengine_debug" (current value: "0") Enable debugging output for the gridengine ras component MCA ras: parameter "ras_gridengine_priority" (current value: "100")
                          Priority of the gridengine ras component
MCA ras: parameter "ras_gridengine_verbose" (current value: "0") Enable verbose output for the gridengine ras component MCA ras: parameter "ras_gridengine_show_jobid" (current value: "0")
                          Show the JOB_ID of the Grid Engine job
MCA ras: parameter "ras_localhost_priority" (current value: "0") Selection priority for the localhost RAS component MCA ras: parameter "ras_slurm_priority" (current value: "75")
                          Priority of the slurm ras component
                 MCA rds: parameter "rds" (current value: <none>)
MCA rds: parameter "rds_hostfile_debug" (current value: "0") Toggle debug output for hostfile RDS component MCA rds: parameter "rds_hostfile_path" (current value: "/usr/local/openmpi-1.2.2/etc/openmpi-default-hostfile")
                          ORTE Host filename
MCA rds: parameter "rds_hostfile_priority" (current value: "0") MCA rds: parameter "rds_proxy_priority" (current value: "0") MCA rds: parameter "rds_resfile_debug" (current value: "0")
                          Toggle debug output for resfile RDS component
MCA rds: parameter "rds_resfile_name" (current value: <none>)
                          ORTE Resource filename
MCA rds: parameter "rds_resfile_priority" (current value: "0") MCA rmaps: parameter "rmaps_base_verbose" (current value: "0")
                          Verbosity level for the rmaps framework
MCA rmaps: parameter "rmaps_base_schedule_policy" (current value: "unspec")
                          Scheduling Policy for RMAPS. [slot | node]
MCA rmaps: parameter "rmaps_base_pernode" (current value: "0")
                          Launch one ppn as directed
MCA rmaps: parameter "rmaps_base_n_pernode" (current value: "-1")
                          Launch n procs/node
MCA rmaps: parameter "rmaps_base_no_schedule_local" (current value: "0") If false, allow scheduling MPI applications on the same node as mpirun (default). If true, do not schedule any MPI applications on the same node as mpirun MCA rmaps: parameter "rmaps_base_no_oversubscribe" (current value: "0") If true, then do not allow oversubscription of nodes - mpirun will return an error if there aren't enough nodes to launch all processes without oversubscribing
               MCA rmaps: parameter "rmaps" (current value: <none>)
Default selection set of components for the rmaps framework (<none> means "use all components that can be found") MCA rmaps: parameter "rmaps_round_robin_debug" (current value: "1") Toggle debug output for Round Robin RMAPS component MCA rmaps: parameter "rmaps_round_robin_priority" (current value: "1") Selection priority for Round Robin RMAPS component
                MCA rmgr: parameter "rmgr" (current value: <none>)
Default selection set of components for the rmgr framework (<none> means "use all components that can be found") MCA rmgr: parameter "rmgr_proxy_priority" (current value: "0") MCA rmgr: parameter "rmgr_urm_priority" (current value: "0")
                 MCA rml: parameter "rml" (current value: <none>)
Default selection set of components for the rml framework (<none> means "use all components that can be found") MCA rml: parameter "rml_base_verbose" (current value: "0") Verbosity level for the rml framework (0 = no verbosity) MCA rml: parameter "rml_oob_priority" (current value: "0") MCA pls: parameter "pls_base_reuse_daemons" (current value: "0") If nonzero, reuse daemons to launch dynamically spawned processes. If zero, do not reuse daemons (default)
                 MCA pls: parameter "pls" (current value: <none>)
Default selection set of components for the pls framework (<none> means "use all components that can be found") MCA pls: parameter "pls_base_verbose" (current value: "0") Verbosity level for the pls framework (0 = no verbosity) MCA pls: parameter "pls_gridengine_debug" (current value: "0")
                          Enable debugging of gridengine pls component
MCA pls: parameter "pls_gridengine_verbose" (current value: "0") Enable verbose output of the gridengine qrsh -inherit command MCA pls: parameter "pls_gridengine_priority" (current value: "100")
                          Priority of the gridengine pls component
MCA pls: parameter "pls_gridengine_orted" (current value: "orted") The command name that the gridengine pls component will invoke for the ORTE daemon MCA pls: parameter "pls_proxy_priority" (current value: "0") MCA pls: parameter "pls_rsh_debug" (current value: "0") Whether or not to enable debugging output for the rsh pls component (0 or 1) MCA pls: parameter "pls_rsh_num_concurrent" (current value: "128") How many pls_rsh_agent instances to invoke concurrently (must be > 0) MCA pls: parameter "pls_rsh_force_rsh" (current value: "0") Force the launcher to always use rsh, even for local daemons MCA pls: parameter "pls_rsh_orted" (current value: "orted") The command name that the rsh pls component will invoke for the ORTE daemon MCA pls: parameter "pls_rsh_priority" (current value: "10")
                          Priority of the rsh pls component
MCA pls: parameter "pls_rsh_delay" (current value: "1") Delay (in seconds) between invocations of the remote agent, but only used when the "debug" MCA parameter is true, or the top-level MCA debugging is enabled (otherwise this value is ignored)
                 MCA pls: parameter "pls_rsh_reap" (current value: "1")
If set to 1, wait for all the processes to complete before exiting. Otherwise, quit immediately -- without waiting for confirmation that all other processes in the job have completed. MCA pls: parameter "pls_rsh_assume_same_shell" (current value: "1") If set to 1, assume that the shell on the remote node is the same as the shell on the local node. Otherwise, probe for what the remote shell. MCA pls: parameter "pls_rsh_agent" (current value: "rsh") The command used to launch executables on remote nodes (typically either "ssh" or "rsh") MCA pls: parameter "pls_slurm_debug" (current value: "0")
                          Enable debugging of slurm pls
MCA pls: parameter "pls_slurm_priority" (current value: "75")
                          Default selection priority
MCA pls: parameter "pls_slurm_orted" (current value: "orted")
                          Command to use to start proxy orted
MCA pls: parameter "pls_slurm_args" (current value: <none>)
                          Custom arguments to srun
                 MCA sds: parameter "sds" (current value: <none>)
Default selection set of components for the sds framework (<none> means "use all components that can be found") MCA sds: parameter "sds_base_verbose" (current value: "0") Verbosity level for the sds framework (0 = no verbosity) MCA sds: parameter "sds_env_priority" (current value: "0") MCA sds: parameter "sds_pipe_priority" (current value: "0") MCA sds: parameter "sds_seed_priority" (current value: "0") MCA sds: parameter "sds_singleton_priority" (current value: "0") MCA sds: parameter "sds_slurm_priority" (current value: "0")
-bash-3.00$

Reply via email to