Hello!
Here is my time results:
$time mpirun -n 1 ./hello_c
Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI 
semenov@compiler-2 Distribution, ident: 1.9a1r32570, repo rev: r32570, Aug 21, 
2014 (nightly snapshot tarball), 146)
real 1m3.985s
user 0m0.031s
sys 0m0.083s


Fri, 22 Aug 2014 07:43:03 -0700 от Ralph Castain <r...@open-mpi.org>:
>I'm also puzzled by your timing statement - I can't replicate it:
>
>07:41:43  $ time mpirun -n 1 ./hello_c
>Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI rhc@bend001 
>Distribution, ident: 1.9a1r32577, repo rev: r32577, Unreleased developer copy, 
>125)
>
>real 0m0.547s
>user 0m0.043s
>sys 0m0.046s
>
>The entire thing ran in 0.5 seconds
>
>
>On Aug 22, 2014, at 6:33 AM, Mike Dubman < mi...@dev.mellanox.co.il > wrote:
>>Hi,
>>The default delimiter is ";" . You can change delimiter with 
>>mca_base_env_list_delimiter.
>>
>>
>>
>>On Fri, Aug 22, 2014 at 2:59 PM, Timur Ismagilov  < tismagi...@mail.ru > 
>>wrote:
>>>Hello!
>>>If i use latest night snapshot:
>>>$ ompi_info -V
>>>Open MPI v1.9a1r32570
>>>*  In programm hello_c initialization takes ~1 min
>>>In ompi 1.8.2rc4 and ealier it takes ~1 sec(or less)
>>>*  if i use 
>>>$mpirun  --mca mca_base_env_list 'MXM_SHM_KCOPY_MODE=off,OMP_NUM_THREADS=8' 
>>>--map-by slot:pe=8 -np 1 ./hello_c
>>>i got error 
>>>config_parser.c:657  MXM  ERROR Invalid value for SHM_KCOPY_MODE: 
>>>'off,OMP_NUM_THREADS=8'. Expected: [off|knem|cma|autodetect]
>>>but with -x all works fine (but with warn)
>>>$mpirun  -x MXM_SHM_KCOPY_MODE=off -x OMP_NUM_THREADS=8 -np 1 ./hello_c
>>>WARNING: The mechanism by which environment variables are explicitly
>>>..............
>>>..............
>>>..............
>>>Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI 
>>>semenov@compiler-2 Distribution, ident: 1.9a1r32570, repo rev: r32570, Aug 
>>>21, 2014 (nightly snapshot tarball), 146)
>>>Thu, 21 Aug 2014 06:26:13 -0700 от Ralph Castain < r...@open-mpi.org >:
>>>>Not sure I understand. The problem has been fixed in both the trunk and the 
>>>>1.8 branch now, so you should be able to work with either of those nightly 
>>>>builds.
>>>>
>>>>On Aug 21, 2014, at 12:02 AM, Timur Ismagilov < tismagi...@mail.ru > wrote:
>>>>>Have i I any opportunity to run mpi jobs?
>>>>>
>>>>>
>>>>>Wed, 20 Aug 2014 10:48:38 -0700 от Ralph Castain < r...@open-mpi.org >:
>>>>>>yes, i know - it is cmr'd
>>>>>>
>>>>>>On Aug 20, 2014, at 10:26 AM, Mike Dubman < mi...@dev.mellanox.co.il > 
>>>>>>wrote:
>>>>>>>btw, we get same error in v1.8 branch as well.
>>>>>>>
>>>>>>>
>>>>>>>On Wed, Aug 20, 2014 at 8:06 PM, Ralph Castain   < r...@open-mpi.org >   
>>>>>>>wrote:
>>>>>>>>It was not yet fixed - but should be now.
>>>>>>>>
>>>>>>>>On Aug 20, 2014, at 6:39 AM, Timur Ismagilov < tismagi...@mail.ru > 
>>>>>>>>wrote:
>>>>>>>>>Hello!
>>>>>>>>>
>>>>>>>>>As i can see, the bug is fixed, but in Open MPI v1.9a1r32516  i still 
>>>>>>>>>have the problem
>>>>>>>>>
>>>>>>>>>a)
>>>>>>>>>$ mpirun  -np 1 ./hello_c
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>An ORTE daemon has unexpectedly failed after launch and before
>>>>>>>>>communicating back to mpirun. This could be caused by a number
>>>>>>>>>of factors, including an inability to create a connection back
>>>>>>>>>to mpirun due to a lack of common network interfaces and/or no
>>>>>>>>>route found between them. Please check network connectivity
>>>>>>>>>(including firewalls and network routing requirements).
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>b)
>>>>>>>>>$ mpirun --mca oob_tcp_if_include ib0 -np 1 ./hello_c
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>An ORTE daemon has unexpectedly failed after launch and before
>>>>>>>>>communicating back to mpirun. This could be caused by a number
>>>>>>>>>of factors, including an inability to create a connection back
>>>>>>>>>to mpirun due to a lack of common network interfaces and/or no
>>>>>>>>>route found between them. Please check network connectivity
>>>>>>>>>(including firewalls and network routing requirements).
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>
>>>>>>>>>c)
>>>>>>>>>
>>>>>>>>>$ mpirun --mca oob_tcp_if_include ib0 -debug-daemons --mca 
>>>>>>>>>plm_base_verbose 5 -mca oob_base_verbose 10 -mca rml_base_verbose 10 
>>>>>>>>>-np 1 ./hello_c
>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Querying component [isolated]
>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Query of component 
>>>>>>>>>[isolated] set priority to 0
>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Querying component [rsh]
>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Query of component [rsh] set 
>>>>>>>>>priority to 10
>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Querying component [slurm]
>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Query of component [slurm] 
>>>>>>>>>set priority to 75
>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Selected component [slurm]
>>>>>>>>>[compiler-2:14673] mca: base: components_register: registering oob 
>>>>>>>>>components
>>>>>>>>>[compiler-2:14673] mca: base: components_register: found loaded 
>>>>>>>>>component tcp
>>>>>>>>>[compiler-2:14673] mca: base: components_register: component tcp 
>>>>>>>>>register function successful
>>>>>>>>>[compiler-2:14673] mca: base: components_open: opening oob components
>>>>>>>>>[compiler-2:14673] mca: base: components_open: found loaded component 
>>>>>>>>>tcp
>>>>>>>>>[compiler-2:14673] mca: base: components_open: component tcp open 
>>>>>>>>>function successful
>>>>>>>>>[compiler-2:14673] mca:oob:select: checking available component tcp
>>>>>>>>>[compiler-2:14673] mca:oob:select: Querying component [tcp]
>>>>>>>>>[compiler-2:14673] oob:tcp: component_available called
>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 1 KERNEL INDEX 1 FAMILY: V4
>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 2 KERNEL INDEX 3 FAMILY: V4
>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 3 KERNEL INDEX 4 FAMILY: V4
>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 4 KERNEL INDEX 5 FAMILY: V4
>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 5 KERNEL INDEX 6 FAMILY: V4
>>>>>>>>>[compiler-2:14673] [[49095,0],0] oob:tcp:init adding 10.128.0.4 to our 
>>>>>>>>>list of V4 connections
>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 6 KERNEL INDEX 7 FAMILY: V4
>>>>>>>>>[compiler-2:14673] [[49095,0],0] TCP STARTUP
>>>>>>>>>[compiler-2:14673] [[49095,0],0] attempting to bind to IPv4 port 0
>>>>>>>>>[compiler-2:14673] [[49095,0],0] assigned IPv4 port 59460
>>>>>>>>>[compiler-2:14673] mca:oob:select: Adding component to end
>>>>>>>>>[compiler-2:14673] mca:oob:select: Found 1 active transports
>>>>>>>>>[compiler-2:14673] mca: base: components_register: registering rml 
>>>>>>>>>components
>>>>>>>>>[compiler-2:14673] mca: base: components_register: found loaded 
>>>>>>>>>component oob
>>>>>>>>>[compiler-2:14673] mca: base: components_register: component oob has 
>>>>>>>>>no register or open function
>>>>>>>>>[compiler-2:14673] mca: base: components_open: opening rml components
>>>>>>>>>[compiler-2:14673] mca: base: components_open: found loaded component 
>>>>>>>>>oob
>>>>>>>>>[compiler-2:14673] mca: base: components_open: component oob open 
>>>>>>>>>function successful
>>>>>>>>>[compiler-2:14673] orte_rml_base_select: initializing rml component oob
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 30 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 15 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 32 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 33 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 5 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 10 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 12 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 9 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 34 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 2 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 21 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 22 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 45 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 46 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 1 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 27 for 
>>>>>>>>>peer [[WILDCARD],WILDCARD]
>>>>>>>>>Daemon was launched on node1-128-01 - beginning to initialize
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>WARNING: An invalid value was given for oob_tcp_if_include. This
>>>>>>>>>value will be ignored.
>>>>>>>>>Local host: node1-128-01
>>>>>>>>>Value: "ib0"
>>>>>>>>>Message: Invalid specification (missing "/")
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>None of the TCP networks specified to be included for out-of-band 
>>>>>>>>>communications
>>>>>>>>>could be found:
>>>>>>>>>Value given:
>>>>>>>>>Please revise the specification and try again.
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>No network interfaces were found for out-of-band communications. We 
>>>>>>>>>require
>>>>>>>>>at least one available network for out-of-band messaging.
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>It looks like orte_init failed for some reason; your parallel process 
>>>>>>>>>is
>>>>>>>>>likely to abort. There are many reasons that a parallel process can
>>>>>>>>>fail during orte_init; some of which are due to configuration or
>>>>>>>>>environment problems. This failure appears to be an internal failure;
>>>>>>>>>here's some additional information (which may only be relevant to an
>>>>>>>>>Open MPI developer):
>>>>>>>>>orte_oob_base_select failed
>>>>>>>>>--> Returned value (null) (-43) instead of ORTE_SUCCESS
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>srun: error: node1-128-01: task 0: Exited with exit code 213
>>>>>>>>>srun: Terminating job step 661215.0
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>An ORTE daemon has unexpectedly failed after launch and before
>>>>>>>>>communicating back to mpirun. This could be caused by a number
>>>>>>>>>of factors, including an inability to create a connection back
>>>>>>>>>to mpirun due to a lack of common network interfaces and/or no
>>>>>>>>>route found between them. Please check network connectivity
>>>>>>>>>(including firewalls and network routing requirements).
>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>[compiler-2:14673] [[49095,0],0] orted_cmd: received halt_vm cmd
>>>>>>>>>[compiler-2:14673] mca: base: close: component oob closed
>>>>>>>>>[compiler-2:14673] mca: base: close: unloading component oob
>>>>>>>>>[compiler-2:14673] [[49095,0],0] TCP SHUTDOWN
>>>>>>>>>[compiler-2:14673] mca: base: close: component tcp closed
>>>>>>>>>[compiler-2:14673] mca: base: close: unloading component tcp
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>Tue, 12 Aug 2014 18:33:24 +0000 от "Jeff Squyres (jsquyres)" < 
>>>>>>>>>jsquy...@cisco.com >:
>>>>>>>>>>I filed the following ticket:
>>>>>>>>>>
>>>>>>>>>>     https://svn.open-mpi.org/trac/ompi/ticket/4857
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>On Aug 12, 2014, at 12:39 PM, Jeff Squyres (jsquyres) < 
>>>>>>>>>>jsquy...@cisco.com > wrote:
>>>>>>>>>>
>>>>>>>>>>> (please keep the users list CC'ed)
>>>>>>>>>>>  
>>>>>>>>>>> We talked about this on the weekly engineering call today. Ralph 
>>>>>>>>>>> has an idea what is happening -- I need to do a little 
>>>>>>>>>>> investigation today and file a bug. I'll make sure you're CC'ed on 
>>>>>>>>>>> the bug ticket.
>>>>>>>>>>>  
>>>>>>>>>>>  
>>>>>>>>>>>  
>>>>>>>>>>> On Aug 12, 2014, at 12:27 PM, Timur Ismagilov < tismagi...@mail.ru 
>>>>>>>>>>> > wrote:
>>>>>>>>>>>  
>>>>>>>>>>>> I don't have this error in OMPI 1.9a1r32252 and OMPI 1.8.1 (with 
>>>>>>>>>>>> --mca oob_tcp_if_include ib0), but in all latest night snapshots i 
>>>>>>>>>>>> got this error.
>>>>>>>>>>>>  
>>>>>>>>>>>>  
>>>>>>>>>>>> Tue, 12 Aug 2014 13:08:12 +0000 от "Jeff Squyres (jsquyres)" < 
>>>>>>>>>>>> jsquy...@cisco.com >:
>>>>>>>>>>>> Are you running any kind of firewall on the node where mpirun is 
>>>>>>>>>>>> invoked? Open MPI needs to be able to use arbitrary TCP ports 
>>>>>>>>>>>> between the servers on which it runs.
>>>>>>>>>>>>  
>>>>>>>>>>>> This second mail seems to imply a bug in OMPI's oob_tcp_if_include 
>>>>>>>>>>>> param handling, however -- it's supposed to be able to handle an 
>>>>>>>>>>>> interface name (not just a network specification).
>>>>>>>>>>>>  
>>>>>>>>>>>> Ralph -- can you have a look?
>>>>>>>>>>>>  
>>>>>>>>>>>>  
>>>>>>>>>>>> On Aug 12, 2014, at 8:41 AM, Timur Ismagilov < tismagi...@mail.ru 
>>>>>>>>>>>> > wrote:
>>>>>>>>>>>>  
>>>>>>>>>>>>> When i add --mca oob_tcp_if_include ib0 (infiniband interface) to 
>>>>>>>>>>>>> mpirun (as it was here:   
>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/07/24857.php   
>>>>>>>>>>>>> ) i got this output:
>>>>>>>>>>>>>  
>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Querying component 
>>>>>>>>>>>>> [isolated]
>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Query of component 
>>>>>>>>>>>>> [isolated] set priority to 0
>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Querying component [rsh]
>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Query of component 
>>>>>>>>>>>>> [rsh] set priority to 10
>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Querying component 
>>>>>>>>>>>>> [slurm]
>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Query of component 
>>>>>>>>>>>>> [slurm] set priority to 75
>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Selected component 
>>>>>>>>>>>>> [slurm]
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: registering 
>>>>>>>>>>>>> oob components
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: found loaded 
>>>>>>>>>>>>> component tcp
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: component tcp 
>>>>>>>>>>>>> register function successful
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: opening oob 
>>>>>>>>>>>>> components
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: found loaded 
>>>>>>>>>>>>> component tcp
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: component tcp open 
>>>>>>>>>>>>> function successful
>>>>>>>>>>>>> [compiler-2:08792] mca:oob:select: checking available component 
>>>>>>>>>>>>> tcp
>>>>>>>>>>>>> [compiler-2:08792] mca:oob:select: Querying component [tcp]
>>>>>>>>>>>>> [compiler-2:08792] oob:tcp: component_available called
>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 1 KERNEL INDEX 1 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 2 KERNEL INDEX 3 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 3 KERNEL INDEX 4 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 4 KERNEL INDEX 5 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 5 KERNEL INDEX 6 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] oob:tcp:init adding 10.128.0.4 
>>>>>>>>>>>>> to our list of V4 connections
>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 6 KERNEL INDEX 7 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] TCP STARTUP
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] attempting to bind to IPv4 port 0
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] assigned IPv4 port 53883
>>>>>>>>>>>>> [compiler-2:08792] mca:oob:select: Adding component to end
>>>>>>>>>>>>> [compiler-2:08792] mca:oob:select: Found 1 active transports
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: registering 
>>>>>>>>>>>>> rml components
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: found loaded 
>>>>>>>>>>>>> component oob
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: component oob 
>>>>>>>>>>>>> has no register or open function
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: opening rml 
>>>>>>>>>>>>> components
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: found loaded 
>>>>>>>>>>>>> component oob
>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: component oob open 
>>>>>>>>>>>>> function successful
>>>>>>>>>>>>> [compiler-2:08792] orte_rml_base_select: initializing rml 
>>>>>>>>>>>>> component oob
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 30 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 15 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 32 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 33 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 5 
>>>>>>>>>>>>> for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 10 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 12 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 9 
>>>>>>>>>>>>> for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 34 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 2 
>>>>>>>>>>>>> for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 21 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 22 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 45 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 46 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 1 
>>>>>>>>>>>>> for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 27 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> Daemon was launched on node1-128-01 - beginning to initialize
>>>>>>>>>>>>> Daemon was launched on node1-128-02 - beginning to initialize
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> WARNING: An invalid value was given for oob_tcp_if_include. This
>>>>>>>>>>>>> value will be ignored.
>>>>>>>>>>>>>  
>>>>>>>>>>>>> Local host: node1-128-01
>>>>>>>>>>>>> Value: "ib0"
>>>>>>>>>>>>> Message: Invalid specification (missing "/")
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> WARNING: An invalid value was given for oob_tcp_if_include. This
>>>>>>>>>>>>> value will be ignored.
>>>>>>>>>>>>>  
>>>>>>>>>>>>> Local host: node1-128-02
>>>>>>>>>>>>> Value: "ib0"
>>>>>>>>>>>>> Message: Invalid specification (missing "/")
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> None of the TCP networks specified to be included for out-of-band 
>>>>>>>>>>>>> communications
>>>>>>>>>>>>> could be found:
>>>>>>>>>>>>>  
>>>>>>>>>>>>> Value given:
>>>>>>>>>>>>>  
>>>>>>>>>>>>> Please revise the specification and try again.
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> None of the TCP networks specified to be included for out-of-band 
>>>>>>>>>>>>> communications
>>>>>>>>>>>>> could be found:
>>>>>>>>>>>>>  
>>>>>>>>>>>>> Value given:
>>>>>>>>>>>>>  
>>>>>>>>>>>>> Please revise the specification and try again.
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> No network interfaces were found for out-of-band communications. 
>>>>>>>>>>>>> We require
>>>>>>>>>>>>> at least one available network for out-of-band messaging.
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> No network interfaces were found for out-of-band communications. 
>>>>>>>>>>>>> We require
>>>>>>>>>>>>> at least one available network for out-of-band messaging.
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> It looks like orte_init failed for some reason; your parallel 
>>>>>>>>>>>>> process is
>>>>>>>>>>>>> likely to abort. There are many reasons that a parallel process 
>>>>>>>>>>>>> can
>>>>>>>>>>>>> fail during orte_init; some of which are due to configuration or
>>>>>>>>>>>>> environment problems. This failure appears to be an internal 
>>>>>>>>>>>>> failure;
>>>>>>>>>>>>> here's some additional information (which may only be relevant to 
>>>>>>>>>>>>> an
>>>>>>>>>>>>> Open MPI developer):
>>>>>>>>>>>>>  
>>>>>>>>>>>>> orte_oob_base_select failed
>>>>>>>>>>>>> --> Returned value (null) (-43) instead of ORTE_SUCCESS
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> It looks like orte_init failed for some reason; your parallel 
>>>>>>>>>>>>> process is
>>>>>>>>>>>>> likely to abort. There are many reasons that a parallel process 
>>>>>>>>>>>>> can
>>>>>>>>>>>>> fail during orte_init; some of which are due to configuration or
>>>>>>>>>>>>> environment problems. This failure appears to be an internal 
>>>>>>>>>>>>> failure;
>>>>>>>>>>>>> here's some additional information (which may only be relevant to 
>>>>>>>>>>>>> an
>>>>>>>>>>>>> Open MPI developer):
>>>>>>>>>>>>>  
>>>>>>>>>>>>> orte_oob_base_select failed
>>>>>>>>>>>>> --> Returned value (null) (-43) instead of ORTE_SUCCESS
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> srun: error: node1-128-02: task 1: Exited with exit code 213
>>>>>>>>>>>>> srun: Terminating job step 657300.0
>>>>>>>>>>>>> srun: error: node1-128-01: task 0: Exited with exit code 213
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> An ORTE daemon has unexpectedly failed after launch and before
>>>>>>>>>>>>> communicating back to mpirun. This could be caused by a number
>>>>>>>>>>>>> of factors, including an inability to create a connection back
>>>>>>>>>>>>> to mpirun due to a lack of common network interfaces and/or no
>>>>>>>>>>>>> route found between them. Please check network connectivity
>>>>>>>>>>>>> (including firewalls and network routing requirements).
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] orted_cmd: received halt_vm cmd
>>>>>>>>>>>>> [compiler-2:08792] mca: base: close: component oob closed
>>>>>>>>>>>>> [compiler-2:08792] mca: base: close: unloading component oob
>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] TCP SHUTDOWN
>>>>>>>>>>>>> [compiler-2:08792] mca: base: close: component tcp closed
>>>>>>>>>>>>> [compiler-2:08792] mca: base: close: unloading component tcp
>>>>>>>>>>>>>  
>>>>>>>>>>>>>  
>>>>>>>>>>>>>  
>>>>>>>>>>>>> Tue, 12 Aug 2014 16:14:58 +0400 от Timur Ismagilov < 
>>>>>>>>>>>>> tismagi...@mail.ru >:
>>>>>>>>>>>>> Hello!
>>>>>>>>>>>>>  
>>>>>>>>>>>>> I have Open MPI v1.8.2rc4r32485
>>>>>>>>>>>>>  
>>>>>>>>>>>>> When i run hello_c, I got this error message
>>>>>>>>>>>>> $mpirun -np 2 hello_c
>>>>>>>>>>>>>  
>>>>>>>>>>>>> An ORTE daemon has unexpectedly failed after launch and before
>>>>>>>>>>>>>  
>>>>>>>>>>>>> communicating back to mpirun. This could be caused by a number
>>>>>>>>>>>>> of factors, including an inability to create a connection back
>>>>>>>>>>>>> to mpirun due to a lack of common network interfaces and/or no
>>>>>>>>>>>>> route found between them. Please check network connectivity
>>>>>>>>>>>>> (including firewalls and network routing requirements).
>>>>>>>>>>>>>  
>>>>>>>>>>>>> When i run with --debug-daemons --mca plm_base_verbose 5 -mca 
>>>>>>>>>>>>> oob_base_verbose 10 -mca rml_base_verbose 10 i got this output:
>>>>>>>>>>>>> $mpirun --debug-daemons --mca plm_base_verbose 5 -mca 
>>>>>>>>>>>>> oob_base_verbose 10 -mca rml_base_verbose 10 -np 2 hello_c
>>>>>>>>>>>>>  
>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Querying component 
>>>>>>>>>>>>> [isolated]
>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Query of component 
>>>>>>>>>>>>> [isolated] set priority to 0
>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Querying component [rsh]
>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Query of component 
>>>>>>>>>>>>> [rsh] set priority to 10
>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Querying component 
>>>>>>>>>>>>> [slurm]
>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Query of component 
>>>>>>>>>>>>> [slurm] set priority to 75
>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Selected component 
>>>>>>>>>>>>> [slurm]
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: registering 
>>>>>>>>>>>>> oob components
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: found loaded 
>>>>>>>>>>>>> component tcp
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: component tcp 
>>>>>>>>>>>>> register function successful
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: opening oob 
>>>>>>>>>>>>> components
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: found loaded 
>>>>>>>>>>>>> component tcp
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: component tcp open 
>>>>>>>>>>>>> function successful
>>>>>>>>>>>>> [compiler-2:08780] mca:oob:select: checking available component 
>>>>>>>>>>>>> tcp
>>>>>>>>>>>>> [compiler-2:08780] mca:oob:select: Querying component [tcp]
>>>>>>>>>>>>> [compiler-2:08780] oob:tcp: component_available called
>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 1 KERNEL INDEX 1 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 2 KERNEL INDEX 3 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 10.0.251.53 
>>>>>>>>>>>>> to our list of V4 connections
>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 3 KERNEL INDEX 4 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 10.0.0.4 to 
>>>>>>>>>>>>> our list of V4 connections
>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 4 KERNEL INDEX 5 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 10.2.251.14 
>>>>>>>>>>>>> to our list of V4 connections
>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 5 KERNEL INDEX 6 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 10.128.0.4 
>>>>>>>>>>>>> to our list of V4 connections
>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 6 KERNEL INDEX 7 FAMILY: V4
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 93.180.7.38 
>>>>>>>>>>>>> to our list of V4 connections
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] TCP STARTUP
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] attempting to bind to IPv4 port 0
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] assigned IPv4 port 38420
>>>>>>>>>>>>> [compiler-2:08780] mca:oob:select: Adding component to end
>>>>>>>>>>>>> [compiler-2:08780] mca:oob:select: Found 1 active transports
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: registering 
>>>>>>>>>>>>> rml components
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: found loaded 
>>>>>>>>>>>>> component oob
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: component oob 
>>>>>>>>>>>>> has no register or open function
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: opening rml 
>>>>>>>>>>>>> components
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: found loaded 
>>>>>>>>>>>>> component oob
>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: component oob open 
>>>>>>>>>>>>> function successful
>>>>>>>>>>>>> [compiler-2:08780] orte_rml_base_select: initializing rml 
>>>>>>>>>>>>> component oob
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 30 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 15 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 32 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 33 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 5 
>>>>>>>>>>>>> for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 10 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 12 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 9 
>>>>>>>>>>>>> for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 34 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 2 
>>>>>>>>>>>>> for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 21 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 22 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 45 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 46 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 1 
>>>>>>>>>>>>> for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 
>>>>>>>>>>>>> 27 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>> Daemon was launched on node1-130-08 - beginning to initialize
>>>>>>>>>>>>> Daemon was launched on node1-130-03 - beginning to initialize
>>>>>>>>>>>>> Daemon was launched on node1-130-05 - beginning to initialize
>>>>>>>>>>>>> Daemon was launched on node1-130-02 - beginning to initialize
>>>>>>>>>>>>> Daemon was launched on node1-130-01 - beginning to initialize
>>>>>>>>>>>>> Daemon was launched on node1-130-04 - beginning to initialize
>>>>>>>>>>>>> Daemon was launched on node1-130-07 - beginning to initialize
>>>>>>>>>>>>> Daemon was launched on node1-130-06 - beginning to initialize
>>>>>>>>>>>>> Daemon [[42202,0],3] checking in as pid 7178 on host node1-130-03
>>>>>>>>>>>>> [node1-130-03:07178] [[42202,0],3] orted: up and running - 
>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>> Daemon [[42202,0],2] checking in as pid 13581 on host node1-130-02
>>>>>>>>>>>>> [node1-130-02:13581] [[42202,0],2] orted: up and running - 
>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>> Daemon [[42202,0],1] checking in as pid 17220 on host node1-130-01
>>>>>>>>>>>>> [node1-130-01:17220] [[42202,0],1] orted: up and running - 
>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>> Daemon [[42202,0],5] checking in as pid 6663 on host node1-130-05
>>>>>>>>>>>>> [node1-130-05:06663] [[42202,0],5] orted: up and running - 
>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>> Daemon [[42202,0],8] checking in as pid 6683 on host node1-130-08
>>>>>>>>>>>>> [node1-130-08:06683] [[42202,0],8] orted: up and running - 
>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>> Daemon [[42202,0],7] checking in as pid 7877 on host node1-130-07
>>>>>>>>>>>>> [node1-130-07:07877] [[42202,0],7] orted: up and running - 
>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>> Daemon [[42202,0],4] checking in as pid 7735 on host node1-130-04
>>>>>>>>>>>>> [node1-130-04:07735] [[42202,0],4] orted: up and running - 
>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>> Daemon [[42202,0],6] checking in as pid 8451 on host node1-130-06
>>>>>>>>>>>>> [node1-130-06:08451] [[42202,0],6] orted: up and running - 
>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>> srun: error: node1-130-03: task 2: Exited with exit code 1
>>>>>>>>>>>>> srun: Terminating job step 657040.1
>>>>>>>>>>>>> srun: error: node1-130-02: task 1: Exited with exit code 1
>>>>>>>>>>>>> slurmd[node1-130-04]: *** STEP 657040.1 KILLED AT 
>>>>>>>>>>>>> 2014-08-12T12:59:07 WITH SIGNAL 9 ***
>>>>>>>>>>>>> slurmd[node1-130-07]: *** STEP 657040.1 KILLED AT 
>>>>>>>>>>>>> 2014-08-12T12:59:07 WITH SIGNAL 9 ***
>>>>>>>>>>>>> slurmd[node1-130-06]: *** STEP 657040.1 KILLED AT 
>>>>>>>>>>>>> 2014-08-12T12:59:07 WITH SIGNAL 9 ***
>>>>>>>>>>>>> srun: Job step aborted: Waiting up to 2 seconds for job step to 
>>>>>>>>>>>>> finish.
>>>>>>>>>>>>> srun: error: node1-130-01: task 0: Exited with exit code 1
>>>>>>>>>>>>> srun: error: node1-130-05: task 4: Exited with exit code 1
>>>>>>>>>>>>> srun: error: node1-130-08: task 7: Exited with exit code 1
>>>>>>>>>>>>> srun: error: node1-130-07: task 6: Exited with exit code 1
>>>>>>>>>>>>> srun: error: node1-130-04: task 3: Killed
>>>>>>>>>>>>> srun: error: node1-130-06: task 5: Killed
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> An ORTE daemon has unexpectedly failed after launch and before
>>>>>>>>>>>>> communicating back to mpirun. This could be caused by a number
>>>>>>>>>>>>> of factors, including an inability to create a connection back
>>>>>>>>>>>>> to mpirun due to a lack of common network interfaces and/or no
>>>>>>>>>>>>> route found between them. Please check network connectivity
>>>>>>>>>>>>> (including firewalls and network routing requirements).
>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] orted_cmd: received halt_vm cmd
>>>>>>>>>>>>> [compiler-2:08780] mca: base: close: component oob closed
>>>>>>>>>>>>> [compiler-2:08780] mca: base: close: unloading component oob
>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] TCP SHUTDOWN
>>>>>>>>>>>>> [compiler-2:08780] mca: base: close: component tcp closed
>>>>>>>>>>>>> [compiler-2:08780] mca: base: close: unloading component tcp
>>>>>>>>>>>>>  
>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>> users mailing list
>>>>>>>>>>>>>   us...@open-mpi.org
>>>>>>>>>>>>> Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>> Link to this post:   
>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/08/24987.php
>>>>>>>>>>>>>  
>>>>>>>>>>>>>  
>>>>>>>>>>>>>  
>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>> users mailing list
>>>>>>>>>>>>>   us...@open-mpi.org
>>>>>>>>>>>>> Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>> Link to this post:   
>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/08/24988.php
>>>>>>>>>>>>  
>>>>>>>>>>>>  
>>>>>>>>>>>> --  
>>>>>>>>>>>> Jeff Squyres
>>>>>>>>>>>>   jsquy...@cisco.com
>>>>>>>>>>>> For corporate legal information go to:   
>>>>>>>>>>>> http://www.cisco.com/web/about/doing_business/legal/cri/
>>>>>>>>>>>>  
>>>>>>>>>>>>  
>>>>>>>>>>>>  
>>>>>>>>>>>>  
>>>>>>>>>>>>  
>>>>>>>>>>>  
>>>>>>>>>>>  
>>>>>>>>>>> --  
>>>>>>>>>>> Jeff Squyres
>>>>>>>>>>>   jsquy...@cisco.com
>>>>>>>>>>> For corporate legal information go to:   
>>>>>>>>>>> http://www.cisco.com/web/about/doing_business/legal/cri/
>>>>>>>>>>>  
>>>>>>>>>>> _______________________________________________
>>>>>>>>>>> users mailing list
>>>>>>>>>>>   us...@open-mpi.org
>>>>>>>>>>> Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>> Link to this post:   
>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/08/25001.php
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>--  
>>>>>>>>>>Jeff Squyres
>>>>>>>>>>jsquy...@cisco.com
>>>>>>>>>>For corporate legal information go to:   
>>>>>>>>>>http://www.cisco.com/web/about/doing_business/legal/cri/
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>_______________________________________________
>>>>>>>>>users mailing list
>>>>>>>>>us...@open-mpi.org
>>>>>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>Link to this post:   
>>>>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25086.php
>>>>>>>>
>>>>>>>>_______________________________________________
>>>>>>>>users mailing list
>>>>>>>>us...@open-mpi.org
>>>>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>Link to this post:   
>>>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25093.php
>>>>>>>
>>>>>>>_______________________________________________
>>>>>>>users mailing list
>>>>>>>us...@open-mpi.org
>>>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>Link to this post:   
>>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25094.php
>>>>>>_______________________________________________
>>>>>>users mailing list
>>>>>>us...@open-mpi.org
>>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>Link to this post:   
>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25095.php
>>>>>
>>>>>
>>>>>
>>>>>_______________________________________________
>>>>>users mailing list
>>>>>us...@open-mpi.org
>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>Link to this post:   
>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25105.php
>>>
>>>
>>>
>>>
>>>_______________________________________________
>>>users mailing list
>>>us...@open-mpi.org
>>>Subscription:  http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>Link to this post:  
>>>http://www.open-mpi.org/community/lists/users/2014/08/25127.php
>>
>>
>>
>>-- 
>>
>>Kind Regards,
>>
>>M. _______________________________________________
>>users mailing list
>>us...@open-mpi.org
>>Subscription:  http://www.open-mpi.org/mailman/listinfo.cgi/users
>>Link to this post:  
>>http://www.open-mpi.org/community/lists/users/2014/08/25128.php
>_______________________________________________
>users mailing list
>us...@open-mpi.org
>Subscription:  http://www.open-mpi.org/mailman/listinfo.cgi/users
>Link to this post:  
>http://www.open-mpi.org/community/lists/users/2014/08/25129.php




Reply via email to