Your message dated Mon, 29 Apr 2024 09:49:22 +0200
with message-id <zi9raolszjbjp...@smaug.dr-blatt.de>
and subject line OpenMPI bug resolved
has caused the Debian Bug report #1069492,
regarding dune-common: FTBFS on armhf: tests fail
to be marked as done.

This means that you claim that the problem has been dealt with.
If this is not the case it is now your responsibility to reopen the
Bug report if necessary, and/or fix the problem forthwith.

(NB: If you are a system administrator and have no idea what this
message is talking about, this may indicate a serious mail system
misconfiguration somewhere. Please contact ow...@bugs.debian.org
immediately.)


-- 
1069492: https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=1069492
Debian Bug Tracking System
Contact ow...@bugs.debian.org with problems
--- Begin Message ---
Source: dune-common
Version: 2.9.0-6
Severity: serious
Justification: FTBFS
Tags: trixie sid ftbfs
User: lu...@debian.org
Usertags: ftbfs-20240420 ftbfs-trixie ftbfs-t64-armhf

Hi,

During a rebuild of all packages in sid, your package failed to build
on armhf.


Relevant part (hopefully):
> make[5]: Entering directory '/<<PKGBUILDDIR>>/build'
> make[5]: Nothing to be done for 'CMakeFiles/build_tests.dir/build'.
> make[5]: Leaving directory '/<<PKGBUILDDIR>>/build'
> [100%] Built target build_tests
> make[4]: Leaving directory '/<<PKGBUILDDIR>>/build'
> /usr/bin/cmake -E cmake_progress_start /<<PKGBUILDDIR>>/build/CMakeFiles 0
> make[3]: Leaving directory '/<<PKGBUILDDIR>>/build'
> make[2]: Leaving directory '/<<PKGBUILDDIR>>/build'
> cd build; PATH=/<<PKGBUILDDIR>>/debian/tmp-test:$PATH 
> /<<PKGBUILDDIR>>/bin/dune-ctest 
>    Site: ip-10-84-234-171
>    Build name: Linux-c++
> Create new tag: 20240420-0332 - Experimental
> Test project /<<PKGBUILDDIR>>/build
>         Start   1: communicationtest
>   1/115 Test   #1: communicationtest ......................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976535] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976534] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976534] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976534] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start   2: communicationtest-mpi-2
>   2/115 Test   #2: communicationtest-mpi-2 ................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976536] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start   3: indexsettest
>   3/115 Test   #3: indexsettest ...........................   Passed    0.00 
> sec
>         Start   4: remoteindicestest
>   4/115 Test   #4: remoteindicestest ......................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976539] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976538] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976538] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976538] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start   5: remoteindicestest-mpi-2
>   5/115 Test   #5: remoteindicestest-mpi-2 ................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976540] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start   6: selectiontest
>   6/115 Test   #6: selectiontest ..........................   Passed    0.17 
> sec
>         Start   7: syncertest
>   7/115 Test   #7: syncertest .............................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976543] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976542] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976542] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976542] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start   8: syncertest-mpi-2
>   8/115 Test   #8: syncertest-mpi-2 .......................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976544] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start   9: variablesizecommunicatortest
>   9/115 Test   #9: variablesizecommunicatortest ...........***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976546] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976545] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976545] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976545] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  10: variablesizecommunicatortest-mpi-2
>  10/115 Test  #10: variablesizecommunicatortest-mpi-2 .....***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976547] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start  11: mpidatatest-mpi-2
>  11/115 Test  #11: mpidatatest-mpi-2 ......................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976548] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start  12: mpifuturetest
>  12/115 Test  #12: mpifuturetest ..........................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976550] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976549] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976549] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976549] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  13: mpifuturetest-mpi-2
>  13/115 Test  #13: mpifuturetest-mpi-2 ....................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976551] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start  14: mpipacktest-mpi-2
>  14/115 Test  #14: mpipacktest-mpi-2 ......................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976552] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start  15: mpigatherscattertest-mpi-2
>  15/115 Test  #15: mpigatherscattertest-mpi-2 .............***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976553] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start  16: looptest
>  16/115 Test  #16: looptest ...............................   Passed    0.03 
> sec
>         Start  17: standardtest
>  17/115 Test  #17: standardtest ...........................   Passed    0.00 
> sec
>         Start  18: vcarraytest
>  18/115 Test  #18: vcarraytest ............................***Skipped   0.00 
> sec
>         Start  19: vcvectortest
>  19/115 Test  #19: vcvectortest ...........................***Skipped   0.00 
> sec
>         Start  20: arithmetictestsuitetest
>  20/115 Test  #20: arithmetictestsuitetest ................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976559] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976558] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976558] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976558] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  21: arraylisttest
>  21/115 Test  #21: arraylisttest ..........................   Passed    0.00 
> sec
>         Start  22: arraytest
>  22/115 Test  #22: arraytest ..............................   Passed    0.00 
> sec
>         Start  23: assertandreturntest
>  23/115 Test  #23: assertandreturntest ....................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976563] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976562] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976562] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976562] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  24: assertandreturntest_compiletime_fail
>  24/115 Test  #24: assertandreturntest_compiletime_fail ...   Passed    1.08 
> sec
>         Start  25: assertandreturntest_ndebug
>  25/115 Test  #25: assertandreturntest_ndebug .............***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976588] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976587] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976587] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976587] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  26: autocopytest
>  26/115 Test  #26: autocopytest ...........................   Passed    0.00 
> sec
>         Start  27: bigunsignedinttest
>  27/115 Test  #27: bigunsignedinttest .....................   Passed    0.00 
> sec
>         Start  28: bitsetvectortest
>  28/115 Test  #28: bitsetvectortest .......................   Passed    0.00 
> sec
>         Start  29: boundscheckingtest
>  29/115 Test  #29: boundscheckingtest .....................   Passed    0.00 
> sec
>         Start  30: boundscheckingmvtest
>  30/115 Test  #30: boundscheckingmvtest ...................   Passed    0.00 
> sec
>         Start  31: boundscheckingoptest
>  31/115 Test  #31: boundscheckingoptest ...................   Passed    0.00 
> sec
>         Start  32: calloncetest
>  32/115 Test  #32: calloncetest ...........................   Passed    0.00 
> sec
>         Start  33: check_fvector_size
>  33/115 Test  #33: check_fvector_size .....................   Passed    0.00 
> sec
>         Start  34: check_fvector_size_fail1
>  34/115 Test  #34: check_fvector_size_fail1 ...............   Passed    0.72 
> sec
>         Start  35: check_fvector_size_fail2
>  35/115 Test  #35: check_fvector_size_fail2 ...............   Passed    0.73 
> sec
>         Start  36: classnametest-demangled
>  36/115 Test  #36: classnametest-demangled ................   Passed    0.01 
> sec
>         Start  37: classnametest-fallback
>  37/115 Test  #37: classnametest-fallback .................   Passed    0.01 
> sec
>         Start  38: concept
>  38/115 Test  #38: concept ................................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976634] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976633] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976633] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976633] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  39: constexprifelsetest
>  39/115 Test  #39: constexprifelsetest ....................   Passed    0.00 
> sec
>         Start  40: debugaligntest
>  40/115 Test  #40: debugaligntest .........................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976637] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976636] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976636] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976636] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  41: debugalignsimdtest
>  41/115 Test  #41: debugalignsimdtest .....................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976639] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976638] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976638] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976638] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  42: densematrixassignmenttest
>  42/115 Test  #42: densematrixassignmenttest ..............   Passed    0.00 
> sec
>         Start  43: densematrixassignmenttest_fail0
>  43/115 Test  #43: densematrixassignmenttest_fail0 ........   Passed    1.21 
> sec
>         Start  44: densematrixassignmenttest_fail1
>  44/115 Test  #44: densematrixassignmenttest_fail1 ........   Passed    1.20 
> sec
>         Start  45: densematrixassignmenttest_fail2
>  45/115 Test  #45: densematrixassignmenttest_fail2 ........   Passed    1.20 
> sec
>         Start  46: densematrixassignmenttest_fail3
>  46/115 Test  #46: densematrixassignmenttest_fail3 ........   Passed    1.21 
> sec
>         Start  47: densematrixassignmenttest_fail4
>  47/115 Test  #47: densematrixassignmenttest_fail4 ........   Passed    1.22 
> sec
>         Start  48: densematrixassignmenttest_fail5
>  48/115 Test  #48: densematrixassignmenttest_fail5 ........   Passed    1.23 
> sec
>         Start  49: densematrixassignmenttest_fail6
>  49/115 Test  #49: densematrixassignmenttest_fail6 ........   Passed    1.21 
> sec
>         Start  50: densevectorassignmenttest
>  50/115 Test  #50: densevectorassignmenttest ..............   Passed    0.00 
> sec
>         Start  51: diagonalmatrixtest
>  51/115 Test  #51: diagonalmatrixtest .....................   Passed    0.00 
> sec
>         Start  52: dynmatrixtest
>  52/115 Test  #52: dynmatrixtest ..........................   Passed    0.00 
> sec
>         Start  53: dynvectortest
>  53/115 Test  #53: dynvectortest ..........................   Passed    0.00 
> sec
>         Start  54: densevectortest
>  54/115 Test  #54: densevectortest ........................   Passed    0.00 
> sec
>         Start  55: enumsettest
>  55/115 Test  #55: enumsettest ............................   Passed    0.00 
> sec
>         Start  56: filledarraytest
>  56/115 Test  #56: filledarraytest ........................   Passed    0.00 
> sec
>         Start  57: fmatrixtest
>  57/115 Test  #57: fmatrixtest ............................   Passed    0.00 
> sec
>         Start  58: functiontest
>  58/115 Test  #58: functiontest ...........................   Passed    0.00 
> sec
>         Start  59: fvectortest
>  59/115 Test  #59: fvectortest ............................   Passed    0.00 
> sec
>         Start  60: fvectorconversion1d
>  60/115 Test  #60: fvectorconversion1d ....................   Passed    0.00 
> sec
>         Start  61: genericiterator_compile_fail
>  61/115 Test  #61: genericiterator_compile_fail ...........   Passed    0.76 
> sec
>         Start  62: hybridutilitiestest
>  62/115 Test  #62: hybridutilitiestest ....................   Passed    0.00 
> sec
>         Start  63: indicestest
>  63/115 Test  #63: indicestest ............................   Passed    0.00 
> sec
>         Start  64: iscallabletest
>  64/115 Test  #64: iscallabletest .........................   Passed    0.00 
> sec
>         Start  65: iteratorfacadetest2
>  65/115 Test  #65: iteratorfacadetest2 ....................   Passed    0.00 
> sec
>         Start  66: iteratorfacadetest
>  66/115 Test  #66: iteratorfacadetest .....................   Passed    0.00 
> sec
>         Start  67: lrutest
>  67/115 Test  #67: lrutest ................................   Passed    0.00 
> sec
>         Start  68: mathclassifierstest
>  68/115 Test  #68: mathclassifierstest ....................   Passed    0.00 
> sec
>         Start  69: metistest
>  69/115 Test  #69: metistest ..............................***Skipped   0.00 
> sec
>         Start  70: mpicommunicationtest
>  70/115 Test  #70: mpicommunicationtest ...................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976839] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976838] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976838] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976838] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  71: mpicommunicationtest-mpi-2
>  71/115 Test  #71: mpicommunicationtest-mpi-2 .............***Failed    0.01 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976840] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start  72: mpiguardtest
>  72/115 Test  #72: mpiguardtest ...........................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976842] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976841] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976841] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976841] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  73: mpiguardtest-mpi-2
>  73/115 Test  #73: mpiguardtest-mpi-2 .....................***Failed    0.01 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976843] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start  74: mpihelpertest
>  74/115 Test  #74: mpihelpertest ..........................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976845] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976844] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976844] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976844] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  75: mpihelpertest-mpi-2
>  75/115 Test  #75: mpihelpertest-mpi-2 ....................***Failed    0.01 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976846] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start  76: mpihelpertest2
>  76/115 Test  #76: mpihelpertest2 .........................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976848] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976847] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976847] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976847] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  77: mpihelpertest2-mpi-2
>  77/115 Test  #77: mpihelpertest2-mpi-2 ...................***Failed    0.01 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976849] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> 
>         Start  78: overloadsettest
>  78/115 Test  #78: overloadsettest ........................   Passed    0.00 
> sec
>         Start  79: parameterizedobjecttest
>  79/115 Test  #79: parameterizedobjecttest ................   Passed    0.00 
> sec
>         Start  80: parametertreelocaletest
>  80/115 Test  #80: parametertreelocaletest ................***Skipped   0.00 
> sec
>         Start  81: parametertreetest
>  81/115 Test  #81: parametertreetest ......................   Passed    0.00 
> sec
>         Start  82: pathtest
>  82/115 Test  #82: pathtest ...............................   Passed    0.00 
> sec
>         Start  83: poolallocatortest
>  83/115 Test  #83: poolallocatortest ......................   Passed    0.00 
> sec
>         Start  84: powertest
>  84/115 Test  #84: powertest ..............................   Passed    0.00 
> sec
>         Start  85: quadmathtest
>  85/115 Test  #85: quadmathtest ...........................***Skipped   0.00 
> sec
>         Start  86: rangeutilitiestest
>  86/115 Test  #86: rangeutilitiestest .....................   Passed    0.00 
> sec
>         Start  87: referencehelpertest
>  87/115 Test  #87: referencehelpertest ....................***Failed    0.02 
> sec
> --------------------------------------------------------------------------
> Sorry!  You were supposed to get help about:
>     pmix_init:startup:internal-failure
> But I couldn't open the help file:
>     /usr/share/pmix/help-pmix-runtime.txt: No such file or directory.  Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976860] PMIX ERROR: NOT-FOUND in file 
> ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at 
> line 237
> [ip-10-84-234-171:1976859] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976859] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to 
> start a daemon on the local node in file 
> ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) 
> instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of 
> "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-10-84-234-171:1976859] Local abort before MPI_INIT completed completed 
> successfully, but am not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> 
>         Start  88: reservedvectortest
>  88/115 Test  #88: reservedvectortest .....................   Passed    0.00 
> sec
>         Start  89: scotchtest
>  89/115 Test  #89: scotchtest .............................***Skipped   0.00 
> sec
>         Start  90: shared_ptrtest
>  90/115 Test  #90: shared_ptrtest .........................   Passed    0.00 
> sec
>         Start  91: singletontest
>  91/115 Test  #91: singletontest ..........................   Passed    0.00 
> sec
>         Start  92: sllisttest
>  92/115 Test  #92: sllisttest .............................   Passed    0.00 
> sec
>         Start  93: stdidentity
>  93/115 Test  #93: stdidentity ............................   Passed    0.00 
> sec
>         Start  94: stdapplytest
>  94/115 Test  #94: stdapplytest ...........................   Passed    0.00 
> sec
>         Start  95: stdchecktypes
>  95/115 Test  #95: stdchecktypes ..........................   Passed    0.00 
> sec
>         Start  96: streamoperatorstest
>  96/115 Test  #96: streamoperatorstest ....................   Passed    0.00 
> sec
>         Start  97: streamtest
>  97/115 Test  #97: streamtest .............................   Passed    0.00 
> sec
>         Start  98: stringutilitytest
>  98/115 Test  #98: stringutilitytest ......................   Passed    0.00 
> sec
>         Start  99: testdebugallocator
>  99/115 Test  #99: testdebugallocator .....................   Passed    0.00 
> sec
>         Start 100: testdebugallocator_fail1
> 100/115 Test #100: testdebugallocator_fail1 ...............   Passed    0.00 
> sec
>         Start 101: testdebugallocator_fail2
> 101/115 Test #101: testdebugallocator_fail2 ...............   Passed    0.00 
> sec
>         Start 102: testdebugallocator_fail3
> 102/115 Test #102: testdebugallocator_fail3 ...............   Passed    0.00 
> sec
>         Start 103: testdebugallocator_fail4
> 103/115 Test #103: testdebugallocator_fail4 ...............   Passed    0.00 
> sec
>         Start 104: testdebugallocator_fail5
> 104/115 Test #104: testdebugallocator_fail5 ...............   Passed    0.00 
> sec
>         Start 105: testfloatcmp
> 105/115 Test #105: testfloatcmp ...........................   Passed    0.00 
> sec
>         Start 106: transposetest
> 106/115 Test #106: transposetest ..........................   Passed    0.00 
> sec
>         Start 107: tupleutilitytest
> 107/115 Test #107: tupleutilitytest .......................   Passed    0.00 
> sec
>         Start 108: typeutilitytest
> 108/115 Test #108: typeutilitytest ........................   Passed    0.00 
> sec
>         Start 109: typelisttest
> 109/115 Test #109: typelisttest ...........................   Passed    0.00 
> sec
>         Start 110: utilitytest
> 110/115 Test #110: utilitytest ............................   Passed    0.00 
> sec
>         Start 111: eigenvaluestest
> 111/115 Test #111: eigenvaluestest ........................   Passed    0.66 
> sec
>         Start 112: versiontest
> 112/115 Test #112: versiontest ............................   Passed    0.00 
> sec
>         Start 113: mathtest
> 113/115 Test #113: mathtest ...............................   Passed    0.00 
> sec
>         Start 114: vcexpectedimpltest
> 114/115 Test #114: vcexpectedimpltest .....................***Skipped   0.00 
> sec
>         Start 115: alignedallocatortest
> 115/115 Test #115: alignedallocatortest ...................   Passed    0.00 
> sec
> 
> 76% tests passed, 28 tests failed out of 115
> 
> Label Time Summary:
> quick    =  13.46 sec*proc (107 tests)
> 
> Total Test time (real) =  13.35 sec
> 
> The following tests did not run:
>        18 - vcarraytest (Skipped)
>        19 - vcvectortest (Skipped)
>        69 - metistest (Skipped)
>        80 - parametertreelocaletest (Skipped)
>        85 - quadmathtest (Skipped)
>        89 - scotchtest (Skipped)
>       114 - vcexpectedimpltest (Skipped)
> 
> The following tests FAILED:
>         1 - communicationtest (Failed)
>         2 - communicationtest-mpi-2 (Failed)
>         4 - remoteindicestest (Failed)
>         5 - remoteindicestest-mpi-2 (Failed)
>         7 - syncertest (Failed)
>         8 - syncertest-mpi-2 (Failed)
>         9 - variablesizecommunicatortest (Failed)
>        10 - variablesizecommunicatortest-mpi-2 (Failed)
>        11 - mpidatatest-mpi-2 (Failed)
>        12 - mpifuturetest (Failed)
>        13 - mpifuturetest-mpi-2 (Failed)
>        14 - mpipacktest-mpi-2 (Failed)
>        15 - mpigatherscattertest-mpi-2 (Failed)
>        20 - arithmetictestsuitetest (Failed)
>        23 - assertandreturntest (Failed)
>        25 - assertandreturntest_ndebug (Failed)
>        38 - concept (Failed)
>        40 - debugaligntest (Failed)
>        41 - debugalignsimdtest (Failed)
>        70 - mpicommunicationtest (Failed)
>        71 - mpicommunicationtest-mpi-2 (Failed)
>        72 - mpiguardtest (Failed)
>        73 - mpiguardtest-mpi-2 (Failed)
>        74 - mpihelpertest (Failed)
>        75 - mpihelpertest-mpi-2 (Failed)
>        76 - mpihelpertest2 (Failed)
>        77 - mpihelpertest2-mpi-2 (Failed)
>        87 - referencehelpertest (Failed)
> Errors while running CTest
> ======================================================================
> Name:      communicationtest
> FullName:  ./dune/common/parallel/test/communicationtest
> Status:    FAILED
> 
> ======================================================================
> Name:      communicationtest-mpi-2
> FullName:  ./dune/common/parallel/test/communicationtest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      remoteindicestest
> FullName:  ./dune/common/parallel/test/remoteindicestest
> Status:    FAILED
> 
> ======================================================================
> Name:      remoteindicestest-mpi-2
> FullName:  ./dune/common/parallel/test/remoteindicestest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      syncertest
> FullName:  ./dune/common/parallel/test/syncertest
> Status:    FAILED
> 
> ======================================================================
> Name:      syncertest-mpi-2
> FullName:  ./dune/common/parallel/test/syncertest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      variablesizecommunicatortest
> FullName:  ./dune/common/parallel/test/variablesizecommunicatortest
> Status:    FAILED
> 
> ======================================================================
> Name:      variablesizecommunicatortest-mpi-2
> FullName:  ./dune/common/parallel/test/variablesizecommunicatortest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpidatatest-mpi-2
> FullName:  ./dune/common/parallel/test/mpidatatest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpifuturetest
> FullName:  ./dune/common/parallel/test/mpifuturetest
> Status:    FAILED
> 
> ======================================================================
> Name:      mpifuturetest-mpi-2
> FullName:  ./dune/common/parallel/test/mpifuturetest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpipacktest-mpi-2
> FullName:  ./dune/common/parallel/test/mpipacktest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpigatherscattertest-mpi-2
> FullName:  ./dune/common/parallel/test/mpigatherscattertest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      arithmetictestsuitetest
> FullName:  ./dune/common/test/arithmetictestsuitetest
> Status:    FAILED
> 
> ======================================================================
> Name:      assertandreturntest
> FullName:  ./dune/common/test/assertandreturntest
> Status:    FAILED
> 
> ======================================================================
> Name:      assertandreturntest_ndebug
> FullName:  ./dune/common/test/assertandreturntest_ndebug
> Status:    FAILED
> 
> ======================================================================
> Name:      concept
> FullName:  ./dune/common/test/concept
> Status:    FAILED
> 
> ======================================================================
> Name:      debugaligntest
> FullName:  ./dune/common/test/debugaligntest
> Status:    FAILED
> 
> ======================================================================
> Name:      debugalignsimdtest
> FullName:  ./dune/common/test/debugalignsimdtest
> Status:    FAILED
> 
> ======================================================================
> Name:      mpicommunicationtest
> FullName:  ./dune/common/test/mpicommunicationtest
> Status:    FAILED
> 
> ======================================================================
> Name:      mpicommunicationtest-mpi-2
> FullName:  ./dune/common/test/mpicommunicationtest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpiguardtest
> FullName:  ./dune/common/test/mpiguardtest
> Status:    FAILED
> 
> ======================================================================
> Name:      mpiguardtest-mpi-2
> FullName:  ./dune/common/test/mpiguardtest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpihelpertest
> FullName:  ./dune/common/test/mpihelpertest
> Status:    FAILED
> 
> ======================================================================
> Name:      mpihelpertest-mpi-2
> FullName:  ./dune/common/test/mpihelpertest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpihelpertest2
> FullName:  ./dune/common/test/mpihelpertest2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpihelpertest2-mpi-2
> FullName:  ./dune/common/test/mpihelpertest2-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      referencehelpertest
> FullName:  ./dune/common/test/referencehelpertest
> Status:    FAILED
> 
> JUnit report for CTest results written to 
> /<<PKGBUILDDIR>>/build/junit/cmake.xml
> make[1]: *** [debian/dune-debian.mk:39: override_dh_auto_test] Error 1


The full build log is available from:
http://qa-logs.debian.net/2024/04/20/dune-common_2.9.0-6_unstable-armhf.log

All bugs filed during this archive rebuild are listed at:
https://bugs.debian.org/cgi-bin/pkgreport.cgi?tag=ftbfs-20240420;users=lu...@debian.org
or:
https://udd.debian.org/bugs/?release=na&merged=ign&fnewerval=7&flastmodval=7&fusertag=only&fusertagtag=ftbfs-20240420&fusertaguser=lu...@debian.org&allbugs=1&cseverity=1&ctags=1&caffected=1#results

A list of current common problems and possible solutions is available at
http://wiki.debian.org/qa.debian.org/FTBFS . You're welcome to contribute!

If you reassign this bug to another package, please mark it as 'affects'-ing
this package. See https://www.debian.org/Bugs/server-control#affects

If you fail to reproduce this, please provide a build log and diff it with mine
so that we can identify if something relevant changed in the meantime.

--- End Message ---
--- Begin Message ---
Hi,

this bug is actually the same as  bug 1069433 [1] in src:openmpi. That bug has 
been fixed with the upload of version 4.1.6-13 to unstable.

Best,

Markus

[1] https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=1069433

--- End Message ---

Reply via email to