Bug#1069492: dune-common: FTBFS on armhf: tests fail
Lucas Nussbaum
lucas at debian.org
Sat Apr 20 13:50:06 BST 2024
Source: dune-common
Version: 2.9.0-6
Severity: serious
Justification: FTBFS
Tags: trixie sid ftbfs
User: lucas at debian.org
Usertags: ftbfs-20240420 ftbfs-trixie ftbfs-t64-armhf
Hi,
During a rebuild of all packages in sid, your package failed to build
on armhf.
Relevant part (hopefully):
> make[5]: Entering directory '/<<PKGBUILDDIR>>/build'
> make[5]: Nothing to be done for 'CMakeFiles/build_tests.dir/build'.
> make[5]: Leaving directory '/<<PKGBUILDDIR>>/build'
> [100%] Built target build_tests
> make[4]: Leaving directory '/<<PKGBUILDDIR>>/build'
> /usr/bin/cmake -E cmake_progress_start /<<PKGBUILDDIR>>/build/CMakeFiles 0
> make[3]: Leaving directory '/<<PKGBUILDDIR>>/build'
> make[2]: Leaving directory '/<<PKGBUILDDIR>>/build'
> cd build; PATH=/<<PKGBUILDDIR>>/debian/tmp-test:$PATH /<<PKGBUILDDIR>>/bin/dune-ctest
> Site: ip-10-84-234-171
> Build name: Linux-c++
> Create new tag: 20240420-0332 - Experimental
> Test project /<<PKGBUILDDIR>>/build
> Start 1: communicationtest
> 1/115 Test #1: communicationtest ......................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976535] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976534] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976534] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976534] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 2: communicationtest-mpi-2
> 2/115 Test #2: communicationtest-mpi-2 ................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976536] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 3: indexsettest
> 3/115 Test #3: indexsettest ........................... Passed 0.00 sec
> Start 4: remoteindicestest
> 4/115 Test #4: remoteindicestest ......................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976539] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976538] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976538] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976538] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 5: remoteindicestest-mpi-2
> 5/115 Test #5: remoteindicestest-mpi-2 ................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976540] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 6: selectiontest
> 6/115 Test #6: selectiontest .......................... Passed 0.17 sec
> Start 7: syncertest
> 7/115 Test #7: syncertest .............................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976543] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976542] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976542] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976542] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 8: syncertest-mpi-2
> 8/115 Test #8: syncertest-mpi-2 .......................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976544] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 9: variablesizecommunicatortest
> 9/115 Test #9: variablesizecommunicatortest ...........***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976546] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976545] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976545] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976545] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 10: variablesizecommunicatortest-mpi-2
> 10/115 Test #10: variablesizecommunicatortest-mpi-2 .....***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976547] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 11: mpidatatest-mpi-2
> 11/115 Test #11: mpidatatest-mpi-2 ......................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976548] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 12: mpifuturetest
> 12/115 Test #12: mpifuturetest ..........................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976550] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976549] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976549] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976549] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 13: mpifuturetest-mpi-2
> 13/115 Test #13: mpifuturetest-mpi-2 ....................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976551] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 14: mpipacktest-mpi-2
> 14/115 Test #14: mpipacktest-mpi-2 ......................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976552] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 15: mpigatherscattertest-mpi-2
> 15/115 Test #15: mpigatherscattertest-mpi-2 .............***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976553] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 16: looptest
> 16/115 Test #16: looptest ............................... Passed 0.03 sec
> Start 17: standardtest
> 17/115 Test #17: standardtest ........................... Passed 0.00 sec
> Start 18: vcarraytest
> 18/115 Test #18: vcarraytest ............................***Skipped 0.00 sec
> Start 19: vcvectortest
> 19/115 Test #19: vcvectortest ...........................***Skipped 0.00 sec
> Start 20: arithmetictestsuitetest
> 20/115 Test #20: arithmetictestsuitetest ................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976559] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976558] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976558] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976558] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 21: arraylisttest
> 21/115 Test #21: arraylisttest .......................... Passed 0.00 sec
> Start 22: arraytest
> 22/115 Test #22: arraytest .............................. Passed 0.00 sec
> Start 23: assertandreturntest
> 23/115 Test #23: assertandreturntest ....................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976563] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976562] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976562] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976562] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 24: assertandreturntest_compiletime_fail
> 24/115 Test #24: assertandreturntest_compiletime_fail ... Passed 1.08 sec
> Start 25: assertandreturntest_ndebug
> 25/115 Test #25: assertandreturntest_ndebug .............***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976588] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976587] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976587] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976587] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 26: autocopytest
> 26/115 Test #26: autocopytest ........................... Passed 0.00 sec
> Start 27: bigunsignedinttest
> 27/115 Test #27: bigunsignedinttest ..................... Passed 0.00 sec
> Start 28: bitsetvectortest
> 28/115 Test #28: bitsetvectortest ....................... Passed 0.00 sec
> Start 29: boundscheckingtest
> 29/115 Test #29: boundscheckingtest ..................... Passed 0.00 sec
> Start 30: boundscheckingmvtest
> 30/115 Test #30: boundscheckingmvtest ................... Passed 0.00 sec
> Start 31: boundscheckingoptest
> 31/115 Test #31: boundscheckingoptest ................... Passed 0.00 sec
> Start 32: calloncetest
> 32/115 Test #32: calloncetest ........................... Passed 0.00 sec
> Start 33: check_fvector_size
> 33/115 Test #33: check_fvector_size ..................... Passed 0.00 sec
> Start 34: check_fvector_size_fail1
> 34/115 Test #34: check_fvector_size_fail1 ............... Passed 0.72 sec
> Start 35: check_fvector_size_fail2
> 35/115 Test #35: check_fvector_size_fail2 ............... Passed 0.73 sec
> Start 36: classnametest-demangled
> 36/115 Test #36: classnametest-demangled ................ Passed 0.01 sec
> Start 37: classnametest-fallback
> 37/115 Test #37: classnametest-fallback ................. Passed 0.01 sec
> Start 38: concept
> 38/115 Test #38: concept ................................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976634] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976633] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976633] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976633] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 39: constexprifelsetest
> 39/115 Test #39: constexprifelsetest .................... Passed 0.00 sec
> Start 40: debugaligntest
> 40/115 Test #40: debugaligntest .........................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976637] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976636] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976636] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976636] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 41: debugalignsimdtest
> 41/115 Test #41: debugalignsimdtest .....................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976639] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976638] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976638] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976638] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 42: densematrixassignmenttest
> 42/115 Test #42: densematrixassignmenttest .............. Passed 0.00 sec
> Start 43: densematrixassignmenttest_fail0
> 43/115 Test #43: densematrixassignmenttest_fail0 ........ Passed 1.21 sec
> Start 44: densematrixassignmenttest_fail1
> 44/115 Test #44: densematrixassignmenttest_fail1 ........ Passed 1.20 sec
> Start 45: densematrixassignmenttest_fail2
> 45/115 Test #45: densematrixassignmenttest_fail2 ........ Passed 1.20 sec
> Start 46: densematrixassignmenttest_fail3
> 46/115 Test #46: densematrixassignmenttest_fail3 ........ Passed 1.21 sec
> Start 47: densematrixassignmenttest_fail4
> 47/115 Test #47: densematrixassignmenttest_fail4 ........ Passed 1.22 sec
> Start 48: densematrixassignmenttest_fail5
> 48/115 Test #48: densematrixassignmenttest_fail5 ........ Passed 1.23 sec
> Start 49: densematrixassignmenttest_fail6
> 49/115 Test #49: densematrixassignmenttest_fail6 ........ Passed 1.21 sec
> Start 50: densevectorassignmenttest
> 50/115 Test #50: densevectorassignmenttest .............. Passed 0.00 sec
> Start 51: diagonalmatrixtest
> 51/115 Test #51: diagonalmatrixtest ..................... Passed 0.00 sec
> Start 52: dynmatrixtest
> 52/115 Test #52: dynmatrixtest .......................... Passed 0.00 sec
> Start 53: dynvectortest
> 53/115 Test #53: dynvectortest .......................... Passed 0.00 sec
> Start 54: densevectortest
> 54/115 Test #54: densevectortest ........................ Passed 0.00 sec
> Start 55: enumsettest
> 55/115 Test #55: enumsettest ............................ Passed 0.00 sec
> Start 56: filledarraytest
> 56/115 Test #56: filledarraytest ........................ Passed 0.00 sec
> Start 57: fmatrixtest
> 57/115 Test #57: fmatrixtest ............................ Passed 0.00 sec
> Start 58: functiontest
> 58/115 Test #58: functiontest ........................... Passed 0.00 sec
> Start 59: fvectortest
> 59/115 Test #59: fvectortest ............................ Passed 0.00 sec
> Start 60: fvectorconversion1d
> 60/115 Test #60: fvectorconversion1d .................... Passed 0.00 sec
> Start 61: genericiterator_compile_fail
> 61/115 Test #61: genericiterator_compile_fail ........... Passed 0.76 sec
> Start 62: hybridutilitiestest
> 62/115 Test #62: hybridutilitiestest .................... Passed 0.00 sec
> Start 63: indicestest
> 63/115 Test #63: indicestest ............................ Passed 0.00 sec
> Start 64: iscallabletest
> 64/115 Test #64: iscallabletest ......................... Passed 0.00 sec
> Start 65: iteratorfacadetest2
> 65/115 Test #65: iteratorfacadetest2 .................... Passed 0.00 sec
> Start 66: iteratorfacadetest
> 66/115 Test #66: iteratorfacadetest ..................... Passed 0.00 sec
> Start 67: lrutest
> 67/115 Test #67: lrutest ................................ Passed 0.00 sec
> Start 68: mathclassifierstest
> 68/115 Test #68: mathclassifierstest .................... Passed 0.00 sec
> Start 69: metistest
> 69/115 Test #69: metistest ..............................***Skipped 0.00 sec
> Start 70: mpicommunicationtest
> 70/115 Test #70: mpicommunicationtest ...................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976839] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976838] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976838] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976838] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 71: mpicommunicationtest-mpi-2
> 71/115 Test #71: mpicommunicationtest-mpi-2 .............***Failed 0.01 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976840] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 72: mpiguardtest
> 72/115 Test #72: mpiguardtest ...........................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976842] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976841] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976841] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976841] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 73: mpiguardtest-mpi-2
> 73/115 Test #73: mpiguardtest-mpi-2 .....................***Failed 0.01 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976843] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 74: mpihelpertest
> 74/115 Test #74: mpihelpertest ..........................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976845] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976844] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976844] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976844] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 75: mpihelpertest-mpi-2
> 75/115 Test #75: mpihelpertest-mpi-2 ....................***Failed 0.01 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976846] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 76: mpihelpertest2
> 76/115 Test #76: mpihelpertest2 .........................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976848] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976847] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976847] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976847] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 77: mpihelpertest2-mpi-2
> 77/115 Test #77: mpihelpertest2-mpi-2 ...................***Failed 0.01 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976849] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
>
> Start 78: overloadsettest
> 78/115 Test #78: overloadsettest ........................ Passed 0.00 sec
> Start 79: parameterizedobjecttest
> 79/115 Test #79: parameterizedobjecttest ................ Passed 0.00 sec
> Start 80: parametertreelocaletest
> 80/115 Test #80: parametertreelocaletest ................***Skipped 0.00 sec
> Start 81: parametertreetest
> 81/115 Test #81: parametertreetest ...................... Passed 0.00 sec
> Start 82: pathtest
> 82/115 Test #82: pathtest ............................... Passed 0.00 sec
> Start 83: poolallocatortest
> 83/115 Test #83: poolallocatortest ...................... Passed 0.00 sec
> Start 84: powertest
> 84/115 Test #84: powertest .............................. Passed 0.00 sec
> Start 85: quadmathtest
> 85/115 Test #85: quadmathtest ...........................***Skipped 0.00 sec
> Start 86: rangeutilitiestest
> 86/115 Test #86: rangeutilitiestest ..................... Passed 0.00 sec
> Start 87: referencehelpertest
> 87/115 Test #87: referencehelpertest ....................***Failed 0.02 sec
> --------------------------------------------------------------------------
> Sorry! You were supposed to get help about:
> pmix_init:startup:internal-failure
> But I couldn't open the help file:
> /usr/share/pmix/help-pmix-runtime.txt: No such file or directory. Sorry!
> --------------------------------------------------------------------------
> [ip-10-84-234-171:1976860] PMIX ERROR: NOT-FOUND in file ../../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at line 237
> [ip-10-84-234-171:1976859] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-10-84-234-171:1976859] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems. This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>
> orte_ess_init failed
> --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>
> ompi_mpi_init: ompi_rte_init failed
> --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> *** and potentially your MPI job)
> [ip-10-84-234-171:1976859] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
>
> Start 88: reservedvectortest
> 88/115 Test #88: reservedvectortest ..................... Passed 0.00 sec
> Start 89: scotchtest
> 89/115 Test #89: scotchtest .............................***Skipped 0.00 sec
> Start 90: shared_ptrtest
> 90/115 Test #90: shared_ptrtest ......................... Passed 0.00 sec
> Start 91: singletontest
> 91/115 Test #91: singletontest .......................... Passed 0.00 sec
> Start 92: sllisttest
> 92/115 Test #92: sllisttest ............................. Passed 0.00 sec
> Start 93: stdidentity
> 93/115 Test #93: stdidentity ............................ Passed 0.00 sec
> Start 94: stdapplytest
> 94/115 Test #94: stdapplytest ........................... Passed 0.00 sec
> Start 95: stdchecktypes
> 95/115 Test #95: stdchecktypes .......................... Passed 0.00 sec
> Start 96: streamoperatorstest
> 96/115 Test #96: streamoperatorstest .................... Passed 0.00 sec
> Start 97: streamtest
> 97/115 Test #97: streamtest ............................. Passed 0.00 sec
> Start 98: stringutilitytest
> 98/115 Test #98: stringutilitytest ...................... Passed 0.00 sec
> Start 99: testdebugallocator
> 99/115 Test #99: testdebugallocator ..................... Passed 0.00 sec
> Start 100: testdebugallocator_fail1
> 100/115 Test #100: testdebugallocator_fail1 ............... Passed 0.00 sec
> Start 101: testdebugallocator_fail2
> 101/115 Test #101: testdebugallocator_fail2 ............... Passed 0.00 sec
> Start 102: testdebugallocator_fail3
> 102/115 Test #102: testdebugallocator_fail3 ............... Passed 0.00 sec
> Start 103: testdebugallocator_fail4
> 103/115 Test #103: testdebugallocator_fail4 ............... Passed 0.00 sec
> Start 104: testdebugallocator_fail5
> 104/115 Test #104: testdebugallocator_fail5 ............... Passed 0.00 sec
> Start 105: testfloatcmp
> 105/115 Test #105: testfloatcmp ........................... Passed 0.00 sec
> Start 106: transposetest
> 106/115 Test #106: transposetest .......................... Passed 0.00 sec
> Start 107: tupleutilitytest
> 107/115 Test #107: tupleutilitytest ....................... Passed 0.00 sec
> Start 108: typeutilitytest
> 108/115 Test #108: typeutilitytest ........................ Passed 0.00 sec
> Start 109: typelisttest
> 109/115 Test #109: typelisttest ........................... Passed 0.00 sec
> Start 110: utilitytest
> 110/115 Test #110: utilitytest ............................ Passed 0.00 sec
> Start 111: eigenvaluestest
> 111/115 Test #111: eigenvaluestest ........................ Passed 0.66 sec
> Start 112: versiontest
> 112/115 Test #112: versiontest ............................ Passed 0.00 sec
> Start 113: mathtest
> 113/115 Test #113: mathtest ............................... Passed 0.00 sec
> Start 114: vcexpectedimpltest
> 114/115 Test #114: vcexpectedimpltest .....................***Skipped 0.00 sec
> Start 115: alignedallocatortest
> 115/115 Test #115: alignedallocatortest ................... Passed 0.00 sec
>
> 76% tests passed, 28 tests failed out of 115
>
> Label Time Summary:
> quick = 13.46 sec*proc (107 tests)
>
> Total Test time (real) = 13.35 sec
>
> The following tests did not run:
> 18 - vcarraytest (Skipped)
> 19 - vcvectortest (Skipped)
> 69 - metistest (Skipped)
> 80 - parametertreelocaletest (Skipped)
> 85 - quadmathtest (Skipped)
> 89 - scotchtest (Skipped)
> 114 - vcexpectedimpltest (Skipped)
>
> The following tests FAILED:
> 1 - communicationtest (Failed)
> 2 - communicationtest-mpi-2 (Failed)
> 4 - remoteindicestest (Failed)
> 5 - remoteindicestest-mpi-2 (Failed)
> 7 - syncertest (Failed)
> 8 - syncertest-mpi-2 (Failed)
> 9 - variablesizecommunicatortest (Failed)
> 10 - variablesizecommunicatortest-mpi-2 (Failed)
> 11 - mpidatatest-mpi-2 (Failed)
> 12 - mpifuturetest (Failed)
> 13 - mpifuturetest-mpi-2 (Failed)
> 14 - mpipacktest-mpi-2 (Failed)
> 15 - mpigatherscattertest-mpi-2 (Failed)
> 20 - arithmetictestsuitetest (Failed)
> 23 - assertandreturntest (Failed)
> 25 - assertandreturntest_ndebug (Failed)
> 38 - concept (Failed)
> 40 - debugaligntest (Failed)
> 41 - debugalignsimdtest (Failed)
> 70 - mpicommunicationtest (Failed)
> 71 - mpicommunicationtest-mpi-2 (Failed)
> 72 - mpiguardtest (Failed)
> 73 - mpiguardtest-mpi-2 (Failed)
> 74 - mpihelpertest (Failed)
> 75 - mpihelpertest-mpi-2 (Failed)
> 76 - mpihelpertest2 (Failed)
> 77 - mpihelpertest2-mpi-2 (Failed)
> 87 - referencehelpertest (Failed)
> Errors while running CTest
> ======================================================================
> Name: communicationtest
> FullName: ./dune/common/parallel/test/communicationtest
> Status: FAILED
>
> ======================================================================
> Name: communicationtest-mpi-2
> FullName: ./dune/common/parallel/test/communicationtest-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: remoteindicestest
> FullName: ./dune/common/parallel/test/remoteindicestest
> Status: FAILED
>
> ======================================================================
> Name: remoteindicestest-mpi-2
> FullName: ./dune/common/parallel/test/remoteindicestest-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: syncertest
> FullName: ./dune/common/parallel/test/syncertest
> Status: FAILED
>
> ======================================================================
> Name: syncertest-mpi-2
> FullName: ./dune/common/parallel/test/syncertest-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: variablesizecommunicatortest
> FullName: ./dune/common/parallel/test/variablesizecommunicatortest
> Status: FAILED
>
> ======================================================================
> Name: variablesizecommunicatortest-mpi-2
> FullName: ./dune/common/parallel/test/variablesizecommunicatortest-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: mpidatatest-mpi-2
> FullName: ./dune/common/parallel/test/mpidatatest-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: mpifuturetest
> FullName: ./dune/common/parallel/test/mpifuturetest
> Status: FAILED
>
> ======================================================================
> Name: mpifuturetest-mpi-2
> FullName: ./dune/common/parallel/test/mpifuturetest-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: mpipacktest-mpi-2
> FullName: ./dune/common/parallel/test/mpipacktest-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: mpigatherscattertest-mpi-2
> FullName: ./dune/common/parallel/test/mpigatherscattertest-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: arithmetictestsuitetest
> FullName: ./dune/common/test/arithmetictestsuitetest
> Status: FAILED
>
> ======================================================================
> Name: assertandreturntest
> FullName: ./dune/common/test/assertandreturntest
> Status: FAILED
>
> ======================================================================
> Name: assertandreturntest_ndebug
> FullName: ./dune/common/test/assertandreturntest_ndebug
> Status: FAILED
>
> ======================================================================
> Name: concept
> FullName: ./dune/common/test/concept
> Status: FAILED
>
> ======================================================================
> Name: debugaligntest
> FullName: ./dune/common/test/debugaligntest
> Status: FAILED
>
> ======================================================================
> Name: debugalignsimdtest
> FullName: ./dune/common/test/debugalignsimdtest
> Status: FAILED
>
> ======================================================================
> Name: mpicommunicationtest
> FullName: ./dune/common/test/mpicommunicationtest
> Status: FAILED
>
> ======================================================================
> Name: mpicommunicationtest-mpi-2
> FullName: ./dune/common/test/mpicommunicationtest-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: mpiguardtest
> FullName: ./dune/common/test/mpiguardtest
> Status: FAILED
>
> ======================================================================
> Name: mpiguardtest-mpi-2
> FullName: ./dune/common/test/mpiguardtest-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: mpihelpertest
> FullName: ./dune/common/test/mpihelpertest
> Status: FAILED
>
> ======================================================================
> Name: mpihelpertest-mpi-2
> FullName: ./dune/common/test/mpihelpertest-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: mpihelpertest2
> FullName: ./dune/common/test/mpihelpertest2
> Status: FAILED
>
> ======================================================================
> Name: mpihelpertest2-mpi-2
> FullName: ./dune/common/test/mpihelpertest2-mpi-2
> Status: FAILED
>
> ======================================================================
> Name: referencehelpertest
> FullName: ./dune/common/test/referencehelpertest
> Status: FAILED
>
> JUnit report for CTest results written to /<<PKGBUILDDIR>>/build/junit/cmake.xml
> make[1]: *** [debian/dune-debian.mk:39: override_dh_auto_test] Error 1
The full build log is available from:
http://qa-logs.debian.net/2024/04/20/dune-common_2.9.0-6_unstable-armhf.log
All bugs filed during this archive rebuild are listed at:
https://bugs.debian.org/cgi-bin/pkgreport.cgi?tag=ftbfs-20240420;users=lucas@debian.org
or:
https://udd.debian.org/bugs/?release=na&merged=ign&fnewerval=7&flastmodval=7&fusertag=only&fusertagtag=ftbfs-20240420&fusertaguser=lucas@debian.org&allbugs=1&cseverity=1&ctags=1&caffected=1#results
A list of current common problems and possible solutions is available at
http://wiki.debian.org/qa.debian.org/FTBFS . You're welcome to contribute!
If you reassign this bug to another package, please mark it as 'affects'-ing
this package. See https://www.debian.org/Bugs/server-control#affects
If you fail to reproduce this, please provide a build log and diff it with mine
so that we can identify if something relevant changed in the meantime.
More information about the debian-science-maintainers
mailing list