Bug#978211: dune-common: FTBFS: tests failed

Lucas Nussbaum lucas at debian.org
Sat Dec 26 21:11:29 GMT 2020


Source: dune-common
Version: 2.7.0-5
Severity: serious
Justification: FTBFS on amd64
Tags: bullseye sid ftbfs
Usertags: ftbfs-20201226 ftbfs-bullseye

Hi,

During a rebuild of all packages in sid, your package failed to build
on amd64.

Relevant part (hopefully):
> make[5]: Entering directory '/<<PKGBUILDDIR>>/build'
> make[5]: Nothing to be done for 'CMakeFiles/build_tests.dir/build'.
> make[5]: Leaving directory '/<<PKGBUILDDIR>>/build'
> [100%] Built target build_tests
> make[4]: Leaving directory '/<<PKGBUILDDIR>>/build'
> /usr/bin/cmake -E cmake_progress_start /<<PKGBUILDDIR>>/build/CMakeFiles 0
> make[3]: Leaving directory '/<<PKGBUILDDIR>>/build'
> make[2]: Leaving directory '/<<PKGBUILDDIR>>/build'
> cd build; PATH=/<<PKGBUILDDIR>>/debian/tmp-test:$PATH /<<PKGBUILDDIR>>/bin/dune-ctest 
>    Site: ip-172-31-8-9
>    Build name: Linux-c++
> Create new tag: 20201226-1836 - Experimental
> Test project /<<PKGBUILDDIR>>/build
>         Start   1: indexsettest
>   1/112 Test   #1: indexsettest ...........................   Passed    0.00 sec
>         Start   2: remoteindicestest
>   2/112 Test   #2: remoteindicestest ......................***Failed    0.04 sec
> [ip-172-31-8-9:13404] [[33605,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13403] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13403] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13403] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start   3: remoteindicestest-mpi-2
>   3/112 Test   #3: remoteindicestest-mpi-2 ................***Failed    0.01 sec
> [ip-172-31-8-9:13405] [[33604,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> 
>         Start   4: selectiontest
>   4/112 Test   #4: selectiontest ..........................   Passed    0.17 sec
>         Start   5: syncertest
>   5/112 Test   #5: syncertest .............................***Failed    0.02 sec
> [ip-172-31-8-9:13408] [[33657,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13407] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13407] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13407] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start   6: syncertest-mpi-2
>   6/112 Test   #6: syncertest-mpi-2 .......................***Failed    0.01 sec
> [ip-172-31-8-9:13409] [[33656,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> 
>         Start   7: variablesizecommunicatortest
>   7/112 Test   #7: variablesizecommunicatortest ...........***Failed    0.02 sec
> [ip-172-31-8-9:13411] [[33658,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13410] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13410] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13410] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start   8: variablesizecommunicatortest-mpi-2
>   8/112 Test   #8: variablesizecommunicatortest-mpi-2 .....***Failed    0.01 sec
> [ip-172-31-8-9:13412] [[33661,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> 
>         Start   9: mpidatatest-mpi-2
>   9/112 Test   #9: mpidatatest-mpi-2 ......................***Failed    0.01 sec
> [ip-172-31-8-9:13413] [[33660,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> 
>         Start  10: mpifuturetest
>  10/112 Test  #10: mpifuturetest ..........................***Failed    0.02 sec
> [ip-172-31-8-9:13415] [[33662,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13414] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13414] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13414] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start  11: mpifuturetest-mpi-2
>  11/112 Test  #11: mpifuturetest-mpi-2 ....................***Failed    0.01 sec
> [ip-172-31-8-9:13416] [[33649,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> 
>         Start  12: mpipacktest-mpi-2
>  12/112 Test  #12: mpipacktest-mpi-2 ......................***Failed    0.01 sec
> [ip-172-31-8-9:13417] [[33648,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> 
>         Start  13: looptest
>  13/112 Test  #13: looptest ...............................   Passed    0.01 sec
>         Start  14: standardtest
>  14/112 Test  #14: standardtest ...........................   Passed    0.01 sec
>         Start  15: vcarraytest
>  15/112 Test  #15: vcarraytest ............................***Skipped   0.00 sec
>         Start  16: vcvectortest
>  16/112 Test  #16: vcvectortest ...........................***Skipped   0.00 sec
>         Start  17: arithmetictestsuitetest
>  17/112 Test  #17: arithmetictestsuitetest ................***Failed    0.02 sec
> [ip-172-31-8-9:13423] [[33654,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13422] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13422] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13422] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start  18: arraylisttest
>  18/112 Test  #18: arraylisttest ..........................   Passed    0.00 sec
>         Start  19: arraytest
>  19/112 Test  #19: arraytest ..............................   Passed    0.00 sec
>         Start  20: assertandreturntest
>  20/112 Test  #20: assertandreturntest ....................***Failed    0.02 sec
> [ip-172-31-8-9:13427] [[33642,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13426] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13426] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13426] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start  21: assertandreturntest_compiletime_fail
>  21/112 Test  #21: assertandreturntest_compiletime_fail ...   Passed    0.72 sec
>         Start  22: assertandreturntest_ndebug
>  22/112 Test  #22: assertandreturntest_ndebug .............***Failed    0.02 sec
> [ip-172-31-8-9:13452] [[33685,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13451] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13451] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13451] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start  23: autocopytest
>  23/112 Test  #23: autocopytest ...........................   Passed    0.00 sec
>         Start  24: bigunsignedinttest
>  24/112 Test  #24: bigunsignedinttest .....................   Passed    0.00 sec
>         Start  25: bitsetvectortest
>  25/112 Test  #25: bitsetvectortest .......................   Passed    0.00 sec
>         Start  26: boundscheckingtest
>  26/112 Test  #26: boundscheckingtest .....................   Passed    0.00 sec
>         Start  27: boundscheckingmvtest
>  27/112 Test  #27: boundscheckingmvtest ...................   Passed    0.00 sec
>         Start  28: boundscheckingoptest
>  28/112 Test  #28: boundscheckingoptest ...................   Passed    0.00 sec
>         Start  29: calloncetest
>  29/112 Test  #29: calloncetest ...........................   Passed    0.00 sec
>         Start  30: check_fvector_size
>  30/112 Test  #30: check_fvector_size .....................   Passed    0.00 sec
>         Start  31: check_fvector_size_fail1
>  31/112 Test  #31: check_fvector_size_fail1 ...............   Passed    0.61 sec
>         Start  32: check_fvector_size_fail2
>  32/112 Test  #32: check_fvector_size_fail2 ...............   Passed    0.61 sec
>         Start  33: classnametest-demangled
>  33/112 Test  #33: classnametest-demangled ................   Passed    0.01 sec
>         Start  34: classnametest-fallback
>  34/112 Test  #34: classnametest-fallback .................   Passed    0.01 sec
>         Start  35: concept
>  35/112 Test  #35: concept ................................***Failed    0.03 sec
> [ip-172-31-8-9:13499] [[33698,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13498] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13498] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13498] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start  36: constexprifelsetest
>  36/112 Test  #36: constexprifelsetest ....................   Passed    0.00 sec
>         Start  37: debugaligntest
>  37/112 Test  #37: debugaligntest .........................***Failed    0.02 sec
> [ip-172-31-8-9:13502] [[33703,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13501] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13501] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13501] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start  38: debugalignsimdtest
>  38/112 Test  #38: debugalignsimdtest .....................***Failed    0.02 sec
> [ip-172-31-8-9:13504] [[33753,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13503] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13503] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13503] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start  39: densematrixassignmenttest
>  39/112 Test  #39: densematrixassignmenttest ..............   Passed    0.00 sec
>         Start  40: densematrixassignmenttest_fail0
>  40/112 Test  #40: densematrixassignmenttest_fail0 ........   Passed    0.97 sec
>         Start  41: densematrixassignmenttest_fail1
>  41/112 Test  #41: densematrixassignmenttest_fail1 ........   Passed    0.97 sec
>         Start  42: densematrixassignmenttest_fail2
>  42/112 Test  #42: densematrixassignmenttest_fail2 ........   Passed    1.01 sec
>         Start  43: densematrixassignmenttest_fail3
>  43/112 Test  #43: densematrixassignmenttest_fail3 ........   Passed    1.02 sec
>         Start  44: densematrixassignmenttest_fail4
>  44/112 Test  #44: densematrixassignmenttest_fail4 ........   Passed    1.00 sec
>         Start  45: densematrixassignmenttest_fail5
>  45/112 Test  #45: densematrixassignmenttest_fail5 ........   Passed    1.03 sec
>         Start  46: densematrixassignmenttest_fail6
>  46/112 Test  #46: densematrixassignmenttest_fail6 ........   Passed    0.98 sec
>         Start  47: densevectorassignmenttest
>  47/112 Test  #47: densevectorassignmenttest ..............   Passed    0.00 sec
>         Start  48: diagonalmatrixtest
>  48/112 Test  #48: diagonalmatrixtest .....................   Passed    0.00 sec
>         Start  49: dynmatrixtest
>  49/112 Test  #49: dynmatrixtest ..........................   Passed    0.00 sec
>         Start  50: dynvectortest
>  50/112 Test  #50: dynvectortest ..........................   Passed    0.00 sec
>         Start  51: densevectortest
>  51/112 Test  #51: densevectortest ........................   Passed    0.00 sec
>         Start  52: enumsettest
>  52/112 Test  #52: enumsettest ............................   Passed    0.00 sec
>         Start  53: filledarraytest
>  53/112 Test  #53: filledarraytest ........................   Passed    0.00 sec
>         Start  54: fmatrixtest
>  54/112 Test  #54: fmatrixtest ............................   Passed    0.00 sec
>         Start  55: functiontest
>  55/112 Test  #55: functiontest ...........................   Passed    0.00 sec
>         Start  56: fvectortest
>  56/112 Test  #56: fvectortest ............................   Passed    0.00 sec
>         Start  57: fvectorconversion1d
>  57/112 Test  #57: fvectorconversion1d ....................   Passed    0.00 sec
>         Start  58: genericiterator_compile_fail
>  58/112 Test  #58: genericiterator_compile_fail ...........   Passed    0.64 sec
>         Start  59: gcdlcmtest
>  59/112 Test  #59: gcdlcmtest .............................   Passed    0.00 sec
>         Start  60: hybridutilitiestest
>  60/112 Test  #60: hybridutilitiestest ....................   Passed    0.00 sec
>         Start  61: indicestest
>  61/112 Test  #61: indicestest ............................   Passed    0.00 sec
>         Start  62: integersequence
>  62/112 Test  #62: integersequence ........................   Passed    0.00 sec
>         Start  63: iteratorfacadetest2
>  63/112 Test  #63: iteratorfacadetest2 ....................   Passed    0.00 sec
>         Start  64: iteratorfacadetest
>  64/112 Test  #64: iteratorfacadetest .....................   Passed    0.00 sec
>         Start  65: lrutest
>  65/112 Test  #65: lrutest ................................   Passed    0.00 sec
>         Start  66: mathclassifierstest
>  66/112 Test  #66: mathclassifierstest ....................   Passed    0.00 sec
>         Start  67: mpicommunicationtest
>  67/112 Test  #67: mpicommunicationtest ...................***Failed    0.04 sec
> [ip-172-31-8-9:13704] [[33425,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13703] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13703] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13703] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start  68: mpicommunicationtest-mpi-2
>  68/112 Test  #68: mpicommunicationtest-mpi-2 .............***Failed    0.01 sec
> [ip-172-31-8-9:13705] [[33424,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> 
>         Start  69: mpiguardtest
>  69/112 Test  #69: mpiguardtest ...........................***Failed    0.02 sec
> [ip-172-31-8-9:13707] [[33426,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13706] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13706] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13706] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start  70: mpiguardtest-mpi-2
>  70/112 Test  #70: mpiguardtest-mpi-2 .....................***Failed    0.01 sec
> [ip-172-31-8-9:13708] [[33429,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> 
>         Start  71: mpihelpertest
>  71/112 Test  #71: mpihelpertest ..........................***Failed    0.02 sec
> [ip-172-31-8-9:13710] [[33431,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13709] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13709] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13709] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start  72: mpihelpertest-mpi-2
>  72/112 Test  #72: mpihelpertest-mpi-2 ....................***Failed    0.01 sec
> [ip-172-31-8-9:13711] [[33430,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> 
>         Start  73: mpihelpertest2
>  73/112 Test  #73: mpihelpertest2 .........................***Failed    0.02 sec
> [ip-172-31-8-9:13713] [[33416,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13712] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13712] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13712] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start  74: mpihelpertest2-mpi-2
>  74/112 Test  #74: mpihelpertest2-mpi-2 ...................***Failed    0.01 sec
> [ip-172-31-8-9:13714] [[33419,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> 
>         Start  75: overloadsettest
>  75/112 Test  #75: overloadsettest ........................   Passed    0.00 sec
>         Start  76: parameterizedobjecttest
>  76/112 Test  #76: parameterizedobjecttest ................   Passed    0.00 sec
>         Start  77: parametertreelocaletest
>  77/112 Test  #77: parametertreelocaletest ................***Skipped   0.00 sec
>         Start  78: parametertreetest
>  78/112 Test  #78: parametertreetest ......................   Passed    0.00 sec
>         Start  79: pathtest
>  79/112 Test  #79: pathtest ...............................   Passed    0.00 sec
>         Start  80: poolallocatortest
>  80/112 Test  #80: poolallocatortest ......................   Passed    0.00 sec
>         Start  81: powertest
>  81/112 Test  #81: powertest ..............................   Passed    0.00 sec
>         Start  82: quadmathtest
>  82/112 Test  #82: quadmathtest ...........................   Passed    0.00 sec
>         Start  83: rangeutilitiestest
>  83/112 Test  #83: rangeutilitiestest .....................   Passed    0.00 sec
>         Start  84: reservedvectortest
>  84/112 Test  #84: reservedvectortest .....................   Passed    0.00 sec
>         Start  85: shared_ptrtest
>  85/112 Test  #85: shared_ptrtest .........................   Passed    0.00 sec
>         Start  86: singletontest
>  86/112 Test  #86: singletontest ..........................   Passed    0.00 sec
>         Start  87: sllisttest
>  87/112 Test  #87: sllisttest .............................   Passed    0.00 sec
>         Start  88: stdidentity
>  88/112 Test  #88: stdidentity ............................   Passed    0.00 sec
>         Start  89: stdapplytest
>  89/112 Test  #89: stdapplytest ...........................   Passed    0.00 sec
>         Start  90: stdtypetraitstest
>  90/112 Test  #90: stdtypetraitstest ......................   Passed    0.00 sec
>         Start  91: streamoperatorstest
>  91/112 Test  #91: streamoperatorstest ....................   Passed    0.00 sec
>         Start  92: streamtest
>  92/112 Test  #92: streamtest .............................   Passed    0.00 sec
>         Start  93: stringutilitytest
>  93/112 Test  #93: stringutilitytest ......................   Passed    0.00 sec
>         Start  94: testdebugallocator
>  94/112 Test  #94: testdebugallocator .....................   Passed    0.00 sec
>         Start  95: testdebugallocator_fail1
>  95/112 Test  #95: testdebugallocator_fail1 ...............   Passed    0.00 sec
>         Start  96: testdebugallocator_fail2
>  96/112 Test  #96: testdebugallocator_fail2 ...............   Passed    0.00 sec
>         Start  97: testdebugallocator_fail3
>  97/112 Test  #97: testdebugallocator_fail3 ...............   Passed    0.00 sec
>         Start  98: testdebugallocator_fail4
>  98/112 Test  #98: testdebugallocator_fail4 ...............   Passed    0.00 sec
>         Start  99: testdebugallocator_fail5
>  99/112 Test  #99: testdebugallocator_fail5 ...............   Passed    0.00 sec
>         Start 100: testfloatcmp
> 100/112 Test #100: testfloatcmp ...........................   Passed    0.00 sec
>         Start 101: to_unique_ptrtest
> 101/112 Test #101: to_unique_ptrtest ......................   Passed    0.00 sec
>         Start 102: tupleutilitytest
> 102/112 Test #102: tupleutilitytest .......................   Passed    0.00 sec
>         Start 103: typeutilitytest
> 103/112 Test #103: typeutilitytest ........................   Passed    0.00 sec
>         Start 104: typelisttest
> 104/112 Test #104: typelisttest ...........................   Passed    0.00 sec
>         Start 105: utilitytest
> 105/112 Test #105: utilitytest ............................   Passed    0.00 sec
>         Start 106: eigenvaluestest
> 106/112 Test #106: eigenvaluestest ........................   Passed    0.00 sec
>         Start 107: optionaltest
> 107/112 Test #107: optionaltest ...........................   Passed    0.00 sec
>         Start 108: versiontest
> 108/112 Test #108: versiontest ............................   Passed    0.00 sec
>         Start 109: mathtest
> 109/112 Test #109: mathtest ...............................   Passed    0.00 sec
>         Start 110: varianttest
> 110/112 Test #110: varianttest ............................***Failed    0.02 sec
> [ip-172-31-8-9:13751] [[33454,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   opal_pmix_base_select failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> [ip-172-31-8-9:13750] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 716
> [ip-172-31-8-9:13750] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ../../../../../../orte/mca/ess/singleton/ess_singleton_module.c at line 172
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>   orte_ess_init failed
>   --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> 
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [ip-172-31-8-9:13750] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
> 
>         Start 111: vcexpectedimpltest
> 111/112 Test #111: vcexpectedimpltest .....................***Skipped   0.00 sec
>         Start 112: alignedallocatortest
> 112/112 Test #112: alignedallocatortest ...................   Passed    0.00 sec
> 
> 78% tests passed, 25 tests failed out of 112
> 
> Label Time Summary:
> quick    =  10.46 sec*proc (105 tests)
> 
> Total Test time (real) =  10.39 sec
> 
> The following tests did not run:
> 	 15 - vcarraytest (Skipped)
> 	 16 - vcvectortest (Skipped)
> 	 77 - parametertreelocaletest (Skipped)
> 	111 - vcexpectedimpltest (Skipped)
> 
> The following tests FAILED:
> 	  2 - remoteindicestest (Failed)
> 	  3 - remoteindicestest-mpi-2 (Failed)
> 	  5 - syncertest (Failed)
> 	  6 - syncertest-mpi-2 (Failed)
> 	  7 - variablesizecommunicatortest (Failed)
> 	  8 - variablesizecommunicatortest-mpi-2 (Failed)
> 	  9 - mpidatatest-mpi-2 (Failed)
> 	 10 - mpifuturetest (Failed)
> 	 11 - mpifuturetest-mpi-2 (Failed)
> 	 12 - mpipacktest-mpi-2 (Failed)
> 	 17 - arithmetictestsuitetest (Failed)
> 	 20 - assertandreturntest (Failed)
> 	 22 - assertandreturntest_ndebug (Failed)
> 	 35 - concept (Failed)
> 	 37 - debugaligntest (Failed)
> 	 38 - debugalignsimdtest (Failed)
> 	 67 - mpicommunicationtest (Failed)
> 	 68 - mpicommunicationtest-mpi-2 (Failed)
> 	 69 - mpiguardtest (Failed)
> 	 70 - mpiguardtest-mpi-2 (Failed)
> 	 71 - mpihelpertest (Failed)
> 	 72 - mpihelpertest-mpi-2 (Failed)
> 	 73 - mpihelpertest2 (Failed)
> 	 74 - mpihelpertest2-mpi-2 (Failed)
> 	110 - varianttest (Failed)
> Errors while running CTest
> ======================================================================
> Name:      remoteindicestest
> FullName:  ./dune/common/parallel/test/remoteindicestest
> Status:    FAILED
> 
> ======================================================================
> Name:      remoteindicestest-mpi-2
> FullName:  ./dune/common/parallel/test/remoteindicestest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      syncertest
> FullName:  ./dune/common/parallel/test/syncertest
> Status:    FAILED
> 
> ======================================================================
> Name:      syncertest-mpi-2
> FullName:  ./dune/common/parallel/test/syncertest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      variablesizecommunicatortest
> FullName:  ./dune/common/parallel/test/variablesizecommunicatortest
> Status:    FAILED
> 
> ======================================================================
> Name:      variablesizecommunicatortest-mpi-2
> FullName:  ./dune/common/parallel/test/variablesizecommunicatortest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpidatatest-mpi-2
> FullName:  ./dune/common/parallel/test/mpidatatest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpifuturetest
> FullName:  ./dune/common/parallel/test/mpifuturetest
> Status:    FAILED
> 
> ======================================================================
> Name:      mpifuturetest-mpi-2
> FullName:  ./dune/common/parallel/test/mpifuturetest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpipacktest-mpi-2
> FullName:  ./dune/common/parallel/test/mpipacktest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      arithmetictestsuitetest
> FullName:  ./dune/common/test/arithmetictestsuitetest
> Status:    FAILED
> 
> ======================================================================
> Name:      assertandreturntest
> FullName:  ./dune/common/test/assertandreturntest
> Status:    FAILED
> 
> ======================================================================
> Name:      assertandreturntest_ndebug
> FullName:  ./dune/common/test/assertandreturntest_ndebug
> Status:    FAILED
> 
> ======================================================================
> Name:      concept
> FullName:  ./dune/common/test/concept
> Status:    FAILED
> 
> ======================================================================
> Name:      debugaligntest
> FullName:  ./dune/common/test/debugaligntest
> Status:    FAILED
> 
> ======================================================================
> Name:      debugalignsimdtest
> FullName:  ./dune/common/test/debugalignsimdtest
> Status:    FAILED
> 
> ======================================================================
> Name:      mpicommunicationtest
> FullName:  ./dune/common/test/mpicommunicationtest
> Status:    FAILED
> 
> ======================================================================
> Name:      mpicommunicationtest-mpi-2
> FullName:  ./dune/common/test/mpicommunicationtest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpiguardtest
> FullName:  ./dune/common/test/mpiguardtest
> Status:    FAILED
> 
> ======================================================================
> Name:      mpiguardtest-mpi-2
> FullName:  ./dune/common/test/mpiguardtest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpihelpertest
> FullName:  ./dune/common/test/mpihelpertest
> Status:    FAILED
> 
> ======================================================================
> Name:      mpihelpertest-mpi-2
> FullName:  ./dune/common/test/mpihelpertest-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpihelpertest2
> FullName:  ./dune/common/test/mpihelpertest2
> Status:    FAILED
> 
> ======================================================================
> Name:      mpihelpertest2-mpi-2
> FullName:  ./dune/common/test/mpihelpertest2-mpi-2
> Status:    FAILED
> 
> ======================================================================
> Name:      varianttest
> FullName:  ./dune/common/test/varianttest
> Status:    FAILED
> 
> JUnit report for CTest results written to /<<PKGBUILDDIR>>/build/junit/cmake.xml
> make[1]: *** [debian/dune-debian.mk:39: override_dh_auto_test] Error 1

The full build log is available from:
   http://qa-logs.debian.net/2020/12/26/dune-common_2.7.0-5_unstable.log

A list of current common problems and possible solutions is available at
http://wiki.debian.org/qa.debian.org/FTBFS . You're welcome to contribute!

If you reassign this bug to another package, please marking it as 'affects'-ing
this package. See https://www.debian.org/Bugs/server-control#affects

If you fail to reproduce this, please provide a build log and diff it with me
so that we can identify if something relevant changed in the meantime.

About the archive rebuild: The rebuild was done on EC2 VM instances from
Amazon Web Services, using a clean, minimal and up-to-date chroot. Every
failed build was retried once to eliminate random failures.



More information about the debian-science-maintainers mailing list