Bug#978260: xmds2: FTBFS: AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;

Lucas Nussbaum lucas at debian.org
Sat Dec 26 21:46:05 GMT 2020


Source: xmds2
Version: 3.0.0+dfsg-4
Severity: serious
Justification: FTBFS on amd64
Tags: bullseye sid ftbfs
Usertags: ftbfs-20201226 ftbfs-bullseye

Hi,

During a rebuild of all packages in sid, your package failed to build
on amd64.

Relevant part (hopefully):
> make[1]: Entering directory '/<<PKGBUILDDIR>>'
> debian/tests/run-tests -b
> xmds2 version 3.0.0 "Release the Kraken" (Debian package 3.0.0+dfsg-4)
> Copyright 2000-2019 Graham Dennis, Joseph Hope, Mattias Johnsson
>                     and the xmds team
> 
> Configuring for single-process scripts...
> Checking for 'g++' (C++ compiler)        : /usr/bin/g++ 
> Checking whether the compiler works      : yes 
> Checking that we have a C++ compiler     : yes 
> Checking whether we are cross-compiling  : no 
> Checking whether we can link to only static libraries : yes 
> Trying to make compiler optimise for this machine     : yes 
> Trying to make compiler tune for this machine         : yes 
> Checking for compiler flags -O3                       : yes 
> Checking for compiler flags -ffast-math               : yes 
> Checking for compiler flags -funroll-all-loops        : yes 
> Checking for compiler flags -fomit-frame-pointer      : yes 
> Checking for compiler flags -falign-loops             : yes 
> Checking for compiler flags -fstrict-aliasing         : yes 
> Checking for compiler flags -momit-leaf-frame-pointer : yes 
> Checking for cautious math flags                      : yes 
> Checking for Autovectorisation                        : yes 
> Checking for OpenMP                                   : yes 
> Checking for pthreads                                 : yes 
> Checking for compiler debug flags                     : yes 
> Checking for srandomdev                               : yes 
> Checking for /dev/urandom                             : yes 
> Checking for program 'h5cc'                           : /usr/bin/h5cc 
> Checking for HDF5 (static library)                    : yes 
> Checking for HDF5 High-level library (static library) : yes 
> Checking for header hdf5.h                            : yes 
> Checking for libxmds (static library)                 : no (will try dynamic library instead) 
> Checking for libxmds (dynamic library)                : no (it's optional anyway) 
> Checking for Intel's Vector Math Library (static library) : no (will try dynamic library instead) 
> Checking for Intel's Vector Math Library (dynamic library) : no (it's optional anyway) 
> Checking safer dSFMT compile flags                         : yes 
> Checking for Intel's Math Kernel Library (static library)  : no (will try dynamic library instead) 
> Checking for Intel's Math Kernel Library (dynamic library) : no (it's optional anyway) 
> Checking for ATLAS's CBLAS. (static library)               : no (will try dynamic library instead) 
> Checking for ATLAS's CBLAS. (dynamic library)              : yes 
> Checking for GSL (static library)                          : yes 
> Checking for FFTW3 (static library)                        : yes 
> Checking for single-precision FFTW3 (static library)       : yes 
> Checking for threading support in FFTW3 (static library)   : yes 
> Checking for OpenMP support in FFTW3 (static library)      : yes 
> Checking for threading support in single-precision FFTW3 (static library) : yes 
> Checking for OpenMP support in single-precision FFTW3 (static library)    : yes 
> 
> Configuring for MPI scripts...
> Checking for program 'mpic++'                                             : /usr/bin/mpic++ 
> Checking for 'g++' (C++ compiler)                                         : /usr/bin/mpic++ 
> Checking whether the compiler works                                       : yes 
> Checking that we have a C++ compiler                                      : yes 
> Checking whether we are cross-compiling                                   : no 
> Checking whether we can link to only static libraries                     : yes 
> Trying to make compiler optimise for this machine                         : yes 
> Trying to make compiler tune for this machine                             : yes 
> Checking for compiler flags -O3                                           : yes 
> Checking for compiler flags -ffast-math                                   : yes 
> Checking for compiler flags -funroll-all-loops                            : yes 
> Checking for compiler flags -fomit-frame-pointer                          : yes 
> Checking for compiler flags -falign-loops                                 : yes 
> Checking for compiler flags -fstrict-aliasing                             : yes 
> Checking for compiler flags -momit-leaf-frame-pointer                     : yes 
> Checking for cautious math flags                                          : yes 
> Checking for Autovectorisation                                            : yes 
> Checking for OpenMP                                                       : yes 
> Checking for pthreads                                                     : yes 
> Checking for compiler debug flags                                         : yes 
> Checking for srandomdev                                                   : yes 
> Checking for /dev/urandom                                                 : yes 
> Checking for program 'h5cc'                                               : /usr/bin/h5cc 
> Checking for HDF5 (static library)                                        : yes 
> Checking for HDF5 High-level library (static library)                     : yes 
> Checking for header hdf5.h                                                : yes 
> Checking for libxmds (static library)                                     : no (will try dynamic library instead) 
> Checking for libxmds (dynamic library)                                    : no (it's optional anyway) 
> Checking for Intel's Vector Math Library (static library)                 : no (will try dynamic library instead) 
> Checking for Intel's Vector Math Library (dynamic library)                : no (it's optional anyway) 
> Checking safer dSFMT compile flags                                        : yes 
> Checking for Intel's Math Kernel Library (static library)                 : no (will try dynamic library instead) 
> Checking for Intel's Math Kernel Library (dynamic library)                : no (it's optional anyway) 
> Checking for ATLAS's CBLAS. (static library)                              : no (will try dynamic library instead) 
> Checking for ATLAS's CBLAS. (dynamic library)                             : yes 
> Checking for GSL (static library)                                         : yes 
> Checking for FFTW3 (static library)                                       : yes 
> Checking for single-precision FFTW3 (static library)                      : yes 
> Checking for threading support in FFTW3 (static library)                  : yes 
> Checking for OpenMP support in FFTW3 (static library)                     : yes 
> Checking for threading support in single-precision FFTW3 (static library) : yes 
> Checking for OpenMP support in single-precision FFTW3 (static library)    : yes 
> Checking for FFTW3 with MPI (static library)                              : yes 
> Checking for single-precision FFTW3 with MPI (static library)             : yes 
> ('Config log saved to ', '/<<PKGBUILDDIR>>/debian/xmds-user-data/waf_configure/config.log')
> test_bug_adaptive_timestep_hang (__main__.main.<locals>.ScriptTestCase)
> integrators/bug_adaptive_timestep_hang.xmds ... ok
> test_vibstring_ark45 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_ark45.xmds ... ok
> test_vibstring_ark89 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_ark89.xmds ... ok
> test_vibstring_mm (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_mm.xmds ... ok
> test_vibstring_re (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_re.xmds ... ok
> test_vibstring_rk4 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_rk4.xmds ... ok
> test_vibstring_rk45 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_rk45.xmds ... ok
> test_vibstring_rk89 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_rk89.xmds ... ok
> test_vibstring_rk9 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_rk9.xmds ... ok
> test_vibstring_si (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_si.xmds ... ok
> test_RbGSdipoles (__main__.main.<locals>.ScriptTestCase)
> fast/RbGSdipoles.xmds ... FAIL
> test_bessel_cosine_evolution (__main__.main.<locals>.ScriptTestCase)
> fast/bessel_cosine_evolution.xmds ... ok
> test_bessel_cosine_groundstate (__main__.main.<locals>.ScriptTestCase)
> fast/bessel_cosine_groundstate.xmds ... ok
> test_cpc_example4 (__main__.main.<locals>.ScriptTestCase)
> fast/cpc_example4.xmds ... FAIL
> test_eigenvalues (__main__.main.<locals>.ScriptTestCase)
> fast/eigenvalues.xmds ... ok
> test_groundstate (__main__.main.<locals>.ScriptTestCase)
> fast/groundstate.xmds ... ok
> test_lorenz (__main__.main.<locals>.ScriptTestCase)
> fast/lorenz.xmds ... ok
> test_tla (__main__.main.<locals>.ScriptTestCase)
> fast/tla.xmds ... ok
> test_transverse_integration_in_vector_initialisation (__main__.main.<locals>.ScriptTestCase)
> fast/transverse_integration_in_vector_initialisation.xmds ... ok
> test_vibstring_circle_spectral (__main__.main.<locals>.ScriptTestCase)
> fast/vibstring_circle_spectral.xmds ... ok
> test_constant_complex_ip (__main__.main.<locals>.ScriptTestCase)
> operators/constant_complex_ip.xmds ... ok
> test_constant_complex_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/constant_complex_ip_2d.xmds ... ok
> test_constant_complex_separated_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/constant_complex_separated_ip_2d.xmds ... ok
> test_constant_double_ip (__main__.main.<locals>.ScriptTestCase)
> operators/constant_double_ip.xmds ... ok
> test_constant_ex (__main__.main.<locals>.ScriptTestCase)
> operators/constant_ex.xmds ... ok
> test_constant_ex_arbitrary_code (__main__.main.<locals>.ScriptTestCase)
> operators/constant_ex_arbitrary_code.xmds ... ok
> test_constant_ex_arbitrary_order (__main__.main.<locals>.ScriptTestCase)
> operators/constant_ex_arbitrary_order.xmds ... ok
> test_constant_ex_arbitrary_order2 (__main__.main.<locals>.ScriptTestCase)
> operators/constant_ex_arbitrary_order2.xmds ... ok
> test_constant_real_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/constant_real_ip_2d.xmds ... ok
> test_cross_propagation (__main__.main.<locals>.ScriptTestCase)
> operators/cross_propagation.xmds ... ok
> test_cross_propagation2 (__main__.main.<locals>.ScriptTestCase)
> operators/cross_propagation2.xmds ... ok
> test_cross_propagation_right (__main__.main.<locals>.ScriptTestCase)
> operators/cross_propagation_right.xmds ... ok
> test_cross_propagation_sic (__main__.main.<locals>.ScriptTestCase)
> operators/cross_propagation_sic.xmds ... ok
> test_cross_propagation_sic_right (__main__.main.<locals>.ScriptTestCase)
> operators/cross_propagation_sic_right.xmds ... ok
> test_highdimcrossprop (__main__.main.<locals>.ScriptTestCase)
> operators/highdimcrossprop.xmds ... ok
> test_nonconstant_complex_ip (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_complex_ip.xmds ... ok
> test_nonconstant_complex_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_complex_ip_2d.xmds ... ok
> test_nonconstant_complex_separated_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_complex_separated_ip_2d.xmds ... ok
> test_nonconstant_double_ip (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_double_ip.xmds ... ok
> test_nonconstant_ex (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_ex.xmds ... ok
> test_nonconstant_real_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_real_ip_2d.xmds ... ok
> test_bessel_neumann_wave_equation (__main__.main.<locals>.ScriptTestCase)
> transforms/bessel_neumann_wave_equation.xmds ... ok
> test_bessel_transform (__main__.main.<locals>.ScriptTestCase)
> transforms/bessel_transform.xmds ... ok
> test_bessel_transform_rectangular (__main__.main.<locals>.ScriptTestCase)
> transforms/bessel_transform_rectangular.xmds ... ok
> test_diffusion_bessel (__main__.main.<locals>.ScriptTestCase)
> transforms/diffusion_bessel.xmds ... ok
> test_diffusion_dst (__main__.main.<locals>.ScriptTestCase)
> transforms/diffusion_dst.xmds ... ok
> test_disc (__main__.main.<locals>.ScriptTestCase)
> transforms/disc.xmds ... ok
> test_hermitegauss_fourier (__main__.main.<locals>.ScriptTestCase)
> transforms/hermitegauss_fourier.xmds ... ok
> test_hermitegauss_fourier_loading (__main__.main.<locals>.ScriptTestCase)
> transforms/hermitegauss_fourier_loading.xmds ... ok
> test_hermitegauss_transform_2d (__main__.main.<locals>.ScriptTestCase)
> transforms/hermitegauss_transform_2d.xmds ... ok
> test_hermitegauss_transform_2d_chunked (__main__.main.<locals>.ScriptTestCase)
> transforms/hermitegauss_transform_2d_chunked.xmds ... ok
> test_spherical_ball (__main__.main.<locals>.ScriptTestCase)
> transforms/spherical_ball.xmds ... ok
> test_vibstring_dct (__main__.main.<locals>.ScriptTestCase)
> transforms/vibstring_dct.xmds ... ok
> test_diffusion_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/diffusion_mpi.xmds ... FAIL
> test_diffusion_mpi_chunked (__main__.main.<locals>.ScriptTestCase)
> mpi/diffusion_mpi_chunked.xmds ... FAIL
> test_eigenvalues (__main__.main.<locals>.ScriptTestCase)
> mpi/eigenvalues.xmds ... FAIL
> test_fibre_integer_dimensions_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/fibre_integer_dimensions_mpi.xmds ... FAIL
> test_hermitegauss_transform_2d_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/hermitegauss_transform_2d_mpi.xmds ... FAIL
> test_hermitegauss_transform_2d_mpi_small (__main__.main.<locals>.ScriptTestCase)
> mpi/hermitegauss_transform_2d_mpi_small.xmds ... FAIL
> test_kubo_adaptive_mpi_paths (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_adaptive_mpi_paths.xmds ... FAIL
> test_kubo_integer_dimensions_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_integer_dimensions_mpi.xmds ... FAIL
> test_kubo_mpi_paths (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_mpi_paths.xmds ... FAIL
> test_lorenz_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/lorenz_mpi.xmds ... FAIL
> test_mpi_dft (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft.xmds ... FAIL
> test_mpi_dft_hdf5 (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft_hdf5.xmds ... FAIL
> test_mpi_dft_small (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft_small.xmds ... FAIL
> test_mpi_forward_plan_bug (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_forward_plan_bug.xmds ... FAIL
> test_mpi_highdimcrossprop (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_highdimcrossprop.xmds ... FAIL
> test_partial_integration_computed_vector (__main__.main.<locals>.ScriptTestCase)
> mpi/partial_integration_computed_vector.xmds ... FAIL
> test_vibstring_dst_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_dst_mpi.xmds ... FAIL
> test_vibstring_dst_mpi_chunked (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_dst_mpi_chunked.xmds ... FAIL
> test_vibstring_mpi_aliases (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_mpi_aliases.xmds ... FAIL
> test_integer_dimensions (__main__.main.<locals>.ScriptTestCase)
> geometry/integer_dimensions.xmds ... ok
> test_integer_dimensions_reordered (__main__.main.<locals>.ScriptTestCase)
> geometry/integer_dimensions_reordered.xmds ... ok
> test_nonlocal_access_multiple_components (__main__.main.<locals>.ScriptTestCase)
> geometry/nonlocal_access_multiple_components.xmds ... ok
> test_nonlocal_edge_uniform_access (__main__.main.<locals>.ScriptTestCase)
> geometry/nonlocal_edge_uniform_access.xmds ... ok
> test_nonlocal_index_access (__main__.main.<locals>.ScriptTestCase)
> geometry/nonlocal_index_access.xmds ... ok
> test_nonlocal_negative_uniform_access (__main__.main.<locals>.ScriptTestCase)
> geometry/nonlocal_negative_uniform_access.xmds ... ok
> test_nonlocal_split_uniform_access (__main__.main.<locals>.ScriptTestCase)
> geometry/nonlocal_split_uniform_access.xmds ... ok
> test_bessel_cosine_stochastic_groundstate (__main__.main.<locals>.ScriptTestCase)
> stochastic/bessel_cosine_stochastic_groundstate.xmds ... ok
> test_double_precision_noise_tests (__main__.main.<locals>.ScriptTestCase)
> stochastic/double_precision_noise_tests.xmds ... ok
> test_dsfmt_single_precision (__main__.main.<locals>.ScriptTestCase)
> stochastic/dsfmt_single_precision.xmds ... ok
> test_fibre (__main__.main.<locals>.ScriptTestCase)
> stochastic/fibre.xmds ... ok
> test_fibre_with_correlation_functions (__main__.main.<locals>.ScriptTestCase)
> stochastic/fibre_with_correlation_functions.xmds ... ok
> test_kubo (__main__.main.<locals>.ScriptTestCase)
> stochastic/kubo.xmds ... ok
> test_kubo_fixedstep (__main__.main.<locals>.ScriptTestCase)
> stochastic/kubo_fixedstep.xmds ... ok
> test_photodetector (__main__.main.<locals>.ScriptTestCase)
> stochastic/photodetector.xmds ... ok
> test_photodetector_bessel (__main__.main.<locals>.ScriptTestCase)
> stochastic/photodetector_bessel.xmds ... ok
> test_photodetector_linear (__main__.main.<locals>.ScriptTestCase)
> stochastic/photodetector_linear.xmds ... ok
> test_wigner_cool_HO (__main__.main.<locals>.ScriptTestCase)
> stochastic/wigner_cool_HO.xmds ... ok
> test_bessel_cosine_groundstate (__main__.main.<locals>.ScriptTestCase)
> openmp/bessel_cosine_groundstate.xmds ... ok
> test_diffusion_openmp (__main__.main.<locals>.ScriptTestCase)
> openmp/diffusion_openmp.xmds ... ok
> test_diffusion_openmp_chunked (__main__.main.<locals>.ScriptTestCase)
> openmp/diffusion_openmp_chunked.xmds ... ok
> test_eigenvalues (__main__.main.<locals>.ScriptTestCase)
> openmp/eigenvalues.xmds ... ok
> test_hermitegauss_transform_2d_openmp (__main__.main.<locals>.ScriptTestCase)
> openmp/hermitegauss_transform_2d_openmp.xmds ... ok
> test_hermitegauss_transform_2d_openmp_small (__main__.main.<locals>.ScriptTestCase)
> openmp/hermitegauss_transform_2d_openmp_small.xmds ... ok
> test_kubo_integer_dimensions_openmp (__main__.main.<locals>.ScriptTestCase)
> openmp/kubo_integer_dimensions_openmp.xmds ... ok
> test_lorenz_openmp (__main__.main.<locals>.ScriptTestCase)
> openmp/lorenz_openmp.xmds ... ok
> test_openmp_dft (__main__.main.<locals>.ScriptTestCase)
> openmp/openmp_dft.xmds ... ok
> test_openmp_dft_hdf5 (__main__.main.<locals>.ScriptTestCase)
> openmp/openmp_dft_hdf5.xmds ... ok
> test_openmp_dft_small (__main__.main.<locals>.ScriptTestCase)
> openmp/openmp_dft_small.xmds ... ok
> test_openmp_highdimcrossprop (__main__.main.<locals>.ScriptTestCase)
> openmp/openmp_highdimcrossprop.xmds ... ok
> test_partial_integration_computed_vector (__main__.main.<locals>.ScriptTestCase)
> openmp/partial_integration_computed_vector.xmds ... ok
> test_vibstring_dst_openmp (__main__.main.<locals>.ScriptTestCase)
> openmp/vibstring_dst_openmp.xmds ... ok
> test_vibstring_dst_openmp_chunked (__main__.main.<locals>.ScriptTestCase)
> openmp/vibstring_dst_openmp_chunked.xmds ... ok
> test_breakpoints_hdf5 (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/breakpoints_hdf5.xmds ... ok
> test_integer_dimensions (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/integer_dimensions.xmds ... ok
> test_integer_dimensions_with_fixed_lattice (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/integer_dimensions_with_fixed_lattice.xmds ... ok
> test_integer_dimensions_with_runtime_lattice (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/integer_dimensions_with_runtime_lattice.xmds ... ok
> test_runtime_lattice_diffusion_dst (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_diffusion_dst.xmds ... ok
> test_runtime_lattice_initialisation_order (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_initialisation_order.xmds ... ok
> test_runtime_lattice_mpi_dft_small (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_mpi_dft_small.xmds ... FAIL
> test_runtime_lattice_nonlocal_split_uniform_access (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_nonlocal_split_uniform_access.xmds ... ok
> test_runtime_lattice_vibstring_ark89 (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_vibstring_ark89.xmds ... ok
> test_runtime_lattice_xsilloading_hdf5_loose2 (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_xsilloading_hdf5_loose2.xmds ... ok
> test_initialisation_order (__main__.main.<locals>.ScriptTestCase)
> vectors/initialisation_order.xmds ... ok
> test_initialisation_order_chunked (__main__.main.<locals>.ScriptTestCase)
> vectors/initialisation_order_chunked.xmds ... ok
> test_partial_integration_computed_vector (__main__.main.<locals>.ScriptTestCase)
> vectors/partial_integration_computed_vector.xmds ... ok
> test_breakpoints (__main__.main.<locals>.ScriptTestCase)
> io/breakpoints.xmds ... ok
> test_breakpoints_hdf5 (__main__.main.<locals>.ScriptTestCase)
> io/breakpoints_hdf5.xmds ... ok
> test_mpi_xsilloading_hdf5 (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5.xmds ... FAIL
> test_mpi_xsilloading_hdf5_loose (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5_loose.xmds ... FAIL
> test_mpi_xsilloading_hdf5_loose2 (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5_loose2.xmds ... FAIL
> test_nlse_sampling (__main__.main.<locals>.ScriptTestCase)
> io/nlse_sampling.xmds ... ok
> test_xsilloading_hdf5 (__main__.main.<locals>.ScriptTestCase)
> io/xsilloading_hdf5.xmds ... ok
> test_xsilloading_hdf5_loose (__main__.main.<locals>.ScriptTestCase)
> io/xsilloading_hdf5_loose.xmds ... ok
> test_xsilloading_hdf5_loose2 (__main__.main.<locals>.ScriptTestCase)
> io/xsilloading_hdf5_loose2.xmds ... ok
> test_arguments (__main__.main.<locals>.ScriptTestCase)
> features/arguments.xmds ... ok
> test_arguments_append_args_to_output_filename (__main__.main.<locals>.ScriptTestCase)
> features/arguments_append_args_to_output_filename.xmds ... ok
> test_arguments_with_similar_names (__main__.main.<locals>.ScriptTestCase)
> features/arguments_with_similar_names.xmds ... ok
> test_error_check_multipath (__main__.main.<locals>.ScriptTestCase)
> features/error_check_multipath.xmds ... ok
> test_halt_non_finite (__main__.main.<locals>.ScriptTestCase)
> features/halt_non_finite.xmds ... ok
> test_hermitegauss_transform_2d_chunked_breakpoints (__main__.main.<locals>.ScriptTestCase)
> features/hermitegauss_transform_2d_chunked_breakpoints.xmds ... ok
> test_hermitegauss_transform_2d_chunked_breakpoints_hdf5 (__main__.main.<locals>.ScriptTestCase)
> features/hermitegauss_transform_2d_chunked_breakpoints_hdf5.xmds ... ok
> test_realistic_Rb_and_fields (__main__.main.<locals>.ScriptTestCase)
> features/realistic_Rb_and_fields.xmds ... ok
> test_runtime_paths (__main__.main.<locals>.ScriptTestCase)
> features/runtime_paths.xmds ... ok
> test_space in filename (__main__.main.<locals>.ScriptTestCase)
> features/space in filename.xmds ... ok
> test_assignmentToIncorrectVariable (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_basic (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_combined (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_complicatedSafeOperation (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_hiddenUnsafeOperation (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_missingAssignment (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_realExample (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_safeBinaryOperation (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_safeSubtraction (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_unsafeBinaryOperation (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_unsafeSubtraction (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_unsafeUnaryOperation (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_doubleDivisionByInteger (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_floatDivision (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_ignoreComments (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_ignoreStrings (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_integerDivision (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_integerDivisionByDouble (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_symbolDivision (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_accessDifferentVariables (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_accessMultipleTimes (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_basic (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_combined (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_multipleAccess (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_notGreedy (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_withPrintf (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_combined (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_ignoreChildComment (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_ignoreSiblingComment (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_ignoreSiblingQuotedString (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_invalidSyntax (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_nestedOperators (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_notGreedy (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_unbalancedString (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_withPrintf (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> 
> ======================================================================
> FAIL: test_RbGSdipoles (__main__.main.<locals>.ScriptTestCase)
> fast/RbGSdipoles.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:11215] [[18447,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_cpc_example4 (__main__.main.<locals>.ScriptTestCase)
> fast/cpc_example4.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:11268] [[20420,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_diffusion_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/diffusion_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:11950] [[19822,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_diffusion_mpi_chunked (__main__.main.<locals>.ScriptTestCase)
> mpi/diffusion_mpi_chunked.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:11969] [[19713,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_eigenvalues (__main__.main.<locals>.ScriptTestCase)
> mpi/eigenvalues.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:11988] [[19732,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_fibre_integer_dimensions_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/fibre_integer_dimensions_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12007] [[19751,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_hermitegauss_transform_2d_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/hermitegauss_transform_2d_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12026] [[19770,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_hermitegauss_transform_2d_mpi_small (__main__.main.<locals>.ScriptTestCase)
> mpi/hermitegauss_transform_2d_mpi_small.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12045] [[19661,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_kubo_adaptive_mpi_paths (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_adaptive_mpi_paths.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12064] [[19680,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_kubo_integer_dimensions_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_integer_dimensions_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12083] [[19699,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_kubo_mpi_paths (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_mpi_paths.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12102] [[19590,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_lorenz_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/lorenz_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12121] [[19609,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_mpi_dft (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12140] [[19628,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_mpi_dft_hdf5 (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft_hdf5.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12159] [[19647,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_mpi_dft_small (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft_small.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12178] [[19538,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_mpi_forward_plan_bug (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_forward_plan_bug.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12197] [[19557,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_mpi_highdimcrossprop (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_highdimcrossprop.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12216] [[19576,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_partial_integration_computed_vector (__main__.main.<locals>.ScriptTestCase)
> mpi/partial_integration_computed_vector.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12235] [[19467,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_vibstring_dst_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_dst_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12254] [[19486,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_vibstring_dst_mpi_chunked (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_dst_mpi_chunked.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12273] [[19505,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_vibstring_mpi_aliases (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_mpi_aliases.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12292] [[21444,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_runtime_lattice_mpi_dft_small (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_mpi_dft_small.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:13021] [[20765,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_mpi_xsilloading_hdf5 (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:13176] [[20664,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_mpi_xsilloading_hdf5_loose (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5_loose.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:13195] [[20555,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ======================================================================
> FAIL: test_mpi_xsilloading_hdf5_loose2 (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5_loose2.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
>     return func(*(args + fargs), **newkeywords)
>   File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
>     self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:13214] [[20574,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort.  There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems.  This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n  opal_pmix_base_select failed\n  --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
> 
> 
> ----------------------------------------------------------------------
> Ran 172 tests in 406.785s
> 
> FAILED (failures=25)
> Saving test results in /<<PKGBUILDDIR>>/testsuite_results
> make[1]: *** [debian/rules:15: override_dh_auto_test] Error 1

The full build log is available from:
   http://qa-logs.debian.net/2020/12/26/xmds2_3.0.0+dfsg-4_unstable.log

A list of current common problems and possible solutions is available at
http://wiki.debian.org/qa.debian.org/FTBFS . You're welcome to contribute!

If you reassign this bug to another package, please marking it as 'affects'-ing
this package. See https://www.debian.org/Bugs/server-control#affects

If you fail to reproduce this, please provide a build log and diff it with me
so that we can identify if something relevant changed in the meantime.

About the archive rebuild: The rebuild was done on EC2 VM instances from
Amazon Web Services, using a clean, minimal and up-to-date chroot. Every
failed build was retried once to eliminate random failures.



More information about the debian-science-maintainers mailing list