Bug#978260: xmds2: FTBFS: AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
Lucas Nussbaum
lucas at debian.org
Sat Dec 26 21:46:05 GMT 2020
Source: xmds2
Version: 3.0.0+dfsg-4
Severity: serious
Justification: FTBFS on amd64
Tags: bullseye sid ftbfs
Usertags: ftbfs-20201226 ftbfs-bullseye
Hi,
During a rebuild of all packages in sid, your package failed to build
on amd64.
Relevant part (hopefully):
> make[1]: Entering directory '/<<PKGBUILDDIR>>'
> debian/tests/run-tests -b
> xmds2 version 3.0.0 "Release the Kraken" (Debian package 3.0.0+dfsg-4)
> Copyright 2000-2019 Graham Dennis, Joseph Hope, Mattias Johnsson
> and the xmds team
>
> Configuring for single-process scripts...
> Checking for 'g++' (C++ compiler) : /usr/bin/g++
> Checking whether the compiler works : yes
> Checking that we have a C++ compiler : yes
> Checking whether we are cross-compiling : no
> Checking whether we can link to only static libraries : yes
> Trying to make compiler optimise for this machine : yes
> Trying to make compiler tune for this machine : yes
> Checking for compiler flags -O3 : yes
> Checking for compiler flags -ffast-math : yes
> Checking for compiler flags -funroll-all-loops : yes
> Checking for compiler flags -fomit-frame-pointer : yes
> Checking for compiler flags -falign-loops : yes
> Checking for compiler flags -fstrict-aliasing : yes
> Checking for compiler flags -momit-leaf-frame-pointer : yes
> Checking for cautious math flags : yes
> Checking for Autovectorisation : yes
> Checking for OpenMP : yes
> Checking for pthreads : yes
> Checking for compiler debug flags : yes
> Checking for srandomdev : yes
> Checking for /dev/urandom : yes
> Checking for program 'h5cc' : /usr/bin/h5cc
> Checking for HDF5 (static library) : yes
> Checking for HDF5 High-level library (static library) : yes
> Checking for header hdf5.h : yes
> Checking for libxmds (static library) : no (will try dynamic library instead)
> Checking for libxmds (dynamic library) : no (it's optional anyway)
> Checking for Intel's Vector Math Library (static library) : no (will try dynamic library instead)
> Checking for Intel's Vector Math Library (dynamic library) : no (it's optional anyway)
> Checking safer dSFMT compile flags : yes
> Checking for Intel's Math Kernel Library (static library) : no (will try dynamic library instead)
> Checking for Intel's Math Kernel Library (dynamic library) : no (it's optional anyway)
> Checking for ATLAS's CBLAS. (static library) : no (will try dynamic library instead)
> Checking for ATLAS's CBLAS. (dynamic library) : yes
> Checking for GSL (static library) : yes
> Checking for FFTW3 (static library) : yes
> Checking for single-precision FFTW3 (static library) : yes
> Checking for threading support in FFTW3 (static library) : yes
> Checking for OpenMP support in FFTW3 (static library) : yes
> Checking for threading support in single-precision FFTW3 (static library) : yes
> Checking for OpenMP support in single-precision FFTW3 (static library) : yes
>
> Configuring for MPI scripts...
> Checking for program 'mpic++' : /usr/bin/mpic++
> Checking for 'g++' (C++ compiler) : /usr/bin/mpic++
> Checking whether the compiler works : yes
> Checking that we have a C++ compiler : yes
> Checking whether we are cross-compiling : no
> Checking whether we can link to only static libraries : yes
> Trying to make compiler optimise for this machine : yes
> Trying to make compiler tune for this machine : yes
> Checking for compiler flags -O3 : yes
> Checking for compiler flags -ffast-math : yes
> Checking for compiler flags -funroll-all-loops : yes
> Checking for compiler flags -fomit-frame-pointer : yes
> Checking for compiler flags -falign-loops : yes
> Checking for compiler flags -fstrict-aliasing : yes
> Checking for compiler flags -momit-leaf-frame-pointer : yes
> Checking for cautious math flags : yes
> Checking for Autovectorisation : yes
> Checking for OpenMP : yes
> Checking for pthreads : yes
> Checking for compiler debug flags : yes
> Checking for srandomdev : yes
> Checking for /dev/urandom : yes
> Checking for program 'h5cc' : /usr/bin/h5cc
> Checking for HDF5 (static library) : yes
> Checking for HDF5 High-level library (static library) : yes
> Checking for header hdf5.h : yes
> Checking for libxmds (static library) : no (will try dynamic library instead)
> Checking for libxmds (dynamic library) : no (it's optional anyway)
> Checking for Intel's Vector Math Library (static library) : no (will try dynamic library instead)
> Checking for Intel's Vector Math Library (dynamic library) : no (it's optional anyway)
> Checking safer dSFMT compile flags : yes
> Checking for Intel's Math Kernel Library (static library) : no (will try dynamic library instead)
> Checking for Intel's Math Kernel Library (dynamic library) : no (it's optional anyway)
> Checking for ATLAS's CBLAS. (static library) : no (will try dynamic library instead)
> Checking for ATLAS's CBLAS. (dynamic library) : yes
> Checking for GSL (static library) : yes
> Checking for FFTW3 (static library) : yes
> Checking for single-precision FFTW3 (static library) : yes
> Checking for threading support in FFTW3 (static library) : yes
> Checking for OpenMP support in FFTW3 (static library) : yes
> Checking for threading support in single-precision FFTW3 (static library) : yes
> Checking for OpenMP support in single-precision FFTW3 (static library) : yes
> Checking for FFTW3 with MPI (static library) : yes
> Checking for single-precision FFTW3 with MPI (static library) : yes
> ('Config log saved to ', '/<<PKGBUILDDIR>>/debian/xmds-user-data/waf_configure/config.log')
> test_bug_adaptive_timestep_hang (__main__.main.<locals>.ScriptTestCase)
> integrators/bug_adaptive_timestep_hang.xmds ... ok
> test_vibstring_ark45 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_ark45.xmds ... ok
> test_vibstring_ark89 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_ark89.xmds ... ok
> test_vibstring_mm (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_mm.xmds ... ok
> test_vibstring_re (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_re.xmds ... ok
> test_vibstring_rk4 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_rk4.xmds ... ok
> test_vibstring_rk45 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_rk45.xmds ... ok
> test_vibstring_rk89 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_rk89.xmds ... ok
> test_vibstring_rk9 (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_rk9.xmds ... ok
> test_vibstring_si (__main__.main.<locals>.ScriptTestCase)
> integrators/vibstring_si.xmds ... ok
> test_RbGSdipoles (__main__.main.<locals>.ScriptTestCase)
> fast/RbGSdipoles.xmds ... FAIL
> test_bessel_cosine_evolution (__main__.main.<locals>.ScriptTestCase)
> fast/bessel_cosine_evolution.xmds ... ok
> test_bessel_cosine_groundstate (__main__.main.<locals>.ScriptTestCase)
> fast/bessel_cosine_groundstate.xmds ... ok
> test_cpc_example4 (__main__.main.<locals>.ScriptTestCase)
> fast/cpc_example4.xmds ... FAIL
> test_eigenvalues (__main__.main.<locals>.ScriptTestCase)
> fast/eigenvalues.xmds ... ok
> test_groundstate (__main__.main.<locals>.ScriptTestCase)
> fast/groundstate.xmds ... ok
> test_lorenz (__main__.main.<locals>.ScriptTestCase)
> fast/lorenz.xmds ... ok
> test_tla (__main__.main.<locals>.ScriptTestCase)
> fast/tla.xmds ... ok
> test_transverse_integration_in_vector_initialisation (__main__.main.<locals>.ScriptTestCase)
> fast/transverse_integration_in_vector_initialisation.xmds ... ok
> test_vibstring_circle_spectral (__main__.main.<locals>.ScriptTestCase)
> fast/vibstring_circle_spectral.xmds ... ok
> test_constant_complex_ip (__main__.main.<locals>.ScriptTestCase)
> operators/constant_complex_ip.xmds ... ok
> test_constant_complex_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/constant_complex_ip_2d.xmds ... ok
> test_constant_complex_separated_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/constant_complex_separated_ip_2d.xmds ... ok
> test_constant_double_ip (__main__.main.<locals>.ScriptTestCase)
> operators/constant_double_ip.xmds ... ok
> test_constant_ex (__main__.main.<locals>.ScriptTestCase)
> operators/constant_ex.xmds ... ok
> test_constant_ex_arbitrary_code (__main__.main.<locals>.ScriptTestCase)
> operators/constant_ex_arbitrary_code.xmds ... ok
> test_constant_ex_arbitrary_order (__main__.main.<locals>.ScriptTestCase)
> operators/constant_ex_arbitrary_order.xmds ... ok
> test_constant_ex_arbitrary_order2 (__main__.main.<locals>.ScriptTestCase)
> operators/constant_ex_arbitrary_order2.xmds ... ok
> test_constant_real_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/constant_real_ip_2d.xmds ... ok
> test_cross_propagation (__main__.main.<locals>.ScriptTestCase)
> operators/cross_propagation.xmds ... ok
> test_cross_propagation2 (__main__.main.<locals>.ScriptTestCase)
> operators/cross_propagation2.xmds ... ok
> test_cross_propagation_right (__main__.main.<locals>.ScriptTestCase)
> operators/cross_propagation_right.xmds ... ok
> test_cross_propagation_sic (__main__.main.<locals>.ScriptTestCase)
> operators/cross_propagation_sic.xmds ... ok
> test_cross_propagation_sic_right (__main__.main.<locals>.ScriptTestCase)
> operators/cross_propagation_sic_right.xmds ... ok
> test_highdimcrossprop (__main__.main.<locals>.ScriptTestCase)
> operators/highdimcrossprop.xmds ... ok
> test_nonconstant_complex_ip (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_complex_ip.xmds ... ok
> test_nonconstant_complex_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_complex_ip_2d.xmds ... ok
> test_nonconstant_complex_separated_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_complex_separated_ip_2d.xmds ... ok
> test_nonconstant_double_ip (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_double_ip.xmds ... ok
> test_nonconstant_ex (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_ex.xmds ... ok
> test_nonconstant_real_ip_2d (__main__.main.<locals>.ScriptTestCase)
> operators/nonconstant_real_ip_2d.xmds ... ok
> test_bessel_neumann_wave_equation (__main__.main.<locals>.ScriptTestCase)
> transforms/bessel_neumann_wave_equation.xmds ... ok
> test_bessel_transform (__main__.main.<locals>.ScriptTestCase)
> transforms/bessel_transform.xmds ... ok
> test_bessel_transform_rectangular (__main__.main.<locals>.ScriptTestCase)
> transforms/bessel_transform_rectangular.xmds ... ok
> test_diffusion_bessel (__main__.main.<locals>.ScriptTestCase)
> transforms/diffusion_bessel.xmds ... ok
> test_diffusion_dst (__main__.main.<locals>.ScriptTestCase)
> transforms/diffusion_dst.xmds ... ok
> test_disc (__main__.main.<locals>.ScriptTestCase)
> transforms/disc.xmds ... ok
> test_hermitegauss_fourier (__main__.main.<locals>.ScriptTestCase)
> transforms/hermitegauss_fourier.xmds ... ok
> test_hermitegauss_fourier_loading (__main__.main.<locals>.ScriptTestCase)
> transforms/hermitegauss_fourier_loading.xmds ... ok
> test_hermitegauss_transform_2d (__main__.main.<locals>.ScriptTestCase)
> transforms/hermitegauss_transform_2d.xmds ... ok
> test_hermitegauss_transform_2d_chunked (__main__.main.<locals>.ScriptTestCase)
> transforms/hermitegauss_transform_2d_chunked.xmds ... ok
> test_spherical_ball (__main__.main.<locals>.ScriptTestCase)
> transforms/spherical_ball.xmds ... ok
> test_vibstring_dct (__main__.main.<locals>.ScriptTestCase)
> transforms/vibstring_dct.xmds ... ok
> test_diffusion_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/diffusion_mpi.xmds ... FAIL
> test_diffusion_mpi_chunked (__main__.main.<locals>.ScriptTestCase)
> mpi/diffusion_mpi_chunked.xmds ... FAIL
> test_eigenvalues (__main__.main.<locals>.ScriptTestCase)
> mpi/eigenvalues.xmds ... FAIL
> test_fibre_integer_dimensions_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/fibre_integer_dimensions_mpi.xmds ... FAIL
> test_hermitegauss_transform_2d_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/hermitegauss_transform_2d_mpi.xmds ... FAIL
> test_hermitegauss_transform_2d_mpi_small (__main__.main.<locals>.ScriptTestCase)
> mpi/hermitegauss_transform_2d_mpi_small.xmds ... FAIL
> test_kubo_adaptive_mpi_paths (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_adaptive_mpi_paths.xmds ... FAIL
> test_kubo_integer_dimensions_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_integer_dimensions_mpi.xmds ... FAIL
> test_kubo_mpi_paths (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_mpi_paths.xmds ... FAIL
> test_lorenz_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/lorenz_mpi.xmds ... FAIL
> test_mpi_dft (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft.xmds ... FAIL
> test_mpi_dft_hdf5 (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft_hdf5.xmds ... FAIL
> test_mpi_dft_small (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft_small.xmds ... FAIL
> test_mpi_forward_plan_bug (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_forward_plan_bug.xmds ... FAIL
> test_mpi_highdimcrossprop (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_highdimcrossprop.xmds ... FAIL
> test_partial_integration_computed_vector (__main__.main.<locals>.ScriptTestCase)
> mpi/partial_integration_computed_vector.xmds ... FAIL
> test_vibstring_dst_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_dst_mpi.xmds ... FAIL
> test_vibstring_dst_mpi_chunked (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_dst_mpi_chunked.xmds ... FAIL
> test_vibstring_mpi_aliases (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_mpi_aliases.xmds ... FAIL
> test_integer_dimensions (__main__.main.<locals>.ScriptTestCase)
> geometry/integer_dimensions.xmds ... ok
> test_integer_dimensions_reordered (__main__.main.<locals>.ScriptTestCase)
> geometry/integer_dimensions_reordered.xmds ... ok
> test_nonlocal_access_multiple_components (__main__.main.<locals>.ScriptTestCase)
> geometry/nonlocal_access_multiple_components.xmds ... ok
> test_nonlocal_edge_uniform_access (__main__.main.<locals>.ScriptTestCase)
> geometry/nonlocal_edge_uniform_access.xmds ... ok
> test_nonlocal_index_access (__main__.main.<locals>.ScriptTestCase)
> geometry/nonlocal_index_access.xmds ... ok
> test_nonlocal_negative_uniform_access (__main__.main.<locals>.ScriptTestCase)
> geometry/nonlocal_negative_uniform_access.xmds ... ok
> test_nonlocal_split_uniform_access (__main__.main.<locals>.ScriptTestCase)
> geometry/nonlocal_split_uniform_access.xmds ... ok
> test_bessel_cosine_stochastic_groundstate (__main__.main.<locals>.ScriptTestCase)
> stochastic/bessel_cosine_stochastic_groundstate.xmds ... ok
> test_double_precision_noise_tests (__main__.main.<locals>.ScriptTestCase)
> stochastic/double_precision_noise_tests.xmds ... ok
> test_dsfmt_single_precision (__main__.main.<locals>.ScriptTestCase)
> stochastic/dsfmt_single_precision.xmds ... ok
> test_fibre (__main__.main.<locals>.ScriptTestCase)
> stochastic/fibre.xmds ... ok
> test_fibre_with_correlation_functions (__main__.main.<locals>.ScriptTestCase)
> stochastic/fibre_with_correlation_functions.xmds ... ok
> test_kubo (__main__.main.<locals>.ScriptTestCase)
> stochastic/kubo.xmds ... ok
> test_kubo_fixedstep (__main__.main.<locals>.ScriptTestCase)
> stochastic/kubo_fixedstep.xmds ... ok
> test_photodetector (__main__.main.<locals>.ScriptTestCase)
> stochastic/photodetector.xmds ... ok
> test_photodetector_bessel (__main__.main.<locals>.ScriptTestCase)
> stochastic/photodetector_bessel.xmds ... ok
> test_photodetector_linear (__main__.main.<locals>.ScriptTestCase)
> stochastic/photodetector_linear.xmds ... ok
> test_wigner_cool_HO (__main__.main.<locals>.ScriptTestCase)
> stochastic/wigner_cool_HO.xmds ... ok
> test_bessel_cosine_groundstate (__main__.main.<locals>.ScriptTestCase)
> openmp/bessel_cosine_groundstate.xmds ... ok
> test_diffusion_openmp (__main__.main.<locals>.ScriptTestCase)
> openmp/diffusion_openmp.xmds ... ok
> test_diffusion_openmp_chunked (__main__.main.<locals>.ScriptTestCase)
> openmp/diffusion_openmp_chunked.xmds ... ok
> test_eigenvalues (__main__.main.<locals>.ScriptTestCase)
> openmp/eigenvalues.xmds ... ok
> test_hermitegauss_transform_2d_openmp (__main__.main.<locals>.ScriptTestCase)
> openmp/hermitegauss_transform_2d_openmp.xmds ... ok
> test_hermitegauss_transform_2d_openmp_small (__main__.main.<locals>.ScriptTestCase)
> openmp/hermitegauss_transform_2d_openmp_small.xmds ... ok
> test_kubo_integer_dimensions_openmp (__main__.main.<locals>.ScriptTestCase)
> openmp/kubo_integer_dimensions_openmp.xmds ... ok
> test_lorenz_openmp (__main__.main.<locals>.ScriptTestCase)
> openmp/lorenz_openmp.xmds ... ok
> test_openmp_dft (__main__.main.<locals>.ScriptTestCase)
> openmp/openmp_dft.xmds ... ok
> test_openmp_dft_hdf5 (__main__.main.<locals>.ScriptTestCase)
> openmp/openmp_dft_hdf5.xmds ... ok
> test_openmp_dft_small (__main__.main.<locals>.ScriptTestCase)
> openmp/openmp_dft_small.xmds ... ok
> test_openmp_highdimcrossprop (__main__.main.<locals>.ScriptTestCase)
> openmp/openmp_highdimcrossprop.xmds ... ok
> test_partial_integration_computed_vector (__main__.main.<locals>.ScriptTestCase)
> openmp/partial_integration_computed_vector.xmds ... ok
> test_vibstring_dst_openmp (__main__.main.<locals>.ScriptTestCase)
> openmp/vibstring_dst_openmp.xmds ... ok
> test_vibstring_dst_openmp_chunked (__main__.main.<locals>.ScriptTestCase)
> openmp/vibstring_dst_openmp_chunked.xmds ... ok
> test_breakpoints_hdf5 (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/breakpoints_hdf5.xmds ... ok
> test_integer_dimensions (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/integer_dimensions.xmds ... ok
> test_integer_dimensions_with_fixed_lattice (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/integer_dimensions_with_fixed_lattice.xmds ... ok
> test_integer_dimensions_with_runtime_lattice (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/integer_dimensions_with_runtime_lattice.xmds ... ok
> test_runtime_lattice_diffusion_dst (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_diffusion_dst.xmds ... ok
> test_runtime_lattice_initialisation_order (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_initialisation_order.xmds ... ok
> test_runtime_lattice_mpi_dft_small (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_mpi_dft_small.xmds ... FAIL
> test_runtime_lattice_nonlocal_split_uniform_access (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_nonlocal_split_uniform_access.xmds ... ok
> test_runtime_lattice_vibstring_ark89 (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_vibstring_ark89.xmds ... ok
> test_runtime_lattice_xsilloading_hdf5_loose2 (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_xsilloading_hdf5_loose2.xmds ... ok
> test_initialisation_order (__main__.main.<locals>.ScriptTestCase)
> vectors/initialisation_order.xmds ... ok
> test_initialisation_order_chunked (__main__.main.<locals>.ScriptTestCase)
> vectors/initialisation_order_chunked.xmds ... ok
> test_partial_integration_computed_vector (__main__.main.<locals>.ScriptTestCase)
> vectors/partial_integration_computed_vector.xmds ... ok
> test_breakpoints (__main__.main.<locals>.ScriptTestCase)
> io/breakpoints.xmds ... ok
> test_breakpoints_hdf5 (__main__.main.<locals>.ScriptTestCase)
> io/breakpoints_hdf5.xmds ... ok
> test_mpi_xsilloading_hdf5 (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5.xmds ... FAIL
> test_mpi_xsilloading_hdf5_loose (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5_loose.xmds ... FAIL
> test_mpi_xsilloading_hdf5_loose2 (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5_loose2.xmds ... FAIL
> test_nlse_sampling (__main__.main.<locals>.ScriptTestCase)
> io/nlse_sampling.xmds ... ok
> test_xsilloading_hdf5 (__main__.main.<locals>.ScriptTestCase)
> io/xsilloading_hdf5.xmds ... ok
> test_xsilloading_hdf5_loose (__main__.main.<locals>.ScriptTestCase)
> io/xsilloading_hdf5_loose.xmds ... ok
> test_xsilloading_hdf5_loose2 (__main__.main.<locals>.ScriptTestCase)
> io/xsilloading_hdf5_loose2.xmds ... ok
> test_arguments (__main__.main.<locals>.ScriptTestCase)
> features/arguments.xmds ... ok
> test_arguments_append_args_to_output_filename (__main__.main.<locals>.ScriptTestCase)
> features/arguments_append_args_to_output_filename.xmds ... ok
> test_arguments_with_similar_names (__main__.main.<locals>.ScriptTestCase)
> features/arguments_with_similar_names.xmds ... ok
> test_error_check_multipath (__main__.main.<locals>.ScriptTestCase)
> features/error_check_multipath.xmds ... ok
> test_halt_non_finite (__main__.main.<locals>.ScriptTestCase)
> features/halt_non_finite.xmds ... ok
> test_hermitegauss_transform_2d_chunked_breakpoints (__main__.main.<locals>.ScriptTestCase)
> features/hermitegauss_transform_2d_chunked_breakpoints.xmds ... ok
> test_hermitegauss_transform_2d_chunked_breakpoints_hdf5 (__main__.main.<locals>.ScriptTestCase)
> features/hermitegauss_transform_2d_chunked_breakpoints_hdf5.xmds ... ok
> test_realistic_Rb_and_fields (__main__.main.<locals>.ScriptTestCase)
> features/realistic_Rb_and_fields.xmds ... ok
> test_runtime_paths (__main__.main.<locals>.ScriptTestCase)
> features/runtime_paths.xmds ... ok
> test_space in filename (__main__.main.<locals>.ScriptTestCase)
> features/space in filename.xmds ... ok
> test_assignmentToIncorrectVariable (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_basic (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_combined (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_complicatedSafeOperation (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_hiddenUnsafeOperation (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_missingAssignment (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_realExample (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_safeBinaryOperation (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_safeSubtraction (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_unsafeBinaryOperation (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_unsafeSubtraction (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_unsafeUnaryOperation (xpdeint.CodeParser.IPOperatorSanityCheckTests) ... ok
> test_doubleDivisionByInteger (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_floatDivision (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_ignoreComments (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_ignoreStrings (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_integerDivision (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_integerDivisionByDouble (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_symbolDivision (xpdeint.CodeParser.IntegerDivisionTests) ... ok
> test_accessDifferentVariables (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_accessMultipleTimes (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_basic (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_combined (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_multipleAccess (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_notGreedy (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_withPrintf (xpdeint.CodeParser.NonlocalDimensionAccessForComponentsTests) ... ok
> test_combined (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_ignoreChildComment (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_ignoreSiblingComment (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_ignoreSiblingQuotedString (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_invalidSyntax (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_nestedOperators (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_notGreedy (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_unbalancedString (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
> test_withPrintf (xpdeint.CodeParser.TargetComponentsForOperatorsInStringTests) ... ok
>
> ======================================================================
> FAIL: test_RbGSdipoles (__main__.main.<locals>.ScriptTestCase)
> fast/RbGSdipoles.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:11215] [[18447,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_cpc_example4 (__main__.main.<locals>.ScriptTestCase)
> fast/cpc_example4.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:11268] [[20420,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_diffusion_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/diffusion_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:11950] [[19822,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_diffusion_mpi_chunked (__main__.main.<locals>.ScriptTestCase)
> mpi/diffusion_mpi_chunked.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:11969] [[19713,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_eigenvalues (__main__.main.<locals>.ScriptTestCase)
> mpi/eigenvalues.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:11988] [[19732,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_fibre_integer_dimensions_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/fibre_integer_dimensions_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12007] [[19751,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_hermitegauss_transform_2d_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/hermitegauss_transform_2d_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12026] [[19770,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_hermitegauss_transform_2d_mpi_small (__main__.main.<locals>.ScriptTestCase)
> mpi/hermitegauss_transform_2d_mpi_small.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12045] [[19661,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_kubo_adaptive_mpi_paths (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_adaptive_mpi_paths.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12064] [[19680,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_kubo_integer_dimensions_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_integer_dimensions_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12083] [[19699,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_kubo_mpi_paths (__main__.main.<locals>.ScriptTestCase)
> mpi/kubo_mpi_paths.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12102] [[19590,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_lorenz_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/lorenz_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12121] [[19609,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_mpi_dft (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12140] [[19628,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_mpi_dft_hdf5 (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft_hdf5.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12159] [[19647,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_mpi_dft_small (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_dft_small.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12178] [[19538,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_mpi_forward_plan_bug (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_forward_plan_bug.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12197] [[19557,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_mpi_highdimcrossprop (__main__.main.<locals>.ScriptTestCase)
> mpi/mpi_highdimcrossprop.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12216] [[19576,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_partial_integration_computed_vector (__main__.main.<locals>.ScriptTestCase)
> mpi/partial_integration_computed_vector.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12235] [[19467,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_vibstring_dst_mpi (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_dst_mpi.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12254] [[19486,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_vibstring_dst_mpi_chunked (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_dst_mpi_chunked.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12273] [[19505,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_vibstring_mpi_aliases (__main__.main.<locals>.ScriptTestCase)
> mpi/vibstring_mpi_aliases.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:12292] [[21444,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_runtime_lattice_mpi_dft_small (__main__.main.<locals>.ScriptTestCase)
> runtime_lattice/runtime_lattice_mpi_dft_small.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:13021] [[20765,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_mpi_xsilloading_hdf5 (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:13176] [[20664,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_mpi_xsilloading_hdf5_loose (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5_loose.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:13195] [[20555,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ======================================================================
> FAIL: test_mpi_xsilloading_hdf5_loose2 (__main__.main.<locals>.ScriptTestCase)
> io/mpi_xsilloading_hdf5_loose2.xmds
> ----------------------------------------------------------------------
> Traceback (most recent call last):
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 328, in newfunc
> return func(*(args + fargs), **newkeywords)
> File "/<<PKGBUILDDIR>>/./run_tests.py", line 172, in scriptTestingFunction
> self.assertTrue(returnCode == 0, "Failed to execute compiled simulation correctly. Got returnCode %(returnCode)i;\nstdout = %(stdout)s;\nstderr = %(stderr)s\n" % locals())
> AssertionError: False is not true : Failed to execute compiled simulation correctly. Got returnCode 1;
> stdout = b'';
> stderr = b"[ip-172-31-1-222:13214] [[20574,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../../orte/mca/ess/hnp/ess_hnp_module.c at line 320\n--------------------------------------------------------------------------\nIt looks like orte_init failed for some reason; your parallel process is\nlikely to abort. There are many reasons that a parallel process can\nfail during orte_init; some of which are due to configuration or\nenvironment problems. This failure appears to be an internal failure;\nhere's some additional information (which may only be relevant to an\nOpen MPI developer):\n\n opal_pmix_base_select failed\n --> Returned value Not found (-13) instead of ORTE_SUCCESS\n--------------------------------------------------------------------------\n"
>
>
> ----------------------------------------------------------------------
> Ran 172 tests in 406.785s
>
> FAILED (failures=25)
> Saving test results in /<<PKGBUILDDIR>>/testsuite_results
> make[1]: *** [debian/rules:15: override_dh_auto_test] Error 1
The full build log is available from:
http://qa-logs.debian.net/2020/12/26/xmds2_3.0.0+dfsg-4_unstable.log
A list of current common problems and possible solutions is available at
http://wiki.debian.org/qa.debian.org/FTBFS . You're welcome to contribute!
If you reassign this bug to another package, please marking it as 'affects'-ing
this package. See https://www.debian.org/Bugs/server-control#affects
If you fail to reproduce this, please provide a build log and diff it with me
so that we can identify if something relevant changed in the meantime.
About the archive rebuild: The rebuild was done on EC2 VM instances from
Amazon Web Services, using a clean, minimal and up-to-date chroot. Every
failed build was retried once to eliminate random failures.
More information about the debian-science-maintainers
mailing list