Bug#944769: python3-h5py fails to import if offline due to apparent MPI failure

Jameson Graef Rollins jrollins at finestructure.net
Fri Nov 15 01:36:47 GMT 2019


Package: python3-h5py
Version: 2.10.0-2
Severity: serious

The latest version of h5py compiled against MPI completely fails to
import if there is no network connection:

servo:~ 0$ python3 -c 'import h5py; print("foo")'
--------------------------------------------------------------------------
No network interfaces were found for out-of-band communications. We require
at least one available network for out-of-band messaging.
--------------------------------------------------------------------------
[servo:32114] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ess_singleton_module.c at line 532
[servo:32114] [[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the local node in file ess_singleton_module.c at line 166
--------------------------------------------------------------------------
It looks like orte_init failed for some reason; your parallel process is
likely to abort.  There are many reasons that a parallel process can
fail during orte_init; some of which are due to configuration or
environment problems.  This failure appears to be an internal failure;
here's some additional information (which may only be relevant to an
Open MPI developer):

  orte_ess_init failed
    --> Returned value Unable to start a daemon on the local node (-127) instead of ORTE_SUCCESS
--------------------------------------------------------------------------
--------------------------------------------------------------------------
It looks like MPI_INIT failed for some reason; your parallel process is
likely to abort.  There are many reasons that a parallel process can
fail during MPI_INIT; some of which are due to configuration or environment
problems.  This failure appears to be an internal failure; here's some
additional information (which may only be relevant to an Open MPI
developer):

  ompi_mpi_init: ompi_rte_init failed
    --> Returned "Unable to start a daemon on the local node" (-127) instead of "Success" (0)
--------------------------------------------------------------------------
*** An error occurred in MPI_Init_thread
*** on a NULL communicator
*** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
***    and potentially your MPI job)
[servo:32114] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
servo:~ 1$

Note the return code and that the print command was not executed.  The
import actually caused python to crash without throwing an exception.
This is obviously unacceptable and very major usability issue.

As mention in other similar reports, please drop the MPI support in
the default build!

jamie.


-- System Information:
Debian Release: bullseye/sid
  APT prefers testing
  APT policy: (600, 'testing'), (500, 'stable'), (200, 'unstable'), (101, 'experimental')
Architecture: amd64 (x86_64)

Kernel: Linux 5.2.0-3-amd64 (SMP w/8 CPU cores)
Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8), LANGUAGE=en_US.UTF-8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash
Init: systemd (via /run/systemd/system)
LSM: AppArmor: enabled

Versions of packages python3-h5py depends on:
ii  libc6                               2.29-3
ii  libhdf5-openmpi-103                 1.10.4+repack-10
ii  python3                             3.7.5-1
ii  python3-mpi4py                      3.0.2-13
ii  python3-numpy [python3-numpy-abi9]  1:1.16.5-1
ii  python3-six                         1.12.0-2

python3-h5py recommends no packages.

Versions of packages python3-h5py suggests:
pn  python-h5py-doc  <none>

-- no debconf information



More information about the debian-science-maintainers mailing list