Package: libopenmpi-dev
Version: 2.0.2-2
Severity: important
File: openmpi
Dear Maintainer,
Openmpi segfaut on stretch(oldstable) distrib.
Here is an exemple that reproduces the problem:
cat mpi_tst.cc
#include <mpi.h>
int main(int argc, char ** argv) {
MPI_Init(&argc, &argv);
MPI_Finalize();
}
mpic++ mpi_tst.cc
./a.out
*** Process received signal ***
Signal: Segmentation fault (11)
Signal code: Address not mapped (1)
Failing at address: (nil)
[ 0]
/lib/x86_64-linux-gnu/libpthread.so.0(+0x110e0)[0x7f68c4b7c0e0]
[ 1]
/lib/x86_64-linux-gnu/libc.so.6(strlen+0x26)[0x7f68c484c676]
[ 2]
/lib/x86_64-linux-gnu/libc.so.6(__strdup+0xe)[0x7f68c484c3ae]
[ 3]
/usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi/mca_rtc_freq.so(+0x2b95)[0x7f68bc422b95]
[ 4]
/usr/lib/x86_64-linux-gnu/libopen-rte.so.20(orte_rtc_base_select+0xeb)[0x7f68c59eef4b]
[ 5]
/usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi/mca_ess_hnp.so(+0x5391)[0x7f68c37a2391]
[ 6]
/usr/lib/x86_64-linux-gnu/libopen-rte.so.20(orte_init+0x235)[0x7f68c599aaa5]
[ 7]
/usr/lib/x86_64-linux-gnu/libopen-rte.so.20(orte_daemon+0x488)[0x7f68c59b9608]
[ 8] orted(+0x8ea)[0x55e95c37b8ea]
[ 9]
/lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xf1)[0x7f68c47ec2e1]
[10] orted(+0x94a)[0x55e95c37b94a]
*** End of error message ***
[[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the
local node in file ess_singleton_module.c at line 575
[[INVALID],INVALID] ORTE_ERROR_LOG: Unable to start a daemon on the
local node in file ess_singleton_module.c at line 165
--------------------------------------------------------------------------
It looks like orte_init failed for some reason; your parallel
process is
likely to abort. There are many reasons that a parallel
process can
fail during orte_init; some of which are due to configuration or
environment problems. This failure appears to be an internal
failure;
here's some additional information (which may only be relevant
to an
Open MPI developer):
orte_ess_init failed
--> Returned value Unable to start a daemon on the local node
(-127) instead of ORTE_SUCCESS
--------------------------------------------------------------------------
--------------------------------------------------------------------------
It looks like MPI_INIT failed for some reason; your parallel
process is
likely to abort. There are many reasons that a parallel
process can
fail during MPI_INIT; some of which are due to configuration or
environment
problems. This failure appears to be an internal failure;
here's some
additional information (which may only be relevant to an Open
MPI
developer):
ompi_mpi_init: ompi_rte_init failed
--> Returned "Unable to start a daemon on the local node" (-127)
instead of "Success" (0)
--------------------------------------------------------------------------
*** An error occurred in MPI_Init
*** on a NULL communicator
*** MPI_ERRORS_ARE_FATAL (processes in this communicator will
now abort,
*** and potentially your MPI job)
Local abort before MPI_INIT completed completed successfully, but am
not able to aggregate error messages, and not able to guarantee that
all other processes were killed!
-- System Information:
Debian Release: 9.12
APT prefers oldstable
APT policy: (500, 'oldstable')
Architecture: amd64 (x86_64)
Kernel: Linux 5.4.0-4-amd64 (SMP w/4 CPU cores)
Locale: LANG=C, LC_CTYPE=C (charmap=ANSI_X3.4-1968) (ignored: LC_ALL
set to C), LANGUAGE=C (charmap=ANSI_X3.4-1968) (ignored: LC_ALL set to
C)
Shell: /bin/sh linked to /bin/dash
Init: unable to detect
Versions of packages libopenmpi-dev depends on:
ii libc6 2.24-11+deb9u4
ii libhwloc-dev 1.11.5-1
ii libhwloc5 1.11.5-1
ii libibverbs-dev 1.2.1-2
ii libopenmpi2 2.0.2-2
ii openmpi-common 2.0.2-2
libopenmpi-dev recommends no packages.
Versions of packages libopenmpi-dev suggests:
pn openmpi-doc <none>
-- no debconf information
--
pierre.saram...@imag.fr
Directeur de Recherche CNRS
Laboratoire Jean Kuntzmann, Grenoble, France
http://ljk.imag.fr/membres/Pierre.Saramito