RHEL 8 with OpenMPI 4.1.5a1 on a HPC cluster compute node Singularity
version 3.7.1. I see the error in another issue mentioned
<https://github.com/open-mpi/ompi/issues/4437> at the Git page an on SO
<https://stackoverflow.com/questions/65634590/data-unpack-would-read-past-end-of-buffer-in-file-util-show-help-c-at-line-501>
where
a suggestion for setting -mca orte_base_help_aggregate 0 suppresses it.
ompi_info | head
Package: Open MPI root@c-141-88-1-005 Distribution
Open MPI: 4.1.5a1
Open MPI repo revision: v4.1.4-32-g5abd86c
Open MPI release date: Sep 05, 2022
Open RTE: 4.1.5a1
Open RTE repo revision: v4.1.4-32-g5abd86c
Open RTE release date: Sep 05, 2022
OPAL: 4.1.5a1
OPAL repo revision: v4.1.4-32-g5abd86c
OPAL release date: Sep 05, 2022
mpirun -np 8 -debug-devel -v snappyHexMesh -overwrite -parallel >
snappyHexMesh.out
[g279:2750943] procdir: /tmp/ompi.g279.547289/pid.2750943/0/0
[g279:2750943] jobdir: /tmp/ompi.g279.547289/pid.2750943/0
[g279:2750943] top: /tmp/ompi.g279.547289/pid.2750943
[g279:2750943] top: /tmp/ompi.g279.547289
[g279:2750943] tmp: /tmp
[g279:2750943] sess_dir_cleanup: job session dir does not exist
[g279:2750943] sess_dir_cleanup: top session dir not empty - leaving
[g279:2750943] procdir: /tmp/ompi.g279.547289/pid.2750943/0/0
[g279:2750943] jobdir: /tmp/ompi.g279.547289/pid.2750943/0
[g279:2750943] top: /tmp/ompi.g279.547289/pid.2750943
[g279:2750943] top: /tmp/ompi.g279.547289
[g279:2750943] tmp: /tmp
[g279:2750943] [[20506,0],0] Releasing job data for [INVALID]
--------------------------------------------------------------------------
By default, for Open MPI 4.0 and later, infiniband ports on a device
are not used by default. The intent is to use UCX for these devices.
You can override this policy by setting the btl_openib_allow_ib MCA
parameter
to true.
Local host: g279
Local adapter: mlx5_0
Local port: 1
--------------------------------------------------------------------------
[g279:2750947] procdir: /tmp/ompi.g279.547289/pid.2750943/1/0
[g279:2750947] jobdir: /tmp/ompi.g279.547289/pid.2750943/1
[g279:2750947] top: /tmp/ompi.g279.547289/pid.2750943
[g279:2750947] top: /tmp/ompi.g279.547289
[g279:2750947] tmp: /tmp
[g279:2750949] procdir: /tmp/ompi.g279.547289/pid.2750943/1/2
[g279:2750949] jobdir: /tmp/ompi.g279.547289/pid.2750943/1
[g279:2750949] top: /tmp/ompi.g279.547289/pid.2750943
[g279:2750949] top: /tmp/ompi.g279.547289
[g279:2750949] tmp: /tmp
--------------------------------------------------------------------------
WARNING: There was an error initializing an OpenFabrics device.
Local host: g279
Local device: mlx5_0
--------------------------------------------------------------------------
[g279:2750948] procdir: /tmp/ompi.g279.547289/pid.2750943/1/1
[g279:2750948] jobdir: /tmp/ompi.g279.547289/pid.2750943/1
[g279:2750948] top: /tmp/ompi.g279.547289/pid.2750943
[g279:2750948] top: /tmp/ompi.g279.547289
[g279:2750948] tmp: /tmp
[g279:2750953] procdir: /tmp/ompi.g279.547289/pid.2750943/1/4
[g279:2750953] jobdir: /tmp/ompi.g279.547289/pid.2750943/1
[g279:2750953] top: /tmp/ompi.g279.547289/pid.2750943
[g279:2750953] top: /tmp/ompi.g279.547289
[g279:2750953] tmp: /tmp
[g279:2750950] procdir: /tmp/ompi.g279.547289/pid.2750943/1/3
[g279:2750950] jobdir: /tmp/ompi.g279.547289/pid.2750943/1
[g279:2750950] top: /tmp/ompi.g279.547289/pid.2750943
[g279:2750950] top: /tmp/ompi.g279.547289
[g279:2750950] tmp: /tmp
[g279:2750954] procdir: /tmp/ompi.g279.547289/pid.2750943/1/5
[g279:2750954] jobdir: /tmp/ompi.g279.547289/pid.2750943/1
[g279:2750954] top: /tmp/ompi.g279.547289/pid.2750943
[g279:2750954] top: /tmp/ompi.g279.547289
[g279:2750954] tmp: /tmp
[g279:2750955] procdir: /tmp/ompi.g279.547289/pid.2750943/1/6
[g279:2750955] jobdir: /tmp/ompi.g279.547289/pid.2750943/1
[g279:2750955] top: /tmp/ompi.g279.547289/pid.2750943
[g279:2750955] top: /tmp/ompi.g279.547289
[g279:2750955] tmp: /tmp
MPIR_being_debugged = 0
MPIR_debug_state = 1
MPIR_partial_attach_ok = 1
MPIR_i_am_starter = 0
MPIR_forward_output = 0
MPIR_proctable_size = 8
MPIR_proctable:
(i, host, exe, pid) = (0, g279,
/usr/lib/openfoam/openfoam2212/platforms/linux64GccDPInt32Opt/bin/snappyHexMesh,
2750947)
(i, host, exe, pid) = (1, g279,
/usr/lib/openfoam/openfoam2212/platforms/linux64GccDPInt32Opt/bin/snappyHexMesh,
2750948)
(i, host, exe, pid) = (2, g279,
/usr/lib/openfoam/openfoam2212/platforms/linux64GccDPInt32Opt/bin/snappyHexMesh,
2750949)
(i, host, exe, pid) = (3, g279,
/usr/lib/openfoam/openfoam2212/platforms/linux64GccDPInt32Opt/bin/snappyHexMesh,
2750950)
(i, host, exe, pid) = (4, g279,
/usr/lib/openfoam/openfoam2212/platforms/linux64GccDPInt32Opt/bin/snappyHexMesh,
2750953)
(i, host, exe, pid) = (5, g279,
/usr/lib/openfoam/openfoam2212/platforms/linux64GccDPInt32Opt/bin/snappyHexMesh,
2750954)
(i, host, exe, pid) = (6, g279,
/usr/lib/openfoam/openfoam2212/platforms/linux64GccDPInt32Opt/bin/snappyHexMesh,
2750955)
(i, host, exe, pid) = (7, g279,
/usr/lib/openfoam/openfoam2212/platforms/linux64GccDPInt32Opt/bin/snappyHexMesh,
2750956)
MPIR_executable_path: NULL
MPIR_server_arguments: NULL
[g279:2750956] procdir: /tmp/ompi.g279.547289/pid.2750943/1/7
[g279:2750956] jobdir: /tmp/ompi.g279.547289/pid.2750943/1
[g279:2750956] top: /tmp/ompi.g279.547289/pid.2750943
[g279:2750956] top: /tmp/ompi.g279.547289
[g279:2750956] tmp: /tmp
[g279:2750943] 7 more processes have sent help message
help-mpi-btl-openib.txt / ib port not selected
[g279:2750943] Set MCA parameter "orte_base_help_aggregate" to 0 to see all
help / error messages
[g279:2750943] 7 more processes have sent help message
help-mpi-btl-openib.txt / error in device init
[g279:2750943] sess_dir_finalize: proc session dir does not exist
[g279:2750943] sess_dir_finalize: job session dir does not exist
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: top session dir not empty - leaving
[g279:2750943] sess_dir_finalize: proc session dir does not exist
[g279:2750943] sess_dir_finalize: job session dir does not exist
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: top session dir not empty - leaving
[g279:2750943] sess_dir_finalize: proc session dir does not exist
[g279:2750943] sess_dir_finalize: job session dir does not exist
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: top session dir not empty - leaving
[g279:2750943] sess_dir_finalize: proc session dir does not exist
[g279:2750943] sess_dir_finalize: job session dir does not exist
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: top session dir not empty - leaving
[g279:2750943] sess_dir_finalize: proc session dir does not exist
[g279:2750943] sess_dir_finalize: job session dir does not exist
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: top session dir not empty - leaving
[g279:2750943] sess_dir_finalize: proc session dir does not exist
[g279:2750943] sess_dir_finalize: job session dir does not exist
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: top session dir not empty - leaving
[g279:2750943] sess_dir_finalize: proc session dir does not exist
[g279:2750943] sess_dir_finalize: job session dir does not exist
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: top session dir not empty - leaving
[g279:2750943] sess_dir_finalize: proc session dir does not exist
[g279:2750943] sess_dir_finalize: job session dir does not exist
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: jobfam session dir not empty - leaving
[g279:2750943] sess_dir_finalize: top session dir not empty - leaving
[g279:2750943] Job UNKNOWN has launched
[g279:2750943] [[20506,0],0] Releasing job data for [20506,1]
[g279:2750943] sess_dir_finalize: proc session dir does not exist
[g279:2750943] sess_dir_finalize: job session dir does not exist
[g279:2750943] sess_dir_finalize: jobfam session dir does not exist
[g279:2750943] sess_dir_finalize: jobfam session dir does not exist
[g279:2750943] sess_dir_finalize: top session dir not empty - leaving
[g279:2750943] sess_dir_cleanup: job session dir does not exist
[g279:2750943] sess_dir_cleanup: top session dir not empty - leaving
[g279:2750943] [[20506,0],0] Releasing job data for [20506,0]
[g279:2750943] sess_dir_cleanup: job session dir does not exist
[g279:2750943] sess_dir_cleanup: top session dir not empty - leaving
Running mpirun without the debug and verbose options gets this:
[g279:2738279] [[24994,0],0] ORTE_ERROR_LOG: Data unpack would read past
end of buffer in file util/show_help.c at line 501
[g279:2738279] 7 more processes have sent help message
help-mpi-btl-openib.txt / ib port not selected
[g279:2738279] Set MCA parameter "orte_base_help_aggregate" to 0 to see all
help / error messages
[g279:2738279] 6 more processes have sent help message
help-mpi-btl-openib.txt / error in device init
Mesh created. It is advised to check in paraview to confirm mesh of
porespace is reasonable before running flow
Are these just warnings?