This is on an ARM processor? I suspect that is the root of the problems as we aren’t seeing anything like this elsewhere.
> On Jun 18, 2018, at 1:27 PM, Bennet Fauber <ben...@umich.edu> wrote: > > If it's of any use, 3.0.0 seems to hang at > > Making check in class > make[2]: Entering directory `/tmp/build/openmpi-3.0.0/test/class' > make ompi_rb_tree opal_bitmap opal_hash_table opal_proc_table > opal_tree opal_list opal_value_array opal_pointer_array opal_lifo > opal_fifo > make[3]: Entering directory `/tmp/build/openmpi-3.0.0/test/class' > make[3]: `ompi_rb_tree' is up to date. > make[3]: `opal_bitmap' is up to date. > make[3]: `opal_hash_table' is up to date. > make[3]: `opal_proc_table' is up to date. > make[3]: `opal_tree' is up to date. > make[3]: `opal_list' is up to date. > make[3]: `opal_value_array' is up to date. > make[3]: `opal_pointer_array' is up to date. > make[3]: `opal_lifo' is up to date. > make[3]: `opal_fifo' is up to date. > make[3]: Leaving directory `/tmp/build/openmpi-3.0.0/test/class' > make check-TESTS > make[3]: Entering directory `/tmp/build/openmpi-3.0.0/test/class' > make[4]: Entering directory `/tmp/build/openmpi-3.0.0/test/class' > > I have to interrupt it, but it's been many minutes, and usually these > have not been behaving this way. > > -- bennet > > On Mon, Jun 18, 2018 at 4:21 PM Bennet Fauber <ben...@umich.edu> wrote: >> >> No such luck. If it matters, mpirun does seem to work with processes >> on the local node that have no internal MPI code. That is, >> >> [bennet@cavium-hpc ~]$ mpirun -np 4 hello >> Hello, ARM >> Hello, ARM >> Hello, ARM >> Hello, ARM >> >> but it fails with a similar error if run while a SLURM job is active; i.e., >> >> [bennet@cavium-hpc ~]$ mpirun hello >> -------------------------------------------------------------------------- >> ORTE has lost communication with a remote daemon. >> >> HNP daemon : [[26589,0],0] on node cavium-hpc >> Remote daemon: [[26589,0],1] on node cav01 >> >> This is usually due to either a failure of the TCP network >> connection to the node, or possibly an internal failure of >> the daemon itself. We cannot recover from this failure, and >> therefore will terminate the job. >> -------------------------------------------------------------------------- >> >> That makes sense, I guess. >> >> I'll keep you posted as to what happens with 3.0.0 and downgrading SLURM. >> >> >> Thanks, -- bennet >> >> >> On Mon, Jun 18, 2018 at 4:05 PM r...@open-mpi.org <r...@open-mpi.org> wrote: >>> >>> I doubt Slurm is the issue. For grins, lets try adding “--mca plm rsh” to >>> your mpirun cmd line and see if that works. >>> >>> >>>> On Jun 18, 2018, at 12:57 PM, Bennet Fauber <ben...@umich.edu> wrote: >>>> >>>> To eliminate possibilities, I removed all other versions of OpenMPI >>>> from the system, and rebuilt using the same build script as was used >>>> to generate the prior report. >>>> >>>> [bennet@cavium-hpc bennet]$ ./ompi-3.1.0bd.sh >>>> Checking compilers and things >>>> OMPI is ompi >>>> COMP_NAME is gcc_7_1_0 >>>> SRC_ROOT is /sw/arcts/centos7/src >>>> PREFIX_ROOT is /sw/arcts/centos7 >>>> PREFIX is /sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-bd >>>> CONFIGURE_FLAGS are >>>> COMPILERS are CC=gcc CXX=g++ FC=gfortran >>>> >>>> Currently Loaded Modules: >>>> 1) gcc/7.1.0 >>>> >>>> gcc (ARM-build-14) 7.1.0 >>>> Copyright (C) 2017 Free Software Foundation, Inc. >>>> This is free software; see the source for copying conditions. There is NO >>>> warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. >>>> >>>> Using the following configure command >>>> >>>> ./configure --prefix=/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-bd >>>> --mandir=/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-bd/share/man >>>> --with-pmix=/opt/pmix/2.0.2 --with-libevent=external >>>> --with-hwloc=external --with-slurm --disable-dlopen >>>> --enable-debug CC=gcc CXX=g++ FC=gfortran >>>> >>>> The tar ball is >>>> >>>> 2e783873f6b206aa71f745762fa15da5 >>>> /sw/arcts/centos7/src/ompi/openmpi-3.1.0.tar.gz >>>> >>>> I still get >>>> >>>> [bennet@cavium-hpc ~]$ salloc -N 1 --ntasks-per-node=24 >>>> salloc: Pending job allocation 165 >>>> salloc: job 165 queued and waiting for resources >>>> salloc: job 165 has been allocated resources >>>> salloc: Granted job allocation 165 >>>> [bennet@cavium-hpc ~]$ srun ./test_mpi >>>> The sum = 0.866386 >>>> Elapsed time is: 5.425549 >>>> The sum = 0.866386 >>>> Elapsed time is: 5.422826 >>>> The sum = 0.866386 >>>> Elapsed time is: 5.427676 >>>> The sum = 0.866386 >>>> Elapsed time is: 5.424928 >>>> The sum = 0.866386 >>>> Elapsed time is: 5.422060 >>>> The sum = 0.866386 >>>> Elapsed time is: 5.425431 >>>> The sum = 0.866386 >>>> Elapsed time is: 5.424350 >>>> The sum = 0.866386 >>>> Elapsed time is: 5.423037 >>>> The sum = 0.866386 >>>> Elapsed time is: 5.427727 >>>> The sum = 0.866386 >>>> Elapsed time is: 5.424922 >>>> The sum = 0.866386 >>>> Elapsed time is: 5.424279 >>>> Total time is: 59.672992 >>>> >>>> [bennet@cavium-hpc ~]$ mpirun ./test_mpi >>>> -------------------------------------------------------------------------- >>>> An ORTE daemon has unexpectedly failed after launch and before >>>> communicating back to mpirun. This could be caused by a number >>>> of factors, including an inability to create a connection back >>>> to mpirun due to a lack of common network interfaces and/or no >>>> route found between them. Please check network connectivity >>>> (including firewalls and network routing requirements). >>>> -------------------------------------------------------------------------- >>>> >>>> I reran with >>>> >>>> [bennet@cavium-hpc ~]$ mpirun --mca plm_base_verbose 10 ./test_mpi >>>> 2>&1 | tee debug3.log >>>> >>>> and the gzipped log is attached. >>>> >>>> I thought to try it with a different test program, which spits the error >>>> [cavium-hpc.arc-ts.umich.edu:42853] [[58987,1],0] ORTE_ERROR_LOG: Not >>>> found in file base/ess_base_std_app.c at line 219 >>>> [cavium-hpc.arc-ts.umich.edu:42854] [[58987,1],1] ORTE_ERROR_LOG: Not >>>> found in file base/ess_base_std_app.c at line 219 >>>> -------------------------------------------------------------------------- >>>> It looks like orte_init failed for some reason; your parallel process is >>>> likely to abort. There are many reasons that a parallel process can >>>> fail during orte_init; some of which are due to configuration or >>>> environment problems. This failure appears to be an internal failure; >>>> here's some additional information (which may only be relevant to an >>>> Open MPI developer): >>>> >>>> store DAEMON URI failed >>>> --> Returned value Not found (-13) instead of ORTE_SUCCESS >>>> >>>> >>>> At one point, I am almost certain that OMPI mpirun did work, and I am >>>> at a loss to explain why it no longer does. >>>> >>>> I have also tried the 3.1.1rc1 version. I am now going to try 3.0.0, >>>> and we'll try downgrading SLURM to a prior version. >>>> >>>> -- bennet >>>> >>>> >>>> -- bennetOn Mon, Jun 18, 2018 at 10:56 AM r...@open-mpi.org >>>> <r...@open-mpi.org> wrote: >>>>> >>>>> Hmmm...well, the error has changed from your initial report. Turning off >>>>> the firewall was the solution to that problem. >>>>> >>>>> This problem is different - it isn’t the orted that failed in the log you >>>>> sent, but the application proc that couldn’t initialize. It looks like >>>>> that app was compiled against some earlier version of OMPI? It is looking >>>>> for something that no longer exists. I saw that you compiled it with a >>>>> simple “gcc” instead of our wrapper compiler “mpicc” - any particular >>>>> reason? My guess is that your compile picked up some older version of >>>>> OMPI on the system. >>>>> >>>>> Ralph >>>>> >>>>> >>>>>> On Jun 17, 2018, at 2:51 PM, Bennet Fauber <ben...@umich.edu> wrote: >>>>>> >>>>>> I rebuilt with --enable-debug, then ran with >>>>>> >>>>>> [bennet@cavium-hpc ~]$ salloc -N 1 --ntasks-per-node=24 >>>>>> salloc: Pending job allocation 158 >>>>>> salloc: job 158 queued and waiting for resources >>>>>> salloc: job 158 has been allocated resources >>>>>> salloc: Granted job allocation 158 >>>>>> >>>>>> [bennet@cavium-hpc ~]$ srun ./test_mpi >>>>>> The sum = 0.866386 >>>>>> Elapsed time is: 5.426759 >>>>>> The sum = 0.866386 >>>>>> Elapsed time is: 5.424068 >>>>>> The sum = 0.866386 >>>>>> Elapsed time is: 5.426195 >>>>>> The sum = 0.866386 >>>>>> Elapsed time is: 5.426059 >>>>>> The sum = 0.866386 >>>>>> Elapsed time is: 5.423192 >>>>>> The sum = 0.866386 >>>>>> Elapsed time is: 5.426252 >>>>>> The sum = 0.866386 >>>>>> Elapsed time is: 5.425444 >>>>>> The sum = 0.866386 >>>>>> Elapsed time is: 5.423647 >>>>>> The sum = 0.866386 >>>>>> Elapsed time is: 5.426082 >>>>>> The sum = 0.866386 >>>>>> Elapsed time is: 5.425936 >>>>>> The sum = 0.866386 >>>>>> Elapsed time is: 5.423964 >>>>>> Total time is: 59.677830 >>>>>> >>>>>> [bennet@cavium-hpc ~]$ mpirun --mca plm_base_verbose 10 ./test_mpi >>>>>> 2>&1 | tee debug2.log >>>>>> >>>>>> The zipped debug log should be attached. >>>>>> >>>>>> I did that after using systemctl to turn off the firewall on the login >>>>>> node from which the mpirun is executed, as well as on the host on >>>>>> which it runs. >>>>>> >>>>>> [bennet@cavium-hpc ~]$ mpirun hostname >>>>>> -------------------------------------------------------------------------- >>>>>> An ORTE daemon has unexpectedly failed after launch and before >>>>>> communicating back to mpirun. This could be caused by a number >>>>>> of factors, including an inability to create a connection back >>>>>> to mpirun due to a lack of common network interfaces and/or no >>>>>> route found between them. Please check network connectivity >>>>>> (including firewalls and network routing requirements). >>>>>> -------------------------------------------------------------------------- >>>>>> >>>>>> [bennet@cavium-hpc ~]$ squeue >>>>>> JOBID PARTITION NAME USER ST TIME NODES >>>>>> NODELIST(REASON) >>>>>> 158 standard bash bennet R 14:30 1 cav01 >>>>>> [bennet@cavium-hpc ~]$ srun hostname >>>>>> cav01.arc-ts.umich.edu >>>>>> [ repeated 23 more times ] >>>>>> >>>>>> As always, your help is much appreciated, >>>>>> >>>>>> -- bennet >>>>>> >>>>>> On Sun, Jun 17, 2018 at 1:06 PM r...@open-mpi.org <r...@open-mpi.org> >>>>>> wrote: >>>>>>> >>>>>>> Add --enable-debug to your OMPI configure cmd line, and then add --mca >>>>>>> plm_base_verbose 10 to your mpirun cmd line. For some reason, the >>>>>>> remote daemon isn’t starting - this will give you some info as to why. >>>>>>> >>>>>>> >>>>>>>> On Jun 17, 2018, at 9:07 AM, Bennet Fauber <ben...@umich.edu> wrote: >>>>>>>> >>>>>>>> I have a compiled binary that will run with srun but not with mpirun. >>>>>>>> The attempts to run with mpirun all result in failures to initialize. >>>>>>>> I have tried this on one node, and on two nodes, with firewall turned >>>>>>>> on and with it off. >>>>>>>> >>>>>>>> Am I missing some command line option for mpirun? >>>>>>>> >>>>>>>> OMPI built from this configure command >>>>>>>> >>>>>>>> $ ./configure --prefix=/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b >>>>>>>> --mandir=/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b/share/man >>>>>>>> --with-pmix=/opt/pmix/2.0.2 --with-libevent=external >>>>>>>> --with-hwloc=external --with-slurm --disable-dlopen CC=gcc CXX=g++ >>>>>>>> FC=gfortran >>>>>>>> >>>>>>>> All tests from `make check` passed, see below. >>>>>>>> >>>>>>>> [bennet@cavium-hpc ~]$ mpicc --show >>>>>>>> gcc -I/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b/include -pthread >>>>>>>> -L/opt/pmix/2.0.2/lib -Wl,-rpath -Wl,/opt/pmix/2.0.2/lib -Wl,-rpath >>>>>>>> -Wl,/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b/lib >>>>>>>> -Wl,--enable-new-dtags >>>>>>>> -L/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b/lib -lmpi >>>>>>>> >>>>>>>> The test_mpi was compiled with >>>>>>>> >>>>>>>> $ gcc -o test_mpi test_mpi.c -lm >>>>>>>> >>>>>>>> This is the runtime library path >>>>>>>> >>>>>>>> [bennet@cavium-hpc ~]$ echo $LD_LIBRARY_PATH >>>>>>>> /opt/slurm/lib64:/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0-b/lib:/opt/arm/gcc-7.1.0_Generic-AArch64_RHEL-7_aarch64-linux/lib64:/opt/arm/gcc-7.1.0_Generic-AArch64_RHEL-7_aarch64-linux/lib:/opt/slurm/lib64:/opt/pmix/2.0.2/lib:/sw/arcts/centos7/hpc-utils/lib >>>>>>>> >>>>>>>> >>>>>>>> These commands are given in exact sequence in which they were entered >>>>>>>> at a console. >>>>>>>> >>>>>>>> [bennet@cavium-hpc ~]$ salloc -N 1 --ntasks-per-node=24 >>>>>>>> salloc: Pending job allocation 156 >>>>>>>> salloc: job 156 queued and waiting for resources >>>>>>>> salloc: job 156 has been allocated resources >>>>>>>> salloc: Granted job allocation 156 >>>>>>>> >>>>>>>> [bennet@cavium-hpc ~]$ mpirun ./test_mpi >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> An ORTE daemon has unexpectedly failed after launch and before >>>>>>>> communicating back to mpirun. This could be caused by a number >>>>>>>> of factors, including an inability to create a connection back >>>>>>>> to mpirun due to a lack of common network interfaces and/or no >>>>>>>> route found between them. Please check network connectivity >>>>>>>> (including firewalls and network routing requirements). >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> >>>>>>>> [bennet@cavium-hpc ~]$ srun ./test_mpi >>>>>>>> The sum = 0.866386 >>>>>>>> Elapsed time is: 5.425439 >>>>>>>> The sum = 0.866386 >>>>>>>> Elapsed time is: 5.427427 >>>>>>>> The sum = 0.866386 >>>>>>>> Elapsed time is: 5.422579 >>>>>>>> The sum = 0.866386 >>>>>>>> Elapsed time is: 5.424168 >>>>>>>> The sum = 0.866386 >>>>>>>> Elapsed time is: 5.423951 >>>>>>>> The sum = 0.866386 >>>>>>>> Elapsed time is: 5.422414 >>>>>>>> The sum = 0.866386 >>>>>>>> Elapsed time is: 5.427156 >>>>>>>> The sum = 0.866386 >>>>>>>> Elapsed time is: 5.424834 >>>>>>>> The sum = 0.866386 >>>>>>>> Elapsed time is: 5.425103 >>>>>>>> The sum = 0.866386 >>>>>>>> Elapsed time is: 5.422415 >>>>>>>> The sum = 0.866386 >>>>>>>> Elapsed time is: 5.422948 >>>>>>>> Total time is: 59.668622 >>>>>>>> >>>>>>>> Thanks, -- bennet >>>>>>>> >>>>>>>> >>>>>>>> make check results >>>>>>>> ---------------------------------------------- >>>>>>>> >>>>>>>> make check-TESTS >>>>>>>> make[3]: Entering directory `/tmp/build/openmpi-3.1.0/ompi/debuggers' >>>>>>>> make[4]: Entering directory `/tmp/build/openmpi-3.1.0/ompi/debuggers' >>>>>>>> PASS: predefined_gap_test >>>>>>>> PASS: predefined_pad_test >>>>>>>> SKIP: dlopen_test >>>>>>>> ============================================================================ >>>>>>>> Testsuite summary for Open MPI 3.1.0 >>>>>>>> ============================================================================ >>>>>>>> # TOTAL: 3 >>>>>>>> # PASS: 2 >>>>>>>> # SKIP: 1 >>>>>>>> # XFAIL: 0 >>>>>>>> # FAIL: 0 >>>>>>>> # XPASS: 0 >>>>>>>> # ERROR: 0 >>>>>>>> ============================================================================ >>>>>>>> [ elided ] >>>>>>>> PASS: atomic_cmpset_noinline >>>>>>>> - 5 threads: Passed >>>>>>>> PASS: atomic_cmpset_noinline >>>>>>>> - 8 threads: Passed >>>>>>>> ============================================================================ >>>>>>>> Testsuite summary for Open MPI 3.1.0 >>>>>>>> ============================================================================ >>>>>>>> # TOTAL: 8 >>>>>>>> # PASS: 8 >>>>>>>> # SKIP: 0 >>>>>>>> # XFAIL: 0 >>>>>>>> # FAIL: 0 >>>>>>>> # XPASS: 0 >>>>>>>> # ERROR: 0 >>>>>>>> ============================================================================ >>>>>>>> [ elided ] >>>>>>>> make[4]: Entering directory `/tmp/build/openmpi-3.1.0/test/class' >>>>>>>> PASS: ompi_rb_tree >>>>>>>> PASS: opal_bitmap >>>>>>>> PASS: opal_hash_table >>>>>>>> PASS: opal_proc_table >>>>>>>> PASS: opal_tree >>>>>>>> PASS: opal_list >>>>>>>> PASS: opal_value_array >>>>>>>> PASS: opal_pointer_array >>>>>>>> PASS: opal_lifo >>>>>>>> PASS: opal_fifo >>>>>>>> ============================================================================ >>>>>>>> Testsuite summary for Open MPI 3.1.0 >>>>>>>> ============================================================================ >>>>>>>> # TOTAL: 10 >>>>>>>> # PASS: 10 >>>>>>>> # SKIP: 0 >>>>>>>> # XFAIL: 0 >>>>>>>> # FAIL: 0 >>>>>>>> # XPASS: 0 >>>>>>>> # ERROR: 0 >>>>>>>> ============================================================================ >>>>>>>> [ elided ] >>>>>>>> make opal_thread opal_condition >>>>>>>> make[3]: Entering directory `/tmp/build/openmpi-3.1.0/test/threads' >>>>>>>> CC opal_thread.o >>>>>>>> CCLD opal_thread >>>>>>>> CC opal_condition.o >>>>>>>> CCLD opal_condition >>>>>>>> make[3]: Leaving directory `/tmp/build/openmpi-3.1.0/test/threads' >>>>>>>> make check-TESTS >>>>>>>> make[3]: Entering directory `/tmp/build/openmpi-3.1.0/test/threads' >>>>>>>> make[4]: Entering directory `/tmp/build/openmpi-3.1.0/test/threads' >>>>>>>> ============================================================================ >>>>>>>> Testsuite summary for Open MPI 3.1.0 >>>>>>>> ============================================================================ >>>>>>>> # TOTAL: 0 >>>>>>>> # PASS: 0 >>>>>>>> # SKIP: 0 >>>>>>>> # XFAIL: 0 >>>>>>>> # FAIL: 0 >>>>>>>> # XPASS: 0 >>>>>>>> # ERROR: 0 >>>>>>>> ============================================================================ >>>>>>>> [ elided ] >>>>>>>> make[4]: Entering directory `/tmp/build/openmpi-3.1.0/test/datatype' >>>>>>>> PASS: opal_datatype_test >>>>>>>> PASS: unpack_hetero >>>>>>>> PASS: checksum >>>>>>>> PASS: position >>>>>>>> PASS: position_noncontig >>>>>>>> PASS: ddt_test >>>>>>>> PASS: ddt_raw >>>>>>>> PASS: unpack_ooo >>>>>>>> PASS: ddt_pack >>>>>>>> PASS: external32 >>>>>>>> ============================================================================ >>>>>>>> Testsuite summary for Open MPI 3.1.0 >>>>>>>> ============================================================================ >>>>>>>> # TOTAL: 10 >>>>>>>> # PASS: 10 >>>>>>>> # SKIP: 0 >>>>>>>> # XFAIL: 0 >>>>>>>> # FAIL: 0 >>>>>>>> # XPASS: 0 >>>>>>>> # ERROR: 0 >>>>>>>> ============================================================================ >>>>>>>> [ elided ] >>>>>>>> make[4]: Entering directory `/tmp/build/openmpi-3.1.0/test/util' >>>>>>>> PASS: opal_bit_ops >>>>>>>> PASS: opal_path_nfs >>>>>>>> PASS: bipartite_graph >>>>>>>> ============================================================================ >>>>>>>> Testsuite summary for Open MPI 3.1.0 >>>>>>>> ============================================================================ >>>>>>>> # TOTAL: 3 >>>>>>>> # PASS: 3 >>>>>>>> # SKIP: 0 >>>>>>>> # XFAIL: 0 >>>>>>>> # FAIL: 0 >>>>>>>> # XPASS: 0 >>>>>>>> # ERROR: 0 >>>>>>>> ============================================================================ >>>>>>>> [ elided ] >>>>>>>> make[4]: Entering directory `/tmp/build/openmpi-3.1.0/test/dss' >>>>>>>> PASS: dss_buffer >>>>>>>> PASS: dss_cmp >>>>>>>> PASS: dss_payload >>>>>>>> PASS: dss_print >>>>>>>> ============================================================================ >>>>>>>> Testsuite summary for Open MPI 3.1.0 >>>>>>>> ============================================================================ >>>>>>>> # TOTAL: 4 >>>>>>>> # PASS: 4 >>>>>>>> # SKIP: 0 >>>>>>>> # XFAIL: 0 >>>>>>>> # FAIL: 0 >>>>>>>> # XPASS: 0 >>>>>>>> # ERROR: 0 >>>>>>>> ============================================================================ >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> users@lists.open-mpi.org >>>>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> users@lists.open-mpi.org >>>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>>>> <debug2.log.gz>_______________________________________________ >>>>>> users mailing list >>>>>> users@lists.open-mpi.org >>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> users@lists.open-mpi.org >>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>> <debug3.log.gz>_______________________________________________ >>>> users mailing list >>>> users@lists.open-mpi.org >>>> https://lists.open-mpi.org/mailman/listinfo/users >>> >>> _______________________________________________ >>> users mailing list >>> users@lists.open-mpi.org >>> https://lists.open-mpi.org/mailman/listinfo/users > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://lists.open-mpi.org/mailman/listinfo/users _______________________________________________ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users