Hi, Artem,

Thanks for the reply.  I'll answer a couple of questions inline below.

One odd thing that I see in the error output that you have provided is that
pmix2x_client.c is active.

> Looking into the v3.1.x branch (
> https://github.com/open-mpi/ompi/tree/v3.1.x/opal/mca/pmix) I see the
> following components:
> * ext1x
> * ext2x
> ...
> *pmix2x
>
> Pmix2x_client is in internal pmix2x component that shouldn't be built if
> external ext2x component was configured. At least it was the case before.
> According to the output it fails on PMIx_Init().
> Can you please do "$ ls mca_pmix_*" in the <ompi-prefix>/lib/openmpi
> directory?
>

$ ls mca_pmix*
mca_pmix_flux.la  mca_pmix_isolated.la  mca_pmix_pmix2x.la
mca_pmix_flux.so  mca_pmix_isolated.so  mca_pmix_pmix2x.so

Another thing that caught my eye: you say that OMPI searches of PMIx 3.x:
> ...
> >>>>> It fails on the test for PMIx 3, which is expected, but then
> >>>>> reports
> >>>>>
> >>>>>
> >>>>> configure:12843: checking version 2x
> >>>>> configure:12861: gcc -E -I/opt/pmix/2.0.2/include  conftest.c
> >>>>> configure:12861: $? = 0
> >>>>> configure:12862: result: found
> >>>>>
>
> But OMPI v3.1.x doesn't have such a component. Can you provide the related
> lines from config.log?
>


Here are the relevant lines.

configure:12680: checking if user requested external PMIx
support(/opt/pmix/2.0.2)
configure:12690: result: yes
configure:12701: checking --with-external-pmix value
configure:12725: result: sanity check ok (/opt/pmix/2.0.2/include)
configure:12768: checking libpmix.* in /opt/pmix/2.0.2/lib64
configure:12774: checking libpmix.* in /opt/pmix/2.0.2/lib
configure:12794: checking PMIx version
configure:12804: result: version file found
configure:12812: checking version 3x
configure:12830: gcc -E -I/opt/pmix/2.0.2/include  conftest.c
conftest.c:95:56: error: #error "not version 3"

I believe that is a red herring.  Some time in the past, I was told that
there is an anticipatory test for pmix3, and since there isn't such a
thing, this is expected to fail.



> Now about debugging of what is happening:
> 1. I'd like to see results with PMIx debug on:
> $ env PMIX_DEBUG=100 srun --mpi=pmix_v2 ...
>

Here is that output, which seems little changed from what was before.  I
include only that from the first communicator, as it repeats almost
verbatim for the others.

srun: Step created for job 99
[cav02.arc-ts.umich.edu:41373] psec: native init
[cav02.arc-ts.umich.edu:41373] psec: none init
[cav02.arc-ts.umich.edu:41374] psec: native init
[cav02.arc-ts.umich.edu:41374] psec: none init
[cav02.arc-ts.umich.edu:41373] pmix: init called
[cav02.arc-ts.umich.edu:41373] PMIX ERROR: OUT-OF-RESOURCE in file
client/pmix_client.c at line 234
[cav02.arc-ts.umich.edu:41373] OPAL ERROR: Error in file pmix2x_client.c at
line 109
[cav02.arc-ts.umich.edu:41374] pmix: init called
[cav02.arc-ts.umich.edu:41374] PMIX ERROR: OUT-OF-RESOURCE in file
client/pmix_client.c at line 234
[cav02.arc-ts.umich.edu:41374] OPAL ERROR: Error in file pmix2x_client.c at
line 109
--------------------------------------------------------------------------
The application appears to have been direct launched using "srun",
but OMPI was not built with SLURM's PMI support and therefore cannot
execute. There are several options for building PMI support under
SLURM, depending upon the SLURM version you are using:

  version 16.05 or later: you can use SLURM's PMIx support. This
  requires that you configure and build SLURM --with-pmix.

  Versions earlier than 16.05: you must use either SLURM's PMI-1 or
  PMI-2 support. SLURM builds PMI-1 by default, or you can manually
  install PMI-2. You must then build Open MPI using --with-pmi pointing
  to the SLURM PMI library location.

Please configure as appropriate and try again.
--------------------------------------------------------------------------
*** An error occurred in MPI_Init
*** on a NULL communicator
*** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
--------------------------------------------------------------------------


The second through fourth also have a line about

[cav02.arc-ts.umich.edu:41373] Local abort before MPI_INIT completed
completed successfully, but am not able to aggregate error messages, and
not able to guarantee that all other processes were killed!


> 2. Can you set SlurmdDebug option in slurm.conf to 10, run the test and
> provide the content of slurmd.log?
>

I will reply separately with this, as I have to coordinate with the cluster
administrator, who is not in yet.

Please note, also, that I was able to build this successfully after install
the hwlock-devel package and adding the --disable-dlopen and
--enable-shared options to configure.

Thanks,    -- bennet



>
> Today's Topics:
>
>    1. Re: Fwd: OpenMPI 3.1.0 on aarch64 (r...@open-mpi.org)
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Thu, 7 Jun 2018 08:05:30 -0700
> From: "r...@open-mpi.org" <r...@open-mpi.org>
> To: Open MPI Users <users@lists.open-mpi.org>
> Subject: Re: [OMPI users] Fwd: OpenMPI 3.1.0 on aarch64
> Message-ID: <b96e575b-51e4-47b7-996d-20998d1da...@open-mpi.org>
> Content-Type: text/plain;       charset=utf-8
>
> Odd - Artem, do you have any suggestions?
>
> > On Jun 7, 2018, at 7:41 AM, Bennet Fauber <ben...@umich.edu> wrote:
> >
> > Thanks, Ralph,
> >
> > I just tried it with
> >
> >    srun --mpi=pmix_v2 ./test_mpi
> >
> > and got these messages
> >
> >
> > srun: Step created for job 89
> > [cav02.arc-ts.umich.edu:92286] PMIX ERROR: OUT-OF-RESOURCE in file
> > client/pmix_client.c at line 234 [cav02.arc-ts.umich.edu:92286] OPAL
> > ERROR: Error in file pmix2x_client.c at line 109
> > [cav02.arc-ts.umich.edu:92287] PMIX ERROR: OUT-OF-RESOURCE in file
> > client/pmix_client.c at line 234 [cav02.arc-ts.umich.edu:92287] OPAL
> > ERROR: Error in file pmix2x_client.c at line 109
> > ----------------------------------------------------------------------
> > ---- The application appears to have been direct launched using
> > "srun", but OMPI was not built with SLURM's PMI support and therefore
> > cannot execute. There are several options for building PMI support
> > under SLURM, depending upon the SLURM version you are using:
> >
> >  version 16.05 or later: you can use SLURM's PMIx support. This
> > requires that you configure and build SLURM --with-pmix.
> >
> >  Versions earlier than 16.05: you must use either SLURM's PMI-1 or
> >  PMI-2 support. SLURM builds PMI-1 by default, or you can manually
> > install PMI-2. You must then build Open MPI using --with-pmi pointing
> > to the SLURM PMI library location.
> >
> > Please configure as appropriate and try again.
> > ----------------------------------------------------------------------
> > ----
> >
> >
> > Just to be complete, I checked the library path,
> >
> >
> > $ ldconfig -p | egrep 'slurm|pmix'
> >    libpmi2.so.1 (libc6,AArch64) => /opt/pmix/2.0.2/lib/libpmi2.so.1
> >    libpmi2.so (libc6,AArch64) => /opt/pmix/2.0.2/lib/libpmi2.so
> >    libpmix.so.2 (libc6,AArch64) => /opt/pmix/2.0.2/lib/libpmix.so.2
> >    libpmix.so (libc6,AArch64) => /opt/pmix/2.0.2/lib/libpmix.so
> >    libpmi.so.1 (libc6,AArch64) => /opt/pmix/2.0.2/lib/libpmi.so.1
> >    libpmi.so (libc6,AArch64) => /opt/pmix/2.0.2/lib/libpmi.so
> >
> >
> > and libpmi* does appear there.
> >
> >
> > I also tried explicitly listing the slurm directory from the slurm
> > library installation in LD_LIBRARY_PATH, just in case it wasn't
> > traversing correctly.  that is, both
> >
> > $ echo $LD_LIBRARY_PATH
> > /sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0/lib:/opt/arm/gcc-7.1.0_Gener
> > ic-AArch64_RHEL-7_aarch64-linux/lib64:/opt/arm/gcc-7.1.0_Generic-AArch
> > 64_RHEL-7_aarch64-linux/lib:/opt/slurm/lib64:/sw/arcts/centos7/hpc-uti
> > ls/lib
> >
> > and
> >
> > $ echo $LD_LIBRARY_PATH
> > /opt/slurm/lib64/slurm:/opt/pmix/2.0.2/lib:/sw/arcts/centos7/gcc_7_1_0
> > /openmpi/3.1.0/lib:/opt/arm/gcc-7.1.0_Generic-AArch64_RHEL-7_aarch64-l
> > inux/lib64:/opt/arm/gcc-7.1.0_Generic-AArch64_RHEL-7_aarch64-linux/lib
> > :/opt/slurm/lib64:/sw/arcts/centos7/hpc-utils/lib
> >
> >
> > I don't have a saved build log, but I can rebuild this and save the
> > build logs, in case any information in those logs would help.
> >
> > I will also mention that we have, in the past, used the
> > --disable-dlopen and --enable-shared flags, which we did not use here.
> > Just in case that makes any difference.
> >
> > -- bennet
> >
> >
> >
> >
> >
> >
> >
> > On Thu, Jun 7, 2018 at 10:01 AM, r...@open-mpi.org <r...@open-mpi.org>
> wrote:
> >> I think you need to set your MPIDefault to pmix_v2 since you are
> >> using a PMIx v2 library
> >>
> >>
> >>> On Jun 7, 2018, at 6:25 AM, Bennet Fauber <ben...@umich.edu> wrote:
> >>>
> >>> Hi, Ralph,
> >>>
> >>> Thanks for the reply, and sorry for the missing information.  I hope
> >>> this fills in the picture better.
> >>>
> >>> $ srun --version
> >>> slurm 17.11.7
> >>>
> >>> $ srun --mpi=list
> >>> srun: MPI types are...
> >>> srun: pmix_v2
> >>> srun: openmpi
> >>> srun: none
> >>> srun: pmi2
> >>> srun: pmix
> >>>
> >>> We have pmix configured as the default in /opt/slurm/etc/slurm.conf
> >>>
> >>>   MpiDefault=pmix
> >>>
> >>> and on the x86_64 system configured the same way, a bare 'srun
> >>> ./test_mpi' is sufficient and runs.
> >>>
> >>> I have tried all of the following srun variations with no joy
> >>>
> >>>
> >>> srun ./test_mpi
> >>> srun --mpi=pmix ./test_mpi
> >>> srun --mpi=pmi2 ./test_mpi
> >>> srun --mpi=openmpi ./test_mpi
> >>>
> >>>
> >>> I believe we are using the spec files that come with both pmix and
> >>> with slurm, and the following to build the .rpm files used at
> >>> installation
> >>>
> >>>
> >>> $ rpmbuild --define '_prefix /opt/pmix/2.0.2' \
> >>>   -ba pmix-2.0.2.spec
> >>>
> >>> $ rpmbuild --define '_prefix /opt/slurm' \
> >>>   --define '_with-pmix --with-pmix=/opt/pmix/2.0.2' \
> >>>   -ta slurm-17.11.7.tar.bz2
> >>>
> >>>
> >>> I did use the '--with-pmix=/opt/pmix/2.0.2' option when building
> OpenMPI.
> >>>
> >>>
> >>> In case it helps, we have these libraries on the aarch64 in
> >>> /opt/slurm/lib64/slurm/mpi*
> >>>
> >>> -rwxr-xr-x 1 root root 257288 May 30 15:27
> >>> /opt/slurm/lib64/slurm/mpi_none.so
> >>> -rwxr-xr-x 1 root root 257240 May 30 15:27
> >>> /opt/slurm/lib64/slurm/mpi_openmpi.so
> >>> -rwxr-xr-x 1 root root 668808 May 30 15:27
> /opt/slurm/lib64/slurm/mpi_pmi2.so
> >>> lrwxrwxrwx 1 root root     16 Jun  1 08:38
> >>> /opt/slurm/lib64/slurm/mpi_pmix.so -> ./mpi_pmix_v2.so -rwxr-xr-x 1
> >>> root root 841312 May 30 15:27 /opt/slurm/lib64/slurm/mpi_pmix_v2.so
> >>>
> >>> and on the x86_64, where it runs, we have a comparable list,
> >>>
> >>> -rwxr-xr-x 1 root root 193192 May 30 15:20
> >>> /opt/slurm/lib64/slurm/mpi_none.so
> >>> -rwxr-xr-x 1 root root 193192 May 30 15:20
> >>> /opt/slurm/lib64/slurm/mpi_openmpi.so
> >>> -rwxr-xr-x 1 root root 622848 May 30 15:20
> /opt/slurm/lib64/slurm/mpi_pmi2.so
> >>> lrwxrwxrwx 1 root root     16 Jun  1 08:32
> >>> /opt/slurm/lib64/slurm/mpi_pmix.so -> ./mpi_pmix_v2.so -rwxr-xr-x 1
> >>> root root 828232 May 30 15:20 /opt/slurm/lib64/slurm/mpi_pmix_v2.so
> >>>
> >>>
> >>> Let me know if anything else would be helpful.
> >>>
> >>> Thanks,    -- bennet
> >>>
> >>> On Thu, Jun 7, 2018 at 8:56 AM, r...@open-mpi.org <r...@open-mpi.org>
> wrote:
> >>>> You didn?t show your srun direct launch cmd line or what version of
> Slurm is being used (and how it was configured), so I can only provide some
> advice. If you want to use PMIx, then you have to do two things:
> >>>>
> >>>> 1. Slurm must be configured to use PMIx - depending on the version,
> >>>> that might be there by default in the rpm
> >>>>
> >>>> 2. you have to tell srun to use the pmix plugin (IIRC you add --mpi
> >>>> pmix to the cmd line - you should check that)
> >>>>
> >>>> If your intent was to use Slurm?s PMI-1 or PMI-2, then you need to
> >>>> configure OMPI --with-pmi=<path-to-those-libraries>
> >>>>
> >>>> Ralph
> >>>>
> >>>>
> >>>>> On Jun 7, 2018, at 5:21 AM, Bennet Fauber <ben...@umich.edu> wrote:
> >>>>>
> >>>>> We are trying out MPI on an aarch64 cluster.
> >>>>>
> >>>>> Our system administrators installed SLURM and PMIx 2.0.2 from .rpm.
> >>>>>
> >>>>> I compiled OpenMPI using the ARM distributed gcc/7.1.0 using the
> >>>>> configure flags shown in this snippet from the top of config.log
> >>>>>
> >>>>> It was created by Open MPI configure 3.1.0, which was generated by
> >>>>> GNU Autoconf 2.69.  Invocation command line was
> >>>>>
> >>>>> $ ./configure --prefix=/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0
> >>>>> --mandir=/sw/arcts/centos7/gcc_7_1_0/openmpi/3.1.0/share/man
> >>>>> --with-pmix=/opt/pmix/2.0.2 --with-libevent=external
> >>>>> --with-hwloc=external --with-slurm CC=gcc CXX=g++ FC=gfortran
> >>>>>
> >>>>> ## --------- ##
> >>>>> ## Platform. ##
> >>>>> ## --------- ##
> >>>>>
> >>>>> hostname = cavium-hpc.arc-ts.umich.edu uname -m = aarch64 uname -r
> >>>>> = 4.11.0-45.4.1.el7a.aarch64 uname -s = Linux uname -v = #1 SMP
> >>>>> Fri Feb 2 17:11:57 UTC 2018
> >>>>>
> >>>>> /usr/bin/uname -p = aarch64
> >>>>>
> >>>>>
> >>>>> It checks for pmi and reports it found,
> >>>>>
> >>>>>
> >>>>> configure:12680: checking if user requested external PMIx
> >>>>> support(/opt/pmix/2.0.2)
> >>>>> configure:12690: result: yes
> >>>>> configure:12701: checking --with-external-pmix value
> >>>>> configure:12725: result: sanity check ok (/opt/pmix/2.0.2/include)
> >>>>> configure:12768: checking libpmix.* in /opt/pmix/2.0.2/lib64
> >>>>> configure:12774: checking libpmix.* in /opt/pmix/2.0.2/lib
> >>>>> configure:12794: checking PMIx version
> >>>>> configure:12804: result: version file found
> >>>>>
> >>>>>
> >>>>> It fails on the test for PMIx 3, which is expected, but then
> >>>>> reports
> >>>>>
> >>>>>
> >>>>> configure:12843: checking version 2x
> >>>>> configure:12861: gcc -E -I/opt/pmix/2.0.2/include  conftest.c
> >>>>> configure:12861: $? = 0
> >>>>> configure:12862: result: found
> >>>>>
> >>>>>
> >>>>> I have a small, test MPI program that I run, and it runs when run
> >>>>> with mpirun using mpirun.  The processes running on the first node
> >>>>> of a two node job are
> >>>>>
> >>>>>
> >>>>> [bennet@cav02 ~]$ ps -ef | grep bennet | egrep 'test_mpi|srun'
> >>>>>
> >>>>> bennet   20340 20282  0 08:04 ?        00:00:00 mpirun ./test_mpi
> >>>>>
> >>>>> bennet   20346 20340  0 08:04 ?        00:00:00 srun
> >>>>> --ntasks-per-node=1 --kill-on-bad-exit --cpu_bind=none --nodes=1
> >>>>> --nodelist=cav03 --ntasks=1 orted -mca ess "slurm" -mca
> >>>>> ess_base_jobid "3609657344" -mca ess_base_vpid "1" -mca
> >>>>> ess_base_num_procs "2" -mca orte_node_regex "cav[2:2-3]@0(2)" -mca
> >>>>> orte_hnp_uri "3609657344.0;tcp://10.242.15.36:58681"
> >>>>>
> >>>>> bennet   20347 20346  0 08:04 ?        00:00:00 srun
> >>>>> --ntasks-per-node=1 --kill-on-bad-exit --cpu_bind=none --nodes=1
> >>>>> --nodelist=cav03 --ntasks=1 orted -mca ess "slurm" -mca
> >>>>> ess_base_jobid "3609657344" -mca ess_base_vpid "1" -mca
> >>>>> ess_base_num_procs "2" -mca orte_node_regex "cav[2:2-3]@0(2)" -mca
> >>>>> orte_hnp_uri "3609657344.0;tcp://10.242.15.36:58681"
> >>>>>
> >>>>> bennet   20352 20340 98 08:04 ?        00:01:50 ./test_mpi
> >>>>>
> >>>>> bennet   20353 20340 98 08:04 ?        00:01:50 ./test_mpi
> >>>>>
> >>>>>
> >>>>> However, when I run it using srun directly, I get the following
> output:
> >>>>>
> >>>>>
> >>>>> srun: Step created for job 87
> >>>>> [cav02.arc-ts.umich.edu:19828] OPAL ERROR: Not initialized in file
> >>>>> pmix2x_client.c at line 109
> >>>>> ------------------------------------------------------------------
> >>>>> -------- The application appears to have been direct launched
> >>>>> using "srun", but OMPI was not built with SLURM's PMI support and
> >>>>> therefore cannot execute. There are several options for building
> >>>>> PMI support under SLURM, depending upon the SLURM version you are
> >>>>> using:
> >>>>>
> >>>>> version 16.05 or later: you can use SLURM's PMIx support. This
> >>>>> requires that you configure and build SLURM --with-pmix.
> >>>>>
> >>>>> Versions earlier than 16.05: you must use either SLURM's PMI-1 or
> >>>>> PMI-2 support. SLURM builds PMI-1 by default, or you can manually
> >>>>> install PMI-2. You must then build Open MPI using --with-pmi
> >>>>> pointing to the SLURM PMI library location.
> >>>>>
> >>>>> Please configure as appropriate and try again.
> >>>>> ------------------------------------------------------------------
> >>>>> --------
> >>>>> *** An error occurred in MPI_Init
> >>>>> *** on a NULL communicator
> >>>>> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now
> abort,
> >>>>> ***    and potentially your MPI job)
> >>>>> [cav02.arc-ts.umich.edu:19828] Local abort before MPI_INIT
> >>>>> completed completed successfully, but am not able to aggregate
> >>>>> error messages, and not able to guarantee that all other processes
> were killed!
> >>>>>
> >>>>>
> >>>>> Using the same scheme to set this up on x86_64 worked, and I am
> >>>>> taking installation parameters, test files, and job parameters
> >>>>> from the working x86_64 installation.
> >>>>>
> >>>>> Other than the architecture, the main difference between the two
> >>>>> clusters is that the aarch64 has only ethernet networking, whereas
> >>>>> there is infiniband on the x86_64 cluster.  I removed the
> >>>>> --with-verbs from the configure line, though, and I thought that
> >>>>> would be sufficient.
> >>>>>
> >>>>> Anyone have suggestions what might be wrong, how to fix it, or for
> >>>>> further diagnostics?
> >>>>>
> >>>>> Thank you,    -- bennet
> >>>>> _______________________________________________
> >>>>> users mailing list
> >>>>> users@lists.open-mpi.org
> >>>>> https://emea01.safelinks.protection.outlook.com/?url=https%3A%2F%2
> >>>>> Flists.open-mpi.org%2Fmailman%2Flistinfo%2Fusers&data=02%7C01%7Car
> >>>>> temp%40mellanox.com%7Ca94f8437c0f147e8631a08d5cca0b3db%7Ca652971c7
> >>>>> d2e4d9ba6a4d149256f461b%7C0%7C0%7C636639912948859600&sdata=p6aGuFT
> >>>>> eHL3YElvHpWHgUFKDtOqXekIBpmEF8L43Jfo%3D&reserved=0
> >>>>
> >>>> _______________________________________________
> >>>> users mailing list
> >>>> users@lists.open-mpi.org
> >>>> https://emea01.safelinks.protection.outlook.com/?url=https%3A%2F%2F
> >>>> lists.open-mpi.org%2Fmailman%2Flistinfo%2Fusers&data=02%7C01%7Carte
> >>>> mp%40mellanox.com%7Ca94f8437c0f147e8631a08d5cca0b3db%7Ca652971c7d2e
> >>>> 4d9ba6a4d149256f461b%7C0%7C0%7C636639912948859600&sdata=p6aGuFTeHL3
> >>>> YElvHpWHgUFKDtOqXekIBpmEF8L43Jfo%3D&reserved=0
> >>> _______________________________________________
> >>> users mailing list
> >>> users@lists.open-mpi.org
> >>> https://emea01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fl
> >>> ists.open-mpi.org%2Fmailman%2Flistinfo%2Fusers&data=02%7C01%7Cartemp
> >>> %40mellanox.com%7Ca94f8437c0f147e8631a08d5cca0b3db%7Ca652971c7d2e4d9
> >>> ba6a4d149256f461b%7C0%7C0%7C636639912948859600&sdata=p6aGuFTeHL3YElv
> >>> HpWHgUFKDtOqXekIBpmEF8L43Jfo%3D&reserved=0
> >>
> >> _______________________________________________
> >> users mailing list
> >> users@lists.open-mpi.org
> >> https://emea01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fli
> >> sts.open-mpi.org%2Fmailman%2Flistinfo%2Fusers&data=02%7C01%7Cartemp%4
> >> 0mellanox.com%7Ca94f8437c0f147e8631a08d5cca0b3db%7Ca652971c7d2e4d9ba6
> >> a4d149256f461b%7C0%7C0%7C636639912948859600&sdata=p6aGuFTeHL3YElvHpWH
> >> gUFKDtOqXekIBpmEF8L43Jfo%3D&reserved=0
> > _______________________________________________
> > users mailing list
> > users@lists.open-mpi.org
> > https://emea01.safelinks.protection.outlook.com/?url=https%3A%2F%2Flis
> > ts.open-mpi.org%2Fmailman%2Flistinfo%2Fusers&data=02%7C01%7Cartemp%40m
> > ellanox.com%7Ca94f8437c0f147e8631a08d5cca0b3db%7Ca652971c7d2e4d9ba6a4d
> > 149256f461b%7C0%7C0%7C636639912948859600&sdata=p6aGuFTeHL3YElvHpWHgUFK
> > DtOqXekIBpmEF8L43Jfo%3D&reserved=0
>
>
>
> ------------------------------
>
> Subject: Digest Footer
>
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
>
> https://emea01.safelinks.protection.outlook.com/?url=https%3A%2F%2Flists.open-mpi.org%2Fmailman%2Flistinfo%2Fusers&data=02%7C01%7Cartemp%40mellanox.com%7Ca94f8437c0f147e8631a08d5cca0b3db%7Ca652971c7d2e4d9ba6a4d149256f461b%7C0%7C0%7C636639912948859600&sdata=p6aGuFTeHL3YElvHpWHgUFKDtOqXekIBpmEF8L43Jfo%3D&reserved=0
>
> ------------------------------
>
> End of users Digest, Vol 4059, Issue 2
> **************************************
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://lists.open-mpi.org/mailman/listinfo/users
>
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to