Re: [OMPI users] Tracing of openmpi internal functions

2022-11-16 Thread Chandran, Arun via users
Hi Jeff, Thanks, I will check flamegraphs. Sample generation with perf could be a problem, I don't think I can do 'mpirun -np <> perf record ' and get the sampling done on all the cores and store each cores data (perf.data) separately to analyze it. Is it possible to do? Came to know that

Re: [OMPI users] What is the best choice of pml and btl for intranode communication

2023-03-06 Thread Chandran, Arun via users
nment variable or something? ____________ From: users mailto:users-boun...@lists.open-mpi.org>> on behalf of Chandran, Arun via users mailto:users@lists.open-mpi.org>> Sent: Monday, March 6, 2023 3:33 AM To: Open MPI Users mailto:users@lists.open-mpi.org&

Re: [OMPI users] What is the best choice of pml and btl for intranode communication

2023-03-06 Thread Chandran, Arun via users
users-boun...@lists.open-mpi.org>> on behalf of Chandran, Arun via users mailto:users@lists.open-mpi.org>> Sent: Monday, March 6, 2023 3:33 AM To: Open MPI Users mailto:users@lists.open-mpi.org>> Cc: Chandran, Arun mailto:arun.chand...@amd.com>> Subject: Re: [OMPI users] What

[OMPI users] What is the best choice of pml and btl for intranode communication

2023-03-02 Thread Chandran, Arun via users
[Public] Hi Folks, As the number of cores in a socket is keep on increasing, the right pml,btl (ucx, ob1, uct, vader, etc) that gives the best performance in "intra-node" scenario is important. For openmpi-4.1.4, which pml, btl combination is the best for intra-node communication in the case

Re: [OMPI users] What is the best choice of pml and btl for intranode communication

2023-03-06 Thread Chandran, Arun via users
Mon, Mar 6, 2023 at 3:12 PM Chandran, Arun via users mailto:users@lists.open-mpi.org>> wrote: [Public] Hi Folks, I can run benchmarks and find the pml+btl (ob1, ucx, uct, vader, etc) combination that gives the best performance, but I wanted to hear from the community about what is gener

Re: [OMPI users] What is the best choice of pml and btl for intranode communication

2023-03-06 Thread Chandran, Arun via users
rs, Gilles On Mon, Mar 6, 2023 at 3:12 PM Chandran, Arun via users mailto:users@lists.open-mpi.org>> wrote: [Public] Hi Folks, I can run benchmarks and find the pml+btl (ob1, ucx, uct, vader, etc) combination that gives the best performance, but I wanted to hear from the commu

Re: [OMPI users] What is the best choice of pml and btl for intranode communication

2023-03-05 Thread Chandran, Arun via users
UCX for IB) --Arun -Original Message- From: users On Behalf Of Chandran, Arun via users Sent: Thursday, March 2, 2023 4:01 PM To: users@lists.open-mpi.org Cc: Chandran, Arun Subject: [OMPI users] What is the best choice of pml and btl for intranode communication Hi Folks, As the numbe

[OMPI users] How to use hugetlbfs with openmpi and ucx

2023-07-19 Thread Chandran, Arun via users
Hi All, I am trying to see whether hugetlbfs is improving the latency of communication with a small send receive program. mpirun -np 2 --map-by core --bind-to core --mca pml ucx --mca opal_common_ucx_tls any --mca opal_common_ucx_devices any -mca pml_base_verbose 10 --mca mtl_base_verbose 10

Re: [OMPI users] How to use hugetlbfs with openmpi and ucx

2023-07-19 Thread Chandran, Arun via users
B consumption. It may be useful for large and frequently used buffers. Regards, Florent -Message d'origine- De : users De la part de Chandran, Arun via users Envoyé : mercredi 19 juillet 2023 15:44 À : users@lists.open-mpi.org Cc : Chandran, Arun Objet : [OMPI users] How to use hugetlbfs wit

Re: [OMPI users] [EXTERNAL] Re: How to use hugetlbfs with openmpi and ucx

2023-07-21 Thread Chandran, Arun via users
n 7/19/23, 9:24 PM, "users on behalf of Chandran, Arun via users" mailto:users-boun...@lists.open-mpi.org> on behalf of users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>> wrote: Good luck, Howard Hi, I am trying to use static huge pages, not transparent huge p

[OMPI users] Warnings while building openmpi-5.0.1

2024-01-05 Thread Chandran, Arun via users
Hi All, I am building openmpi-5.0.1 as shown below. a) Download (wget https://download.open-mpi.org/release/open-mpi/v5.0/openmpi-5.0.1.tar.gz) and untar b) make distclean; ./configure --prefix=~/install_path --enable-mca-no-build=btl-uct --with-hwloc=internal --with-libevent=internal