The poor performance when compiling with --disable-dlopen also
occurs in OpenMPI 3.0.0 in addition to 2.1.1 and 2.0.2 that I reported
earlier. My understanding is that the EasyBuild group is looking into
simply removing --disable-dlopen from their build, which is what we've
done on our system.
Dave Turner
On Wed, Jan 24, 2018 at 1:00 PM, <[email protected]> wrote:
> Send devel mailing list submissions to
> [email protected]
>
> To subscribe or unsubscribe via the World Wide Web, visit
> https://lists.open-mpi.org/mailman/listinfo/devel
> or, via email, send a message with subject or body 'help' to
> [email protected]
>
> You can reach the person managing the list at
> [email protected]
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of devel digest..."
>
>
> Today's Topics:
>
> 1. Open MPI 3.0.1rc2 available for testing (Barrett, Brian)
> 2. Open MPI 3.1.0 pre-release available (Barrett, Brian)
> 3. Poor performance when compiling with --disable-dlopen
> (Dave Turner)
> 4. Re: Poor performance when compiling with --disable-dlopen
> (Gilles Gouaillardet)
> 5. Re: Poor performance when compiling with --disable-dlopen
> (Gilles Gouaillardet)
> 6. Re: Poor performance when compiling with --disable-dlopen
> (Gilles Gouaillardet)
> 7. Re: Poor performance when compiling with --disable-dlopen
> (Gilles Gouaillardet)
> 8. Re: Poor performance when compiling with --disable-dlopen
> (Paul Hargrove)
> 9. Re: Poor performance when compiling with --disable-dlopen
> (Gilles Gouaillardet)
> 10. Re: Poor performance when compiling with --disable-dlopen
> (Dave Turner)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Wed, 24 Jan 2018 01:04:10 +0000
> From: "Barrett, Brian" <[email protected]>
> To: Open MPI Developers <[email protected]>
> Subject: [OMPI devel] Open MPI 3.0.1rc2 available for testing
> Message-ID: <[email protected]>
> Content-Type: text/plain; charset="utf-8"
>
> I?ve posted the first public release candidate of Open MPI 3.0.1 this
> evening. It can be downloaded for testing from:
>
> https://www.open-mpi.org/software/ompi/v3.0/
>
> We appreciate any testing you can do in preparation for a release in the
> next week or two.
>
>
> Thanks,
>
> Brian & Howard
>
> ------------------------------
>
> Message: 2
> Date: Wed, 24 Jan 2018 01:24:06 +0000
> From: "Barrett, Brian" <[email protected]>
> To: Open MPI Developers <[email protected]>
> Subject: [OMPI devel] Open MPI 3.1.0 pre-release available
> Message-ID: <[email protected]>
> Content-Type: text/plain; charset="us-ascii"
>
> The Open MPI team is pleased to announce the first pre-release of the Open
> MPI 3.1 series, available at:
>
> https://www.open-mpi.org/software/ompi/v3.1/
>
> RC1 has two known issues:
>
> - We did not complete work to support hwloc 2.x, even when hwloc is
> built as an external library. This may or may not be complete before 3.1.0
> is shipped.
> - 3.1.0 is shipping with a pre-release version of PMIx 2.1. We will
> finish the update to PMIx 2.1 before 3.1.0 is released.
>
> We look forward to any other issues you may find in testing.
>
> Thanks,
>
> Brian
>
> ------------------------------
>
> Message: 3
> Date: Tue, 23 Jan 2018 21:55:38 -0600
> From: Dave Turner <[email protected]>
> To: Open MPI Developers <[email protected]>, [email protected]
> Subject: [OMPI devel] Poor performance when compiling with
> --disable-dlopen
> Message-ID:
> <CAFGXdkyNOWkEDxVnrjg4k24Og5MRnN_9KDLSQ73M3TXH_WtUfQ@mail.
> gmail.com>
> Content-Type: text/plain; charset="utf-8"
>
> We compiled OpenMPI 2.1.1 using the EasyBuild configuration
> for CentOS as below and tested on Mellanox QDR hardware.
>
> ./configure --prefix=/homes/daveturner/libs/openmpi-2.1.1c
> --enable-shared
> --enable-mpi-thread-multiple
> --with-verbs
> --enable-mpirun-prefix-by-default
> --with-mpi-cxx
> --enable-mpi-cxx
> --with-hwloc=$EBROOTHWLOC
> --disable-dlopen
>
> The red curve in the attached NetPIPE graph shows the poor performance
> above
> 8 kB for the uni-directional tests with bi-directional and aggregate
> tests also showing similar problems. When I compile using the same
> configuration but with the --disable-dlopen parameter removed then the
> performance is very good as the green curve in the graph shows.
>
> We see the same problems with OpenMPI 2.0.2.
> Replacing --disable-dlopen with --disable-mca-dso showed good performance.
> Replacing --disable-dlopen with --enable-static showed good performance.
> So it's only --disable-dlopen that leads to poor performance.
>
> http://netpipe.cs.ksu.edu
>
> Dave Turner
>
> --
> Work: [email protected] (785) 532-7791
> 2219 Engineering Hall, Manhattan KS 66506
> Home: [email protected]
> cell: (785) 770-5929
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <https://lists.open-mpi.org/mailman/private/devel/
> attachments/20180123/6d4537ad/attachment.html>
> -------------- next part --------------
> A non-text attachment was scrubbed...
> Name: MPI_on_QDR_dlopen_paramter.pdf
> Type: application/pdf
> Size: 16813 bytes
> Desc: not available
> URL: <https://lists.open-mpi.org/mailman/private/devel/
> attachments/20180123/6d4537ad/attachment.pdf>
>
> ------------------------------
>
> Message: 4
> Date: Wed, 24 Jan 2018 13:03:13 +0900
> From: Gilles Gouaillardet <[email protected]>
> To: Dave Turner <[email protected]>, Open MPI Developers
> <[email protected]>
> Subject: Re: [OMPI devel] Poor performance when compiling with
> --disable-dlopen
> Message-ID:
> <CAAkFZ5uXZGsbeg0vVBJ4HiLKmDMVfvUV1tu6HOT-gUaEi5OW8Q@mail.
> gmail.com>
> Content-Type: text/plain; charset="UTF-8"
>
> Dave,
>
> At first glance, that looks pretty odd, and I'll have a look at it.
>
> Which benchmark are you using to measure the bandwidth ?
> Does your benchmark MPI_Init_thread(MPI_THREAD_MULTIPLE) ?
> Have you tried without --enable-mpi-thread-multiple ?
>
> Cheers,
>
> Gilles
>
> On Wed, Jan 24, 2018 at 12:55 PM, Dave Turner <[email protected]>
> wrote:
> >
> > We compiled OpenMPI 2.1.1 using the EasyBuild configuration
> > for CentOS as below and tested on Mellanox QDR hardware.
> >
> > ./configure --prefix=/homes/daveturner/libs/openmpi-2.1.1c
> > --enable-shared
> > --enable-mpi-thread-multiple
> > --with-verbs
> > --enable-mpirun-prefix-by-default
> > --with-mpi-cxx
> > --enable-mpi-cxx
> > --with-hwloc=$EBROOTHWLOC
> > --disable-dlopen
> >
> > The red curve in the attached NetPIPE graph shows the poor performance
> above
> > 8 kB for the uni-directional tests with bi-directional and aggregate
> > tests also showing similar problems. When I compile using the same
> > configuration but with the --disable-dlopen parameter removed then the
> > performance is very good as the green curve in the graph shows.
> >
> > We see the same problems with OpenMPI 2.0.2.
> > Replacing --disable-dlopen with --disable-mca-dso showed good
> performance.
> > Replacing --disable-dlopen with --enable-static showed good performance.
> > So it's only --disable-dlopen that leads to poor performance.
> >
> > http://netpipe.cs.ksu.edu
> >
> > Dave Turner
> >
> > --
> > Work: [email protected] (785) 532-7791
> > 2219 Engineering Hall, Manhattan KS 66506
> > Home: [email protected]
> > cell: (785) 770-5929
> >
> > _______________________________________________
> > devel mailing list
> > [email protected]
> > https://lists.open-mpi.org/mailman/listinfo/devel
>
>
> ------------------------------
>
> Message: 5
> Date: Wed, 24 Jan 2018 13:16:25 +0900
> From: Gilles Gouaillardet <[email protected]>
> To: [email protected]
> Subject: Re: [OMPI devel] Poor performance when compiling with
> --disable-dlopen
> Message-ID: <[email protected]>
> Content-Type: text/plain; charset=utf-8; format=flowed
>
> Dave,
>
>
> one more question, are you running the openib/btl ? or other libraries
> such as MXM or UCX ?
>
>
> Cheers,
>
>
> Gilles
>
>
> On 1/24/2018 12:55 PM, Dave Turner wrote:
> >
> > ? ?We compiled OpenMPI 2.1.1 using the EasyBuild configuration
> > for CentOS as below and tested on Mellanox QDR hardware.
> >
> > ./configure --prefix=/homes/daveturner/libs/openmpi-2.1.1c
> > ? ? ? ? ? ? ? ? ?--enable-shared
> > ? ? ? ? ? ? ? ? ?--enable-mpi-thread-multiple
> > ? ? ? ? ? ? ? ? ?--with-verbs
> > ? ? ? ? ? ? ? ? ?--enable-mpirun-prefix-by-default
> > ? ? ? ? ? ? ? ? ?--with-mpi-cxx
> > ? ? ? ? ? ? ? ? ?--enable-mpi-cxx
> > ? ? ? ? ? ? ? ? ?--with-hwloc=$EBROOTHWLOC
> > ? ? ? ? ? ? ? ? ?--disable-dlopen
> >
> > The red curve in the attached NetPIPE graph shows the poor performance
> > above
> > 8 kB for the uni-directional tests with bi-directional and aggregate
> > tests also showing similar problems.? When I compile using the same
> > configuration but with the --disable-dlopen parameter removed then the
> > performance is very good as the green curve in the graph shows.
> >
> > We see the same problems with OpenMPI 2.0.2.
> > Replacing --disable-dlopen with --disable-mca-dso showed good
> performance.
> > Replacing --disable-dlopen with --enable-static showed good performance.
> > So it's only --disable-dlopen that leads to poor performance.
> >
> > http://netpipe.cs.ksu.edu
> >
> > ? ? ? ? ? ? ? ? ? ?Dave Turner
> >
> > --
> > Work: [email protected] <mailto:[email protected]> ? ? (785) 532-7791
> > ? ? ? ? ? ?? 2219 Engineering Hall, Manhattan KS ?66506
> > Home: [email protected] <mailto:[email protected]>
> > ?????????? ?? cell: (785) 770-5929
> >
> >
> > _______________________________________________
> > devel mailing list
> > [email protected]
> > https://lists.open-mpi.org/mailman/listinfo/devel
>
>
>
> ------------------------------
>
> Message: 6
> Date: Wed, 24 Jan 2018 13:29:09 +0900
> From: Gilles Gouaillardet <[email protected]>
> To: Open MPI Developers <[email protected]>
> Subject: Re: [OMPI devel] Poor performance when compiling with
> --disable-dlopen
> Message-ID:
> <CAAkFZ5vJQ0FvkVtcVmayNsDQpq2oDXpcY2kAMGNUw_pZw3WTCQ@mail.
> gmail.com>
> Content-Type: text/plain; charset="UTF-8"
>
> Dave,
>
> i can reproduce the issue with btl/openib and the IMB benchmark, that
> is known to MPI_Init_thread(MPI_THREAD_MULTIPLE)
>
> note performance is ok with OSU benchmark that does not require
> MPI_THREAD_MULTIPLE
>
> Cheers,
>
> Gilles
>
> On Wed, Jan 24, 2018 at 1:16 PM, Gilles Gouaillardet <[email protected]>
> wrote:
> > Dave,
> >
> >
> > one more question, are you running the openib/btl ? or other libraries
> such
> > as MXM or UCX ?
> >
> >
> > Cheers,
> >
> >
> > Gilles
> >
> >
> > On 1/24/2018 12:55 PM, Dave Turner wrote:
> >>
> >>
> >> We compiled OpenMPI 2.1.1 using the EasyBuild configuration
> >> for CentOS as below and tested on Mellanox QDR hardware.
> >>
> >> ./configure --prefix=/homes/daveturner/libs/openmpi-2.1.1c
> >> --enable-shared
> >> --enable-mpi-thread-multiple
> >> --with-verbs
> >> --enable-mpirun-prefix-by-default
> >> --with-mpi-cxx
> >> --enable-mpi-cxx
> >> --with-hwloc=$EBROOTHWLOC
> >> --disable-dlopen
> >>
> >> The red curve in the attached NetPIPE graph shows the poor performance
> >> above
> >> 8 kB for the uni-directional tests with bi-directional and aggregate
> >> tests also showing similar problems. When I compile using the same
> >> configuration but with the --disable-dlopen parameter removed then the
> >> performance is very good as the green curve in the graph shows.
> >>
> >> We see the same problems with OpenMPI 2.0.2.
> >> Replacing --disable-dlopen with --disable-mca-dso showed good
> performance.
> >> Replacing --disable-dlopen with --enable-static showed good performance.
> >> So it's only --disable-dlopen that leads to poor performance.
> >>
> >> http://netpipe.cs.ksu.edu
> >>
> >> Dave Turner
> >>
> >> --
> >> Work: [email protected] <mailto:[email protected]> (785) 532-7791
> >> 2219 Engineering Hall, Manhattan KS 66506
> >> Home: [email protected] <mailto:[email protected]>
> >> cell: (785) 770-5929
> >>
> >>
> >> _______________________________________________
> >> devel mailing list
> >> [email protected]
> >> https://lists.open-mpi.org/mailman/listinfo/devel
> >
> >
> > _______________________________________________
> > devel mailing list
> > [email protected]
> > https://lists.open-mpi.org/mailman/listinfo/devel
>
>
> ------------------------------
>
> Message: 7
> Date: Wed, 24 Jan 2018 14:17:56 +0900
> From: Gilles Gouaillardet <[email protected]>
> To: Open MPI Developers <[email protected]>
> Subject: Re: [OMPI devel] Poor performance when compiling with
> --disable-dlopen
> Message-ID:
> <CAAkFZ5uhz9gaNfqVMieDr4nkkCNJF+cbRS36i2fGC+o25wRJiw@mail.
> gmail.com>
> Content-Type: text/plain; charset="UTF-8"
>
> Dave,
>
> here is what I found
>
> - MPI_THREAD_MULTIPLE is not part of the equation (I just found it is
> no more required by IMB by default)
> - patcher/overwrite is not built when Open MPI is configure'd with
> --disable-dlopen
> - when configure'd without --disable-dlopen, performances are way
> worst for the IMB (PingPong) benchmark when ran with
> mpirun --mca patcher ^overwrite
> - OSU (osu_bw) performances are not impacted by the patcher/overwrite
> component being blacklisted
>
> I am afraid that's all I can do ...
>
>
> Nathan,
>
> could you please shed some light ?
>
>
> Cheers,
>
> Gilles
>
> On Wed, Jan 24, 2018 at 1:29 PM, Gilles Gouaillardet
> <[email protected]> wrote:
> > Dave,
> >
> > i can reproduce the issue with btl/openib and the IMB benchmark, that
> > is known to MPI_Init_thread(MPI_THREAD_MULTIPLE)
> >
> > note performance is ok with OSU benchmark that does not require
> > MPI_THREAD_MULTIPLE
> >
> > Cheers,
> >
> > Gilles
> >
> > On Wed, Jan 24, 2018 at 1:16 PM, Gilles Gouaillardet <[email protected]>
> wrote:
> >> Dave,
> >>
> >>
> >> one more question, are you running the openib/btl ? or other libraries
> such
> >> as MXM or UCX ?
> >>
> >>
> >> Cheers,
> >>
> >>
> >> Gilles
> >>
> >>
> >> On 1/24/2018 12:55 PM, Dave Turner wrote:
> >>>
> >>>
> >>> We compiled OpenMPI 2.1.1 using the EasyBuild configuration
> >>> for CentOS as below and tested on Mellanox QDR hardware.
> >>>
> >>> ./configure --prefix=/homes/daveturner/libs/openmpi-2.1.1c
> >>> --enable-shared
> >>> --enable-mpi-thread-multiple
> >>> --with-verbs
> >>> --enable-mpirun-prefix-by-default
> >>> --with-mpi-cxx
> >>> --enable-mpi-cxx
> >>> --with-hwloc=$EBROOTHWLOC
> >>> --disable-dlopen
> >>>
> >>> The red curve in the attached NetPIPE graph shows the poor performance
> >>> above
> >>> 8 kB for the uni-directional tests with bi-directional and aggregate
> >>> tests also showing similar problems. When I compile using the same
> >>> configuration but with the --disable-dlopen parameter removed then the
> >>> performance is very good as the green curve in the graph shows.
> >>>
> >>> We see the same problems with OpenMPI 2.0.2.
> >>> Replacing --disable-dlopen with --disable-mca-dso showed good
> performance.
> >>> Replacing --disable-dlopen with --enable-static showed good
> performance.
> >>> So it's only --disable-dlopen that leads to poor performance.
> >>>
> >>> http://netpipe.cs.ksu.edu
> >>>
> >>> Dave Turner
> >>>
> >>> --
> >>> Work: [email protected] <mailto:[email protected]> (785)
> 532-7791
> >>> 2219 Engineering Hall, Manhattan KS 66506
> >>> Home: [email protected] <mailto:[email protected]>
> >>> cell: (785) 770-5929
> >>>
> >>>
> >>> _______________________________________________
> >>> devel mailing list
> >>> [email protected]
> >>> https://lists.open-mpi.org/mailman/listinfo/devel
> >>
> >>
> >> _______________________________________________
> >> devel mailing list
> >> [email protected]
> >> https://lists.open-mpi.org/mailman/listinfo/devel
>
>
> ------------------------------
>
> Message: 8
> Date: Tue, 23 Jan 2018 21:28:32 -0800
> From: Paul Hargrove <[email protected]>
> To: Open MPI Developers <[email protected]>
> Subject: Re: [OMPI devel] Poor performance when compiling with
> --disable-dlopen
> Message-ID:
> <CAAvDA177bOSj_5W9oCt_8NTpZsF75V=9Jkpkg1SFhsov_Q=obg
> @mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
>
> Ah, this sounds familiar.
>
> I believe that the issue Dave sees is that without patcher/overwrite the
> "leave pinned" protocol is OFF by default.
>
> Use of '-mca mpi_leave_pinned 1' may help if my guess is right.
> HOWEVER, w/o the memory management hooks provided using patcher/overwrite,
> leave pinned can give incorrect results.
>
> -Paul
>
> On Tue, Jan 23, 2018 at 9:17 PM, Gilles Gouaillardet <
> [email protected]> wrote:
>
> > Dave,
> >
> > here is what I found
> >
> > - MPI_THREAD_MULTIPLE is not part of the equation (I just found it is
> > no more required by IMB by default)
> > - patcher/overwrite is not built when Open MPI is configure'd with
> > --disable-dlopen
> > - when configure'd without --disable-dlopen, performances are way
> > worst for the IMB (PingPong) benchmark when ran with
> > mpirun --mca patcher ^overwrite
> > - OSU (osu_bw) performances are not impacted by the patcher/overwrite
> > component being blacklisted
> >
> > I am afraid that's all I can do ...
> >
> >
> > Nathan,
> >
> > could you please shed some light ?
> >
> >
> > Cheers,
> >
> > Gilles
> >
> > On Wed, Jan 24, 2018 at 1:29 PM, Gilles Gouaillardet
> > <[email protected]> wrote:
> > > Dave,
> > >
> > > i can reproduce the issue with btl/openib and the IMB benchmark, that
> > > is known to MPI_Init_thread(MPI_THREAD_MULTIPLE)
> > >
> > > note performance is ok with OSU benchmark that does not require
> > > MPI_THREAD_MULTIPLE
> > >
> > > Cheers,
> > >
> > > Gilles
> > >
> > > On Wed, Jan 24, 2018 at 1:16 PM, Gilles Gouaillardet <
> [email protected]>
> > wrote:
> > >> Dave,
> > >>
> > >>
> > >> one more question, are you running the openib/btl ? or other libraries
> > such
> > >> as MXM or UCX ?
> > >>
> > >>
> > >> Cheers,
> > >>
> > >>
> > >> Gilles
> > >>
> > >>
> > >> On 1/24/2018 12:55 PM, Dave Turner wrote:
> > >>>
> > >>>
> > >>> We compiled OpenMPI 2.1.1 using the EasyBuild configuration
> > >>> for CentOS as below and tested on Mellanox QDR hardware.
> > >>>
> > >>> ./configure --prefix=/homes/daveturner/libs/openmpi-2.1.1c
> > >>> --enable-shared
> > >>> --enable-mpi-thread-multiple
> > >>> --with-verbs
> > >>> --enable-mpirun-prefix-by-default
> > >>> --with-mpi-cxx
> > >>> --enable-mpi-cxx
> > >>> --with-hwloc=$EBROOTHWLOC
> > >>> --disable-dlopen
> > >>>
> > >>> The red curve in the attached NetPIPE graph shows the poor
> performance
> > >>> above
> > >>> 8 kB for the uni-directional tests with bi-directional and aggregate
> > >>> tests also showing similar problems. When I compile using the same
> > >>> configuration but with the --disable-dlopen parameter removed then
> the
> > >>> performance is very good as the green curve in the graph shows.
> > >>>
> > >>> We see the same problems with OpenMPI 2.0.2.
> > >>> Replacing --disable-dlopen with --disable-mca-dso showed good
> > performance.
> > >>> Replacing --disable-dlopen with --enable-static showed good
> > performance.
> > >>> So it's only --disable-dlopen that leads to poor performance.
> > >>>
> > >>> http://netpipe.cs.ksu.edu
> > >>>
> > >>> Dave Turner
> > >>>
> > >>> --
> > >>> Work: [email protected] <mailto:[email protected]> (785)
> > 532-7791
> > >>> 2219 Engineering Hall, Manhattan KS 66506
> > >>> Home: [email protected] <mailto:[email protected]>
> > >>> cell: (785) 770-5929
> > >>>
> > >>>
> > >>> _______________________________________________
> > >>> devel mailing list
> > >>> [email protected]
> > >>> https://lists.open-mpi.org/mailman/listinfo/devel
> > >>
> > >>
> > >> _______________________________________________
> > >> devel mailing list
> > >> [email protected]
> > >> https://lists.open-mpi.org/mailman/listinfo/devel
> > _______________________________________________
> > devel mailing list
> > [email protected]
> > https://lists.open-mpi.org/mailman/listinfo/devel
> >
>
>
>
> --
> Paul H. Hargrove <[email protected]>
> Computer Languages & Systems Software (CLaSS) Group
> Computer Science Department
> Lawrence Berkeley National Laboratory
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <https://lists.open-mpi.org/mailman/private/devel/
> attachments/20180123/ce0ba332/attachment.html>
>
> ------------------------------
>
> Message: 9
> Date: Wed, 24 Jan 2018 14:39:28 +0900
> From: Gilles Gouaillardet <[email protected]>
> To: Open MPI Developers <[email protected]>
> Subject: Re: [OMPI devel] Poor performance when compiling with
> --disable-dlopen
> Message-ID:
> <CAAkFZ5tCh0dw6=+RQXeLE_qw06S46qpireSEf5uaLa4=fysQWA@
> mail.gmail.com>
> Content-Type: text/plain; charset="UTF-8"
>
> Thanks Paul,
>
> unfortunately, that did not help :-(
>
> performance are just as bad even with --mca mpi_leave_pinned 1
>
> and surprisingly, when patcher/overwrite is used, performances are not
> worse with --mca mpi_leave_pinned 0
>
>
> Cheers,
>
> Gilles
>
> On Wed, Jan 24, 2018 at 2:28 PM, Paul Hargrove <[email protected]> wrote:
> > Ah, this sounds familiar.
> >
> > I believe that the issue Dave sees is that without patcher/overwrite the
> > "leave pinned" protocol is OFF by default.
> >
> > Use of '-mca mpi_leave_pinned 1' may help if my guess is right.
> > HOWEVER, w/o the memory management hooks provided using
> patcher/overwrite,
> > leave pinned can give incorrect results.
> >
> > -Paul
> >
> > On Tue, Jan 23, 2018 at 9:17 PM, Gilles Gouaillardet
> > <[email protected]> wrote:
> >>
> >> Dave,
> >>
> >> here is what I found
> >>
> >> - MPI_THREAD_MULTIPLE is not part of the equation (I just found it is
> >> no more required by IMB by default)
> >> - patcher/overwrite is not built when Open MPI is configure'd with
> >> --disable-dlopen
> >> - when configure'd without --disable-dlopen, performances are way
> >> worst for the IMB (PingPong) benchmark when ran with
> >> mpirun --mca patcher ^overwrite
> >> - OSU (osu_bw) performances are not impacted by the patcher/overwrite
> >> component being blacklisted
> >>
> >> I am afraid that's all I can do ...
> >>
> >>
> >> Nathan,
> >>
> >> could you please shed some light ?
> >>
> >>
> >> Cheers,
> >>
> >> Gilles
> >>
> >> On Wed, Jan 24, 2018 at 1:29 PM, Gilles Gouaillardet
> >> <[email protected]> wrote:
> >> > Dave,
> >> >
> >> > i can reproduce the issue with btl/openib and the IMB benchmark, that
> >> > is known to MPI_Init_thread(MPI_THREAD_MULTIPLE)
> >> >
> >> > note performance is ok with OSU benchmark that does not require
> >> > MPI_THREAD_MULTIPLE
> >> >
> >> > Cheers,
> >> >
> >> > Gilles
> >> >
> >> > On Wed, Jan 24, 2018 at 1:16 PM, Gilles Gouaillardet <
> [email protected]>
> >> > wrote:
> >> >> Dave,
> >> >>
> >> >>
> >> >> one more question, are you running the openib/btl ? or other
> libraries
> >> >> such
> >> >> as MXM or UCX ?
> >> >>
> >> >>
> >> >> Cheers,
> >> >>
> >> >>
> >> >> Gilles
> >> >>
> >> >>
> >> >> On 1/24/2018 12:55 PM, Dave Turner wrote:
> >> >>>
> >> >>>
> >> >>> We compiled OpenMPI 2.1.1 using the EasyBuild configuration
> >> >>> for CentOS as below and tested on Mellanox QDR hardware.
> >> >>>
> >> >>> ./configure --prefix=/homes/daveturner/libs/openmpi-2.1.1c
> >> >>> --enable-shared
> >> >>> --enable-mpi-thread-multiple
> >> >>> --with-verbs
> >> >>> --enable-mpirun-prefix-by-default
> >> >>> --with-mpi-cxx
> >> >>> --enable-mpi-cxx
> >> >>> --with-hwloc=$EBROOTHWLOC
> >> >>> --disable-dlopen
> >> >>>
> >> >>> The red curve in the attached NetPIPE graph shows the poor
> performance
> >> >>> above
> >> >>> 8 kB for the uni-directional tests with bi-directional and aggregate
> >> >>> tests also showing similar problems. When I compile using the same
> >> >>> configuration but with the --disable-dlopen parameter removed then
> the
> >> >>> performance is very good as the green curve in the graph shows.
> >> >>>
> >> >>> We see the same problems with OpenMPI 2.0.2.
> >> >>> Replacing --disable-dlopen with --disable-mca-dso showed good
> >> >>> performance.
> >> >>> Replacing --disable-dlopen with --enable-static showed good
> >> >>> performance.
> >> >>> So it's only --disable-dlopen that leads to poor performance.
> >> >>>
> >> >>> http://netpipe.cs.ksu.edu
> >> >>>
> >> >>> Dave Turner
> >> >>>
> >> >>> --
> >> >>> Work: [email protected] <mailto:[email protected]> (785)
> >> >>> 532-7791
> >> >>> 2219 Engineering Hall, Manhattan KS 66506
> >> >>> Home: [email protected] <mailto:[email protected]>
> >> >>> cell: (785) 770-5929
> >> >>>
> >> >>>
> >> >>> _______________________________________________
> >> >>> devel mailing list
> >> >>> [email protected]
> >> >>> https://lists.open-mpi.org/mailman/listinfo/devel
> >> >>
> >> >>
> >> >> _______________________________________________
> >> >> devel mailing list
> >> >> [email protected]
> >> >> https://lists.open-mpi.org/mailman/listinfo/devel
> >> _______________________________________________
> >> devel mailing list
> >> [email protected]
> >> https://lists.open-mpi.org/mailman/listinfo/devel
> >
> >
> >
> >
> > --
> > Paul H. Hargrove <[email protected]>
> > Computer Languages & Systems Software (CLaSS) Group
> > Computer Science Department
> > Lawrence Berkeley National Laboratory
> >
> > _______________________________________________
> > devel mailing list
> > [email protected]
> > https://lists.open-mpi.org/mailman/listinfo/devel
>
>
> ------------------------------
>
> Message: 10
> Date: Tue, 23 Jan 2018 23:55:35 -0600
> From: Dave Turner <[email protected]>
> To: Gilles Gouaillardet <[email protected]>, Open MPI
> Developers <[email protected]>
> Subject: Re: [OMPI devel] Poor performance when compiling with
> --disable-dlopen
> Message-ID:
> <CAFGXdkyf5QJ46RrTT1=fXZr0uBg8xAHNV=C0GPAV3OEnzupxx
> [email protected]>
> Content-Type: text/plain; charset="utf-8"
>
> Gilles,
>
> I'm using NetPIPE which is available at http://netpipe.cs.ksu.edu
> My base test is uni-directional with 1 process on a node communicating
> with a process on a second node.
>
> make mpi
> mpirun -np 2 --hostfile=hf.2p2n NPmpi
> cat hf.2p2n
> node0 slots=1
> node1 slots=1
>
> NetPIPE does not do any MPI_Init_thread().
> Tests on the configs below give good performance with and without the
> --enable-mpi-thread-multiple
> so I don't think that's the issue.
>
> configure --prefix=/homes/daveturner/libs/openmpi-2.1.1
> --enable-mpi-fortran=all --with-verbs --enable-ipv6 --enable-mpi-cxx
> configure --prefix=/homes/daveturner/libs/openmpi-2.1.1
> --enable-mpi-fortran=all --with-verbs --enable-ipv6 --enable-mpi-cxx
> --enable-mpi-thread-multiple
>
> Dave
>
> On Tue, Jan 23, 2018 at 10:03 PM, Gilles Gouaillardet <
> [email protected]> wrote:
>
> > Dave,
> >
> > At first glance, that looks pretty odd, and I'll have a look at it.
> >
> > Which benchmark are you using to measure the bandwidth ?
> > Does your benchmark MPI_Init_thread(MPI_THREAD_MULTIPLE) ?
> > Have you tried without --enable-mpi-thread-multiple ?
> >
> > Cheers,
> >
> > Gilles
> >
> > On Wed, Jan 24, 2018 at 12:55 PM, Dave Turner <[email protected]>
> > wrote:
> > >
> > > We compiled OpenMPI 2.1.1 using the EasyBuild configuration
> > > for CentOS as below and tested on Mellanox QDR hardware.
> > >
> > > ./configure --prefix=/homes/daveturner/libs/openmpi-2.1.1c
> > > --enable-shared
> > > --enable-mpi-thread-multiple
> > > --with-verbs
> > > --enable-mpirun-prefix-by-default
> > > --with-mpi-cxx
> > > --enable-mpi-cxx
> > > --with-hwloc=$EBROOTHWLOC
> > > --disable-dlopen
> > >
> > > The red curve in the attached NetPIPE graph shows the poor performance
> > above
> > > 8 kB for the uni-directional tests with bi-directional and aggregate
> > > tests also showing similar problems. When I compile using the same
> > > configuration but with the --disable-dlopen parameter removed then the
> > > performance is very good as the green curve in the graph shows.
> > >
> > > We see the same problems with OpenMPI 2.0.2.
> > > Replacing --disable-dlopen with --disable-mca-dso showed good
> > performance.
> > > Replacing --disable-dlopen with --enable-static showed good
> performance.
> > > So it's only --disable-dlopen that leads to poor performance.
> > >
> > > http://netpipe.cs.ksu.edu
> > >
> > > Dave Turner
> > >
> > > --
> > > Work: [email protected] (785) 532-7791
> > > 2219 Engineering Hall, Manhattan KS 66506
> > > Home: [email protected]
> > > cell: (785) 770-5929
> > >
> > > _______________________________________________
> > > devel mailing list
> > > [email protected]
> > > https://lists.open-mpi.org/mailman/listinfo/devel
> >
>
>
>
> --
> Work: [email protected] (785) 532-7791
> 2219 Engineering Hall, Manhattan KS 66506
> Home: [email protected]
> cell: (785) 770-5929
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <https://lists.open-mpi.org/mailman/private/devel/
> attachments/20180123/05fb1735/attachment.html>
>
> ------------------------------
>
> Subject: Digest Footer
>
> _______________________________________________
> devel mailing list
> [email protected]
> https://lists.open-mpi.org/mailman/listinfo/devel
>
> ------------------------------
>
> End of devel Digest, Vol 3576, Issue 1
> **************************************
>
--
Work: [email protected] (785) 532-7791
2219 Engineering Hall, Manhattan KS 66506
Home: [email protected]
cell: (785) 770-5929
_______________________________________________
devel mailing list
[email protected]
https://lists.open-mpi.org/mailman/listinfo/devel