Re: [OMPI devel] RFC: OB1 optimizations

2014-01-10 Thread Nathan Hjelm
request is not correctly released, leading to assert in > >> debug > >>>> mode. This is because you avoid calling > >> MCA_PML_BASE_RECV_REQUEST_FINI, > >>>> fact that leaves the request in an ACTIVE state, condition > >>

Re: [OMPI devel] RFC: OB1 optimizations

2014-01-10 Thread George Bosilca
>>>> >>>> I attached a second patch that fixes the issue above, and >> implement a >>>> similar optimization for the blocking send. >>>> >>>> Unfortunately, this is not enough. The mca_pml_ob1_send_inline >>>>

Re: [OMPI devel] RFC: OB1 optimizations

2014-01-08 Thread Nathan Hjelm
end_inline optimization on the multithreaded >case. All >>> the others optimizations should be safe in all cases. >>> >>> George. >>> >>> On Jan 8, 2014, at 01:15 , Shamis, Pavel <sham...@ornl.gov> &g

Re: [OMPI devel] RFC: OB1 optimizations

2014-01-08 Thread Paul Hargrove
ization for the blocking send. >> >> >> >> Unfortunately, this is not enough. The mca_pml_ob1_send_inline >> >> optimization is horribly wrong in a multithreaded case as it alter >> the >> >> send_sequence without storing it.

Re: [OMPI devel] RFC: OB1 optimizations

2014-01-08 Thread Paul Hargrove
the > send_sequence > >> a deadlock will __definitively__ occur. I strongly suggest you turn > off > >> the mca_pml_ob1_send_inline optimization on the multithreaded case. > All > >> the others optimizations should be safe in all cases

Re: [OMPI devel] RFC: OB1 optimizations

2014-01-08 Thread Nathan Hjelm
e as it alter the > >> send_sequence without storing it. If you create a gap in the > >> send_sequence > >> a deadlock will __definitively__ occur. I strongly suggest you turn off > >> the mca_pml_ob1_send_inline optimization on the multithreaded case. All

Re: [OMPI devel] RFC: OB1 optimizations

2014-01-08 Thread George Bosilca
>> >>> Overall it looks good. It would be helpful to validate performance >> numbers for other interconnects as well. >>> -Pasha >>> >>>> -Original Message- >>>> From: devel [mailto:devel-boun...@open-mpi.org] On Behalf Of N

Re: [OMPI devel] RFC: OB1 optimizations

2014-01-08 Thread Nathan Hjelm
elm > >> Sent: Tuesday, January 07, 2014 6:45 PM >>> To: Open MPI Developers List >>> Subject: [OMPI devel] RFC: OB1 optimizations >>> >>> What: Push some ob1 optimizations to the trunk and 1.7.5. >>> >>&g

Re: [OMPI devel] RFC: OB1 optimizations

2014-01-07 Thread George Bosilca
;> Sent: Tuesday, January 07, 2014 6:45 PM >> To: Open MPI Developers List >> Subject: [OMPI devel] RFC: OB1 optimizations >> >> What: Push some ob1 optimizations to the trunk and 1.7.5. >> >> What: This patch contains two optimizations: >> >>

Re: [OMPI devel] RFC: OB1 optimizations

2014-01-07 Thread Shamis, Pavel
Developers List > Subject: [OMPI devel] RFC: OB1 optimizations > > What: Push some ob1 optimizations to the trunk and 1.7.5. > > What: This patch contains two optimizations: > > - Introduce a fast send path for blocking send calls. This path uses > the btl sendi funct

[OMPI devel] RFC: OB1 optimizations

2014-01-07 Thread Nathan Hjelm
What: Push some ob1 optimizations to the trunk and 1.7.5. What: This patch contains two optimizations: - Introduce a fast send path for blocking send calls. This path uses the btl sendi function to put the data on the wire without the need for setting up a send request. In the case of