Yann,

Please post the test case that evidences the issue.

What is the minimal config required to reproduce it
(E.g. number of nodes and tasks per node)

If more than one node, which interconnect are you using ?
Out of curiosity, what if you
mpirun --mca mpi_leave_pinned 0 ...
or
mpirun --mca btl tcp,self --mca pml ob1 ...


Cheers,

Gilles

Yann Jobic <yann.jo...@univ-amu.fr> wrote:
>Hi all,
>
>I'm going crazy about a possible bug in my code. I'm using a derived mpi 
>datatype in a sendrecv function.
>The problem is that the memory footprint of my code is growing as time 
>increases.
>The problem is not showing with a regular datatype, as MPI_DOUBLE.
>I don't have this problem for openmpi 1.8.4, but it's present for 1.10.1 
>and 2.0.1
>
>The key parts of the code are (i'm using a 1D array with a macro in 
>order to be 3D) :
>
>Definition of the datatype:
>
>   MPI_Type_vector( Ny, 1, Nx, MPI_DOUBLE, &mpi.MPI_COL );
>   MPI_Type_commit( &mpi.MPI_COL ) ;
>
>And the sendrecv part:
>
>       MPI_Sendrecv( &(thebigone[_(1,0,k)])    , 1    , mpi.MPI_COL , 
>mpi.left , 3, \
>               &(thebigone[_(Nx-1,0,k)]) , 1    , mpi.MPI_COL , 
>mpi.right, 3, \
>               mpi.com, &mpi.stat );
>
>Is it coming from my code ?
>
>I isolated the communications in a small code (500 lines). I can give it 
>in order to reproduce the problem.
>
>Thanks,
>
>Yann
>
>
>---
>L'absence de virus dans ce courrier électronique a été vérifiée par le 
>logiciel antivirus Avast.
>https://www.avast.com/antivirus
>
>_______________________________________________
>users mailing list
>users@lists.open-mpi.org
>https://rfd.newmexicoconsortium.org/mailman/listinfo/users
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to