[OMPI users] MPI_WAIT hangs after a call to MPI_CANCEL

2017-04-01 Thread McGrattan, Kevin B. Dr. (Fed)
I am running a large computational fluid dynamics code on a linux cluster (Centos 6.8, Open MPI 1.8.4). The code is written in Fortran and compiled with Intel Fortran 16.0.3. The cluster has 36 nodes, each node has two sockets, each socket has six cores. I have noticed that the code hangs when t

Re: [OMPI users] MPI_WAIT hangs after a call to MPI_CANCEL

2017-04-03 Thread McGrattan, Kevin B. Dr. (Fed)
Users Cc: McGrattan, Kevin B. Dr. (Fed) Subject: Re: [OMPI users] MPI_WAIT hangs after a call to MPI_CANCEL Kevin, In Open MPI we only support cancelling non-yet matched receives. So, you cannot cancel sends nor receive requests that have already been matched. While the latter are supposed to

Re: [OMPI users] MPI_WAIT hangs after a call to MPI_CANCEL

2017-04-05 Thread McGrattan, Kevin B. Dr. (Fed)
just ignore a dropped packet now and again, or try to figure out what’s wrong with our IB. Thanks Kevin From: George Bosilca [mailto:bosi...@icl.utk.edu] Sent: Monday, April 03, 2017 5:59 PM To: McGrattan, Kevin B. Dr. (Fed) Cc: Open MPI Users Subject: Re: [OMPI users] MPI_WAIT hangs after a

Re: [OMPI users] MPI_WAIT hangs after a call to MPI_CANCEL

2017-04-07 Thread McGrattan, Kevin B. Dr. (Fed)
’re going to “diff” our cluster with another here to see if we can find some setting that is different. Thanks for your help. Kevin From: George Bosilca [mailto:bosi...@icl.utk.edu] Sent: Thursday, April 06, 2017 6:31 PM To: McGrattan, Kevin B. Dr. (Fed) Cc: Open MPI Users Subject: Re: [OM

[OMPI users] Memory leak in Open MPI 2.1.1

2017-07-28 Thread McGrattan, Kevin B. Dr. (Fed)
I am using Open MPI 2.1.1 along with Intel Fortran 17 update 4 and I am experiencing what I think is a memory leak with a job that uses 184 MPI processes. The memory used per process appears to be increasing by about 1 to 2 percent per hour. My code uses mostly persistent sends and receives to e

[OMPI users] Problem with MPI_FILE_WRITE_AT

2017-09-15 Thread McGrattan, Kevin B. Dr. (Fed)
I am using MPI_FILE_WRITE_AT to print out the timings of subroutines in a big Fortran code. I have noticed since upgrading to Open MPI 2.1.1 that sometimes the file to be written is corrupted. Each MPI process is supposed to write out a character string that is 159 characters in length, plus a l