Great. I will try the 3.0.x version to see how it goes.

On a side note, I did manage to run EPW without getting memory leaks using 
openmpi-1.8.8 and gcc-4.8.5. These are the tools that apparently worked when 
the code was developed as seen on their Test Farm 
(http://epw.org.uk/Main/TestFarm).

Thanks,

Vahid

On Jan 11, 2018, at 12:50 PM, Jeff Squyres (jsquyres) 
<jsquy...@cisco.com<mailto:jsquy...@cisco.com>> wrote:

You are correct: 3.0.1 has not been released yet.

However, our nightly snapshots of the 3.0.x branch are available for download.  
These are not official releases, but they are great for getting users to test 
what will eventually become an official release (i.e., 3.0.1) to see if 
particular bugs have been fixed.  This is one of the benefits of open source.  
:-)

Here's where the 3.0.1 nightly snapshots are available for download:

   https://www.open-mpi.org/nightly/v3.0.x/

They are organized by date.


On Jan 11, 2018, at 11:34 AM, Vahid Askarpour 
<vh261...@dal.ca<mailto:vh261...@dal.ca>> wrote:

Hi Jeff,

I looked for the 3.0.1 version but I only found the 3.0.0 version available for 
download. So I thought it may take a while for the 3.0.1 to become available. 
Or did I miss something?

Thanks,

Vahid

On Jan 11, 2018, at 12:04 PM, Jeff Squyres (jsquyres) 
<jsquy...@cisco.com<mailto:jsquy...@cisco.com>> wrote:

Vahid --

Were you able to give it a whirl?

Thanks.


On Jan 5, 2018, at 7:58 PM, Vahid Askarpour 
<vh261...@dal.ca<mailto:vh261...@dal.ca>> wrote:

Gilles,

I will try the 3.0.1rc1 version to see how it goes.

Thanks,

Vahid

On Jan 5, 2018, at 8:40 PM, Gilles Gouaillardet 
<gilles.gouaillar...@gmail.com<mailto:gilles.gouaillar...@gmail.com>> wrote:

Vahid,

This looks like the description of the issue reported at 
https://github.com/open-mpi/ompi/issues/4336
The fix is currently available in 3.0.1rc1, and I will back port the fix fo the 
v2.x branch.
A workaround is to use ROM-IO instead of ompio, you can achieve this with
mpirun —mca io ^ompio ...
(FWIW 1.10 series use ROM-IO by default, so there is no leak out of the box)

IIRC, a possible (and ugly) workaround for the compilation issue is to
configure —with-ucx=/usr ...
That being said, you should really upgrade to a supported version of Open MPI 
as previously suggested

Cheers,

Gilles

On Saturday, January 6, 2018, Jeff Squyres (jsquyres) 
<jsquy...@cisco.com<mailto:jsquy...@cisco.com>> wrote:
You can still give Open MPI 2.1.1 a try.  It should be source compatible with 
EPW.  Hopefully the behavior is close enough that it should work.

If not, please encourage the EPW developers to upgrade.  v3.0.x is the current 
stable series; v1.10.x is ancient.



On Jan 5, 2018, at 5:22 PM, Vahid Askarpour 
<vh261...@dal.ca<mailto:vh261...@dal.ca>> wrote:

Thank you Jeff for your suggestion to use the v.2.1 series.

I am attempting to use openmpi with EPW. On the EPW website 
(http://epw.org.uk/Main/DownloadAndInstall), it is stated that:

Compatibility of EPW

EPW is tested and should work on the following compilers and libraries:

  • gcc640 serial
  • gcc640 + openmpi-1.10.7
  • intel 12 + openmpi-1.10.7
  • intel 17 + impi
  • PGI 17 + mvapich2.3
EPW is know to have the following incompatibilities with:

  • openmpi 2.0.2 (but likely on all the 2.x.x version): Works but memory leak. 
If you open and close a file a lot of times with openmpi 2.0.2, the memory 
increase linearly with the number of times the file is open.

So I am hoping to avoid the 2.x.x series and use the 1.10.7 version suggested 
by the EPW developers. However, it appears that this is not possible.

Vahid

On Jan 5, 2018, at 5:06 PM, Jeff Squyres (jsquyres) 
<jsquy...@cisco.com<mailto:jsquy...@cisco.com>> wrote:

I forget what the underlying issue was, but this issue just came up and was 
recently fixed:

https://github.com/open-mpi/ompi/issues/4345

However, the v1.10 series is fairly ancient -- the fix was not applied to that 
series.  The fix was applied to the v2.1.x series, and a snapshot tarball 
containing the fix is available here (generally just take the latest tarball):

https://www.open-mpi.org/nightly/v2.x/

The fix is still pending for the v3.0.x and v3.1.x series (i.e., there are 
pending pull requests that haven't been merged yet, so the nightly snapshots 
for the v3.0.x and v3.1.x branches do not yet contain this fix).



On Jan 5, 2018, at 1:34 PM, Vahid Askarpour 
<vh261...@dal.ca<mailto:vh261...@dal.ca>> wrote:

I am attempting to install openmpi-1.10.7 on CentOS Linux (7.4.1708) using 
GCC-6.4.0.

When compiling, I get the following error:

make[2]: Leaving directory '/home/vaskarpo/bin/openmpi-1.10.7/ompi/mca/pml/ob1'
Making all in mca/pml/ucx
make[2]: Entering directory '/home/vaskarpo/bin/openmpi-1.10.7/ompi/mca/pml/ucx'
CC       pml_ucx.lo
CC       pml_ucx_request.lo
CC       pml_ucx_datatype.lo
CC       pml_ucx_component.lo
CCLD     mca_pml_ucx.la<http://mca_pml_ucx.la>
libtool:   error: require no space between '-L' and '-lrt'
make[2]: *** [Makefile:1725: mca_pml_ucx.la<http://mca_pml_ucx.la>] Error 1
make[2]: Leaving directory '/home/vaskarpo/bin/openmpi-1.10.7/ompi/mca/pml/ucx'
make[1]: *** [Makefile:3261: all-recursive] Error 1
make[1]: Leaving directory '/home/vaskarpo/bin/openmpi-1.10.7/ompi'
make: *** [Makefile:1777: all-recursive] Error 1

Thank you,

Vahid
_______________________________________________
users mailing list
users@lists.open-mpi.org<mailto:users@lists.open-mpi.org>
https://lists.open-mpi.org/mailman/listinfo/users


--
Jeff Squyres
jsquy...@cisco.com<mailto:jsquy...@cisco.com>



_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

_______________________________________________
users mailing list
users@lists.open-mpi.org<mailto:users@lists.open-mpi.org>
https://lists.open-mpi.org/mailman/listinfo/users


--
Jeff Squyres
jsquy...@cisco.com<mailto:jsquy...@cisco.com>



_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

_______________________________________________
users mailing list
users@lists.open-mpi.org<mailto:users@lists.open-mpi.org>
https://lists.open-mpi.org/mailman/listinfo/users


--
Jeff Squyres
jsquy...@cisco.com<mailto:jsquy...@cisco.com>



_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

_______________________________________________
users mailing list
users@lists.open-mpi.org<mailto:users@lists.open-mpi.org>
https://lists.open-mpi.org/mailman/listinfo/users


--
Jeff Squyres
jsquy...@cisco.com<mailto:jsquy...@cisco.com>



_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to