Re: [OMPI users] OMPI users] OMPI users] Possible Memory Leak in simple PingPong-Routine with OpenMPI 1.8.3?

2014-10-28 Thread Gilles Gouaillardet
Yep, will do today Ralph Castain wrote: >Gilles: will you be committing this to trunk and PR to 1.8? > > >> On Oct 28, 2014, at 11:05 AM, Marco Atzeri wrote: >> >> On 10/28/2014 4:41 PM, Gilles Gouaillardet wrote: >>> Thanks Marco, >>> >>>

Re: [OMPI users] OMPI users] Possible Memory Leak in simple PingPong-Routine with OpenMPI 1.8.3?

2014-10-28 Thread Ralph Castain
Gilles: will you be committing this to trunk and PR to 1.8? > On Oct 28, 2014, at 11:05 AM, Marco Atzeri wrote: > > On 10/28/2014 4:41 PM, Gilles Gouaillardet wrote: >> Thanks Marco, >> >> pthread_mutex_init calls calloc under cygwin but does not allocate memory >>

Re: [OMPI users] SIGBUS in openmpi-dev-178-ga16c1e4 on Solaris 10 Sparc

2014-10-28 Thread Siegmar Gross
Hi Gilles, > From the jvm logs, there is an alignment error in native_get_attr > but i could not find it by reading the source code. > > Could you please do > ulimit -c unlimited > mpiexec ... > and then > gdb /bin/java core > And run bt on all threads until you get a line number in

Re: [OMPI users] OMPI users] Possible Memory Leak in simple PingPong-Routine with OpenMPI 1.8.3?

2014-10-28 Thread Marco Atzeri
On 10/28/2014 4:41 PM, Gilles Gouaillardet wrote: Thanks Marco, pthread_mutex_init calls calloc under cygwin but does not allocate memory under linux, so not invoking pthread_mutex_destroy causes a memory leak only under cygwin. Gilles thanks for the work . uploading 1.8.3-2 on

[OMPI users] Allgather in OpenMPI 1.4.3

2014-10-28 Thread Sebastian Rettenberger
Hi, I know 1.4.3 is really old but I am currently stuck with it. However, there seems to be a bug in Allgather. I have attached the source of an example program. The output I would expect is: rettenbs@hpcsccs4:/tmp$ mpiexec -np 5 ./a.out 0 0 1 2 1 0 1 2 2 0 1 2 3 0 1 2 4 0 1 2 But what I

Re: [OMPI users] OMPI users] Possible Memory Leak in simple PingPong-Routine with OpenMPI 1.8.3?

2014-10-28 Thread Gilles Gouaillardet
Thanks Marco, pthread_mutex_init calls calloc under cygwin but does not allocate memory under linux, so not invoking pthread_mutex_destroy causes a memory leak only under cygwin. Gilles Marco Atzeri wrote: >On 10/28/2014 12:04 PM, Gilles Gouaillardet wrote: >> Marco,

Re: [OMPI users] SIGBUS in openmpi-dev-178-ga16c1e4 on Solaris 10 Sparc

2014-10-28 Thread Gilles Gouaillardet
Hi Siegmar, From the jvm logs, there is an alignment error in native_get_attr but i could not find it by reading the source code. Could you please do ulimit -c unlimited mpiexec ... and then gdb /bin/java core And run bt on all threads until you get a line number in native_get_attr Thanks

Re: [OMPI users] Possible Memory Leak in simple PingPong-Routine with OpenMPI 1.8.3?

2014-10-28 Thread Marco Atzeri
On 10/28/2014 12:04 PM, Gilles Gouaillardet wrote: Marco, here is attached a patch that fixes the issue /* i could not find yet why this does not occurs on Linux ... */ could you please give it a try ? Cheers, Gilles It solves the issue on 64 bit. I see no growing memory usage anymore I

Re: [OMPI users] Java FAQ Page out of date

2014-10-28 Thread Jeff Squyres (jsquyres)
Thanks Brock; I opened https://github.com/open-mpi/ompi/issues/254 to track the issue. On Oct 27, 2014, at 12:57 AM, Brock Palen wrote: > I think a lot of the information on this page: > > http://www.open-mpi.org/faq/?category=java > > Is out of date with the 1.8 release.

Re: [OMPI users] MPI_Init seems to hang, but works after a, minute or two

2014-10-28 Thread Jeff Squyres (jsquyres)
On Oct 28, 2014, at 9:02 AM, maxinator333 wrote: > It doesn't seem to work. (switching off wlan still works) > mpicc mpiinit.c -o mpiinit.exe; time mpirun --mca btl sm,self -n 2 > ./mpiinit.exe > > real0m43.733s > user0m0.888s > sys 0m0.824s Ah, this

Re: [OMPI users] MPI_Init seems to hang, but works after a, minute or two

2014-10-28 Thread maxinator333
It doesn't seem to work. (switching off wlan still works) mpicc mpiinit.c -o mpiinit.exe; time mpirun --mca btl sm,self -n 2 ./mpiinit.exe real0m43.733s user0m0.888s sys 0m0.824s Am 28.10.2014 13:40, schrieb Jeff Squyres (jsquyres): On Oct 27, 2014, at 1:25 PM, maxinator333

Re: [OMPI users] MPI_Init seems to hang, but works after a, minute or two

2014-10-28 Thread Jeff Squyres (jsquyres)
On Oct 27, 2014, at 1:25 PM, maxinator333 wrote: > Deactivating my WLAN did indeed the trick! > It also seems to not work, if a LAN-cable is plugged in. No difference if I > am correctly connected (to the internet/gateway) or not (wrong IP, e.g. > static given IP

Re: [OMPI users] OpenMPI 1.8.3 configure fails, Mac OS X 10.9.5, Intel Compilers

2014-10-28 Thread Jeff Squyres (jsquyres)
It sounds like your intel compiler installation is broken -- these types of "present but not compilable" kinds of errors usually indicate that the compiler itself has some kind of local conflict that is unrelated to Open MPI (that's why we put those tests in OMPI's configure -- so that we can

[OMPI users] SIGBUS in openmpi-dev-178-ga16c1e4 on Solaris 10 Sparc

2014-10-28 Thread Siegmar Gross
Hi, today I installed openmpi-dev-178-ga16c1e4 on Solaris 10 Sparc with gcc-4.9.1 and Java 8. Now a very simple Java program works as expected, but other Java programs still break. I removed the warnings about "shmem.jar" and used the following configure command. tyr

Re: [OMPI users] Possible Memory Leak in simple PingPong-Routine with OpenMPI 1.8.3?

2014-10-28 Thread Gilles Gouaillardet
Marco, here is attached a patch that fixes the issue /* i could not find yet why this does not occurs on Linux ... */ could you please give it a try ? Cheers, Gilles On 2014/10/27 18:45, Marco Atzeri wrote: > > > On 10/27/2014 10:30 AM, Gilles Gouaillardet wrote: >> Hi, >> >> i tested on a

Re: [OMPI users] OMPI users] OMPI users] OMPI users] which info is needed for SIGSEGV inJava foropenmpi-dev-124-g91e9686on Solaris

2014-10-28 Thread Ralph Castain
> On Oct 27, 2014, at 7:21 PM, Gilles Gouaillardet > wrote: > > Ralph, > > On 2014/10/28 0:46, Ralph Castain wrote: >> Actually, I propose to also remove that issue. Simple enough to use a >> hash_table_32 to handle the jobids, and let that point to a >>