Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Tom Kacvinsky via users
ok.  But deleting the whole source tree and re-configuring + re-building also works.  🙂  From: Jeffrey Layton <layto...@gmail.com> Sent: Tuesday, July 18, 2023 11:38 AM To: Jeff Squyres (jsquyres) <jsquy...@cisco.com> Cc: Open MPI Users <users@lists.open-mpi.org> Subject: Re: [OMPI u

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Bernstein, Noam CIV USN NRL (6393) Washington DC (USA) via users
-ran configure, it probably would have built ok. But deleting the whole source tree and re-configuring + re-building also works. 🙂 From: Jeffrey Layton mailto:layto...@gmail.com>> Sent: Tuesday, July 18, 2023 11:38 AM To: Jeff Squyres (jsquyres) mailto:jsquy..

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Jeff Squyres (jsquyres) via users
n Sent: Tuesday, July 18, 2023 12:51 PM To: Jeff Squyres (jsquyres) Cc: Open MPI Users Subject: Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3 As soon as you pointed out /usr/lib/gcc/x86_64-linux-gnu/9/include/float.h that made me think of the previous build. I did "make clean"

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Jeffrey Layton via users
source tree and re-configuring > + re-building also works. 🙂 > -- > *From:* Jeffrey Layton > *Sent:* Tuesday, July 18, 2023 11:38 AM > *To:* Jeff Squyres (jsquyres) > *Cc:* Open MPI Users > *Subject:* Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3 > > Jeff, >

Re: [OMPI users] Error handling

2023-07-18 Thread George Bosilca via users
Alex, How are your values "random" if you provide correct values ? Even for negative values you could use MIN to pick one value and return it. What is the problem with `MPI_Abort` ? it does seem to do what you want. George. On Tue, Jul 18, 2023 at 4:38 AM Alexander Stadik via users < users@li

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Jeff Squyres (jsquyres) via users
n MPI Users Subject: Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3 Jeff, Thanks for the tip - it started me thinking a bit. I was using a directory in my /home account with 4.1.5 that I had previously built using GCC 9.4 (Ubuntu 20.04). I rebuilt the system with Ubuntu-22.04 but I d

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Jeffrey Layton via users
> makefiles? > > Did you run configure on one machine and make on a different machine, > perchance? > -- > *From:* users on behalf of Jeffrey > Layton via users > *Sent:* Monday, July 17, 2023 2:05 PM > *To:* Open MPI Users > *Cc:* Jeffrey

[OMPI users] Error handling

2023-07-18 Thread Alexander Stadik via users
Hey everyone, I am working for longer time now with cuda-aware OpenMPI, and developed longer time back a small exceptions handling framework including MPI and CUDA exceptions. Currently I am using MPI_Abort with costum error numbers, to terminate everything elegantly, which works well, by just

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-17 Thread Jeff Squyres (jsquyres) via users
ut then how did it get into Open MPI's makefiles? Did you run configure on one machine and make on a different machine, perchance? From: users on behalf of Jeffrey Layton via users Sent: Monday, July 17, 2023 2:05 PM To: Open MPI Users Cc: Jeffrey Layton

[OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-17 Thread Jeffrey Layton via users
Good afternoon, I'm trying to build Open MPI 4.1.5 using GCC 11.3. However, I get an error that I'm not sure how to correct. The error is, ... CC pscatter.lo CC piscatter.lo CC pscatterv.lo CC piscatterv.lo CC psend.lo CC psend_init.lo CC ps

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application (users Digest, Vol 4715, Issue 1)

2022-02-07 Thread David Perozzi via users
Hi Bernd, Thanks for your valuable input! Your suggested approach indeed seems like the correct one and is actually what I've always wanted to do. In the past, I've also asked our cluster support if there was this possibility, but they always suggested the following approach: export OMP_NUM

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-03 Thread David Perozzi via users
, 2019 [1,0]:opal:version:full:4.0.2 [1,0]:opal:version:repo:v4.0.2 [1,0]:opal:version:release_date:Oct 07, 2019 [1,0]:mpi-api:version:full:3.1.0 [1,0]:ident:4.0.2 Best Christoph - Original Message - From: "Open MPI Users" To: "Open MPI Users" Cc: "Ralph Castain&q

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-03 Thread Ralph Castain via users
release_date:Oct 07, 2019 [1,0]:opal:version:full:4.0.2 [1,0]:opal:version:repo:v4.0.2 [1,0]:opal:version:release_date:Oct 07, 2019 [1,0]:mpi-api:version:full:3.1.0 [1,0]:ident:4.0.2 Best Christoph - Original Message - From: "Open MPI Users" To: "Open MPI Users" Cc:

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application (users Digest, Vol 4715, Issue 1)

2022-02-03 Thread Bernd Dammann via users
Hi David, On 03/02/2022 00:03 , David Perozzi wrote: Helo, I'm trying to run a code implemented with OpenMPI and OpenMP (for threading) on a large cluster that uses LSF for the job scheduling and dispatch. The problem with LSF is that it is not very straightforward to allocate and bind the r

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-03 Thread David Perozzi via users
pi:version:repo:v4.0.2 [1,0]:ompi:version:release_date:Oct 07, 2019 [1,0]:orte:version:full:4.0.2 [1,0]:orte:version:repo:v4.0.2 [1,0]:orte:version:release_date:Oct 07, 2019 [1,0]:opal:version:full:4.0.2 [1,0]:opal:version:repo:v4.0.2 [1,0]:opal:version:release_date:Oct 07, 2019 [1,0]:mpi-api:versio

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-03 Thread Ralph Castain via users
ackage:Open MPI spackapps@eu-c7-042-03 Distribution >>> [1,0]:ompi:version:full:4.0.2 >>> [1,0]:ompi:version:repo:v4.0.2 >>> [1,0]:ompi:version:release_date:Oct 07, 2019 >>> [1,0]:orte:version:full:4.0.2 >>> [1,0]:orte:version:repo:v4.0.2 >>> [1,0]

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-03 Thread David Perozzi via users
2 [1,0]:opal:version:repo:v4.0.2 [1,0]:opal:version:release_date:Oct 07, 2019 [1,0]:mpi-api:version:full:3.1.0 [1,0]:ident:4.0.2 Best Christoph - Original Message - From: "Open MPI Users" To: "Open MPI Users" Cc: "Ralph Castain" Sent: Thursday, 3 February,

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-02 Thread Ralph Castain via users
al:version:full:4.0.2 > [1,0]:opal:version:repo:v4.0.2 > [1,0]:opal:version:release_date:Oct 07, 2019 > [1,0]:mpi-api:version:full:3.1.0 > [1,0]:ident:4.0.2 > > Best > Christoph > > - Original Message - > From: "Open MPI Users" > To: "Open MPI Users

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-02 Thread Christoph Niethammer via users
rs" Cc: "Ralph Castain" Sent: Thursday, 3 February, 2022 00:22:30 Subject: Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application Errr...what version OMPI are you using? > On Feb 2, 2022, at 3:03 PM, David Perozzi via use

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-02 Thread Ralph Castain via users
Errr...what version OMPI are you using? > On Feb 2, 2022, at 3:03 PM, David Perozzi via users > wrote: > > Helo, > > I'm trying to run a code implemented with OpenMPI and OpenMP (for threading) > on a large cluster that uses LSF for the job scheduling and dispatch. The > problem with LSF is

[OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-02 Thread David Perozzi via users
Helo, I'm trying to run a code implemented with OpenMPI and OpenMP (for threading) on a large cluster that uses LSF for the job scheduling and dispatch. The problem with LSF is that it is not very straightforward to allocate and bind the right amount of threads to an MPI rank inside a single

Re: [OMPI users] Error code for I/O operations

2021-07-01 Thread Aurelien Bouteiller via users
Eric, Error codes are **not** mandated to be consistent across ranks in the MPI standard. Although file errors are not expected to cause the program to abort, there is no special case for I/O in that regard in the MPI standard. The Open MPI implementation acts accordingly and will not automati

[OMPI users] Error code for I/O operations

2021-06-30 Thread Eric Chamberland via users
Hi, I have a simple question about error codes returned by MPI_File_*_all* and MPI_File_open/close functions: If an error is returned will it be the same for *all* processes? In other worlds, are error codes communicated under the hood so we, end users, can avoid to add "reduce" on those err

[OMPI users] Error Signal code: Address not mapped (1)

2021-06-21 Thread Arturo Fernandez via users
Hello, I'm getting the error message (with either v4.1.0 or v4.1.1) *** Process received signal *** Signal: Segmentation fault (11) Signal code: Address not mapped (1) Failing at address: (nil) *** End of error message *** Segmentation fault (core dumped) The AWS system is running CentOS8 but I do

Re: [OMPI users] Error intialising an OpenFabrics device.

2021-03-13 Thread Heinz, Michael William via users
I’ve begun getting this annoyingly generic warning, too. It appears to be coming from the openib provider. If you disable it with -mtl ^openib the warning goes away. Sent from my iPad > On Mar 13, 2021, at 3:28 PM, Bob Beattie via users > wrote: > > Hi everyone, > > To be honest, as an MPI

[OMPI users] Error intialising an OpenFabrics device.

2021-03-13 Thread Bob Beattie via users
Hi everyone, To be honest, as an MPI / IB noob, I don't know if this falls under OpenMPI or Mellanox Am running a small cluster of HP DL380 G6/G7 machines. Each runs Ubuntu server 20.04 and has a Mellanox ConnectX-3 card, connected by an IS dumb switch. When I begin my MPI program (snappy

Re: [OMPI users] Error with building OMPI with PGI

2021-01-19 Thread Passant A. Hafez via users
". make[2]: *** [param.o] Error 2 make[2]: Leaving directory `/tmp/openmpi-4.0.3/orte/tools/orte-info' make[1]: *** [install-recursive] Error 1 make[1]: Leaving directory `/tmp/openmpi-4.0.3/orte' make: *** [install-recursive] Error 1 Best, Passant _________

Re: [OMPI users] Error with building OMPI with PGI

2021-01-19 Thread Gilles Gouaillardet via users
n switch: +03 > > > Please advise. > > All the best, > Passant > > From: users on behalf of Gus Correa via > users > Sent: Friday, January 15, 2021 2:36 AM > To: Open MPI Users > Cc: Gus Correa > Subject: Re: [OMPI users] Err

Re: [OMPI users] Error with building OMPI with PGI

2021-01-19 Thread Passant A. Hafez via users
__ From: users on behalf of Gus Correa via users Sent: Friday, January 15, 2021 2:36 AM To: Open MPI Users Cc: Gus Correa Subject: Re: [OMPI users] Error with building OMPI with PGI Hi Passant, list This is an old problem with PGI. There are many threads in the OpenMPI mailing list archiv

Re: [OMPI users] Error with building OMPI with PGI

2021-01-14 Thread Gus Correa via users
Hi Passant, list This is an old problem with PGI. There are many threads in the OpenMPI mailing list archives about this, with workarounds. The simplest is to use FC="pgf90 -noswitcherror". Here are two out of many threads ... well, not pthreads! :) https://www.mail-archive.com/users@lists.open

[OMPI users] Error with building OMPI with PGI

2021-01-14 Thread Passant A. Hafez via users
Hello, I'm having an error when trying to build OMPI 4.0.3 (also tried 4.1) with PGI 20.1 ./configure CPP=cpp CC=pgcc CXX=pgc++ F77=pgf77 FC=pgf90 --prefix=$PREFIX --with-ucx=$UCX_HOME --with-slurm --with-pmi=/opt/slurm/cluster/ibex/install --with-cuda=$CUDATOOLKIT_HOME in the make install

Re: [OMPI users] error while running mpirun

2020-07-13 Thread Srijan Chatterjee via users
Hi Gilles, Thanks for the suggestion. If I do exclude sm in the command line in the following way it seems to work. @chimera examples]$ mpirun -np 1 --mca btl ^sm hello_c libibverbs: Warning: no userspace device-specific driver found for /sys/class/infiniband_verbs/uverbs0 libibverbs: Warning: n

Re: [OMPI users] error while running mpirun

2020-07-12 Thread Gilles Gouaillardet via users
Srijan, The logs suggest you explicitly request the btl/sm component, and this typically occurs via a openmpi-mca-params.conf (that contains a line such as btl = sm,openib,self), or the OMPI_MCA_btl environment variable Cheers, Gilles On Mon, Jul 13, 2020 at 1:50 AM Srijan Chatterjee via users

[OMPI users] error while running mpirun

2020-07-12 Thread Srijan Chatterjee via users
Dear Open MPI users, I am using the following system CentOS release 6.6 Rocks 6.2 I have been trying to install openmpi-3.1.3. After installing it, in the example folder if want to test run I got the following error I run command mpirun -np 1 hello_c [user@chimera examples]$ mpirun -np 1 hello_c

Re: [OMPI users] Error with MPI_GET_ADDRESS and MPI_TYPE_CREATE_RESIZED?

2020-05-18 Thread Jeff Squyres (jsquyres) via users
Was your Open MPI compiled with -r8? We definitely recommend using the same compiler flags to compile Open MPI as your application. As George noted, -r8 can definitely cause issues if one was compiled with it and the other was not. On May 18, 2020, at 12:24 AM, George Bosilca via users mail

Re: [OMPI users] Error with MPI_GET_ADDRESS and MPI_TYPE_CREATE_RESIZED?

2020-05-17 Thread George Bosilca via users
Diego, I see nothing wrong with the way you create the datatype. In fact this is the perfect example on how to almost do it right in FORTRAN. The almost is because your code is highly dependent on the -r8 compiler option (otherwise the REAL in your type will not match the MPI_DOUBLE_PRECISION you

Re: [OMPI users] Error with MPI_GET_ADDRESS and MPI_TYPE_CREATE_RESIZED?

2020-05-17 Thread Diego Avesani via users
Dear Gilles, dear All, as far I remember no. The compiler is the same as the options which I use. Maybe, the error is some other places of my code. However, the results look errors in allocation of sent and received vector of datatype. The import is that at least mydata type definitions in corre

Re: [OMPI users] Error with MPI_GET_ADDRESS and MPI_TYPE_CREATE_RESIZED?

2020-05-17 Thread Gilles Gouaillardet via users
Diego, Did you change your compiler options? Cheers, Gilles - Original Message - > Dear all, > > I would like to share with what I have done in oder to create my own MPI > data type. The strange thing is that it worked until some day ago and then > it stopped working. This because pr

[OMPI users] Error with MPI_GET_ADDRESS and MPI_TYPE_CREATE_RESIZED ?

2020-05-17 Thread Diego Avesani via users
Dear all, I would like to share with what I have done in oder to create my own MPI data type. The strange thing is that it worked until some day ago and then it stopped working. This because probably I have changed my data type and I miss some knowledge about MPI data type This is my data type: *

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-21 Thread Sangam B via users
Thanks Jeff for the resolution. On Mon, Aug 19, 2019 at 7:45 PM Jeff Squyres (jsquyres) wrote: > On Aug 19, 2019, at 6:15 AM, Sangam B via users > wrote: > > > > subroutine recv(this,lmb) > > class(some__example6), intent(inout) :: this > > integer, intent(in) :: lmb(2,2) > > > > i

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Jeff Squyres (jsquyres) via users
On Aug 19, 2019, at 6:15 AM, Sangam B via users wrote: > > subroutine recv(this,lmb) > class(some__example6), intent(inout) :: this > integer, intent(in) :: lmb(2,2) > > integer :: cs3, ierr > integer(kind=C_LONG) :: size This ^^ is your problem. More below. > ! receive

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Sangam B via users
Hi, Evan after recompiling OpenMPI with -fdefault-real-8, it fails with same error. It seems to me that, it's an issue with OpenMPI itself, because: Intel MPI + gnu compiler --- works Intel MPI + intel compiler --- works Open MPI + gnu compiler --- fails Open MPI + AOCC compiler --- fails Wil

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Gilles Gouaillardet via users
Thanks, but this is not really helping. Could you please build a Minimal, Reproducible Example as described at https://stackoverflow.com/help/minimal-reproducible-example ? Cheers, Gilles On Mon, Aug 19, 2019 at 7:19 PM Sangam B via users wrote: > > Hi, > > Here is the sample program snippet:

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Sangam B via users
Hi, Here is the sample program snippet: #include "intrinsic_sizes.h" #include "redef.h" module module1_m use mod1_m, only: some__example2 use mod2_m, only: some__example3 use mod3_m, only: some__example4 use mpi use, intrinsic :: iso_c_binding implicit none private

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Gilles Gouaillardet via users
I am not questioning whether you are facing an issue with Open MPI or not. I am just asking for "the same application" (read minimal source code) so I can reproduce the issue, investigate it and hopefully help you. Meanwhile, try rebuilding Open MPI with '-fdefault-real-8' in your FCFLAGS (since t

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Sangam B via users
Hi, I've tried both gcc-8.1.0 and AOCC-2.0 compilers with openmpi-3.1.1. It fails for both the compilers. Posted error message was from OpenMPI-3.1.1 + AOCC-2.0 compiler. To cross-check whether it is problem with OpenMPI or the base compiler, compiled the same application with Intel MPI using ba

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Gilles Gouaillardet via users
One more thing ... Your initial message mentioned a failure with gcc 8.2.0, but your follow-up message mentions LLVM compiler. So which compiler did you use to build Open MPI that fails to build your test ? Cheers, Gilles On Mon, Aug 19, 2019 at 6:49 PM Gilles Gouaillardet wrote: > > Thanks,

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Gilles Gouaillardet via users
Thanks, and your reproducer is ? Cheers, Gilles On Mon, Aug 19, 2019 at 6:42 PM Sangam B via users wrote: > > Hi, > > OpenMPI is configured as follows: > > export CC=`which clang` > export CXX=`which clang++` > export FC=`which flang` > export F90=`which flang` > > ../configure --prefix=/sw/op

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Sangam B via users
Hi, OpenMPI is configured as follows: export CC=`which clang` export CXX=`which clang++` export FC=`which flang` export F90=`which flang` ../configure --prefix=/sw/openmpi/3.1.1/aocc20hpcx210-mpifort --enable-mpi-fortran --enable-mpi-cxx --without-psm --without-psm2 --without-knem --without-libf

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Gilles Gouaillardet via users
Hi, Can you please post a full but minimal example that evidences the issue? Also please post your Open MPI configure command line. Cheers, Gilles Sent from my iPod > On Aug 19, 2019, at 18:13, Sangam B via users > wrote: > > Hi, > > I get following error if the application is compiled

[OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Sangam B via users
Hi, I get following error if the application is compiled with openmpi-3.1.1: mpifort -O3 -march=native -funroll-loops -finline-aggressive -flto -J./bin/obj_amd64aocc20 -std=f2008 -O3 -march=native -funroll-loops -finline-aggressive -flto -fallow-fortran-gnu-ext -ffree-form -fdefault-real-8 exampl

Re: [OMPI users] error running mpirun command

2019-05-16 Thread Eric F. Alemany via users
Hi Jeff, Thank you for picking up on the thread. I followed your suggestion and I also did some research. I think I fixed the issue. My problem was two things: 1) From the mater node I made sure i could do a password-less ssh connection to the other nodes via; IP address of the nodes, host

Re: [OMPI users] error running mpirun command

2019-05-14 Thread Jeff Squyres (jsquyres) via users
Looks like this thread accidentally got dropped; sorry! More below. > On May 4, 2019, at 10:40 AM, Eric F. Alemany via users > wrote: > > Hi Gilles, > > Thank you for your message and your suggestion. As you suggested i tried > mpirun -np 84 - -hostfile hostsfile --mca routed direct ./open

Re: [OMPI users] error running mpirun command

2019-05-04 Thread Eric F. Alemany via users
Hi Gilles, Thank you for your message and your suggestion. As you suggested i tried mpirun -np 84 - -hostfile hostsfile --mca routed direct ./openmpi_hello.c The command hangs with no message or error message until i hit "control + z". Then i have the same error message as before. To answer

Re: [OMPI users] error running mpirun command

2019-05-03 Thread Gilles Gouaillardet via users
Eric, which version of Open MPI are you using ? how many hosts in your hostsfile ? The error message suggests this could be a bug within Open MPI, and a potential workaround for you would be to try mpirun -np 84 - -hostfile hostsfile --mca routed direct ./openmpi_hello.c You might also want to

[OMPI users] error running mpirun command

2019-05-03 Thread Eric F. Alemany via users
Hello everyone, I am new to Open MPI please forgive me for my beginner mistake. I read through the FAQ of open-mpi.org website and built a small cluster (9 nodes - including a master node). I thought i followed the instructions accordingly but i am having issue running a si

[OMPI users] Error When Using Open MPI SHMEM with UCX

2019-04-23 Thread Benjamin Brock via users
And, to provide more details, I'm using a fresh vanilla build of Open MPI 4.0.1 with UCX 1.5.1 (`./configure --with-ucx=$DIR/ucx-1.5.1`). Ben ___ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users

[OMPI users] Error When Using Open MPI SHMEM with UCX

2019-04-23 Thread Benjamin Brock via users
I get the following error when trying to run SHMEM programs using UCX. [xiii@shini dir]$ oshrun -n 1 ./target/debug/main [1556046469.890238] [shini:19769:0]sys.c:619 UCX ERROR shmget(size=2097152 flags=0xfb0) for mm_recv_desc failed: Operation not permitted, please check shared memor

Re: [OMPI users] Error when Building an MPI Java program

2019-04-11 Thread Kawashima, Takahiro
Benson, I think OpenJDK 9 is enough. I tested Open MPI Java bindings with OpenJDK 8, 9, 10, and 11 several months ago. Compiling your MPI Java program requires mpi.jar, which should have been installed in $OMPI_INSTALL_DIR/lib. Probably your compilation error message indicates mpi.jar is not f

Re: [OMPI users] Error when Building an MPI Java program

2019-04-10 Thread Benson Muite
Thanks. The code is here: https://github.com/mboysan/ping-pong-mpi-tcp OpenMPI is built using: ./configure --prefix=/PATH_TO_HOME_DIRECTORY/openmpi-4.0.1-install \     --enable-mpi-java The code builds on a number of other machines. Are there any minimum java build requirements for Op

Re: [OMPI users] Error when Building an MPI Java program

2019-04-09 Thread Jeff Squyres (jsquyres) via users
Can you provide a small, standalone example + recipe to show the problem? > On Apr 8, 2019, at 6:45 AM, Benson Muite wrote: > > Hi > > Am trying to build an MPI Java program using OpenMPI 4.0.1: > > I get the following error: > > Compilation failure > /MY_DIRECTORY/ping-pong-mpi-tcp/src/mai

[OMPI users] Error when Building an MPI Java program

2019-04-08 Thread Benson Muite
Hi Am trying to build an MPI Java program using OpenMPI 4.0.1: I get the following error: Compilation failure /MY_DIRECTORY/ping-pong-mpi-tcp/src/main/java/SocketMainMultiJVM.java:[2,10] error: cannot access MPIException Any suggestions on what I am doing incorrectly? Thanks, Benson _

Re: [OMPI users] error "unacceptable operand for unary &" for openmpi-master-201903260242-dfbc144 on Linux with Sun C

2019-03-26 Thread Nathan Hjelm via users
This really looks like a compiler bug. There is no & @ osc_pt2pt.h line 579. There is one at line 577 but there is no “unacceptable operand” on that line. If I have time this week I will try to find a workaround but it might be worth filing a bug with Oracle and see what they say. -Nathan > On

[OMPI users] error "unacceptable operand for unary &" for openmpi-master-201903260242-dfbc144 on Linux with Sun C

2019-03-26 Thread Siegmar Gross
Hi, I've tried to install openmpi-master-201903260242-dfbc144 on my "SUSE Linux Enterprise Server 12.3 with Sun C 5.15 (Oracle Developer Studio 12.6). Unfortunately, I still get the following error that I reported some time ago: https://github.com/open-mpi/ompi/issues/6180. I'm able to build open

Re: [OMPI users] Error initializing an UCX / OpenFabrics device. #6300

2019-03-22 Thread Jeff Squyres (jsquyres) via users
Greetings Charlie. Yes, it looks like you replied to a closed issue on Github -- would you mind opening a new issue about it? You can certainly refer to the old issue for context. But replying to closed issues is a bit dangerous: if we miss the initial email from GitHub (and all of us have ov

[OMPI users] Error initializing an UCX / OpenFabrics device. #6300

2019-03-22 Thread Charles A Taylor
Anyone else running into the issue below with OpenMPI 4.0.0? https://github.com/open-mpi/ompi/issues/6300 (Error initializing an UCX / OpenFabrics device) I’m hitting it and don’t really see why. I posted to the bug but maybe I need to just o

Re: [OMPI users] error while using open mpi to compile parallel-netcdf-1.8.1

2019-03-07 Thread Jeff Squyres (jsquyres) via users
FWIW, the error message is telling you exactly what is wrong and provides a link to the FAQ item on how to fix it. It's a bit inelegant that it's segv'ing after that, but the real issue is what is described in the help message. > On Mar 6, 2019, at 3:07 PM, Zhifeng Yang wrote: > > Hi > > I

[OMPI users] error while using open mpi to compile parallel-netcdf-1.8.1

2019-03-06 Thread Zhifeng Yang
Hi I compiled openmpi-3.1.2 using pgi 18.10 on our IBM power9 machine. After that, I used openmpi-3.1.2 to compile parallel-netcdf-1.8.1. However, I got the following error. ./nc_test -c-d . [c699login01:12104] mca_base_component_repository_open: unable to open mca_plm_lsf: libbat.so: cannot

[OMPI users] error "unacceptable operand for unary &" for openmpi-master-201810310352-a1e85b0 on Linux with Sun C

2018-11-01 Thread Siegmar Gross
Hi, I've tried to install openmpi-master-201810310352-a1e85b0 on my "SUSE Linux Enterprise Server 12.3 (x86_64)" with Sun C 5.15 (Oracle Developer Studio 12.6). Unfortunately, I get the folowing error. loki openmpi-master-201810310352-a1e85b0-Linux.x86_64.64_cc 143 head -7 config.log | tail -1

Re: [OMPI users] error building openmpi-master-201810310352-a1e85b0 on Linux with Sun C

2018-11-01 Thread Siegmar Gross
ly, I still get the following error that I reported some time ago. https://users.open-mpi.narkive.com/lvQKElru/ompi-users-error-building-openmpi-master-201810050304-5f1c940-on-linux-with-sun-c loki openmpi-master-201810310352-a1e85b0-Linux.x86_64.64_cc 125 head -7 config.log | tail -1 $ ../

Re: [OMPI users] error building openmpi-master-201810310352-a1e85b0 on Linux with Sun C

2018-10-31 Thread Kawashima, Takahiro
ot; with Sun C 5.15 (Oracle Developer Studio > 12.6). Unfortunately, I still get the following error that I reported some > time ago. > > https://users.open-mpi.narkive.com/lvQKElru/ompi-users-error-building-openmpi-master-201810050304-5f1c940-on-linux-with-sun-c > > > loki

[OMPI users] error building openmpi-master-201810310352-a1e85b0 on Linux with Sun C

2018-10-31 Thread Siegmar Gross
Hi, I've tried to install openmpi-master-201810310352-a1e85b0 on my "SUSE Linux Enterprise Server 12.3 (x86_64)" with Sun C 5.15 (Oracle Developer Studio 12.6). Unfortunately, I still get the following error that I reported some time ago. https://users.open-mpi.narkive.com/lvQK

Re: [OMPI users] error building Java api for openmpi-v4.0.x-201810090241-2124192 and openmpi-master-201810090329-e9e4d2a

2018-10-09 Thread Kawashima, Takahiro
I confirmed the patch resolved the issue with OpenJDK 11. I've created a PR for the master branch. I'll also create PRs for release branches. Thanks to your bug report! https://github.com/open-mpi/ompi/pull/5870 Takahiro Kawashima, Fujitsu > Siegmar, > > I think you are using Java 11 and it c

Re: [OMPI users] error building Java api for openmpi-v4.0.x-201810090241-2124192 and openmpi-master-201810090329-e9e4d2a

2018-10-09 Thread Kawashima, Takahiro
Siegmar, I think you are using Java 11 and it changed the default output HTML version. I'll take a look. But downloading OpenJDK 11 takes time... Probably the following patch will resolve the issue. diff --git ompi/mpi/java/java/Com

Re: [OMPI users] error building Java api for openmpi-v4.0.x-201810090241-2124192 and openmpi-master-201810090329-e9e4d2a

2018-10-09 Thread Kawashima, Takahiro
What version of Java are you using? Could you type "java -version" and show the output? Takahiro Kawashima, Fujitsu > today I've tried to build openmpi-v4.0.x-201810090241-2124192 and > openmpi-master-201810090329-e9e4d2a on my "SUSE Linux Enterprise Server > 12.3 (x86_64)" with Sun C 5.15, gcc 6

[OMPI users] error building Java api for openmpi-v4.0.x-201810090241-2124192 and openmpi-master-201810090329-e9e4d2a

2018-10-09 Thread Siegmar Gross
Hi, today I've tried to build openmpi-v4.0.x-201810090241-2124192 and openmpi-master-201810090329-e9e4d2a on my "SUSE Linux Enterprise Server 12.3 (x86_64)" with Sun C 5.15, gcc 6.4.0, Intel icc 18.0.3, and Portland Group pgcc 18.4-0. Unfortunately, I get the following error for all seven version

[OMPI users] error building openmpi-master-201810050304-5f1c940 on Linux with Sun C

2018-10-05 Thread Siegmar Gross
Hi, I've tried to install openmpi-master-201810050304-5f1c940 on my "SUSE Linux Enterprise Server 12.3 (x86_64)" with Sun C 5.15 (Oracle Developer Studio 12.6). Unfortunately, I get the following error. loki openmpi-master-201810050304-5f1c940-Linux.x86_64.64_cc 128 head -7 config.log | tail -

Re: [OMPI users] error building openmpi-master-201809290304-73075b8 with Sun C 5.15

2018-10-01 Thread Jeff Squyres (jsquyres) via users
Siegmar -- I created a GitHub issue for this: https://github.com/open-mpi/ompi/issues/5814 Nathan posted a test program on there for you to try; can you try it and reply on the issue? Thanks. > On Oct 1, 2018, at 9:23 AM, Siegmar Gross > wrote: > > Hi, > > I've tried to install openm

[OMPI users] error building openmpi-master-201809290304-73075b8 with Sun C 5.15

2018-10-01 Thread Siegmar Gross
Hi, I've tried to install openmpi-master-201809290304-73075b8 on my "SUSE Linux Enterprise Server 12.3 (x86_64)" with Sun C 5.15 (Oracle Developer Studio 12.6). Unfortunately, I get the following error. loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_cc 151 head -7 config.log | tail -

Re: [OMPI users] *** Error in `orted': double free or corruption (out): 0x00002aaab4001680 ***, in some node combos.

2018-09-13 Thread Jeff Squyres (jsquyres) via users
On Sep 12, 2018, at 4:54 AM, Balázs Hajgató wrote: > > Setting mca oob to tcp works. I will stick to this solution in our production > environment. Great! > I am not sure that it is relevant, but I also tried the patch on a > non-procduction OpenMPI 3.1.1, and "mpirun -host nic114,nic151 host

Re: [OMPI users] *** Error in `orted': double free or corruption (out): 0x00002aaab4001680 ***, in some node combos.

2018-09-12 Thread Balázs Hajgató
Dear Jeff, Setting mca oob to tcp works. I will stick to this solution in our production environment. I am not sure that it is relevant, but I also tried the patch on a non-procduction OpenMPI 3.1.1, and "mpirun -host nic114,nic151 hostname" works without any parameters, but issuing the libi

Re: [OMPI users] *** Error in `orted': double free or corruption (out): 0x00002aaab4001680 ***, in some node combos.

2018-09-11 Thread Jeff Squyres (jsquyres) via users
Thanks for reporting the issue. First, you can workaround the issue by using: mpirun --mca oob tcp ... This uses a different out-of-band plugin (TCP) instead of verbs unreliable datagrams. Second, I just filed a fix for our current release branches (v2.1.x, v3.0.x, and v3.1.x): https

[OMPI users] *** Error in `orted': double free or corruption (out): 0x00002aaab4001680 ***, in some node combos.

2018-09-10 Thread Balazs HAJGATO
Dear list readers, I have some problems with OpenMPI 3.1.1. In some node combos, I got the error (libibverbs: GRH is mandatory For RoCE address handle; *** Error in `/apps/brussel/CO7/ivybridge-ib/software/OpenMPI/3.1.1-GCC-7.3.0-2.30/bin/orted': double free or corruption (out): 0x2aaab4001

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-28 Thread Lovelace III, Henry
:18 PM To: Open MPI Users Subject: Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO Hi Gilles, If I understand the instruction correctly by invoking the -showme command to give addition information. I did use the -showme command, however the files that are being

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-27 Thread Lovelace III, Henry
updated that is the reason behind the directory change. Locally I have no problem with the build. Thanks, Henry From: users on behalf of gil...@rist.or.jp Sent: Tuesday, July 24, 2018 10:00:51 PM To: Open MPI Users Subject: Re: [OMPI users] Error in file base/plm_

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-24 Thread gilles
or from the MPIfort? Thanks, Henry From: users on behalf of Brian Smith Sent: Monday, July 23, 2018 2:06:14 PM To: Open MPI Users Subject: Re: [OMPI users] Error in file base/plm_base_launch_support. c: OPAL_HWLOC_TOPO On Sat, Jul 21, 2018 at 9:13 PM, Gill

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-24 Thread Lovelace III, Henry
. The build is not seeing the mpi libraries even though before the build I defined export LD_LIBRARY_PATH=${LD_LIBRARY_PATH}:$HOME/BMAD/bmad_dist_2018_0717/production/lib Is this error from the MPIfort? Thanks, Henry From: users on behalf of Brian Smith Sent: M

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-23 Thread Brian Smith
On Sat, Jul 21, 2018 at 9:13 PM, Gilles Gouaillardet wrote: > Brian, > > As Ralph already stated, this is likely a hwloc API issue. > From debian9, you can > lstopo --of xml | ssh debian8 lstopo --if xml -i - > > that will likely confirm the API error. > > If you are willing to get a bit more deta

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-21 Thread Gilles Gouaillardet
Brian, As Ralph already stated, this is likely a hwloc API issue. From debian9, you can lstopo --of xml | ssh debian8 lstopo --if xml -i - that will likely confirm the API error. If you are willing to get a bit more details, you can add some printf in opal_hwloc_unpack (from opal/mca/hwloc/base/

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-21 Thread r...@open-mpi.org
More than likely the problem is the difference in hwloc versions - sounds like the topology to/from xml is different between the two versions, and the older one doesn’t understand the new one. > On Jul 21, 2018, at 12:04 PM, Brian Smith > wrote: > > Greetings, > > I'm having trouble getting

[OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-21 Thread Brian Smith
Greetings, I'm having trouble getting openmpi 2.1.2 to work when launching a process from debian 8 on a remote debian 9 host. To keep things simple in this example, I'm just launching date on the remote host. deb8host$ mpirun -H deb9host date [deb8host:01552] [[32763,0],0] ORTE_ERROR_LOG: Error i

[OMPI users] Error when using Open MPI shared library in a program compiled by g++

2018-07-01 Thread lille stor
Hi,   I have successfully built Open MPI version 2.1.3 from scratch in Ubuntu 14.04 64 bit and using GCC 4.9. The result was the following shared libraries (needed for a program to use Open MPI):         dummy@machine:~/$ ldd /home/dummy/openmpi/build/lib/libmpi.so /home/dummy/openmpi/build/lib

Re: [OMPI users] error building openmpi-master-201806060243-64a5baa on Linux with Sun C

2018-06-07 Thread Siegmar Gross
Hi, it seems that the problem results from a compiler bug, because today I was able to build the package with the compilers from "Intel Parallel Studio XE 2018" and "Portland Group Community Edition 2018". Unfortunately, I have no contact person at Oracle to report the bug, so that it can be fixe

Re: [OMPI users] error building openmpi-master-201806060243-64a5baa on Linux with Sun C

2018-06-06 Thread Nathan Hjelm
The bindings in v3.1.0 are incorrect. They are missing the asynchronous attribute. That will be fixed in v3.1.1. > On Jun 6, 2018, at 12:06 PM, Siegmar Gross > wrote: > > Hi Jeff, > >> I asked some Fortran gurus, and they don't think that there >> is any restriction on having ASYNCHRONOUS

Re: [OMPI users] error building openmpi-master-201806060243-64a5baa on Linux with Sun C

2018-06-06 Thread Siegmar Gross
Hi Jeff, I asked some Fortran gurus, and they don't think that there is any restriction on having ASYNCHRONOUS and INTENT on the same line. Indeed, Open MPI's definition of MPI_ACCUMULATE seems to agree with what is in MPI-3.1. Is this a new version of a Fortran compiler that you're using, per

Re: [OMPI users] error building openmpi-master-201806060243-64a5baa on Linux with Sun C

2018-06-06 Thread Jeff Squyres (jsquyres) via users
Siegmar -- I asked some Fortran gurus, and they don't think that there is any restriction on having ASYNCHRONOUS and INTENT on the same line. Indeed, Open MPI's definition of MPI_ACCUMULATE seems to agree with what is in MPI-3.1. Is this a new version of a Fortran compiler that you're using, p

Re: [OMPI users] error building openmpi-master-201806060243-64a5baa on Linux with Sun C

2018-06-06 Thread Nathan Hjelm
I put in a PR to "fix" this but if you look at the standard it has both intent(in) and asynchronous. Might be a compiler problem? -Nathan > On Jun 6, 2018, at 5:11 AM, Siegmar Gross > wrote: > > Hi, > > I've tried to install openmpi-master-201806060243-64a5baa on my "SUSE Linux > Enterprise

[OMPI users] error building openmpi-master-201806060243-64a5baa on Linux with Sun C

2018-06-06 Thread Siegmar Gross
Hi, I've tried to install openmpi-master-201806060243-64a5baa on my "SUSE Linux Enterprise Server 12.3 (x86_64)" with Sun C 5.15 (Oracle Developer Studio 12.6). Unfortunately I still get the following error that I already reported on April 12th and May 5th. loki openmpi-master-201806060243-64a5b

[OMPI users] error building openmpi-master-201805030307-c22c485 on Linux with Sun C

2018-05-04 Thread Siegmar Gross
Hi, I've tried to install openmpi-master-201805030307-c22c485 on my "SUSE Linux Enterprise Server 12.3 (x86_64)" with Sun C 5.15 (Oracle Developer Studio 12.6). Unfortunately I still get the following error that I already reported on April 12th. loki openmpi-master-201805030307-c22c485-Linux.x8

Re: [OMPI users] Error in hello_cxx.cc

2018-04-23 Thread r...@open-mpi.org
and 3) we no longer support Windows. You could try using the cygwin port instead. > On Apr 23, 2018, at 7:52 PM, Nathan Hjelm wrote: > > Two things. 1) 1.4 is extremely old and you will not likely get much help > with it, and 2) the c++ bindings were deprecated in MPI-2.2 (2009) and > removed

  1   2   3   4   5   6   7   8   >