Just checked the head of the 1.8 branch (soon to be released as 1.8.4), and
confirmed the same results. I know the thread-multiple option is still
broken there, but will test that once we get the final fix committed.


On Mon, Nov 17, 2014 at 7:29 PM, Ralph Castain <r...@open-mpi.org> wrote:

> FWIW: I don't have access to a Linux box right now, but I built the OMPI
> devel master on my Mac using Intel 2015 compilers and was able to build/run
> all of the Fortran examples in our "examples" directory.
>
> I suspect the problem here is your use of the --enable-mpi-thread-multiple 
> option.
> The 1.8 series had an issue with that option - we are in the process of
> fixing it (I'm waiting for an updated patch), and you might be hitting it.
>
> If you remove that configure option, do things then work?
> Ralph
>
>
> On Mon, Nov 17, 2014 at 5:56 PM, Gilles Gouaillardet <
> gilles.gouaillar...@iferc.org> wrote:
>
>>  Hi John,
>>
>> do you MPI_Init() or do you MPI_Init_thread(MPI_THREAD_MULTIPLE) ?
>>
>> does your program calls MPI anywhere from an OpenMP region ?
>> does your program calls MPI only within an !$OMP MASTER section ?
>> does your program does not invoke MPI at all from any OpenMP region ?
>>
>> can you reproduce this issue with a simple fortran program ? or can you
>> publish all your files ?
>>
>> Cheers,
>>
>> Gilles
>>
>>
>> On 2014/11/18 1:41, John Bray wrote:
>>
>> I have succesfully been using OpenMPI 1.8.3 compiled with Intel-14, using
>>
>> ./configure --prefix=/usr/local/mpi/$(basename $PWD) --with-threads=posix
>> --enable-mpi-thread-multiple --disable-vt --with-scif=no
>>
>> I have now switched to Intel 15.0.1, and configuring with the same options,
>> I get minor changes in config.log about warning spotting, but it makes all
>> the binaries, and I can compile my own fortran code with mpif90/mpicc
>>
>> but a command 'mpiexec --verbose -n 12 ./fortran_binary' does nothing
>>
>> I checked the FAQ and started using
>>
>> ./configure --prefix=/usr/local/mpi/$(basename $PWD) --with-threads=posix
>> --enable-mpi-thread-multiple --disable-vt --with-scif=no CC=icc CXX=icpc
>> F77=ifort FC=ifort
>>
>> but that makes no difference.
>>
>> Only with -d do I get any more information
>>
>> mpirun -d --verbose -n 12
>> /home/jbray/5.0/mic2/one/intel-15_openmpi-1.8.3/one_f_debug.exe
>> [mic2:21851] procdir: /tmp/openmpi-sessions-jbray@mic2_0/27642/0/0
>> [mic2:21851] jobdir: /tmp/openmpi-sessions-jbray@mic2_0/27642/0
>> [mic2:21851] top: openmpi-sessions-jbray@mic2_0
>> [mic2:21851] tmp: /tmp
>> [mic2:21851] sess_dir_cleanup: job session dir does not exist
>> [mic2:21851] procdir: /tmp/openmpi-sessions-jbray@mic2_0/27642/0/0
>> [mic2:21851] jobdir: /tmp/openmpi-sessions-jbray@mic2_0/27642/0
>> [mic2:21851] top: openmpi-sessions-jbray@mic2_0
>> [mic2:21851] tmp: /tmp
>> [mic2:21851] sess_dir_finalize: proc session dir does not exist
>> <12 times>
>>
>>
>> [mic2:21851] sess_dir_cleanup: job session dir does not exist
>> exiting with status 139
>>
>> My C codes do not have this problem
>>
>> Compiler options are
>>
>> mpicxx -g -O0 -fno-inline-functions -openmp -o one_c_debug.exe async.c
>> collective.c compute.c memory.c one.c openmp.c p2p.c variables.c
>> auditmpi.c   control.c inout.c perfio.c ring.c wave.c io.c   leak.c mpiio.c
>> pthreads.c -openmp -lpthread
>>
>> mpif90 -g -O0  -fno-inline-functions -openmp -o one_f_debug.exe control.o
>> io.f90 leak.f90 memory.f90 one.f90 ring.f90 slow.f90 swapbounds.f90
>> variables.f90 wave.f90 *.F90 -openmp
>>
>> Any suggestions as to what is upsetting Fortran with Intel-15
>>
>> John
>>
>>
>>
>>
>> _______________________________________________
>> users mailing listus...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/users/2014/11/25823.php
>>
>>
>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post:
>> http://www.open-mpi.org/community/lists/users/2014/11/25829.php
>>
>
>

Reply via email to