Sorry. I just realized that you must mean the log file created by the
configure script. I was looking for a OpenMPI runtime log file. Is there
such a thing?

Dennis

> -----Original Message-----
> From: users-boun...@open-mpi.org 
> [mailto:users-boun...@open-mpi.org] On Behalf Of Dennis McRitchie
> Sent: Friday, February 02, 2007 6:12 PM
> To: Open MPI Users
> Subject: Re: [OMPI users] Can't run simple job with openmpi 
> using the Intelcompiler
> 
> Also, I see mention in your FAQ about config.log. My openmpi does not
> appear to be generating it, at least not anywhere in the install tree.
> How can I enable the creation of the log file?
> 
> Thanks,
>        Dennis
> 
> -----Original Message-----
> From: Dennis McRitchie 
> Sent: Friday, February 02, 2007 6:08 PM
> To: 'Open MPI Users'
> Subject: Can't run simple job with openmpi using the Intel compiler
> 
> When I submit a simple job (described below) using PBS, I 
> always get one
> of the following two errors:
> 1) [adroit-28:03945] [0,0,1]-[0,0,0] mca_oob_tcp_peer_recv_blocking:
> recv() failed with errno=104
> 
> 2) [adroit-30:03770] [0,0,3]-[0,0,0] 
> mca_oob_tcp_peer_complete_connect:
> connection failed (errno=111) - retrying (pid=3770)
> 
> The program does a uname and prints out results to standard out. The
> only MPI calls it makes are MPI_Init, MPI_Comm_size, 
> MPI_Comm_rank, and
> MPI_Finalize. I have tried it with both openmpi v 1.1.2 and 
> 1.1.4, built
> with Intel C compiler 9.1.045, and get the same results. But 
> if I build
> the same versions of openmpi using gcc, the test program always works
> fine. The app itself is built with mpicc.
> 
> It runs successfully if run from the command line with "mpiexec -n X
> <test-program-name>", where X is 1 to 8, but if I wrap it in the
> following qsub command file:
> ---------------------------------------------------
> #PBS -l pmem=512mb,nodes=1:ppn=1,walltime=0:10:00
> #PBS -m abe
> # #PBS -o /home0/dmcr/my_mpi/curt/uname_test.gcc.stdout
> # #PBS -e /home0/dmcr/my_mpi/curt/uname_test.gcc.stderr
> 
> cd /home/dmcr/my_mpi/openmpi
> echo "About to call mpiexec"
> module list
> mpiexec -n 1 uname_test.intel
> echo "After call to mpiexec"
> ----------------------------------------------------
> 
> it fails on any number of processors from 1 to 8, and the application
> segfaults.
> 
> The complete standard error of an 8-processsor job follows (note that
> mpiexec ran on adroit-31, but usually there is no info about adroit-31
> in standard error):
> -------------------------
> Currently Loaded Modulefiles:
>   1) intel/9.1/32/C/9.1.045         4) intel/9.1/32/default
>   2) intel/9.1/32/Fortran/9.1.040   5) openmpi/intel/1.1.2/32
>   3) intel/9.1/32/Iidb/9.1.045
> Signal:11 info.si_errno:0(Success) si_code:1(SEGV_MAPERR)
> Failing at addr:0x5
> [0] func:/usr/local/openmpi/1.1.4/intel/i386/lib/libopal.so.0 
> [0xb72c5b]
> *** End of error message ***
> ^@[adroit-29:03934] [0,0,2]-[0,0,0] mca_oob_tcp_peer_recv_blocking:
> recv() failed with errno=104
> [adroit-28:03945] [0,0,1]-[0,0,0] 
> mca_oob_tcp_peer_recv_blocking: recv()
> failed with errno=104
> [adroit-30:03770] [0,0,3]-[0,0,0] mca_oob_tcp_peer_complete_connect:
> connection failed (errno=111) - retrying (pid=3770)
> --------------------------
> 
> The complete standard error of an 1-processsor job follows:
> --------------------------
> Currently Loaded Modulefiles:
>   1) intel/9.1/32/C/9.1.045         4) intel/9.1/32/default
>   2) intel/9.1/32/Fortran/9.1.040   5) openmpi/intel/1.1.2/32
>   3) intel/9.1/32/Iidb/9.1.045
> Signal:11 info.si_errno:0(Success) si_code:1(SEGV_MAPERR)
> Failing at addr:0x2
> [0] func:/usr/local/openmpi/1.1.2/intel/i386/lib/libopal.so.0 
> [0x27d847]
> *** End of error message ***
> ^@[adroit-31:08840] [0,0,1]-[0,0,0] mca_oob_tcp_peer_complete_connect:
> connection failed (errno=111) - retrying (pid=8840)
> ---------------------------
> 
> Any thoughts as to why this might be failing?
> 
> Thanks,
>        Dennis
> 
> Dennis McRitchie
> Computational Science and Engineering Support (CSES)
> Academic Services Department
> Office of Information Technology
> Princeton University
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 

Reply via email to