I tried out the TCP connection and here is what the error file came out as.
[wasatch-29:05042] [0,0,0] ORTE_ERROR_LOG: Timeout in file ../../../../orte/mca\ /pls/base/pls_base_orted_cmds.c at line 275 [wasatch-29:05042] [0,0,0] ORTE_ERROR_LOG: Timeout in file ../../../../../orte/\ mca/pls/tm/pls_tm_module.c at line 572 [wasatch-29:05042] [0,0,0] ORTE_ERROR_LOG: Timeout in file ../../../../../orte/\ mca/errmgr/hnp/errmgr_hnp.c at line 90 [wasatch-29:05042] [0,0,0] ORTE_ERROR_LOG: Timeout in file ../../../../orte/mca\ /pls/base/pls_base_orted_cmds.c at line 188 [wasatch-29:05042] [0,0,0] ORTE_ERROR_LOG: Timeout in file ../../../../../orte/\ mca/pls/tm/pls_tm_module.c at line 603 -------------------------------------------------------------------------- mpirun was unable to cleanly terminate the daemons for this job. Returned value\ Timeout instead of ORTE_SUCCESS. -------------------------------------------------------------------------- [wasatch-29:05044] OOB: Connection to HNP lost On Sat, Nov 20, 2010 at 2:03 PM, Addepalli, Srirangam V < srirangam.v.addepa...@ttu.edu> wrote: > mpirun --mca btl self,sm,tcp --mca btl_base_verbose 30 -np 8 > /home/A00945081/SWMF_v2.3/run/SWMF.exe > run.log > > to run using tcp interface in job submission script. > > Rangam > _______________________________________ > From: users-boun...@open-mpi.org [users-boun...@open-mpi.org] On Behalf Of > Tushar Andriyas [thugnomic...@gmail.com] > Sent: Saturday, November 20, 2010 1:36 PM > To: Open MPI Users > Subject: Re: [OMPI users] Unable to find the following executable > > Ya sure, here is the list > > > Open MPI: 1.2.7 > Open MPI SVN revision: r19401 > Open RTE: 1.2.7 > Open RTE SVN revision: r19401 > OPAL: 1.2.7 > OPAL SVN revision: r19401 > Prefix: /opt/libraries/openmpi/openmpi-1.2.7-pgi > Configured architecture: x86_64-unknown-linux-gnu > Configured by: A00017402 > Configured on: Thu Sep 18 15:00:05 MDT 2008 > Configure host: volvox.hpc.usu.edu<http://volvox.hpc.usu.edu> > Built by: A00017402 > Built on: Thu Sep 18 15:20:06 MDT 2008 > Built host: volvox.hpc.usu.edu<http://volvox.hpc.usu.edu> > C bindings: yes > C++ bindings: yes > Fortran77 bindings: yes (all) > Fortran90 bindings: yes > Fortran90 bindings size: large > C compiler: pgcc > C compiler absolute: /opt/apps/pgi/linux86-64/7.2/bin/pgcc > C++ compiler: pgCC > C++ compiler absolute: /opt/apps/pgi/linux86-64/7.2/bin/pgCC > Fortran77 compiler: pgf77 > Fortran77 compiler abs: /opt/apps/pgi/linux86-64/7.2/bin/pgf77 > Fortran90 compiler: pgf90 > Fortran90 compiler abs: /opt/apps/pgi/linux86-64/7.2/bin/pgf90 > C profiling: yes > C++ profiling: yes > Fortran77 profiling: yes > Fortran90 profiling: yes > C++ exceptions: no > Thread support: posix (mpi: no, progress: no) > Internal debug support: no > MPI parameter check: runtime > Memory profiling support: no > Memory debugging support: no > libltdl support: yes > Heterogeneous support: yes > mpirun default --prefix: no > MCA backtrace: execinfo (MCA v1.0, API v1.0, Component v1.2.7) > MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component v1.2.7) > MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.2.7) > MCA maffinity: first_use (MCA v1.0, API v1.0, Component v1.2.7) > MCA maffinity: libnuma (MCA v1.0, API v1.0, Component v1.2.7) > MCA timer: linux (MCA v1.0, API v1.0, Component v1.2.7) > MCA installdirs: env (MCA v1.0, API v1.0, Component v1.2.7) > MCA installdirs: config (MCA v1.0, API v1.0, Component v1.2.7) > MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0) > MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0) > MCA coll: basic (MCA v1.0, API v1.0, Component v1.2.7) > MCA coll: self (MCA v1.0, API v1.0, Component v1.2.7) > MCA coll: sm (MCA v1.0, API v1.0, Component v1.2.7) > MCA coll: tuned (MCA v1.0, API v1.0, Component v1.2.7) > MCA io: romio (MCA v1.0, API v1.0, Component v1.2.7) > MCA mpool: rdma (MCA v1.0, API v1.0, Component v1.2.7) > MCA mpool: sm (MCA v1.0, API v1.0, Component v1.2.7) > MCA pml: cm (MCA v1.0, API v1.0, Component v1.2.7) > MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.2.7) > MCA bml: r2 (MCA v1.0, API v1.0, Component v1.2.7) > MCA rcache: vma (MCA v1.0, API v1.0, Component v1.2.7) > MCA btl: gm (MCA v1.0, API v1.0.1, Component v1.2.7) > MCA btl: self (MCA v1.0, API v1.0.1, Component v1.2.7) > MCA btl: sm (MCA v1.0, API v1.0.1, Component v1.2.7) > MCA btl: tcp (MCA v1.0, API v1.0.1, Component v1.0) > MCA topo: unity (MCA v1.0, API v1.0, Component v1.2.7) > MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.2.7) > MCA errmgr: hnp (MCA v1.0, API v1.3, Component v1.2.7) > MCA errmgr: orted (MCA v1.0, API v1.3, Component v1.2.7) > MCA errmgr: proxy (MCA v1.0, API v1.3, Component v1.2.7) > MCA gpr: null (MCA v1.0, API v1.0, Component v1.2.7) > MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.2.7) > MCA gpr: replica (MCA v1.0, API v1.0, Component v1.2.7) > MCA iof: proxy (MCA v1.0, API v1.0, Component v1.2.7) > MCA iof: svc (MCA v1.0, API v1.0, Component v1.2.7) > MCA ns: proxy (MCA v1.0, API v2.0, Component v1.2.7) > MCA ns: replica (MCA v1.0, API v2.0, Component v1.2.7) > MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0) > MCA ras: dash_host (MCA v1.0, API v1.3, Component v1.2.7) > MCA ras: gridengine (MCA v1.0, API v1.3, Component v1.2.7) > MCA ras: localhost (MCA v1.0, API v1.3, Component v1.2.7) > MCA ras: slurm (MCA v1.0, API v1.3, Component v1.2.7) > MCA ras: tm (MCA v1.0, API v1.3, Component v1.2.7) > MCA rds: hostfile (MCA v1.0, API v1.3, Component v1.2.7) > MCA rds: proxy (MCA v1.0, API v1.3, Component v1.2.7) > MCA rds: resfile (MCA v1.0, API v1.3, Component v1.2.7) > MCA rmaps: round_robin (MCA v1.0, API v1.3, Component v1.2.7) > MCA rmgr: proxy (MCA v1.0, API v2.0, Component v1.2.7) > MCA rmgr: urm (MCA v1.0, API v2.0, Component v1.2.7) > MCA rml: oob (MCA v1.0, API v1.0, Component v1.2.7) > MCA pls: gridengine (MCA v1.0, API v1.3, Component v1.2.7) > MCA pls: proxy (MCA v1.0, API v1.3, Component v1.2.7) > MCA pls: rsh (MCA v1.0, API v1.3, Component v1.2.7) > MCA pls: slurm (MCA v1.0, API v1.3, Component v1.2.7) > MCA pls: tm (MCA v1.0, API v1.3, Component v1.2.7) > MCA sds: env (MCA v1.0, API v1.0, Component v1.2.7) > MCA sds: pipe (MCA v1.0, API v1.0, Component v1.2.7) > MCA sds: seed (MCA v1.0, API v1.0, Component v1.2.7) > MCA sds: singleton (MCA v1.0, API v1.0, Component v1.2.7) > MCA sds: slurm (MCA v1.0, API v1.0, Component v1.2.7) > > How do you invoke tcp? I no for sure that the launcher on the clusters is > torque. > > Tushar > > On Sat, Nov 20, 2010 at 11:28 AM, Addepalli, Srirangam V < > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu>> > wrote: > Hello Tushar, > Can you send me the output of ompi_info. > Have you tried using just tcp instead of IB to narrow down. > Rangam > > #!/bin/sh > #PBS -V > #PBS -q wasatch > #PBS -N SWMF > #PBS -l nodes=1:ppn=8 > # change to the run directory > #cd $SWMF_v2.3/run > cat `echo ${PBS_NODEFILE}` > list_of_nodes > > mpirun --mca btl self,sm,tcp --mca btl_base_verbose 30 -np 8 > /home/A00945081/SWMF_v2.3/run/SWMF.exe > run.log > > > ________________________________________ > From: users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org> [ > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>] On Behalf > Of Tushar Andriyas [thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>] > Sent: Saturday, November 20, 2010 12:11 PM > To: Open MPI Users > Subject: Re: [OMPI users] Unable to find the following executable > > Rangam, > > It does not want to run at all. Attached is the log file from the batch > file run u sent. > > On Sat, Nov 20, 2010 at 10:32 AM, Addepalli, Srirangam V < > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu > ><mailto:srirangam.v.addepa...@ttu.edu<mailto: > srirangam.v.addepa...@ttu.edu>>> wrote: > Hello Tushar, > MPIRUN is not able to spawn processes on the node allocated. This should > help > > #!/bin/sh > #PBS -V > #PBS -q wasatch > #PBS -N SWMF > #PBS -l nodes=2:ppn=8 > # change to the run directory > #cd $SWMF_v2.3/run > cat `echo ${PBS_NODEFILE}` > list_of_nodes > mpirun -np 8 /home/A00945081/SWMF_v2.3/run/SWMF.exe > run.log > > > Rangam > > > ________________________________________ > From: users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>> [ > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>] On Behalf > Of Tushar Andriyas [thugnomic...@gmail.com<mailto:thugnomic...@gmail.com > ><mailto:thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>>] > Sent: Saturday, November 20, 2010 10:48 AM > To: Open MPI Users > Subject: Re: [OMPI users] Unable to find the following executable > > Hi Rangam, > > I ran the batch file that you gave and have attached the error file. Also, > since the WASATCH cluster is kind of small, people usually run on UINTA. So, > if possible could you look at the uinta error files? > Tushar > > On Fri, Nov 19, 2010 at 12:31 PM, Addepalli, Srirangam V < > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu > ><mailto:srirangam.v.addepa...@ttu.edu<mailto: > srirangam.v.addepa...@ttu.edu>><mailto:srirangam.v.addepa...@ttu.edu > <mailto:srirangam.v.addepa...@ttu.edu><mailto: > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu>>>> > wrote: > Hello Tushar, > After looking at the log files you attached it appears that there are > multiple issues. > > [0,1,11]: Myrinet/GM on host wasatch-55 was unable to find any NICs. > Another transport will be used instead, although this may result in > lower performance. > > Usually they occur if there is a mismatch in mpirun version and mca blt > selection. I suggest the following order to check if the job actually works > on a single node > > #!/bin/sh > #PBS -V > #PBS -q wasatch > #PBS -N SWMF > #PBS -l nodes=2:ppn=8 > # change to the run directory > #cd $SWMF_v2.3/run > cat `echo ${PBS_NODEFILE}` > list_of_nodes > mpirun -np 8 -machinefile list_of_nodes > /home/A00945081/SWMF_v2.3/run/SWMF.exe > run.log > > > Rangam > > > ________________________________________ > From: users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>> [ > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>>] On Behalf > Of Tushar Andriyas [thugnomic...@gmail.com<mailto:thugnomic...@gmail.com > ><mailto:thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>>>] > Sent: Friday, November 19, 2010 1:11 PM > To: Open MPI Users > Subject: Re: [OMPI users] Unable to find the following executable > > Hey Rangam, > > I tried out the batch script and the error file comes out empty and the > output file has /home/A00945081/SWM_v2.3/run/SWMF.exe (WHEN RUN ON A SINGLE > MACHINE) and the same with multiple machines in the run. So, does that mean > that the exe is auto mounted ? What should I do next? > > Tushar > > On Fri, Nov 19, 2010 at 10:05 AM, Addepalli, Srirangam V < > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu > ><mailto:srirangam.v.addepa...@ttu.edu<mailto: > srirangam.v.addepa...@ttu.edu>><mailto:srirangam.v.addepa...@ttu.edu > <mailto:srirangam.v.addepa...@ttu.edu><mailto: > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu > >>><mailto:srirangam.v.addepa...@ttu.edu<mailto: > srirangam.v.addepa...@ttu.edu><mailto:srirangam.v.addepa...@ttu.edu > <mailto:srirangam.v.addepa...@ttu.edu>><mailto: > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu > ><mailto:srirangam.v.addepa...@ttu.edu<mailto: > srirangam.v.addepa...@ttu.edu>>>>> wrote: > Hello Tushar, > > Try the following script. > > #!/bin/sh > #PBS -V > #PBS -q wasatch > #PBS -N SWMF > #PBS -l nodes=1:ppn=8 > # change to the run directory > #cd $SWMF_v2.3/run > cat `echo ${PBS_NODEFILE}` > list_of_nodes > > > > > The objective is to check if your user directories are auto mounted on > compute nodes and are available during run time. > > If the job returns information about SWMF.exe then it can be safely assumed > that user directories are being auto mounted. > > Rangam > > > > ________________________________________ > From: users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > >>><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>>> > [users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: > users-bounce! > s...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>>>] On > Behalf Of Tushar Andriyas [thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>>>>] > Sent: Friday, November 19, 2010 8:35 AM > To: Open MPI Users > Subject: Re: [OMPI users] Unable to find the following executable > > It just gives back the info on folders in my home directory. Dont get me > wrong but i m kinda new in this. So, could u type out d full command which i > need to give? > > Tushar > > On Thu, Nov 18, 2010 at 8:35 AM, Ralph Castain <r...@open-mpi.org<mailto: > r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org > >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: > r...@open-mpi.org<mailto:r...@open-mpi.org>>><mailto:r...@open-mpi.org > <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org > >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: > r...@open-mpi.org<mailto:r...@open-mpi.org>>>><mailto:r...@open-mpi.org > <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org > >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: > r...@open-mpi.org<mailto:r...@open-mpi.org>>><mailto:r...@open-mpi.org > <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org > >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: > r...@open-mpi.org<mailto:r...@open-mpi.org>>>>>> wrote: > You can qsub a simple "ls" on that path - that will tell you if the path is > valid on all machines in that allocation. > > What typically happens is that home directories aren't remotely mounted, or > are mounted on a different location. > > > On Thu, Nov 18, 2010 at 8:31 AM, Tushar Andriyas <thugnomic...@gmail.com > <mailto:thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>>>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomics28@! > gmail.com<mailto:thugnomic...@gmail.com>>>>>> wrote: > no its not in the same directory as SWMF. I guess the path is the same > since all the machines in a cluster are configured d same way. How do I know > if this is not the case? > > > On Thu, Nov 18, 2010 at 8:25 AM, Ralph Castain <r...@open-mpi.org<mailto: > r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org > >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: > r...@open-mpi.org<mailto:r...@open-mpi.org>>><mailto:r...@open-mpi.org > <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org > >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: > r...@open-mpi.org<mailto:r...@open-mpi.org>>>><mailto:r...@open-mpi.org > <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org > >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: > r...@open-mpi.org<mailto:r...@open-mpi.org>>><mailto:r...@open-mpi.org > <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org > >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: > r...@open-mpi.org<mailto:r...@open-mpi.org>>>>>> wrote: > Is you "hello world" test program in the same directory as SWMF? Is it > possible that the path you are specifying is not available on all of the > remote machines? That's the most common problem we see. > > > On Thu, Nov 18, 2010 at 7:59 AM, Tushar Andriyas <thugnomic...@gmail.com > <mailto:thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>>>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: > thugnomic...@gmail.com><mailto:thugnomics28@! > gmail.com<mailto:thugnomic...@gmail.com>>>>>> wrote: > Hi there, > > Thanks for the expedite reply. The thing is that although the mpirun is > setup correctly (since a simple hello world works), when I run the main > SWMF.exe executable, the cluster machines somehow fail to find the > executable (SWMF.exe). > > So, I have attached the sample error file from one of the runs > (SWMF.e143438) and also the MAKEFILES so that you could better gauge the > problem. The makefiles have Linux as the OS and pgf90 as compiler with > mpif90 as the linker. I am using openmpi-1.2.7-pgi. Job is submitted using a > batch file (job.bats) and the scheduler is Torque (version I am not sure but > I can see three on the machines viz 2.0.0, 2.2.1, 2.5.2). > > I have also attached an error file from one of the clusters (WASATCH viz > SWMF.e143439) and UINTA (SWMF.e143440) with the whole path of the exe as > Srirangam mentioned as follows (in the batch file). > > mpirun --prefix /opt/libraries/openmpi/openmpi-1.2.7-pgi > /home/A00945081/SWMF_v2.3/run/SWMF.exe > runlog_`date +%y%m%d%H%M` > > I have tried both mpirun and mpiexec but nothing seems to work. > > Tushar > > > On Wed, Nov 17, 2010 at 8:12 PM, Addepalli, Srirangam V < > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu > ><mailto:srirangam.v.addepa...@ttu.edu<mailto: > srirangam.v.addepa...@ttu.edu>><mailto:srirangam.v.addepa...@ttu.edu > <mailto:srirangam.v.addepa...@ttu.edu><mailto: > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu > >>><mailto:srirangam.v.addepa...@ttu.edu<mailto: > srirangam.v.addepa...@ttu.edu><mailto:srirangam.v.addepa...@ttu.edu > <mailto:srirangam.v.addepa...@ttu.edu>><mailto: > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu > ><mailto:srirangam.v.addepa...@ttu.edu<mailto: > srirangam.v.addepa...@ttu.edu>>>><mailto:srirangam.v.addepa...@ttu.edu > <mailto:srirangam.v.addepa...@ttu.edu><mailto: > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu > >><mailto:srirangam.v.addepa...@ttu.edu<mailto: > srirangam.v.addepa...@ttu.edu><mailto:srirangam.v.addepa...@ttu.edu > <mailto:srirangam.v.addepa...@ttu.edu>>><mailto:srirangam.v.addepalli@! > ttu.edu<mailto:srirangam.v.addepa...@ttu.edu><mailto: > srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu > >><mailto:srirangam.v.addepa...@ttu.edu<mailto: > srirangam.v.addepa...@ttu.edu><mailto:srirangam.v.addepa...@ttu.edu > <mailto:srirangam.v.addepa...@ttu.edu>>>>>> wrote: > Hello Tushar, > Have you tried supplying the full path of the executable just to check ? > Rangam > ________________________________________ > From: users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > >>><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > >>>><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > >>><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org > >><mailto:users-! > boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>>>> [ > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>>><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>><mailto: > users-bounce<mailto:! > users-bounce>! > s...@open-mpi.org<mailto:s...@open-mpi.org><mailto:users-boun...@open-mpi.org > <mailto:users-boun...@open-mpi.org>><mailto:users-boun...@open-mpi.org > <mailto:users-boun...@open-mpi.org><mailto:users-boun...@open-mpi.org > <mailto:users-boun...@open-mpi.org>>>>>] On Behalf Of Tushar Andriyas [ > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>>><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>>>><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: > thugnomic...@gmail.com<mailt! > o:thugnomic...@gmail.com <o%3athugnomic...@gmail.com>>>><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: > thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>>>>>] > Sent: Wednesday, November 17, 2010 8:49 PM > To: us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org>>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> > Subject: [OMPI users] Unable to find the following executable > > Hi there, > > I am new to using mpi commands and was stuck in problem with running a > code. When I submit my job through a batch file, the job exits with the > message that the executable could not be found on the machines. I have tried > a lot of options such as PBS -V and so on on but the problem persists. If > someone is interested, I can send the full info on the cluster, the compiler > and openmpi settings and other stuff. BTW the launcher is torque (which you > might have guessed). The code does not have a forum so I am in a deep mire. > > Thanks, > Tushar > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org > >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: > us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org>>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: > us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org > <mailto:us...@open-mpi.org>> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >