On Fri, 2007-01-19 at 13:25 -0700, Greg Watson wrote: > Bluesteel is a 64bit bproc machine. I configured with: > > ./configure --with-devel-headers --disable-shared --enable-static > > When I attempt to run an MPI program: > > [bluesteel.lanl.gov:28663] [0,0,0] ORTE_ERROR_LOG: Not available in > file ras_bjs.c at line 247 > ------------------------------------------------------------------------ > -- > The bproc PLS component was not able to launch all the processes on > the remote > nodes and therefore cannot continue. > > On node 0 the process pid was -2 and errno was set to 11. >
Shared lib? Ollie > For reference, we tried to launch ./x > ------------------------------------------------------------------------ > -- > [bluesteel.lanl.gov:28663] [0,0,0] ORTE_ERROR_LOG: Error in file > pls_bproc.c at line 943 > [bluesteel.lanl.gov:28663] [0,0,0] ORTE_ERROR_LOG: Error in file > pls_bproc.c at line 1141 > [bluesteel.lanl.gov:28663] [0,0,0] ORTE_ERROR_LOG: Error in file > rmgr_urm.c at line 460 > [bluesteel.lanl.gov:28663] mpirun: spawn failed with errno=-1 > [n0:28664] OOB: Connection to HNP lost > > Output from ompi_info: > > Open MPI: 1.2b3 > Open MPI SVN revision: r13112 > Open RTE: 1.2b3 > Open RTE SVN revision: r13112 > OPAL: 1.2b3 > OPAL SVN revision: r13112 > Prefix: /users/gwatson/ompi_1.2b3 > Configured architecture: x86_64-unknown-linux-gnu > Configured by: gwatson > Configured on: Fri Jan 19 12:52:21 MST 2007 > Configure host: bluesteel.lanl.gov > Built by: gwatson > Built on: Fri Jan 19 13:07:21 MST 2007 > Built host: bluesteel.lanl.gov > C bindings: yes > C++ bindings: yes > Fortran77 bindings: yes (all) > Fortran90 bindings: yes > Fortran90 bindings size: small > C compiler: gcc > C compiler absolute: /usr/bin/gcc > C++ compiler: g++ > C++ compiler absolute: /usr/bin/g++ > Fortran77 compiler: gfortran > Fortran77 compiler abs: /usr/bin/gfortran > Fortran90 compiler: gfortran > Fortran90 compiler abs: /usr/bin/gfortran > C profiling: yes > C++ profiling: yes > Fortran77 profiling: yes > Fortran90 profiling: yes > C++ exceptions: no > Thread support: posix (mpi: no, progress: no) > Internal debug support: no > MPI parameter check: runtime > Memory profiling support: no > Memory debugging support: no > libltdl support: yes > Heterogeneous support: yes > mpirun default --prefix: no > MCA backtrace: execinfo (MCA v1.0, API v1.0, Component v1.2) > MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component > v1.2) > MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.2) > MCA maffinity: first_use (MCA v1.0, API v1.0, Component > v1.2) > MCA timer: linux (MCA v1.0, API v1.0, Component v1.2) > MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0) > MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0) > MCA coll: basic (MCA v1.0, API v1.0, Component v1.2) > MCA coll: self (MCA v1.0, API v1.0, Component v1.2) > MCA coll: sm (MCA v1.0, API v1.0, Component v1.2) > MCA coll: tuned (MCA v1.0, API v1.0, Component v1.2) > MCA io: romio (MCA v1.0, API v1.0, Component v1.2) > MCA mpool: sm (MCA v1.0, API v1.0, Component v1.2) > MCA pml: cm (MCA v1.0, API v1.0, Component v1.2) > MCA pml: dr (MCA v1.0, API v1.0, Component v1.2) > MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.2) > MCA bml: r2 (MCA v1.0, API v1.0, Component v1.2) > MCA rcache: rb (MCA v1.0, API v1.0, Component v1.2) > MCA rcache: vma (MCA v1.0, API v1.0, Component v1.2) > MCA btl: self (MCA v1.0, API v1.0.1, Component v1.2) > MCA btl: sm (MCA v1.0, API v1.0.1, Component v1.2) > MCA btl: tcp (MCA v1.0, API v1.0.1, Component v1.0) > MCA topo: unity (MCA v1.0, API v1.0, Component v1.2) > MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.2) > MCA osc: rdma (MCA v1.0, API v1.0, Component v1.2) > MCA errmgr: hnp (MCA v1.0, API v1.3, Component v1.2) > MCA errmgr: orted (MCA v1.0, API v1.3, Component v1.2) > MCA errmgr: proxy (MCA v1.0, API v1.3, Component v1.2) > MCA errmgr: bproc (MCA v1.0, API v1.3, Component v1.2) > MCA gpr: null (MCA v1.0, API v1.0, Component v1.2) > MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.2) > MCA gpr: replica (MCA v1.0, API v1.0, Component v1.2) > MCA iof: proxy (MCA v1.0, API v1.0, Component v1.2) > MCA iof: svc (MCA v1.0, API v1.0, Component v1.2) > MCA ns: proxy (MCA v1.0, API v2.0, Component v1.2) > MCA ns: replica (MCA v1.0, API v2.0, Component v1.2) > MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0) > MCA ras: dash_host (MCA v1.0, API v1.3, Component > v1.2) > MCA ras: localhost (MCA v1.0, API v1.3, Component > v1.2) > MCA ras: bjs (MCA v1.0, API v1.3, Component v1.2) > MCA ras: gridengine (MCA v1.0, API v1.3, Component > v1.2) > MCA ras: lsf_bproc (MCA v1.0, API v1.3, Component > v1.2) > MCA ras: slurm (MCA v1.0, API v1.3, Component v1.2) > MCA rds: hostfile (MCA v1.0, API v1.3, Component v1.2) > MCA rds: proxy (MCA v1.0, API v1.3, Component v1.2) > MCA rds: resfile (MCA v1.0, API v1.3, Component v1.2) > MCA rmaps: round_robin (MCA v1.0, API v1.3, Component > v1.2) > MCA rmgr: proxy (MCA v1.0, API v2.0, Component v1.2) > MCA rmgr: urm (MCA v1.0, API v2.0, Component v1.2) > MCA rml: oob (MCA v1.0, API v1.0, Component v1.2) > MCA pls: proxy (MCA v1.0, API v1.3, Component v1.2) > MCA pls: bproc (MCA v1.0, API v1.3, Component v1.2) > MCA pls: gridengine (MCA v1.0, API v1.3, Component > v1.2) > MCA pls: rsh (MCA v1.0, API v1.3, Component v1.2) > MCA pls: slurm (MCA v1.0, API v1.3, Component v1.2) > MCA sds: env (MCA v1.0, API v1.0, Component v1.2) > MCA sds: seed (MCA v1.0, API v1.0, Component v1.2) > MCA sds: singleton (MCA v1.0, API v1.0, Component > v1.2) > MCA sds: bproc (MCA v1.0, API v1.0, Component v1.2) > MCA sds: pipe (MCA v1.0, API v1.0, Component v1.2) > MCA sds: slurm (MCA v1.0, API v1.0, Component v1.2) > > Any suggestions? > > Greg > _______________________________________________ > devel mailing list > de...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/devel