Did you configure OMPI —with-tm=<path-to-PBS-libs>? It looks like we didn’t build PBS support and so we only see one node with a single slot allocated to it.
> On Oct 3, 2018, at 12:02 PM, Castellana Michele <michele.castell...@curie.fr> > wrote: > > Dear all, > I am having trouble running an MPI code across multiple cores on a new > computer cluster, which uses PBS. Here is a minimal example, where I want to > run two MPI processes, each on a different node. The PBS script is > > #!/bin/bash > #PBS -l walltime=00:01:00 > #PBS -l mem=1gb > #PBS -l nodes=2:ppn=1 > #PBS -q batch > #PBS -N test > mpirun -np 2 ./code.o > > and when I submit it with > > $qsub script.sh > > I get the following message in the PBS error file > > $ cat test.e1234 > [shbli040:08879] mca_base_component_repository_open: unable to open > mca_plm_tm: libcrypto.so.0.9.8: cannot open shared object file: No such file > or directory (ignored) > [shbli040:08879] mca_base_component_repository_open: unable to open > mca_oob_ud: libibverbs.so.1: cannot open shared object file: No such file or > directory (ignored) > [shbli040:08879] mca_base_component_repository_open: unable to open > mca_ras_tm: libcrypto.so.0.9.8: cannot open shared object file: No such file > or directory (ignored) > -------------------------------------------------------------------------- > There are not enough slots available in the system to satisfy the 2 slots > that were requested by the application: > ./code.o > > Either request fewer slots for your application, or make more slots available > for use. > ————————————————————————————————————— > > The PBS version is > > $ qstat --version > Version: 6.1.2 > > and here is some additional information on the MPI version > > $ mpicc -v > Using built-in specs. > COLLECT_GCC=/bin/gcc > COLLECT_LTO_WRAPPER=/usr/libexec/gcc/x86_64-redhat-linux/4.8.5/lto-wrapper > Target: x86_64-redhat-linux > […] > Thread model: posix > gcc version 4.8.5 20150623 (Red Hat 4.8.5-28) (GCC) > > Do you guys know what may be the issue here? > > Thank you > Best, > > > > > > > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://lists.open-mpi.org/mailman/listinfo/users
_______________________________________________ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users