Did you configure OMPI —with-tm=<path-to-PBS-libs>? It looks like we didn’t 
build PBS support and so we only see one node with a single slot allocated to 
it.


> On Oct 3, 2018, at 12:02 PM, Castellana Michele <michele.castell...@curie.fr> 
> wrote:
> 
> Dear all,
> I am having trouble running an MPI code across multiple cores on a new 
> computer cluster, which uses PBS. Here is a minimal example, where I want to 
> run two MPI processes, each on  a different node. The PBS script is 
> 
> #!/bin/bash
> #PBS -l walltime=00:01:00
> #PBS -l mem=1gb
> #PBS -l nodes=2:ppn=1
> #PBS -q batch
> #PBS -N test
> mpirun -np 2 ./code.o
> 
> and when I submit it with 
> 
> $qsub script.sh
> 
> I get the following message in the PBS error file
> 
> $ cat test.e1234 
> [shbli040:08879] mca_base_component_repository_open: unable to open 
> mca_plm_tm: libcrypto.so.0.9.8: cannot open shared object file: No such file 
> or directory (ignored)
> [shbli040:08879] mca_base_component_repository_open: unable to open 
> mca_oob_ud: libibverbs.so.1: cannot open shared object file: No such file or 
> directory (ignored)
> [shbli040:08879] mca_base_component_repository_open: unable to open 
> mca_ras_tm: libcrypto.so.0.9.8: cannot open shared object file: No such file 
> or directory (ignored)
> --------------------------------------------------------------------------
> There are not enough slots available in the system to satisfy the 2 slots
> that were requested by the application:
>   ./code.o
> 
> Either request fewer slots for your application, or make more slots available
> for use.
> —————————————————————————————————————
> 
> The PBS version is
> 
> $ qstat --version
> Version: 6.1.2
> 
> and here is some additional information on the MPI version
> 
> $ mpicc -v
> Using built-in specs.
> COLLECT_GCC=/bin/gcc
> COLLECT_LTO_WRAPPER=/usr/libexec/gcc/x86_64-redhat-linux/4.8.5/lto-wrapper
> Target: x86_64-redhat-linux
> […]
> Thread model: posix
> gcc version 4.8.5 20150623 (Red Hat 4.8.5-28) (GCC) 
> 
> Do you guys know what may be the issue here? 
> 
> Thank you
> Best,
> 
> 
> 
> 
> 
> 
> 
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://lists.open-mpi.org/mailman/listinfo/users

_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to