In this case, some Open MPI plugins are missing some third party libraries,
so you would have to ldd all the plugins (e.g. the .so files) located
in <prefix>/lib/openmpi
in order to evidence any issue.

Cheers,

Gilles

On Thu, Oct 4, 2018 at 4:34 PM John Hearns via users
<users@lists.open-mpi.org> wrote:
>
> Michele  one tip:   log into a compute node using ssh and as your own 
> username.
> If you use the Modules envirnonment then load the modules you use in
> the job script
> then use the  ldd  utility to check if you can load all the libraries
> in the code.io executable
>
> Actually you are better to submit a short batch job which does not use
> mpirun but uses ldd
> A proper batch job will duplicate the environment you wish to run in.
>
>     ldd ./code.io
>
> By the way, is the batch system PBSPro or OpenPBS?  Version 6 seems a bit old.
> Can you say what version of Redhat or CentOS this cluster is installed with?
>
>
>
> On Thu, 4 Oct 2018 at 00:02, Castellana Michele
> <michele.castell...@curie.fr> wrote:
> >
> > I fixed it, the correct file was in /lib64, not in /lib.
> >
> > Thank you for your help.
> >
> > On Oct 3, 2018, at 11:30 PM, Castellana Michele 
> > <michele.castell...@curie.fr> wrote:
> >
> > Thank you, I found some libcrypto files in /usr/lib indeed:
> >
> > $ ls libcry*
> > libcrypt-2.17.so  libcrypto.so.10  libcrypto.so.1.0.2k  libcrypt.so.1
> >
> > but I could not find libcrypto.so.0.9.8. Here they suggest to create a 
> > hyperlink, but if I do I still get an error from MPI. Is there another way 
> > around this?
> >
> > Best,
> >
> > On Oct 3, 2018, at 11:00 PM, Jeff Squyres (jsquyres) via users 
> > <users@lists.open-mpi.org> wrote:
> >
> > It's probably in your Linux distro somewhere -- I'd guess you're missing a 
> > package (e.g., an RPM or a deb) out on your compute nodes...?
> >
> >
> > On Oct 3, 2018, at 4:24 PM, Castellana Michele 
> > <michele.castell...@curie.fr> wrote:
> >
> > Dear Ralph,
> > Thank you for your reply. Do you know where I could find libcrypto.so.0.9.8 
> > ?
> >
> > Best,
> >
> > On Oct 3, 2018, at 9:41 PM, Ralph H Castain <r...@open-mpi.org> wrote:
> >
> > Actually, I see that you do have the tm components built, but they cannot 
> > be loaded because you are missing libcrypto from your LD_LIBRARY_PATH
> >
> >
> > On Oct 3, 2018, at 12:33 PM, Ralph H Castain <r...@open-mpi.org> wrote:
> >
> > Did you configure OMPI —with-tm=<path-to-PBS-libs>? It looks like we didn’t 
> > build PBS support and so we only see one node with a single slot allocated 
> > to it.
> >
> >
> > On Oct 3, 2018, at 12:02 PM, Castellana Michele 
> > <michele.castell...@curie.fr> wrote:
> >
> > Dear all,
> > I am having trouble running an MPI code across multiple cores on a new 
> > computer cluster, which uses PBS. Here is a minimal example, where I want 
> > to run two MPI processes, each on  a different node. The PBS script is
> >
> > #!/bin/bash
> > #PBS -l walltime=00:01:00
> > #PBS -l mem=1gb
> > #PBS -l nodes=2:ppn=1
> > #PBS -q batch
> > #PBS -N test
> > mpirun -np 2 ./code.o
> >
> > and when I submit it with
> >
> > $qsub script.sh
> >
> > I get the following message in the PBS error file
> >
> > $ cat test.e1234
> > [shbli040:08879] mca_base_component_repository_open: unable to open 
> > mca_plm_tm: libcrypto.so.0.9.8: cannot open shared object file: No such 
> > file or directory (ignored)
> > [shbli040:08879] mca_base_component_repository_open: unable to open 
> > mca_oob_ud: libibverbs.so.1: cannot open shared object file: No such file 
> > or directory (ignored)
> > [shbli040:08879] mca_base_component_repository_open: unable to open 
> > mca_ras_tm: libcrypto.so.0.9.8: cannot open shared object file: No such 
> > file or directory (ignored)
> > --------------------------------------------------------------------------
> > There are not enough slots available in the system to satisfy the 2 slots
> > that were requested by the application:
> >  ./code.o
> >
> > Either request fewer slots for your application, or make more slots 
> > available
> > for use.
> > —————————————————————————————————————
> >
> > The PBS version is
> >
> > $ qstat --version
> > Version: 6.1.2
> >
> > and here is some additional information on the MPI version
> >
> > $ mpicc -v
> > Using built-in specs.
> > COLLECT_GCC=/bin/gcc
> > COLLECT_LTO_WRAPPER=/usr/libexec/gcc/x86_64-redhat-linux/4.8.5/lto-wrapper
> > Target: x86_64-redhat-linux
> > […]
> > Thread model: posix
> > gcc version 4.8.5 20150623 (Red Hat 4.8.5-28) (GCC)
> >
> > Do you guys know what may be the issue here?
> >
> > Thank you
> > Best,
> >
> >
> >
> >
> >
> >
> >
> > _______________________________________________
> > users mailing list
> > users@lists.open-mpi.org
> > https://lists.open-mpi.org/mailman/listinfo/users
> >
> >
> > _______________________________________________
> > users mailing list
> > users@lists.open-mpi.org
> > https://lists.open-mpi.org/mailman/listinfo/users
> >
> >
> > _______________________________________________
> > users mailing list
> > users@lists.open-mpi.org
> > https://lists.open-mpi.org/mailman/listinfo/users
> >
> >
> > _______________________________________________
> > users mailing list
> > users@lists.open-mpi.org
> > https://lists.open-mpi.org/mailman/listinfo/users
> >
> >
> >
> > --
> > Jeff Squyres
> > jsquy...@cisco.com
> >
> > _______________________________________________
> > users mailing list
> > users@lists.open-mpi.org
> > https://lists.open-mpi.org/mailman/listinfo/users
> >
> >
> > _______________________________________________
> > users mailing list
> > users@lists.open-mpi.org
> > https://lists.open-mpi.org/mailman/listinfo/users
> >
> >
> > _______________________________________________
> > users mailing list
> > users@lists.open-mpi.org
> > https://lists.open-mpi.org/mailman/listinfo/users
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://lists.open-mpi.org/mailman/listinfo/users
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to