Actually, I see that you do have the tm components built, but they cannot be loaded because you are missing libcrypto from your LD_LIBRARY_PATH
> On Oct 3, 2018, at 12:33 PM, Ralph H Castain <r...@open-mpi.org> wrote: > > Did you configure OMPI —with-tm=<path-to-PBS-libs>? It looks like we didn’t > build PBS support and so we only see one node with a single slot allocated to > it. > > >> On Oct 3, 2018, at 12:02 PM, Castellana Michele <michele.castell...@curie.fr >> <mailto:michele.castell...@curie.fr>> wrote: >> >> Dear all, >> I am having trouble running an MPI code across multiple cores on a new >> computer cluster, which uses PBS. Here is a minimal example, where I want to >> run two MPI processes, each on a different node. The PBS script is >> >> #!/bin/bash >> #PBS -l walltime=00:01:00 >> #PBS -l mem=1gb >> #PBS -l nodes=2:ppn=1 >> #PBS -q batch >> #PBS -N test >> mpirun -np 2 ./code.o >> >> and when I submit it with >> >> $qsub script.sh >> >> I get the following message in the PBS error file >> >> $ cat test.e1234 >> [shbli040:08879] mca_base_component_repository_open: unable to open >> mca_plm_tm: libcrypto.so.0.9.8: cannot open shared object file: No such file >> or directory (ignored) >> [shbli040:08879] mca_base_component_repository_open: unable to open >> mca_oob_ud: libibverbs.so.1: cannot open shared object file: No such file or >> directory (ignored) >> [shbli040:08879] mca_base_component_repository_open: unable to open >> mca_ras_tm: libcrypto.so.0.9.8: cannot open shared object file: No such file >> or directory (ignored) >> -------------------------------------------------------------------------- >> There are not enough slots available in the system to satisfy the 2 slots >> that were requested by the application: >> ./code.o >> >> Either request fewer slots for your application, or make more slots available >> for use. >> ————————————————————————————————————— >> >> The PBS version is >> >> $ qstat --version >> Version: 6.1.2 >> >> and here is some additional information on the MPI version >> >> $ mpicc -v >> Using built-in specs. >> COLLECT_GCC=/bin/gcc >> COLLECT_LTO_WRAPPER=/usr/libexec/gcc/x86_64-redhat-linux/4.8.5/lto-wrapper >> Target: x86_64-redhat-linux >> […] >> Thread model: posix >> gcc version 4.8.5 20150623 (Red Hat 4.8.5-28) (GCC) >> >> Do you guys know what may be the issue here? >> >> Thank you >> Best, >> >> >> >> >> >> >> >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >> https://lists.open-mpi.org/mailman/listinfo/users > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://lists.open-mpi.org/mailman/listinfo/users
_______________________________________________ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users