OS: Centos 7 Infiniband Packages from OS repos Mellanox HCA
Compiled openmpi 1.10.7 on centos7 with the following config ./configure --prefix=/usr/local/software/OpenMPI/openmpi-1.10.7 --with-tm=/opt/pbs --with-verbs Snippet from config.log seems to indicate that the infiniband header files were located btl_openib_CPPFLAGS=' -I/usr/include/infiniband' common_verbs_CPPFLAGS=' -I/usr/include/infiniband' oshmem_verbs_CPPFLAGS=' -I/usr/include/infiniband' Everthing seems to have compiled correctly, but when I try to run any program using mpirun I am receiving the following error: mpirun -np 8 ./a.out -------------------------------------------------------------------------- [[18431,1],2]: A high-performance Open MPI point-to-point messaging module was unable to find any relevant network interfaces: Module: OpenFabrics (openib) Host: host-name Another transport will be used instead, although this may result in lower performance. -------------------------------------------------------------------------- [:13959] 7 more processes have sent help message help-mpi-btl-base.txt / btl:no-nics [:13959] Set MCA parameter "orte_base_help_aggre gate" to 0 to see all help / error messages I am unsure as to where to go from here. Any help would be appreciated as to how to troubleshoot this issue. Thank you, Sajesh
_______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users