Hi,
this part has been revamped recently.
at first, i would recommend you make a fresh install
remove the install directory, and the build directory if you use VPATH,
re-run configure && make && make install
that should hopefully fix the issue
Cheers,
Gilles
On 9/1/2015 9:35 AM, Cabral, Matias A wrote:
Hi,
Before submitting a pull req I decided to test some changes on ompi
master branch but I’m facing an unrelated runtime error with ess pmi
not being found. I confirmed PATH and LD_LIBRARY_PATH are set
correctly and also that mca_ess_pmi.so where it should. Any suggestions?
Thanks,
Regards,
s-7 ~/devel/ompi> ls ./lib/openmpi/ |grep pmi
mca_ess_pmi.la
mca_ess_pmi.so
mca_pmix_pmix1xx.la
mca_pmix_pmix1xx.so
s-7 ~/devel/ompi> cat ~/.bashrc |grep -e PATH -e LD_LIBRARY_PATH
export PATH=$HOME/devel/ompi/bin/:$PATH
export LD_LIBRARY_PATH=$HOME/devel/ompi/lib
s-7 ~ ./bin/mpirun -host s-7,s-8 -np 2 ./osu_latency
*** An error occurred in MPI_Init
*** on a NULL communicator
*** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
*** and potentially your MPI job)
[s-7.abc.com:56614] Local abort before MPI_INIT completed
successfully; not able to aggregate error messages, and not able to
guarantee that all other processes were killed!
[s-7.abc.com:56614] [[INVALID],INVALID] ORTE_ERROR_LOG: Not found in
file runtime/orte_init.c at line 129
--------------------------------------------------------------------------
A requested component was not found, or was unable to be opened. This
means that this component is either not installed or is unable to be
used on your system (e.g., sometimes this means that shared libraries
that the component requires are unable to be found/loaded). Note that
Open MPI stopped checking at the first component that it did not find.
Host: s-7.abc.com
Framework: ess
Component: pmi
--------------------------------------------------------------------------
--------------------------------------------------------------------------
It looks like orte_init failed for some reason; your parallel process is
likely to abort. There are many reasons that a parallel process can
fail during orte_init; some of which are due to configuration or
environment problems. This failure appears to be an internal failure;
here's some additional information (which may only be relevant to an
Open MPI developer):
orte_ess_base_open failed
--> Returned value Not found (-13) instead of ORTE_SUCCESS
--------------------------------------------------------------------------
--------------------------------------------------------------------------
It looks like MPI_INIT failed for some reason; your parallel process is
likely to abort. There are many reasons that a parallel process can
fail during MPI_INIT; some of which are due to configuration or
environment
problems. This failure appears to be an internal failure; here's some
additional information (which may only be relevant to an Open MPI
developer):
ompi_mpi_init: ompi_rte_init failed
--> Returned "Not found" (-13) instead of "Success" (0)
--------------------------------------------------------------------------
-------------------------------------------------------
Primary job terminated normally, but 1 process returned
a non-zero exit code.. Per user-direction, the job has been aborted.
-------------------------------------------------------
--------------------------------------------------------------------------
mpirun detected that one or more processes exited with non-zero
status, thus causing
the job to be terminated. The first process to do so was:
Process name: [[2886,1],0]
Exit code: 1
--------------------------------------------------------------------------
_______________________________________________
devel mailing list
de...@open-mpi.org
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
Link to this post:
http://www.open-mpi.org/community/lists/devel/2015/08/17908.php