Hi Ralph, RHEL 7.0, building in the repo location.
Yes, running autogen.pl to generate configure. I suspect this is unrelated, but I saw this while make install: WARNING! Common symbols found: btl_openib_lex.o: 0000000000000008 C btl_openib_ini_yyleng btl_openib_lex.o: 0000000000000008 C btl_openib_ini_yytext keyval_lex.o: 0000000000000008 C opal_util_keyval_yyleng keyval_lex.o: 0000000000000008 C opal_util_keyval_yytext show_help_lex.o: 0000000000000008 C opal_show_help_yyleng show_help_lex.o: 0000000000000008 C opal_show_help_yytext rmaps_rank_file_lex.o: 0000000000000008 C orte_rmaps_rank_file_leng rmaps_rank_file_lex.o: 0000000000000008 C orte_rmaps_rank_file_text hostfile_lex.o: 0000000000000008 C orte_util_hostfile_leng hostfile_lex.o: 0000000000000008 C orte_util_hostfile_text make[3]: [install-exec-hook] Error 1 (ignored) Thanks, _MAC From: devel [mailto:devel-boun...@open-mpi.org] On Behalf Of Ralph Castain Sent: Tuesday, September 01, 2015 10:43 AM To: Open MPI Developers Subject: Re: [OMPI devel] Problem running from ompi master What system is this on? CentOS7? Are you doing a VPATH build, or doing the build in the repo location? Also, I assume you remembered to run autogen.pl before configure, yes? On Sep 1, 2015, at 10:11 AM, Cabral, Matias A <matias.a.cab...@intel.com<mailto:matias.a.cab...@intel.com>> wrote: Hi Gilles, I deleted everything, re-cloned and re-built (without my patch), but still see the same issue. The only option I’m using with configure is --prefix. I even tried building with --enable-mpirun-prefix-by-default, and also passing the prefix at runtime (mpirun –prefix =/…), but I always end with the same issue. Is it possible that the issue is related to configure --prefix ? Thanks, _MAC From: devel [mailto:devel-boun...@open-mpi.org] On Behalf Of Gilles Gouaillardet Sent: Monday, August 31, 2015 5:46 PM To: Open MPI Developers Subject: Re: [OMPI devel] Problem running from ompi master Hi, this part has been revamped recently. at first, i would recommend you make a fresh install remove the install directory, and the build directory if you use VPATH, re-run configure && make && make install that should hopefully fix the issue Cheers, Gilles On 9/1/2015 9:35 AM, Cabral, Matias A wrote: Hi, Before submitting a pull req I decided to test some changes on ompi master branch but I’m facing an unrelated runtime error with ess pmi not being found. I confirmed PATH and LD_LIBRARY_PATH are set correctly and also that mca_ess_pmi.so where it should. Any suggestions? Thanks, Regards, s-7 ~/devel/ompi> ls ./lib/openmpi/ |grep pmi mca_ess_pmi.la mca_ess_pmi.so mca_pmix_pmix1xx.la mca_pmix_pmix1xx.so s-7 ~/devel/ompi> cat ~/.bashrc |grep -e PATH -e LD_LIBRARY_PATH export PATH=$HOME/devel/ompi/bin/:$PATH export LD_LIBRARY_PATH=$HOME/devel/ompi/lib s-7 ~ ./bin/mpirun -host s-7,s-8 -np 2 ./osu_latency *** An error occurred in MPI_Init *** on a NULL communicator *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort, *** and potentially your MPI job) [s-7.abc.com<http://s-7.abc.com/>:56614] Local abort before MPI_INIT completed successfully; not able to aggregate error messages, and not able to guarantee that all other processes were killed! [s-7.abc.com<http://s-7.abc.com/>:56614] [[INVALID],INVALID] ORTE_ERROR_LOG: Not found in file runtime/orte_init.c at line 129 -------------------------------------------------------------------------- A requested component was not found, or was unable to be opened. This means that this component is either not installed or is unable to be used on your system (e.g., sometimes this means that shared libraries that the component requires are unable to be found/loaded). Note that Open MPI stopped checking at the first component that it did not find. Host: s-7.abc.com<http://s-7.abc.com/> Framework: ess Component: pmi -------------------------------------------------------------------------- -------------------------------------------------------------------------- It looks like orte_init failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during orte_init; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): orte_ess_base_open failed --> Returned value Not found (-13) instead of ORTE_SUCCESS -------------------------------------------------------------------------- -------------------------------------------------------------------------- It looks like MPI_INIT failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during MPI_INIT; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): ompi_mpi_init: ompi_rte_init failed --> Returned "Not found" (-13) instead of "Success" (0) -------------------------------------------------------------------------- ------------------------------------------------------- Primary job terminated normally, but 1 process returned a non-zero exit code.. Per user-direction, the job has been aborted. ------------------------------------------------------- -------------------------------------------------------------------------- mpirun detected that one or more processes exited with non-zero status, thus causing the job to be terminated. The first process to do so was: Process name: [[2886,1],0] Exit code: 1 -------------------------------------------------------------------------- _______________________________________________ devel mailing list de...@open-mpi.org<mailto:de...@open-mpi.org> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel Link to this post: http://www.open-mpi.org/community/lists/devel/2015/08/17908.php _______________________________________________ devel mailing list de...@open-mpi.org<mailto:de...@open-mpi.org> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel Link to this post: http://www.open-mpi.org/community/lists/devel/2015/09/17912.php