Those are just warnings that have no impact on things - just a reminder to developers about cleanup we promised to do over time.
I’ll try to reproduce this here > On Sep 1, 2015, at 11:11 AM, Cabral, Matias A <matias.a.cab...@intel.com> > wrote: > > Hi Ralph, <> > > RHEL 7.0, building in the repo location. > > Yes, running autogen.pl to generate configure. > > I suspect this is unrelated, but I saw this while make install: > > WARNING! Common symbols found: > btl_openib_lex.o: 0000000000000008 C btl_openib_ini_yyleng > btl_openib_lex.o: 0000000000000008 C btl_openib_ini_yytext > keyval_lex.o: 0000000000000008 C opal_util_keyval_yyleng > keyval_lex.o: 0000000000000008 C opal_util_keyval_yytext > show_help_lex.o: 0000000000000008 C opal_show_help_yyleng > show_help_lex.o: 0000000000000008 C opal_show_help_yytext > rmaps_rank_file_lex.o: 0000000000000008 C orte_rmaps_rank_file_leng > rmaps_rank_file_lex.o: 0000000000000008 C orte_rmaps_rank_file_text > hostfile_lex.o: 0000000000000008 C orte_util_hostfile_leng > hostfile_lex.o: 0000000000000008 C orte_util_hostfile_text > make[3]: [install-exec-hook] Error 1 (ignored) > > Thanks, > > _MAC > > From: devel [mailto:devel-boun...@open-mpi.org] On Behalf Of Ralph Castain > Sent: Tuesday, September 01, 2015 10:43 AM > To: Open MPI Developers > Subject: Re: [OMPI devel] Problem running from ompi master > > What system is this on? CentOS7? Are you doing a VPATH build, or doing the > build in the repo location? > > Also, I assume you remembered to run autogen.pl before configure, yes? > > > On Sep 1, 2015, at 10:11 AM, Cabral, Matias A <matias.a.cab...@intel.com > <mailto:matias.a.cab...@intel.com>> wrote: > > Hi Gilles, > > I deleted everything, re-cloned and re-built (without my patch), but still > see the same issue. The only option I’m using with configure is --prefix. I > even tried building with --enable-mpirun-prefix-by-default, and also passing > the prefix at runtime (mpirun –prefix =/…), but I always end with the same > issue. Is it possible that the issue is related to configure --prefix ? > > Thanks, > > _MAC > > From: devel [mailto:devel-boun...@open-mpi.org > <mailto:devel-boun...@open-mpi.org>] On Behalf Of Gilles Gouaillardet > Sent: Monday, August 31, 2015 5:46 PM > To: Open MPI Developers > Subject: Re: [OMPI devel] Problem running from ompi master > > Hi, > > this part has been revamped recently. > > at first, i would recommend you make a fresh install > remove the install directory, and the build directory if you use VPATH, > re-run configure && make && make install > that should hopefully fix the issue > > Cheers, > > Gilles > > On 9/1/2015 9:35 AM, Cabral, Matias A wrote: > Hi, > > Before submitting a pull req I decided to test some changes on ompi master > branch but I’m facing an unrelated runtime error with ess pmi not being > found. I confirmed PATH and LD_LIBRARY_PATH are set correctly and also that > mca_ess_pmi.so where it should. Any suggestions? > > Thanks, > Regards, > > s-7 ~/devel/ompi> ls ./lib/openmpi/ |grep pmi > mca_ess_pmi.la > mca_ess_pmi.so > mca_pmix_pmix1xx.la > mca_pmix_pmix1xx.so > > s-7 ~/devel/ompi> cat ~/.bashrc |grep -e PATH -e LD_LIBRARY_PATH > export PATH=$HOME/devel/ompi/bin/:$PATH > export LD_LIBRARY_PATH=$HOME/devel/ompi/lib > > > s-7 ~ ./bin/mpirun -host s-7,s-8 -np 2 ./osu_latency > *** An error occurred in MPI_Init > *** on a NULL communicator > *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort, > *** and potentially your MPI job) > [s-7.abc.com <http://s-7.abc.com/>:56614] Local abort before MPI_INIT > completed successfully; not able to aggregate error messages, and not able to > guarantee that all other processes were killed! > [s-7.abc.com <http://s-7.abc.com/>:56614] [[INVALID],INVALID] ORTE_ERROR_LOG: > Not found in file runtime/orte_init.c at line 129 > -------------------------------------------------------------------------- > A requested component was not found, or was unable to be opened. This > means that this component is either not installed or is unable to be > used on your system (e.g., sometimes this means that shared libraries > that the component requires are unable to be found/loaded). Note that > Open MPI stopped checking at the first component that it did not find. > > Host: s-7.abc.com <http://s-7.abc.com/> > Framework: ess > Component: pmi > -------------------------------------------------------------------------- > -------------------------------------------------------------------------- > It looks like orte_init failed for some reason; your parallel process is > likely to abort. There are many reasons that a parallel process can > fail during orte_init; some of which are due to configuration or > environment problems. This failure appears to be an internal failure; > here's some additional information (which may only be relevant to an > Open MPI developer): > > orte_ess_base_open failed > --> Returned value Not found (-13) instead of ORTE_SUCCESS > -------------------------------------------------------------------------- > -------------------------------------------------------------------------- > It looks like MPI_INIT failed for some reason; your parallel process is > likely to abort. There are many reasons that a parallel process can > fail during MPI_INIT; some of which are due to configuration or environment > problems. This failure appears to be an internal failure; here's some > additional information (which may only be relevant to an Open MPI > developer): > > ompi_mpi_init: ompi_rte_init failed > --> Returned "Not found" (-13) instead of "Success" (0) > -------------------------------------------------------------------------- > ------------------------------------------------------- > Primary job terminated normally, but 1 process returned > a non-zero exit code.. Per user-direction, the job has been aborted. > ------------------------------------------------------- > -------------------------------------------------------------------------- > mpirun detected that one or more processes exited with non-zero status, thus > causing > the job to be terminated. The first process to do so was: > > Process name: [[2886,1],0] > Exit code: 1 > -------------------------------------------------------------------------- > > > > > _______________________________________________ > devel mailing list > de...@open-mpi.org <mailto:de...@open-mpi.org> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel > <http://www.open-mpi.org/mailman/listinfo.cgi/devel> > Link to this post: > http://www.open-mpi.org/community/lists/devel/2015/08/17908.php > <http://www.open-mpi.org/community/lists/devel/2015/08/17908.php> > > _______________________________________________ > devel mailing list > de...@open-mpi.org <mailto:de...@open-mpi.org> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel > <http://www.open-mpi.org/mailman/listinfo.cgi/devel> > Link to this post: > http://www.open-mpi.org/community/lists/devel/2015/09/17912.php > <http://www.open-mpi.org/community/lists/devel/2015/09/17912.php> > > _______________________________________________ > devel mailing list > de...@open-mpi.org <mailto:de...@open-mpi.org> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel > <http://www.open-mpi.org/mailman/listinfo.cgi/devel> > Link to this post: > http://www.open-mpi.org/community/lists/devel/2015/09/17914.php > <http://www.open-mpi.org/community/lists/devel/2015/09/17914.php>