What system is this on? CentOS7? Are you doing a VPATH build, or doing the 
build in the repo location?

Also, I assume you remembered to run autogen.pl before configure, yes?


> On Sep 1, 2015, at 10:11 AM, Cabral, Matias A <matias.a.cab...@intel.com> 
> wrote:
> 
> Hi Gilles, <>
>  
> I deleted everything, re-cloned and re-built (without my patch), but still 
> see the same issue.  The only option I’m using with configure is --prefix. I 
> even tried building with --enable-mpirun-prefix-by-default, and also passing 
> the prefix at runtime  (mpirun –prefix =/…), but I always end with the same 
> issue. Is it possible that the issue is related to configure --prefix ?
>  
> Thanks,
>  
> _MAC
>  
> From: devel [mailto:devel-boun...@open-mpi.org 
> <mailto:devel-boun...@open-mpi.org>] On Behalf Of Gilles Gouaillardet
> Sent: Monday, August 31, 2015 5:46 PM
> To: Open MPI Developers
> Subject: Re: [OMPI devel] Problem running from ompi master
>  
> Hi,
> 
> this part has been revamped recently.
> 
> at first, i would recommend you make a fresh install
> remove the install directory, and the build directory if you use VPATH, 
> re-run configure && make && make install
> that should hopefully fix the issue
> 
> Cheers,
> 
> Gilles
> 
> On 9/1/2015 9:35 AM, Cabral, Matias A wrote:
> Hi,
>  
> Before submitting a pull req I decided to test some changes on ompi master 
> branch but I’m facing an unrelated runtime error with ess pmi not being 
> found. I confirmed PATH and LD_LIBRARY_PATH are set correctly and also that 
> mca_ess_pmi.so where it should.  Any suggestions?
>  
> Thanks,
> Regards,
>  
> s-7  ~/devel/ompi> ls ./lib/openmpi/ |grep pmi
> mca_ess_pmi.la
> mca_ess_pmi.so
> mca_pmix_pmix1xx.la
> mca_pmix_pmix1xx.so
>  
> s-7 ~/devel/ompi> cat ~/.bashrc |grep -e PATH -e LD_LIBRARY_PATH
> export PATH=$HOME/devel/ompi/bin/:$PATH
> export LD_LIBRARY_PATH=$HOME/devel/ompi/lib
>  
>  
> s-7 ~ ./bin/mpirun  -host s-7,s-8 -np 2  ./osu_latency
> *** An error occurred in MPI_Init
> *** on a NULL communicator
> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
> ***    and potentially your MPI job)
> [s-7.abc.com <http://s-7.abc.com/>:56614] Local abort before MPI_INIT 
> completed successfully; not able to aggregate error messages, and not able to 
> guarantee that all other processes were killed!
> [s-7.abc.com <http://s-7.abc.com/>:56614] [[INVALID],INVALID] ORTE_ERROR_LOG: 
> Not found in file runtime/orte_init.c at line 129
> --------------------------------------------------------------------------
> A requested component was not found, or was unable to be opened.  This
> means that this component is either not installed or is unable to be
> used on your system (e.g., sometimes this means that shared libraries
> that the component requires are unable to be found/loaded).  Note that
> Open MPI stopped checking at the first component that it did not find.
>  
> Host:      s-7.abc.com <http://s-7.abc.com/>
> Framework: ess
> Component: pmi
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
>  
>   orte_ess_base_open failed
>   --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
>  
>   ompi_mpi_init: ompi_rte_init failed
>   --> Returned "Not found" (-13) instead of "Success" (0)
> --------------------------------------------------------------------------
> -------------------------------------------------------
> Primary job  terminated normally, but 1 process returned
> a non-zero exit code.. Per user-direction, the job has been aborted.
> -------------------------------------------------------
> --------------------------------------------------------------------------
> mpirun detected that one or more processes exited with non-zero status, thus 
> causing
> the job to be terminated. The first process to do so was:
>  
>   Process name: [[2886,1],0]
>   Exit code:    1
> --------------------------------------------------------------------------
> 
> 
> 
> _______________________________________________
> devel mailing list
> de...@open-mpi.org <mailto:de...@open-mpi.org>
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
> <http://www.open-mpi.org/mailman/listinfo.cgi/devel>
> Link to this post: 
> http://www.open-mpi.org/community/lists/devel/2015/08/17908.php 
> <http://www.open-mpi.org/community/lists/devel/2015/08/17908.php>
>  
> _______________________________________________
> devel mailing list
> de...@open-mpi.org <mailto:de...@open-mpi.org>
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
> <http://www.open-mpi.org/mailman/listinfo.cgi/devel>
> Link to this post: 
> http://www.open-mpi.org/community/lists/devel/2015/09/17912.php 
> <http://www.open-mpi.org/community/lists/devel/2015/09/17912.php>

Reply via email to