This can also happen when you compile your application with one MPI 
implementation (e.g., Open MPI), but then mistakenly use the "mpirun" (or 
"mpiexec") from a different MPI implementation (e.g., MPICH).


On Apr 14, 2014, at 2:32 PM, Djordje Romanic <djord...@gmail.com> wrote:

> I compiled it with: x86_64 Linux, gfortran compiler with gcc   (dmpar). dmpar 
> - distributed memory option. 
> 
> Attached is the self-generated configuration file. The architecture 
> specification settings start at line 107. I didn't use Open MPI (shared 
> memory option). 
> 
> 
> On Mon, Apr 14, 2014 at 1:23 PM, Dave Goodell (dgoodell) <dgood...@cisco.com> 
> wrote:
> On Apr 14, 2014, at 12:15 PM, Djordje Romanic <djord...@gmail.com> wrote:
> 
> > When I start wrf with mpirun -np 4 ./wrf.exe, I get this:
> > -------------------------------------------------
> >  starting wrf task            0  of            1
> >  starting wrf task            0  of            1
> >  starting wrf task            0  of            1
> >  starting wrf task            0  of            1
> > -------------------------------------------------
> > This indicates that it is not using 4 processors, but 1.
> >
> > Any idea what might be the problem?
> 
> It could be that you compiled WRF with a different MPI implementation than 
> you are using to run it (e.g., MPICH vs. Open MPI).
> 
> -Dave
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> <configure.wrf>_______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/

Reply via email to