Hi,

from the double output it looks like two identical mdruns, 
each with 1 PP process and 10 OpenMP threads, are started. 
Maybe there is something wrong with your MPI setup (did
you by mistake compile with thread-MPI instead of MPI?)

Carsten


On 09 Sep 2014, at 09:06, Albert <mailmd2...@gmail.com> wrote:

> Here are more informations from log file:
> 
> mpirun -np 2 mdrun_mpi -v -s npt2.tpr -c npt2.gro -x npt2.xtc -g
> npt2.log -gpu_id 01 -ntomp 0
> 
> 
> Number of hardware threads detected (20) does not match the number
> reported by OpenMP (10).
> Consider setting the launch configuration manually!
> 
> Number of hardware threads detected (20) does not match the number
> reported by OpenMP (10).
> Consider setting the launch configuration manually!
> Reading file npt2.tpr, VERSION 5.0.1 (single precision)
> Reading file npt2.tpr, VERSION 5.0.1 (single precision)
> Using 1 MPI process
> Using 10 OpenMP threads
> 
> 2 GPUs detected on host cudaB:
> #0: NVIDIA GeForce GTX 780 Ti, compute cap.: 3.5, ECC: no, stat: compatible
> #1: NVIDIA GeForce GTX 780 Ti, compute cap.: 3.5, ECC: no, stat: compatible
> 
> 2 GPUs user-selected for this run.
> Mapping of GPUs to the 1 PP rank in this node: #0, #1
> 
> 
> -------------------------------------------------------
> Program mdrun_mpi, VERSION 5.0.1
> Source code file:
> /soft2/plumed-2.2/gromacs-5.0.1/src/gromacs/gmxlib/gmx_detect_hardware.c, 
> line:
> 359
> 
> Fatal error:
> Incorrect launch configuration: mismatching number of PP MPI processes
> and GPUs per node.
> mdrun_mpi was started with 1 PP MPI process per node, but you provided 2
> GPUs.
> For more information and tips for troubleshooting, please check the GROMACS
> website at http://www.gromacs.org/Documentation/Errors
> -------------------------------------------------------
> 
> Halting program mdrun_mpi
> 
> gcq#314: "Do You Have Sex Maniacs or Schizophrenics or Astrophysicists
> in Your Family?" (Gogol Bordello)
> 
> --------------------------------------------------------------------------
> MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD
> with errorcode -1.
> 
> NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
> You may or may not see output from other processes, depending on
> exactly when Open MPI kills them.
> --------------------------------------------------------------------------
> Using 1 MPI process
> Using 10 OpenMP threads
> 
> 2 GPUs detected on host cudaB:
> #0: NVIDIA GeForce GTX 780 Ti, compute cap.: 3.5, ECC: no, stat: compatible
> #1: NVIDIA GeForce GTX 780 Ti, compute cap.: 3.5, ECC: no, stat: compatible
> 
> 2 GPUs user-selected for this run.
> Mapping of GPUs to the 1 PP rank in this node: #0, #1
> 
> 
> -------------------------------------------------------
> Program mdrun_mpi, VERSION 5.0.1
> Source code file:
> /soft2/plumed-2.2/gromacs-5.0.1/src/gromacs/gmxlib/gmx_detect_hardware.c, 
> line:
> 359
> 
> Fatal error:
> Incorrect launch configuration: mismatching number of PP MPI processes
> and GPUs per node.
> mdrun_mpi was started with 1 PP MPI process per node, but you provided 2
> GPUs.
> For more information and tips for troubleshooting, please check the GROMACS
> website at http://www.gromacs.org/Documentation/Errors
> -------------------------------------------------------
> 
> Halting program mdrun_mpi
> 
> gcq#56: "Lunatics On Pogo Sticks" (Red Hot Chili Peppers)
> 
> --------------------------------------------------------------------------
> MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD
> with errorcode -1.
> 
> NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
> You may or may not see output from other processes, depending on
> exactly when Open MPI kills them.
> --------------------------------------------------------------------------
> 
> 
> 
> 
> 
> 
> 
> 
> 
> On 09/08/2014 11:59 PM, Yunlong Liu wrote:
>> Same idea with Szilard.
>> 
>> How many nodes are you using?
>> On one nodes, how many MPI ranks do you have? The error is complaining about 
>> you assigned two GPUs to only one MPI process on one node. If you spread 
>> your two MPI ranks on two nodes, that means you only have one at each. Then 
>> you can't assign two GPU for only one MPI rank.
>> 
>> How many GPU do you have on one node? If there are two, you can either 
>> launch two PPMPI processes on one node and assign two GPU for them. If you 
>> only want to launch one MPI rank on each node, you can assign only one GPU 
>> for each node ( by -gpu_id 0 )
>> 
>> Yunlong
> 
> -- 
> Gromacs Users mailing list
> 
> * Please search the archive at 
> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before posting!
> 
> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
> 
> * For (un)subscribe requests visit
> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or send a 
> mail to gmx-users-requ...@gromacs.org.


--
Dr. Carsten Kutzner
Max Planck Institute for Biophysical Chemistry
Theoretical and Computational Biophysics
Am Fassberg 11, 37077 Goettingen, Germany
Tel. +49-551-2012313, Fax: +49-551-2012302
http://www.mpibpc.mpg.de/grubmueller/kutzner
http://www.mpibpc.mpg.de/grubmueller/sppexa

-- 
Gromacs Users mailing list

* Please search the archive at 
http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before posting!

* Can't post? Read http://www.gromacs.org/Support/Mailing_Lists

* For (un)subscribe requests visit
https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or send a 
mail to gmx-users-requ...@gromacs.org.

Reply via email to