[gmx-users] Mismatching number of PP MPI processes and GPUs per node

2013-03-11 Thread George Patargias
Hello

Sorry for posting this again.

I am trying to run GROMACS 4.6 compiled with MPI and GPU acceleration
(CUDA 5.0 lib) using the following SGE batch script.

#!/bin/sh
#$ -V
#$ -S /bin/sh
#$ -N test-gpus
#$ -l h=xgrid-node02
#$ -pe mpi_fill_up 12
#$ -cwd

source /opt/NetUsers/pgkeka/gromacs-4.6_gpu_mpi/bin/GMXRC
export DYLD_LIBRARY_PATH=/Developer/NVIDIA/CUDA-5.0/lib:$DYLD_LIBRARY_PATH

mpirun -np 12 mdrun_mpi -s test.tpr -deffnm test_out -nb gpu

After detection of the installed GPU card

1 GPU detected on host xgrid-node02.xgrid:
  #0: NVIDIA Quadro 4000, compute cap.: 2.0, ECC:  no, stat: compatible

GROMACS issues the following error

Incorrect launch configuration: mismatching number of PP MPI processes and
GPUs per node. mdrun_mpi was started with 12 PP MPI processes per node,
but only 1 GPU were detected.

It can't be that we need to run GROMACS only on a single core so that it
matches the single GPU card.

Do you have any idea what has to be done?

Many thanks.

Dr. George Patargias
Postdoctoral Researcher
Biomedical Research Foundation
Academy of Athens
4, Soranou Ephessiou
115 27
Athens
Greece

Office: +302106597568

-- 
gmx-users mailing listgmx-users@gromacs.org
http://lists.gromacs.org/mailman/listinfo/gmx-users
* Please search the archive at 
http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
* Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to gmx-users-requ...@gromacs.org.
* Can't post? Read http://www.gromacs.org/Support/Mailing_Lists


Re: [gmx-users] Mismatching number of PP MPI processes and GPUs per node

2013-03-11 Thread Carsten Kutzner
Hi,

On Mar 11, 2013, at 10:50 AM, George Patargias g...@bioacademy.gr wrote:

 Hello
 
 Sorry for posting this again.
 
 I am trying to run GROMACS 4.6 compiled with MPI and GPU acceleration
 (CUDA 5.0 lib) using the following SGE batch script.
 
 #!/bin/sh
 #$ -V
 #$ -S /bin/sh
 #$ -N test-gpus
 #$ -l h=xgrid-node02
 #$ -pe mpi_fill_up 12
 #$ -cwd
 
 source /opt/NetUsers/pgkeka/gromacs-4.6_gpu_mpi/bin/GMXRC
 export DYLD_LIBRARY_PATH=/Developer/NVIDIA/CUDA-5.0/lib:$DYLD_LIBRARY_PATH
 
 mpirun -np 12 mdrun_mpi -s test.tpr -deffnm test_out -nb gpu
 
 After detection of the installed GPU card
 
 1 GPU detected on host xgrid-node02.xgrid:
  #0: NVIDIA Quadro 4000, compute cap.: 2.0, ECC:  no, stat: compatible
 
 GROMACS issues the following error
 
 Incorrect launch configuration: mismatching number of PP MPI processes and
 GPUs per node. mdrun_mpi was started with 12 PP MPI processes per node,
 but only 1 GPU were detected.
 
 It can't be that we need to run GROMACS only on a single core so that it
 matches the single GPU card.
Have you compiled mdrun_mpi with OpenMP threads support? Then, if you
do 

mpirun -np 1 mdrun_mpi …

it should start one MPI process with 12 OpenMP threads, which should give
you what you want. You can also manually specify the number of OpenMP threads
by adding 

-ntomp 12

Carsten

 

 
 Do you have any idea what has to be done?
 
 Many thanks.
 
 Dr. George Patargias
 Postdoctoral Researcher
 Biomedical Research Foundation
 Academy of Athens
 4, Soranou Ephessiou
 115 27
 Athens
 Greece
 
 Office: +302106597568
 
 -- 
 gmx-users mailing listgmx-users@gromacs.org
 http://lists.gromacs.org/mailman/listinfo/gmx-users
 * Please search the archive at 
 http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
 * Please don't post (un)subscribe requests to the list. Use the 
 www interface or send it to gmx-users-requ...@gromacs.org.
 * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists


--
Dr. Carsten Kutzner
Max Planck Institute for Biophysical Chemistry
Theoretical and Computational Biophysics
Am Fassberg 11, 37077 Goettingen, Germany
Tel. +49-551-2012313, Fax: +49-551-2012302
http://www.mpibpc.mpg.de/grubmueller/kutzner
http://www.mpibpc.mpg.de/grubmueller/sppexa

--
gmx-users mailing listgmx-users@gromacs.org
http://lists.gromacs.org/mailman/listinfo/gmx-users
* Please search the archive at 
http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
* Please don't post (un)subscribe requests to the list. Use the
www interface or send it to gmx-users-requ...@gromacs.org.
* Can't post? Read http://www.gromacs.org/Support/Mailing_Lists


RE: [gmx-users] mdrun WARING and crash

2013-03-11 Thread L.Liu
Hallo Justin,

Thank you for your comments. 
Taking your suggestions, I set nrexcl=1, and comment the [pairs] section, 
because there is no special case of non-bonded interactions to declare,  then 
we try to see what happens.
Now we minimize it by steep, then by cg, both of the processor are very quick 
done, because after around 7000 steps, the energy can not go further down any 
more.
Then we finish the mdrun, the energies output are like:

   Step   Time Lambda
  1   10.00.0

   Energies (kJ/mol)
LJ (SR)   Coulomb (SR)  PotentialKinetic En.   Total Energy
1.49468e+050.0e+001.49468e+058.73890e+031.58207e+05
Temperature Pressure (bar)
4.38208e+029.93028e+04

Although this time no running error, we find the outputs are extremely wired, 
for example through 
VMD conf.gro traj.xtc
we find 0 frame a homogeneous box, starting from the first step, the box 
becomes a lattice, which is far away from our expected the Polymer melt system 
should be.

The force parameters are taken from literature PRL 85(5), 1128(2000), I am 
still very worried about the format of my input files. Could you please give 
me, a very beginner a help. 

Thanks a lot.

Kind regards,
Li

From: gmx-users-boun...@gromacs.org [gmx-users-boun...@gromacs.org] on behalf 
of Justin Lemkul [jalem...@vt.edu]
Sent: Thursday, February 28, 2013 3:02 PM
To: Discussion list for GROMACS users
Subject: Re: [gmx-users] mdrun WARING and crash

On 2/28/13 6:59 AM, l@utwente.nl wrote:
 Hallo Justin,

 Thank you for you help. I have read the previous discussions on this topic, 
 which is very helpful.
 The link is: 
 http://gromacs.5086.n6.nabble.com/What-is-the-purpose-of-the-pairs-section-td5003598.html
 Well, there are still something I want to make sure, which might be the 
 reason of mdrun crash of my system.

 ###Introduction of system##
 Linear Polyethylene melt:  each chain contains 16 beads, each bead coarse 
 grained 3 monomers. Number of  chain in the box is 64.

 Force Field##
 ffbonded.itp
 [ bondtypes ]
 ; FENE, Ro = 1.5 sigma and kb = 30 epsilon/sigma^2
 ;   ij funcb0 (nm) kb (kJ/mol nm^2)
   CH2   CH27   0.795   393.

 ffnonbonded.itp
 [ atomtypes ]
 ; epsilon / kB = 443K
 ;name  at.num  mass (au)   charge   ptype sigma (nm)epsilon 
 (kJ/mol)
 CH2  142.3   0.000   A   0.5300   
3.68133

 [ nonbond_params ]
; i  jfunc  sigma   epsilon
 CH2   CH210.5303.68133

 [ pairtypes ]
;  i  jfunc  sigma   epsilon
 CH2   CH21  0.533.68133

 topology##
 [ defaults ]
 ; nbfunccomb-rule   gen-pairs   fudgeLJ   fudgeQQ
  1  2no  1.0  1.0

 ; The force field files to be included
 #include ../forcefield/forcefield.itp

 [ moleculetype ]
 ; name  nrexcl
 PE  0
 [atoms]
 ;   nrtype   resnr  residuatomcgnr  charge
   1 CH2   1PE  C   1  0.0
   2 CH2   1PE  C   2  0.0
   3 CH2   1PE  C   3  0.0
   4 CH2   1PE  C   4  0.0
  ..
   15CH2   1PE  C   15 0.0
   16CH2   1PE  C   16 0.0

 [ bonds ]
 ;  aiaj  funct   c0   c1
  1 2  7  0.795 393.
  2 3 7  0.795 393.
  3 4 7  0.795 393.
 ..
  1415 7  0.795 393.
  1516 7  0.795 393.

 [ pairs ]
 ;  aiaj  funct   c0   c1
  1 2   1 0.53 3.68133
  1 3   1 0.53 3.68133
  1 4   1 0.53 3.68133
  2 3   1 0.53 3.68133
  2 4   1 0.53 3.68133
  2 5   1 0.53 3.68133
  ..
  1415  1 0.53 3.68133
  1416  1 0.53 3.68133
  1516  1 0.53 3.68133

 ##mdp###

 ;directories to include in your topology. Format
 include  = -I/home/otterw/Install/gromacs-4.6/src/gmxlib
 integrator   = md
 ;
 ;  RUN CONTROL  *
 ;
 tinit = 0; start time to run
 dt   = 1e-3   

Re: [gmx-users] mdrun WARING and crash

2013-03-11 Thread Justin Lemkul
On Monday, March 11, 2013, wrote:

 Hallo Justin,

 Thank you for your comments.
 Taking your suggestions, I set nrexcl=1, and comment the [pairs] section,
 because there is no special case of non-bonded interactions to declare,
  then we try to see what happens.
 Now we minimize it by steep, then by cg, both of the processor are very
 quick done, because after around 7000 steps, the energy can not go further
 down any more.
 Then we finish the mdrun, the energies output are like:

Step   Time Lambda
   1   10.00.0

Energies (kJ/mol)
 LJ (SR)   Coulomb (SR)  PotentialKinetic En.   Total Energy
 1.49468e+050.0e+001.49468e+058.73890e+031.58207e+05
 Temperature Pressure (bar)
 4.38208e+029.93028e+04

 Although this time no running error, we find the outputs are extremely
 wired, for example through
 VMD conf.gro traj.xtc
 we find 0 frame a homogeneous box, starting from the first step, the box
 becomes a lattice, which is far away from our expected the Polymer melt
 system should be.

 The force parameters are taken from literature PRL 85(5), 1128(2000), I am
 still very worried about the format of my input files. Could you please
 give me, a very beginner a help.


Please provide links to images. This is probably not a big deal as long as
the simulation is actually running, since a triclinic representation of the
unit cell is used.

-Justin



 Thanks a lot.
 Kind regards,
 Li
 
 From: gmx-users-boun...@gromacs.org javascript:; [
 gmx-users-boun...@gromacs.org javascript:;] on behalf of Justin Lemkul [
 jalem...@vt.edu javascript:;]
 Sent: Thursday, February 28, 2013 3:02 PM
 To: Discussion list for GROMACS users
 Subject: Re: [gmx-users] mdrun WARING and crash

 On 2/28/13 6:59 AM, l@utwente.nl wrote:
  Hallo Justin,
 
  Thank you for you help. I have read the previous discussions on this
 topic, which is very helpful.
  The link is:
 http://gromacs.5086.n6.nabble.com/What-is-the-purpose-of-the-pairs-section-td5003598.html
  Well, there are still something I want to make sure, which might be the
 reason of mdrun crash of my system.
 
  ###Introduction of system##
  Linear Polyethylene melt:  each chain contains 16 beads, each bead
 coarse grained 3 monomers. Number of  chain in the box is 64.
 
  Force Field##
  ffbonded.itp
  [ bondtypes ]
  ; FENE, Ro = 1.5 sigma and kb = 30 epsilon/sigma^2
  ;   ij funcb0 (nm) kb (kJ/mol nm^2)
CH2   CH27   0.795   393.
 
  ffnonbonded.itp
  [ atomtypes ]
  ; epsilon / kB = 443K
  ;name  at.num  mass (au)   charge   ptype sigma (nm)
  epsilon (kJ/mol)
  CH2  142.3   0.000   A   0.5300
  3.68133
 
  [ nonbond_params ]
 ; i  jfunc  sigma   epsilon
  CH2   CH210.5303.68133
 
  [ pairtypes ]
 ;  i  jfunc  sigma   epsilon
  CH2   CH21  0.533.68133
 
  topology##
  [ defaults ]
  ; nbfunccomb-rule   gen-pairs   fudgeLJ   fudgeQQ
   1  2no  1.0  1.0
 
  ; The force field files to be included
  #include ../forcefield/forcefield.itp
 
  [ moleculetype ]
  ; name  nrexcl
  PE  0
  [atoms]
  ;   nrtype   resnr  residuatomcgnr  charge
1 CH2   1PE  C   1  0.0
2 CH2   1PE  C   2  0.0
3 CH2   1PE  C   3  0.0
4 CH2   1PE  C   4  0.0
   ..
15CH2   1PE  C   15 0.0
16CH2   1PE  C   16 0.0
 
  [ bonds ]
  ;  aiaj  funct   c0   c1
   1 2  7  0.795 393.
   2 3 7  0.795 393.
   3 4 7  0.795 393.
  ..
   1415 7  0.795 393.
   1516 7  0.795 393.
 
  [ pairs ]
  ;  ai   --
 gmx-users mailing listgmx-users@gromacs.org javascript:;
 http://lists.gromacs.org/mailman/listinfo/gmx-users
 * Please search the archive at
 http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
 * Please don't post (un)subscribe requests to the list. Use the
 www interface or send it to gmx-users-requ...@gromacs.org javascript:;.
 * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
 --
 gmx-users mailing listgmx-users@gromacs.org javascript:;
 http://lists.gromacs.org/mailman/listinfo/gmx-users
 * Please search the archive at
 http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
 * Please don't post (un)subscribe requests to the list. Use the
 www interface or send it to 

Re: [gmx-users] only 4 zero eigvals for NMA (with 10^-6 emtol for l-bfgs)

2013-03-11 Thread Hyuntae Na
I printed all small 50 eigenvalues, and their first 9 are listed as follows as 
before:* -1.8476593e-06* -1.2164109e-06* -1.4830436e-07* 1.3104790e-06* 
1.1364704e+00* 1.7418209e+00* 2.2836573e+00* 3.7218206e+00* 4.4875873e+00 As 
you see, the first three of them are negative and close to zero, which can be 
considered as zero. do you know any reason of this,  having only 4 (close to) 
zero eigenvalues ?? -- Hyuntae  Message: 3
 Date: Sun, 10 Mar 2013 23:10:05 +0100
 From: lloyd riggs lloyd.ri...@gmx.ch
 Subject: Re: [gmx-users] only 4 zero eigvals for NMA (with 10^-6 emtol
   for l-bfgs)
 To: Discussion list for GROMACS users gmx-users@gromacs.org
 Message-ID: 20130310221005.162...@gmx.net
 Content-Type: text/plain; charset=utf-8
 
 
 I can imagine why you would go past the first few, but does it print the 
 zero's or just negate them from the equation, as there 0?
 
 Stephan
  Original-Nachricht 
  Datum: Sun, 10 Mar 2013 14:34:52 -0500
  Von: Hyuntae Na h...@hotmail.com
  An: gmx-users@gromacs.org gmx-users@gromacs.org
  Betreff: [gmx-users] only 4 zero eigvals for NMA (with 10^-6 emtol for  
  l-bfgs)
 
  Dear, This is the same issue that I asked last time. I wonder (1) if I
  setup the test wrong, (2) if NMA *can* have smaller than 6 eigenvalues even
  though the conformation of a protein is in equilibrium  in the enough 
  degree,
  or (2) if it gromacs does not make it in equilibrium because of the limit
  of the tolerance that gromacs can take care of. Is there anyone who can
  give me an advice for this? I updated the sample location in the case that 
  the
  previous Skype link is not available. I am studying several protein
  fluctuation proteins using NMA (Normal Mode Analysis). In the mean time, I 
  found
  out in many cases that NMA hessian matrix has about 3-4 zero eigenvalues
  even though having 6 zero eigenvalues are expected. I tried to reduce the
  tolerance level lesser than 10^-6 for the l-bfgs minimization, but still it
  have only 4 zero eigenvalues. Following is the 9 smallest eigenvalues (you
  can see that it has only 4 close-to-zero eigenvalues):
  -1.8476593e-06-1.2164109e-06-1.4830436e-071.3104790e-061.1364704e+001.7418209e+002.2836573e+003.7218206e+004.4875873e+00
  Theoretically, NMA should have 6 zero eigenvalues which represents the
  freedom of the rotation and translation. I wonder why this happens.
   
  Is there anyone who can help me for this problem?
   
  In order to specify the case, I leave a link to download a zip file that
  contains files for one NMA tests having the problem: script, mdp, pdb, gro,
  topol, etc. The zip file also includes the eigenvalue/vector files.
  http://129.186.69.109/NMA-test.zip
   
  Thank you very much.
   
  -- Hyuntae
  --
gmx-users mailing listgmx-users@gromacs.org
http://lists.gromacs.org/mailman/listinfo/gmx-users
* Please search the archive at 
http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
* Please don't post (un)subscribe requests to the list. Use the
www interface or send it to gmx-users-requ...@gromacs.org.
* Can't post? Read http://www.gromacs.org/Support/Mailing_Lists


[gmx-users] removing rotational and translational motion (holonomic contraint)

2013-03-11 Thread Evelyne Deplazes
Hi guys

I am running a set of umbrella sampling simulations using gromacs 333 and the 
pull code. The idea is to apply an umbrella potential to restrain the 
distance between a protein and a ligand in the x direction. Unfortunately, the 
rotational and translational movement of the protein during the MD run 
interferes with my restraining of the protein-ligand distance. I am therefore 
interested in applying a rotational and translational constraint to my protein.

Amadei et al described a solution to this problem in their J Chem Phys paper 
(see below). The problem was also discussed in the following gromacs post:

http://lists.gromacs.org/pipermail/gmx-users/2003-May/005591.html

Extract from the post:

For a simulation in solvent it is less trivial but feasible as was shown
by Andrea Amadei and coworkers. They devised a method which constrains
roto-translational movements of a solute, described in J. Chem. Phys.
112(1) pp. 9-23.

They've implemented the method in Gromacs and a patch can be obtained
from Danilo Roccatano ( roccata at caspur dot it ) from that group. We
will be trying to implement things properly into Gromacs in the near
future here in Groningen, but it has to wait a bit. If it's done we'll
put it forward for a future release.

There is another post in 2006 that refers to a patch for removing 
roto-translational motion but the link no longer works.

http://gromacs.5086.n6.nabble.com/Protein-rotation-in-triclinic-waterbox-td4413734.htmhttp://gromacs.5086.n6.nabble.com/Protein-rotation-in-triclinic-waterbox-td4413734.html

Finally, there is a more recent post (2012) stating it has been implemented in 
version 4.5.1 and 4.5.3 but not released.

http://gromacs.5086.n6.nabble.com/roto-translational-constraints-td4390750.html

Has this feature been implemented in the official gromacs version? Can I obtain 
the patch and use it with gromacs 333?

Thanks a lot
Evelyne


--
gmx-users mailing listgmx-users@gromacs.org
http://lists.gromacs.org/mailman/listinfo/gmx-users
* Please search the archive at 
http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
* Please don't post (un)subscribe requests to the list. Use the
www interface or send it to gmx-users-requ...@gromacs.org.
* Can't post? Read http://www.gromacs.org/Support/Mailing_Lists


Re: [gmx-users] query for gromacs-4.5.4

2013-03-11 Thread Chaitali Chandratre
Hello Sir,

Actually I have been given work to setup gromacs-4.5.4 in our cluster which
is being used
by users.I am not gromacs user and not aware of its internal details.
I have got only .tpr file from user and I need to test my installation
using that .tpr file.

It works fine for 2 nodes 8 processes , 1 node 8 processes.
 But when number of processes are equal to 16 it gives segmentation fault
and
 if number of processes are equal to 32 it gives
error message like
 HYD_pmcd_pmiserv_send_signal (./pm/pmiserv/pmiserv_cb.c:221): assert
(!closed) failed
 ui_cmd_cb (./pm/pmiserv/pmiserv_pmci.c:128): unable to send SIGUSR1
downstream
 HYDT_dmxu_poll_wait_for_event (./tools/demux/demux_poll.c:77): callback
returned error status
 HYD_pmci_wait_for_completion (./pm/pmiserv/pmiserv_pmci.c:388): error
waiting for event
[ main (./ui/mpich/mpiexec.c:718): process manager error waiting for
completion

I am not clear like whether problem is there in my installation or what?

Thanks and Regards,
   Chaitalij

On Wed, Mar 6, 2013 at 5:41 PM, Justin Lemkul jalem...@vt.edu wrote:



 On 3/6/13 4:20 AM, Chaitali Chandratre wrote:

 Dear Sir ,

 I am new to this installation and setup area. I need some information for
 -stepout option for


 What more information do you need?


  mdrun_mpi and also probable causes for segmentation fault in
  gromacs-4.5.4.
 (my node having 64 GB mem running with 16 processes, nsteps = 2000)


 There are too many causes to name.  Please consult http://www.gromacs.org/
 **Documentation/Terminology/**Blowing_Uphttp://www.gromacs.org/Documentation/Terminology/Blowing_Up.
  If you need further help, please be more specific, including a description
 of the system, steps taken to minimize and/or equilibrate it, and any
 complete .mdp file(s) that you are using.

 -Justin

 --
 ==**==

 Justin A. Lemkul, Ph.D.
 Research Scientist
 Department of Biochemistry
 Virginia Tech
 Blacksburg, VA
 jalemkul[at]vt.edu | (540) 231-9080
 http://www.bevanlab.biochem.**vt.edu/Pages/Personal/justinhttp://www.bevanlab.biochem.vt.edu/Pages/Personal/justin

 ==**==
 --
 gmx-users mailing listgmx-users@gromacs.org
 http://lists.gromacs.org/**mailman/listinfo/gmx-usershttp://lists.gromacs.org/mailman/listinfo/gmx-users
 * Please search the archive at http://www.gromacs.org/**
 Support/Mailing_Lists/Searchhttp://www.gromacs.org/Support/Mailing_Lists/Searchbefore
  posting!
 * Please don't post (un)subscribe requests to the list. Use the www
 interface or send it to gmx-users-requ...@gromacs.org.
 * Can't post? Read 
 http://www.gromacs.org/**Support/Mailing_Listshttp://www.gromacs.org/Support/Mailing_Lists

-- 
gmx-users mailing listgmx-users@gromacs.org
http://lists.gromacs.org/mailman/listinfo/gmx-users
* Please search the archive at 
http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
* Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to gmx-users-requ...@gromacs.org.
* Can't post? Read http://www.gromacs.org/Support/Mailing_Lists