Carsten Kutzner a écrit :
On Jan 10, 2009, at 8:32 PM, Nicolas wrote:
Berk Hess a écrit :
Hi,
Setting -npme 2 is ridicolous.
mdrun estimates the number of PME nodes by itself when you do not
specify -npme.
In most cases you need 1/3 or 1/4 of the nodes doing pme.
The default -npme guess of mdrun is usually not bad,
but might need to tuned a bit.
At the end of the md.log file you find the relative PP/PME load
so you can see in which direction you might need to change -npme,
if necessary.
Actually, I have tested npme ranging from 0 to 5, but 2 is well
representative of what happens. For example with 5 cores for the PME,
the perfs reach a plateau at 14-15 cores. So, setting npme to 0
systematically gives the best results. I have also tested -1. With,
npme set to -1, the performances are the same than for 0 until 8
cores. Above that, the guess is not so efficient.
Hi Nicolas,
as Berk mentioned, you should expect a different optimal number of PME
nodes for
each number of total nodes you test on. So the way to go is to fix the
number of total
nodes and vary the number of PME nodes until you find the best
performance for that
number of nodes. Then move on to another number of total nodes. I have
written
a small tool that does a part of this job for you by finding out the
optimum number
of PME nodes for a given number of total nodes. If you want to give it
a try, I can
send it to you. Typically the optimum number of PME nodes should not
be too far
off the mdrun estimate. If it is far off, this could point out some
network or MPI
problem. Note that separate PME nodes can only work if the MPI ranks
are not scattered
among the nodes, i.e. on 4-core nodes the ranks 0-3 should be on the
same node
as well as ranks 4-7 and so on. This is printed at the very start of a
parallel
simulation.
Carsten
If you could send me that script, that would be great. I've got a couple
of clusters to benchmark.
By the way, I probably didn't explain my case well enough. I'm trying to
benchmark Gromacs 4 as a function of the number of corse AND the number
of cores dedicated to the PME in a same time. My point was that I didn't
obtain better performance with npme >= 1 compared to npme=0, whatever
the cores and PME cores is. So, I was not sure whether the result was
related to the cluster itself or from a wrong usage of Gromacs.
Nicolas
--
Dr. Carsten Kutzner
Max Planck Institute for Biophysical Chemistry
Theoretical and Computational Biophysics
Am Fassberg 11
37077 Goettingen, Germany
Tel. +49-551-2012313, Fax: +49-551-2012302
http://www.mpibpc.mpg.de/home/grubmueller/ihp/ckutzne
_______________________________________________
gmx-users mailing list gmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before
posting!
Please don't post (un)subscribe requests to the list. Use thewww
interface or send it to gmx-users-requ...@gromacs.org.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php
begin:vcard
fn:Nicolas Sapay
n:Sapay;Nicolas
org:University of Calgary;Biological department
adr:;;2500 University drive NW;Calgary;AB;T2N 1N4;Canada
email;internet:nsa...@ucalgary.ca
title:Post-doctoral fellow
tel;work:403-220-6869
x-mozilla-html:TRUE
url:http://moose.bio.ucalgary.ca/
version:2.1
end:vcard
_______________________________________________
gmx-users mailing list gmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to gmx-users-requ...@gromacs.org.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php