Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-12 Thread Martin Höfling
Am Mittwoch 12 November 2008 06:18:14 schrieb vivek sharma:

> Everybody thanks for your usefull suggestions..
> What do you mean by % imbalance reported in log file. I don't know how to
> assign the specific load to PME, but I can see that around 37% of the
> computation is being used by PME.
> I am not assigning PME nodes separately. I have no idea of dynamic load
> balancing and how to use it  ?

The switch is -dlb
Normally it's turned on automatically (auto), but you can turn it on/off with 
this switch as well.

-dds might be of interest as well...

Best
Martin
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php


Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-11 Thread Mark Abraham

vivek sharma wrote:



2008/11/11 Justin A. Lemkul <[EMAIL PROTECTED] >



vivek sharma wrote:

HI MArtin,
I am using here the infiniband having speed more than 10
gbps..Can you suggest some option to scale better in this case.


What % imbalance is being reported in the log file?  What fraction
of the load is being assigned to PME, from grompp?  How many
processors are you assigning to the PME calculation?  Are you using
dynamic load balancing?


Everybody thanks for your usefull suggestions..
What do you mean by % imbalance reported in log file. I don't know how 
to assign the specific load to PME, but I can see that around 37% of the 
computation is being used by PME.
I am not assigning PME nodes separately. I have no idea of dynamic load 
balancing and how to use it  ?


For the moment, you just need to look for information in your grompp 
output and your log file, like Justin asked.


To find out how to use things, check out mdrun -h and the version 4 
manual. There's some relevant discussion in this thread 
http://www.gromacs.org/pipermail/gmx-users/2008-October/036856.html


Mark
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]

Can't post? Read http://www.gromacs.org/mailing_lists/users.php


Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-11 Thread vivek sharma
2008/11/11 Justin A. Lemkul <[EMAIL PROTECTED]>

>
>
> vivek sharma wrote:
>
>> HI MArtin,
>> I am using here the infiniband having speed more than 10 gbps..Can you
>> suggest some option to scale better in this case.
>>
>>
> What % imbalance is being reported in the log file?  What fraction of the
> load is being assigned to PME, from grompp?  How many processors are you
> assigning to the PME calculation?  Are you using dynamic load balancing?


Everybody thanks for your usefull suggestions..
What do you mean by % imbalance reported in log file. I don't know how to
assign the specific load to PME, but I can see that around 37% of the
computation is being used by PME.
I am not assigning PME nodes separately. I have no idea of dynamic load
balancing and how to use it  ?

Looking forward for answers...

With Thanks,
Vivek

>
>
> All of these factors affect performance.
>
> -Justin
>
>  With Thanks,
>> Vivek
>>
>> 2008/11/11 Martin Höfling <[EMAIL PROTECTED] > [EMAIL PROTECTED]>>
>>
>>
>>Am Dienstag 11 November 2008 12:06:06 schrieb vivek sharma:
>>
>>
>> > I have also tried scaling gromacs for a number of nodes but
>>was not
>> > able to optimize it beyond 20 processor..on 20 nodes i.e. 1
>>processor per
>>
>>As mentioned before, performance strongly depends on the type of
>>interconnect
>>you're using between your processes. Shared Memory, Ethernet,
>>Infiniband,
>>NumaLink, whatever...
>>
>>I assume you're using ethernet (100/1000 MBit?), you can tune here
>>to some
>>extend as described in:
>>
>>Kutzner, C.; Spoel, D. V. D.; Fechner, M.; Lindahl, E.; Schmitt, U.
>>W.; Groot,
>>B. L. D. & Grubmüller, H. Speeding up parallel GROMACS on high-latency
>>networks Journal of Computational Chemistry, 2007
>>
>>...but be aware that principal limitations of ethernet remain. To
>>come around
>>this, you might consider to invest in the interconnect. If you can
>>come out
>>with <16 cores, shared memory nodes will give you the "biggest bang
>>for the
>>buck".
>>
>>Best
>> Martin
>>___
>>gmx-users mailing listgmx-users@gromacs.org
>>
>>http://www.gromacs.org/mailman/listinfo/gmx-users
>>Please search the archive at http://www.gromacs.org/search before
>>posting!
>>Please don't post (un)subscribe requests to the list. Use the
>>www interface or send it to [EMAIL PROTECTED]
>>.
>>Can't post? Read http://www.gromacs.org/mailing_lists/users.php
>>
>>
>>
>> 
>>
>> ___
>> gmx-users mailing listgmx-users@gromacs.org
>> http://www.gromacs.org/mailman/listinfo/gmx-users
>> Please search the archive at http://www.gromacs.org/search before
>> posting!
>> Please don't post (un)subscribe requests to the list. Use the www
>> interface or send it to [EMAIL PROTECTED]
>> Can't post? Read http://www.gromacs.org/mailing_lists/users.php
>>
>
> --
> 
>
> Justin A. Lemkul
> Graduate Research Assistant
> Department of Biochemistry
> Virginia Tech
> Blacksburg, VA
> jalemkul[at]vt.edu | (540) 231-9080
> http://www.bevanlab.biochem.vt.edu/Pages/Personal/justin
>
> 
> ___
> gmx-users mailing listgmx-users@gromacs.org
> http://www.gromacs.org/mailman/listinfo/gmx-users
> Please search the archive at http://www.gromacs.org/search before posting!
> Please don't post (un)subscribe requests to the list. Use the www interface
> or send it to [EMAIL PROTECTED]
> Can't post? Read http://www.gromacs.org/mailing_lists/users.php
>
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php

Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-11 Thread Christian Seifert
A page on the wiki with further information and hints would be nice.
Topic: "improving performance with GMX4" or "Pimp my GMX4" ;-)

The beta manualpage of mdrun (version4) is not very comprehensible/user
friendly in my eyes.

- Christian


On Tue, 2008-11-11 at 09:12 -0500, Justin A. Lemkul wrote:
> 
> vivek sharma wrote:
> > HI MArtin,
> > I am using here the infiniband having speed more than 10 gbps..Can you 
> > suggest some option to scale better in this case.
> > 
> 
> What % imbalance is being reported in the log file?  What fraction of the 
> load 
> is being assigned to PME, from grompp?  How many processors are you assigning 
> to 
> the PME calculation?  Are you using dynamic load balancing?
> 
> All of these factors affect performance.
> 
> -Justin
> 
> > With Thanks,
> > Vivek
> > 
> > 2008/11/11 Martin Höfling <[EMAIL PROTECTED] 
> > >
> > 
> > Am Dienstag 11 November 2008 12:06:06 schrieb vivek sharma:
> > 
> > 
> >  > I have also tried scaling gromacs for a number of nodes but
> > was not
> >  > able to optimize it beyond 20 processor..on 20 nodes i.e. 1
> > processor per
> > 
> > As mentioned before, performance strongly depends on the type of
> > interconnect
> > you're using between your processes. Shared Memory, Ethernet,
> > Infiniband,
> > NumaLink, whatever...
> > 
> > I assume you're using ethernet (100/1000 MBit?), you can tune here
> > to some
> > extend as described in:
> > 
> > Kutzner, C.; Spoel, D. V. D.; Fechner, M.; Lindahl, E.; Schmitt, U.
> > W.; Groot,
> > B. L. D. & Grubmüller, H. Speeding up parallel GROMACS on high-latency
> > networks Journal of Computational Chemistry, 2007
> > 
> > ...but be aware that principal limitations of ethernet remain. To
> > come around
> > this, you might consider to invest in the interconnect. If you can
> > come out
> > with <16 cores, shared memory nodes will give you the "biggest bang
> > for the
> > buck".
> > 
> > Best
> >  Martin
> > ___
> > gmx-users mailing listgmx-users@gromacs.org
> > 
> > http://www.gromacs.org/mailman/listinfo/gmx-users
> > Please search the archive at http://www.gromacs.org/search before
> > posting!
> > Please don't post (un)subscribe requests to the list. Use the
> > www interface or send it to [EMAIL PROTECTED]
> > .
> > Can't post? Read http://www.gromacs.org/mailing_lists/users.php
> > 
> > 
> > 
> > 
> > 
> > ___
> > gmx-users mailing listgmx-users@gromacs.org
> > http://www.gromacs.org/mailman/listinfo/gmx-users
> > Please search the archive at http://www.gromacs.org/search before posting!
> > Please don't post (un)subscribe requests to the list. Use the 
> > www interface or send it to [EMAIL PROTECTED]
> > Can't post? Read http://www.gromacs.org/mailing_lists/users.php
> 
-- 
M. Sc. Christian Seifert
Department of Biophysics
University of Bochum
ND 04/67
44780 Bochum
Germany
Tel: +49 (0)234 32 28363
Fax: +49 (0)234 32 14626
E-Mail: [EMAIL PROTECTED]
Web: http://www.bph.rub.de

___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php


Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-11 Thread Justin A. Lemkul



vivek sharma wrote:

HI MArtin,
I am using here the infiniband having speed more than 10 gbps..Can you 
suggest some option to scale better in this case.




What % imbalance is being reported in the log file?  What fraction of the load 
is being assigned to PME, from grompp?  How many processors are you assigning to 
the PME calculation?  Are you using dynamic load balancing?


All of these factors affect performance.

-Justin


With Thanks,
Vivek

2008/11/11 Martin Höfling <[EMAIL PROTECTED] 
>


Am Dienstag 11 November 2008 12:06:06 schrieb vivek sharma:


 > I have also tried scaling gromacs for a number of nodes but
was not
 > able to optimize it beyond 20 processor..on 20 nodes i.e. 1
processor per

As mentioned before, performance strongly depends on the type of
interconnect
you're using between your processes. Shared Memory, Ethernet,
Infiniband,
NumaLink, whatever...

I assume you're using ethernet (100/1000 MBit?), you can tune here
to some
extend as described in:

Kutzner, C.; Spoel, D. V. D.; Fechner, M.; Lindahl, E.; Schmitt, U.
W.; Groot,
B. L. D. & Grubmüller, H. Speeding up parallel GROMACS on high-latency
networks Journal of Computational Chemistry, 2007

...but be aware that principal limitations of ethernet remain. To
come around
this, you might consider to invest in the interconnect. If you can
come out
with <16 cores, shared memory nodes will give you the "biggest bang
for the
buck".

Best
 Martin
___
gmx-users mailing listgmx-users@gromacs.org

http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before
posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php





___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]

Can't post? Read http://www.gromacs.org/mailing_lists/users.php


--


Justin A. Lemkul
Graduate Research Assistant
Department of Biochemistry
Virginia Tech
Blacksburg, VA
jalemkul[at]vt.edu | (540) 231-9080
http://www.bevanlab.biochem.vt.edu/Pages/Personal/justin


___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php


Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-11 Thread Mark Abraham

vivek sharma wrote:

Hi All,
one thing I forgot to mention I am getting here around 6 ns/day...for a 
protein of size around 2600 atoms..


Much more relevant is how much water...

You can also be rate limited by I/O if you have poor hardware and/or are 
writing to disk excessively.


Mark
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]

Can't post? Read http://www.gromacs.org/mailing_lists/users.php


Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-11 Thread vivek sharma
Hi All,
one thing I forgot to mention I am getting here around 6 ns/day...for a
protein of size around 2600 atoms..

With Thanks,
Vivek

2008/11/11 vivek sharma <[EMAIL PROTECTED]>

> HI MArtin,
> I am using here the infiniband having speed more than 10 gbps..Can you
> suggest some option to scale better in this case.
>
> With Thanks,
> Vivek
>
> 2008/11/11 Martin Höfling <[EMAIL PROTECTED]>
>
> Am Dienstag 11 November 2008 12:06:06 schrieb vivek sharma:
>>
>>
>> > I have also tried scaling gromacs for a number of nodes but was not
>> > able to optimize it beyond 20 processor..on 20 nodes i.e. 1 processor
>> per
>>
>> As mentioned before, performance strongly depends on the type of
>> interconnect
>> you're using between your processes. Shared Memory, Ethernet, Infiniband,
>> NumaLink, whatever...
>>
>> I assume you're using ethernet (100/1000 MBit?), you can tune here to some
>> extend as described in:
>>
>> Kutzner, C.; Spoel, D. V. D.; Fechner, M.; Lindahl, E.; Schmitt, U. W.;
>> Groot,
>> B. L. D. & Grubmüller, H. Speeding up parallel GROMACS on high-latency
>> networks Journal of Computational Chemistry, 2007
>>
>> ...but be aware that principal limitations of ethernet remain. To come
>> around
>> this, you might consider to invest in the interconnect. If you can come
>> out
>> with <16 cores, shared memory nodes will give you the "biggest bang for
>> the
>> buck".
>>
>> Best
>>   Martin
>> ___
>> gmx-users mailing listgmx-users@gromacs.org
>> http://www.gromacs.org/mailman/listinfo/gmx-users
>> Please search the archive at http://www.gromacs.org/search before
>> posting!
>> Please don't post (un)subscribe requests to the list. Use the
>> www interface or send it to [EMAIL PROTECTED]
>> Can't post? Read http://www.gromacs.org/mailing_lists/users.php
>>
>
>
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php

Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-11 Thread vivek sharma
HI MArtin,
I am using here the infiniband having speed more than 10 gbps..Can you
suggest some option to scale better in this case.

With Thanks,
Vivek

2008/11/11 Martin Höfling <[EMAIL PROTECTED]>

> Am Dienstag 11 November 2008 12:06:06 schrieb vivek sharma:
>
>
> > I have also tried scaling gromacs for a number of nodes but was not
> > able to optimize it beyond 20 processor..on 20 nodes i.e. 1 processor per
>
> As mentioned before, performance strongly depends on the type of
> interconnect
> you're using between your processes. Shared Memory, Ethernet, Infiniband,
> NumaLink, whatever...
>
> I assume you're using ethernet (100/1000 MBit?), you can tune here to some
> extend as described in:
>
> Kutzner, C.; Spoel, D. V. D.; Fechner, M.; Lindahl, E.; Schmitt, U. W.;
> Groot,
> B. L. D. & Grubmüller, H. Speeding up parallel GROMACS on high-latency
> networks Journal of Computational Chemistry, 2007
>
> ...but be aware that principal limitations of ethernet remain. To come
> around
> this, you might consider to invest in the interconnect. If you can come out
> with <16 cores, shared memory nodes will give you the "biggest bang for the
> buck".
>
> Best
>   Martin
> ___
> gmx-users mailing listgmx-users@gromacs.org
> http://www.gromacs.org/mailman/listinfo/gmx-users
> Please search the archive at http://www.gromacs.org/search before posting!
> Please don't post (un)subscribe requests to the list. Use the
> www interface or send it to [EMAIL PROTECTED]
> Can't post? Read http://www.gromacs.org/mailing_lists/users.php
>
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php

Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-11 Thread Carsten Kutzner

Hi Vivek,

if you use separate PME nodes (-npme) then one group of the processors  
will calculate the long-range (reciprocal space) part while the  
remaining processors do the short-range (direct space) part of the  
Coulomb forces. The goal is to choose the number of nodes in both  
groups such that the LR calculation takes the same amount of time as  
the SR calculation - otherwise one of the groups just does nothing for  
part of the time step.


Also it is advisable not to have too many nodes in the LR group - a  
quarter or a third is a good number as  pointed out by Justin and  
Berk. This you can achieve by shifting work between LR and SR part by  
enlarging the grid spacing and at the same time enlarging the Coulomb  
radius.


Carsten

On Nov 11, 2008, at 12:06 PM, vivek sharma wrote:


Hi Carsten,
I have also tried scaling gromacs for a number of nodes but was  
not able to optimize it beyond 20 processor..on 20 nodes i.e. 1  
processor per node..
I am not getting the point of optimizing PME for the number of  
nodes, is it like we can change the parameters for PME for MDS or  
using some other coulomb type for the same, please explain.

and suggest me the way to do it.

With Thanks,
Vivek

2008/11/10 Carsten Kutzner <[EMAIL PROTECTED]>
Hi,

most likely the Ethernet is the problem here. I compiled some  
numbers for the DPPC
benchmark in the paper "Speeding up parallel GROMACS on high-latency  
networks",

http://www3.interscience.wiley.com/journal/114205207/abstract?CRETRY=1&SRETRY=0
which are for version 3.3, but PME will behave similarly. If you did  
not already use
separate PME nodes, this is worth a try, since on Ethernet the  
performance will drastically
depend on the number of nodes involved in the FFT. I also have a  
tool which finds the
optimal PME settings for a given number of nodes, by varying the  
number of PME nodes

and the fourier grid settings. I can send it to you if you want.

Carsten


On Nov 9, 2008, at 10:30 PM, Yawar JQ wrote:

I was wondering if anyone could comment on these benchmark results  
for the d.dppc benchmark?


Nodes   Cutoff (ns/day) PME (ns/day)
4   1.331   0.797
8   2.564   1.497
16  4.5 1.92
32  8.308   0.575
64  13.50.275
128 20.093  -
192 21.6-

It seems to scale relatively well up to 32-64 nodes without PME.  
This seems slightly better than the benchmark results for Gromacs 3  
on www.gromacs.org.


Can someone comment on the magnitude of the performance hit and  
lack of scaling with PME is worrying me.


For the PME runs, I set rlist,rvdw,rouloumb=1.2 and the rest set to  
the defaults. I can try it with some other settings, larger spacing  
for the grid, but I'm not sure how much more that would help. Is  
there a more standardized system I should use for testing PME  
scaling?


This is with GNU compilers and parallelization with OpenMPI 1.2.  
I'm not sure what we're using for the FFTW The compute nodes are  
Dell m600 blades w/ 16GB of RAM and dual quad core Intel Xeon 3GHz  
processors. I believe it's all ethernet interconnects.


Thanks,
YQ
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before  
posting!

Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php



___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before  
posting!

Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php

___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before  
posting!

Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php


___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php

Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-11 Thread Martin Höfling
Am Dienstag 11 November 2008 12:06:06 schrieb vivek sharma:


> I have also tried scaling gromacs for a number of nodes but was not
> able to optimize it beyond 20 processor..on 20 nodes i.e. 1 processor per

As mentioned before, performance strongly depends on the type of interconnect 
you're using between your processes. Shared Memory, Ethernet, Infiniband, 
NumaLink, whatever...

I assume you're using ethernet (100/1000 MBit?), you can tune here to some 
extend as described in:

Kutzner, C.; Spoel, D. V. D.; Fechner, M.; Lindahl, E.; Schmitt, U. W.; Groot, 
B. L. D. & Grubmüller, H. Speeding up parallel GROMACS on high-latency 
networks Journal of Computational Chemistry, 2007 

...but be aware that principal limitations of ethernet remain. To come around 
this, you might consider to invest in the interconnect. If you can come out 
with <16 cores, shared memory nodes will give you the "biggest bang for the 
buck".

Best
  Martin
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php


Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-11 Thread Justin A. Lemkul



vivek sharma wrote:

Hi Carsten,
I have also tried scaling gromacs for a number of nodes but was not 
able to optimize it beyond 20 processor..on 20 nodes i.e. 1 processor 
per node..
I am not getting the point of optimizing PME for the number of nodes, is 
it like we can change the parameters for PME for MDS or using some other 
coulomb type for the same, please explain.


This is something I played with for a while; see the thread I started here:

http://www.gromacs.org/pipermail/gmx-users/2008-October/036856.html

I got some great advice there.  A big factor is the PME/PP balance, which grompp 
will estimate for you.  For simple rectangular boxes, the goal is to shoot for 
0.25 for the PME load (this is printed out by grompp).  In the thread above, 
Berk shared with me some tips on how to get this to happen.


Then you should be able to set -npme using mdrun to however many processors is 
appropriate.  I believe mdrun will try to guess, but I'm in the habit of 
specifying it myself just for my own satisfaction :)


-Justin


and suggest me the way to do it.

With Thanks,
Vivek

2008/11/10 Carsten Kutzner <[EMAIL PROTECTED] >

Hi, 


most likely the Ethernet is the problem here. I compiled some
numbers for the DPPC 
benchmark in the paper "Speeding up parallel GROMACS on high-latency
networks", 
http://www3.interscience.wiley.com/journal/114205207/abstract?CRETRY=1&SRETRY=0



which are for version 3.3, but PME will behave similarly. If you did
not already use 
separate PME nodes, this is worth a try, since on Ethernet the

performance will drastically
depend on the number of nodes involved in the FFT. I also have a
tool which finds the
optimal PME settings for a given number of nodes, by varying the
number of PME nodes
and the fourier grid settings. I can send it to you if you want.

Carsten


On Nov 9, 2008, at 10:30 PM, Yawar JQ wrote:


I was wondering if anyone could comment on these benchmark results
for the d.dppc benchmark?
 
Nodes 	Cutoff (ns/day) 	PME (ns/day)

4   1.331   0.797
8   2.564   1.497
16  4.5 1.92
32  8.308   0.575
64  13.50.275
128 20.093  -
192 21.6-

 
It seems to scale relatively well up to 32-64 nodes without PME.

This seems slightly better than the benchmark results for Gromacs
3 on www.gromacs.org . 
 
Can someone comment on the magnitude of the performance hit and

lack of scaling with PME is worrying me.
 
For the PME runs, I set rlist,rvdw,rouloumb=1.2 and the rest set

to the defaults. I can try it with some other settings, larger
spacing for the grid, but I'm not sure how much more that would
help. Is there a more standardized system I should use for testing
PME scaling?
 
This is with GNU compilers and parallelization with OpenMPI 1.2.

I'm not sure what we're using for the FFTW The compute nodes are
Dell m600 blades w/ 16GB of RAM and dual quad core Intel Xeon 3GHz
processors. I believe it's all ethernet interconnects.
 
Thanks,

YQ
___
gmx-users mailing listgmx-users@gromacs.org

http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before
posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php



___
gmx-users mailing listgmx-users@gromacs.org

http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before
posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php





___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]

Can't post? Read http://www.gromacs.org/mailing_lists/users.php


--


Justin A. Lemkul
Graduate Research Assistant
Department of Biochemistry
Virginia Tech
Blacksburg, VA
jalemkul[at]vt.edu | (540) 231-9080
http://www.bevanlab.biochem.vt.edu/Pages/Personal/justin

===

Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-11 Thread vivek sharma
Hi Carsten,
I have also tried scaling gromacs for a number of nodes but was not able
to optimize it beyond 20 processor..on 20 nodes i.e. 1 processor per node..
I am not getting the point of optimizing PME for the number of nodes, is it
like we can change the parameters for PME for MDS or using some other
coulomb type for the same, please explain.
and suggest me the way to do it.

With Thanks,
Vivek

2008/11/10 Carsten Kutzner <[EMAIL PROTECTED]>

> Hi,
> most likely the Ethernet is the problem here. I compiled some numbers for
> the DPPC
> benchmark in the paper "Speeding up parallel GROMACS on high-latency
> networks",
>
> http://www3.interscience.wiley.com/journal/114205207/abstract?CRETRY=1&SRETRY=0
> which are for version 3.3, but PME will behave similarly. If you did not
> already use
> separate PME nodes, this is worth a try, since on Ethernet the performance
> will drastically
> depend on the number of nodes involved in the FFT. I also have a tool which
> finds the
> optimal PME settings for a given number of nodes, by varying the number of
> PME nodes
> and the fourier grid settings. I can send it to you if you want.
>
> Carsten
>
>
> On Nov 9, 2008, at 10:30 PM, Yawar JQ wrote:
>
> I was wondering if anyone could comment on these benchmark results for the
> d.dppc benchmark?
>
> Nodes Cutoff (ns/day) PME (ns/day) 4 1.331 0.797 8 2.564 1.497 16 4.5
> 1.92 32 8.308 0.575 64 13.5 0.275 128 20.093 - 192 21.6 -
>
> It seems to scale relatively well up to 32-64 nodes without PME. This seems
> slightly better than the benchmark results for Gromacs 3 on
> www.gromacs.org.
>
> Can someone comment on the magnitude of the performance hit and lack of
> scaling with PME is worrying me.
>
> For the PME runs, I set rlist,rvdw,rouloumb=1.2 and the rest set to the
> defaults. I can try it with some other settings, larger spacing for the
> grid, but I'm not sure how much more that would help. Is there a more
> standardized system I should use for testing PME scaling?
>
> This is with GNU compilers and parallelization with OpenMPI 1.2. I'm not
> sure what we're using for the FFTW The compute nodes are Dell m600 blades w/
> 16GB of RAM and dual quad core Intel Xeon 3GHz processors. I believe it's
> all ethernet interconnects.
>
> Thanks,
> YQ
> ___
> gmx-users mailing listgmx-users@gromacs.org
> http://www.gromacs.org/mailman/listinfo/gmx-users
> Please search the archive at http://www.gromacs.org/search before posting!
> Please don't post (un)subscribe requests to the list. Use the
> www interface or send it to [EMAIL PROTECTED]
> Can't post? Read http://www.gromacs.org/mailing_lists/users.php
>
>
>
> ___
> gmx-users mailing listgmx-users@gromacs.org
> http://www.gromacs.org/mailman/listinfo/gmx-users
> Please search the archive at http://www.gromacs.org/search before posting!
> Please don't post (un)subscribe requests to the list. Use the
> www interface or send it to [EMAIL PROTECTED]
> Can't post? Read http://www.gromacs.org/mailing_lists/users.php
>
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php

RE: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-10 Thread Mike Hanby
The fftw used during compilation was FFTW 3.1.2 compiled using the GNU
compilers.

 

From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Yawar JQ
Sent: Sunday, November 09, 2008 3:31 PM
To: gmx-users@gromacs.org
Subject: [gmx-users] Gromacs 4 Scaling Benchmarks...

 

I was wondering if anyone could comment on these benchmark results for
the d.dppc benchmark?

 

Nodes

Cutoff (ns/day)

PME (ns/day)

4

1.331

0.797

8

2.564

1.497

16

4.5

1.92

32

8.308

0.575

64

13.5

0.275

128

20.093

-

192

21.6

-

 

It seems to scale relatively well up to 32-64 nodes without PME. This
seems slightly better than the benchmark results for Gromacs 3 on
www.gromacs.org  . 

 

Can someone comment on the magnitude of the performance hit and lack of
scaling with PME is worrying me. 

 

For the PME runs, I set rlist,rvdw,rouloumb=1.2 and the rest set to the
defaults. I can try it with some other settings, larger spacing for the
grid, but I'm not sure how much more that would help. Is there a more
standardized system I should use for testing PME scaling?

 

This is with GNU compilers and parallelization with OpenMPI 1.2. I'm not
sure what we're using for the FFTW The compute nodes are Dell m600
blades w/ 16GB of RAM and dual quad core Intel Xeon 3GHz processors. I
believe it's all ethernet interconnects.

 

Thanks,

YQ

___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php

Re: [gmx-users] Gromacs 4 Scaling Benchmarks...

2008-11-10 Thread Carsten Kutzner

Hi,

most likely the Ethernet is the problem here. I compiled some numbers  
for the DPPC
benchmark in the paper "Speeding up parallel GROMACS on high-latency  
networks",

http://www3.interscience.wiley.com/journal/114205207/abstract?CRETRY=1&SRETRY=0
which are for version 3.3, but PME will behave similarly. If you did  
not already use
separate PME nodes, this is worth a try, since on Ethernet the  
performance will drastically
depend on the number of nodes involved in the FFT. I also have a tool  
which finds the
optimal PME settings for a given number of nodes, by varying the  
number of PME nodes

and the fourier grid settings. I can send it to you if you want.

Carsten


On Nov 9, 2008, at 10:30 PM, Yawar JQ wrote:

I was wondering if anyone could comment on these benchmark results  
for the d.dppc benchmark?


Nodes   Cutoff (ns/day) PME (ns/day)
4   1.331   0.797
8   2.564   1.497
16  4.5 1.92
32  8.308   0.575
64  13.50.275
128 20.093  -
192 21.6-

It seems to scale relatively well up to 32-64 nodes without PME.  
This seems slightly better than the benchmark results for Gromacs 3  
on www.gromacs.org.


Can someone comment on the magnitude of the performance hit and lack  
of scaling with PME is worrying me.


For the PME runs, I set rlist,rvdw,rouloumb=1.2 and the rest set to  
the defaults. I can try it with some other settings, larger spacing  
for the grid, but I'm not sure how much more that would help. Is  
there a more standardized system I should use for testing PME scaling?


This is with GNU compilers and parallelization with OpenMPI 1.2. I'm  
not sure what we're using for the FFTW The compute nodes are Dell  
m600 blades w/ 16GB of RAM and dual quad core Intel Xeon 3GHz  
processors. I believe it's all ethernet interconnects.


Thanks,
YQ
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before  
posting!

Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php


___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php