RE: [gmx-users] Strange assignment of atoms to processors with pd

2009-05-28 Thread Berk Hess

Well, David is the person who should fix this.
You can submit a bugzilla or bug him directly.

Berk

> Date: Thu, 28 May 2009 10:41:12 +0200
> From: er...@xray.bmc.uu.se
> To: gmx-users@gromacs.org
> Subject: Re: [gmx-users] Strange assignment of atoms to processors with pd
> 
> Berk Hess skrev:
> > Hi,
> >
> > This is plain 4.0 code is presume?
> > This problem should be fixed then.
> Should I commit a buzilla?
> >
> > But I now also made vacuum without cut-off working with domain 
> > decomposition in CVS head.
> > Compared to a not-unbalanced PD (for instance only a protein, no 
> > water) DD is slightly slower.
> > But DD will be faster than a badly balanced PD system.
> >
> > Berk
> >
> > > Date: Wed, 27 May 2009 11:04:49 +0200
> > > From: sp...@xray.bmc.uu.se
> > > To: gmx-users@gromacs.org
> > > Subject: Re: [gmx-users] Strange assignment of atoms to processors 
> > with pd
> > >
> > > Erik Marklund wrote:
> > > > David van der Spoel skrev:
> > > >> Erik Marklund wrote:
> > > >>> I should add that this problem only seem to arise when the 
> > analyte is
> > > >>> covered with a thin sheet of water. When simulating a dry analyte I
> > > >>> get good scaling. In the latter case the charges, and therefore the
> > > >>> topology, is slightly different.
> > > >> How about vsites? Did you happen to turn them off as well in the
> > > >> vacuum case?
> > > > Turned off in all cases. The VSites mentioned in the log file is the
> > > > 4:th particle on the tip4p-water molecules.
> > > OK. Did you try a one step run with -debug?
> > > It may give more info on the partitioning.
> > >
> > > >>>
> > > >>> /Erik
> > > >>>
> > > >>> Erik Marklund skrev:
> > > >>>> Hi,
> > > >>>>
> > > >>>> I'm simulating non-periodic systems in vacuo, using constrained
> > > >>>> h-bonds and particle decomposition. For some of my simulations the
> > > >>>> cpu-usage seem far from optimal. The first cpu gets no atoms, 
> > while
> > > >>>> the second one gets plenty and the remaining cpus get less than I
> > > >>>> expected. Is this a bug?
> > > >>>>
> > > >>>>
> > > >>>> An excerpt from the log file:
> > > >>>>
> > > >>>> There are: 2911 Atoms
> > > >>>> There are: 317 VSites
> > > >>>> splitting topology...
> > > >>>> There are 999 charge group borders and 318 shake borders
> > > >>>> There are 318 total borders
> > > >>>> Division over nodes in atoms:
> > > >>>> 0 1960 212 212 212 212 212 208
> > > >>>> Walking down the molecule graph to make constraint-blocks
> > > >>>> CPU= 0, lastcg= -1, targetcg= 499, myshift= 1
> > > >>>> CPU= 1, lastcg= 681, targetcg= 182, myshift= 0
> > > >>>> CPU= 2, lastcg= 734, targetcg= 235, myshift= 7
> > > >>>> CPU= 3, lastcg= 787, targetcg= 288, myshift= 6
> > > >>>> CPU= 4, lastcg= 840, targetcg= 341, myshift= 5
> > > >>>> CPU= 5, lastcg= 893, targetcg= 394, myshift= 4
> > > >>>> CPU= 6, lastcg= 946, targetcg= 447, myshift= 3
> > > >>>> CPU= 7, lastcg= 998, targetcg= 499, myshift= 2
> > > >>>> pd->shift = 7, pd->bshift= 0
> > > >>>> Division of bonded forces over processors
> > > >>>> CPU 0 1 2 3 4 5 6 7
> > > >>>> Workload division
> > > >>>> nnodes: 8
> > > >>>> pd->shift: 7
> > > >>>> pd->bshift: 0
> > > >>>> Nodeid atom0 #atom cg0 #cg
> > > >>>> 0 0 0 0 0
> > > >>>> 1 0 1960 0 682
> > > >>>> 2 1960 212 682 53
> > > >>>> 3 2172 212 735 53
> > > >>>> 4 2384 212 788 53
> > > >>>> 5 2596 212 841 53
> > > >>>> 6 2808 212 894 53
> > > >>>> 7 3020 208 947 52
> > > >>>>
> > > >>>> …
> > > >>>> Total Scaling: 18% of max performance
> > > >>>>
> > > >>>
> > > >>>

Re: [gmx-users] Strange assignment of atoms to processors with pd

2009-05-28 Thread Erik Marklund

Berk Hess skrev:

Hi,

This is plain 4.0 code is presume?
This problem should be fixed then.

Should I commit a buzilla?


But I now also made vacuum without cut-off working with domain 
decomposition in CVS head.
Compared to a not-unbalanced PD (for instance only a protein, no 
water) DD is slightly slower.

But DD will be faster than a badly balanced PD system.

Berk

> Date: Wed, 27 May 2009 11:04:49 +0200
> From: sp...@xray.bmc.uu.se
> To: gmx-users@gromacs.org
> Subject: Re: [gmx-users] Strange assignment of atoms to processors 
with pd

>
> Erik Marklund wrote:
> > David van der Spoel skrev:
> >> Erik Marklund wrote:
> >>> I should add that this problem only seem to arise when the 
analyte is

> >>> covered with a thin sheet of water. When simulating a dry analyte I
> >>> get good scaling. In the latter case the charges, and therefore the
> >>> topology, is slightly different.
> >> How about vsites? Did you happen to turn them off as well in the
> >> vacuum case?
> > Turned off in all cases. The VSites mentioned in the log file is the
> > 4:th particle on the tip4p-water molecules.
> OK. Did you try a one step run with -debug?
> It may give more info on the partitioning.
>
> >>>
> >>> /Erik
> >>>
> >>> Erik Marklund skrev:
> >>>> Hi,
> >>>>
> >>>> I'm simulating non-periodic systems in vacuo, using constrained
> >>>> h-bonds and particle decomposition. For some of my simulations the
> >>>> cpu-usage seem far from optimal. The first cpu gets no atoms, 
while

> >>>> the second one gets plenty and the remaining cpus get less than I
> >>>> expected. Is this a bug?
> >>>>
> >>>>
> >>>> An excerpt from the log file:
> >>>>
> >>>> There are: 2911 Atoms
> >>>> There are: 317 VSites
> >>>> splitting topology...
> >>>> There are 999 charge group borders and 318 shake borders
> >>>> There are 318 total borders
> >>>> Division over nodes in atoms:
> >>>> 0 1960 212 212 212 212 212 208
> >>>> Walking down the molecule graph to make constraint-blocks
> >>>> CPU= 0, lastcg= -1, targetcg= 499, myshift= 1
> >>>> CPU= 1, lastcg= 681, targetcg= 182, myshift= 0
> >>>> CPU= 2, lastcg= 734, targetcg= 235, myshift= 7
> >>>> CPU= 3, lastcg= 787, targetcg= 288, myshift= 6
> >>>> CPU= 4, lastcg= 840, targetcg= 341, myshift= 5
> >>>> CPU= 5, lastcg= 893, targetcg= 394, myshift= 4
> >>>> CPU= 6, lastcg= 946, targetcg= 447, myshift= 3
> >>>> CPU= 7, lastcg= 998, targetcg= 499, myshift= 2
> >>>> pd->shift = 7, pd->bshift= 0
> >>>> Division of bonded forces over processors
> >>>> CPU 0 1 2 3 4 5 6 7
> >>>> Workload division
> >>>> nnodes: 8
> >>>> pd->shift: 7
> >>>> pd->bshift: 0
> >>>> Nodeid atom0 #atom cg0 #cg
> >>>> 0 0 0 0 0
> >>>> 1 0 1960 0 682
> >>>> 2 1960 212 682 53
> >>>> 3 2172 212 735 53
> >>>> 4 2384 212 788 53
> >>>> 5 2596 212 841 53
> >>>> 6 2808 212 894 53
> >>>> 7 3020 208 947 52
> >>>>
> >>>> …
> >>>> Total Scaling: 18% of max performance
> >>>>
> >>>
> >>>
> >>
> >>
> >
> >
>
>
> --
> David van der Spoel, Ph.D., Professor of Biology
> Molec. Biophys. group, Dept. of Cell & Molec. Biol., Uppsala University.
> Box 596, 75124 Uppsala, Sweden. Phone: +46184714205. Fax: +4618511755.
> sp...@xray.bmc.uu.se sp...@gromacs.org http://folding.bmc.uu.se
> ___
> gmx-users mailing list gmx-users@gromacs.org
> http://www.gromacs.org/mailman/listinfo/gmx-users
> Please search the archive at http://www.gromacs.org/search before 
posting!

> Please don't post (un)subscribe requests to the list. Use the
> www interface or send it to gmx-users-requ...@gromacs.org.
> Can't post? Read http://www.gromacs.org/mailing_lists/users.php


What can you do with the new Windows Live? Find out 
<http://www.microsoft.com/windows/windowslive/default.aspx>



___
gmx-users mailing listgmx-users@gromacs.org
http

Re: [gmx-users] Strange assignment of atoms to processors with pd

2009-05-27 Thread Erik Marklund
This was done with verision 4.0.4, taken straight from the website. I've 
already simulated a great deal with this version and I'd hate to swap 
versions in the middle of the project.


As for the -debug runs, I did it for the dry analyte and for one with a 
6 Å water layer. It seems that the protein is indeed divided over the 
processors if no other molecules are present. When there are other 
molecules, such as water, in the system then I suspect that the 
molecules are divided over the processors. This doesn't explain why the 
first processor got no atoms when I ran the simulations on 8 cpus though.


 without water ===-,
There are: 1960 Atoms
splitting topology...
There are 698 charge group borders and 1 shake borders
There are 698 total borders
Division over nodes in atoms:
492 489 488 491
Walking down the molecule graph to make constraint-blocks
CPU=  0, lastcg=  171, targetcg=  520, myshift=2
CPU=  1, lastcg=  354, targetcg=6, myshift=3
CPU=  2, lastcg=  521, targetcg=  172, myshift=2
CPU=  3, lastcg=  697, targetcg=  348, myshift=2
pd->shift =   3, pd->bshift=  0
Division of bonded forces over processors
CPU  0 1 2 3
BONDS  259   263   250   253
ANGLES 894   867   901   870
PDIHS  110   11787   107
RBDIHS1002   967  1015   960
LJ14  1289  1260  1318  1234
CONSTR 242   232   244   241
Workload division
nnodes:   4
pd->shift:3
pd->bshift:   0
Nodeid   atom0   #atom cg0   #cg
0   0 492   0   172
1 492 489 172   183
2 981 488 355   167
31469 491 522   176

===-'

== with a bit of water ===-,
There are: 5086 Atoms
There are: 1042 VSites
splitting topology...
There are 1724 charge group borders and 1043 shake borders
There are 1043 total borders
Division over nodes in atoms:
   1960138813921388
Walking down the molecule graph to make constraint-blocks
CPU=  0, lastcg=  681, targetcg= 1543, myshift=3
CPU=  1, lastcg= 1028, targetcg=  166, myshift=3
CPU=  2, lastcg= 1376, targetcg=  514, myshift=2
CPU=  3, lastcg= 1723, targetcg=  862, myshift=2
pd->shift =   3, pd->bshift=  0
Division of bonded forces over processors
CPU  0 1 2 3
BONDS 1025 0 0 0
ANGLES3547 0 0 0
PDIHS  426 0 0 0
RBDIHS3950 0 0 0
LJ14  5106 0 0 0
CONSTR 959 0 0 0
Workload division
nnodes:   4
pd->shift:3
pd->bshift:   0
Nodeid   atom0   #atom cg0   #cg
0   01960   0   682
119601388 682   347
2334813921029   348
3474013881377   347




Berk Hess skrev:

Hi,

This is plain 4.0 code is presume?
This problem should be fixed then.

But I now also made vacuum without cut-off working with domain 
decomposition in CVS head.
Compared to a not-unbalanced PD (for instance only a protein, no 
water) DD is slightly slower.

But DD will be faster than a badly balanced PD system.

Berk

> Date: Wed, 27 May 2009 11:04:49 +0200
> From: sp...@xray.bmc.uu.se
> To: gmx-users@gromacs.org
> Subject: Re: [gmx-users] Strange assignment of atoms to processors 
with pd

>
> Erik Marklund wrote:
> > David van der Spoel skrev:
> >> Erik Marklund wrote:
> >>> I should add that this problem only seem to arise when the 
analyte is

> >>> covered with a thin sheet of water. When simulating a dry analyte I
> >>> get good scaling. In the latter case the charges, and therefore the
> >>> topology, is slightly different.
> >> How about vsites? Did you happen to turn them off as well in the
> >> vacuum case?
> > Turned off in all cases. The VSites mentioned in the log file is the
> > 4:th particle on the tip4p-water molecules.
> OK. Did you try a one step run with -debug?
> It may give more info on the partitioning.
>
> >>>
> >>> /Erik
> >>>
> >>> Erik Marklund skrev:
> >>>> Hi,
> >>>>
> >>>> I'm simulating non-periodic systems in vacuo, using constrained
> >>>> h-bonds and particle decomposition. For some of my simulations the
> >>>> cpu-usage seem far from optimal. The first cpu gets no atoms, 
while

> >>>> the second one gets plenty and the remaining cpus get less than I
> >>>> expected. Is this a bug?
> >>>>
> >>>>
> >>>> An excerpt from the log file:
> >>>>
> >>>> There are: 2911 Atoms
> >>>> There are: 317 VSites
>

RE: [gmx-users] Strange assignment of atoms to processors with pd

2009-05-27 Thread Berk Hess

Hi,

This is plain 4.0 code is presume?
This problem should be fixed then.

But I now also made vacuum without cut-off working with domain decomposition in 
CVS head.
Compared to a not-unbalanced PD (for instance only a protein, no water) DD is 
slightly slower.
But DD will be faster than a badly balanced PD system.

Berk

> Date: Wed, 27 May 2009 11:04:49 +0200
> From: sp...@xray.bmc.uu.se
> To: gmx-users@gromacs.org
> Subject: Re: [gmx-users] Strange assignment of atoms to processors with pd
> 
> Erik Marklund wrote:
> > David van der Spoel skrev:
> >> Erik Marklund wrote:
> >>> I should add that this problem only seem to arise when the analyte is 
> >>> covered with a thin sheet of water. When simulating a dry analyte I 
> >>> get good scaling. In the latter case the charges, and therefore the 
> >>> topology, is slightly different.
> >> How about vsites? Did you happen to turn them off as well in the 
> >> vacuum case?
> > Turned off in all cases. The VSites mentioned in the log file is the 
> > 4:th particle on the tip4p-water molecules.
> OK. Did you try a one step run with -debug?
> It may give more info on the partitioning.
> 
> >>>
> >>> /Erik
> >>>
> >>> Erik Marklund skrev:
> >>>> Hi,
> >>>>
> >>>> I'm simulating non-periodic systems in vacuo, using constrained 
> >>>> h-bonds and particle decomposition. For some of my simulations the 
> >>>> cpu-usage seem far from optimal. The first cpu gets no atoms, while 
> >>>> the second one gets plenty and the remaining cpus get less than I 
> >>>> expected. Is this a bug?
> >>>>
> >>>>
> >>>> An excerpt from the log file:
> >>>>
> >>>> There are: 2911 Atoms
> >>>> There are: 317 VSites
> >>>> splitting topology...
> >>>> There are 999 charge group borders and 318 shake borders
> >>>> There are 318 total borders
> >>>> Division over nodes in atoms:
> >>>> 0 1960 212 212 212 212 212 208
> >>>> Walking down the molecule graph to make constraint-blocks
> >>>> CPU= 0, lastcg= -1, targetcg= 499, myshift= 1
> >>>> CPU= 1, lastcg= 681, targetcg= 182, myshift= 0
> >>>> CPU= 2, lastcg= 734, targetcg= 235, myshift= 7
> >>>> CPU= 3, lastcg= 787, targetcg= 288, myshift= 6
> >>>> CPU= 4, lastcg= 840, targetcg= 341, myshift= 5
> >>>> CPU= 5, lastcg= 893, targetcg= 394, myshift= 4
> >>>> CPU= 6, lastcg= 946, targetcg= 447, myshift= 3
> >>>> CPU= 7, lastcg= 998, targetcg= 499, myshift= 2
> >>>> pd->shift = 7, pd->bshift= 0
> >>>> Division of bonded forces over processors
> >>>> CPU 0 1 2 3 4 5 6 7
> >>>> Workload division
> >>>> nnodes: 8
> >>>> pd->shift: 7
> >>>> pd->bshift: 0
> >>>> Nodeid atom0 #atom cg0 #cg
> >>>> 0 0 0 0 0
> >>>> 1 0 1960 0 682
> >>>> 2 1960 212 682 53
> >>>> 3 2172 212 735 53
> >>>> 4 2384 212 788 53
> >>>> 5 2596 212 841 53
> >>>> 6 2808 212 894 53
> >>>> 7 3020 208 947 52
> >>>>
> >>>> …
> >>>> Total Scaling: 18% of max performance
> >>>>
> >>>
> >>>
> >>
> >>
> > 
> > 
> 
> 
> -- 
> David van der Spoel, Ph.D., Professor of Biology
> Molec. Biophys. group, Dept. of Cell & Molec. Biol., Uppsala University.
> Box 596, 75124 Uppsala, Sweden. Phone:+46184714205. Fax: +4618511755.
> sp...@xray.bmc.uu.se  sp...@gromacs.org   http://folding.bmc.uu.se
> ___
> gmx-users mailing listgmx-users@gromacs.org
> http://www.gromacs.org/mailman/listinfo/gmx-users
> Please search the archive at http://www.gromacs.org/search before posting!
> Please don't post (un)subscribe requests to the list. Use the 
> www interface or send it to gmx-users-requ...@gromacs.org.
> Can't post? Read http://www.gromacs.org/mailing_lists/users.php

_
What can you do with the new Windows Live? Find out
http://www.microsoft.com/windows/windowslive/default.aspx___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to gmx-users-requ...@gromacs.org.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php

Re: [gmx-users] Strange assignment of atoms to processors with pd

2009-05-27 Thread David van der Spoel

Erik Marklund wrote:

David van der Spoel skrev:

Erik Marklund wrote:
I should add that this problem only seem to arise when the analyte is 
covered with a thin sheet of water. When simulating a dry analyte I 
get good scaling. In the latter case the charges, and therefore the 
topology, is slightly different.
How about vsites? Did you happen to turn them off as well in the 
vacuum case?
Turned off in all cases. The VSites mentioned in the log file is the 
4:th particle on the tip4p-water molecules.

OK. Did you try a one step run with -debug?
It may give more info on the partitioning.



/Erik

Erik Marklund skrev:

Hi,

I'm simulating non-periodic systems in vacuo, using constrained 
h-bonds and particle decomposition. For some of my simulations the 
cpu-usage seem far from optimal. The first cpu gets no atoms, while 
the second one gets plenty and the remaining cpus get less than I 
expected. Is this a bug?



An excerpt from the log file:

There are: 2911 Atoms
There are: 317 VSites
splitting topology...
There are 999 charge group borders and 318 shake borders
There are 318 total borders
Division over nodes in atoms:
0 1960 212 212 212 212 212 208
Walking down the molecule graph to make constraint-blocks
CPU= 0, lastcg= -1, targetcg= 499, myshift= 1
CPU= 1, lastcg= 681, targetcg= 182, myshift= 0
CPU= 2, lastcg= 734, targetcg= 235, myshift= 7
CPU= 3, lastcg= 787, targetcg= 288, myshift= 6
CPU= 4, lastcg= 840, targetcg= 341, myshift= 5
CPU= 5, lastcg= 893, targetcg= 394, myshift= 4
CPU= 6, lastcg= 946, targetcg= 447, myshift= 3
CPU= 7, lastcg= 998, targetcg= 499, myshift= 2
pd->shift = 7, pd->bshift= 0
Division of bonded forces over processors
CPU 0 1 2 3 4 5 6 7
Workload division
nnodes: 8
pd->shift: 7
pd->bshift: 0
Nodeid atom0 #atom cg0 #cg
0 0 0 0 0
1 0 1960 0 682
2 1960 212 682 53
3 2172 212 735 53
4 2384 212 788 53
5 2596 212 841 53
6 2808 212 894 53
7 3020 208 947 52

…
Total Scaling: 18% of max performance













--
David van der Spoel, Ph.D., Professor of Biology
Molec. Biophys. group, Dept. of Cell & Molec. Biol., Uppsala University.
Box 596, 75124 Uppsala, Sweden. Phone:  +46184714205. Fax: +4618511755.
sp...@xray.bmc.uu.sesp...@gromacs.org   http://folding.bmc.uu.se
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to gmx-users-requ...@gromacs.org.

Can't post? Read http://www.gromacs.org/mailing_lists/users.php


Re: [gmx-users] Strange assignment of atoms to processors with pd

2009-05-27 Thread Erik Marklund

David van der Spoel skrev:

Erik Marklund wrote:
I should add that this problem only seem to arise when the analyte is 
covered with a thin sheet of water. When simulating a dry analyte I 
get good scaling. In the latter case the charges, and therefore the 
topology, is slightly different.
How about vsites? Did you happen to turn them off as well in the 
vacuum case?
Turned off in all cases. The VSites mentioned in the log file is the 
4:th particle on the tip4p-water molecules.


/Erik

Erik Marklund skrev:

Hi,

I'm simulating non-periodic systems in vacuo, using constrained 
h-bonds and particle decomposition. For some of my simulations the 
cpu-usage seem far from optimal. The first cpu gets no atoms, while 
the second one gets plenty and the remaining cpus get less than I 
expected. Is this a bug?



An excerpt from the log file:

There are: 2911 Atoms
There are: 317 VSites
splitting topology...
There are 999 charge group borders and 318 shake borders
There are 318 total borders
Division over nodes in atoms:
0 1960 212 212 212 212 212 208
Walking down the molecule graph to make constraint-blocks
CPU= 0, lastcg= -1, targetcg= 499, myshift= 1
CPU= 1, lastcg= 681, targetcg= 182, myshift= 0
CPU= 2, lastcg= 734, targetcg= 235, myshift= 7
CPU= 3, lastcg= 787, targetcg= 288, myshift= 6
CPU= 4, lastcg= 840, targetcg= 341, myshift= 5
CPU= 5, lastcg= 893, targetcg= 394, myshift= 4
CPU= 6, lastcg= 946, targetcg= 447, myshift= 3
CPU= 7, lastcg= 998, targetcg= 499, myshift= 2
pd->shift = 7, pd->bshift= 0
Division of bonded forces over processors
CPU 0 1 2 3 4 5 6 7
Workload division
nnodes: 8
pd->shift: 7
pd->bshift: 0
Nodeid atom0 #atom cg0 #cg
0 0 0 0 0
1 0 1960 0 682
2 1960 212 682 53
3 2172 212 735 53
4 2384 212 788 53
5 2596 212 841 53
6 2808 212 894 53
7 3020 208 947 52

…
Total Scaling: 18% of max performance










--
---
Erik Marklund, PhD student
Laboratory of Molecular Biophysics,
Dept. of Cell and Molecular Biology, Uppsala University.
Husargatan 3, Box 596,75124 Uppsala, Sweden
phone:+46 18 471 4537fax: +46 18 511 755
er...@xray.bmc.uu.sehttp://xray.bmc.uu.se/molbiophys

___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to gmx-users-requ...@gromacs.org.

Can't post? Read http://www.gromacs.org/mailing_lists/users.php


Re: [gmx-users] Strange assignment of atoms to processors with pd

2009-05-27 Thread David van der Spoel

Erik Marklund wrote:
I should add that this problem only seem to arise when the analyte is 
covered with a thin sheet of water. When simulating a dry analyte I get 
good scaling. In the latter case the charges, and therefore the 
topology, is slightly different.
How about vsites? Did you happen to turn them off as well in the vacuum 
case?


/Erik

Erik Marklund skrev:

Hi,

I'm simulating non-periodic systems in vacuo, using constrained 
h-bonds and particle decomposition. For some of my simulations the 
cpu-usage seem far from optimal. The first cpu gets no atoms, while 
the second one gets plenty and the remaining cpus get less than I 
expected. Is this a bug?



An excerpt from the log file:

There are: 2911 Atoms
There are: 317 VSites
splitting topology...
There are 999 charge group borders and 318 shake borders
There are 318 total borders
Division over nodes in atoms:
0 1960 212 212 212 212 212 208
Walking down the molecule graph to make constraint-blocks
CPU= 0, lastcg= -1, targetcg= 499, myshift= 1
CPU= 1, lastcg= 681, targetcg= 182, myshift= 0
CPU= 2, lastcg= 734, targetcg= 235, myshift= 7
CPU= 3, lastcg= 787, targetcg= 288, myshift= 6
CPU= 4, lastcg= 840, targetcg= 341, myshift= 5
CPU= 5, lastcg= 893, targetcg= 394, myshift= 4
CPU= 6, lastcg= 946, targetcg= 447, myshift= 3
CPU= 7, lastcg= 998, targetcg= 499, myshift= 2
pd->shift = 7, pd->bshift= 0
Division of bonded forces over processors
CPU 0 1 2 3 4 5 6 7
Workload division
nnodes: 8
pd->shift: 7
pd->bshift: 0
Nodeid atom0 #atom cg0 #cg
0 0 0 0 0
1 0 1960 0 682
2 1960 212 682 53
3 2172 212 735 53
4 2384 212 788 53
5 2596 212 841 53
6 2808 212 894 53
7 3020 208 947 52

…
Total Scaling: 18% of max performance







--
David van der Spoel, Ph.D., Professor of Biology
Molec. Biophys. group, Dept. of Cell & Molec. Biol., Uppsala University.
Box 596, 75124 Uppsala, Sweden. Phone:  +46184714205. Fax: +4618511755.
sp...@xray.bmc.uu.sesp...@gromacs.org   http://folding.bmc.uu.se
___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to gmx-users-requ...@gromacs.org.

Can't post? Read http://www.gromacs.org/mailing_lists/users.php


Re: [gmx-users] Strange assignment of atoms to processors with pd

2009-05-27 Thread Erik Marklund
I should add that this problem only seem to arise when the analyte is 
covered with a thin sheet of water. When simulating a dry analyte I get 
good scaling. In the latter case the charges, and therefore the 
topology, is slightly different.


/Erik

Erik Marklund skrev:

Hi,

I'm simulating non-periodic systems in vacuo, using constrained 
h-bonds and particle decomposition. For some of my simulations the 
cpu-usage seem far from optimal. The first cpu gets no atoms, while 
the second one gets plenty and the remaining cpus get less than I 
expected. Is this a bug?



An excerpt from the log file:

There are: 2911 Atoms
There are: 317 VSites
splitting topology...
There are 999 charge group borders and 318 shake borders
There are 318 total borders
Division over nodes in atoms:
0 1960 212 212 212 212 212 208
Walking down the molecule graph to make constraint-blocks
CPU= 0, lastcg= -1, targetcg= 499, myshift= 1
CPU= 1, lastcg= 681, targetcg= 182, myshift= 0
CPU= 2, lastcg= 734, targetcg= 235, myshift= 7
CPU= 3, lastcg= 787, targetcg= 288, myshift= 6
CPU= 4, lastcg= 840, targetcg= 341, myshift= 5
CPU= 5, lastcg= 893, targetcg= 394, myshift= 4
CPU= 6, lastcg= 946, targetcg= 447, myshift= 3
CPU= 7, lastcg= 998, targetcg= 499, myshift= 2
pd->shift = 7, pd->bshift= 0
Division of bonded forces over processors
CPU 0 1 2 3 4 5 6 7
Workload division
nnodes: 8
pd->shift: 7
pd->bshift: 0
Nodeid atom0 #atom cg0 #cg
0 0 0 0 0
1 0 1960 0 682
2 1960 212 682 53
3 2172 212 735 53
4 2384 212 788 53
5 2596 212 841 53
6 2808 212 894 53
7 3020 208 947 52

…
Total Scaling: 18% of max performance




--
---
Erik Marklund, PhD student
Laboratory of Molecular Biophysics,
Dept. of Cell and Molecular Biology, Uppsala University.
Husargatan 3, Box 596,75124 Uppsala, Sweden
phone:+46 18 471 4537fax: +46 18 511 755
er...@xray.bmc.uu.sehttp://xray.bmc.uu.se/molbiophys

___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to gmx-users-requ...@gromacs.org.

Can't post? Read http://www.gromacs.org/mailing_lists/users.php


[gmx-users] Strange assignment of atoms to processors with pd

2009-05-27 Thread Erik Marklund

Hi,

I'm simulating non-periodic systems in vacuo, using constrained h-bonds 
and particle decomposition. For some of my simulations the cpu-usage 
seem far from optimal. The first cpu gets no atoms, while the second one 
gets plenty and the remaining cpus get less than I expected. Is this a bug?



An excerpt from the log file:

There are: 2911 Atoms
There are: 317 VSites
splitting topology...
There are 999 charge group borders and 318 shake borders
There are 318 total borders
Division over nodes in atoms:
0 1960 212 212 212 212 212 208
Walking down the molecule graph to make constraint-blocks
CPU= 0, lastcg= -1, targetcg= 499, myshift= 1
CPU= 1, lastcg= 681, targetcg= 182, myshift= 0
CPU= 2, lastcg= 734, targetcg= 235, myshift= 7
CPU= 3, lastcg= 787, targetcg= 288, myshift= 6
CPU= 4, lastcg= 840, targetcg= 341, myshift= 5
CPU= 5, lastcg= 893, targetcg= 394, myshift= 4
CPU= 6, lastcg= 946, targetcg= 447, myshift= 3
CPU= 7, lastcg= 998, targetcg= 499, myshift= 2
pd->shift = 7, pd->bshift= 0
Division of bonded forces over processors
CPU 0 1 2 3 4 5 6 7
Workload division
nnodes: 8
pd->shift: 7
pd->bshift: 0
Nodeid atom0 #atom cg0 #cg
0 0 0 0 0
1 0 1960 0 682
2 1960 212 682 53
3 2172 212 735 53
4 2384 212 788 53
5 2596 212 841 53
6 2808 212 894 53
7 3020 208 947 52

…
Total Scaling: 18% of max performance

--
---
Erik Marklund, PhD student
Laboratory of Molecular Biophysics,
Dept. of Cell and Molecular Biology, Uppsala University.
Husargatan 3, Box 596,75124 Uppsala, Sweden
phone:+46 18 471 4537fax: +46 18 511 755
er...@xray.bmc.uu.sehttp://xray.bmc.uu.se/molbiophys

___
gmx-users mailing listgmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to gmx-users-requ...@gromacs.org.

Can't post? Read http://www.gromacs.org/mailing_lists/users.php