RE: [gmx-users] Strange assignment of atoms to processors with pd
Well, David is the person who should fix this. You can submit a bugzilla or bug him directly. Berk > Date: Thu, 28 May 2009 10:41:12 +0200 > From: er...@xray.bmc.uu.se > To: gmx-users@gromacs.org > Subject: Re: [gmx-users] Strange assignment of atoms to processors with pd > > Berk Hess skrev: > > Hi, > > > > This is plain 4.0 code is presume? > > This problem should be fixed then. > Should I commit a buzilla? > > > > But I now also made vacuum without cut-off working with domain > > decomposition in CVS head. > > Compared to a not-unbalanced PD (for instance only a protein, no > > water) DD is slightly slower. > > But DD will be faster than a badly balanced PD system. > > > > Berk > > > > > Date: Wed, 27 May 2009 11:04:49 +0200 > > > From: sp...@xray.bmc.uu.se > > > To: gmx-users@gromacs.org > > > Subject: Re: [gmx-users] Strange assignment of atoms to processors > > with pd > > > > > > Erik Marklund wrote: > > > > David van der Spoel skrev: > > > >> Erik Marklund wrote: > > > >>> I should add that this problem only seem to arise when the > > analyte is > > > >>> covered with a thin sheet of water. When simulating a dry analyte I > > > >>> get good scaling. In the latter case the charges, and therefore the > > > >>> topology, is slightly different. > > > >> How about vsites? Did you happen to turn them off as well in the > > > >> vacuum case? > > > > Turned off in all cases. The VSites mentioned in the log file is the > > > > 4:th particle on the tip4p-water molecules. > > > OK. Did you try a one step run with -debug? > > > It may give more info on the partitioning. > > > > > > >>> > > > >>> /Erik > > > >>> > > > >>> Erik Marklund skrev: > > > >>>> Hi, > > > >>>> > > > >>>> I'm simulating non-periodic systems in vacuo, using constrained > > > >>>> h-bonds and particle decomposition. For some of my simulations the > > > >>>> cpu-usage seem far from optimal. The first cpu gets no atoms, > > while > > > >>>> the second one gets plenty and the remaining cpus get less than I > > > >>>> expected. Is this a bug? > > > >>>> > > > >>>> > > > >>>> An excerpt from the log file: > > > >>>> > > > >>>> There are: 2911 Atoms > > > >>>> There are: 317 VSites > > > >>>> splitting topology... > > > >>>> There are 999 charge group borders and 318 shake borders > > > >>>> There are 318 total borders > > > >>>> Division over nodes in atoms: > > > >>>> 0 1960 212 212 212 212 212 208 > > > >>>> Walking down the molecule graph to make constraint-blocks > > > >>>> CPU= 0, lastcg= -1, targetcg= 499, myshift= 1 > > > >>>> CPU= 1, lastcg= 681, targetcg= 182, myshift= 0 > > > >>>> CPU= 2, lastcg= 734, targetcg= 235, myshift= 7 > > > >>>> CPU= 3, lastcg= 787, targetcg= 288, myshift= 6 > > > >>>> CPU= 4, lastcg= 840, targetcg= 341, myshift= 5 > > > >>>> CPU= 5, lastcg= 893, targetcg= 394, myshift= 4 > > > >>>> CPU= 6, lastcg= 946, targetcg= 447, myshift= 3 > > > >>>> CPU= 7, lastcg= 998, targetcg= 499, myshift= 2 > > > >>>> pd->shift = 7, pd->bshift= 0 > > > >>>> Division of bonded forces over processors > > > >>>> CPU 0 1 2 3 4 5 6 7 > > > >>>> Workload division > > > >>>> nnodes: 8 > > > >>>> pd->shift: 7 > > > >>>> pd->bshift: 0 > > > >>>> Nodeid atom0 #atom cg0 #cg > > > >>>> 0 0 0 0 0 > > > >>>> 1 0 1960 0 682 > > > >>>> 2 1960 212 682 53 > > > >>>> 3 2172 212 735 53 > > > >>>> 4 2384 212 788 53 > > > >>>> 5 2596 212 841 53 > > > >>>> 6 2808 212 894 53 > > > >>>> 7 3020 208 947 52 > > > >>>> > > > >>>> … > > > >>>> Total Scaling: 18% of max performance > > > >>>> > > > >>> > > > >>>
Re: [gmx-users] Strange assignment of atoms to processors with pd
Berk Hess skrev: Hi, This is plain 4.0 code is presume? This problem should be fixed then. Should I commit a buzilla? But I now also made vacuum without cut-off working with domain decomposition in CVS head. Compared to a not-unbalanced PD (for instance only a protein, no water) DD is slightly slower. But DD will be faster than a badly balanced PD system. Berk > Date: Wed, 27 May 2009 11:04:49 +0200 > From: sp...@xray.bmc.uu.se > To: gmx-users@gromacs.org > Subject: Re: [gmx-users] Strange assignment of atoms to processors with pd > > Erik Marklund wrote: > > David van der Spoel skrev: > >> Erik Marklund wrote: > >>> I should add that this problem only seem to arise when the analyte is > >>> covered with a thin sheet of water. When simulating a dry analyte I > >>> get good scaling. In the latter case the charges, and therefore the > >>> topology, is slightly different. > >> How about vsites? Did you happen to turn them off as well in the > >> vacuum case? > > Turned off in all cases. The VSites mentioned in the log file is the > > 4:th particle on the tip4p-water molecules. > OK. Did you try a one step run with -debug? > It may give more info on the partitioning. > > >>> > >>> /Erik > >>> > >>> Erik Marklund skrev: > >>>> Hi, > >>>> > >>>> I'm simulating non-periodic systems in vacuo, using constrained > >>>> h-bonds and particle decomposition. For some of my simulations the > >>>> cpu-usage seem far from optimal. The first cpu gets no atoms, while > >>>> the second one gets plenty and the remaining cpus get less than I > >>>> expected. Is this a bug? > >>>> > >>>> > >>>> An excerpt from the log file: > >>>> > >>>> There are: 2911 Atoms > >>>> There are: 317 VSites > >>>> splitting topology... > >>>> There are 999 charge group borders and 318 shake borders > >>>> There are 318 total borders > >>>> Division over nodes in atoms: > >>>> 0 1960 212 212 212 212 212 208 > >>>> Walking down the molecule graph to make constraint-blocks > >>>> CPU= 0, lastcg= -1, targetcg= 499, myshift= 1 > >>>> CPU= 1, lastcg= 681, targetcg= 182, myshift= 0 > >>>> CPU= 2, lastcg= 734, targetcg= 235, myshift= 7 > >>>> CPU= 3, lastcg= 787, targetcg= 288, myshift= 6 > >>>> CPU= 4, lastcg= 840, targetcg= 341, myshift= 5 > >>>> CPU= 5, lastcg= 893, targetcg= 394, myshift= 4 > >>>> CPU= 6, lastcg= 946, targetcg= 447, myshift= 3 > >>>> CPU= 7, lastcg= 998, targetcg= 499, myshift= 2 > >>>> pd->shift = 7, pd->bshift= 0 > >>>> Division of bonded forces over processors > >>>> CPU 0 1 2 3 4 5 6 7 > >>>> Workload division > >>>> nnodes: 8 > >>>> pd->shift: 7 > >>>> pd->bshift: 0 > >>>> Nodeid atom0 #atom cg0 #cg > >>>> 0 0 0 0 0 > >>>> 1 0 1960 0 682 > >>>> 2 1960 212 682 53 > >>>> 3 2172 212 735 53 > >>>> 4 2384 212 788 53 > >>>> 5 2596 212 841 53 > >>>> 6 2808 212 894 53 > >>>> 7 3020 208 947 52 > >>>> > >>>> … > >>>> Total Scaling: 18% of max performance > >>>> > >>> > >>> > >> > >> > > > > > > > -- > David van der Spoel, Ph.D., Professor of Biology > Molec. Biophys. group, Dept. of Cell & Molec. Biol., Uppsala University. > Box 596, 75124 Uppsala, Sweden. Phone: +46184714205. Fax: +4618511755. > sp...@xray.bmc.uu.se sp...@gromacs.org http://folding.bmc.uu.se > ___ > gmx-users mailing list gmx-users@gromacs.org > http://www.gromacs.org/mailman/listinfo/gmx-users > Please search the archive at http://www.gromacs.org/search before posting! > Please don't post (un)subscribe requests to the list. Use the > www interface or send it to gmx-users-requ...@gromacs.org. > Can't post? Read http://www.gromacs.org/mailing_lists/users.php What can you do with the new Windows Live? Find out <http://www.microsoft.com/windows/windowslive/default.aspx> ___ gmx-users mailing listgmx-users@gromacs.org http
Re: [gmx-users] Strange assignment of atoms to processors with pd
This was done with verision 4.0.4, taken straight from the website. I've already simulated a great deal with this version and I'd hate to swap versions in the middle of the project. As for the -debug runs, I did it for the dry analyte and for one with a 6 Å water layer. It seems that the protein is indeed divided over the processors if no other molecules are present. When there are other molecules, such as water, in the system then I suspect that the molecules are divided over the processors. This doesn't explain why the first processor got no atoms when I ran the simulations on 8 cpus though. without water ===-, There are: 1960 Atoms splitting topology... There are 698 charge group borders and 1 shake borders There are 698 total borders Division over nodes in atoms: 492 489 488 491 Walking down the molecule graph to make constraint-blocks CPU= 0, lastcg= 171, targetcg= 520, myshift=2 CPU= 1, lastcg= 354, targetcg=6, myshift=3 CPU= 2, lastcg= 521, targetcg= 172, myshift=2 CPU= 3, lastcg= 697, targetcg= 348, myshift=2 pd->shift = 3, pd->bshift= 0 Division of bonded forces over processors CPU 0 1 2 3 BONDS 259 263 250 253 ANGLES 894 867 901 870 PDIHS 110 11787 107 RBDIHS1002 967 1015 960 LJ14 1289 1260 1318 1234 CONSTR 242 232 244 241 Workload division nnodes: 4 pd->shift:3 pd->bshift: 0 Nodeid atom0 #atom cg0 #cg 0 0 492 0 172 1 492 489 172 183 2 981 488 355 167 31469 491 522 176 ===-' == with a bit of water ===-, There are: 5086 Atoms There are: 1042 VSites splitting topology... There are 1724 charge group borders and 1043 shake borders There are 1043 total borders Division over nodes in atoms: 1960138813921388 Walking down the molecule graph to make constraint-blocks CPU= 0, lastcg= 681, targetcg= 1543, myshift=3 CPU= 1, lastcg= 1028, targetcg= 166, myshift=3 CPU= 2, lastcg= 1376, targetcg= 514, myshift=2 CPU= 3, lastcg= 1723, targetcg= 862, myshift=2 pd->shift = 3, pd->bshift= 0 Division of bonded forces over processors CPU 0 1 2 3 BONDS 1025 0 0 0 ANGLES3547 0 0 0 PDIHS 426 0 0 0 RBDIHS3950 0 0 0 LJ14 5106 0 0 0 CONSTR 959 0 0 0 Workload division nnodes: 4 pd->shift:3 pd->bshift: 0 Nodeid atom0 #atom cg0 #cg 0 01960 0 682 119601388 682 347 2334813921029 348 3474013881377 347 Berk Hess skrev: Hi, This is plain 4.0 code is presume? This problem should be fixed then. But I now also made vacuum without cut-off working with domain decomposition in CVS head. Compared to a not-unbalanced PD (for instance only a protein, no water) DD is slightly slower. But DD will be faster than a badly balanced PD system. Berk > Date: Wed, 27 May 2009 11:04:49 +0200 > From: sp...@xray.bmc.uu.se > To: gmx-users@gromacs.org > Subject: Re: [gmx-users] Strange assignment of atoms to processors with pd > > Erik Marklund wrote: > > David van der Spoel skrev: > >> Erik Marklund wrote: > >>> I should add that this problem only seem to arise when the analyte is > >>> covered with a thin sheet of water. When simulating a dry analyte I > >>> get good scaling. In the latter case the charges, and therefore the > >>> topology, is slightly different. > >> How about vsites? Did you happen to turn them off as well in the > >> vacuum case? > > Turned off in all cases. The VSites mentioned in the log file is the > > 4:th particle on the tip4p-water molecules. > OK. Did you try a one step run with -debug? > It may give more info on the partitioning. > > >>> > >>> /Erik > >>> > >>> Erik Marklund skrev: > >>>> Hi, > >>>> > >>>> I'm simulating non-periodic systems in vacuo, using constrained > >>>> h-bonds and particle decomposition. For some of my simulations the > >>>> cpu-usage seem far from optimal. The first cpu gets no atoms, while > >>>> the second one gets plenty and the remaining cpus get less than I > >>>> expected. Is this a bug? > >>>> > >>>> > >>>> An excerpt from the log file: > >>>> > >>>> There are: 2911 Atoms > >>>> There are: 317 VSites >
RE: [gmx-users] Strange assignment of atoms to processors with pd
Hi, This is plain 4.0 code is presume? This problem should be fixed then. But I now also made vacuum without cut-off working with domain decomposition in CVS head. Compared to a not-unbalanced PD (for instance only a protein, no water) DD is slightly slower. But DD will be faster than a badly balanced PD system. Berk > Date: Wed, 27 May 2009 11:04:49 +0200 > From: sp...@xray.bmc.uu.se > To: gmx-users@gromacs.org > Subject: Re: [gmx-users] Strange assignment of atoms to processors with pd > > Erik Marklund wrote: > > David van der Spoel skrev: > >> Erik Marklund wrote: > >>> I should add that this problem only seem to arise when the analyte is > >>> covered with a thin sheet of water. When simulating a dry analyte I > >>> get good scaling. In the latter case the charges, and therefore the > >>> topology, is slightly different. > >> How about vsites? Did you happen to turn them off as well in the > >> vacuum case? > > Turned off in all cases. The VSites mentioned in the log file is the > > 4:th particle on the tip4p-water molecules. > OK. Did you try a one step run with -debug? > It may give more info on the partitioning. > > >>> > >>> /Erik > >>> > >>> Erik Marklund skrev: > >>>> Hi, > >>>> > >>>> I'm simulating non-periodic systems in vacuo, using constrained > >>>> h-bonds and particle decomposition. For some of my simulations the > >>>> cpu-usage seem far from optimal. The first cpu gets no atoms, while > >>>> the second one gets plenty and the remaining cpus get less than I > >>>> expected. Is this a bug? > >>>> > >>>> > >>>> An excerpt from the log file: > >>>> > >>>> There are: 2911 Atoms > >>>> There are: 317 VSites > >>>> splitting topology... > >>>> There are 999 charge group borders and 318 shake borders > >>>> There are 318 total borders > >>>> Division over nodes in atoms: > >>>> 0 1960 212 212 212 212 212 208 > >>>> Walking down the molecule graph to make constraint-blocks > >>>> CPU= 0, lastcg= -1, targetcg= 499, myshift= 1 > >>>> CPU= 1, lastcg= 681, targetcg= 182, myshift= 0 > >>>> CPU= 2, lastcg= 734, targetcg= 235, myshift= 7 > >>>> CPU= 3, lastcg= 787, targetcg= 288, myshift= 6 > >>>> CPU= 4, lastcg= 840, targetcg= 341, myshift= 5 > >>>> CPU= 5, lastcg= 893, targetcg= 394, myshift= 4 > >>>> CPU= 6, lastcg= 946, targetcg= 447, myshift= 3 > >>>> CPU= 7, lastcg= 998, targetcg= 499, myshift= 2 > >>>> pd->shift = 7, pd->bshift= 0 > >>>> Division of bonded forces over processors > >>>> CPU 0 1 2 3 4 5 6 7 > >>>> Workload division > >>>> nnodes: 8 > >>>> pd->shift: 7 > >>>> pd->bshift: 0 > >>>> Nodeid atom0 #atom cg0 #cg > >>>> 0 0 0 0 0 > >>>> 1 0 1960 0 682 > >>>> 2 1960 212 682 53 > >>>> 3 2172 212 735 53 > >>>> 4 2384 212 788 53 > >>>> 5 2596 212 841 53 > >>>> 6 2808 212 894 53 > >>>> 7 3020 208 947 52 > >>>> > >>>> … > >>>> Total Scaling: 18% of max performance > >>>> > >>> > >>> > >> > >> > > > > > > > -- > David van der Spoel, Ph.D., Professor of Biology > Molec. Biophys. group, Dept. of Cell & Molec. Biol., Uppsala University. > Box 596, 75124 Uppsala, Sweden. Phone:+46184714205. Fax: +4618511755. > sp...@xray.bmc.uu.se sp...@gromacs.org http://folding.bmc.uu.se > ___ > gmx-users mailing listgmx-users@gromacs.org > http://www.gromacs.org/mailman/listinfo/gmx-users > Please search the archive at http://www.gromacs.org/search before posting! > Please don't post (un)subscribe requests to the list. Use the > www interface or send it to gmx-users-requ...@gromacs.org. > Can't post? Read http://www.gromacs.org/mailing_lists/users.php _ What can you do with the new Windows Live? Find out http://www.microsoft.com/windows/windowslive/default.aspx___ gmx-users mailing listgmx-users@gromacs.org http://www.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. Can't post? Read http://www.gromacs.org/mailing_lists/users.php
Re: [gmx-users] Strange assignment of atoms to processors with pd
Erik Marklund wrote: David van der Spoel skrev: Erik Marklund wrote: I should add that this problem only seem to arise when the analyte is covered with a thin sheet of water. When simulating a dry analyte I get good scaling. In the latter case the charges, and therefore the topology, is slightly different. How about vsites? Did you happen to turn them off as well in the vacuum case? Turned off in all cases. The VSites mentioned in the log file is the 4:th particle on the tip4p-water molecules. OK. Did you try a one step run with -debug? It may give more info on the partitioning. /Erik Erik Marklund skrev: Hi, I'm simulating non-periodic systems in vacuo, using constrained h-bonds and particle decomposition. For some of my simulations the cpu-usage seem far from optimal. The first cpu gets no atoms, while the second one gets plenty and the remaining cpus get less than I expected. Is this a bug? An excerpt from the log file: There are: 2911 Atoms There are: 317 VSites splitting topology... There are 999 charge group borders and 318 shake borders There are 318 total borders Division over nodes in atoms: 0 1960 212 212 212 212 212 208 Walking down the molecule graph to make constraint-blocks CPU= 0, lastcg= -1, targetcg= 499, myshift= 1 CPU= 1, lastcg= 681, targetcg= 182, myshift= 0 CPU= 2, lastcg= 734, targetcg= 235, myshift= 7 CPU= 3, lastcg= 787, targetcg= 288, myshift= 6 CPU= 4, lastcg= 840, targetcg= 341, myshift= 5 CPU= 5, lastcg= 893, targetcg= 394, myshift= 4 CPU= 6, lastcg= 946, targetcg= 447, myshift= 3 CPU= 7, lastcg= 998, targetcg= 499, myshift= 2 pd->shift = 7, pd->bshift= 0 Division of bonded forces over processors CPU 0 1 2 3 4 5 6 7 Workload division nnodes: 8 pd->shift: 7 pd->bshift: 0 Nodeid atom0 #atom cg0 #cg 0 0 0 0 0 1 0 1960 0 682 2 1960 212 682 53 3 2172 212 735 53 4 2384 212 788 53 5 2596 212 841 53 6 2808 212 894 53 7 3020 208 947 52 … Total Scaling: 18% of max performance -- David van der Spoel, Ph.D., Professor of Biology Molec. Biophys. group, Dept. of Cell & Molec. Biol., Uppsala University. Box 596, 75124 Uppsala, Sweden. Phone: +46184714205. Fax: +4618511755. sp...@xray.bmc.uu.sesp...@gromacs.org http://folding.bmc.uu.se ___ gmx-users mailing listgmx-users@gromacs.org http://www.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. Can't post? Read http://www.gromacs.org/mailing_lists/users.php
Re: [gmx-users] Strange assignment of atoms to processors with pd
David van der Spoel skrev: Erik Marklund wrote: I should add that this problem only seem to arise when the analyte is covered with a thin sheet of water. When simulating a dry analyte I get good scaling. In the latter case the charges, and therefore the topology, is slightly different. How about vsites? Did you happen to turn them off as well in the vacuum case? Turned off in all cases. The VSites mentioned in the log file is the 4:th particle on the tip4p-water molecules. /Erik Erik Marklund skrev: Hi, I'm simulating non-periodic systems in vacuo, using constrained h-bonds and particle decomposition. For some of my simulations the cpu-usage seem far from optimal. The first cpu gets no atoms, while the second one gets plenty and the remaining cpus get less than I expected. Is this a bug? An excerpt from the log file: There are: 2911 Atoms There are: 317 VSites splitting topology... There are 999 charge group borders and 318 shake borders There are 318 total borders Division over nodes in atoms: 0 1960 212 212 212 212 212 208 Walking down the molecule graph to make constraint-blocks CPU= 0, lastcg= -1, targetcg= 499, myshift= 1 CPU= 1, lastcg= 681, targetcg= 182, myshift= 0 CPU= 2, lastcg= 734, targetcg= 235, myshift= 7 CPU= 3, lastcg= 787, targetcg= 288, myshift= 6 CPU= 4, lastcg= 840, targetcg= 341, myshift= 5 CPU= 5, lastcg= 893, targetcg= 394, myshift= 4 CPU= 6, lastcg= 946, targetcg= 447, myshift= 3 CPU= 7, lastcg= 998, targetcg= 499, myshift= 2 pd->shift = 7, pd->bshift= 0 Division of bonded forces over processors CPU 0 1 2 3 4 5 6 7 Workload division nnodes: 8 pd->shift: 7 pd->bshift: 0 Nodeid atom0 #atom cg0 #cg 0 0 0 0 0 1 0 1960 0 682 2 1960 212 682 53 3 2172 212 735 53 4 2384 212 788 53 5 2596 212 841 53 6 2808 212 894 53 7 3020 208 947 52 … Total Scaling: 18% of max performance -- --- Erik Marklund, PhD student Laboratory of Molecular Biophysics, Dept. of Cell and Molecular Biology, Uppsala University. Husargatan 3, Box 596,75124 Uppsala, Sweden phone:+46 18 471 4537fax: +46 18 511 755 er...@xray.bmc.uu.sehttp://xray.bmc.uu.se/molbiophys ___ gmx-users mailing listgmx-users@gromacs.org http://www.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. Can't post? Read http://www.gromacs.org/mailing_lists/users.php
Re: [gmx-users] Strange assignment of atoms to processors with pd
Erik Marklund wrote: I should add that this problem only seem to arise when the analyte is covered with a thin sheet of water. When simulating a dry analyte I get good scaling. In the latter case the charges, and therefore the topology, is slightly different. How about vsites? Did you happen to turn them off as well in the vacuum case? /Erik Erik Marklund skrev: Hi, I'm simulating non-periodic systems in vacuo, using constrained h-bonds and particle decomposition. For some of my simulations the cpu-usage seem far from optimal. The first cpu gets no atoms, while the second one gets plenty and the remaining cpus get less than I expected. Is this a bug? An excerpt from the log file: There are: 2911 Atoms There are: 317 VSites splitting topology... There are 999 charge group borders and 318 shake borders There are 318 total borders Division over nodes in atoms: 0 1960 212 212 212 212 212 208 Walking down the molecule graph to make constraint-blocks CPU= 0, lastcg= -1, targetcg= 499, myshift= 1 CPU= 1, lastcg= 681, targetcg= 182, myshift= 0 CPU= 2, lastcg= 734, targetcg= 235, myshift= 7 CPU= 3, lastcg= 787, targetcg= 288, myshift= 6 CPU= 4, lastcg= 840, targetcg= 341, myshift= 5 CPU= 5, lastcg= 893, targetcg= 394, myshift= 4 CPU= 6, lastcg= 946, targetcg= 447, myshift= 3 CPU= 7, lastcg= 998, targetcg= 499, myshift= 2 pd->shift = 7, pd->bshift= 0 Division of bonded forces over processors CPU 0 1 2 3 4 5 6 7 Workload division nnodes: 8 pd->shift: 7 pd->bshift: 0 Nodeid atom0 #atom cg0 #cg 0 0 0 0 0 1 0 1960 0 682 2 1960 212 682 53 3 2172 212 735 53 4 2384 212 788 53 5 2596 212 841 53 6 2808 212 894 53 7 3020 208 947 52 … Total Scaling: 18% of max performance -- David van der Spoel, Ph.D., Professor of Biology Molec. Biophys. group, Dept. of Cell & Molec. Biol., Uppsala University. Box 596, 75124 Uppsala, Sweden. Phone: +46184714205. Fax: +4618511755. sp...@xray.bmc.uu.sesp...@gromacs.org http://folding.bmc.uu.se ___ gmx-users mailing listgmx-users@gromacs.org http://www.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. Can't post? Read http://www.gromacs.org/mailing_lists/users.php
Re: [gmx-users] Strange assignment of atoms to processors with pd
I should add that this problem only seem to arise when the analyte is covered with a thin sheet of water. When simulating a dry analyte I get good scaling. In the latter case the charges, and therefore the topology, is slightly different. /Erik Erik Marklund skrev: Hi, I'm simulating non-periodic systems in vacuo, using constrained h-bonds and particle decomposition. For some of my simulations the cpu-usage seem far from optimal. The first cpu gets no atoms, while the second one gets plenty and the remaining cpus get less than I expected. Is this a bug? An excerpt from the log file: There are: 2911 Atoms There are: 317 VSites splitting topology... There are 999 charge group borders and 318 shake borders There are 318 total borders Division over nodes in atoms: 0 1960 212 212 212 212 212 208 Walking down the molecule graph to make constraint-blocks CPU= 0, lastcg= -1, targetcg= 499, myshift= 1 CPU= 1, lastcg= 681, targetcg= 182, myshift= 0 CPU= 2, lastcg= 734, targetcg= 235, myshift= 7 CPU= 3, lastcg= 787, targetcg= 288, myshift= 6 CPU= 4, lastcg= 840, targetcg= 341, myshift= 5 CPU= 5, lastcg= 893, targetcg= 394, myshift= 4 CPU= 6, lastcg= 946, targetcg= 447, myshift= 3 CPU= 7, lastcg= 998, targetcg= 499, myshift= 2 pd->shift = 7, pd->bshift= 0 Division of bonded forces over processors CPU 0 1 2 3 4 5 6 7 Workload division nnodes: 8 pd->shift: 7 pd->bshift: 0 Nodeid atom0 #atom cg0 #cg 0 0 0 0 0 1 0 1960 0 682 2 1960 212 682 53 3 2172 212 735 53 4 2384 212 788 53 5 2596 212 841 53 6 2808 212 894 53 7 3020 208 947 52 … Total Scaling: 18% of max performance -- --- Erik Marklund, PhD student Laboratory of Molecular Biophysics, Dept. of Cell and Molecular Biology, Uppsala University. Husargatan 3, Box 596,75124 Uppsala, Sweden phone:+46 18 471 4537fax: +46 18 511 755 er...@xray.bmc.uu.sehttp://xray.bmc.uu.se/molbiophys ___ gmx-users mailing listgmx-users@gromacs.org http://www.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. Can't post? Read http://www.gromacs.org/mailing_lists/users.php
[gmx-users] Strange assignment of atoms to processors with pd
Hi, I'm simulating non-periodic systems in vacuo, using constrained h-bonds and particle decomposition. For some of my simulations the cpu-usage seem far from optimal. The first cpu gets no atoms, while the second one gets plenty and the remaining cpus get less than I expected. Is this a bug? An excerpt from the log file: There are: 2911 Atoms There are: 317 VSites splitting topology... There are 999 charge group borders and 318 shake borders There are 318 total borders Division over nodes in atoms: 0 1960 212 212 212 212 212 208 Walking down the molecule graph to make constraint-blocks CPU= 0, lastcg= -1, targetcg= 499, myshift= 1 CPU= 1, lastcg= 681, targetcg= 182, myshift= 0 CPU= 2, lastcg= 734, targetcg= 235, myshift= 7 CPU= 3, lastcg= 787, targetcg= 288, myshift= 6 CPU= 4, lastcg= 840, targetcg= 341, myshift= 5 CPU= 5, lastcg= 893, targetcg= 394, myshift= 4 CPU= 6, lastcg= 946, targetcg= 447, myshift= 3 CPU= 7, lastcg= 998, targetcg= 499, myshift= 2 pd->shift = 7, pd->bshift= 0 Division of bonded forces over processors CPU 0 1 2 3 4 5 6 7 Workload division nnodes: 8 pd->shift: 7 pd->bshift: 0 Nodeid atom0 #atom cg0 #cg 0 0 0 0 0 1 0 1960 0 682 2 1960 212 682 53 3 2172 212 735 53 4 2384 212 788 53 5 2596 212 841 53 6 2808 212 894 53 7 3020 208 947 52 … Total Scaling: 18% of max performance -- --- Erik Marklund, PhD student Laboratory of Molecular Biophysics, Dept. of Cell and Molecular Biology, Uppsala University. Husargatan 3, Box 596,75124 Uppsala, Sweden phone:+46 18 471 4537fax: +46 18 511 755 er...@xray.bmc.uu.sehttp://xray.bmc.uu.se/molbiophys ___ gmx-users mailing listgmx-users@gromacs.org http://www.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. Can't post? Read http://www.gromacs.org/mailing_lists/users.php