>On 23/12/2010 10:01 PM, Wojtyczka, André wrote: >> Dear Gromacs Enthusiasts. >> >> I am experiencing problems with mdrun_mpi (4.5.3) on a Nehalem cluster. >> >> Problem: >> This runs fine: >> mpiexec -np 72 /../mdrun_mpi -pd -s full031K_mdrun_ions.tpr >> >> This produces a segmentation fault: >> mpiexec -np 128 /../mdrun_mpi -pd -s full031K_mdrun_ions.tpr > >Unless you know you need it, don't use -pd. DD will be faster and is >probably better bug-tested too. > >Mark
Hi Mark thanks for the push into that direction, but I am in the unfortunate situation where I really need -pd because I have long bonds which is the reason why my large system is decomposable just into a little number of domains. > >> So the only difference is the number of cores I am using. >> >> mdrun_mpi was compiled using the intel compiler 11.1.072 with my own fftw3 >> installation. >> >> While configuring and make mdrun / make install-mdrun no errors came >> up. >> >> Is there some issue with threading or mpi? >> >> If someone has a clue please give me a hint. >> >> >> integrator = md >> dt = 0.004 >> nsteps = 25000000 >> nstxout = 0 >> nstvout = 0 >> nstlog = 250000 >> nstenergy = 250000 >> nstxtcout = 12500 >> xtc_grps = protein >> energygrps = protein non-protein >> nstlist = 2 >> ns_type = grid >> rlist = 0.9 >> coulombtype = PME >> rcoulomb = 0.9 >> fourierspacing = 0.12 >> pme_order = 4 >> ewald_rtol = 1e-5 >> rvdw = 0.9 >> pbc = xyz >> periodic_molecules = yes >> tcoupl = nose-hoover >> nsttcouple = 1 >> tc-grps = protein non-protein >> tau_t = 0.1 0.1 >> ref_t = 310 310 >> Pcoupl = no >> gen_vel = yes >> gen_temp = 310 >> gen_seed = 173529 >> constraints = all-bonds >> >> >> >> Error: >> Getting Loaded... >> Reading file full031K_mdrun_ions.tpr, VERSION 4.5.3 (single precision) >> Loaded with Money >> >> >> NOTE: The load imbalance in PME FFT and solve is 48%. >> For optimal PME load balancing >> PME grid_x (144) and grid_y (144) should be divisible by #PME_nodes_x >> (128) >> and PME grid_y (144) and grid_z (144) should be divisible by >> #PME_nodes_y (1) >> >> >> Step 0, time 0 (ps) >> PSIlogger: Child with rank 82 exited on signal 11: Segmentation fault >> PSIlogger: Child with rank 79 exited on signal 11: Segmentation fault >> PSIlogger: Child with rank 2 exited on signal 11: Segmentation fault >> PSIlogger: Child with rank 1 exited on signal 11: Segmentation fault >> PSIlogger: Child with rank 100 exited on signal 11: Segmentation fault >> PSIlogger: Child with rank 97 exited on signal 11: Segmentation fault >> PSIlogger: Child with rank 98 exited on signal 11: Segmentation fault >> PSIlogger: Child with rank 96 exited on signal 6: Aborted >> ... >> >> Ps, for now I don't care about the imbalanced PME load unless it's >> independent from my problem. >> >> Cheers >> André ------------------------------------------------------------------------------------------------ ------------------------------------------------------------------------------------------------ Forschungszentrum Juelich GmbH 52425 Juelich Sitz der Gesellschaft: Juelich Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498 Vorsitzender des Aufsichtsrats: MinDirig Dr. Karl Eugen Huthmacher Geschaeftsfuehrung: Prof. Dr. Achim Bachem (Vorsitzender), Dr. Ulrich Krafft (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt, Prof. Dr. Sebastian M. Schmidt ------------------------------------------------------------------------------------------------ ------------------------------------------------------------------------------------------------ -- gmx-users mailing list gmx-users@gromacs.org http://lists.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/Search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. Can't post? Read http://www.gromacs.org/Support/Mailing_Lists