On Mon, Mar 4, 2013 at 6:02 AM, Reid Van Lehn <rvanl...@mit.edu> wrote:
> Hello users, > > I ran into a bug I do not understand today upon upgrading from v. 4.5.5 to > v 4.6. I'm using older 8 core Intel Xeon E5430 machines, and when I > submitted a job for 8 cores to one of the nodes I received the following > error: > > NOTE: In thread-MPI thread #3: Affinity setting failed. > This can cause performance degradation! > > NOTE: In thread-MPI thread #2: Affinity setting failed. > This can cause performance degradation! > > NOTE: In thread-MPI thread #1: Affinity setting failed. > This can cause performance degradation! > > I ran mdrun simply with the flags: > > mdrun -v -ntmpi 8 -deffnm em > > Using the top command, I confirmed that no other programs were running and > that mdrun was in fact only using 5 cores. Reducing -ntmpi to 7, however, > resulted in no error (only a warning about not using all of the logical > cores) and mdrun used 7 cores correctly. Since it warned about thread > affinity settings, I tried setting -pin on -pinoffset 0 even though I was > using all the cores on the machine. This resulted in the same error. > However, turning pinning off explicitly with -pin off (rather than -pin > auto) did correctly give me the all 8 cores again. > > While I figured out a solution in this particular instance, my question is > whether I should be have known from my hardware/settings that pinning > should be turned off (for future reference), or if this is a bug? > I'm not sure - those are 2007-era processors, so there may be some limitations in what they could do (or how well the kernel and system libraries support it). So investing time into working out the real problem is not really worthwhile. Thanks for reporting your work-around, however, others might benefit from it. If you plan on doing lengthy simulations, you might like to verify that you get linear scaling with increasing -ntmpi, and/or compare performance with the MPI version on the same hardware. Mark -- gmx-users mailing list gmx-users@gromacs.org http://lists.gromacs.org/mailman/listinfo/gmx-users * Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/Search before posting! * Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists