Hi,

Did you end up coming to a conclusion on this?  I have TaskAffinity turned
off in cgroup.conf, but was thinking that it makes sense to turn it on.

Thanks.

Chris



On 2/11/15, 2:01 PM, "Jonathan Perkins" <perki...@cse.ohio-state.edu>
wrote:

>Hi Ryan, do you still have SLURM's TaskAffinity turned off?  If this is
>turned on I would expect your binding problem to be resolved since you're
>now using srun to launch the jobs.
>
>On Wed, Feb 11, 2015 at 3:34 PM, Ryan Novosielski
><novos...@ca.rutgers.edu> wrote:
>
>
>-----BEGIN PGP SIGNED MESSAGE-----
>Hash: SHA1
>
>So I've rebuilt MVAPICH2 2.0.1, with the flags for SLURM PMI. Then I
>rebuilt AMBER 14. It solved one problem, which was that every time I
>ran pmemd.cuda.MPI with srun, it would run all of the processes
>against the same GPU card. That no longer happens and it uses
>different cards. However, the CPU affinity thing is still a problem.
>It will use the same CPU's for both jobs. This does not seem to be an
>issue when running jobs via Torque, so I'm not really sure what is
>happening. Could it be that it is relying on the Torque cpuset in
>order to cause certain CPU's to be used and that MVAPICH2 is not
>really doing it?
>
>I'm actually not sure what mailing list this belongs on at this point.
>It does seem as is this works right with Torque and not SLURM, which
>would seem to implicate SLURM. But it seems that MVAPICH2 should be
>making this happen and isn't. I guess if I had some pointers for where
>to look here, I could figure out what's going on.
>
>On 02/10/2015 09:24 PM, Novosielski, Ryan wrote:
>>
>> So it certainly could be related to affinity. Here is the
>> affinity-related output from the two PBS jobs:
>>
>> run001: -------------CPU AFFINITY------------- RANK:0  CPU_SET:
>> 0 RANK:1  CPU_SET:   1 -------------------------------------
>> -------------CPU AFFINITY------------- RANK:0  CPU_SET:   0 RANK:1
>> CPU_SET:   1 -------------------------------------
>>
>> run002: -------------CPU AFFINITY------------- RANK:0  CPU_SET:
>> 4 RANK:1  CPU_SET:   5 -------------------------------------
>> -------------CPU AFFINITY------------- RANK:0  CPU_SET:   4 RANK:1
>> CPU_SET:   5 -------------------------------------
>>
>> Now the two SLURM jobs: run001: -------------CPU
>> AFFINITY------------- RANK:0  CPU_SET:   0 RANK:1  CPU_SET:   1
>> -------------------------------------
>>
>> run002: -------------CPU AFFINITY------------- RANK:0  CPU_SET:
>> 0 RANK:1  CPU_SET:   1 -------------------------------------
>>
>> Both jobs are running on an MVAPICH2 that was built without setting
>> SLURM as the PMI. The jobs spawn the processes via MVAPICH2's
>> mpiexec. I tried srun, but it seemed to run all of the jobs on the
>> same GPU, so I was waiting to recompile MVAPICH2 and then AMBER
>> using the SLURM PMI. I guess it's possible that will solve the
>> problem, but this is still peculiar.
>>
>> -- ____ *Note: UMDNJ is now Rutgers-Biomedical and Health
>> Sciences* || \\UTGERS
>> |---------------------*O*--------------------- ||_// Biomedical |
>> Ryan Novosielski - Senior Technologist || \\ and Health |
>> novos...@rutgers.edu -
>973/972.0922 <tel:973%2F972.0922> (2x0922) ||  \\  Sciences |
>> OIRT/High Perf & Res Comp - MSB C630, Newark `'
>> ________________________________________ From: Jonathan Perkins
>> [perki...@cse.ohio-state.edu] Sent: Tuesday, February 10, 2015 4:42
>> PM To: slurm-dev Cc: Novosielski, Ryan Subject: [slurm-dev] Amber +
>> MVAPICH2 slower with SLURM vs PBS
>>
>> Do you have both environments available to do this comparision.  If
>> so, is only SLURM vs Torque the only difference?
>>
>> I do think that it'll be good to provide the output of the MPI job
>> with those two variables that I mentioned in the earlier post.
>> Maybe it will show a difference in affinity.  Otherwise it can be
>> something else at play.
>>
>> Between your two jobs with SLURM, did you did you only flip the
>> Task Affinity setting?  It seems that affinity in MVAPICH2 was
>> enabled in both runs so I would expect the second run to not
>> perform so badly.
>>
>> On Sat, Feb 07, 2015 at 10:48:31AM -0800, Novosielski, Ryan wrote:
>>> So I turned off TaskAffinity (=none) and we ran two CUDA/GPU jobs
>>> on one node. Apparently the performance with PBS/Torque is good
>>> and with Slurm it is not. I'm confused as to why it would make
>>> any difference:
>>>
>>> Running 1 MPI job with slurm Gpu utilization 74-99% Cpu  — 4 CPU
>>> cores, 0-3 64-84% utilization
>>>
>>> Speed: 21ns/day vs previously reported 25.6ns/day with PBS
>>>
>>> Submitting the second MPI job Gpu utilization down to 9-13%
>>> (slightly better, it was 1-3% before [with TaskAffinity
>>> enabled]) Cpu  — 4 CPU cores, the same 0-3 99%  utilization
>>>
>>> Speed:  slow..   Okay, finally got it  1.45ns/day
>>>
>>> Would that variable still be helpful to try?
>>>
>>> We're using Slurm 14.11.3, MVAPICH2 2.0, Intel Compiler 15.0.1,
>>> and AMBER 14 for these performance numbers. GPU's are M2090's I
>>> think. I'd have to check that he wasn't using the K20's.
>>>
>>> ____ *Note: UMDNJ is now Rutgers-Biomedical and Health Sciences*
>>> || \\UTGERS      |---------------------*O*---------------------
>>> ||_// Biomedical | Ryan Novosielski - Senior Technologist || \\
>>> and Health | novos...@rutgers.edu<mailto:novos...@rutgers.edu>-
>>> 973/972.0922 <tel:973%2F972.0922> (2x0922) ||  \\  Sciences |
>>>OIRT/High Perf & Res
>>> Comp - MSB C630, Newark `'
>>>
>>> On Feb 7, 2015, at 12:34, Jonathan Perkins
>>> <perki...@cse.ohio-state.edu<mailto:perki...@cse.ohio-state.edu>>
>>> wrote:
>>>
>>>
>>> Can you set MV2_SHOW_CPU_BINDING equal to 1 when running your
>>> job?  This should show whether affinity is causing your processes
>>> to be oversubscribed on a set of cores.
>>>
>>> If this is the case you can disable the affinity from the library
>>> by setting MV2_ENABLE_AFFINITY to 0.
>>>
>>> On Fri, Feb 06, 2015 at 03:18:48PM -0800, Novosielski, Ryan
>>> wrote: I am running into a similar problem, with Slurm 14.11.3
>>> and MVAPICH2 2.0. I am wondering if perhaps having CPU affinity
>>> configured in MVAPICH2 and Slurm at the same time isn't a bad
>>> idea (I've also since realized that it uses cgroups and that the
>>> 2.6.18 kernel in RHEL5 does not support it anyway -- but it
>>> didn't seem to be harming anything. Maybe it was?).
>>>
>>> ____ *Note: UMDNJ is now Rutgers-Biomedical and Health Sciences*
>>> || \\UTGERS      |---------------------*O*---------------------
>>> ||_// Biomedical | Ryan Novosielski - Senior Technologist || \\
>>> and Health |
>>> 
>>>novos...@rutgers.edu<mailto:novos...@rutgers.edu><mailto:novosirj@rutger
>>>s.edu>-
>>> 973/972.0922 <tel:973%2F972.0922> (2x0922) ||  \\  Sciences |
>>>OIRT/High Perf & Res
>>> Comp - MSB C630, Newark `'
>>
>> -- Jonathan Perkins
>>
>
>
>
>- --
>____ *Note: UMDNJ is now Rutgers-Biomedical and Health Sciences*
>|| \\UTGERS      |---------------------*O*---------------------
>||_// Biomedical | Ryan Novosielski - Senior Technologist
>|| \\ and Health | novos...@rutgers.edu -
>973/972.0922 <tel:973%2F972.0922> (2x0922)
>||  \\  Sciences | OIRT/High Perf & Res Comp - MSB C630, Newark
>     `'
>-----BEGIN PGP SIGNATURE-----
>Version: GnuPG v1
>
>iEYEARECAAYFAlTbuiwACgkQmb+gadEcsb4HQwCdG8nMs/Qxt2JqVpyBtqoS1IvQ
>UScAoNpGLd3AhH0Zkyh0J0XRIFwg66FN
>=HGsw
>-----END PGP SIGNATURE-----
>
>
>
>
>
>
>
>-- 
>Jonathan Perkins
>http://www.cse.ohio-state.edu/~perkinjo
><http://www.cse.ohio-state.edu/%7Eperkinjo>
>
>

Reply via email to