We will rely on the per CPU fault statistics and its shared/private derivative even more in the future, so stabilize this metric even better.
The staged updates introduced in commit: sched: Introduce staged average NUMA faults Already stabilized this key metric significantly, but in real workloads it was still reacting to temporary load balancing transients too quickly. Slow down by weighting the average. The weighting value was found via experimentation. Cc: Linus Torvalds <torva...@linux-foundation.org> Cc: Andrew Morton <a...@linux-foundation.org> Cc: Peter Zijlstra <a.p.zijls...@chello.nl> Cc: Andrea Arcangeli <aarca...@redhat.com> Cc: Rik van Riel <r...@redhat.com> Cc: Mel Gorman <mgor...@suse.de> Cc: Hugh Dickins <hu...@google.com> Signed-off-by: Ingo Molnar <mi...@kernel.org> --- kernel/sched/fair.c | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 24a5588..a5f3ad7 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -914,8 +914,8 @@ static void task_numa_placement(struct task_struct *p) p->numa_faults_curr[idx] = 0; /* Keep a simple running average: */ - p->numa_faults[idx] += new_faults; - p->numa_faults[idx] /= 2; + p->numa_faults[idx] = p->numa_faults[idx]*7 + new_faults; + p->numa_faults[idx] /= 8; faults += p->numa_faults[idx]; total[priv] += p->numa_faults[idx]; -- 1.7.11.7 -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/