On 02-Apr 11:41, Patrick Bellasi wrote:
> Because of bucketization, different task-specific clamp values are
> tracked in the same bucket.  For example, with 20% bucket size and
> assuming to have:
>   Task1: util_min=25%
>   Task2: util_min=35%
> both tasks will be refcounted in the [20..39]% bucket and always boosted
> only up to 20% thus implementing a simple floor aggregation normally
> used in histograms.
> 
> In systems with only few and well-defined clamp values, it would be
> useful to track the exact clamp value required by a task whenever
> possible. For example, if a system requires only 23% and 47% boost
> values then it's possible to track the exact boost required by each
> task using only 3 buckets of ~33% size each.
> 
> Introduce a mechanism to max aggregate the requested clamp values of
> RUNNABLE tasks in the same bucket. Keep it simple by resetting the
> bucket value to its base value only when a bucket becomes inactive.
> Allow a limited and controlled overboosting margin for tasks recounted
> in the same bucket.
> 
> In systems where the boost values are not known in advance, it is still
> possible to control the maximum acceptable overboosting margin by tuning
> the number of clamp groups. For example, 20 groups ensure a 5% maximum
> overboost.
> 
> Remove the rq bucket initialization code since a correct bucket value
> is now computed when a task is refcounted into a CPU's rq.
> 
> Signed-off-by: Patrick Bellasi <patrick.bell...@arm.com>
> Cc: Ingo Molnar <mi...@redhat.com>
> Cc: Peter Zijlstra <pet...@infradead.org>
> 
> --
> Changes in v8:
>  Message-ID: <20190313193916.gq2...@worktop.programming.kicks-ass.net>
>  - split this code out from the previous patch
> ---
>  kernel/sched/core.c | 46 ++++++++++++++++++++++++++-------------------
>  1 file changed, 27 insertions(+), 19 deletions(-)
> 
> diff --git a/kernel/sched/core.c b/kernel/sched/core.c
> index 032211b72110..6e1beae5f348 100644
> --- a/kernel/sched/core.c
> +++ b/kernel/sched/core.c
> @@ -778,6 +778,11 @@ unsigned int uclamp_rq_max_value(struct rq *rq, unsigned 
> int clamp_id)
>   * When a task is enqueued on a rq, the clamp bucket currently defined by the
>   * task's uclamp::bucket_id is refcounted on that rq. This also immediately
>   * updates the rq's clamp value if required.
> + *
> + * Tasks can have a task-specific value requested from user-space, track
> + * within each bucket the maximum value for tasks refcounted in it.
> + * This "local max aggregation" allows to track the exact "requested" value
> + * for each bucket when all its RUNNABLE tasks require the same clamp.
>   */
>  static inline void uclamp_rq_inc_id(struct task_struct *p, struct rq *rq,
>                                   unsigned int clamp_id)
> @@ -789,8 +794,15 @@ static inline void uclamp_rq_inc_id(struct task_struct 
> *p, struct rq *rq,
>       bucket = &uc_rq->bucket[uc_se->bucket_id];
>       bucket->tasks++;
>  
> +     /*
> +      * Local max aggregation: rq buckets always track the max
> +      * "requested" clamp value of its RUNNABLE tasks.
> +      */
> +     if (uc_se->value > bucket->value)
> +             bucket->value = uc_se->value;
> +
>       if (uc_se->value > READ_ONCE(uc_rq->value))
> -             WRITE_ONCE(uc_rq->value, bucket->value);
> +             WRITE_ONCE(uc_rq->value, uc_se->value);
>  }
>  
>  /*
> @@ -815,6 +827,12 @@ static inline void uclamp_rq_dec_id(struct task_struct 
> *p, struct rq *rq,
>       if (likely(bucket->tasks))
>               bucket->tasks--;
>  
> +     /*
> +      * Keep "local max aggregation" simple and accept to (possibly)
> +      * overboost some RUNNABLE tasks in the same bucket.
> +      * The rq clamp bucket value is reset to its base value whenever
> +      * there are no more RUNNABLE tasks refcounting it.
> +      */
>       if (likely(bucket->tasks))
>               return;
>  
> @@ -824,8 +842,14 @@ static inline void uclamp_rq_dec_id(struct task_struct 
> *p, struct rq *rq,
>        * e.g. due to future modification, warn and fixup the expected value.
>        */
>       SCHED_WARN_ON(bucket->value > rq_clamp);
> -     if (bucket->value >= rq_clamp)
> +     if (bucket->value >= rq_clamp) {
> +             /*
> +              * Reset clamp bucket value to its nominal value whenever
> +              * there are anymore RUNNABLE tasks refcounting it.
> +              */
> +             bucket->value = uclamp_bucket_base_value(bucket->value);

While running tests on Android, I found that the snipped above should
be better done in uclamp_rq_inc_id() for two main reasons:
 1. because of the early return in this function, we skip the reset in
    case the task is not the last one running on a CPU, thus
    triggering the above SCHED_WARN_ON
 2. since a non active bucket is already ignored, we don't care about
    resetting its "local max value"

Will move the "max local update" in uclamp_rq_inc_id() where something
like:


    /*
     * Local max aggregation: rq buckets always track the max
     * "requested" clamp value of its RUNNABLE tasks.
     */
    if (bucket->tasks == 1 || uc_se->value > bucket->value)
        bucket->value = uc_se->value;


Should grant to always have an updated local max every time a bucket
is active.


>               WRITE_ONCE(uc_rq->value, uclamp_rq_max_value(rq, clamp_id));
> +     }
>  }
>  
>  static inline void uclamp_rq_inc(struct rq *rq, struct task_struct *p)
> @@ -855,25 +879,9 @@ static void __init init_uclamp(void)
>       unsigned int clamp_id;
>       int cpu;
>  
> -     for_each_possible_cpu(cpu) {
> -             struct uclamp_bucket *bucket;
> -             struct uclamp_rq *uc_rq;
> -             unsigned int bucket_id;
> -
> +     for_each_possible_cpu(cpu)
>               memset(&cpu_rq(cpu)->uclamp, 0, sizeof(struct uclamp_rq));
>  
> -             for (clamp_id = 0; clamp_id < UCLAMP_CNT; ++clamp_id) {
> -                     uc_rq = &cpu_rq(cpu)->uclamp[clamp_id];
> -
> -                     bucket_id = 1;
> -                     while (bucket_id < UCLAMP_BUCKETS) {
> -                             bucket = &uc_rq->bucket[bucket_id];
> -                             bucket->value = bucket_id * UCLAMP_BUCKET_DELTA;
> -                             ++bucket_id;
> -                     }
> -             }
> -     }
> -
>       for (clamp_id = 0; clamp_id < UCLAMP_CNT; ++clamp_id) {
>               struct uclamp_se *uc_se = &init_task.uclamp[clamp_id];
>  
> -- 
> 2.20.1
> 

-- 
#include <best/regards.h>

Patrick Bellasi

Reply via email to