Hi Dietmar, On Wed, 31 Oct 2018 at 08:20, Dietmar Eggemann <dietmar.eggem...@arm.com> wrote: > > On 10/26/18 6:11 PM, Vincent Guittot wrote: > > [...] > > > static int select_idle_sibling(struct task_struct *p, int prev_cpu, int > > cpu); > > static unsigned long task_h_load(struct task_struct *p); > > @@ -764,7 +763,7 @@ void post_init_entity_util_avg(struct sched_entity *se) > > * such that the next switched_to_fair() has the > > * expected state. > > */ > > - se->avg.last_update_time = cfs_rq_clock_task(cfs_rq); > > + se->avg.last_update_time = cfs_rq_clock_pelt(cfs_rq); > > return; > > } > > } > > There is this 1/cpu scaling of se->avg.util_sum (running_sum) in > update_tg_cfs_runnable() so it can be used to calculate > se->avg.runnable_load_sum (runnable_sum). I guess with your approach > this should be removed.
Yes good catch > > > @@ -3466,7 +3465,7 @@ static void detach_entity_load_avg(struct cfs_rq > > *cfs_rq, struct sched_entity *s > > /* Update task and its cfs_rq load average */ > > static inline void update_load_avg(struct cfs_rq *cfs_rq, struct > > sched_entity *se, int flags) > > { > > - u64 now = cfs_rq_clock_task(cfs_rq); > > + u64 now = cfs_rq_clock_pelt(cfs_rq); > > struct rq *rq = rq_of(cfs_rq); > > int cpu = cpu_of(rq); > > int decayed; > > @@ -6694,6 +6693,12 @@ done: __maybe_unused; > > if (new_tasks > 0) > > goto again; > > > > + /* > > + * rq is about to be idle, check if we need to update the > > + * lost_idle_time of clock_pelt > > + */ > > + update_idle_rq_clock_pelt(rq); > > + > > return NULL; > > } > > [...]