On 07/24/20 11:12, Peter Zijlstra wrote:
> On Thu, Jul 16, 2020 at 12:03:47PM +0100, Qais Yousef wrote:
> 
> I've trimmed the Changelog to read like:

+1

Should we mention the ordering issue too? Or maybe I misinterpreted the
'Possible unsafe locking scenario' part?

> 
> ---
> Subject: sched/uclamp: Fix a deadlock when enabling uclamp static key
> From: Qais Yousef <[email protected]>
> Date: Thu, 16 Jul 2020 12:03:47 +0100
> 
> From: Qais Yousef <[email protected]>
> 
> The following splat was caught when setting uclamp value of a task:
> 
>   BUG: sleeping function called from invalid context at 
> ./include/linux/percpu-rwsem.h:49
> 
>    cpus_read_lock+0x68/0x130
>    static_key_enable+0x1c/0x38
>    __sched_setscheduler+0x900/0xad8
> 
> Fix by ensuring we enable the key outside of the critical section in
> __sched_setscheduler()
> 
> Fixes: 46609ce22703 ("sched/uclamp: Protect uclamp fast path code with static 
> key")
> Signed-off-by: Qais Yousef <[email protected]>
> Signed-off-by: Peter Zijlstra (Intel) <[email protected]>
> Link: https://lkml.kernel.org/r/[email protected]
> ---
> 
> And placed this patch first in the series
> 
> That said; don't we have a slight problem with enabling the key late
> like this? It means the uclamp will not actually take effect immediately
> and we'll have to wait for the next context switch ... whenever that
> might be.

The enabling is racy inherently. Your suggestion below is better though.
I should have scrolled more screens up!

> 
> Should we not have enabled the key early, instead of late?
> 
> something like so perhaps?

This should work, but you'll need to sprinkle ifdef around the key. Or move it
to uclamp_validate()

---
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index e1578c3ad40c..cd3b7a25ac59 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -1345,6 +1345,21 @@ static int uclamp_validate(struct task_struct *p,
        if (upper_bound > SCHED_CAPACITY_SCALE)
                return -EINVAL;
 
+       /*
+        * Enable uclamp static key outside the critical section of
+        * __sched_setschuler().
+        *
+        * static_branch_enable() will hold cpu_hotplug_lock; if done from
+        * critical section, __setscheduler_uclamp(), which holds other locks
+        * (rq->lock namely), it could lead to deadlock scenarios as both are
+        * popular locks and could be acquired from different paths in
+        * different orders.
+        *
+        * Besides cpu_hotplug_lock could sleep, which is not allowed inside
+        * the critical section of __sched_setscheduler().
+        */
+       static_branch_enable(&sched_uclamp_used);
+
        return 0;
 }
 
@@ -1378,8 +1393,6 @@ static void __setscheduler_uclamp(struct task_struct *p,
        if (likely(!(attr->sched_flags & SCHED_FLAG_UTIL_CLAMP)))
                return;
 
-       static_branch_enable(&sched_uclamp_used);
-
        if (attr->sched_flags & SCHED_FLAG_UTIL_CLAMP_MIN) {
                uclamp_se_set(&p->uclamp_req[UCLAMP_MIN],
                              attr->sched_util_min, true);

Reply via email to