Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-20 Thread Chen Yu
On 2024-06-19 at 00:03:30 +0530, K Prateek Nayak wrote:
> Hello Chenyu,
> 
> On 6/18/2024 1:19 PM, Chen Yu wrote:
> > [..snip..]
> > > > > > 
> > > > > > > Vincent [5] pointed out a case where the idle load kick will fail 
> > > > > > > to
> > > > > > > run on an idle CPU since the IPI handler launching the ILB will 
> > > > > > > check
> > > > > > > for need_resched(). In such cases, the idle CPU relies on
> > > > > > > newidle_balance() to pull tasks towards itself.
> > > > > > 
> > > > > > Is this the need_resched() in _nohz_idle_balance() ? Should we 
> > > > > > change
> > > > > > this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or
> > > > > > something long those lines?
> > > > > 
> > > > > It's not only this but also in do_idle() as well which exits the loop
> > > > > to look for tasks to schedule
> > > > > 
> > > > > > 
> > > > > > I mean, it's fairly trivial to figure out if there really is going 
> > > > > > to be
> > > > > > work there.
> > > > > > 
> > > > > > > Using an alternate flag instead of NEED_RESCHED to indicate a 
> > > > > > > pending
> > > > > > > IPI was suggested as the correct approach to solve this problem 
> > > > > > > on the
> > > > > > > same thread.
> > > > > > 
> > > > > > So adding per-arch changes for this seems like something we 
> > > > > > shouldn't
> > > > > > unless there really is no other sane options.
> > > > > > 
> > > > > > That is, I really think we should start with something like the 
> > > > > > below
> > > > > > and then fix any fallout from that.
> > > > > 
> > > > > The main problem is that need_resched becomes somewhat meaningless
> > > > > because it doesn't  only mean "I need to resched a task" and we have
> > > > > to add more tests around even for those not using polling
> > > > > 
> > > > > > 
> > > > > > diff --git a/kernel/sched/core.c b/kernel/sched/core.c
> > > > > > index 0935f9d4bb7b..cfa45338ae97 100644
> > > > > > --- a/kernel/sched/core.c
> > > > > > +++ b/kernel/sched/core.c
> > > > > > @@ -5799,7 +5800,7 @@ static inline struct task_struct *
> > > > > >__pick_next_task(struct rq *rq, struct task_struct *prev, struct 
> > > > > > rq_flags *rf)
> > > > > >{
> > > > > >   const struct sched_class *class;
> > > > > > -   struct task_struct *p;
> > > > > > +   struct task_struct *p = NULL;
> > > > > > 
> > > > > >   /*
> > > > > >* Optimization: we know that if all tasks are in the 
> > > > > > fair class we can
> > > > > > @@ -5810,9 +5811,11 @@ __pick_next_task(struct rq *rq, struct 
> > > > > > task_struct *prev, struct rq_flags *rf)
> > > > > >   if (likely(!sched_class_above(prev->sched_class, 
> > > > > > _sched_class) &&
> > > > > >  rq->nr_running == rq->cfs.h_nr_running)) {
> > > > > > 
> > > > > > -   p = pick_next_task_fair(rq, prev, rf);
> > > > > > -   if (unlikely(p == RETRY_TASK))
> > > > > > -   goto restart;
> > > > > > +   if (rq->nr_running) {
> > > > > 
> > > > > How do you make the diff between a spurious need_resched() because of
> > > > > polling and a cpu becoming idle ? isn't rq->nr_running null in both
> > > > > cases ?
> > > > > In the later case, we need to call sched_balance_newidle() but not in 
> > > > > the former
> > > > > 
> > > > 
> > > > Not sure if I understand correctly, if the goal of 
> > > > smp_call_function_single() is to
> > > > kick the idle CPU and do not force it to launch the 
> > > > schedule()->sched_balance_newidle(),
> > > > can we set the _TIF_POLLING_NRFLAG rather than _TIF_NEED_RESCHED in 
> > > > set_nr_if_polling()?
> > > > I think writing any value to the monitor address would wakeup the idle 
> > > > CPU. And _TIF_POLLING_NRFLAG
> > > > will be cleared once that idle CPU exit the idle loop, so we don't 
> > > > introduce arch-wide flag.
> > > Although this might work for MWAIT, there is no way for the generic idle
> > > path to know if there is a pending interrupt within a TIF_POLLING_NRFLAG
> > > section. do_idle() sets TIF_POLLING_NRFLAG and relies on a bunch of
> > > need_resched() checks along the way to bail early until finally doing a
> > > current_clr_polling_and_test() before handing off to the cpuidle driver
> > > in call_cpuidle(). I believe this section will necessarily need the sender
> > > to indicate a pending interrupt via TIF_NEED_RESCHED flag to enable the
> > > early bail out before going into the cpuidle driver since this case cannot
> > > be considered the same as a break from MWAIT.
> > > 
> > 
> > I see, this is a good point. So you mean with only TIF_POLLING_NRFLAG there 
> > is
> > possibility that the 'ipi kick CPU out of idle' is lost after the CPU enters
> > do_idle() and before finally entering the idle state. While setting 
> > _TIF_NEED_RESCHED
> > could help the do_idle() loop to detect pending request easier.
> 
> Yup, that is correct.
> 
> > BTW, before the
> > commit b2a02fc43a1f ("smp: Optimize 

Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-18 Thread K Prateek Nayak

Hello Chenyu,

On 6/18/2024 1:19 PM, Chen Yu wrote:

[..snip..]



Vincent [5] pointed out a case where the idle load kick will fail to
run on an idle CPU since the IPI handler launching the ILB will check
for need_resched(). In such cases, the idle CPU relies on
newidle_balance() to pull tasks towards itself.


Is this the need_resched() in _nohz_idle_balance() ? Should we change
this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or
something long those lines?


It's not only this but also in do_idle() as well which exits the loop
to look for tasks to schedule



I mean, it's fairly trivial to figure out if there really is going to be
work there.


Using an alternate flag instead of NEED_RESCHED to indicate a pending
IPI was suggested as the correct approach to solve this problem on the
same thread.


So adding per-arch changes for this seems like something we shouldn't
unless there really is no other sane options.

That is, I really think we should start with something like the below
and then fix any fallout from that.


The main problem is that need_resched becomes somewhat meaningless
because it doesn't  only mean "I need to resched a task" and we have
to add more tests around even for those not using polling



diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index 0935f9d4bb7b..cfa45338ae97 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -5799,7 +5800,7 @@ static inline struct task_struct *
   __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags 
*rf)
   {
  const struct sched_class *class;
-   struct task_struct *p;
+   struct task_struct *p = NULL;

  /*
   * Optimization: we know that if all tasks are in the fair class we 
can
@@ -5810,9 +5811,11 @@ __pick_next_task(struct rq *rq, struct task_struct 
*prev, struct rq_flags *rf)
  if (likely(!sched_class_above(prev->sched_class, _sched_class) &&
 rq->nr_running == rq->cfs.h_nr_running)) {

-   p = pick_next_task_fair(rq, prev, rf);
-   if (unlikely(p == RETRY_TASK))
-   goto restart;
+   if (rq->nr_running) {


How do you make the diff between a spurious need_resched() because of
polling and a cpu becoming idle ? isn't rq->nr_running null in both
cases ?
In the later case, we need to call sched_balance_newidle() but not in the former



Not sure if I understand correctly, if the goal of smp_call_function_single() 
is to
kick the idle CPU and do not force it to launch the 
schedule()->sched_balance_newidle(),
can we set the _TIF_POLLING_NRFLAG rather than _TIF_NEED_RESCHED in 
set_nr_if_polling()?
I think writing any value to the monitor address would wakeup the idle CPU. And 
_TIF_POLLING_NRFLAG
will be cleared once that idle CPU exit the idle loop, so we don't introduce 
arch-wide flag.

Although this might work for MWAIT, there is no way for the generic idle
path to know if there is a pending interrupt within a TIF_POLLING_NRFLAG
section. do_idle() sets TIF_POLLING_NRFLAG and relies on a bunch of
need_resched() checks along the way to bail early until finally doing a
current_clr_polling_and_test() before handing off to the cpuidle driver
in call_cpuidle(). I believe this section will necessarily need the sender
to indicate a pending interrupt via TIF_NEED_RESCHED flag to enable the
early bail out before going into the cpuidle driver since this case cannot
be considered the same as a break from MWAIT.



I see, this is a good point. So you mean with only TIF_POLLING_NRFLAG there is
possibility that the 'ipi kick CPU out of idle' is lost after the CPU enters
do_idle() and before finally entering the idle state. While setting 
_TIF_NEED_RESCHED
could help the do_idle() loop to detect pending request easier.


Yup, that is correct.


BTW, before the
commit b2a02fc43a1f ("smp: Optimize send_call_function_single_ipi()"), the
lost of ipi after entering do_idle() and before entering driver idle state
is also possible, right(the local irq is disabled)?


From what I understand, the IPI remains pending until the interrupts
are enabled again. Before the optimization, the interrupts would be
disabled all the way until the instruction that is used to put the CPU
to sleep which is what __sti_mwait() and native_safe_halt() does. The
CPU would have received the IPI then and broke out of idle before
Peter's optimization went in. There is an elaborate comment on this in
do_idle() function above the call to local_irq_disable(). In  commit
edc8fc01f608 ("x86: Fix CPUIDLE_FLAG_IRQ_ENABLE leaking timer
reprogram") Peter describes a case of actually missing the break from
an interrupt as the driver enabled interrupts much earlier than
executing the sleep instruction.

Since the CPU was in TIF_POLLING_NRFLAG state, one could simply get away
by setting TIF_NEED_RESCHED and not sending an actual IPI which the
need_resched() checks in the idle path would catch and the
flush_smp_call_function_queue() on 

Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-18 Thread Chen Yu
On 2024-06-17 at 14:03:41 +0530, K Prateek Nayak wrote:
> Hello Chenyu,
> 
> On 6/14/2024 10:01 PM, Chen Yu wrote:
> > On 2024-06-14 at 12:48:37 +0200, Vincent Guittot wrote:
> > > On Fri, 14 Jun 2024 at 11:28, Peter Zijlstra  wrote:
> > > > 
> > > > On Thu, Jun 13, 2024 at 06:15:59PM +, K Prateek Nayak wrote:
> > > > > Effects of call_function_single_prep_ipi()
> > > > > ==
> > > > > 
> > > > > To pull a TIF_POLLING thread out of idle to process an IPI, the sender
> > > > > sets the TIF_NEED_RESCHED bit in the idle task's thread info in
> > > > > call_function_single_prep_ipi() and avoids sending an actual IPI to 
> > > > > the
> > > > > target. As a result, the scheduler expects a task to be enqueued when
> > > > > exiting the idle path. This is not the case with non-polling idle 
> > > > > states
> > > > > where the idle CPU exits the non-polling idle state to process the
> > > > > interrupt, and since need_resched() returns false, soon goes back to
> > > > > idle again.
> > > > > 
> > > > > When TIF_NEED_RESCHED flag is set, do_idle() will call 
> > > > > schedule_idle(),
> > > > > a large part of which runs with local IRQ disabled. In case of 
> > > > > ipistorm,
> > > > > when measuring IPI throughput, this large IRQ disabled section delays
> > > > > processing of IPIs. Further auditing revealed that in absence of any
> > > > > runnable tasks, pick_next_task_fair(), which is called from the
> > > > > pick_next_task() fast path, will always call newidle_balance() in this
> > > > > scenario, further increasing the time spent in the IRQ disabled 
> > > > > section.
> > > > > 
> > > > > Following is the crude visualization of the problem with relevant
> > > > > functions expanded:
> > > > > --
> > > > > CPU0  CPU1
> > > > >   
> > > > >do_idle() {
> > > > >
> > > > > __current_set_polling();
> > > > >...
> > > > >
> > > > > monitor(addr);
> > > > >if 
> > > > > (!need_resched())
> > > > >   
> > > > >  mwait() {
> > > > >   
> > > > >  /* Waiting */
> > > > > smp_call_function_single(CPU1, func, wait = 1) {  
> > > > > ...
> > > > >...
> > > > >  ...
> > > > >set_nr_if_polling(CPU1) {  
> > > > >  ...
> > > > >/* Realizes CPU1 is polling */ 
> > > > >  ...
> > > > >try_cmpxchg(addr,  
> > > > >  ...
> > > > >,  
> > > > >  ...
> > > > >val | _TIF_NEED_RESCHED);  
> > > > >  ...
> > > > >} /* Does not send an IPI */   
> > > > >  ...
> > > > >...
> > > > >  } /* mwait exit due to write at addr */
> > > > >csd_lock_wait() {   }
> > > > >/* Waiting */   
> > > > > preempt_set_need_resched();
> > > > >... 
> > > > > __current_clr_polling();
> > > > >... 
> > > > > flush_smp_call_function_queue() {
> > > > >...
> > > > >  func();
> > > > >} /* End of wait */ }
> > > > > } 
> > > > > schedule_idle() {
> > > > >   
> > > > >  ...
> > > > >   
> > > > >  local_irq_disable();
> > > > > smp_call_function_single(CPU1, func, wait = 1) {  
> > > > > ...
> > > > >...
> > > > >  ...
> > > > >arch_send_call_function_single_ipi(CPU1);  
> > > > >  ...
> > > > >\  
> > > > >  ...
> > > > > \ 
> > > > >  newidle_balance() {
> > > > >  \

Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-17 Thread K Prateek Nayak

Hello Chenyu,

On 6/14/2024 10:01 PM, Chen Yu wrote:

On 2024-06-14 at 12:48:37 +0200, Vincent Guittot wrote:

On Fri, 14 Jun 2024 at 11:28, Peter Zijlstra  wrote:


On Thu, Jun 13, 2024 at 06:15:59PM +, K Prateek Nayak wrote:

Effects of call_function_single_prep_ipi()
==

To pull a TIF_POLLING thread out of idle to process an IPI, the sender
sets the TIF_NEED_RESCHED bit in the idle task's thread info in
call_function_single_prep_ipi() and avoids sending an actual IPI to the
target. As a result, the scheduler expects a task to be enqueued when
exiting the idle path. This is not the case with non-polling idle states
where the idle CPU exits the non-polling idle state to process the
interrupt, and since need_resched() returns false, soon goes back to
idle again.

When TIF_NEED_RESCHED flag is set, do_idle() will call schedule_idle(),
a large part of which runs with local IRQ disabled. In case of ipistorm,
when measuring IPI throughput, this large IRQ disabled section delays
processing of IPIs. Further auditing revealed that in absence of any
runnable tasks, pick_next_task_fair(), which is called from the
pick_next_task() fast path, will always call newidle_balance() in this
scenario, further increasing the time spent in the IRQ disabled section.

Following is the crude visualization of the problem with relevant
functions expanded:
--
CPU0  CPU1
  
   do_idle() {
   
__current_set_polling();
   ...
   monitor(addr);
   if 
(!need_resched())
   mwait() {
   /* 
Waiting */
smp_call_function_single(CPU1, func, wait = 1) {  
...
   ... 
...
   set_nr_if_polling(CPU1) {   
...
   /* Realizes CPU1 is polling */  
...
   try_cmpxchg(addr,   
...
   ,   
...
   val | _TIF_NEED_RESCHED);   
...
   } /* Does not send an IPI */
...
   ... } /* 
mwait exit due to write at addr */
   csd_lock_wait() {   }
   /* Waiting */   
preempt_set_need_resched();
   ... 
__current_clr_polling();
   ... 
flush_smp_call_function_queue() {
   ... func();
   } /* End of wait */ }
} schedule_idle() {
   ...
   
local_irq_disable();
smp_call_function_single(CPU1, func, wait = 1) {  ...
   ... ...
   arch_send_call_function_single_ipi(CPU1);   ...
   \   ...
\  
newidle_balance() {
 \ 
...
 /* Delay */   
...
   \   }
\  ...
 \-->  
local_irq_enable();
   /* 
Processes the IPI */
--


Skipping newidle_balance()
==

In an earlier attempt to solve the challenge of the long IRQ disabled
section, newidle_balance() was skipped when a CPU waking up from idle
was found to have no runnable tasks, and was transitioning back to
idle [2]. Tim [3] and David [4] had pointed out that newidle_balance()
may be viable for CPUs that are idling with tick enabled, where the
newidle_balance() has the opportunity to pull tasks onto the idle CPU.


I don't think we should be relying on 

Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-17 Thread K Prateek Nayak

Hello Vincent, Peter,

On 6/16/2024 8:27 PM, Vincent Guittot wrote:

On Sat, 15 Jun 2024 at 03:28, Peter Zijlstra  wrote:


On Fri, Jun 14, 2024 at 12:48:37PM +0200, Vincent Guittot wrote:

On Fri, 14 Jun 2024 at 11:28, Peter Zijlstra  wrote:



Vincent [5] pointed out a case where the idle load kick will fail to
run on an idle CPU since the IPI handler launching the ILB will check
for need_resched(). In such cases, the idle CPU relies on
newidle_balance() to pull tasks towards itself.


Is this the need_resched() in _nohz_idle_balance() ? Should we change
this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or
something long those lines?


It's not only this but also in do_idle() as well which exits the loop
to look for tasks to schedule


Is that really a problem? Reading the initial email the problem seems to
be newidle balance, not hitting schedule. Schedule should be fairly
quick if there's nothing to do, no?


There are 2 problems:
- Because of NEED_RESCHED being set, we go through the full schedule
path for no reason and we finally do a sched_balance_newidle()


Peter's patch up in the thread seems to improve the above case by
speeding up the schedule() loop similar to the very first solution
I tried with
https://lore.kernel.org/lkml/20240119084548.2788-1-kprateek.na...@amd.com/

I do see same level of improvements (if not better) with Peter's
SM_IDLE solution:

  ==
  Test  : ipistorm (modified)
  Units : Normalized runtime
  Interpretation: Lower is better
  Statistic : AMean
  ==
  kernel:   time [pct imp]
  tip:sched/core1.00 [baseline]
  tip:sched/core + revert   0.40 [60.26%]
  tip:sched/core + TIF_NOTIFY_IPI   0.46 [54.88%]
  tip:sched/core + SM_IDLE  0.38 [72.64%]


- Because of need_resched being set o wake up the cpu, we will not
kick the softirq to run the nohz idle load balance and get a chance to
pull a task on an idle CPU


However, this issues with need_resched() still remains. Any
need_resched() check within an interrupt context will return true if the
target CPU is perceived to be in a polling idle state by the sender as a
result of the optimization in commit b2a02fc43a1f ("smp: Optimize
send_call_function_single_ipi()").

If TIF_POLLING_NRFLAG is defined by an arch, do_idle() will set the
flag until the path hits call_cpuidle() where the flag is cleared just
before handing off the state entry to the cpuidle driver. An incoming
interrupt in this window will allow the idle path to bail early and
return before calling the driver specific routine since it'll be
indicated by TIF_NEED_RESCHED being set in the idle task's thread info.
Beyond that point, the cpuidle driver handles the idle entry.

I think an arch may define TIF_POLLING_NRFLAG just to utilize this
optimization in the generic idle path to answer Vincent's observation
on ARM32 having TIF_POLLING_NRFLAG.






I mean, it's fairly trivial to figure out if there really is going to be
work there.


Using an alternate flag instead of NEED_RESCHED to indicate a pending
IPI was suggested as the correct approach to solve this problem on the
same thread.


So adding per-arch changes for this seems like something we shouldn't
unless there really is no other sane options.

That is, I really think we should start with something like the below
and then fix any fallout from that.


The main problem is that need_resched becomes somewhat meaningless
because it doesn't  only mean "I need to resched a task" and we have
to add more tests around even for those not using polling


True, however we already had some of that by having the wakeup list,
that made nr_running less 'reliable'.

The thing is, most architectures seem to have the TIF_POLLING_NRFLAG
bit, even if their main idle routine isn't actually using it, much of


Yes, I'm surprised that Arm arch has the TIF_POLLING_NRFLAG whereas it
has never been supported by the arch


the idle loop until it hits the arch idle will be having it set and will
thus tickle these cases *sometimes*.
[..snip..]


--
Thanks and Regards,
Prateek


Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-17 Thread K Prateek Nayak

Hello Russell,

On 6/15/2024 7:56 PM, Russell King (Oracle) wrote:

On Thu, Jun 13, 2024 at 06:15:59PM +, K Prateek Nayak wrote:

o Dropping the ARM results since I never got my hands on the ARM64
   system I used in my last testing. If I do manage to get my hands on it
   again, I'll rerun the experiments and share the results on the thread.
   To test the case where TIF_NOTIFY_IPI is not enabled for a particular
   architecture, I applied the series only until Patch 3 and tested the
   same on my x86 machine with a WARN_ON_ONCE() in do_idle() to check if
   tif_notify_ipi() ever return true and then repeated the same with
   Patch 4 applied.


Confused. ARM (32-bit) or ARM64? You patch 32-bit ARM, but you don't
touch 64-bit Arm. "ARM" on its own in the context above to me suggests
32-bit, since you refer to ARM64 later.



In my first RFC posting, I had shared the results for ipistorm on an
ARM64 server [1]. Vincent and Linus Walleij brought to my attention that
ARM32 and ARM64 do not share the thread info flags and I probably saw a
one-off behavior during my testing. Since then, it has been slightly
challenging to get my hands on that machine again in a stable condition
to see if there was any scenario that I might have missed but I tried a
bunch of things on my x86 machine to confirm that an arch that does not
define the TIF_NOTIFY_IPI would not hit these changes.

Rest assured, Patch 5 is for ARM32 machines that currently define
TIF_POLLING_NRFLAG

[1] https://lore.kernel.org/lkml/20240220171457.703-6-kprateek.na...@amd.com/

--
Thanks and Regards,
Prateek


Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-17 Thread Vincent Guittot
On Sat, 15 Jun 2024 at 03:28, Peter Zijlstra  wrote:
>
> On Fri, Jun 14, 2024 at 12:48:37PM +0200, Vincent Guittot wrote:
> > On Fri, 14 Jun 2024 at 11:28, Peter Zijlstra  wrote:
>
> > > > Vincent [5] pointed out a case where the idle load kick will fail to
> > > > run on an idle CPU since the IPI handler launching the ILB will check
> > > > for need_resched(). In such cases, the idle CPU relies on
> > > > newidle_balance() to pull tasks towards itself.
> > >
> > > Is this the need_resched() in _nohz_idle_balance() ? Should we change
> > > this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or
> > > something long those lines?
> >
> > It's not only this but also in do_idle() as well which exits the loop
> > to look for tasks to schedule
>
> Is that really a problem? Reading the initial email the problem seems to
> be newidle balance, not hitting schedule. Schedule should be fairly
> quick if there's nothing to do, no?

There are 2 problems:
- Because of NEED_RESCHED being set, we go through the full schedule
path for no reason and we finally do a sched_balance_newidle()
- Because of need_resched being set o wake up the cpu, we will not
kick the softirq to run the nohz idle load balance and get a chance to
pull a task on an idle CPU

>
> > > I mean, it's fairly trivial to figure out if there really is going to be
> > > work there.
> > >
> > > > Using an alternate flag instead of NEED_RESCHED to indicate a pending
> > > > IPI was suggested as the correct approach to solve this problem on the
> > > > same thread.
> > >
> > > So adding per-arch changes for this seems like something we shouldn't
> > > unless there really is no other sane options.
> > >
> > > That is, I really think we should start with something like the below
> > > and then fix any fallout from that.
> >
> > The main problem is that need_resched becomes somewhat meaningless
> > because it doesn't  only mean "I need to resched a task" and we have
> > to add more tests around even for those not using polling
>
> True, however we already had some of that by having the wakeup list,
> that made nr_running less 'reliable'.
>
> The thing is, most architectures seem to have the TIF_POLLING_NRFLAG
> bit, even if their main idle routine isn't actually using it, much of

Yes, I'm surprised that Arm arch has the TIF_POLLING_NRFLAG whereas it
has never been supported by the arch

> the idle loop until it hits the arch idle will be having it set and will
> thus tickle these cases *sometimes*.
>
> > > diff --git a/kernel/sched/core.c b/kernel/sched/core.c
> > > index 0935f9d4bb7b..cfa45338ae97 100644
> > > --- a/kernel/sched/core.c
> > > +++ b/kernel/sched/core.c
> > > @@ -5799,7 +5800,7 @@ static inline struct task_struct *
> > >  __pick_next_task(struct rq *rq, struct task_struct *prev, struct 
> > > rq_flags *rf)
> > >  {
> > > const struct sched_class *class;
> > > -   struct task_struct *p;
> > > +   struct task_struct *p = NULL;
> > >
> > > /*
> > >  * Optimization: we know that if all tasks are in the fair class 
> > > we can
> > > @@ -5810,9 +5811,11 @@ __pick_next_task(struct rq *rq, struct task_struct 
> > > *prev, struct rq_flags *rf)
> > > if (likely(!sched_class_above(prev->sched_class, 
> > > _sched_class) &&
> > >rq->nr_running == rq->cfs.h_nr_running)) {
> > >
> > > -   p = pick_next_task_fair(rq, prev, rf);
> > > -   if (unlikely(p == RETRY_TASK))
> > > -   goto restart;
> > > +   if (rq->nr_running) {
> >
> > How do you make the diff between a spurious need_resched() because of
> > polling and a cpu becoming idle ? isn't rq->nr_running null in both
> > cases ?
>
> Bah, true. It should also check current being idle, which then makes a
> mess of things again. Still, we shouldn't be calling newidle from idle,
> that's daft.
>
> I should probably not write code at 3am, but the below horror is what
> I came up with.
>
> ---
>
> diff --git a/kernel/sched/core.c b/kernel/sched/core.c
> index 0935f9d4bb7b..cfe8d3350819 100644
> --- a/kernel/sched/core.c
> +++ b/kernel/sched/core.c
> @@ -6343,19 +6344,12 @@ pick_next_task(struct rq *rq, struct task_struct 
> *prev, struct rq_flags *rf)
>   * Constants for the sched_mode argument of __schedule().
>   *
>   * The mode argument allows RT enabled kernels to differentiate a
> - * preemption from blocking on an 'sleeping' spin/rwlock. Note that
> - * SM_MASK_PREEMPT for !RT has all bits set, which allows the compiler to
> - * optimize the AND operation out and just check for zero.
> + * preemption from blocking on an 'sleeping' spin/rwlock.
>   */
> -#define SM_NONE0x0
> -#define SM_PREEMPT 0x1
> -#define SM_RTLOCK_WAIT 0x2
> -
> -#ifndef CONFIG_PREEMPT_RT
> -# define SM_MASK_PREEMPT   (~0U)
> -#else
> -# define SM_MASK_PREEMPT   SM_PREEMPT
> -#endif
> +#define SM_IDLE(-1)
> +#define SM_NONE 

Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-15 Thread Russell King (Oracle)
On Thu, Jun 13, 2024 at 06:15:59PM +, K Prateek Nayak wrote:
> o Dropping the ARM results since I never got my hands on the ARM64
>   system I used in my last testing. If I do manage to get my hands on it
>   again, I'll rerun the experiments and share the results on the thread.
>   To test the case where TIF_NOTIFY_IPI is not enabled for a particular
>   architecture, I applied the series only until Patch 3 and tested the
>   same on my x86 machine with a WARN_ON_ONCE() in do_idle() to check if
>   tif_notify_ipi() ever return true and then repeated the same with
>   Patch 4 applied.

Confused. ARM (32-bit) or ARM64? You patch 32-bit ARM, but you don't
touch 64-bit Arm. "ARM" on its own in the context above to me suggests
32-bit, since you refer to ARM64 later.

-- 
RMK's Patch system: https://www.armlinux.org.uk/developer/patches/
FTTP is here! 80Mbps down 10Mbps up. Decent connectivity at last!


Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-15 Thread Peter Zijlstra
On Sat, Jun 15, 2024 at 03:28:14AM +0200, Peter Zijlstra wrote:
> On Fri, Jun 14, 2024 at 12:48:37PM +0200, Vincent Guittot wrote:

> > The main problem is that need_resched becomes somewhat meaningless
> > because it doesn't  only mean "I need to resched a task" and we have
> > to add more tests around even for those not using polling
> 
> True, however we already had some of that by having the wakeup list,
> that made nr_running less 'reliable'.

Doesn't using !idle_cpu() instead of need_resched() in those balance
paths already do the right thing?

Checking need_resched() as an indicator of it getting work is already a
bit an assumption.

Also, Ingo, idle_cpu() and friends don't really belong in syscalls.c...


Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-15 Thread Peter Zijlstra
On Fri, Jun 14, 2024 at 12:48:37PM +0200, Vincent Guittot wrote:

> The main problem is that need_resched becomes somewhat meaningless
> because it doesn't  only mean "I need to resched a task" and we have
> to add more tests around even for those not using polling

The converse problem is that you're adding a bunch of atomic ops that
might be avoided.

It might now need to set both the RESCHED and IPI flags -- and clear
them again.



Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-15 Thread Peter Zijlstra
On Fri, Jun 14, 2024 at 12:48:37PM +0200, Vincent Guittot wrote:
> On Fri, 14 Jun 2024 at 11:28, Peter Zijlstra  wrote:

> > > Vincent [5] pointed out a case where the idle load kick will fail to
> > > run on an idle CPU since the IPI handler launching the ILB will check
> > > for need_resched(). In such cases, the idle CPU relies on
> > > newidle_balance() to pull tasks towards itself.
> >
> > Is this the need_resched() in _nohz_idle_balance() ? Should we change
> > this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or
> > something long those lines?
> 
> It's not only this but also in do_idle() as well which exits the loop
> to look for tasks to schedule

Is that really a problem? Reading the initial email the problem seems to
be newidle balance, not hitting schedule. Schedule should be fairly
quick if there's nothing to do, no?

> > I mean, it's fairly trivial to figure out if there really is going to be
> > work there.
> >
> > > Using an alternate flag instead of NEED_RESCHED to indicate a pending
> > > IPI was suggested as the correct approach to solve this problem on the
> > > same thread.
> >
> > So adding per-arch changes for this seems like something we shouldn't
> > unless there really is no other sane options.
> >
> > That is, I really think we should start with something like the below
> > and then fix any fallout from that.
> 
> The main problem is that need_resched becomes somewhat meaningless
> because it doesn't  only mean "I need to resched a task" and we have
> to add more tests around even for those not using polling

True, however we already had some of that by having the wakeup list,
that made nr_running less 'reliable'.

The thing is, most architectures seem to have the TIF_POLLING_NRFLAG
bit, even if their main idle routine isn't actually using it, much of
the idle loop until it hits the arch idle will be having it set and will
thus tickle these cases *sometimes*.

> > diff --git a/kernel/sched/core.c b/kernel/sched/core.c
> > index 0935f9d4bb7b..cfa45338ae97 100644
> > --- a/kernel/sched/core.c
> > +++ b/kernel/sched/core.c
> > @@ -5799,7 +5800,7 @@ static inline struct task_struct *
> >  __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags 
> > *rf)
> >  {
> > const struct sched_class *class;
> > -   struct task_struct *p;
> > +   struct task_struct *p = NULL;
> >
> > /*
> >  * Optimization: we know that if all tasks are in the fair class we 
> > can
> > @@ -5810,9 +5811,11 @@ __pick_next_task(struct rq *rq, struct task_struct 
> > *prev, struct rq_flags *rf)
> > if (likely(!sched_class_above(prev->sched_class, _sched_class) 
> > &&
> >rq->nr_running == rq->cfs.h_nr_running)) {
> >
> > -   p = pick_next_task_fair(rq, prev, rf);
> > -   if (unlikely(p == RETRY_TASK))
> > -   goto restart;
> > +   if (rq->nr_running) {
> 
> How do you make the diff between a spurious need_resched() because of
> polling and a cpu becoming idle ? isn't rq->nr_running null in both
> cases ?

Bah, true. It should also check current being idle, which then makes a
mess of things again. Still, we shouldn't be calling newidle from idle,
that's daft.

I should probably not write code at 3am, but the below horror is what
I came up with.

---

diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index 0935f9d4bb7b..cfe8d3350819 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -6343,19 +6344,12 @@ pick_next_task(struct rq *rq, struct task_struct *prev, 
struct rq_flags *rf)
  * Constants for the sched_mode argument of __schedule().
  *
  * The mode argument allows RT enabled kernels to differentiate a
- * preemption from blocking on an 'sleeping' spin/rwlock. Note that
- * SM_MASK_PREEMPT for !RT has all bits set, which allows the compiler to
- * optimize the AND operation out and just check for zero.
+ * preemption from blocking on an 'sleeping' spin/rwlock.
  */
-#define SM_NONE0x0
-#define SM_PREEMPT 0x1
-#define SM_RTLOCK_WAIT 0x2
-
-#ifndef CONFIG_PREEMPT_RT
-# define SM_MASK_PREEMPT   (~0U)
-#else
-# define SM_MASK_PREEMPT   SM_PREEMPT
-#endif
+#define SM_IDLE(-1)
+#define SM_NONE0
+#define SM_PREEMPT 1
+#define SM_RTLOCK_WAIT 2
 
 /*
  * __schedule() is the main scheduler function.
@@ -6396,11 +6390,12 @@ pick_next_task(struct rq *rq, struct task_struct *prev, 
struct rq_flags *rf)
  *
  * WARNING: must be called with preemption disabled!
  */
-static void __sched notrace __schedule(unsigned int sched_mode)
+static void __sched notrace __schedule(int sched_mode)
 {
struct task_struct *prev, *next;
unsigned long *switch_count;
unsigned long prev_state;
+   bool preempt = sched_mode > 0;
struct rq_flags rf;
struct rq *rq;
int cpu;
@@ -6409,13 +6404,13 @@ static void __sched 

Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-14 Thread Chen Yu
On 2024-06-14 at 12:48:37 +0200, Vincent Guittot wrote:
> On Fri, 14 Jun 2024 at 11:28, Peter Zijlstra  wrote:
> >
> > On Thu, Jun 13, 2024 at 06:15:59PM +, K Prateek Nayak wrote:
> > > Effects of call_function_single_prep_ipi()
> > > ==
> > >
> > > To pull a TIF_POLLING thread out of idle to process an IPI, the sender
> > > sets the TIF_NEED_RESCHED bit in the idle task's thread info in
> > > call_function_single_prep_ipi() and avoids sending an actual IPI to the
> > > target. As a result, the scheduler expects a task to be enqueued when
> > > exiting the idle path. This is not the case with non-polling idle states
> > > where the idle CPU exits the non-polling idle state to process the
> > > interrupt, and since need_resched() returns false, soon goes back to
> > > idle again.
> > >
> > > When TIF_NEED_RESCHED flag is set, do_idle() will call schedule_idle(),
> > > a large part of which runs with local IRQ disabled. In case of ipistorm,
> > > when measuring IPI throughput, this large IRQ disabled section delays
> > > processing of IPIs. Further auditing revealed that in absence of any
> > > runnable tasks, pick_next_task_fair(), which is called from the
> > > pick_next_task() fast path, will always call newidle_balance() in this
> > > scenario, further increasing the time spent in the IRQ disabled section.
> > >
> > > Following is the crude visualization of the problem with relevant
> > > functions expanded:
> > > --
> > > CPU0  CPU1
> > >   
> > >   do_idle() {
> > >   
> > > __current_set_polling();
> > >   ...
> > >   
> > > monitor(addr);
> > >   if 
> > > (!need_resched())
> > >   
> > > mwait() {
> > >   /* 
> > > Waiting */
> > > smp_call_function_single(CPU1, func, wait = 1) {  
> > > ...
> > >   ... 
> > > ...
> > >   set_nr_if_polling(CPU1) {   
> > > ...
> > >   /* Realizes CPU1 is polling */  
> > > ...
> > >   try_cmpxchg(addr,   
> > > ...
> > >   ,   
> > > ...
> > >   val | _TIF_NEED_RESCHED);   
> > > ...
> > >   } /* Does not send an IPI */
> > > ...
> > >   ... } 
> > > /* mwait exit due to write at addr */
> > >   csd_lock_wait() {   }
> > >   /* Waiting */   
> > > preempt_set_need_resched();
> > >   ... 
> > > __current_clr_polling();
> > >   ... 
> > > flush_smp_call_function_queue() {
> > >   ... 
> > > func();
> > >   } /* End of wait */ }
> > > } 
> > > schedule_idle() {
> > >   ...
> > >   
> > > local_irq_disable();
> > > smp_call_function_single(CPU1, func, wait = 1) {  ...
> > >   ... ...
> > >   arch_send_call_function_single_ipi(CPU1);   ...
> > >   \   ...
> > >\  
> > > newidle_balance() {
> > > \ 
> > > ...
> > > /* Delay */   
> > > ...
> > >   \   }
> > >\  ...
> > > \-->  
> > > local_irq_enable();
> > >   /* 
> > > Processes the IPI */
> > > --
> > >
> > >
> > > Skipping newidle_balance()
> > > 

Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-14 Thread Vincent Guittot
On Fri, 14 Jun 2024 at 11:28, Peter Zijlstra  wrote:
>
> On Thu, Jun 13, 2024 at 06:15:59PM +, K Prateek Nayak wrote:
> > Effects of call_function_single_prep_ipi()
> > ==
> >
> > To pull a TIF_POLLING thread out of idle to process an IPI, the sender
> > sets the TIF_NEED_RESCHED bit in the idle task's thread info in
> > call_function_single_prep_ipi() and avoids sending an actual IPI to the
> > target. As a result, the scheduler expects a task to be enqueued when
> > exiting the idle path. This is not the case with non-polling idle states
> > where the idle CPU exits the non-polling idle state to process the
> > interrupt, and since need_resched() returns false, soon goes back to
> > idle again.
> >
> > When TIF_NEED_RESCHED flag is set, do_idle() will call schedule_idle(),
> > a large part of which runs with local IRQ disabled. In case of ipistorm,
> > when measuring IPI throughput, this large IRQ disabled section delays
> > processing of IPIs. Further auditing revealed that in absence of any
> > runnable tasks, pick_next_task_fair(), which is called from the
> > pick_next_task() fast path, will always call newidle_balance() in this
> > scenario, further increasing the time spent in the IRQ disabled section.
> >
> > Following is the crude visualization of the problem with relevant
> > functions expanded:
> > --
> > CPU0  CPU1
> >   
> >   do_idle() {
> >   
> > __current_set_polling();
> >   ...
> >   monitor(addr);
> >   if 
> > (!need_resched())
> >   
> > mwait() {
> >   /* 
> > Waiting */
> > smp_call_function_single(CPU1, func, wait = 1) {
> >   ...
> >   ...   
> >   ...
> >   set_nr_if_polling(CPU1) { 
> >   ...
> >   /* Realizes CPU1 is polling */
> >   ...
> >   try_cmpxchg(addr, 
> >   ...
> >   , 
> >   ...
> >   val | _TIF_NEED_RESCHED); 
> >   ...
> >   } /* Does not send an IPI */  
> >   ...
> >   ... } /* 
> > mwait exit due to write at addr */
> >   csd_lock_wait() {   }
> >   /* Waiting */   
> > preempt_set_need_resched();
> >   ... 
> > __current_clr_polling();
> >   ... 
> > flush_smp_call_function_queue() {
> >   ... 
> > func();
> >   } /* End of wait */ }
> > } 
> > schedule_idle() {
> >   ...
> >   
> > local_irq_disable();
> > smp_call_function_single(CPU1, func, wait = 1) {  ...
> >   ... ...
> >   arch_send_call_function_single_ipi(CPU1);   ...
> >   \   ...
> >\  
> > newidle_balance() {
> > \   
> >   ...
> > /* Delay */ 
> >   ...
> >   \   }
> >\  ...
> > \-->  
> > local_irq_enable();
> >   /* 
> > Processes the IPI */
> > --
> >
> >
> > Skipping newidle_balance()
> > ==
> >
> > In an earlier attempt to solve the challenge of the long IRQ disabled
> > section, newidle_balance() was skipped when a CPU waking up from idle
> > was found to have no runnable tasks, and was transitioning back to
> > idle [2]. Tim 

Re: [PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-14 Thread Peter Zijlstra
On Thu, Jun 13, 2024 at 06:15:59PM +, K Prateek Nayak wrote:
> Effects of call_function_single_prep_ipi()
> ==
> 
> To pull a TIF_POLLING thread out of idle to process an IPI, the sender
> sets the TIF_NEED_RESCHED bit in the idle task's thread info in
> call_function_single_prep_ipi() and avoids sending an actual IPI to the
> target. As a result, the scheduler expects a task to be enqueued when
> exiting the idle path. This is not the case with non-polling idle states
> where the idle CPU exits the non-polling idle state to process the
> interrupt, and since need_resched() returns false, soon goes back to
> idle again.
> 
> When TIF_NEED_RESCHED flag is set, do_idle() will call schedule_idle(),
> a large part of which runs with local IRQ disabled. In case of ipistorm,
> when measuring IPI throughput, this large IRQ disabled section delays
> processing of IPIs. Further auditing revealed that in absence of any
> runnable tasks, pick_next_task_fair(), which is called from the
> pick_next_task() fast path, will always call newidle_balance() in this
> scenario, further increasing the time spent in the IRQ disabled section.
> 
> Following is the crude visualization of the problem with relevant
> functions expanded:
> --
> CPU0  CPU1
>   
>   do_idle() {
>   
> __current_set_polling();
>   ...
>   monitor(addr);
>   if 
> (!need_resched())
>   mwait() 
> {
>   /* 
> Waiting */
> smp_call_function_single(CPU1, func, wait = 1) {  
> ...
>   ... 
> ...
>   set_nr_if_polling(CPU1) {   
> ...
>   /* Realizes CPU1 is polling */  
> ...
>   try_cmpxchg(addr,   
> ...
>   ,   
> ...
>   val | _TIF_NEED_RESCHED);   
> ...
>   } /* Does not send an IPI */
> ...
>   ... } /* 
> mwait exit due to write at addr */
>   csd_lock_wait() {   } 
>   /* Waiting */   
> preempt_set_need_resched();
>   ... 
> __current_clr_polling();
>   ... 
> flush_smp_call_function_queue() {
>   ... func();
>   } /* End of wait */ }
> } schedule_idle() 
> {
>   ...
>   
> local_irq_disable();
> smp_call_function_single(CPU1, func, wait = 1) {  ...
>   ... ...
>   arch_send_call_function_single_ipi(CPU1);   ...
>   \   ...
>\  
> newidle_balance() {
> \ 
> ...
> /* Delay */   
> ...
>   \   }
>\  ...
> \-->  
> local_irq_enable();
>   /* 
> Processes the IPI */
> --
> 
> 
> Skipping newidle_balance()
> ==
> 
> In an earlier attempt to solve the challenge of the long IRQ disabled
> section, newidle_balance() was skipped when a CPU waking up from idle
> was found to have no runnable tasks, and was transitioning back to
> idle [2]. Tim [3] and David [4] had pointed out that newidle_balance()
> may be viable for CPUs that are idling with tick enabled, where the
> newidle_balance() has the opportunity to pull tasks onto the idle CPU.

I don't think we should be relying on this 

[PATCH v2 00/14] Introducing TIF_NOTIFY_IPI flag

2024-06-13 Thread K Prateek Nayak
Hello everyone,

Before jumping into the issue, let me clarify the Cc list. Everyone have
been cc'ed on Patch 0 through Patch 3. Respective arch maintainers,
reviewers, and committers returned by scripts/get_maintainer.pl have
been cc'ed on the respective arch side changes. Scheduler and CPU Idle
maintainers and reviewers have been included for the entire series. If I
have missed anyone, please do add them. If you would like to be dropped
from the cc list, wholly or partially, for the future iterations, please
do let me know.

As long as the first three patches are applied in-order, the arch
specific enablement can be applied independently and out-of-order since
the TIF_NOTIFY_IPI flag is not used until Patch 3 and Patch 2 preps the
complete tree to handle a break out of TIF_POLLING_NRFLAG state with the
setting of either TIF_NOTIFY_IPI or TIF_NEED_RESCHED.

Quick changelog and addressing concerns from v1
===

v1: https://lore.kernel.org/lkml/20240220171457.703-1-kprateek.na...@amd.com/

v1..v2:

o Rebased the series on latest tip:sched/core at commit c793a62823d1
  ("sched/core: Drop spinlocks on contention iff kernel is preemptible")
  Fixed a conflict with commit edc8fc01f608 ("x86: Fix
  CPUIDLE_FLAG_IRQ_ENABLE leaking timer reprogram") that touched
  mwait_idle_with_hints() in arch/x86/include/asm/mwait.h

o Dropping the ARM results since I never got my hands on the ARM64
  system I used in my last testing. If I do manage to get my hands on it
  again, I'll rerun the experiments and share the results on the thread.
  To test the case where TIF_NOTIFY_IPI is not enabled for a particular
  architecture, I applied the series only until Patch 3 and tested the
  same on my x86 machine with a WARN_ON_ONCE() in do_idle() to check if
  tif_notify_ipi() ever return true and then repeated the same with
  Patch 4 applied.

o Updated benchmark results based on the latest base.

o Collected the Ack from Guo Ren for CSKY enablement.

o Dropped the RFC tag.

o Unfortunately, the series does not solve the issue highlighted by
  Julia Lawall w.r.t. NUMA Balancing in [0] based on her testing of v1.
  However, she did highlight a possible regression last time around
  where compiling a single file took much longer with the series but I
  could not reproduce it on my end. For sanity, I did rerun the same
  experiment this time around and I could not see any difference.
  Following are the numbers for

  $ make clean
  $ time make kernel/sched/core.o

  ---> tip:sched/core
  
-j1

real0m32.734s
user0m25.158s
sys 0m6.750s

-j256

real0m7.181s
user0m27.509s
sys 0m7.876s

  --> tip:sched/core + TIF_NOTIFY_IPI
  
-j1

real0m32.408s
user0m24.826s
sys 0m6.767s

-j256

real0m7.187s
user0m27.556s
sys 0m7.602s

  [0] 
https://lore.kernel.org/lkml/alpine.DEB.2.22.394.2310032059060.3220@hadrien/

Individual patches have their own changelog to help with review.

With those details out of the way ...

Problem statement
=

When measuring IPI throughput using a modified version of Anton
Blanchard's ipistorm benchmark [1], configured to measure time taken to
perform a fixed number of smp_call_function_single() (with wait set to
1), an increase in benchmark time was observed between v5.7 and the
upstream release v6.7-rc6 (this was the latest upstream kernel at the
time of encountering the issue). The issue persists on v6.10-rc1 as
well.

Bisection pointed to commit b2a02fc43a1f ("smp: Optimize
send_call_function_single_ipi()") as the reason behind this increase in
runtime.


Experiments
===

Since the commit cannot be cleanly reverted on top of the current
tip:sched/core, the effects of the optimizations were reverted by:

1. Removing the check for call_function_single_prep_ipi() in
   send_call_function_single_ipi(). With this change
   send_call_function_single_ipi() always calls
   arch_send_call_function_single_ipi()

2. Removing the call to flush_smp_call_function_queue() in do_idle()
   since every smp_call_function, with (1.), would unconditionally send
   an IPI to an idle CPU in TIF_POLLING mode.

Following is the diff of the above described changes which will be
henceforth referred to as the "revert":

diff --git a/kernel/sched/idle.c b/kernel/sched/idle.c
index 31231925f1ec..735184d98c0f 100644
--- a/kernel/sched/idle.c
+++ b/kernel/sched/idle.c
@@ -332,11 +332,6 @@ static void do_idle(void)
 */
smp_mb__after_atomic();
 
-   /*
-* RCU relies on this call to be done outside of an RCU read-side
-* critical section.
-*/
-   flush_smp_call_function_queue();
schedule_idle();
 
if (unlikely(klp_patch_pending(current)))
diff --git a/kernel/smp.c b/kernel/smp.c
index f085ebcdf9e7..2ff100c41885 100644
--- a/kernel/smp.c
+++ b/kernel/smp.c
@@ -111,11 +111,9 @@ void