* Peter Zijlstra <pet...@infradead.org> wrote:
> On Wed, Apr 10, 2019 at 05:13:54AM -0700, tip-bot for Peter Zijlstra wrote: > > Commit-ID: 86071b11317550d994b55ce5e31aa06bcad783b5 > > Gitweb: > > https://git.kernel.org/tip/86071b11317550d994b55ce5e31aa06bcad783b5 > > Author: Peter Zijlstra <pet...@infradead.org> > > AuthorDate: Thu, 4 Apr 2019 15:03:00 +0200 > > Committer: Ingo Molnar <mi...@kernel.org> > > CommitDate: Wed, 10 Apr 2019 13:47:09 +0200 > > > > perf/core: Fix perf_event_disable_inatomic() race > > > > Thomas-Mich Richter reported he triggered a WARN()ing from > > event_function_local() > > on his s390. The problem boils down to: > > > > CPU-A CPU-B > > > > perf_event_overflow() > > perf_event_disable_inatomic() > > @pending_disable = 1 > > irq_work_queue(); > > > > sched-out > > event_sched_out() > > @pending_disable = 0 > > > > sched-in > > perf_event_overflow() > > perf_event_disable_inatomic() > > @pending_disable = 1; > > irq_work_queue(); // FAILS > > > > irq_work_run() > > perf_pending_event() > > if (@pending_disable) > > perf_event_disable_local(); // WHOOPS > > > > The problem exists in generic, but s390 is particularly sensitive > > because it doesn't implement arch_irq_work_raise(), nor does it call > > irq_work_run() from it's PMU interrupt handler (nor would that be > > sufficient in this case, because s390 also generates > > perf_event_overflow() from pmu::stop). Add to that the fact that s390 > > is a virtual architecture and (virtual) CPU-A can stall long enough > > for the above race to happen, even if it would self-IPI. > > > > Adding a irq_work_sync() to event_sched_in() would work for all hardare > > PMUs that properly use irq_work_run() but fails for software PMUs. > > > > Instead encode the CPU number in @pending_disable, such that we can > > tell which CPU requested the disable. This then allows us to detect > > the above scenario and even redirect the IPI to make up for the failed > > queue. > > Ingo, could you please fold in the below delta? It turns out I > overlooked two insteances :-( > > --- a/kernel/events/ring_buffer.c > +++ b/kernel/events/ring_buffer.c > @@ -392,7 +392,7 @@ void *perf_aux_output_begin(struct perf_ > * store that will be enabled on successful return > */ > if (!handle->size) { /* A, matches D */ > - event->pending_disable = 1; > + event->pending_disable = smp_processor_id(); > perf_output_wakeup(handle); > local_set(&rb->aux_nest, 0); > goto err_put; > @@ -480,7 +480,7 @@ void perf_aux_output_end(struct perf_out > > if (wakeup) { > if (handle->aux_flags & PERF_AUX_FLAG_TRUNCATED) > - handle->event->pending_disable = 1; > + handle->event->pending_disable = smp_processor_id(); > perf_output_wakeup(handle); > } Sure, done! Thanks, Ingo