On 09.04.2018 12:26, Peter Zijlstra wrote: > On Mon, Apr 09, 2018 at 10:25:32AM +0300, Alexey Budankov wrote: >> >> Store preempting context switch out event into Perf trace as a part of >> PERF_RECORD_SWITCH[_CPU_WIDE] record. >> >> Percentage of preempting and non-preempting context switches help >> understanding the nature of workloads (CPU or IO bound) that are running >> on a machine; >> >> The event is treated as preemption one when task->state value of the >> thread being switched out is TASK_RUNNING. Event type encoding is >> implemented using PERF_RECORD_MISC_SWITCH_OUT_PREEMPT bit; >> >> Signed-off-by: Alexey Budankov <alexey.budan...@linux.intel.com> > > Acked-by: Peter Zijlstra (Intel) <pet...@infradead.org> > > Acme, I'm thinking you should route this, since most of the changes are > actually to the tool. > >> diff --git a/kernel/events/core.c b/kernel/events/core.c >> index fc1c330c6bd6..872a5aaa77eb 100644 >> --- a/kernel/events/core.c >> +++ b/kernel/events/core.c >> @@ -7584,6 +7584,10 @@ static void perf_event_switch(struct task_struct >> *task, >> }, >> }; >> >> + if (!sched_in && task->state == TASK_RUNNING) >> + switch_event.event_id.header.misc |= >> + PERF_RECORD_MISC_SWITCH_OUT_PREEMPT; > > I typically prefer {} over any multi-line expression.
Yep, makes sense. Sorry for missing that here. > >> perf_iterate_sb(perf_event_switch_output, >> &switch_event, >> NULL); >