On Thu, Aug 01, 2013 at 03:31:55PM +0200, Peter Zijlstra wrote:
> On Thu, Aug 01, 2013 at 02:46:58PM +0200, Jiri Olsa wrote:
> > On Tue, Jul 23, 2013 at 02:31:04AM +0200, Frederic Weisbecker wrote:
> > > This is going to be used by the full dynticks subsystem
> > > as a finer-grained information to know when to keep and
> > > when to stop the tick.
> > > 
> > > Original-patch-by: Peter Zijlstra <pet...@infradead.org>
> > > Signed-off-by: Frederic Weisbecker <fweis...@gmail.com>
> > > Cc: Jiri Olsa <jo...@redhat.com>
> > > Cc: Peter Zijlstra <pet...@infradead.org>
> > > Cc: Namhyung Kim <namhy...@kernel.org>
> > > Cc: Ingo Molnar <mi...@kernel.org>
> > > Cc: Arnaldo Carvalho de Melo <a...@redhat.com>
> > > Cc: Stephane Eranian <eran...@google.com>
> > > ---
> > >  kernel/events/core.c |    7 +++++++
> > >  1 files changed, 7 insertions(+), 0 deletions(-)
> > > 
> > > diff --git a/kernel/events/core.c b/kernel/events/core.c
> > > index b40c3db..f9bd39b 100644
> > > --- a/kernel/events/core.c
> > > +++ b/kernel/events/core.c
> > > @@ -141,6 +141,7 @@ enum event_type_t {
> > >  struct static_key_deferred perf_sched_events __read_mostly;
> > >  static DEFINE_PER_CPU(atomic_t, perf_cgroup_events);
> > >  static DEFINE_PER_CPU(atomic_t, perf_branch_stack_events);
> > > +static DEFINE_PER_CPU(atomic_t, perf_freq_events);
> > >  
> > >  static atomic_t nr_mmap_events __read_mostly;
> > >  static atomic_t nr_comm_events __read_mostly;
> > > @@ -3139,6 +3140,9 @@ static void unaccount_event_cpu(struct perf_event 
> > > *event, int cpu)
> > >   }
> > >   if (is_cgroup_event(event))
> > >           atomic_dec(&per_cpu(perf_cgroup_events, cpu));
> > > +
> > > + if (event->attr.freq)
> > > +         atomic_dec(&per_cpu(perf_freq_events, cpu));
> > >  }
> > >  
> > >  static void unaccount_event(struct perf_event *event)
> > > @@ -6473,6 +6477,9 @@ static void account_event_cpu(struct perf_event 
> > > *event, int cpu)
> > >   }
> > >   if (is_cgroup_event(event))
> > >           atomic_inc(&per_cpu(perf_cgroup_events, cpu));
> > > +
> > > + if (event->attr.freq)
> > > +         atomic_inc(&per_cpu(perf_freq_events, cpu));
> > 
> > cpu could be -1 in here.. getting:
> 
> Ho humm, right you are. 
> 
> So we have:
> 
> static void account_event_cpu(struct perf_event *event, int cpu)
> {
>       if (event->parent)
>               return;
> 
>       if (has_branch_stack(event)) {
>               if (!(event->attach_state & PERF_ATTACH_TASK))
>                       atomic_inc(&per_cpu(perf_branch_stack_events, cpu));
>       }
>       if (is_cgroup_event(event))
>               atomic_inc(&per_cpu(perf_cgroup_events, cpu));
> 
>       if (event->attr.freq)
>               atomic_inc(&per_cpu(perf_freq_events, cpu));
> }
> 
> Where the freq thing is new and shiney, but we already had the other
> two. Of those, cgroup events must be per cpu so that should be good,
> the branch_stack thing tests ATTACH_TASK, which should also be good, but
> leaves me wonder wth they do for those that are attached to tasks.
> 
> But yes, the frequency thing is borken.

Aie, so the freq thing,  I can either account to all CPUs (inc to all and send 
an IPI to all), or
when the event scheds in/out. Probably we should do the former to avoid sending 
an IPI at all context switches.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to