On Tue, 2016-06-21 at 23:49 +0200, Peter Zijlstra wrote: > On Thu, Jun 16, 2016 at 12:06:07PM -0400, [email protected] wrote: > > > > @@ -53,36 +56,72 @@ DEFINE_PER_CPU(seqcount_t, irq_time_seq); > > * softirq -> hardirq, hardirq -> softirq > > * > > * When exiting hardirq or softirq time, account the elapsed time. > > + * > > + * When exiting softirq time, subtract the amount of hardirq time > > that > > + * interrupted this softirq run, to avoid double accounting of > > that time. > > */ > > void irqtime_account_irq(struct task_struct *curr, int irqtype) > > { > > + u64 prev_softirq_start; > > + u64 prev_hardirq; > > + u64 hardirq_time; > > + s64 delta = 0; > We appear to always assign to delta, so this initialization seems > superfluous. > > > > > int cpu; > > > > if (!sched_clock_irqtime) > > return; > > > > cpu = smp_processor_id(); > Per this smp_processor_id() usage, preemption is disabled.
This code is called from the timer code. Surely preemption
is already disabled?
Should I change this into raw_smp_processor_id()?
> >
> > + /*
> > + * Softirq context may get interrupted by hardirq context,
> > + * on the same CPU. At softirq entry time the amount of
> > time
> > + * spent in hardirq context is stored. At softirq exit
> > time,
> > + * the time spent in hardirq context during the softirq is
> > + * subtracted.
> > + */
> > + prev_hardirq = __this_cpu_read(prev_hardirq_time);
> > + prev_softirq_start = __this_cpu_read(softirq_start_time);
> > +
> > + if (irqtype == HARDIRQ_OFFSET) {
> > + delta = sched_clock_cpu(cpu) -
> > __this_cpu_read(hardirq_start_time);
> > + __this_cpu_add(hardirq_start_time, delta);
> > + } else do {
> > + u64 now = sched_clock_cpu(cpu);
> > + hardirq_time = READ_ONCE(per_cpu(cpu_hardirq_time,
> > cpu));
> Which makes this per_cpu(,cpu) usage somewhat curious. What's wrong
> with
> __this_cpu_read() ?
Is __this_cpu_read() as fast as per_cpu(,cpu) on all
architectures?
> >
> > +
> > + delta = now - prev_softirq_start;
> > + if (in_serving_softirq()) {
> > + /*
> > + * Leaving softirq context. Avoid double
> > counting by
> > + * subtracting hardirq time from this
> > interval.
> > + */
> > + s64 hi_delta = hardirq_time -
> > prev_hardirq;
> > + delta -= hi_delta;
> > + } else {
> > + /* Entering softirq context. Note start
> > times. */
> > + __this_cpu_write(softirq_start_time, now);
> > + __this_cpu_write(prev_hardirq_time,
> > hardirq_time);
> > + }
> > + /*
> > + * If a hardirq happened during this calculation,
> > it may not
> > + * have gotten a consistent snapshot. Try again.
> > + */
> > + } while (hardirq_time !=
> > READ_ONCE(per_cpu(cpu_hardirq_time, cpu)));
> That whole thing is somewhat hard to read; but its far too late for
> me
> to suggest anything more readable :/
I only had 2 1/2 hours of sleep last night, so I will not
try to rewrite it now, but I will see if there is anything
I can do to make it more readable tomorrow.
If you have any ideas before then, please let me know :)
> >
> > + irq_time_write_begin(irqtype);
> > /*
> > * We do not account for softirq time from ksoftirqd here.
> > * We want to continue accounting softirq time to
> > ksoftirqd thread
> > * in that case, so as not to confuse scheduler with a
> > special task
> > * that do not consume any time, but still wants to run.
> > */
> > + if (irqtype == HARDIRQ_OFFSET && hardirq_count())
> > __this_cpu_add(cpu_hardirq_time, delta);
> > + else if (irqtype == SOFTIRQ_OFFSET && in_serving_softirq()
> > &&
> > + curr != this_cpu_ksoftirqd())
> > __this_cpu_add(cpu_softirq_time, delta);
> >
> > + irq_time_write_end(irqtype);
> Maybe split the whole thing on irqtype at the very start, instead of
> the
> endless repeated branches?
Let me try if I can make things more readable that way.
Thanks for the review!
Rik
--
All Rights Reversed.
signature.asc
Description: This is a digitally signed message part

