On Mon, Feb 02, 2015 at 06:32:32PM +0100, Peter Zijlstra wrote:
> On Mon, Feb 02, 2015 at 04:42:40PM +0100, Peter Zijlstra wrote:
> > On Mon, Feb 02, 2015 at 01:33:14AM -0500, Vince Weaver wrote:
> > > [407484.309136] ------------[ cut here ]------------
> 
> > > [407484.588602]  <<EOE>>  <IRQ>  [<ffffffff8115c28c>] 
> > > perf_prepare_sample+0x2ec/0x3c0
> > > [407484.597358]  [<ffffffff8115c46e>] __perf_event_overflow+0x10e/0x270
> > > [407484.604708]  [<ffffffff8115c439>] ? __perf_event_overflow+0xd9/0x270
> > > [407484.612215]  [<ffffffff8115c924>] ? perf_tp_event+0xc4/0x210
> > > [407484.619000]  [<ffffffff8115cfe2>] ? __perf_sw_event+0x72/0x1f0
> > > [407484.625937]  [<ffffffff8115c799>] ? perf_swevent_overflow+0xa9/0xe0
> > > [407484.633287]  [<ffffffff8115c799>] perf_swevent_overflow+0xa9/0xe0
> > > [407484.640467]  [<ffffffff8115c837>] perf_swevent_event+0x67/0x90
> > > [407484.647343]  [<ffffffff8115c924>] perf_tp_event+0xc4/0x210
> > > [407484.653923]  [<ffffffff810b6fa9>] ? lock_acquire+0x119/0x130
> > > [407484.660606]  [<ffffffff810b3cf6>] ? 
> > > perf_trace_lock_acquire+0x146/0x180
> > > [407484.668332]  [<ffffffff810b594f>] ? __lock_acquire.isra.31+0x3af/0xfe0
> > > [407484.675962]  [<ffffffff810b3cf6>] perf_trace_lock_acquire+0x146/0x180
> > > [407484.683490]  [<ffffffff810b6fa9>] ? lock_acquire+0x119/0x130
> > > [407484.690211]  [<ffffffff810b6fa9>] lock_acquire+0x119/0x130
> > > [407484.696750]  [<ffffffff8115b7f5>] ? perf_event_wakeup+0x5/0xf0
> > > [407484.703640]  [<ffffffff811f50ef>] ? kill_fasync+0xf/0xf0
> > > [407484.710008]  [<ffffffff8115b828>] perf_event_wakeup+0x38/0xf0
> > > [407484.716798]  [<ffffffff8115b7f5>] ? perf_event_wakeup+0x5/0xf0
> > > [407484.723696]  [<ffffffff8115b913>] perf_pending_event+0x33/0x60
> > > [407484.730570]  [<ffffffff8114cc7c>] irq_work_run_list+0x4c/0x80
> > > [407484.737392]  [<ffffffff8114ccc8>] irq_work_run+0x18/0x40
> > > [407484.743765]  [<ffffffff8101955f>] 
> > > smp_trace_irq_work_interrupt+0x3f/0xc0
> > > [407484.751579]  [<ffffffff816c01fd>] trace_irq_work_interrupt+0x6d/0x80
> 
> > > [407484.799195] ---[ end trace 55752a03ec8ab979 ]---
> > 
> > That looks like tail recursive fun! An irq work that raises and irq work
> > ad infinitum. Lemme see if I can squash that.. didn't we have something
> > like this before... /me goes look.
> 
> 
> Does this make it go away?
> 
> --- a/kernel/events/core.c
> +++ b/kernel/events/core.c
> @@ -4413,6 +4413,8 @@ static void perf_pending_event(struct ir
>       struct perf_event *event = container_of(entry,
>                       struct perf_event, pending);
>  
> +     int rctx = perf_swevent_get_recursion_context();
> +

hum, you should check the rctx

        if (rctx == -1)
                return;

also this recursion is bound to swevent_htable, should we rather add
separate ctx data for irq_work to limit the clashing with SW events?

jirka

>       if (event->pending_disable) {
>               event->pending_disable = 0;
>               __perf_event_disable(event);
> @@ -4422,6 +4424,8 @@ static void perf_pending_event(struct ir
>               event->pending_wakeup = 0;
>               perf_event_wakeup(event);
>       }
> +
> +     perf_swevent_put_recursion_context(rctx);
>  }
>  
>  /*
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to