Occasionally losing the tick_sched_timer

2018-04-09 Thread Nicholas Piggin
We are seeing rare hard lockup watchdog timeouts, a CPU seems to have no
more timers scheduled, despite hard and soft lockup watchdogs should have
their heart beat timers and probably many others.

The reproducer we have is running a KVM workload. The lockup is in the
host kernel, quite rare but we may be able to slowly test things.

I have a sysrq+q snippet. CPU3 is the stuck one, you can see its tick has
stopped for a long time and no hrtimer active. Included CPU4 for what the
other CPUs look like.

Thomas do you have any ideas on what we might look for, or if we can add
some BUG_ON()s to catch this at its source?

- CPU3 is sitting in its cpuidle loop (polling idle with all other idle
  states disabled).

- `taskset -c 3 ls` basically revived the CPU and got timers running again.

- May not be a new bug because we have just in the past few releases
  enabled the hard lockup detector by default

- KVM is being used. This switches various registers like timebase and
  decrementer. Possibly that's involved, but we can't say the bug does
  not happen without KVM.

cpu: 3
 clock 0:
  .base:   df30f5ab
  .index:  0
  .resolution: 1 nsecs
  .get_time:   
ktime_get

  .offset: 0 nsecs
active timers:
 clock 1:
  .base:   520cc304
  .index:  1
  .resolution: 1 nsecs
  .get_time:   
ktime_get_real

  .offset: 1523263049759155857 nsecs
active timers:
 clock 2:
  .base:   706e6277
  .index:  2
  .resolution: 1 nsecs
  .get_time:   
ktime_get_boottime

  .offset: 0 nsecs
active timers:
 clock 3:
  .base:   e2ae2811
  .index:  3
  .resolution: 1 nsecs
  .get_time:   
ktime_get_clocktai

  .offset: 1523263049759155857 nsecs
active timers:
 clock 4:
  .base:   c93e2f8e
  .index:  4
  .resolution: 1 nsecs
  .get_time:   
ktime_get

  .offset: 0 nsecs
active timers:
 clock 5:
  .base:   7b726c6a
  .index:  5
  .resolution: 1 nsecs
  .get_time:   
ktime_get_real

  .offset: 1523263049759155857 nsecs
active timers:
 clock 6:
  .base:   f17c2d4f
  .index:  6
  .resolution: 1 nsecs
  .get_time:   
ktime_get_boottime

  .offset: 0 nsecs
active timers:
 clock 7:
  .base:   6c57ef89
  .index:  7
  .resolution: 1 nsecs
  .get_time:   
ktime_get_clocktai

  .offset: 1523263049759155857 nsecs
active timers:
  .expires_next   : 9223372036854775807 nsecs
  .hres_active: 1
  .nr_events  : 1446533
  .nr_retries : 1434
  .nr_hangs   : 0
  .max_hang_time  : 0
  .nohz_mode  : 2
  .last_tick  : 1776312000 nsecs
  .tick_stopped   : 1
  .idle_jiffies   : 4296713609
  .idle_calls : 2573133
  .idle_sleeps: 1957794
  .idle_entrytime : 1776312625 nsecs
  .idle_waketime  : 59550238131639 nsecs
  .idle_exittime  : 17763110009176 nsecs
  .idle_sleeptime : 17504617295679 nsecs
  .iowait_sleeptime: 719978688 nsecs
  .last_jiffies   : 4296713608
  .next_timer : 1776313000
  .idle_expires   : 1776313000 nsecs
jiffies: 4300892324

cpu: 4
 clock 0:
  .base:   07d8226b
  .index:  0
  .resolution: 1 nsecs
  .get_time:   
ktime_get

  .offset: 0 nsecs
active timers:
 #0: 

, 
tick_sched_timer
, S:01

 # expires at 5955295000-5955295000 nsecs [in 2685654802 to 2685654802 
nsecs]
 #1: 
<9b4a3b88>
, 
hrtimer_wakeup
, S:01

 # expires at 59602585423025-59602642458243 nsecs [in 52321077827 to 
52378113045 nsecs]
 clock 1:
  .base:   d2ae50c4
  .index:  1
  .resolution: 1 nsecs
  .get_time:   
ktime_get_real

  .offset: 1523263049759155857 nsecs
active timers:
 clock 2:
  .base:   1a80e123
  .index:  2
  .resolution: 1 nsecs
  .get_time:   
ktime_get_boottime

  .offset: 0 nsecs
active timers:
 clock 3:
  .base:   5c97ab69
  .index:  3
  .resolution: 1 nsecs
  .get_time:   
ktime_get_clocktai

  .offset: 1523263049759155857 nsecs
active timers:
 clock 4:
  .base:   75ac8f03
  .index:  4
  .resolution: 1 nsecs
  .get_time:   
ktime_get

  .offset: 0 nsecs
active timers:
 clock 5:
  .base:   db06f6ce
  .index:  5
  .resolution: 1 nsecs
  .get_time:   
ktime_get_real

  .offset: 1523263049759155857 nsecs
active timers:
 clock 6:
  .base:   fa63fbce
  .index:  6
  .resolution: 1 nsecs
  .get_time:   
ktime_get_boottime

  .offset: 0 nsecs
active timers:
 clock 7:
  .base:   41de439c
  .index:  7
  .resolution: 1 nsecs
  .get_time:   
ktime_get_clocktai

  .offset: 1523263049759155857 nsecs
active timers:
  .expires_next   : 5955295000 nsecs
  .hres_active: 1
  .nr_events  : 294282
  .nr_retries : 16138
  .nr_hangs   : 0
  .max_hang_time  : 0
  .nohz_mode  : 2
  .last_tick  : 5954562000 nsecs
  .tick_stopped   : 1
  .idle_jiffies   : 4300891859
  .idle_calls : 553259
  .idle_sleeps: 536396
  .idle_entrytime : 59547990019145 nsecs

Re: Occasionally losing the tick_sched_timer

2018-04-10 Thread Thomas Gleixner
Nick,

On Tue, 10 Apr 2018, Nicholas Piggin wrote:
> We are seeing rare hard lockup watchdog timeouts, a CPU seems to have no
> more timers scheduled, despite hard and soft lockup watchdogs should have
> their heart beat timers and probably many others.
>
> The reproducer we have is running a KVM workload. The lockup is in the
> host kernel, quite rare but we may be able to slowly test things.
> 
> I have a sysrq+q snippet. CPU3 is the stuck one, you can see its tick has
> stopped for a long time and no hrtimer active. Included CPU4 for what the
> other CPUs look like.
> 
> Thomas do you have any ideas on what we might look for, or if we can add
> some BUG_ON()s to catch this at its source?

Not really. Tracing might be a more efficient tool that random BUG_ONs.

> - CPU3 is sitting in its cpuidle loop (polling idle with all other idle
>   states disabled).
> 
> - `taskset -c 3 ls` basically revived the CPU and got timers running again.

Which is not surprising because that kicks the CPU out of idle and starts
the tick timer again.

Does this restart the watchdog timers as well?

> cpu: 3
>  clock 0:
>   .base:   df30f5ab
>   .index:  0
>   .resolution: 1 nsecs
>   .get_time:   ktime_get
>   .offset: 0 nsecs
> active timers:

So in theory the soft lockup watchdog hrtimer should be queued here.

>   .expires_next   : 9223372036854775807 nsecs
>   .hres_active: 1
>   .nr_events  : 1446533
>   .nr_retries : 1434
>   .nr_hangs   : 0
>   .max_hang_time  : 0
>   .nohz_mode  : 2
>   .last_tick  : 1776312000 nsecs
>   .tick_stopped   : 1
>   .idle_jiffies   : 4296713609
>   .idle_calls : 2573133
>   .idle_sleeps: 1957794

>   .idle_waketime  : 59550238131639 nsecs
>   .idle_sleeptime : 17504617295679 nsecs
>   .iowait_sleeptime: 719978688 nsecs
>   .last_jiffies   : 4296713608

So this was the last time when the CPU came out of idle:

>   .idle_exittime  : 17763110009176 nsecs

Here it went back into idle:

>   .idle_entrytime : 1776312625 nsecs

And this was the next timer wheel timer due for expiry:

>   .next_timer : 1776313000
>   .idle_expires   : 1776313000 nsecs

which makes no sense whatsoever, but this might be stale information. Can't
tell.

> cpu: 4
>  clock 0:
>   .base:   07d8226b
>   .index:  0
>   .resolution: 1 nsecs
>   .get_time:   ktime_get
>   .offset: 0 nsecs
> active timers: #0: , tick_sched_timer, S:01
># expires at 5955295000-5955295000 nsecs [in 
> 2685654802 to 2685654802 nsecs]

The tick timer is scheduled because the next timer wheel timer is due at
5955295000, which might be the hard watchdog timer

>#1: <9b4a3b88>, hrtimer_wakeup, S:01
># expires at 59602585423025-59602642458243 nsecs [in 
> 52321077827 to 52378113045 nsecs]

That might be the soft lockup hrtimer.

I'd try to gather more information about the chain of events via tracing
and stop the tracer once the lockup detector hits.

Thanks,

tglx




Re: Occasionally losing the tick_sched_timer

2018-04-10 Thread Nicholas Piggin
On Tue, 10 Apr 2018 09:42:29 +0200 (CEST)
Thomas Gleixner  wrote:

> Nick,
> 
> On Tue, 10 Apr 2018, Nicholas Piggin wrote:
> > We are seeing rare hard lockup watchdog timeouts, a CPU seems to have no
> > more timers scheduled, despite hard and soft lockup watchdogs should have
> > their heart beat timers and probably many others.
> >
> > The reproducer we have is running a KVM workload. The lockup is in the
> > host kernel, quite rare but we may be able to slowly test things.
> > 
> > I have a sysrq+q snippet. CPU3 is the stuck one, you can see its tick has
> > stopped for a long time and no hrtimer active. Included CPU4 for what the
> > other CPUs look like.
> > 
> > Thomas do you have any ideas on what we might look for, or if we can add
> > some BUG_ON()s to catch this at its source?  
> 
> Not really. Tracing might be a more efficient tool that random BUG_ONs.

Sure, we could try that. Any suggestions? timer events?

> 
> > - CPU3 is sitting in its cpuidle loop (polling idle with all other idle
> >   states disabled).
> > 
> > - `taskset -c 3 ls` basically revived the CPU and got timers running again. 
> >  
> 
> Which is not surprising because that kicks the CPU out of idle and starts
> the tick timer again.

Yep.
 
> Does this restart the watchdog timers as well?

I think so, but now you ask I'm not 100% sure we directly observed it.
We can check that next time it locks up.

> > cpu: 3
> >  clock 0:
> >   .base:   df30f5ab
> >   .index:  0
> >   .resolution: 1 nsecs
> >   .get_time:   ktime_get
> >   .offset: 0 nsecs
> > active timers:  
> 
> So in theory the soft lockup watchdog hrtimer should be queued here.
> 
> >   .expires_next   : 9223372036854775807 nsecs
> >   .hres_active: 1
> >   .nr_events  : 1446533
> >   .nr_retries : 1434
> >   .nr_hangs   : 0
> >   .max_hang_time  : 0
> >   .nohz_mode  : 2
> >   .last_tick  : 1776312000 nsecs
> >   .tick_stopped   : 1
> >   .idle_jiffies   : 4296713609
> >   .idle_calls : 2573133
> >   .idle_sleeps: 1957794  
> 
> >   .idle_waketime  : 59550238131639 nsecs
> >   .idle_sleeptime : 17504617295679 nsecs
> >   .iowait_sleeptime: 719978688 nsecs
> >   .last_jiffies   : 4296713608  
> 
> So this was the last time when the CPU came out of idle:
> 
> >   .idle_exittime  : 17763110009176 nsecs  
> 
> Here it went back into idle:
> 
> >   .idle_entrytime : 1776312625 nsecs  
> 
> And this was the next timer wheel timer due for expiry:
> 
> >   .next_timer : 1776313000
> >   .idle_expires   : 1776313000 nsecs  
> 
> which makes no sense whatsoever, but this might be stale information. Can't
> tell.

Wouldn't we expect to see that if there is a timer that was missed
somehow because the tick_sched_timer was not set?

> 
> > cpu: 4
> >  clock 0:
> >   .base:   07d8226b
> >   .index:  0
> >   .resolution: 1 nsecs
> >   .get_time:   ktime_get
> >   .offset: 0 nsecs
> > active timers: #0: , tick_sched_timer, S:01
> >  # expires at 5955295000-5955295000 nsecs [in 
> > 2685654802 to 2685654802 nsecs]  
> 
> The tick timer is scheduled because the next timer wheel timer is due at
> 5955295000, which might be the hard watchdog timer
> 
> >  #1: <9b4a3b88>, hrtimer_wakeup, S:01
> >  # expires at 59602585423025-59602642458243 nsecs [in 
> > 52321077827 to 52378113045 nsecs]  
> 
> That might be the soft lockup hrtimer.
> 
> I'd try to gather more information about the chain of events via tracing
> and stop the tracer once the lockup detector hits.

Okay will do, thanks for taking a look.

Thanks,
Nick


Re: Occasionally losing the tick_sched_timer

2018-04-10 Thread Thomas Gleixner
On Tue, 10 Apr 2018, Nicholas Piggin wrote:
> On Tue, 10 Apr 2018 09:42:29 +0200 (CEST)
> Thomas Gleixner  wrote:
> > > Thomas do you have any ideas on what we might look for, or if we can add
> > > some BUG_ON()s to catch this at its source?  
> > 
> > Not really. Tracing might be a more efficient tool that random BUG_ONs.
> 
> Sure, we could try that. Any suggestions? timer events?

timer, hrtimer and the tick-sched stuff should be a good start. And make
sure to freeze the trace once you hit the fault case. tracing_off() is your
friend.

Thanks,

tglx