On Wed, Oct 12, 2016 at 06:19:27PM -0400, Rich Felker wrote:
> On Wed, Oct 12, 2016 at 01:34:17PM -0700, Paul E. McKenney wrote:
> > On Wed, Oct 12, 2016 at 12:35:43PM -0400, Rich Felker wrote:
> > > On Wed, Oct 12, 2016 at 10:18:02AM +0200, Thomas Gleixner wrote:
> > > > On Tue, 11 Oct 2016, Rich Felker wrote:
> > > > > On Sun, Oct 09, 2016 at 09:23:58PM +0200, Thomas Gleixner wrote:
> > > > > > On Sun, 9 Oct 2016, Rich Felker wrote:
> > > > > > > On Sun, Oct 09, 2016 at 01:03:10PM +0200, Thomas Gleixner wrote:
> > > > > > > My preference would just be to keep the branch, but with your 
> > > > > > > improved
> > > > > > > version that doesn't need a function call:
> > > > > > > 
> > > > > > >   irqd_is_per_cpu(irq_desc_get_irq_data(desc))
> > > > > > >
> > > > > > > While there is some overhead testing this condition every time, I 
> > > > > > > can
> > > > > > > probably come up with several better places to look for a ~10 
> > > > > > > cycle
> > > > > > > improvement in the irq code path without imposing new 
> > > > > > > requirements on
> > > > > > > the DT bindings.
> > > > > > 
> > > > > > Fair enough. Your call.
> > > > > >  
> > > > > > > As noted in my followup to the clocksource stall thread, there's 
> > > > > > > also
> > > > > > > a possibility that it might make sense to consider the current
> > > > > > > behavior of having non-percpu irqs bound to a particular cpu as 
> > > > > > > part
> > > > > > > of what's required by the compatible tag, in which case
> > > > > > > handle_percpu_irq or something similar/equivalent might be 
> > > > > > > suitable
> > > > > > > for both the percpu and non-percpu cases. I don't understand the 
> > > > > > > irq
> > > > > > > subsystem well enough to insist on that but I think it's worth
> > > > > > > consideration since it looks like it would improve performance of
> > > > > > > non-percpu interrupts a bit.
> > > > > > 
> > > > > > Well, you can use handle_percpu_irq() for your device interrupts if 
> > > > > > you
> > > > > > guarantee at the hardware level that there is no reentrancy. Once 
> > > > > > you make
> > > > > > the hardware capable of delivering them on either core the picture 
> > > > > > changes.
> > > > > 
> > > > > One more concern here -- I see that handle_simple_irq is handling the
> > > > > soft-disable / IRQS_PENDING flag behavior, and irq_check_poll stuff
> > > > > that's perhaps important too. Since soft-disable is all we have
> > > > > (there's no hard-disable of interrupts), is this a problem? In other
> > > > > words, can drivers have an expectation of not receiving interrupts
> > > > > when the irq is disabled? I would think anything compatible with irq
> > > > > sharing can't have such an expectation, but perhaps the kernel needs
> > > > > disabling internally for synchronization at module-unload time or
> > > > > similar cases?
> > > > 
> > > > Sure. A driver would be surprised getting an interrupt when it is 
> > > > disabled,
> > > > but with your exceptionally well thought out interrupt controller a 
> > > > pending
> > > > (level) interrupt which is not handled will be reraised forever and just
> > > > hard lock the machine.
> > > 
> > > If you want to criticize the interrupt controller design (not my work
> > > or under my control) for limitations in the type of hardware that can
> > > be hooked up to it, that's okay -- this kind of input will actually be
> > > useful for designing the next iteration of it -- but I don't think
> > > this specific possibility is a concern.
> > 
> > Well, if this scenario does happen, the machine will likely either lock
> > up silently and hard, give you RCU CPU stall warning messages, or give
> > you soft-lockup messages.
> 
> The same situation can happen with badly-behaved hardware under
> software interrupt control too if it keeps generating interrupts
> rapidly (more quickly than the cpu can handle them), unless the kernel
> has some kind of framework for disabling the interrupt and only
> reenabling it later via a timer. It's equivalent to a realtime-prio
> process failing to block/sleep to give lower-priority processes a
> chance to run.

Indeed, there are quite a few scenarios that can lead to silent hard
lockups, RCU CPU stall warning messages, or soft-lockup messages.

                                                        Thanx, Paul

Reply via email to