On 05/04/2015 05:26 AM, Paolo Bonzini wrote: > Isn't this racy? > > synchronize_rcu CPU nohz CPU > --------------------------------------------------------- > set flag = 0 > read flag = 0 > return to userspace > set TIF_NOHZ > > and there's no guarantee that TIF_NOHZ is ever processed by the nohz CPU.
Looking at the code some more, a flag is not going to be enough. An irq can hit while we are in kernel mode, leading to the task's "rcu active" counter being incremented twice. However, currently the RCU code seems to use a much more complex counting scheme, with a different increment for kernel/task use, and irq use. This counter seems to be modeled on the task preempt_counter, where we do care about whether we are in task context, irq context, or softirq context. On the other hand, the RCU code only seems to care about whether or not a CPU is in an extended quiescent state, or is potentially in an RCU critical section. Paul, what is the reason for RCU using a complex counter, instead of a simple increment for each potential kernel/RCU entry, like rcu_read_lock() does with CONFIG_PREEMPT_RCU enabled? In fact, would we be able to simply use tsk->rcu_read_lock_nesting as an indicator of whether or not we should bother waiting on that task or CPU when doing synchronize_rcu? -- All rights reversed -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/