On Wed, Jun 26, 2013 at 01:57:55AM +0530, Srivatsa S. Bhat wrote: > Once stop_machine() is gone from the CPU offline path, we won't be able > to depend on disabling preemption to prevent CPUs from going offline > from under us. > > In RCU code, rcu_implicit_dynticks_qs() checks if a CPU is offline, > while being protected by a spinlock. Use the get/put_online_cpus_atomic() > APIs to prevent CPUs from going offline, while invoking from atomic context.
I am not completely sure that this is needed. Here is a (quite possibly flawed) argument for its not being needed: o rcu_gp_init() holds off CPU-hotplug operations during grace-period initialization. Therefore, RCU will avoid looking for quiescent states from CPUs that were offline (and thus in an extended quiescent state) at the beginning of the grace period. o If force_qs_rnp() is looking for a quiescent state from a given CPU, and if it senses that CPU as being offline, then even without synchronization we know that the CPU was offline some time during the current grace period. After all, it was online at the beginning of the grace period (otherwise, we would not be looking at it at all), and our later sampling of its state must have therefore happened after the start of the grace period. Given that the grace period has not yet ended, it also has to happened before the end of the grace period. o Therefore, we should be able to sample the offline state without synchronization. Possible flaws in this argument: memory ordering, oddnesses in the sampling and updates of the cpumask recording which CPUs are online, and so on. Thoughts? Thanx, Paul > Cc: Dipankar Sarma <dipan...@in.ibm.com> > Cc: "Paul E. McKenney" <paul...@linux.vnet.ibm.com> > Signed-off-by: Srivatsa S. Bhat <srivatsa.b...@linux.vnet.ibm.com> > --- > > kernel/rcutree.c | 4 ++++ > 1 file changed, 4 insertions(+) > > diff --git a/kernel/rcutree.c b/kernel/rcutree.c > index cf3adc6..caeed1a 100644 > --- a/kernel/rcutree.c > +++ b/kernel/rcutree.c > @@ -2107,6 +2107,8 @@ static void force_qs_rnp(struct rcu_state *rsp, int > (*f)(struct rcu_data *)) > rcu_initiate_boost(rnp, flags); /* releases rnp->lock */ > continue; > } > + > + get_online_cpus_atomic(); > cpu = rnp->grplo; > bit = 1; > for (; cpu <= rnp->grphi; cpu++, bit <<= 1) { > @@ -2114,6 +2116,8 @@ static void force_qs_rnp(struct rcu_state *rsp, int > (*f)(struct rcu_data *)) > f(per_cpu_ptr(rsp->rda, cpu))) > mask |= bit; > } > + put_online_cpus_atomic(); > + > if (mask != 0) { > > /* rcu_report_qs_rnp() releases rnp->lock. */ > _______________________________________________ Linuxppc-dev mailing list Linuxppc-dev@lists.ozlabs.org https://lists.ozlabs.org/listinfo/linuxppc-dev