On Tue, Jun 30, 2015 at 08:37:01PM -0700, Paul E. McKenney wrote:
> On Tue, Jun 30, 2015 at 05:42:14PM -0700, Josh Triplett wrote:
> > On Tue, Jun 30, 2015 at 05:15:58PM -0700, Paul E. McKenney wrote:
> > > On Tue, Jun 30, 2015 at 04:46:33PM -0700, j...@joshtriplett.org wrote:
> > > > On Tue, Jun 30, 2015 at 03:12:24PM -0700, Paul E. McKenney wrote:
> > > > > On Tue, Jun 30, 2015 at 03:00:15PM -0700, j...@joshtriplett.org wrote:
> > > > > > On Tue, Jun 30, 2015 at 02:48:05PM -0700, Paul E. McKenney wrote:
> > > > > > > Hello!
> > > > > > > 
> > > > > > > This series contains some highly experimental patches that allow 
> > > > > > > normal
> > > > > > > grace periods to take advantage of the work done by concurrent 
> > > > > > > expedited
> > > > > > > grace periods.  This can reduce the overhead incurred by normal 
> > > > > > > grace
> > > > > > > periods by eliminating the need for force-quiescent-state scans 
> > > > > > > that
> > > > > > > would otherwise have happened after the expedited grace period 
> > > > > > > completed.
> > > > > > > It is not clear whether this is a useful tradeoff.  Nevertheless, 
> > > > > > > this
> > > > > > > series contains the following patches:
> > > > > > 
> > > > > > While it makes sense to avoid unnecessarily delaying a normal grace
> > > > > > period if the expedited machinery has provided the necessary delay, 
> > > > > > I'm
> > > > > > also *deeply* concerned that this will create a new class of
> > > > > > nondeterministic performance issues.  Something that uses RCU may
> > > > > > perform badly due to grace period latency, but then suddenly start
> > > > > > performing well because an unrelated task starts hammering expedited
> > > > > > grace periods.  This seems particularly likely during boot, for
> > > > > > instance, where RCU grace periods can be a significant component of 
> > > > > > boot
> > > > > > time (when you're trying to boot to userspace in small fractions of 
> > > > > > a
> > > > > > second).
> > > > > 
> > > > > I will take that as another vote against.  And for a reason that I had
> > > > > not yet come up with, so good show!  ;-)
> > > > 
> > > > Consider it a fairly weak concern against.  Increasing performance seems
> > > > like a good thing in general; I just don't relish the future "feels less
> > > > responsive" bug reports that take a long time to track down and turn out
> > > > to be "this completely unrelated driver was loaded and started using
> > > > expedited grace periods".
> > > 
> > > From what I can see, this one needs a good reason to go in, as opposed
> > > to a good reason to stay out.
> > > 
> > > > Then again, perhaps the more relevant concern would be why drivers use
> > > > expedited grace periods in the first place.
> > > 
> > > Networking uses expedited grace periods when RTNL is held to reduce
> > > contention on that lock.
> > 
> > Wait, what?  Why is anything using traditional (non-S) RCU while *any*
> > lock is held?
> 
> In their defense, it is a sleeplock that is never taken except when
> rearranging networking configuration.  Sometimes they need a grace period
> under the lock.  So synchronize_net() checks to see if RTNL is held, and
> does a synchronize_rcu_expedited() if so and a synchronize_rcu() if not.
> 
> But maybe I am misunderstanding your question?

No, you understood my question.  It seems wrong that they would need a
grace period *under* the lock, rather than the usual case of making a
change under the lock, dropping the lock, and *then* synchronizing.

- Josh Triplett
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to