On 01/14/2014 07:33 PM, Jason Low wrote:
When running workloads that have high contention in mutexes on an 8 socket
machine, spinners would often spin for a long time with no lock owner.

One of the potential reasons for this is because a thread can be preempted
after clearing lock->owner but before releasing the lock, or preempted after
acquiring the mutex but before setting lock->owner. In those cases, the
spinner cannot check if owner is not on_cpu because lock->owner is NULL.

A solution that would address the preemption part of this problem would
be to disable preemption between acquiring/releasing the mutex and
setting/clearing the lock->owner. However, that will require adding overhead
to the mutex fastpath.

The solution used in this patch is to limit the # of times thread can spin on
lock->count when !owner.

The threshold used in this patch for each spinner was 128, which appeared to
be a generous value, but any suggestions on another method to determine
the threshold are welcomed.

Signed-off-by: Jason Low<jason.l...@hp.com>
---
  kernel/locking/mutex.c |   10 +++++++---
  1 files changed, 7 insertions(+), 3 deletions(-)

diff --git a/kernel/locking/mutex.c b/kernel/locking/mutex.c
index b500cc7..9465604 100644
--- a/kernel/locking/mutex.c
+++ b/kernel/locking/mutex.c
@@ -43,6 +43,7 @@
   * mutex.
   */
  #define       MUTEX_SHOW_NO_WAITER(mutex)     (atomic_read(&(mutex)->count)>= 
0)
+#define        MUTEX_SPIN_THRESHOLD            (128)

  void
  __mutex_init(struct mutex *lock, const char *name, struct lock_class_key *key)
@@ -418,7 +419,7 @@ __mutex_lock_common(struct mutex *lock, long state, 
unsigned int subclass,
        struct task_struct *task = current;
        struct mutex_waiter waiter;
        unsigned long flags;
-       int ret;
+       int ret, nr_spins = 0;
        struct mspin_node node;

        preempt_disable();
@@ -453,6 +454,7 @@ __mutex_lock_common(struct mutex *lock, long state, 
unsigned int subclass,
        mspin_lock(MLOCK(lock),&node);
        for (;;) {
                struct task_struct *owner;
+               nr_spins++;

                if (use_ww_ctx&&  ww_ctx->acquired>  0) {
                        struct ww_mutex *ww;
@@ -502,9 +504,11 @@ __mutex_lock_common(struct mutex *lock, long state, 
unsigned int subclass,
                 * When there's no owner, we might have preempted between the
                 * owner acquiring the lock and setting the owner field. If
                 * we're an RT task that will live-lock because we won't let
-                * the owner complete.
+                * the owner complete. Additionally, when there is no owner,
+                * stop spinning after too many tries.
                 */
-               if (!owner&&  (need_resched() || rt_task(task))) {
+               if (!owner&&  (need_resched() || rt_task(task) ||
+                              nr_spins>  MUTEX_SPIN_THRESHOLD)) {
                        mspin_unlock(MLOCK(lock),&node);
                        goto slowpath;
                }

The time that a thread spent on one iteration of the loop can be highly variable. Instead of a predefined iteration count, you may consider setting an elapsed time limit on how long a thread can spin in the loop using changes in jiffies as a proxy. Let say setting a limit of 40ms. On a 1000Hz system, a changes of 40 or more in jiffies will indicate it is time to leave and go to the slowpath.

-Longman
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to