More aggressive use of optimistic spinning and enabling readers to
participate in spinning may make tasks waiting in the queue harder
to get access to the semaphore, especially for writers who need
exclusive access.

This patch enables a waking writer to go back to the optimistic
spinning loop as long as the owner is running. This allows waiting
writers better access to the semaphore as well as reduce the size
of the waiting queue.

Signed-off-by: Waiman Long <waiman.l...@hp.com>
---
 kernel/locking/rwsem-xadd.c |   38 +++++++++++++++++++++++++++++++-------
 1 files changed, 31 insertions(+), 7 deletions(-)

diff --git a/kernel/locking/rwsem-xadd.c b/kernel/locking/rwsem-xadd.c
index aafc9f0..94b0124 100644
--- a/kernel/locking/rwsem-xadd.c
+++ b/kernel/locking/rwsem-xadd.c
@@ -484,6 +484,11 @@ static bool rwsem_optimistic_spin(struct rw_semaphore *sem,
 {
        return false;
 }
+
+static inline bool rwsem_can_spin_on_owner(struct rw_semaphore *sem)
+{
+       return false;
+}
 #endif
 
 /*
@@ -553,12 +558,14 @@ __visible
 struct rw_semaphore __sched *rwsem_down_write_failed(struct rw_semaphore *sem)
 {
        long count;
-       bool waiting = true; /* any queued threads before us */
+       bool waiting; /* any queued threads before us */
+       bool respin;
        struct rwsem_waiter waiter;
 
        /* undo write bias from down_write operation, stop active locking */
        count = rwsem_atomic_update(-RWSEM_ACTIVE_WRITE_BIAS, sem);
 
+optspin:
        /* do optimistic spinning and steal lock if possible */
        if (rwsem_optimistic_spin(sem, RWSEM_WAITING_FOR_WRITE))
                return sem;
@@ -573,8 +580,7 @@ struct rw_semaphore __sched *rwsem_down_write_failed(struct 
rw_semaphore *sem)
        raw_spin_lock_irq(&sem->wait_lock);
 
        /* account for this before adding a new element to the list */
-       if (list_empty(&sem->wait_list))
-               waiting = false;
+       waiting = !list_empty(&sem->wait_list);
 
        list_add_tail(&waiter.list, &sem->wait_list);
 
@@ -595,23 +601,41 @@ struct rw_semaphore __sched 
*rwsem_down_write_failed(struct rw_semaphore *sem)
 
        /* wait until we successfully acquire the lock */
        set_current_state(TASK_UNINTERRUPTIBLE);
-       while (true) {
+       respin = false;
+       while (!respin) {
                if (rwsem_try_write_lock(count, sem))
                        break;
                raw_spin_unlock_irq(&sem->wait_lock);
 
-               /* Block until there are no active lockers. */
-               do {
+               /*
+                * Block until there are no active lockers or optimistic
+                * spinning is possible.
+                */
+               while (true) {
                        schedule();
                        set_current_state(TASK_UNINTERRUPTIBLE);
-               } while ((count = sem->count) & RWSEM_ACTIVE_MASK);
+                       count = ACCESS_ONCE(sem->count);
+                       if (!(count & RWSEM_ACTIVE_MASK))
+                               break;
+                       /*
+                        * Go back to optimistic spinning if possible
+                        */
+                       if (rwsem_can_spin_on_owner(sem)) {
+                               respin = true;
+                               break;
+                       }
+               }
 
                raw_spin_lock_irq(&sem->wait_lock);
        }
        __set_current_state(TASK_RUNNING);
 
        list_del(&waiter.list);
+       if (respin && list_empty(&sem->wait_list))
+               rwsem_atomic_update(-RWSEM_WAITING_BIAS, sem);
        raw_spin_unlock_irq(&sem->wait_lock);
+       if (respin)
+               goto optspin;
 
        return sem;
 }
-- 
1.7.1

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to