From: Guo Ren <guo...@linux.alibaba.com>

This patch introduces a ticket lock implementation for riscv, along the
same lines as the implementation for arch/arm & arch/csky.

Signed-off-by: Guo Ren <guo...@linux.alibaba.com>
Cc: Catalin Marinas <catalin.mari...@arm.com>
Cc: Will Deacon <will.dea...@arm.com>
Cc: Peter Zijlstra <pet...@infradead.org>
Cc: Palmer Dabbelt <palmerdabb...@google.com>
Cc: Anup Patel <a...@brainfault.org>
Cc: Arnd Bergmann <a...@arndb.de>
---
 arch/riscv/Kconfig                      |   1 +
 arch/riscv/include/asm/Kbuild           |   1 +
 arch/riscv/include/asm/spinlock.h       | 158 ++++++++++++--------------------
 arch/riscv/include/asm/spinlock_types.h |  19 ++--
 4 files changed, 74 insertions(+), 105 deletions(-)

diff --git a/arch/riscv/Kconfig b/arch/riscv/Kconfig
index 87d7b52..7c56a20 100644
--- a/arch/riscv/Kconfig
+++ b/arch/riscv/Kconfig
@@ -30,6 +30,7 @@ config RISCV
        select ARCH_HAS_STRICT_KERNEL_RWX if MMU
        select ARCH_OPTIONAL_KERNEL_RWX if ARCH_HAS_STRICT_KERNEL_RWX
        select ARCH_OPTIONAL_KERNEL_RWX_DEFAULT
+       select ARCH_USE_QUEUED_RWLOCKS
        select ARCH_WANT_DEFAULT_TOPDOWN_MMAP_LAYOUT if MMU
        select ARCH_WANT_FRAME_POINTERS
        select ARCH_WANT_HUGE_PMD_SHARE if 64BIT
diff --git a/arch/riscv/include/asm/Kbuild b/arch/riscv/include/asm/Kbuild
index 445ccc9..e57ef80 100644
--- a/arch/riscv/include/asm/Kbuild
+++ b/arch/riscv/include/asm/Kbuild
@@ -3,5 +3,6 @@ generic-y += early_ioremap.h
 generic-y += extable.h
 generic-y += flat.h
 generic-y += kvm_para.h
+generic-y += qrwlock.h
 generic-y += user.h
 generic-y += vmlinux.lds.h
diff --git a/arch/riscv/include/asm/spinlock.h 
b/arch/riscv/include/asm/spinlock.h
index f4f7fa1..2c81764 100644
--- a/arch/riscv/include/asm/spinlock.h
+++ b/arch/riscv/include/asm/spinlock.h
@@ -7,129 +7,91 @@
 #ifndef _ASM_RISCV_SPINLOCK_H
 #define _ASM_RISCV_SPINLOCK_H
 
-#include <linux/kernel.h>
-#include <asm/current.h>
-#include <asm/fence.h>
-
 /*
- * Simple spin lock operations.  These provide no fairness guarantees.
+ * Ticket-based spin-locking.
  */
+static inline void arch_spin_lock(arch_spinlock_t *lock)
+{
+       arch_spinlock_t lockval;
+       u32 tmp;
+
+       asm volatile (
+               "1:     lr.w    %0, %2          \n"
+               "       mv      %1, %0          \n"
+               "       addw    %0, %0, %3      \n"
+               "       sc.w    %0, %0, %2      \n"
+               "       bnez    %0, 1b          \n"
+               : "=&r" (tmp), "=&r" (lockval), "+A" (lock->lock)
+               : "r" (1 << TICKET_NEXT)
+               : "memory");
 
-/* FIXME: Replace this with a ticket lock, like MIPS. */
-
-#define arch_spin_is_locked(x) (READ_ONCE((x)->lock) != 0)
+       while (lockval.tickets.next != lockval.tickets.owner) {
+               /*
+                * FIXME - we need wfi/wfe here to prevent:
+                *  - cache line bouncing
+                *  - saving cpu pipeline in multi-harts-per-core
+                *    processor
+                */
+               lockval.tickets.owner = READ_ONCE(lock->tickets.owner);
+       }
 
-static inline void arch_spin_unlock(arch_spinlock_t *lock)
-{
-       smp_store_release(&lock->lock, 0);
+       __atomic_acquire_fence();
 }
 
 static inline int arch_spin_trylock(arch_spinlock_t *lock)
 {
-       int tmp = 1, busy;
-
-       __asm__ __volatile__ (
-               "       amoswap.w %0, %2, %1\n"
-               RISCV_ACQUIRE_BARRIER
-               : "=r" (busy), "+A" (lock->lock)
-               : "r" (tmp)
+       u32 tmp, contended, res;
+
+       do {
+               asm volatile (
+               "       lr.w    %0, %3          \n"
+               "       srliw   %1, %0, %5      \n"
+               "       slliw   %2, %0, %5      \n"
+               "       or      %1, %2, %1      \n"
+               "       li      %2, 0           \n"
+               "       sub     %1, %1, %0      \n"
+               "       bnez    %1, 1f          \n"
+               "       addw    %0, %0, %4      \n"
+               "       sc.w    %2, %0, %3      \n"
+               "1:                             \n"
+               : "=&r" (tmp), "=&r" (contended), "=&r" (res),
+                 "+A" (lock->lock)
+               : "r" (1 << TICKET_NEXT), "I" (TICKET_NEXT)
                : "memory");
+       } while (res);
 
-       return !busy;
-}
-
-static inline void arch_spin_lock(arch_spinlock_t *lock)
-{
-       while (1) {
-               if (arch_spin_is_locked(lock))
-                       continue;
-
-               if (arch_spin_trylock(lock))
-                       break;
+       if (!contended) {
+               __atomic_acquire_fence();
+               return 1;
+       } else {
+               return 0;
        }
 }
 
-/***********************************************************/
-
-static inline void arch_read_lock(arch_rwlock_t *lock)
+static inline void arch_spin_unlock(arch_spinlock_t *lock)
 {
-       int tmp;
-
-       __asm__ __volatile__(
-               "1:     lr.w    %1, %0\n"
-               "       bltz    %1, 1b\n"
-               "       addi    %1, %1, 1\n"
-               "       sc.w    %1, %1, %0\n"
-               "       bnez    %1, 1b\n"
-               RISCV_ACQUIRE_BARRIER
-               : "+A" (lock->lock), "=&r" (tmp)
-               :: "memory");
+       smp_store_release(&lock->tickets.owner, lock->tickets.owner + 1);
+       /* FIXME - we need ipi/sev here to notify above */
 }
 
-static inline void arch_write_lock(arch_rwlock_t *lock)
+static inline int arch_spin_value_unlocked(arch_spinlock_t lock)
 {
-       int tmp;
-
-       __asm__ __volatile__(
-               "1:     lr.w    %1, %0\n"
-               "       bnez    %1, 1b\n"
-               "       li      %1, -1\n"
-               "       sc.w    %1, %1, %0\n"
-               "       bnez    %1, 1b\n"
-               RISCV_ACQUIRE_BARRIER
-               : "+A" (lock->lock), "=&r" (tmp)
-               :: "memory");
+       return lock.tickets.owner == lock.tickets.next;
 }
 
-static inline int arch_read_trylock(arch_rwlock_t *lock)
+static inline int arch_spin_is_locked(arch_spinlock_t *lock)
 {
-       int busy;
-
-       __asm__ __volatile__(
-               "1:     lr.w    %1, %0\n"
-               "       bltz    %1, 1f\n"
-               "       addi    %1, %1, 1\n"
-               "       sc.w    %1, %1, %0\n"
-               "       bnez    %1, 1b\n"
-               RISCV_ACQUIRE_BARRIER
-               "1:\n"
-               : "+A" (lock->lock), "=&r" (busy)
-               :: "memory");
-
-       return !busy;
+       return !arch_spin_value_unlocked(READ_ONCE(*lock));
 }
 
-static inline int arch_write_trylock(arch_rwlock_t *lock)
+static inline int arch_spin_is_contended(arch_spinlock_t *lock)
 {
-       int busy;
-
-       __asm__ __volatile__(
-               "1:     lr.w    %1, %0\n"
-               "       bnez    %1, 1f\n"
-               "       li      %1, -1\n"
-               "       sc.w    %1, %1, %0\n"
-               "       bnez    %1, 1b\n"
-               RISCV_ACQUIRE_BARRIER
-               "1:\n"
-               : "+A" (lock->lock), "=&r" (busy)
-               :: "memory");
+       struct __raw_tickets tickets = READ_ONCE(lock->tickets);
 
-       return !busy;
+       return (tickets.next - tickets.owner) > 1;
 }
+#define arch_spin_is_contended arch_spin_is_contended
 
-static inline void arch_read_unlock(arch_rwlock_t *lock)
-{
-       __asm__ __volatile__(
-               RISCV_RELEASE_BARRIER
-               "       amoadd.w x0, %1, %0\n"
-               : "+A" (lock->lock)
-               : "r" (-1)
-               : "memory");
-}
-
-static inline void arch_write_unlock(arch_rwlock_t *lock)
-{
-       smp_store_release(&lock->lock, 0);
-}
+#include <asm/qrwlock.h>
 
 #endif /* _ASM_RISCV_SPINLOCK_H */
diff --git a/arch/riscv/include/asm/spinlock_types.h 
b/arch/riscv/include/asm/spinlock_types.h
index f398e76..d7b38bf 100644
--- a/arch/riscv/include/asm/spinlock_types.h
+++ b/arch/riscv/include/asm/spinlock_types.h
@@ -10,16 +10,21 @@
 # error "please don't include this file directly"
 #endif
 
+#define TICKET_NEXT    16
+
 typedef struct {
-       volatile unsigned int lock;
+       union {
+               u32 lock;
+               struct __raw_tickets {
+                       /* little endian */
+                       u16 owner;
+                       u16 next;
+               } tickets;
+       };
 } arch_spinlock_t;
 
-#define __ARCH_SPIN_LOCK_UNLOCKED      { 0 }
-
-typedef struct {
-       volatile unsigned int lock;
-} arch_rwlock_t;
+#define __ARCH_SPIN_LOCK_UNLOCKED      { { 0 } }
 
-#define __ARCH_RW_LOCK_UNLOCKED                { 0 }
+#include <asm-generic/qrwlock_types.h>
 
 #endif /* _ASM_RISCV_SPINLOCK_TYPES_H */
-- 
2.7.4

Reply via email to