On Tue, Jun 06, 2017 at 04:00:03PM -0700, Palmer Dabbelt wrote: > + * atomic_add - add integer to atomic variable > + * @i: integer value to add > + * @v: pointer of type atomic_t > + * > + * Atomically adds @i to @v. > + */ > +static inline void atomic_add(int i, atomic_t *v) > +{ > + __asm__ __volatile__ ( > + "amoadd.w zero, %1, %0" > + : "+A" (v->counter) > + : "r" (i)); > +} > + > +#define atomic_fetch_add atomic_fetch_add > +static inline int atomic_fetch_add(unsigned int mask, atomic_t *v) > +{ > + int out; > + > + __asm__ __volatile__ ( > + "amoadd.w %2, %1, %0" > + : "+A" (v->counter), "=r" (out) > + : "r" (mask)); > + return out; > +} > + > +/** > + * atomic_sub - subtract integer from atomic variable > + * @i: integer value to subtract > + * @v: pointer of type atomic_t > + * > + * Atomically subtracts @i from @v. > + */ > +static inline void atomic_sub(int i, atomic_t *v) > +{ > + atomic_add(-i, v); > +} > + > +#define atomic_fetch_sub atomic_fetch_sub > +static inline int atomic_fetch_sub(unsigned int mask, atomic_t *v) > +{ > + int out; > + > + __asm__ __volatile__ ( > + "amosub.w %2, %1, %0" > + : "+A" (v->counter), "=r" (out) > + : "r" (mask)); > + return out; > +} > + > +/** > + * atomic_add_return - add integer to atomic variable > + * @i: integer value to add > + * @v: pointer of type atomic_t > + * > + * Atomically adds @i to @v and returns the result > + */ > +static inline int atomic_add_return(int i, atomic_t *v) > +{ > + register int c; > + > + __asm__ __volatile__ ( > + "amoadd.w %0, %2, %1" > + : "=r" (c), "+A" (v->counter) > + : "r" (i)); > + return (c + i); > +} > + > +/** > + * atomic_sub_return - subtract integer from atomic variable > + * @i: integer value to subtract > + * @v: pointer of type atomic_t > + * > + * Atomically subtracts @i from @v and returns the result > + */ > +static inline int atomic_sub_return(int i, atomic_t *v) > +{ > + return atomic_add_return(-i, v); > +} > +
> +/** > + * atomic_and - Atomically clear bits in atomic variable > + * @mask: Mask of the bits to be retained > + * @v: pointer of type atomic_t > + * > + * Atomically retains the bits set in @mask from @v > + */ > +static inline void atomic_and(unsigned int mask, atomic_t *v) > +{ > + __asm__ __volatile__ ( > + "amoand.w zero, %1, %0" > + : "+A" (v->counter) > + : "r" (mask)); > +} > + > +#define atomic_fetch_and atomic_fetch_and > +static inline int atomic_fetch_and(unsigned int mask, atomic_t *v) > +{ > + int out; > + > + __asm__ __volatile__ ( > + "amoand.w %2, %1, %0" > + : "+A" (v->counter), "=r" (out) > + : "r" (mask)); > + return out; > +} > + > +/** > + * atomic_or - Atomically set bits in atomic variable > + * @mask: Mask of the bits to be set > + * @v: pointer of type atomic_t > + * > + * Atomically sets the bits set in @mask in @v > + */ > +static inline void atomic_or(unsigned int mask, atomic_t *v) > +{ > + __asm__ __volatile__ ( > + "amoor.w zero, %1, %0" > + : "+A" (v->counter) > + : "r" (mask)); > +} > + > +#define atomic_fetch_or atomic_fetch_or > +static inline int atomic_fetch_or(unsigned int mask, atomic_t *v) > +{ > + int out; > + > + __asm__ __volatile__ ( > + "amoor.w %2, %1, %0" > + : "+A" (v->counter), "=r" (out) > + : "r" (mask)); > + return out; > +} > + > +/** > + * atomic_xor - Atomically flips bits in atomic variable > + * @mask: Mask of the bits to be flipped > + * @v: pointer of type atomic_t > + * > + * Atomically flips the bits set in @mask in @v > + */ > +static inline void atomic_xor(unsigned int mask, atomic_t *v) > +{ > + __asm__ __volatile__ ( > + "amoxor.w zero, %1, %0" > + : "+A" (v->counter) > + : "r" (mask)); > +} > + > +#define atomic_fetch_xor atomic_fetch_xor > +static inline int atomic_fetch_xor(unsigned int mask, atomic_t *v) > +{ > + int out; > + > + __asm__ __volatile__ ( > + "amoxor.w %2, %1, %0" > + : "+A" (v->counter), "=r" (out) > + : "r" (mask)); > + return out; > +} What pretty much all the other architectures do is something like: #define ATOMIC_OP(op, asm_op, c_op) \ static __always_inline void atomic_##op(int i, atomic_t *v) \ { \ __asm__ __volatile__ ( \ "amo" #asm_op ".w zero, %1, %0" \ : "+A" (v->counter) \ : "r" (i)); \ } #define ATOMIC_FETCH_OP(op, asm_op, c_op) \ static __always_inline int atomic_fetch_##op(int i, atomic_t *v)\ { \ register int ret; \ __asm__ __volatile__ ( \ "amo" #asm_op ".w %2, %1, %0" \ : "+A" (v->counter), "=r" (ret) \ : "r" (mask)); \ return ret; \ } #define ATOMIC_OP_RETURN(op, asm_op, c_op) \ static __always_inline int atomic_##op##_return(int i, atomic_t *v) \ { \ return atomic_fetch_##op(i, v) c_op i; \ } #define ATOMIC_OPS(op, asm_op, c_op) \ ATOMIC_OP(op, asm_op, c_op) \ ATOMIC_OP_RETURN(op, asm_op, c_op) \ ATOMIC_FETCH_OP(op, asm_op, c_op) ATOMIC_OPS(add, add, +) ATOMIC_OPS(sub, sub, -) #undef ATOMIC_OPS #define ATOMIC_OPS(op, asm_op, c_op) \ ATOMIC_OP(op, asm_op, c_op) \ ATOMIC_FETCH_OP(op, asm_op, c_op) ATOMIC_OPS(and, and, &) ATOMIC_OPS(or, or, |) ATOMIC_OPS(xor, xor, ^) #undef ATOMIC_OPS Which is much simpler no?