On Wed, 2007-07-25 at 13:41 +0200, Nick Piggin wrote:
> Add powerpc optimised lock bitops.
>
> Signed-off-by: Nick Piggin <[EMAIL PROTECTED]>
Acked-by: Benjamin Herrenschmidt <[EMAIL PROTECTED]>
> ---
> include/asm-powerpc/bitops.h | 46
> ++++++++++++++++++++++++++++++++++++++++++-
> 1 file changed, 45 insertions(+), 1 deletion(-)
>
> Index: linux-2.6/include/asm-powerpc/bitops.h
> ===================================================================
> --- linux-2.6.orig/include/asm-powerpc/bitops.h
> +++ linux-2.6/include/asm-powerpc/bitops.h
> @@ -86,6 +86,24 @@ static __inline__ void clear_bit(int nr,
> : "cc" );
> }
>
> +static __inline__ void clear_bit_unlock(int nr, volatile unsigned long *addr)
> +{
> + unsigned long old;
> + unsigned long mask = BITOP_MASK(nr);
> + unsigned long *p = ((unsigned long *)addr) + BITOP_WORD(nr);
> +
> + __asm__ __volatile__(
> + LWSYNC_ON_SMP
> +"1:" PPC_LLARX "%0,0,%3 # clear_bit_unlock\n"
> + "andc %0,%0,%2\n"
> + PPC405_ERR77(0,%3)
> + PPC_STLCX "%0,0,%3\n"
> + "bne- 1b"
> + : "=&r" (old), "+m" (*p)
> + : "r" (mask), "r" (p)
> + : "cc", "memory");
> +}
> +
> static __inline__ void change_bit(int nr, volatile unsigned long *addr)
> {
> unsigned long old;
> @@ -125,6 +143,27 @@ static __inline__ int test_and_set_bit(u
> return (old & mask) != 0;
> }
>
> +static __inline__ int test_and_set_bit_lock(unsigned long nr,
> + volatile unsigned long *addr)
> +{
> + unsigned long old, t;
> + unsigned long mask = BITOP_MASK(nr);
> + unsigned long *p = ((unsigned long *)addr) + BITOP_WORD(nr);
> +
> + __asm__ __volatile__(
> +"1:" PPC_LLARX "%0,0,%3 # test_and_set_bit_lock\n"
> + "or %1,%0,%2 \n"
> + PPC405_ERR77(0,%3)
> + PPC_STLCX "%1,0,%3 \n"
> + "bne- 1b"
> + ISYNC_ON_SMP
> + : "=&r" (old), "=&r" (t)
> + : "r" (mask), "r" (p)
> + : "cc", "memory");
> +
> + return (old & mask) != 0;
> +}
> +
> static __inline__ int test_and_clear_bit(unsigned long nr,
> volatile unsigned long *addr)
> {
> @@ -185,6 +224,12 @@ static __inline__ void set_bits(unsigned
>
> #include <asm-generic/bitops/non-atomic.h>
>
> +static __inline__ void __clear_bit_unlock(int nr, volatile unsigned long
> *addr)
> +{
> + __asm__ __volatile__(LWSYNC_ON_SMP ::: "memory");
> + __clear_bit(nr, addr);
> +}
> +
> /*
> * Return the zero-based bit position (LE, not IBM bit numbering) of
> * the most significant 1-bit in a double word.
> @@ -266,7 +311,6 @@ static __inline__ int fls(unsigned int x
> #include <asm-generic/bitops/fls64.h>
>
> #include <asm-generic/bitops/hweight.h>
> -#include <asm-generic/bitops/lock.h>
>
> #define find_first_zero_bit(addr, size) find_next_zero_bit((addr), (size), 0)
> unsigned long find_next_zero_bit(const unsigned long *addr,
> -
> To unsubscribe from this list: send the line "unsubscribe linux-arch" in
> the body of a message to [EMAIL PROTECTED]
> More majordomo info at http://vger.kernel.org/majordomo-info.html
-
To unsubscribe from this list: send the line "unsubscribe linux-arch" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at http://vger.kernel.org/majordomo-info.html