Sorry, I meant MIPS.

* Mathieu Desnoyers ([EMAIL PROTECTED]) wrote:
> ia64 architecture local_t extension.
> 
> Signed-off-by: Mathieu Desnoyers <[EMAIL PROTECTED]>
> 
> --- a/include/asm-mips/system.h
> +++ b/include/asm-mips/system.h
> @@ -253,6 +253,58 @@ static inline unsigned long __cmpxchg_u3
>       return retval;
>  }
>  
> +static inline unsigned long __cmpxchg_u32_local(volatile int * m,
> +     unsigned long old, unsigned long new)
> +{
> +     __u32 retval;
> +
> +     if (cpu_has_llsc && R10000_LLSC_WAR) {
> +             __asm__ __volatile__(
> +             "       .set    push                                    \n"
> +             "       .set    noat                                    \n"
> +             "       .set    mips3                                   \n"
> +             "1:     ll      %0, %2                  # __cmpxchg_u32 \n"
> +             "       bne     %0, %z3, 2f                             \n"
> +             "       .set    mips0                                   \n"
> +             "       move    $1, %z4                                 \n"
> +             "       .set    mips3                                   \n"
> +             "       sc      $1, %1                                  \n"
> +             "       beqzl   $1, 1b                                  \n"
> +             "2:                                                     \n"
> +             "       .set    pop                                     \n"
> +             : "=&r" (retval), "=R" (*m)
> +             : "R" (*m), "Jr" (old), "Jr" (new)
> +             : "memory");
> +     } else if (cpu_has_llsc) {
> +             __asm__ __volatile__(
> +             "       .set    push                                    \n"
> +             "       .set    noat                                    \n"
> +             "       .set    mips3                                   \n"
> +             "1:     ll      %0, %2                  # __cmpxchg_u32 \n"
> +             "       bne     %0, %z3, 2f                             \n"
> +             "       .set    mips0                                   \n"
> +             "       move    $1, %z4                                 \n"
> +             "       .set    mips3                                   \n"
> +             "       sc      $1, %1                                  \n"
> +             "       beqz    $1, 1b                                  \n"
> +             "2:                                                     \n"
> +             "       .set    pop                                     \n"
> +             : "=&r" (retval), "=R" (*m)
> +             : "R" (*m), "Jr" (old), "Jr" (new)
> +             : "memory");
> +     } else {
> +             unsigned long flags;
> +
> +             local_irq_save(flags);
> +             retval = *m;
> +             if (retval == old)
> +                     *m = new;
> +             local_irq_restore(flags);       /* implies memory barrier  */
> +     }
> +
> +     return retval;
> +}
> +
>  #ifdef CONFIG_64BIT
>  static inline unsigned long __cmpxchg_u64(volatile int * m, unsigned long 
> old,
>       unsigned long new)
> @@ -303,10 +355,62 @@ static inline unsigned long __cmpxchg_u6
>  
>       return retval;
>  }
> +
> +static inline unsigned long __cmpxchg_u64_local(volatile int * m,
> +     unsigned long old, unsigned long new)
> +{
> +     __u64 retval;
> +
> +     if (cpu_has_llsc && R10000_LLSC_WAR) {
> +             __asm__ __volatile__(
> +             "       .set    push                                    \n"
> +             "       .set    noat                                    \n"
> +             "       .set    mips3                                   \n"
> +             "1:     lld     %0, %2                  # __cmpxchg_u64 \n"
> +             "       bne     %0, %z3, 2f                             \n"
> +             "       move    $1, %z4                                 \n"
> +             "       scd     $1, %1                                  \n"
> +             "       beqzl   $1, 1b                                  \n"
> +             "2:                                                     \n"
> +             "       .set    pop                                     \n"
> +             : "=&r" (retval), "=R" (*m)
> +             : "R" (*m), "Jr" (old), "Jr" (new)
> +             : "memory");
> +     } else if (cpu_has_llsc) {
> +             __asm__ __volatile__(
> +             "       .set    push                                    \n"
> +             "       .set    noat                                    \n"
> +             "       .set    mips3                                   \n"
> +             "1:     lld     %0, %2                  # __cmpxchg_u64 \n"
> +             "       bne     %0, %z3, 2f                             \n"
> +             "       move    $1, %z4                                 \n"
> +             "       scd     $1, %1                                  \n"
> +             "       beqz    $1, 1b                                  \n"
> +             "2:                                                     \n"
> +             "       .set    pop                                     \n"
> +             : "=&r" (retval), "=R" (*m)
> +             : "R" (*m), "Jr" (old), "Jr" (new)
> +             : "memory");
> +     } else {
> +             unsigned long flags;
> +
> +             local_irq_save(flags);
> +             retval = *m;
> +             if (retval == old)
> +                     *m = new;
> +             local_irq_restore(flags);       /* implies memory barrier  */
> +     }
> +
> +     return retval;
> +}
> +
>  #else
>  extern unsigned long __cmpxchg_u64_unsupported_on_32bit_kernels(
>       volatile int * m, unsigned long old, unsigned long new);
>  #define __cmpxchg_u64 __cmpxchg_u64_unsupported_on_32bit_kernels
> +extern unsigned long __cmpxchg_u64_local_unsupported_on_32bit_kernels(
> +     volatile int * m, unsigned long old, unsigned long new);
> +#define __cmpxchg_u64_local __cmpxchg_u64_local_unsupported_on_32bit_kernels
>  #endif
>  
>  /* This function doesn't exist, so you'll get a linker error
> @@ -326,7 +430,26 @@ static inline unsigned long __cmpxchg(vo
>       return old;
>  }
>  
> -#define cmpxchg(ptr,old,new) ((__typeof__(*(ptr)))__cmpxchg((ptr), (unsigned 
> long)(old), (unsigned long)(new),sizeof(*(ptr))))
> +static inline unsigned long __cmpxchg_local(volatile void * ptr,
> +     unsigned long old, unsigned long new, int size)
> +{
> +     switch (size) {
> +     case 4:
> +             return __cmpxchg_u32_local(ptr, old, new);
> +     case 8:
> +             return __cmpxchg_u64_local(ptr, old, new);
> +     }
> +     __cmpxchg_called_with_bad_pointer();
> +     return old;
> +}
> +
> +#define cmpxchg(ptr,old,new) \
> +     ((__typeof__(*(ptr)))__cmpxchg((ptr), \
> +             (unsigned long)(old), (unsigned long)(new),sizeof(*(ptr))))
> +
> +#define cmpxchg_local(ptr,old,new) \
> +     ((__typeof__(*(ptr)))__cmpxchg_local((ptr), \
> +             (unsigned long)(old), (unsigned long)(new),sizeof(*(ptr))))
>  
>  extern void set_handler (unsigned long offset, void *addr, unsigned long 
> len);
>  extern void set_uncached_handler (unsigned long offset, void *addr, unsigned 
> long len);
> --- a/include/asm-mips/local.h
> +++ b/include/asm-mips/local.h
> @@ -1,60 +1,527 @@
> -#ifndef _ASM_LOCAL_H
> -#define _ASM_LOCAL_H
> +#ifndef _ARCH_POWERPC_LOCAL_H
> +#define _ARCH_POWERPC_LOCAL_H
>  
>  #include <linux/percpu.h>
>  #include <asm/atomic.h>
>  
> -#ifdef CONFIG_32BIT
> +typedef struct
> +{
> +     local_long_t a;
> +} local_t;
>  
> -typedef atomic_t local_t;
> +#define LOCAL_INIT(i)        { local_LONG_INIT(i) }
>  
> -#define LOCAL_INIT(i)        ATOMIC_INIT(i)
> -#define local_read(v)        atomic_read(v)
> -#define local_set(v,i)       atomic_set(v,i)
> +#define local_read(l)        local_long_read(&(l)->a)
> +#define local_set(l,i)       local_long_set(&(l)->a, (i))
>  
> -#define local_inc(v) atomic_inc(v)
> -#define local_dec(v) atomic_dec(v)
> -#define local_add(i, v)      atomic_add(i, v)
> -#define local_sub(i, v)      atomic_sub(i, v)
> +#define local_add(i,l)       local_long_add((i),(&(l)->a))
> +#define local_sub(i,l)       local_long_sub((i),(&(l)->a))
> +#define local_inc(l) local_long_inc(&(l)->a)
> +#define local_dec(l) local_long_dec(&(l)->a)
>  
> -#endif
>  
> -#ifdef CONFIG_64BIT
> +#ifndef CONFIG_64BITS
>  
> -typedef atomic64_t local_t;
> +/*
> + * Same as above, but return the result value
> + */
> +static __inline__ int local_add_return(int i, local_t * l)
> +{
> +     unsigned long result;
> +
> +     if (cpu_has_llsc && R10000_LLSC_WAR) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     ll      %1, %2          # local_add_return      \n"
> +             "       addu    %0, %1, %3                              \n"
> +             "       sc      %0, %2                                  \n"
> +             "       beqzl   %0, 1b                                  \n"
> +             "       addu    %0, %1, %3                              \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else if (cpu_has_llsc) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     ll      %1, %2          # local_add_return      \n"
> +             "       addu    %0, %1, %3                              \n"
> +             "       sc      %0, %2                                  \n"
> +             "       beqz    %0, 1b                                  \n"
> +             "       addu    %0, %1, %3                              \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else {
> +             unsigned long flags;
> +
> +             local_irq_save(flags);
> +             result = &(l->a.counter);
> +             result += i;
> +             &(l->a.counter) = result;
> +             local_irq_restore(flags);
> +     }
> +
> +     return result;
> +}
> +
> +static __inline__ int local_sub_return(int i, local_t * l)
> +{
> +     unsigned long result;
> +
> +     if (cpu_has_llsc && R10000_LLSC_WAR) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     ll      %1, %2          # local_sub_return      \n"
> +             "       subu    %0, %1, %3                              \n"
> +             "       sc      %0, %2                                  \n"
> +             "       beqzl   %0, 1b                                  \n"
> +             "       subu    %0, %1, %3                              \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else if (cpu_has_llsc) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     ll      %1, %2          # local_sub_return      \n"
> +             "       subu    %0, %1, %3                              \n"
> +             "       sc      %0, %2                                  \n"
> +             "       beqz    %0, 1b                                  \n"
> +             "       subu    %0, %1, %3                              \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else {
> +             unsigned long flags;
> +
> +             local_irq_save(flags);
> +             result = &(l->a.counter);
> +             result -= i;
> +             &(l->a.counter) = result;
> +             local_irq_restore(flags);
> +     }
> +
> +     return result;
> +}
> +
> +/*
> + * local_sub_if_positive - conditionally subtract integer from atomic 
> variable
> + * @i: integer value to subtract
> + * @l: pointer of type local_t
> + *
> + * Atomically test @l and subtract @i if @l is greater or equal than @i.
> + * The function returns the old value of @l minus @i.
> + */
> +static __inline__ int local_sub_if_positive(int i, local_t * l)
> +{
> +     unsigned long result;
> +
> +     if (cpu_has_llsc && R10000_LLSC_WAR) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     ll      %1, %2          # local_sub_if_positive\n"
> +             "       subu    %0, %1, %3                              \n"
> +             "       bltz    %0, 1f                                  \n"
> +             "       sc      %0, %2                                  \n"
> +             "       .set    noreorder                               \n"
> +             "       beqzl   %0, 1b                                  \n"
> +             "        subu   %0, %1, %3                              \n"
> +             "       .set    reorder                                 \n"
> +             "1:                                                     \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else if (cpu_has_llsc) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     ll      %1, %2          # local_sub_if_positive\n"
> +             "       subu    %0, %1, %3                              \n"
> +             "       bltz    %0, 1f                                  \n"
> +             "       sc      %0, %2                                  \n"
> +             "       .set    noreorder                               \n"
> +             "       beqz    %0, 1b                                  \n"
> +             "        subu   %0, %1, %3                              \n"
> +             "       .set    reorder                                 \n"
> +             "1:                                                     \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else {
> +             unsigned long flags;
> +
> +             local_irq_save(flags);
> +             result = &(l->a.counter);
> +             result -= i;
> +             if (result >= 0)
> +                     &(l->a.counter) = result;
> +             local_irq_restore(flags);
> +     }
> +
> +     return result;
> +}
> +
> +#define local_cmpxchg(l, o, n) \
> +     ((long)cmpxchg(&((l)->a.counter), (o), (n)))
> +#define local_xchg(l, new) (xchg(&((l)->a.counter), new))
> +
> +/**
> + * local_add_unless - add unless the number is a given value
> + * @l: pointer of type local_t
> + * @a: the amount to add to l...
> + * @u: ...unless l is equal to u.
> + *
> + * Atomically adds @a to @l, so long as it was not @u.
> + * Returns non-zero if @l was not @u, and zero otherwise.
> + */
> +#define local_add_unless(l, a, u)                            \
> +({                                                           \
> +     long c, old;                                            \
> +     c = local_read(l);                                      \
> +     while (c != (u) && (old = local_cmpxchg((l), c, c + (a))) != c) \
> +             c = old;                                        \
> +     c != (u);                                               \
> +})
> +#define local_inc_not_zero(l) local_add_unless((l), 1, 0)
> +
> +#define local_dec_return(l) local_sub_return(1,(l))
> +#define local_inc_return(l) local_add_return(1,(l))
> +
> +/*
> + * local_sub_and_test - subtract value from variable and test result
> + * @i: integer value to subtract
> + * @l: pointer of type local_t
> + *
> + * Atomically subtracts @i from @l and returns
> + * true if the result is zero, or false for all
> + * other cases.
> + */
> +#define local_sub_and_test(i,l) (local_sub_return((i), (l)) == 0)
> +
> +/*
> + * local_inc_and_test - increment and test
> + * @l: pointer of type local_t
> + *
> + * Atomically increments @l by 1
> + * and returns true if the result is zero, or false for all
> + * other cases.
> + */
> +#define local_inc_and_test(l) (local_inc_return(l) == 0)
> +
> +/*
> + * local_dec_and_test - decrement by 1 and test
> + * @l: pointer of type local_t
> + *
> + * Atomically decrements @l by 1 and
> + * returns true if the result is 0, or false for all other
> + * cases.
> + */
> +#define local_dec_and_test(l) (local_sub_return(1, (l)) == 0)
> +
> +/*
> + * local_dec_if_positive - decrement by 1 if old value positive
> + * @l: pointer of type local_t
> + */
> +#define local_dec_if_positive(l)     local_sub_if_positive(1, l)
> +
> +/*
> + * local_add_negative - add and test if negative
> + * @l: pointer of type local_t
> + * @i: integer value to add
> + *
> + * Atomically adds @i to @l and returns true
> + * if the result is negative, or false when
> + * result is greater than or equal to zero.
> + */
> +#define local_add_negative(i,l) (local_add_return(i, (l)) < 0)
> +
> +#else /* CONFIG_64BITS */
>  
> -#define LOCAL_INIT(i)        ATOMIC64_INIT(i)
> -#define local_read(v)        atomic64_read(v)
> -#define local_set(v,i)       atomic64_set(v,i)
> +/*
> + * Same as above, but return the result value
> + */
> +static __inline__ long local_add_return(long i, local_t * l)
> +{
> +     unsigned long result;
> +
> +     if (cpu_has_llsc && R10000_LLSC_WAR) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     lld     %1, %2          # local_add_return      \n"
> +             "       addu    %0, %1, %3                              \n"
> +             "       scd     %0, %2                                  \n"
> +             "       beqzl   %0, 1b                                  \n"
> +             "       addu    %0, %1, %3                              \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else if (cpu_has_llsc) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     lld     %1, %2          # local_add_return      \n"
> +             "       addu    %0, %1, %3                              \n"
> +             "       scd     %0, %2                                  \n"
> +             "       beqz    %0, 1b                                  \n"
> +             "       addu    %0, %1, %3                              \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else {
> +             unsigned long flags;
> +
> +             local_irq_save(flags);
> +             result = &(l->a.counter);
> +             result += i;
> +             &(l->a.counter) = result;
> +             local_irq_restore(flags);
> +     }
>  
> -#define local_inc(v) atomic64_inc(v)
> -#define local_dec(v) atomic64_dec(v)
> -#define local_add(i, v)      atomic64_add(i, v)
> -#define local_sub(i, v)      atomic64_sub(i, v)
> +     return result;
> +}
>  
> -#endif
> +static __inline__ long local_sub_return(long i, local_t * l)
> +{
> +     unsigned long result;
>  
> -#define __local_inc(v)               ((v)->counter++)
> -#define __local_dec(v)               ((v)->counter--)
> -#define __local_add(i,v)     ((v)->counter+=(i))
> -#define __local_sub(i,v)     ((v)->counter-=(i))
> +     if (cpu_has_llsc && R10000_LLSC_WAR) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     lld     %1, %2          # local_sub_return      \n"
> +             "       subu    %0, %1, %3                              \n"
> +             "       scd     %0, %2                                  \n"
> +             "       beqzl   %0, 1b                                  \n"
> +             "       subu    %0, %1, %3                              \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else if (cpu_has_llsc) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     lld     %1, %2          # local_sub_return      \n"
> +             "       subu    %0, %1, %3                              \n"
> +             "       scd     %0, %2                                  \n"
> +             "       beqz    %0, 1b                                  \n"
> +             "       subu    %0, %1, %3                              \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else {
> +             unsigned long flags;
> +
> +             local_irq_save(flags);
> +             result = &(l->a.counter);
> +             result -= i;
> +             &(l->a.counter) = result;
> +             local_irq_restore(flags);
> +     }
> +
> +     return result;
> +}
>  
>  /*
> - * Use these for per-cpu local_t variables: on some archs they are
> + * local_sub_if_positive - conditionally subtract integer from atomic 
> variable
> + * @i: integer value to subtract
> + * @l: pointer of type local_t
> + *
> + * Atomically test @l and subtract @i if @l is greater or equal than @i.
> + * The function returns the old value of @l minus @i.
> + */
> +static __inline__ long local_sub_if_positive(long i, local_t * l)
> +{
> +     unsigned long result;
> +
> +     if (cpu_has_llsc && R10000_LLSC_WAR) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     lld     %1, %2          # local_sub_if_positive\n"
> +             "       dsubu   %0, %1, %3                              \n"
> +             "       bltz    %0, 1f                                  \n"
> +             "       scd     %0, %2                                  \n"
> +             "       .set    noreorder                               \n"
> +             "       beqzl   %0, 1b                                  \n"
> +             "        dsubu  %0, %1, %3                              \n"
> +             "       .set    reorder                                 \n"
> +             "1:                                                     \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else if (cpu_has_llsc) {
> +             unsigned long temp;
> +
> +             __asm__ __volatile__(
> +             "       .set    mips3                                   \n"
> +             "1:     lld     %1, %2          # local_sub_if_positive\n"
> +             "       dsubu   %0, %1, %3                              \n"
> +             "       bltz    %0, 1f                                  \n"
> +             "       scd     %0, %2                                  \n"
> +             "       .set    noreorder                               \n"
> +             "       beqz    %0, 1b                                  \n"
> +             "        dsubu  %0, %1, %3                              \n"
> +             "       .set    reorder                                 \n"
> +             "1:                                                     \n"
> +             "       .set    mips0                                   \n"
> +             : "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +             : "Ir" (i), "m" (&(l->a.counter))
> +             : "memory");
> +     } else {
> +             unsigned long flags;
> +
> +             local_irq_save(flags);
> +             result = &(l->a.counter);
> +             result -= i;
> +             if (result >= 0)
> +                     &(l->a.counter) = result;
> +             local_irq_restore(flags);
> +     }
> +
> +     return result;
> +}
> +
> +
> +#define local_cmpxchg(l, o, n) \
> +     ((long)cmpxchg(&((l)->a.counter), (o), (n)))
> +#define local_xchg(l, new) (xchg(&((l)->a.counter), new))
> +
> +/**
> + * local_add_unless - add unless the number is a given value
> + * @l: pointer of type local_t
> + * @a: the amount to add to l...
> + * @u: ...unless l is equal to u.
> + *
> + * Atomically adds @a to @l, so long as it was not @u.
> + * Returns non-zero if @l was not @u, and zero otherwise.
> + */
> +#define local_add_unless(l, a, u)                            \
> +({                                                           \
> +     long c, old;                                            \
> +     c = local_read(l);                                      \
> +     while (c != (u) && (old = local_cmpxchg((l), c, c + (a))) != c) \
> +             c = old;                                        \
> +     c != (u);                                               \
> +})
> +#define local_inc_not_zero(l) local_add_unless((l), 1, 0)
> +
> +#define local_dec_return(l) local_sub_return(1,(l))
> +#define local_inc_return(l) local_add_return(1,(l))
> +
> +/*
> + * local_sub_and_test - subtract value from variable and test result
> + * @i: integer value to subtract
> + * @l: pointer of type local_t
> + *
> + * Atomically subtracts @i from @l and returns
> + * true if the result is zero, or false for all
> + * other cases.
> + */
> +#define local_sub_and_test(i,l) (local_sub_return((i), (l)) == 0)
> +
> +/*
> + * local_inc_and_test - increment and test
> + * @l: pointer of type local_t
> + *
> + * Atomically increments @l by 1
> + * and returns true if the result is zero, or false for all
> + * other cases.
> + */
> +#define local_inc_and_test(l) (local_inc_return(l) == 0)
> +
> +/*
> + * local_dec_and_test - decrement by 1 and test
> + * @l: pointer of type local_t
> + *
> + * Atomically decrements @l by 1 and
> + * returns true if the result is 0, or false for all other
> + * cases.
> + */
> +#define local_dec_and_test(l) (local_sub_return(1, (l)) == 0)
> +
> +/*
> + * local_dec_if_positive - decrement by 1 if old value positive
> + * @l: pointer of type local_t
> + */
> +#define local_dec_if_positive(l)     local_sub_if_positive(1, l)
> +
> +/*
> + * local_add_negative - add and test if negative
> + * @l: pointer of type local_t
> + * @i: integer value to add
> + *
> + * Atomically adds @i to @l and returns true
> + * if the result is negative, or false when
> + * result is greater than or equal to zero.
> + */
> +#define local_add_negative(i,l) (local_add_return(i, (l)) < 0)
> +
> +#endif /* !CONFIG_64BITS */
> +
> +
> +/* Use these for per-cpu local_t variables: on some archs they are
>   * much more efficient than these naive implementations.  Note they take
>   * a variable, not an address.
> + *
> + * This could be done better if we moved the per cpu data directly
> + * after GS.
>   */
> -#define cpu_local_read(v)    local_read(&__get_cpu_var(v))
> -#define cpu_local_set(v, i)  local_set(&__get_cpu_var(v), (i))
>  
> -#define cpu_local_inc(v)     local_inc(&__get_cpu_var(v))
> -#define cpu_local_dec(v)     local_dec(&__get_cpu_var(v))
> -#define cpu_local_add(i, v)  local_add((i), &__get_cpu_var(v))
> -#define cpu_local_sub(i, v)  local_sub((i), &__get_cpu_var(v))
> +#define __local_inc(l)               ((l)->a.counter++)
> +#define __local_dec(l)               ((l)->a.counter++)
> +#define __local_add(i,l)     ((l)->a.counter+=(i))
> +#define __local_sub(i,l)     ((l)->a.counter-=(i))
> +
> +/* Need to disable preemption for the cpu local counters otherwise we could
> +   still access a variable of a previous CPU in a non atomic way. */
> +#define cpu_local_wrap_v(l)          \
> +     ({ local_t res__;               \
> +        preempt_disable();           \
> +        res__ = (l);                 \
> +        preempt_enable();            \
> +        res__; })
> +#define cpu_local_wrap(l)            \
> +     ({ preempt_disable();           \
> +        l;                           \
> +        preempt_enable(); })         \
> +
> +#define cpu_local_read(l)    cpu_local_wrap_v(local_read(&__get_cpu_var(l)))
> +#define cpu_local_set(l, i)  cpu_local_wrap(local_set(&__get_cpu_var(l), 
> (i)))
> +#define cpu_local_inc(l)     cpu_local_wrap(local_inc(&__get_cpu_var(l)))
> +#define cpu_local_dec(l)     cpu_local_wrap(local_dec(&__get_cpu_var(l)))
> +#define cpu_local_add(i, l)  cpu_local_wrap(local_add((i), 
> &__get_cpu_var(l)))
> +#define cpu_local_sub(i, l)  cpu_local_wrap(local_sub((i), 
> &__get_cpu_var(l)))
>  
> -#define __cpu_local_inc(v)   __local_inc(&__get_cpu_var(v))
> -#define __cpu_local_dec(v)   __local_dec(&__get_cpu_var(v))
> -#define __cpu_local_add(i, v)        __local_add((i), &__get_cpu_var(v))
> -#define __cpu_local_sub(i, v)        __local_sub((i), &__get_cpu_var(v))
> +#define __cpu_local_inc(l)   cpu_local_inc(l)
> +#define __cpu_local_dec(l)   cpu_local_dec(l)
> +#define __cpu_local_add(i, l)        cpu_local_add((i), (l))
> +#define __cpu_local_sub(i, l)        cpu_local_sub((i), (l))
>  
> -#endif /* _ASM_LOCAL_H */
> +#endif /* _ARCH_POWERPC_LOCAL_H */
> 

-- 
OpenPGP public key:              http://krystal.dyndns.org:8080/key/compudj.gpg
Key fingerprint:     8CD5 52C3 8E3C 4140 715F  BA06 3F25 A8FE 3BAE 9A68 
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to