On Thu, Mar 26, 2015 at 02:22:20PM +0000, Peter Zijlstra wrote:
> On Thu, Mar 26, 2015 at 01:27:50PM +0000, Will Deacon wrote:
> > On Thu, Mar 26, 2015 at 10:34:42AM +0000, Peter Zijlstra wrote:
> > > On Thu, Mar 26, 2015 at 07:31:12PM +1100, Stephen Rothwell wrote:
> > > > In function '__read_once_size',
> > > >     inlined from 'lockref_get' at lib/lockref.c:50:2:
> 
> 
> > Yeah, I think it's fine because, as you point out, the cmpxchg can only
> > succeed if the 64-bit load appeared to be single-copy atomic (amongst other
> > things).
> 
> So one option to get rid of this warning is to rely on the fact that all
> CMPXCHG_LOOP users are at the beginning of !pure function calls, which
> already imply a compiler barrier and therefore it must already emit that
> load.
> 
> And as already argued, split loads aren't an issue because the cmpxchg
> will catch those for us.
> 
> So we can either just remove the READ_ONCE(), or replace it with a
> leading barrier() call just to be on the paranoid side of things.

If we remove the READ_ONCE then I think the barrier is a good idea, just in
case the LTO guys get their paws on this and we see subtle breakage.

> Any preferences?
> 
> Something like so, but with a sensible comment I suppose.
> 
> ---
>  lib/lockref.c | 3 ++-
>  1 file changed, 2 insertions(+), 1 deletion(-)
> 
> diff --git a/lib/lockref.c b/lib/lockref.c
> index 494994bf17c8..b5ca1f65c8a3 100644
> --- a/lib/lockref.c
> +++ b/lib/lockref.c
> @@ -18,7 +18,8 @@
>  #define CMPXCHG_LOOP(CODE, SUCCESS) do {                                     
> \
>       struct lockref old;                                                     
> \
>       BUILD_BUG_ON(sizeof(old) != 8);                                         
> \
> -     old.lock_count = READ_ONCE(lockref->lock_count);                        
> \
> +     barrier();                                                              
> \
> +     old.lock_count = lockref->lock_count;                                   
> \
>       while (likely(arch_spin_value_unlocked(old.lock.rlock.raw_lock))) {     
> \
>               struct lockref new = old, prev = old;                           
> \
>               CODE                                                            
> \

Is ACCESS_ONCE actually going away? It has its problems, but I think it's
what we want here and reads better than magic barrier() imo.

Will
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to