----- On Jan 28, 2019, at 5:39 PM, paulmck paul...@linux.ibm.com wrote: > On Mon, Jan 28, 2019 at 05:07:07PM -0500, Mathieu Desnoyers wrote: >> Jann Horn identified a racy access to p->mm in the global expedited >> command of the membarrier system call. >> >> The suggested fix is to hold the task_lock() around the accesses to >> p->mm and to the mm_struct membarrier_state field to guarantee the >> existence of the mm_struct. >> >> Link: >> https://lore.kernel.org/lkml/cag48ez2g8ctf8dhs42tf37pthfr3y0rnooytmxvacm4u8yu...@mail.gmail.com >> Signed-off-by: Mathieu Desnoyers <mathieu.desnoy...@efficios.com> >> Tested-by: Jann Horn <ja...@google.com> >> CC: Jann Horn <ja...@google.com> >> CC: Thomas Gleixner <t...@linutronix.de> >> CC: Peter Zijlstra (Intel) <pet...@infradead.org> >> CC: Ingo Molnar <mi...@kernel.org> >> CC: Andrea Parri <parri.and...@gmail.com> >> CC: Andy Lutomirski <l...@kernel.org> >> CC: Avi Kivity <a...@scylladb.com> >> CC: Benjamin Herrenschmidt <b...@kernel.crashing.org> >> CC: Boqun Feng <boqun.f...@gmail.com> >> CC: Dave Watson <davejwat...@fb.com> >> CC: David Sehr <s...@google.com> >> CC: H. Peter Anvin <h...@zytor.com> >> CC: Linus Torvalds <torva...@linux-foundation.org> >> CC: Maged Michael <maged.mich...@gmail.com> >> CC: Michael Ellerman <m...@ellerman.id.au> >> CC: Paul E. McKenney <paul...@linux.vnet.ibm.com> >> CC: Paul Mackerras <pau...@samba.org> >> CC: Russell King <li...@armlinux.org.uk> >> CC: Will Deacon <will.dea...@arm.com> >> CC: sta...@vger.kernel.org # v4.16+ >> CC: linux-...@vger.kernel.org >> --- >> kernel/sched/membarrier.c | 27 +++++++++++++++++++++------ >> 1 file changed, 21 insertions(+), 6 deletions(-) >> >> diff --git a/kernel/sched/membarrier.c b/kernel/sched/membarrier.c >> index 76e0eaf4654e..305fdcc4c5f7 100644 >> --- a/kernel/sched/membarrier.c >> +++ b/kernel/sched/membarrier.c >> @@ -81,12 +81,27 @@ static int membarrier_global_expedited(void) >> >> rcu_read_lock(); >> p = task_rcu_dereference(&cpu_rq(cpu)->curr); >> - if (p && p->mm && (atomic_read(&p->mm->membarrier_state) & >> - MEMBARRIER_STATE_GLOBAL_EXPEDITED)) { >> - if (!fallback) >> - __cpumask_set_cpu(cpu, tmpmask); >> - else >> - smp_call_function_single(cpu, ipi_mb, NULL, 1); >> + /* >> + * Skip this CPU if the runqueue's current task is NULL or if >> + * it is a kernel thread. >> + */ >> + if (p && READ_ONCE(p->mm)) { >> + bool mm_match; >> + >> + /* >> + * Read p->mm and access membarrier_state while holding >> + * the task lock to ensure existence of mm. >> + */ >> + task_lock(p); >> + mm_match = p->mm && >> (atomic_read(&p->mm->membarrier_state) & > > Are we guaranteed that this p->mm will be the same as the one loaded via > READ_ONCE() above? Either way, wouldn't it be better to READ_ONCE() it a > single time and use the same value everywhere?
The first "READ_ONCE()" above is _outside_ of the task_lock() critical section. Those two accesses _can_ load two different pointers, and this is why we need to re-read the p->mm pointer within the task_lock() critical section to ensure existence of the mm_struct that we use. If we move the READ_ONCE() into the task_lock(), we need to uselessly take a lock before we can skip kernel threads. If we lead the READ_ONCE() outside the task_lock(), then p->mm can be updated between the READ_ONCE() and reference to the mm_struct content within the task_lock(), which is racy and does not guarantee its existence. Or am I missing your point ? Thanks, Mathieu > > Thanx, Paul > >> + MEMBARRIER_STATE_GLOBAL_EXPEDITED); >> + task_unlock(p); >> + if (mm_match) { >> + if (!fallback) >> + __cpumask_set_cpu(cpu, tmpmask); >> + else >> + smp_call_function_single(cpu, ipi_mb, >> NULL, 1); >> + } >> } >> rcu_read_unlock(); >> } >> -- >> 2.17.1 -- Mathieu Desnoyers EfficiOS Inc. http://www.efficios.com