On Sat, May 27, 2017 at 9:00 AM, Andy Lutomirski <l...@kernel.org> wrote: > On Sat, May 27, 2017 at 6:31 AM, kernel test robot > <xiaolong...@intel.com> wrote: >> >> FYI, we noticed the following commit: >> >> commit: e2a7dcce31f10bd7471b4245a6d1f2de344e7adf ("x86/mm: Rework lazy TLB >> to track the actual loaded mm") >> https://git.kernel.org/cgit/linux/kernel/git/luto/linux.git >> x86/tlbflush_cleanup > > Ugh, there's an unpleasant interaction between this patch and > intel_idle. I suspect that the intel_idle code in question is either > wrong or pointless, but I want to investigate further. Ingo, can you > hold off on applying this patch?
I think this is what's going on: intel_idle has an optimization and sometimes calls leave_mm(). This is a rather expensive way of working around x86 Linux's fairly weak lazy mm handling. It also abuses the whole switch_mm state machine. In particular, there's no guarantee that the mm is actually lazy at the time. The old code didn't care, but the new code can oops. The short-term fix is to just reorder the code in leave_mm() to avoid the OOPS. I think the long-term fix is to improve laziness and delete the leave_mm() call from intel_idle entirely. Doing that will be quite simple on top of the PCID. PCID needs to reliably track when the TLB is stale without needing IPIs for every invalidation, and laziness should be able to reuse the same mechanism. Once we do this, I think we may even be able to remove leave_mm() altogether. (Rik should like this, since it will let us suppress host TLB flush IPIs being sent to CPUs in guest mode as well.) --Andy