> On 13 May 2019, at 21:17, Andy Lutomirski <l...@kernel.org> wrote:
> 
>> I expect that the KVM address space can eventually be expanded to include
>> the ioctl syscall entries. By doing so, and also adding the KVM page table
>> to the process userland page table (which should be safe to do because the
>> KVM address space doesn't have any secret), we could potentially handle the
>> KVM ioctl without having to switch to the kernel pagetable (thus effectively
>> eliminating KPTI for KVM). Then the only overhead would be if a VM-Exit has
>> to be handled using the full kernel address space.
>> 
> 
> In the hopefully common case where a VM exits and then gets re-entered
> without needing to load full page tables, what code actually runs?
> I'm trying to understand when the optimization of not switching is
> actually useful.
> 
> Allowing ioctl() without switching to kernel tables sounds...
> extremely complicated.  It also makes the dubious assumption that user
> memory contains no secrets.

Let me attempt to clarify what we were thinking when creating this patch series:

1) It is never safe to execute one hyperthread inside guest while it’s sibling 
hyperthread runs in a virtual address space which contains secrets of host or 
other guests.
This is because we assume that using some speculative gadget (such as 
half-Spectrev2 gadget), it will be possible to populate *some* CPU core 
resource which could then be *somehow* leaked by the hyperthread running inside 
guest. In case of L1TF, this would be data populated to the L1D cache.

2) Because of (1), every time a hyperthread runs inside host kernel, we must 
make sure it’s sibling is not running inside guest. i.e. We must kick the 
sibling hyperthread outside of guest using IPI.

3) From (2), we should have theoretically deduced that for every #VMExit, there 
is a need to kick the sibling hyperthread also outside of guest until the 
#VMExit is completed. Such a patch series was implemented at some point but it 
had (obviously) significant performance hit.

4) The main goal of this patch series is to preserve (2), but to avoid the 
overhead specified in (3).

The way this patch series achieves (4) is by observing that during the run of a 
VM, most #VMExits can be handled rather quickly and locally inside KVM and 
doesn’t need to reference any data that is not relevant to this VM or KVM code. 
Therefore, if we will run these #VMExits in an isolated virtual address space 
(i.e. KVM isolated address space), there is no need to kick the sibling 
hyperthread from guest while these #VMExits handlers run.
The hope is that the very vast majority of #VMExit handlers will be able to 
completely run without requiring to switch to full address space. Therefore, 
avoiding the performance hit of (2).
However, for the very few #VMExits that does require to run in full kernel 
address space, we must first kick the sibling hyperthread outside of guest and 
only then switch to full kernel address space and only once all hyperthreads 
return to KVM address space, then allow then to enter into guest.

From this reason, I think the above paragraph (that was added to my original 
cover letter) is incorrect.
I believe that we should by design treat all exits to userspace VMM (e.g. QEMU) 
as slow-path that should not be optimised and therefore ok to switch address 
space (and therefore also kick sibling hyperthread). Similarly, all IOCTLs 
handlers are also slow-path and therefore it should be ok for them to also not 
run in KVM isolated address space.

-Liran









Reply via email to