Re: [PATCH v2 2/2] KVM: nSVM: improve SYSENTER emulation on AMD
Paolo Bonzini writes: > On 01/04/21 15:03, Vitaly Kuznetsov wrote: >>> + svm->sysenter_eip_hi = guest_cpuid_is_intel(vcpu) ? (data >> >>> 32) : 0; >> >> (Personal taste) I'd suggest we keep the whole 'sysenter_eip'/'sysenter_esp' >> even if we only use the upper 32 bits of it. That would reduce the code >> churn a little bit (no need to change 'struct vcpu_svm'). > > Would there really be less changes? Consider that you'd have to look at > the VMCB anyway because svm_get_msr can be reached not just for guest > RDMSR but also for ioctls. > I was thinking about the hunk in arch/x86/kvm/svm/svm.h tweaking vcpu_svm. My opinion is not strong at all here) -- Vitaly
Re: [PATCH v2 2/2] KVM: nSVM: improve SYSENTER emulation on AMD
Maxim Levitsky writes: > Currently to support Intel->AMD migration, if CPU vendor is GenuineIntel, > we emulate the full 64 value for MSR_IA32_SYSENTER_{EIP|ESP} > msrs, and we also emulate the sysenter/sysexit instruction in long mode. > > (Emulator does still refuse to emulate sysenter in 64 bit mode, on the > ground that the code for that wasn't tested and likely has no users) > > However when virtual vmload/vmsave is enabled, the vmload instruction will > update these 32 bit msrs without triggering their msr intercept, > which will lead to having stale values in kvm's shadow copy of these msrs, > which relies on the intercept to be up to date. > > Fix/optimize this by doing the following: > > 1. Enable the MSR intercepts for SYSENTER MSRs iff vendor=GenuineIntel >(This is both a tiny optimization and also ensures that in case >the guest cpu vendor is AMD, the msrs will be 32 bit wide as >AMD defined). > > 2. Store only high 32 bit part of these msrs on interception and combine >it with hardware msr value on intercepted read/writes >iff vendor=GenuineIntel. > > 3. Disable vmload/vmsave virtualization if vendor=GenuineIntel. >(It is somewhat insane to set vendor=GenuineIntel and still enable >SVM for the guest but well whatever). >Then zero the high 32 bit parts when kvm intercepts and emulates vmload. > > Thanks a lot to Paulo Bonzini for helping me with fixing this in the most s/Paulo/Paolo/ :-) > correct way. > > This patch fixes nested migration of 32 bit nested guests, that was > broken because incorrect cached values of SYSENTER msrs were stored in > the migration stream if L1 changed these msrs with > vmload prior to L2 entry. > > Signed-off-by: Maxim Levitsky > --- > arch/x86/kvm/svm/svm.c | 99 +++--- > arch/x86/kvm/svm/svm.h | 6 +-- > 2 files changed, 68 insertions(+), 37 deletions(-) > > diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c > index 271196400495..6c39b0cd6ec6 100644 > --- a/arch/x86/kvm/svm/svm.c > +++ b/arch/x86/kvm/svm/svm.c > @@ -95,6 +95,8 @@ static const struct svm_direct_access_msrs { > } direct_access_msrs[MAX_DIRECT_ACCESS_MSRS] = { > { .index = MSR_STAR,.always = true }, > { .index = MSR_IA32_SYSENTER_CS,.always = true }, > + { .index = MSR_IA32_SYSENTER_EIP, .always = false }, > + { .index = MSR_IA32_SYSENTER_ESP, .always = false }, > #ifdef CONFIG_X86_64 > { .index = MSR_GS_BASE, .always = true }, > { .index = MSR_FS_BASE, .always = true }, > @@ -1258,16 +1260,6 @@ static void init_vmcb(struct kvm_vcpu *vcpu) > if (kvm_vcpu_apicv_active(vcpu)) > avic_init_vmcb(svm); > > - /* > - * If hardware supports Virtual VMLOAD VMSAVE then enable it > - * in VMCB and clear intercepts to avoid #VMEXIT. > - */ > - if (vls) { > - svm_clr_intercept(svm, INTERCEPT_VMLOAD); > - svm_clr_intercept(svm, INTERCEPT_VMSAVE); > - svm->vmcb->control.virt_ext |= > VIRTUAL_VMLOAD_VMSAVE_ENABLE_MASK; > - } > - > if (vgif) { > svm_clr_intercept(svm, INTERCEPT_STGI); > svm_clr_intercept(svm, INTERCEPT_CLGI); > @@ -2133,9 +2125,11 @@ static int vmload_vmsave_interception(struct kvm_vcpu > *vcpu, bool vmload) > > ret = kvm_skip_emulated_instruction(vcpu); > > - if (vmload) > + if (vmload) { > nested_svm_vmloadsave(vmcb12, svm->vmcb); > - else > + svm->sysenter_eip_hi = 0; > + svm->sysenter_esp_hi = 0; > + } else > nested_svm_vmloadsave(svm->vmcb, vmcb12); Nitpicking: {} are now needed for both branches here. > > kvm_vcpu_unmap(vcpu, , true); > @@ -2677,10 +2671,14 @@ static int svm_get_msr(struct kvm_vcpu *vcpu, struct > msr_data *msr_info) > msr_info->data = svm->vmcb01.ptr->save.sysenter_cs; > break; > case MSR_IA32_SYSENTER_EIP: > - msr_info->data = svm->sysenter_eip; > + msr_info->data = (u32)svm->vmcb01.ptr->save.sysenter_eip; > + if (guest_cpuid_is_intel(vcpu)) > + msr_info->data |= (u64)svm->sysenter_eip_hi << 32; > break; > case MSR_IA32_SYSENTER_ESP: > - msr_info->data = svm->sysenter_esp; > + msr_info->data = svm->vmcb01.ptr->save.sysenter_esp; > + if (guest_cpuid_is_intel(vcpu)) > + msr_info->data |= (u64)svm->sysenter_esp_hi << 32; > break; > case MSR_TSC_AUX: > if (!boot_cpu_has(X86_FEATURE_RDTSCP)) > @@ -2885,12 +2883,19 @@ static int svm_set_msr(struct kvm_vcpu *vcpu, struct > msr_data *msr) > svm->vmcb01.ptr->save.sysenter_cs = data; > break; > case MSR_IA32_SYSENTER_EIP: > - svm->sysenter_eip =
Re: [PATCH v2 2/2] KVM: nSVM: improve SYSENTER emulation on AMD
On 01/04/21 17:31, Vitaly Kuznetsov wrote: + svm->sysenter_eip_hi = guest_cpuid_is_intel(vcpu) ? (data >> 32) : 0; (Personal taste) I'd suggest we keep the whole 'sysenter_eip'/'sysenter_esp' even if we only use the upper 32 bits of it. That would reduce the code churn a little bit (no need to change 'struct vcpu_svm'). Would there really be less changes? Consider that you'd have to look at the VMCB anyway because svm_get_msr can be reached not just for guest RDMSR but also for ioctls. I was thinking about the hunk in arch/x86/kvm/svm/svm.h tweaking vcpu_svm. My opinion is not strong at all here) Ah okay, if it's just that I would consider the change a benefit, not a problem with this patch. Paolo
Re: [PATCH v2 2/2] KVM: nSVM: improve SYSENTER emulation on AMD
On 01/04/21 15:03, Vitaly Kuznetsov wrote: + svm->sysenter_eip_hi = guest_cpuid_is_intel(vcpu) ? (data >> 32) : 0; (Personal taste) I'd suggest we keep the whole 'sysenter_eip'/'sysenter_esp' even if we only use the upper 32 bits of it. That would reduce the code churn a little bit (no need to change 'struct vcpu_svm'). Would there really be less changes? Consider that you'd have to look at the VMCB anyway because svm_get_msr can be reached not just for guest RDMSR but also for ioctls. Paolo
Re: [PATCH v2 2/2] KVM: nSVM: improve SYSENTER emulation on AMD
On Thu, 2021-04-01 at 15:03 +0200, Vitaly Kuznetsov wrote: > Maxim Levitsky writes: > > > Currently to support Intel->AMD migration, if CPU vendor is GenuineIntel, > > we emulate the full 64 value for MSR_IA32_SYSENTER_{EIP|ESP} > > msrs, and we also emulate the sysenter/sysexit instruction in long mode. > > > > (Emulator does still refuse to emulate sysenter in 64 bit mode, on the > > ground that the code for that wasn't tested and likely has no users) > > > > However when virtual vmload/vmsave is enabled, the vmload instruction will > > update these 32 bit msrs without triggering their msr intercept, > > which will lead to having stale values in kvm's shadow copy of these msrs, > > which relies on the intercept to be up to date. > > > > Fix/optimize this by doing the following: > > > > 1. Enable the MSR intercepts for SYSENTER MSRs iff vendor=GenuineIntel > >(This is both a tiny optimization and also ensures that in case > >the guest cpu vendor is AMD, the msrs will be 32 bit wide as > >AMD defined). > > > > 2. Store only high 32 bit part of these msrs on interception and combine > >it with hardware msr value on intercepted read/writes > >iff vendor=GenuineIntel. > > > > 3. Disable vmload/vmsave virtualization if vendor=GenuineIntel. > >(It is somewhat insane to set vendor=GenuineIntel and still enable > >SVM for the guest but well whatever). > >Then zero the high 32 bit parts when kvm intercepts and emulates vmload. > > > > Thanks a lot to Paulo Bonzini for helping me with fixing this in the most > > s/Paulo/Paolo/ :-) Sorry about that! > > > correct way. > > > > This patch fixes nested migration of 32 bit nested guests, that was > > broken because incorrect cached values of SYSENTER msrs were stored in > > the migration stream if L1 changed these msrs with > > vmload prior to L2 entry. > > > > Signed-off-by: Maxim Levitsky > > --- > > arch/x86/kvm/svm/svm.c | 99 +++--- > > arch/x86/kvm/svm/svm.h | 6 +-- > > 2 files changed, 68 insertions(+), 37 deletions(-) > > > > diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c > > index 271196400495..6c39b0cd6ec6 100644 > > --- a/arch/x86/kvm/svm/svm.c > > +++ b/arch/x86/kvm/svm/svm.c > > @@ -95,6 +95,8 @@ static const struct svm_direct_access_msrs { > > } direct_access_msrs[MAX_DIRECT_ACCESS_MSRS] = { > > { .index = MSR_STAR,.always = true }, > > { .index = MSR_IA32_SYSENTER_CS,.always = true }, > > + { .index = MSR_IA32_SYSENTER_EIP, .always = false }, > > + { .index = MSR_IA32_SYSENTER_ESP, .always = false }, > > #ifdef CONFIG_X86_64 > > { .index = MSR_GS_BASE, .always = true }, > > { .index = MSR_FS_BASE, .always = true }, > > @@ -1258,16 +1260,6 @@ static void init_vmcb(struct kvm_vcpu *vcpu) > > if (kvm_vcpu_apicv_active(vcpu)) > > avic_init_vmcb(svm); > > > > - /* > > -* If hardware supports Virtual VMLOAD VMSAVE then enable it > > -* in VMCB and clear intercepts to avoid #VMEXIT. > > -*/ > > - if (vls) { > > - svm_clr_intercept(svm, INTERCEPT_VMLOAD); > > - svm_clr_intercept(svm, INTERCEPT_VMSAVE); > > - svm->vmcb->control.virt_ext |= > > VIRTUAL_VMLOAD_VMSAVE_ENABLE_MASK; > > - } > > - > > if (vgif) { > > svm_clr_intercept(svm, INTERCEPT_STGI); > > svm_clr_intercept(svm, INTERCEPT_CLGI); > > @@ -2133,9 +2125,11 @@ static int vmload_vmsave_interception(struct > > kvm_vcpu *vcpu, bool vmload) > > > > ret = kvm_skip_emulated_instruction(vcpu); > > > > - if (vmload) > > + if (vmload) { > > nested_svm_vmloadsave(vmcb12, svm->vmcb); > > - else > > + svm->sysenter_eip_hi = 0; > > + svm->sysenter_esp_hi = 0; > > + } else > > nested_svm_vmloadsave(svm->vmcb, vmcb12); > > Nitpicking: {} are now needed for both branches here. I didn't knew about this rule, and I'll take this into account next time. Thanks! Best regards, Maxim Levitsky