The pdptrs need to be cached in addition to the shadowed root tables, so
the guest walk can be done properly.
Signed-off-by: Marcelo Tosatti <[EMAIL PROTECTED]>
Index: kvm.first/arch/x86/kvm/mmu.c
===================================================================
--- kvm.first.orig/arch/x86/kvm/mmu.c
+++ kvm.first/arch/x86/kvm/mmu.c
@@ -1320,11 +1320,11 @@ static void mmu_alloc_roots(struct kvm_v
ASSERT(!VALID_PAGE(root));
if (vcpu->arch.mmu.root_level == PT32E_ROOT_LEVEL) {
- if (!is_present_pte(vcpu->arch.pdptrs[i])) {
+ if (!is_present_pte(vcpu->arch.pdptrs[j][i])) {
vcpu->arch.mmu.pae_root[j][i] = 0;
continue;
}
- root_gfn = vcpu->arch.pdptrs[i] >> PAGE_SHIFT;
+ root_gfn = vcpu->arch.pdptrs[j][i] >> PAGE_SHIFT;
} else if (vcpu->arch.mmu.root_level == 0)
root_gfn = 0;
sp = kvm_mmu_get_page(vcpu, root_gfn, i << 30,
Index: kvm.first/arch/x86/kvm/paging_tmpl.h
===================================================================
--- kvm.first.orig/arch/x86/kvm/paging_tmpl.h
+++ kvm.first/arch/x86/kvm/paging_tmpl.h
@@ -136,7 +136,8 @@ walk:
pte = vcpu->arch.cr3;
#if PTTYPE == 64
if (!is_long_mode(vcpu)) {
- pte = vcpu->arch.pdptrs[(addr >> 30) & 3];
+ pte = vcpu->arch.pdptrs[vcpu->arch.cr3_cache_idx]
+ [(addr >> 30) & 3];
if (!is_present_pte(pte))
goto not_present;
--walker->level;
Index: kvm.first/arch/x86/kvm/x86.c
===================================================================
--- kvm.first.orig/arch/x86/kvm/x86.c
+++ kvm.first/arch/x86/kvm/x86.c
@@ -192,13 +192,21 @@ static void __queue_exception(struct kvm
/*
* Load the pae pdptrs. Return true is they are all valid.
*/
-int load_pdptrs(struct kvm_vcpu *vcpu, unsigned long cr3)
+int load_pdptrs(struct kvm_vcpu *vcpu, unsigned long cr3, int cr3_cache_inc)
{
gfn_t pdpt_gfn = cr3 >> PAGE_SHIFT;
unsigned offset = ((cr3 & (PAGE_SIZE-1)) >> 5) << 2;
int i;
int ret;
- u64 pdpte[ARRAY_SIZE(vcpu->arch.pdptrs)];
+ u64 pdpte[ARRAY_SIZE(vcpu->arch.pdptrs[0])];
+ int idx = vcpu->arch.cr3_cache_idx;
+
+ idx++;
+ if (unlikely(idx >= vcpu->arch.cr3_cache_limit))
+ idx = 0;
+
+ if (cr3_cache_inc)
+ vcpu->arch.cr3_cache_idx = idx;
down_read(&vcpu->kvm->slots_lock);
ret = kvm_read_guest_page(vcpu->kvm, pdpt_gfn, pdpte,
@@ -215,7 +223,7 @@ int load_pdptrs(struct kvm_vcpu *vcpu, u
}
ret = 1;
- memcpy(vcpu->arch.pdptrs, pdpte, sizeof(vcpu->arch.pdptrs));
+ memcpy(vcpu->arch.pdptrs[idx], pdpte, sizeof(vcpu->arch.pdptrs[0]));
out:
up_read(&vcpu->kvm->slots_lock);
@@ -225,7 +233,7 @@ EXPORT_SYMBOL_GPL(load_pdptrs);
static bool pdptrs_changed(struct kvm_vcpu *vcpu)
{
- u64 pdpte[ARRAY_SIZE(vcpu->arch.pdptrs)];
+ u64 pdpte[ARRAY_SIZE(vcpu->arch.pdptrs[0])];
bool changed = true;
int r;
@@ -236,7 +244,8 @@ static bool pdptrs_changed(struct kvm_vc
r = kvm_read_guest(vcpu->kvm, vcpu->arch.cr3 & ~31u, pdpte,
sizeof(pdpte));
if (r < 0)
goto out;
- changed = memcmp(pdpte, vcpu->arch.pdptrs, sizeof(pdpte)) != 0;
+ changed = memcmp(pdpte, vcpu->arch.pdptrs[vcpu->arch.cr3_cache_idx],
+ sizeof(pdpte)) != 0;
out:
up_read(&vcpu->kvm->slots_lock);
@@ -286,7 +295,7 @@ void kvm_set_cr0(struct kvm_vcpu *vcpu,
}
} else
#endif
- if (is_pae(vcpu) && !load_pdptrs(vcpu, vcpu->arch.cr3)) {
+ if (is_pae(vcpu) && !load_pdptrs(vcpu, vcpu->arch.cr3, 1)) {
printk(KERN_DEBUG "set_cr0: #GP, pdptrs "
"reserved bits\n");
kvm_inject_gp(vcpu, 0);
@@ -325,7 +334,7 @@ void kvm_set_cr4(struct kvm_vcpu *vcpu,
return;
}
} else if (is_paging(vcpu) && !is_pae(vcpu) && (cr4 & X86_CR4_PAE)
- && !load_pdptrs(vcpu, vcpu->arch.cr3)) {
+ && !load_pdptrs(vcpu, vcpu->arch.cr3, 1)) {
printk(KERN_DEBUG "set_cr4: #GP, pdptrs reserved bits\n");
kvm_inject_gp(vcpu, 0);
return;
@@ -363,7 +372,7 @@ void kvm_set_cr3(struct kvm_vcpu *vcpu,
kvm_inject_gp(vcpu, 0);
return;
}
- if (is_paging(vcpu) && !load_pdptrs(vcpu, cr3)) {
+ if (is_paging(vcpu) && !load_pdptrs(vcpu, cr3, 0)) {
printk(KERN_DEBUG "set_cr3: #GP, pdptrs "
"reserved bits\n");
kvm_inject_gp(vcpu, 0);
@@ -3047,7 +3056,7 @@ int kvm_arch_vcpu_ioctl_set_sregs(struct
mmu_reset_needed |= vcpu->arch.cr4 != sregs->cr4;
kvm_x86_ops->set_cr4(vcpu, sregs->cr4);
if (!is_long_mode(vcpu) && is_pae(vcpu))
- load_pdptrs(vcpu, vcpu->arch.cr3);
+ load_pdptrs(vcpu, vcpu->arch.cr3, 1);
if (mmu_reset_needed)
kvm_mmu_reset_context(vcpu);
Index: kvm.first/include/asm-x86/kvm_host.h
===================================================================
--- kvm.first.orig/include/asm-x86/kvm_host.h
+++ kvm.first/include/asm-x86/kvm_host.h
@@ -211,7 +211,7 @@ struct kvm_vcpu_arch {
unsigned int cr3_cache_limit;
unsigned long cr4;
unsigned long cr8;
- u64 pdptrs[4]; /* pae */
+ u64 pdptrs[KVM_CR3_CACHE_SIZE][4]; /* pae */
u64 shadow_efer;
u64 apic_base;
struct kvm_lapic *apic; /* kernel irqchip context */
@@ -433,7 +433,7 @@ void kvm_mmu_zap_all(struct kvm *kvm);
unsigned int kvm_mmu_calculate_mmu_pages(struct kvm *kvm);
void kvm_mmu_change_mmu_pages(struct kvm *kvm, unsigned int kvm_nr_mmu_pages);
-int load_pdptrs(struct kvm_vcpu *vcpu, unsigned long cr3);
+int load_pdptrs(struct kvm_vcpu *vcpu, unsigned long cr3, int cr3_cache_inc);
int __emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa,
const void *val, int bytes);
@@ -526,7 +526,6 @@ int kvm_mmu_page_fault(struct kvm_vcpu *
void kvm_enable_tdp(void);
-int load_pdptrs(struct kvm_vcpu *vcpu, unsigned long cr3);
int complete_pio(struct kvm_vcpu *vcpu);
static inline struct kvm_mmu_page *page_header(hpa_t shadow_page)
Index: kvm.first/arch/x86/kvm/svm.c
===================================================================
--- kvm.first.orig/arch/x86/kvm/svm.c
+++ kvm.first/arch/x86/kvm/svm.c
@@ -1394,7 +1394,7 @@ static int handle_exit(struct kvm_run *k
vcpu->arch.cr0 = svm->vmcb->save.cr0;
vcpu->arch.cr3 = svm->vmcb->save.cr3;
if (is_paging(vcpu) && is_pae(vcpu) && !is_long_mode(vcpu)) {
- if (!load_pdptrs(vcpu, vcpu->arch.cr3)) {
+ if (!load_pdptrs(vcpu, vcpu->arch.cr3, 1)) {
kvm_inject_gp(vcpu, 0);
return 1;
}
--
-------------------------------------------------------------------------
This SF.net email is sponsored by: Microsoft
Defy all challenges. Microsoft(R) Visual Studio 2008.
http://clk.atdmt.com/MRT/go/vse0120000070mrt/direct/01/
_______________________________________________
kvm-devel mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/kvm-devel