Izik Eidus wrote:
> this patch make the guest non shadowed pages swappedable
>
> ------------------------------------------------------------------------
>
> -------------------------------------------------------------------------
> This SF.net email is sponsored by: Splunk Inc.
> Still grepping through log files to find problems?  Stop.
> Now Search log events and configuration files using AJAX and a browser.
> Download your FREE copy of Splunk now >> http://get.splunk.com/
> ------------------------------------------------------------------------
>
> _______________________________________________
> kvm-devel mailing list
> kvm-devel@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/kvm-devel
> From 8e25e215b8ed95ca4ff51cbfcf5bdc438bb799f4 Mon Sep 17 00:00:00 2001
> From: Izik Eidus <[EMAIL PROTECTED](none)>
> Date: Sat, 13 Oct 2007 04:03:28 +0200
> Subject: [PATCH] make the guest non shadowed memory swappable
>
> Signed-off-by: Izik Eidus <[EMAIL PROTECTED]>
> ---
>  drivers/kvm/kvm.h         |    1 +
>  drivers/kvm/kvm_main.c    |   66 +++++++++++++++++++++++++-------------------
>  drivers/kvm/mmu.c         |   13 ++++++++-
>  drivers/kvm/paging_tmpl.h |   23 +++++++++++++--
>  4 files changed, 70 insertions(+), 33 deletions(-)
>
> diff --git a/drivers/kvm/kvm.h b/drivers/kvm/kvm.h
> index a155c2b..2e83fa7 100644
> --- a/drivers/kvm/kvm.h
> +++ b/drivers/kvm/kvm.h
> @@ -409,6 +409,7 @@ struct kvm_memory_slot {
>       unsigned long *rmap;
>       unsigned long *dirty_bitmap;
>       int user_alloc; /* user allocated memory */
> +     unsigned long userspace_addr;
>  };
>  
>  struct kvm {
> diff --git a/drivers/kvm/kvm_main.c b/drivers/kvm/kvm_main.c
> index bfa201c..0dce93c 100644
> --- a/drivers/kvm/kvm_main.c
> +++ b/drivers/kvm/kvm_main.c
> @@ -321,15 +321,6 @@ static struct kvm *kvm_create_vm(void)
>  
>  static void kvm_free_userspace_physmem(struct kvm_memory_slot *free)
>  {
> -     int i;
> -
> -     for (i = 0; i < free->npages; ++i) {
> -             if (free->phys_mem[i]) {
> -                     if (!PageReserved(free->phys_mem[i]))
> -                             SetPageDirty(free->phys_mem[i]);
> -                     page_cache_release(free->phys_mem[i]);
> -             }
> -     }
>  }
>
>   

If you're making this a nop, then you might as well just delete the 
function and it's callers.

>  static void kvm_free_kernel_physmem(struct kvm_memory_slot *free)
> @@ -771,19 +762,8 @@ static int kvm_vm_ioctl_set_memory_region(struct kvm 
> *kvm,
>               memset(new.phys_mem, 0, npages * sizeof(struct page *));
>               memset(new.rmap, 0, npages * sizeof(*new.rmap));
>               if (user_alloc) {
> -                     unsigned long pages_num;
> -
>                       new.user_alloc = 1;
> -                     down_read(&current->mm->mmap_sem);
> -
> -                     pages_num = get_user_pages(current, current->mm,
> -                                                mem->userspace_addr,
> -                                                npages, 1, 0, new.phys_mem,
> -                                                NULL);
> -
> -                     up_read(&current->mm->mmap_sem);
> -                     if (pages_num != npages)
> -                             goto out_unlock;
> +                     new.userspace_addr = mem->userspace_addr;
>               } else {
>   

You can eliminate the braces in this condition.

>                       for (i = 0; i < npages; ++i) {
>                               new.phys_mem[i] = alloc_page(GFP_HIGHUSER
> @@ -1058,8 +1038,27 @@ struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn)
>  
>       gfn = unalias_gfn(kvm, gfn);
>       slot = __gfn_to_memslot(kvm, gfn);
> -     if (!slot)
> +     if (!slot) {
> +             get_page(bad_page);
>               return bad_page;
> +     }
> +     if (slot->user_alloc) {
> +             struct page *page[1];
> +             int npages;
> +
> +             down_read(&current->mm->mmap_sem);
> +             npages = get_user_pages(current, current->mm,
> +                                     slot->userspace_addr
> +                                     + (gfn - slot->base_gfn) * PAGE_SIZE, 1,
> +                                     1, 0, page, NULL);
> +             up_read(&current->mm->mmap_sem);
> +             if (npages != 1) {
> +                     get_page(bad_page);
> +                     return bad_page;
> +             }
> +             return page[0];
>   

Wouldn't it be necessary to assign page[0] to slot->phys_mem[gfn - 
slot->base_gfn]?

Regards,

Anthony Liguori

> +     }
> +     get_page(slot->phys_mem[gfn - slot->base_gfn]);
>       return slot->phys_mem[gfn - slot->base_gfn];
>  }
>  EXPORT_SYMBOL_GPL(gfn_to_page);
> @@ -1079,13 +1078,16 @@ int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, 
> void *data, int offset,
>       struct page *page;
>  
>       page = gfn_to_page(kvm, gfn);
> -     if (is_error_page(page))
> +     if (is_error_page(page)) {
> +             put_page(page);
>               return -EFAULT;
> +     }
>       page_virt = kmap_atomic(page, KM_USER0);
>  
>       memcpy(data, page_virt + offset, len);
>  
>       kunmap_atomic(page_virt, KM_USER0);
> +     put_page(page);
>       return 0;
>  }
>  EXPORT_SYMBOL_GPL(kvm_read_guest_page);
> @@ -1117,14 +1119,17 @@ int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, 
> const void *data,
>       struct page *page;
>  
>       page = gfn_to_page(kvm, gfn);
> -     if (is_error_page(page))
> +     if (is_error_page(page)) {
> +             put_page(page);
>               return -EFAULT;
> +     }
>       page_virt = kmap_atomic(page, KM_USER0);
>  
>       memcpy(page_virt + offset, data, len);
>  
>       kunmap_atomic(page_virt, KM_USER0);
>       mark_page_dirty(kvm, gfn);
> +     put_page(page);
>       return 0;
>  }
>  EXPORT_SYMBOL_GPL(kvm_write_guest_page);
> @@ -1155,13 +1160,16 @@ int kvm_clear_guest_page(struct kvm *kvm, gfn_t gfn, 
> int offset, int len)
>       struct page *page;
>  
>       page = gfn_to_page(kvm, gfn);
> -     if (is_error_page(page))
> +     if (is_error_page(page)) {
> +             put_page(page);
>               return -EFAULT;
> +     }
>       page_virt = kmap_atomic(page, KM_USER0);
>  
>       memset(page_virt + offset, 0, len);
>  
>       kunmap_atomic(page_virt, KM_USER0);
> +     put_page(page);
>       return 0;
>  }
>  EXPORT_SYMBOL_GPL(kvm_clear_guest_page);
> @@ -2090,13 +2098,12 @@ int kvm_emulate_pio_string(struct kvm_vcpu *vcpu, 
> struct kvm_run *run, int in,
>       for (i = 0; i < nr_pages; ++i) {
>               mutex_lock(&vcpu->kvm->lock);
>               page = gva_to_page(vcpu, address + i * PAGE_SIZE);
> -             if (page)
> -                     get_page(page);
>               vcpu->pio.guest_pages[i] = page;
>               mutex_unlock(&vcpu->kvm->lock);
>               if (!page) {
>                       inject_gp(vcpu);
>                       free_pio_guest_pages(vcpu);
>                       return 1;
>               }
>       }
> @@ -3081,9 +3088,10 @@ static struct page *kvm_vm_nopage(struct 
> vm_area_struct *vma,
>  
>       pgoff = ((address - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff;
>       page = gfn_to_page(kvm, pgoff);
> -     if (is_error_page(page))
> +     if (is_error_page(page)) {
> +             put_page(page);
>               return NOPAGE_SIGBUS;
> -     get_page(page);
> +     }
>       if (type != NULL)
>               *type = VM_FAULT_MINOR;
>  
> diff --git a/drivers/kvm/mmu.c b/drivers/kvm/mmu.c
> index e6a9b4a..19af09a 100644
> --- a/drivers/kvm/mmu.c
> +++ b/drivers/kvm/mmu.c
> @@ -425,6 +425,7 @@ static void rmap_remove(struct kvm *kvm, u64 *spte)
>       if (!is_rmap_pte(*spte))
>               return;
>       page = page_header(__pa(spte));
> +     put_page(pfn_to_page((*spte & PT64_BASE_ADDR_MASK) >> PAGE_SHIFT));
>       rmapp = gfn_to_rmap(kvm, page->gfns[spte - page->spt]);
>       if (!*rmapp) {
>               printk(KERN_ERR "rmap_remove: %p %llx 0->BUG\n", spte, *spte);
> @@ -907,6 +908,9 @@ static int nonpaging_map(struct kvm_vcpu *vcpu, gva_t v, 
> hpa_t p)
>                       table[index] = p | PT_PRESENT_MASK | PT_WRITABLE_MASK |
>                                                               PT_USER_MASK;
>                       rmap_add(vcpu, &table[index], v >> PAGE_SHIFT);
> +                     if (!is_rmap_pte(table[index]))
> +                             put_page(pfn_to_page((v & PT64_BASE_ADDR_MASK)
> +                                      >> PAGE_SHIFT));
>                       return 0;
>               }
>  
> @@ -921,6 +925,8 @@ static int nonpaging_map(struct kvm_vcpu *vcpu, gva_t v, 
> hpa_t p)
>                                                    1, 3, &table[index]);
>                       if (!new_table) {
>                               pgprintk("nonpaging_map: ENOMEM\n");
> +                             put_page(pfn_to_page((v & PT64_BASE_ADDR_MASK)
> +                                                  >> PAGE_SHIFT));
>                               return -ENOMEM;
>                       }
>  
> @@ -1035,8 +1041,11 @@ static int nonpaging_page_fault(struct kvm_vcpu *vcpu, 
> gva_t gva,
>  
>       paddr = gpa_to_hpa(vcpu->kvm, addr & PT64_BASE_ADDR_MASK);
>  
> -     if (is_error_hpa(paddr))
> +     if (is_error_hpa(paddr)) {
> +             put_page(pfn_to_page((paddr & PT64_BASE_ADDR_MASK)
> +                                  >> PAGE_SHIFT));
>               return 1;
> +     }
>  
>       return nonpaging_map(vcpu, addr & PAGE_MASK, paddr);
>  }
> @@ -1515,6 +1524,8 @@ static void audit_mappings_page(struct kvm_vcpu *vcpu, 
> u64 page_pte,
>                                && !is_error_hpa(hpa))
>                               printk(KERN_ERR "audit: (%s) notrap shadow,"
>                                      " valid guest gva %lx\n", audit_msg, va);
> +                     put_page(pfn_to_page((paddr & PT64_BASE_ADDR_MASK)
> +                                          >> PAGE_SHIFT));
>  
>               }
>       }
> diff --git a/drivers/kvm/paging_tmpl.h b/drivers/kvm/paging_tmpl.h
> index 58fd35a..931a308 100644
> --- a/drivers/kvm/paging_tmpl.h
> +++ b/drivers/kvm/paging_tmpl.h
> @@ -175,6 +175,7 @@ static int FNAME(walk_addr)(struct guest_walker *walker,
>               walker->inherited_ar &= walker->table[index];
>               table_gfn = (*ptep & PT_BASE_ADDR_MASK) >> PAGE_SHIFT;
>               kunmap_atomic(walker->table, KM_USER0);
> +             put_page(walker->page);
>               walker->page = gfn_to_page(vcpu->kvm, table_gfn);
>               walker->table = kmap_atomic(walker->page, KM_USER0);
>               --walker->level;
> @@ -183,8 +184,10 @@ static int FNAME(walk_addr)(struct guest_walker *walker,
>                        walker->level - 1, table_gfn);
>       }
>       walker->pte = *ptep;
> -     if (walker->page)
> +     if (walker->page) {
>               walker->ptep = NULL;
> +             put_page(walker->page);
> +     }
>       if (walker->table)
>               kunmap_atomic(walker->table, KM_USER0);
>       pgprintk("%s: pte %llx\n", __FUNCTION__, (u64)*ptep);
> @@ -206,6 +209,8 @@ err:
>               walker->error_code |= PFERR_FETCH_MASK;
>       if (walker->table)
>               kunmap_atomic(walker->table, KM_USER0);
> +     if (walker->page)
> +             put_page(walker->page);
>       return 0;
>  }
>  
> @@ -249,6 +254,8 @@ static void FNAME(set_pte_common)(struct kvm_vcpu *vcpu,
>       if (is_error_hpa(paddr)) {
>               set_shadow_pte(shadow_pte,
>                              shadow_trap_nonpresent_pte | PT_SHADOW_IO_MARK);
> +             put_page(pfn_to_page((paddr & PT64_BASE_ADDR_MASK)
> +                                  >> PAGE_SHIFT));
>               return;
>       }
>  
> @@ -286,9 +293,16 @@ unshadowed:
>       pgprintk("%s: setting spte %llx\n", __FUNCTION__, spte);
>       set_shadow_pte(shadow_pte, spte);
>       page_header_update_slot(vcpu->kvm, shadow_pte, gaddr);
> -     if (!was_rmapped)
> +     if (!was_rmapped) {
>               rmap_add(vcpu, shadow_pte, (gaddr & PT64_BASE_ADDR_MASK)
>                        >> PAGE_SHIFT);
> +             if (!is_rmap_pte(*shadow_pte))
> +                     put_page(pfn_to_page((paddr & PT64_BASE_ADDR_MASK)
> +                                          >> PAGE_SHIFT));
> +     }
> +     else
> +             put_page(pfn_to_page((paddr & PT64_BASE_ADDR_MASK)
> +                                  >> PAGE_SHIFT));
>       if (!ptwrite || !*ptwrite)
>               vcpu->last_pte_updated = shadow_pte;
>  }
> @@ -512,19 +526,22 @@ static void FNAME(prefetch_page)(struct kvm_vcpu *vcpu,
>  {
>       int i;
>       pt_element_t *gpt;
> +     struct page *page;
>  
>       if (sp->role.metaphysical || PTTYPE == 32) {
>               nonpaging_prefetch_page(vcpu, sp);
>               return;
>       }
>  
> -     gpt = kmap_atomic(gfn_to_page(vcpu->kvm, sp->gfn), KM_USER0);
> +     page = gfn_to_page(vcpu->kvm, sp->gfn);
> +     gpt = kmap_atomic(page, KM_USER0);
>       for (i = 0; i < PT64_ENT_PER_PAGE; ++i)
>               if (is_present_pte(gpt[i]))
>                       sp->spt[i] = shadow_trap_nonpresent_pte;
>               else
>                       sp->spt[i] = shadow_notrap_nonpresent_pte;
>       kunmap_atomic(gpt, KM_USER0);
> +     put_page(page);
>  }
>  
>  #undef pt_element_t
> -- 1.5.2.4


-------------------------------------------------------------------------
This SF.net email is sponsored by: Splunk Inc.
Still grepping through log files to find problems?  Stop.
Now Search log events and configuration files using AJAX and a browser.
Download your FREE copy of Splunk now >> http://get.splunk.com/
_______________________________________________
kvm-devel mailing list
kvm-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/kvm-devel

Reply via email to