On 15.12.2016 06:53, David Gibson wrote:
> Currently, kvmppc_alloc_hpt() both allocates a new hashed page table (HPT)
> and sets it up as the active page table for a VM.  For the upcoming HPT
> resize implementation we're going to want to allocate HPTs separately from
> activating them.
> 
> So, split the allocation itself out into kvmppc_allocate_hpt() and perform
> the activation with a new kvmppc_set_hpt() function.  Likewise we split
> kvmppc_free_hpt(), which just frees the HPT, from kvmppc_release_hpt()
> which unsets it as an active HPT, then frees it.
> 
> We also move the logic to fall back to smaller HPT sizes if the first try
> fails into the single caller which used that behaviour,
> kvmppc_hv_setup_htab_rma().  This introduces a slight semantic change, in
> that previously if the initial attempt at CMA allocation failed, we would
> fall back to attempting smaller sizes with the page allocator.  Now, we
> try first CMA, then the page allocator at each size.  As far as I can tell
> this change should be harmless.
> 
> To match, we make kvmppc_free_hpt() just free the actual HPT itself.  The
> call to kvmppc_free_lpid() that was there, we move to the single caller.
> 
> Signed-off-by: David Gibson <da...@gibson.dropbear.id.au>
> ---
>  arch/powerpc/include/asm/kvm_book3s_64.h |  3 ++
>  arch/powerpc/include/asm/kvm_ppc.h       |  5 +-
>  arch/powerpc/kvm/book3s_64_mmu_hv.c      | 90 
> ++++++++++++++++----------------
>  arch/powerpc/kvm/book3s_hv.c             | 18 +++++--
>  4 files changed, 65 insertions(+), 51 deletions(-)
> 
> diff --git a/arch/powerpc/include/asm/kvm_book3s_64.h 
> b/arch/powerpc/include/asm/kvm_book3s_64.h
> index 8810327..6dc4004 100644
> --- a/arch/powerpc/include/asm/kvm_book3s_64.h
> +++ b/arch/powerpc/include/asm/kvm_book3s_64.h
> @@ -22,6 +22,9 @@
>  
>  #include <asm/book3s/64/mmu-hash.h>
>  
> +/* Power architecture requires HPT is at least 256kB */
> +#define PPC_MIN_HPT_ORDER    18
> +
>  #ifdef CONFIG_KVM_BOOK3S_PR_POSSIBLE
>  static inline struct kvmppc_book3s_shadow_vcpu *svcpu_get(struct kvm_vcpu 
> *vcpu)
>  {
> diff --git a/arch/powerpc/include/asm/kvm_ppc.h 
> b/arch/powerpc/include/asm/kvm_ppc.h
> index 3db6be9..41575b8 100644
> --- a/arch/powerpc/include/asm/kvm_ppc.h
> +++ b/arch/powerpc/include/asm/kvm_ppc.h
> @@ -155,9 +155,10 @@ extern void kvmppc_core_destroy_mmu(struct kvm_vcpu 
> *vcpu);
>  extern int kvmppc_kvm_pv(struct kvm_vcpu *vcpu);
>  extern void kvmppc_map_magic(struct kvm_vcpu *vcpu);
>  
> -extern long kvmppc_alloc_hpt(struct kvm *kvm, u32 *htab_orderp);
> +extern int kvmppc_allocate_hpt(struct kvm_hpt_info *info, u32 order);
> +extern void kvmppc_set_hpt(struct kvm *kvm, struct kvm_hpt_info *info);
>  extern long kvmppc_alloc_reset_hpt(struct kvm *kvm, u32 *htab_orderp);
> -extern void kvmppc_free_hpt(struct kvm *kvm);
> +extern void kvmppc_free_hpt(struct kvm_hpt_info *info);
>  extern long kvmppc_prepare_vrma(struct kvm *kvm,
>                               struct kvm_userspace_memory_region *mem);
>  extern void kvmppc_map_vrma(struct kvm_vcpu *vcpu,
> diff --git a/arch/powerpc/kvm/book3s_64_mmu_hv.c 
> b/arch/powerpc/kvm/book3s_64_mmu_hv.c
> index fe88132..68bb228 100644
> --- a/arch/powerpc/kvm/book3s_64_mmu_hv.c
> +++ b/arch/powerpc/kvm/book3s_64_mmu_hv.c
> @@ -40,76 +40,70 @@
>  
>  #include "trace_hv.h"
>  
> -/* Power architecture requires HPT is at least 256kB */
> -#define PPC_MIN_HPT_ORDER    18
> -
>  static long kvmppc_virtmode_do_h_enter(struct kvm *kvm, unsigned long flags,
>                               long pte_index, unsigned long pteh,
>                               unsigned long ptel, unsigned long *pte_idx_ret);
>  static void kvmppc_rmap_reset(struct kvm *kvm);
>  
> -long kvmppc_alloc_hpt(struct kvm *kvm, u32 *htab_orderp)
> +int kvmppc_allocate_hpt(struct kvm_hpt_info *info, u32 order)
>  {
>       unsigned long hpt = 0;
> -     struct revmap_entry *rev;
> +     int cma = 0;
>       struct page *page = NULL;
> -     long order = KVM_DEFAULT_HPT_ORDER;
> -
> -     if (htab_orderp) {
> -             order = *htab_orderp;
> -             if (order < PPC_MIN_HPT_ORDER)
> -                     order = PPC_MIN_HPT_ORDER;
> -     }

Not sure, but as far as I can see, *htab_orderp could still be provided
from userspace via kvm_arch_vm_ioctl_hv() -> kvmppc_alloc_reset_hpt() ?
In that case, I think you should still check that the order is bigger
than PPC_MIN_HPT_ORDER, and return an error code otherwise?
Or if you feel confident that this value can never be supplied by
userspace anymore, add at least a BUG_ON(order < PPC_MIN_HPT_ORDER) here?

> +     struct revmap_entry *rev;
> +     unsigned long npte;
>  
> -     kvm->arch.hpt.cma = 0;
> +     hpt = 0;
> +     cma = 0;

Both hpt and cma are initialized to zero in the variable declarations
already, so the above two lines are redundant.

>       page = kvm_alloc_hpt_cma(1ul << (order - PAGE_SHIFT));
>       if (page) {
>               hpt = (unsigned long)pfn_to_kaddr(page_to_pfn(page));
>               memset((void *)hpt, 0, (1ul << order));
> -             kvm->arch.hpt.cma = 1;
> +             cma = 1;
>       }
>  
> -     /* Lastly try successively smaller sizes from the page allocator */
> -     /* Only do this if userspace didn't specify a size via ioctl */
> -     while (!hpt && order > PPC_MIN_HPT_ORDER && !htab_orderp) {
> -             hpt = __get_free_pages(GFP_KERNEL|__GFP_ZERO|__GFP_REPEAT|
> -                                    __GFP_NOWARN, order - PAGE_SHIFT);
> -             if (!hpt)
> -                     --order;
> -     }
> +     if (!hpt)
> +             hpt = __get_free_pages(GFP_KERNEL|__GFP_ZERO|__GFP_REPEAT
> +                                    |__GFP_NOWARN, order - PAGE_SHIFT);
>  
>       if (!hpt)
>               return -ENOMEM;
>  
> -     kvm->arch.hpt.virt = hpt;
> -     kvm->arch.hpt.order = order;
> -
> -     atomic64_set(&kvm->arch.mmio_update, 0);
> +     /* HPTEs are 2**4 bytes long */
> +     npte = 1ul << (order - 4);
>  
>       /* Allocate reverse map array */
> -     rev = vmalloc(sizeof(struct revmap_entry) * 
> kvmppc_hpt_npte(&kvm->arch.hpt));
> +     rev = vmalloc(sizeof(struct revmap_entry) * npte);
>       if (!rev) {
> -             pr_err("kvmppc_alloc_hpt: Couldn't alloc reverse map array\n");
> +             pr_err("kvmppc_allocate_hpt: Couldn't alloc reverse map 
> array\n");
>               goto out_freehpt;

So here you jump to out_freehpt before setting info->virt ...

>       }
> -     kvm->arch.hpt.rev = rev;
> -     kvm->arch.sdr1 = __pa(hpt) | (order - 18);
>  
> -     pr_info("KVM guest htab at %lx (order %ld), LPID %x\n",
> -             hpt, order, kvm->arch.lpid);
> +     info->order = order;
> +     info->virt = hpt;
> +     info->cma = cma;
> +     info->rev = rev;
>  
> -     if (htab_orderp)
> -             *htab_orderp = order;
>       return 0;
>  
>   out_freehpt:
> -     if (kvm->arch.hpt.cma)
> +     if (cma)
>               kvm_free_hpt_cma(page, 1 << (order - PAGE_SHIFT));
>       else
> -             free_pages(hpt, order - PAGE_SHIFT);
> +             free_pages(info->virt, order - PAGE_SHIFT);

... but here you free info->virt which has not been set in case of the
"goto" above. That's clearly wrong.

>       return -ENOMEM;
>  }
>  
> +void kvmppc_set_hpt(struct kvm *kvm, struct kvm_hpt_info *info)
> +{
> +     atomic64_set(&kvm->arch.mmio_update, 0);
> +     kvm->arch.hpt = *info;
> +     kvm->arch.sdr1 = __pa(info->virt) | (info->order - 18);
> +
> +     pr_info("KVM guest htab at %lx (order %ld), LPID %x\n",
> +             info->virt, (long)info->order, kvm->arch.lpid);

Not directly related to your patch, but these messages pop up in the
dmesg each time I start a guest ... for the normal user who does not
have a clue what "htab" means, this can be pretty much annoying. Could
you maybe turn this into a pr_debug() instead?

> +}
> +
>  long kvmppc_alloc_reset_hpt(struct kvm *kvm, u32 *htab_orderp)
>  {
>       long err = -EBUSY;
> @@ -138,24 +132,28 @@ long kvmppc_alloc_reset_hpt(struct kvm *kvm, u32 
> *htab_orderp)
>               *htab_orderp = order;
>               err = 0;
>       } else {
> -             err = kvmppc_alloc_hpt(kvm, htab_orderp);
> -             order = *htab_orderp;
> +             struct kvm_hpt_info info;
> +
> +             err = kvmppc_allocate_hpt(&info, *htab_orderp);
> +             if (err < 0)
> +                     goto out;
> +             kvmppc_set_hpt(kvm, &info);

Just a matter of taste (I dislike gotos if avoidable), but you could
also do:

        err = kvmppc_allocate_hpt(&info, *htab_orderp);
        if (!err)
                kvmppc_set_hpt(kvm, &info);

... and that's even one line shorter ;-)

>       }
>   out:
>       mutex_unlock(&kvm->lock);
>       return err;
>  }
>  
> -void kvmppc_free_hpt(struct kvm *kvm)
> +void kvmppc_free_hpt(struct kvm_hpt_info *info)
>  {
> -     kvmppc_free_lpid(kvm->arch.lpid);
> -     vfree(kvm->arch.hpt.rev);
> -     if (kvm->arch.hpt.cma)
> -             kvm_free_hpt_cma(virt_to_page(kvm->arch.hpt.virt),
> -                              1 << (kvm->arch.hpt.order - PAGE_SHIFT));
> +     vfree(info->rev);
> +     if (info->cma)
> +             kvm_free_hpt_cma(virt_to_page(info->virt),
> +                              1 << (info->order - PAGE_SHIFT));
>       else
> -             free_pages(kvm->arch.hpt.virt,
> -                        kvm->arch.hpt.order - PAGE_SHIFT);
> +             free_pages(info->virt, info->order - PAGE_SHIFT);
> +     info->virt = 0;
> +     info->order = 0;
>  }
>  
>  /* Bits in first HPTE dword for pagesize 4k, 64k or 16M */
> diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c
> index 78baa2b..71c5adb 100644
> --- a/arch/powerpc/kvm/book3s_hv.c
> +++ b/arch/powerpc/kvm/book3s_hv.c
> @@ -3115,11 +3115,22 @@ static int kvmppc_hv_setup_htab_rma(struct kvm_vcpu 
> *vcpu)
>  
>       /* Allocate hashed page table (if not done already) and reset it */
>       if (!kvm->arch.hpt.virt) {
> -             err = kvmppc_alloc_hpt(kvm, NULL);
> -             if (err) {
> +             int order = KVM_DEFAULT_HPT_ORDER;
> +             struct kvm_hpt_info info;
> +
> +             err = kvmppc_allocate_hpt(&info, order);
> +             /* If we get here, it means userspace didn't specify a
> +              * size explicitly.  So, try successively smaller
> +              * sizes if the default failed. */
> +             while (err < 0 && --order >= PPC_MIN_HPT_ORDER)
> +                     err  = kvmppc_allocate_hpt(&info, order);

Not sure, but maybe replace "err < 0" with "err == -ENOMEM" in the while
condition? Or should it also loop on other future possible errors types?

> +             if (err < 0) {
>                       pr_err("KVM: Couldn't alloc HPT\n");
>                       goto out;
>               }
> +
> +             kvmppc_set_hpt(kvm, &info);
>       }
>  
>       /* Look up the memslot for guest physical address 0 */
> @@ -3357,7 +3368,8 @@ static void kvmppc_core_destroy_vm_hv(struct kvm *kvm)
>  
>       kvmppc_free_vcores(kvm);
>  
> -     kvmppc_free_hpt(kvm);
> +     kvmppc_free_lpid(kvm->arch.lpid);
> +     kvmppc_free_hpt(&kvm->arch.hpt);
>  
>       kvmppc_free_pimap(kvm);
>  }
> 

 Thomas

Reply via email to