CC Catalin

On 2017/9/6 2:58, gengdongjiu wrote:
> when exit from guest, some host PSTATE bits may be lost, such as
> PSTATE.PAN or PSTATE.UAO. It is because host and hypervisor all run
> in the EL2, host PSTATE value cannot be saved and restored via
> SPSR_EL2. So if guest has changed the PSTATE, host continues with
> a wrong value guest has set.
> 
> Signed-off-by: Dongjiu Geng <gengdong...@huawei.com>
> Signed-off-by: Haibin Zhang <zhanghaib...@huawei.com>
> ---
>  arch/arm64/include/asm/kvm_host.h |  8 +++++++
>  arch/arm64/include/asm/kvm_hyp.h  |  2 ++
>  arch/arm64/include/asm/sysreg.h   | 23 +++++++++++++++++++
>  arch/arm64/kvm/hyp/entry.S        |  2 --
>  arch/arm64/kvm/hyp/switch.c       | 24 ++++++++++++++++++--
>  arch/arm64/kvm/hyp/sysreg-sr.c    | 48 
> ++++++++++++++++++++++++++++++++++++---
>  6 files changed, 100 insertions(+), 7 deletions(-)
> 
> diff --git a/arch/arm64/include/asm/kvm_host.h 
> b/arch/arm64/include/asm/kvm_host.h
> index e923b58..cba7d3e 100644
> --- a/arch/arm64/include/asm/kvm_host.h
> +++ b/arch/arm64/include/asm/kvm_host.h
> @@ -193,6 +193,12 @@ struct kvm_cpu_context {
>       };
>  };
>  
> +struct kvm_cpu_host_pstate {
> +     u64 daif;
> +     u64 uao;
> +     u64 pan;
> +};
> +
>  typedef struct kvm_cpu_context kvm_cpu_context_t;
>  
>  struct kvm_vcpu_arch {
> @@ -227,6 +233,8 @@ struct kvm_vcpu_arch {
>  
>       /* Pointer to host CPU context */
>       kvm_cpu_context_t *host_cpu_context;
> +     /* Host PSTATE value */
> +     struct kvm_cpu_host_pstate host_pstate;
>       struct {
>               /* {Break,watch}point registers */
>               struct kvm_guest_debug_arch regs;
> diff --git a/arch/arm64/include/asm/kvm_hyp.h 
> b/arch/arm64/include/asm/kvm_hyp.h
> index 4572a9b..a75587a 100644
> --- a/arch/arm64/include/asm/kvm_hyp.h
> +++ b/arch/arm64/include/asm/kvm_hyp.h
> @@ -134,6 +134,8 @@
>  
>  void __sysreg_save_host_state(struct kvm_cpu_context *ctxt);
>  void __sysreg_restore_host_state(struct kvm_cpu_context *ctxt);
> +void __sysreg_save_host_pstate(struct kvm_vcpu *vcpu);
> +void __sysreg_restore_host_pstate(struct kvm_vcpu *vcpu);
>  void __sysreg_save_guest_state(struct kvm_cpu_context *ctxt);
>  void __sysreg_restore_guest_state(struct kvm_cpu_context *ctxt);
>  void __sysreg32_save_state(struct kvm_vcpu *vcpu);
> diff --git a/arch/arm64/include/asm/sysreg.h b/arch/arm64/include/asm/sysreg.h
> index 248339e..efdcf40 100644
> --- a/arch/arm64/include/asm/sysreg.h
> +++ b/arch/arm64/include/asm/sysreg.h
> @@ -295,6 +295,29 @@
>  #define SYS_ICH_LR14_EL2             __SYS__LR8_EL2(6)
>  #define SYS_ICH_LR15_EL2             __SYS__LR8_EL2(7)
>  
> +#define REG_PSTATE_PAN                       sys_reg(3, 0, 4, 2, 3)
> +#define REG_PSTATE_UAO                       sys_reg(3, 0, 4, 2, 4)
> +
> +#define GET_PSTATE_PAN                                       \
> +     ({                                              \
> +             u64 reg;                                \
> +             asm volatile(ALTERNATIVE("mov %0, #0",  \
> +                             "mrs_s %0, " __stringify(REG_PSTATE_PAN),\
> +                             ARM64_HAS_PAN)\
> +                             : "=r" (reg));\
> +             reg;                            \
> +     })
> +
> +#define GET_PSTATE_UAO                                       \
> +     ({                                              \
> +             u64 reg;                                        \
> +             asm volatile(ALTERNATIVE("mov %0, #0",\
> +                             "mrs_s %0, " __stringify(REG_PSTATE_UAO),\
> +                             ARM64_HAS_UAO)\
> +                             : "=r" (reg));\
> +             reg;                    \
> +     })
> +
>  /* Common SCTLR_ELx flags. */
>  #define SCTLR_ELx_EE    (1 << 25)
>  #define SCTLR_ELx_I  (1 << 12)
> diff --git a/arch/arm64/kvm/hyp/entry.S b/arch/arm64/kvm/hyp/entry.S
> index 12ee62d..7662ef5 100644
> --- a/arch/arm64/kvm/hyp/entry.S
> +++ b/arch/arm64/kvm/hyp/entry.S
> @@ -96,8 +96,6 @@ ENTRY(__guest_exit)
>  
>       add     x1, x1, #VCPU_CONTEXT
>  
> -     ALTERNATIVE(nop, SET_PSTATE_PAN(1), ARM64_HAS_PAN, CONFIG_ARM64_PAN)
> -
>       // Store the guest regs x2 and x3
>       stp     x2, x3,   [x1, #CPU_XREG_OFFSET(2)]
>  
> diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c
> index 945e79c..9b380a1 100644
> --- a/arch/arm64/kvm/hyp/switch.c
> +++ b/arch/arm64/kvm/hyp/switch.c
> @@ -278,6 +278,26 @@ static void __hyp_text __skip_instr(struct kvm_vcpu 
> *vcpu)
>       write_sysreg_el2(*vcpu_pc(vcpu), elr);
>  }
>  
> +static void __hyp_text __save_host_state(struct kvm_vcpu *vcpu)
> +{
> +     struct kvm_cpu_context *host_ctxt;
> +
> +     host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
> +
> +     __sysreg_save_host_state(host_ctxt);
> +     __sysreg_save_host_pstate(vcpu);
> +}
> +
> +static void __hyp_text __restore_host_state(struct kvm_vcpu *vcpu)
> +{
> +     struct kvm_cpu_context *host_ctxt;
> +
> +     host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
> +
> +     __sysreg_restore_host_state(host_ctxt);
> +     __sysreg_restore_host_pstate(vcpu);
> +}
> +
>  int __hyp_text __kvm_vcpu_run(struct kvm_vcpu *vcpu)
>  {
>       struct kvm_cpu_context *host_ctxt;
> @@ -291,7 +311,7 @@ int __hyp_text __kvm_vcpu_run(struct kvm_vcpu *vcpu)
>       host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
>       guest_ctxt = &vcpu->arch.ctxt;
>  
> -     __sysreg_save_host_state(host_ctxt);
> +     __save_host_state(vcpu);
>       __debug_cond_save_host_state(vcpu);
>  
>       __activate_traps(vcpu);
> @@ -374,7 +394,7 @@ int __hyp_text __kvm_vcpu_run(struct kvm_vcpu *vcpu)
>       __deactivate_traps(vcpu);
>       __deactivate_vm(vcpu);
>  
> -     __sysreg_restore_host_state(host_ctxt);
> +     __restore_host_state(vcpu);
>  
>       if (fp_enabled) {
>               __fpsimd_save_state(&guest_ctxt->gp_regs.fp_regs);
> diff --git a/arch/arm64/kvm/hyp/sysreg-sr.c b/arch/arm64/kvm/hyp/sysreg-sr.c
> index 9341376..ea8f437 100644
> --- a/arch/arm64/kvm/hyp/sysreg-sr.c
> +++ b/arch/arm64/kvm/hyp/sysreg-sr.c
> @@ -22,7 +22,11 @@
>  #include <asm/kvm_hyp.h>
>  
>  /* Yes, this does nothing, on purpose */
> -static void __hyp_text __sysreg_do_nothing(struct kvm_cpu_context *ctxt) { }
> +static void __hyp_text __sysreg_do_nothing_state(struct kvm_cpu_context 
> *ctxt)
> +{ }
> +static void __hyp_text __sysreg_do_nothing_pstate(struct kvm_vcpu *vcpu)
> +{ }
> +
>  
>  /*
>   * Non-VHE: Both host and guest must save everything.
> @@ -69,7 +73,7 @@ static void __hyp_text __sysreg_save_state(struct 
> kvm_cpu_context *ctxt)
>  }
>  
>  static hyp_alternate_select(__sysreg_call_save_host_state,
> -                         __sysreg_save_state, __sysreg_do_nothing,
> +                         __sysreg_save_state, __sysreg_do_nothing_state,
>                           ARM64_HAS_VIRT_HOST_EXTN);
>  
>  void __hyp_text __sysreg_save_host_state(struct kvm_cpu_context *ctxt)
> @@ -122,7 +126,7 @@ static void __hyp_text __sysreg_restore_state(struct 
> kvm_cpu_context *ctxt)
>  }
>  
>  static hyp_alternate_select(__sysreg_call_restore_host_state,
> -                         __sysreg_restore_state, __sysreg_do_nothing,
> +                         __sysreg_restore_state, __sysreg_do_nothing_state,
>                           ARM64_HAS_VIRT_HOST_EXTN);
>  
>  void __hyp_text __sysreg_restore_host_state(struct kvm_cpu_context *ctxt)
> @@ -137,6 +141,44 @@ void __hyp_text __sysreg_restore_guest_state(struct 
> kvm_cpu_context *ctxt)
>       __sysreg_restore_common_state(ctxt);
>  }
>  
> +static void __hyp_text __sysreg_save_pstate(struct kvm_vcpu *vcpu)
> +{
> +     vcpu->arch.host_pstate.daif = read_sysreg(daif);
> +     vcpu->arch.host_pstate.pan = GET_PSTATE_PAN;
> +     vcpu->arch.host_pstate.uao = GET_PSTATE_UAO;
> +}
> +
> +static hyp_alternate_select(__sysreg_call_save_host_pstate,
> +                         __sysreg_save_pstate, __sysreg_do_nothing_pstate,
> +                         ARM64_HAS_VIRT_HOST_EXTN);
> +
> +void __hyp_text __sysreg_save_host_pstate(struct kvm_vcpu *vcpu)
> +{
> +     __sysreg_call_save_host_pstate()(vcpu);
> +}
> +
> +static void __hyp_text __sysreg_restore_pstate(struct kvm_vcpu *vcpu)
> +{
> +     u8 value = !!(vcpu->arch.host_pstate.pan);
> +
> +     write_sysreg(vcpu->arch.host_pstate.daif, daif);
> +     asm(ALTERNATIVE("nop", SET_PSTATE_PAN(value), ARM64_HAS_PAN,
> +                     CONFIG_ARM64_PAN));
> +
> +     value = !!(vcpu->arch.host_pstate.uao);
> +     asm(ALTERNATIVE("nop", SET_PSTATE_UAO(value), ARM64_HAS_UAO,
> +                     CONFIG_ARM64_UAO));
> +}
> +
> +static hyp_alternate_select(__sysreg_call_restore_host_pstate,
> +                         __sysreg_restore_pstate, __sysreg_do_nothing_pstate,
> +                         ARM64_HAS_VIRT_HOST_EXTN);
> +
> +void __hyp_text __sysreg_restore_host_pstate(struct kvm_vcpu *vcpu)
> +{
> +     __sysreg_call_restore_host_pstate()(vcpu);
> +}
> +
>  void __hyp_text __sysreg32_save_state(struct kvm_vcpu *vcpu)
>  {
>       u64 *spsr, *sysreg;
> 

Reply via email to