On Wed, 2023-11-08 at 11:17 +0000, Nicolas Saenz Julienne wrote:
> Let user-space handle all hypercalls that fall under the AccessVsm
> partition privilege flag. That is:
>  - HVCALL_MODIFY_VTL_PROTECTION_MASK:
>  - HVCALL_ENABLE_PARTITION_VTL:
>  - HVCALL_ENABLE_VP_VTL:
>  - HVCALL_VTL_CALL:
>  - HVCALL_VTL_RETURN:
> The hypercalls are processed through the KVM_EXIT_HYPERV_HVCALL exit.
> Additionally, expose the cpuid bit.
> 
> Signed-off-by: Nicolas Saenz Julienne <nsa...@amazon.com>
> ---
>  arch/x86/kvm/hyperv.c             | 15 +++++++++++++++
>  include/asm-generic/hyperv-tlfs.h |  7 ++++++-
>  2 files changed, 21 insertions(+), 1 deletion(-)
> 
> diff --git a/arch/x86/kvm/hyperv.c b/arch/x86/kvm/hyperv.c
> index a3970d52eef1..a266c5d393f5 100644
> --- a/arch/x86/kvm/hyperv.c
> +++ b/arch/x86/kvm/hyperv.c
> @@ -2462,6 +2462,11 @@ static bool kvm_hv_is_xmm_output_hcall(u16 code)
>       return false;
>  }
>  
> +static inline bool kvm_hv_is_vtl_call_return(u16 code)
> +{
> +     return code == HVCALL_VTL_CALL || code == HVCALL_VTL_RETURN;
> +}
> +
>  static int kvm_hv_hypercall_complete_userspace(struct kvm_vcpu *vcpu)
>  {
>       bool fast = !!(vcpu->run->hyperv.u.hcall.input & HV_HYPERCALL_FAST_BIT);
> @@ -2471,6 +2476,9 @@ static int kvm_hv_hypercall_complete_userspace(struct 
> kvm_vcpu *vcpu)
>       if (kvm_hv_is_xmm_output_hcall(code) && hv_result_success(result) && 
> fast)
>               kvm_hv_write_xmm(vcpu->run->hyperv.u.hcall.xmm);
>  
> +     if (kvm_hv_is_vtl_call_return(code))
> +             return kvm_skip_emulated_instruction(vcpu);

Can you add justification for this?
If this is justified, does it make sense to move this code to 
kvm_hv_hypercall_complete
(which also calls kvm_skip_emulated_instruction())



> +
>       return kvm_hv_hypercall_complete(vcpu, result);
>  }
>  
> @@ -2525,6 +2533,7 @@ static bool is_xmm_fast_hypercall(struct kvm_hv_hcall 
> *hc)
>       case HVCALL_SEND_IPI_EX:
>       case HVCALL_GET_VP_REGISTERS:
>       case HVCALL_SET_VP_REGISTERS:
> +     case HVCALL_MODIFY_VTL_PROTECTION_MASK:
>               return true;
>       }
>  
> @@ -2745,6 +2754,11 @@ int kvm_hv_hypercall(struct kvm_vcpu *vcpu)
>               goto hypercall_userspace_exit;
>       case HVCALL_GET_VP_REGISTERS:
>       case HVCALL_SET_VP_REGISTERS:
> +     case HVCALL_MODIFY_VTL_PROTECTION_MASK:
> +     case HVCALL_ENABLE_PARTITION_VTL:
> +     case HVCALL_ENABLE_VP_VTL:
> +     case HVCALL_VTL_CALL:
> +     case HVCALL_VTL_RETURN:
>               goto hypercall_userspace_exit;
>       default:

Also those new hypercalls also should be added to hv_check_hypercall_access.

>               ret = HV_STATUS_INVALID_HYPERCALL_CODE;
> @@ -2912,6 +2926,7 @@ int kvm_get_hv_cpuid(struct kvm_vcpu *vcpu, struct 
> kvm_cpuid2 *cpuid,
>                       ent->ebx |= HV_SIGNAL_EVENTS;
>                       ent->ebx |= HV_ENABLE_EXTENDED_HYPERCALLS;
>                       ent->ebx |= HV_ACCESS_VP_REGISTERS;
> +                     ent->ebx |= HV_ACCESS_VSM;
>  
>                       ent->edx |= HV_X64_HYPERCALL_XMM_INPUT_AVAILABLE;
>                       ent->edx |= HV_X64_HYPERCALL_XMM_OUTPUT_AVAILABLE;

Best regards,
        Maxim Levitsky

> diff --git a/include/asm-generic/hyperv-tlfs.h 
> b/include/asm-generic/hyperv-tlfs.h
> index 24ea699a3d8e..a8b5c8a84bbc 100644
> --- a/include/asm-generic/hyperv-tlfs.h
> +++ b/include/asm-generic/hyperv-tlfs.h
> @@ -89,6 +89,7 @@
>  #define HV_ACCESS_STATS                              BIT(8)
>  #define HV_DEBUGGING                         BIT(11)
>  #define HV_CPU_MANAGEMENT                    BIT(12)
> +#define HV_ACCESS_VSM                                BIT(16)
>  #define HV_ACCESS_VP_REGISTERS                       BIT(17)
>  #define HV_ENABLE_EXTENDED_HYPERCALLS                BIT(20)
>  #define HV_ISOLATION                         BIT(22)
> @@ -147,9 +148,13 @@ union hv_reference_tsc_msr {
>  /* Declare the various hypercall operations. */
>  #define HVCALL_FLUSH_VIRTUAL_ADDRESS_SPACE   0x0002
>  #define HVCALL_FLUSH_VIRTUAL_ADDRESS_LIST    0x0003
> -#define HVCALL_ENABLE_VP_VTL                 0x000f
>  #define HVCALL_NOTIFY_LONG_SPIN_WAIT         0x0008
>  #define HVCALL_SEND_IPI                              0x000b
> +#define HVCALL_MODIFY_VTL_PROTECTION_MASK    0x000c
> +#define HVCALL_ENABLE_PARTITION_VTL          0x000d
> +#define HVCALL_ENABLE_VP_VTL                 0x000f
> +#define HVCALL_VTL_CALL                              0x0011
> +#define HVCALL_VTL_RETURN                    0x0012
>  #define HVCALL_FLUSH_VIRTUAL_ADDRESS_SPACE_EX        0x0013
>  #define HVCALL_FLUSH_VIRTUAL_ADDRESS_LIST_EX 0x0014
>  #define HVCALL_SEND_IPI_EX                   0x0015





Reply via email to