This replace the variable names to make them self-explaining. The
tracepoint isn't changed accordingly because they're part of ABI:

   * @hsr to @esr
   * @hsr_ec to @ec
   * Use kvm_vcpu_trap_get_class() helper if possible

Signed-off-by: Gavin Shan <gs...@redhat.com>
---
 arch/arm64/kvm/handle_exit.c | 28 ++++++++++++++--------------
 arch/arm64/kvm/hyp/switch.c  |  9 ++++-----
 arch/arm64/kvm/sys_regs.c    | 30 +++++++++++++++---------------
 3 files changed, 33 insertions(+), 34 deletions(-)

diff --git a/arch/arm64/kvm/handle_exit.c b/arch/arm64/kvm/handle_exit.c
index 00858db82a64..e3b3dcd5b811 100644
--- a/arch/arm64/kvm/handle_exit.c
+++ b/arch/arm64/kvm/handle_exit.c
@@ -123,13 +123,13 @@ static int kvm_handle_wfx(struct kvm_vcpu *vcpu, struct 
kvm_run *run)
  */
 static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run)
 {
-       u32 hsr = kvm_vcpu_get_esr(vcpu);
+       u32 esr = kvm_vcpu_get_esr(vcpu);
        int ret = 0;
 
        run->exit_reason = KVM_EXIT_DEBUG;
-       run->debug.arch.hsr = hsr;
+       run->debug.arch.hsr = esr;
 
-       switch (ESR_ELx_EC(hsr)) {
+       switch (kvm_vcpu_trap_get_class(esr)) {
        case ESR_ELx_EC_WATCHPT_LOW:
                run->debug.arch.far = vcpu->arch.fault.far_el2;
                /* fall through */
@@ -139,8 +139,8 @@ static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, 
struct kvm_run *run)
        case ESR_ELx_EC_BRK64:
                break;
        default:
-               kvm_err("%s: un-handled case hsr: %#08x\n",
-                       __func__, (unsigned int) hsr);
+               kvm_err("%s: un-handled case esr: %#08x\n",
+                       __func__, (unsigned int)esr);
                ret = -1;
                break;
        }
@@ -150,10 +150,10 @@ static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, 
struct kvm_run *run)
 
 static int kvm_handle_unknown_ec(struct kvm_vcpu *vcpu, struct kvm_run *run)
 {
-       u32 hsr = kvm_vcpu_get_esr(vcpu);
+       u32 esr = kvm_vcpu_get_esr(vcpu);
 
-       kvm_pr_unimpl("Unknown exception class: hsr: %#08x -- %s\n",
-                     hsr, esr_get_class_string(hsr));
+       kvm_pr_unimpl("Unknown exception class: esr: %#08x -- %s\n",
+                     esr, esr_get_class_string(esr));
 
        kvm_inject_undefined(vcpu);
        return 1;
@@ -230,10 +230,10 @@ static exit_handle_fn arm_exit_handlers[] = {
 
 static exit_handle_fn kvm_get_exit_handler(struct kvm_vcpu *vcpu)
 {
-       u32 hsr = kvm_vcpu_get_esr(vcpu);
-       u8 hsr_ec = ESR_ELx_EC(hsr);
+       u32 esr = kvm_vcpu_get_esr(vcpu);
+       u8 ec = kvm_vcpu_trap_get_class(esr);
 
-       return arm_exit_handlers[hsr_ec];
+       return arm_exit_handlers[ec];
 }
 
 /*
@@ -273,15 +273,15 @@ int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run 
*run,
 {
        if (ARM_SERROR_PENDING(exception_index)) {
                u32 esr = kvm_vcpu_get_esr(vcpu);
-               u8 hsr_ec = ESR_ELx_EC(esr);
+               u8 ec = kvm_vcpu_trap_get_class(esr);
 
                /*
                 * HVC/SMC already have an adjusted PC, which we need
                 * to correct in order to return to after having
                 * injected the SError.
                 */
-               if (hsr_ec == ESR_ELx_EC_HVC32 || hsr_ec == ESR_ELx_EC_HVC64 ||
-                   hsr_ec == ESR_ELx_EC_SMC32 || hsr_ec == ESR_ELx_EC_SMC64) {
+               if (ec == ESR_ELx_EC_HVC32 || ec == ESR_ELx_EC_HVC64 ||
+                   ec == ESR_ELx_EC_SMC32 || ec == ESR_ELx_EC_SMC64) {
                        u32 adj =  kvm_vcpu_trap_il_is32bit(esr) ? 4 : 2;
                        *vcpu_pc(vcpu) -= adj;
                }
diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c
index 369f22f49f3d..7bf4840bf90e 100644
--- a/arch/arm64/kvm/hyp/switch.c
+++ b/arch/arm64/kvm/hyp/switch.c
@@ -356,8 +356,8 @@ static bool __hyp_text __populate_fault_info(struct 
kvm_vcpu *vcpu)
 static bool __hyp_text __hyp_handle_fpsimd(struct kvm_vcpu *vcpu)
 {
        u32 esr = kvm_vcpu_get_esr(vcpu);
+       u8 ec = kvm_vcpu_trap_get_class(esr);
        bool vhe, sve_guest, sve_host;
-       u8 hsr_ec;
 
        if (!system_supports_fpsimd())
                return false;
@@ -372,14 +372,13 @@ static bool __hyp_text __hyp_handle_fpsimd(struct 
kvm_vcpu *vcpu)
                vhe = has_vhe();
        }
 
-       hsr_ec = kvm_vcpu_trap_get_class(esr);
-       if (hsr_ec != ESR_ELx_EC_FP_ASIMD &&
-           hsr_ec != ESR_ELx_EC_SVE)
+       if (ec != ESR_ELx_EC_FP_ASIMD &&
+           ec != ESR_ELx_EC_SVE)
                return false;
 
        /* Don't handle SVE traps for non-SVE vcpus here: */
        if (!sve_guest)
-               if (hsr_ec != ESR_ELx_EC_FP_ASIMD)
+               if (ec != ESR_ELx_EC_FP_ASIMD)
                        return false;
 
        /* Valid trap.  Switch the context: */
diff --git a/arch/arm64/kvm/sys_regs.c b/arch/arm64/kvm/sys_regs.c
index 012fff834a4b..58f81ab519af 100644
--- a/arch/arm64/kvm/sys_regs.c
+++ b/arch/arm64/kvm/sys_regs.c
@@ -2182,10 +2182,10 @@ static void unhandled_cp_access(struct kvm_vcpu *vcpu,
                                struct sys_reg_params *params)
 {
        u32 esr = kvm_vcpu_get_esr(vcpu);
-       u8 hsr_ec = kvm_vcpu_trap_get_class(esr);
+       u8 ec = kvm_vcpu_trap_get_class(esr);
        int cp = -1;
 
-       switch(hsr_ec) {
+       switch (ec) {
        case ESR_ELx_EC_CP15_32:
        case ESR_ELx_EC_CP15_64:
                cp = 15;
@@ -2216,17 +2216,17 @@ static int kvm_handle_cp_64(struct kvm_vcpu *vcpu,
                            size_t nr_specific)
 {
        struct sys_reg_params params;
-       u32 hsr = kvm_vcpu_get_esr(vcpu);
-       int Rt = kvm_vcpu_sys_get_rt(hsr);
-       int Rt2 = (hsr >> 10) & 0x1f;
+       u32 esr = kvm_vcpu_get_esr(vcpu);
+       int Rt = kvm_vcpu_sys_get_rt(esr);
+       int Rt2 = (esr >> 10) & 0x1f;
 
        params.is_aarch32 = true;
        params.is_32bit = false;
-       params.CRm = (hsr >> 1) & 0xf;
-       params.is_write = ((hsr & 1) == 0);
+       params.CRm = (esr >> 1) & 0xf;
+       params.is_write = ((esr & 1) == 0);
 
        params.Op0 = 0;
-       params.Op1 = (hsr >> 16) & 0xf;
+       params.Op1 = (esr >> 16) & 0xf;
        params.Op2 = 0;
        params.CRn = 0;
 
@@ -2273,18 +2273,18 @@ static int kvm_handle_cp_32(struct kvm_vcpu *vcpu,
                            size_t nr_specific)
 {
        struct sys_reg_params params;
-       u32 hsr = kvm_vcpu_get_esr(vcpu);
-       int Rt  = kvm_vcpu_sys_get_rt(hsr);
+       u32 esr = kvm_vcpu_get_esr(vcpu);
+       int Rt = kvm_vcpu_sys_get_rt(esr);
 
        params.is_aarch32 = true;
        params.is_32bit = true;
-       params.CRm = (hsr >> 1) & 0xf;
+       params.CRm = (esr >> 1) & 0xf;
        params.regval = vcpu_get_reg(vcpu, Rt);
-       params.is_write = ((hsr & 1) == 0);
-       params.CRn = (hsr >> 10) & 0xf;
+       params.is_write = ((esr & 1) == 0);
+       params.CRn = (esr >> 10) & 0xf;
        params.Op0 = 0;
-       params.Op1 = (hsr >> 14) & 0x7;
-       params.Op2 = (hsr >> 17) & 0x7;
+       params.Op1 = (esr >> 14) & 0x7;
+       params.Op2 = (esr >> 17) & 0x7;
 
        if (!emulate_cp(vcpu, &params, target_specific, nr_specific) ||
            !emulate_cp(vcpu, &params, global, nr_global)) {
-- 
2.23.0

_______________________________________________
kvmarm mailing list
kvmarm@lists.cs.columbia.edu
https://lists.cs.columbia.edu/mailman/listinfo/kvmarm

Reply via email to