Those functions are used in the old intel vPMU implementation and are not
needed now.

Signed-off-by: Wei Wang <wei.w.w...@intel.com>
Cc: Paolo Bonzini <pbonz...@redhat.com>
Cc: Andi Kleen <a...@linux.intel.com>
Cc: Peter Zijlstra <pet...@infradead.org>
---
 arch/x86/kvm/pmu_intel.c | 103 -----------------------------------------------
 1 file changed, 103 deletions(-)

diff --git a/arch/x86/kvm/pmu_intel.c b/arch/x86/kvm/pmu_intel.c
index 4c6183b..66d7c09 100644
--- a/arch/x86/kvm/pmu_intel.c
+++ b/arch/x86/kvm/pmu_intel.c
@@ -23,87 +23,6 @@
 #define INTEL_PMU_RDPMC_FIXED_PMC (1ULL << 30)
 #define INTEL_PMU_RDPMC_COUNTER_MASK (INTEL_PMU_RDPMC_FIXED_PMC - 1)
 
-static struct kvm_event_hw_type_mapping intel_arch_events[] = {
-       /* Index must match CPUID 0x0A.EBX bit vector */
-       [0] = { 0x3c, 0x00, PERF_COUNT_HW_CPU_CYCLES },
-       [1] = { 0xc0, 0x00, PERF_COUNT_HW_INSTRUCTIONS },
-       [2] = { 0x3c, 0x01, PERF_COUNT_HW_BUS_CYCLES  },
-       [3] = { 0x2e, 0x4f, PERF_COUNT_HW_CACHE_REFERENCES },
-       [4] = { 0x2e, 0x41, PERF_COUNT_HW_CACHE_MISSES },
-       [5] = { 0xc4, 0x00, PERF_COUNT_HW_BRANCH_INSTRUCTIONS },
-       [6] = { 0xc5, 0x00, PERF_COUNT_HW_BRANCH_MISSES },
-       [7] = { 0x00, 0x30, PERF_COUNT_HW_REF_CPU_CYCLES },
-};
-
-/* mapping between fixed pmc index and intel_arch_events array */
-static int fixed_pmc_events[] = {1, 0, 7};
-
-static void reprogram_fixed_counters(struct kvm_pmu *pmu, u64 data)
-{
-       int i;
-
-       for (i = 0; i < pmu->nr_arch_fixed_counters; i++) {
-               u8 new_ctrl = fixed_ctrl_field(data, i);
-               u8 old_ctrl = fixed_ctrl_field(pmu->fixed_ctr_ctrl, i);
-               struct kvm_pmc *pmc;
-
-               pmc = get_fixed_pmc(pmu, MSR_CORE_PERF_FIXED_CTR0 + i);
-
-               if (old_ctrl == new_ctrl)
-                       continue;
-
-               reprogram_fixed_counter(pmc, new_ctrl, i);
-       }
-
-       pmu->fixed_ctr_ctrl = data;
-}
-
-/* function is called when global control register has been updated. */
-static void global_ctrl_changed(struct kvm_pmu *pmu, u64 data)
-{
-       int bit;
-       u64 diff = pmu->global_ctrl ^ data;
-
-       pmu->global_ctrl = data;
-
-       for_each_set_bit(bit, (unsigned long *)&diff, X86_PMC_IDX_MAX)
-               reprogram_counter(pmu, bit);
-}
-
-static unsigned intel_find_arch_event(struct kvm_pmu *pmu,
-                                     u8 event_select,
-                                     u8 unit_mask)
-{
-       int i;
-
-       for (i = 0; i < ARRAY_SIZE(intel_arch_events); i++)
-               if (intel_arch_events[i].eventsel == event_select
-                   && intel_arch_events[i].unit_mask == unit_mask
-                   && (pmu->available_event_types & (1 << i)))
-                       break;
-
-       if (i == ARRAY_SIZE(intel_arch_events))
-               return PERF_COUNT_HW_MAX;
-
-       return intel_arch_events[i].event_type;
-}
-
-static unsigned intel_find_fixed_event(int idx)
-{
-       if (idx >= ARRAY_SIZE(fixed_pmc_events))
-               return PERF_COUNT_HW_MAX;
-
-       return intel_arch_events[fixed_pmc_events[idx]].event_type;
-}
-
-/* check if a PMC is enabled by comparing it with globl_ctrl bits. */
-static bool intel_pmc_is_enabled(struct kvm_pmc *pmc)
-{
-       struct kvm_pmu *pmu = pmc_to_pmu(pmc);
-
-       return test_bit(pmc->idx, (unsigned long *)&pmu->global_ctrl);
-}
-
 static struct kvm_pmc *intel_pmc_idx_to_pmc(struct kvm_pmu *pmu, int pmc_idx)
 {
        if (pmc_idx < INTEL_PMC_IDX_FIXED)
@@ -128,23 +47,6 @@ static int intel_is_valid_msr_idx(struct kvm_vcpu *vcpu, 
unsigned idx)
                (fixed && idx >= pmu->nr_arch_fixed_counters);
 }
 
-static struct kvm_pmc *intel_msr_idx_to_pmc(struct kvm_vcpu *vcpu,
-                                           unsigned idx)
-{
-       struct kvm_pmu *pmu = vcpu_to_pmu(vcpu);
-       bool fixed = idx & (1u << 30);
-       struct kvm_pmc *counters;
-
-       idx &= ~(3u << 30);
-       if (!fixed && idx >= pmu->nr_arch_gp_counters)
-               return NULL;
-       if (fixed && idx >= pmu->nr_arch_fixed_counters)
-               return NULL;
-       counters = fixed ? pmu->fixed_counters : pmu->gp_counters;
-
-       return &counters[idx];
-}
-
 static bool intel_is_valid_msr(struct kvm_vcpu *vcpu, u32 msr)
 {
        struct kvm_pmu *pmu = vcpu_to_pmu(vcpu);
@@ -547,11 +449,6 @@ struct kvm_perf_switch_msr 
*intel_pmu_get_switch_msrs(struct kvm_vcpu *vcpu,
 }
 
 struct kvm_pmu_ops intel_pmu_ops = {
-       .find_arch_event = intel_find_arch_event,
-       .find_fixed_event = intel_find_fixed_event,
-       .pmc_is_enabled = intel_pmc_is_enabled,
-       .pmc_idx_to_pmc = intel_pmc_idx_to_pmc,
-       .msr_idx_to_pmc = intel_msr_idx_to_pmc,
        .is_valid_msr_idx = intel_is_valid_msr_idx,
        .is_valid_msr = intel_is_valid_msr,
        .get_msr = intel_pmu_get_msr,
-- 
2.7.4

Reply via email to