On Mon, Mar 24, 2025, Mingwei Zhang wrote:
>  static void kvm_pmu_incr_counter(struct kvm_pmc *pmc)
>  {
> -     pmc->emulated_counter++;
> -     kvm_pmu_request_counter_reprogram(pmc);
> +     struct kvm_vcpu *vcpu = pmc->vcpu;
> +
> +     /*
> +      * For perf-based PMUs, accumulate software-emulated events separately
> +      * from pmc->counter, as pmc->counter is offset by the count of the
> +      * associated perf event. Request reprogramming, which will consult
> +      * both emulated and hardware-generated events to detect overflow.
> +      */
> +     if (!kvm_mediated_pmu_enabled(vcpu)) {
> +             pmc->emulated_counter++;
> +             kvm_pmu_request_counter_reprogram(pmc);
> +             return;
> +     }
> +
> +     /*
> +      * For mediated PMUs, pmc->counter is updated when the vCPU's PMU is
> +      * put, and will be loaded into hardware when the PMU is loaded. Simply
> +      * increment the counter and signal overflow if it wraps to zero.
> +      */
> +     pmc->counter = (pmc->counter + 1) & pmc_bitmask(pmc);
> +     if (!pmc->counter) {

Ugh, this is broken for the fastpath.  If kvm_skip_emulated_instruction() is
invoked by handle_fastpath_set_msr_irqoff() or handle_fastpath_hlt(), KVM may
consume stale information (GLOBAL_CTRL, GLOBAL_STATUS and PMCs), and even if KVM
gets lucky and those are all fresh, the PMC and GLOBAL_STATUS changes won't be
propagated back to hardware.

The best idea I have is to track whether or not the guest may be counting 
branches
and/or instruction based on eventsels, and then bail from fastpaths that need to
skip instructions.  That flag would also be useful to further optimize
kvm_pmu_trigger_event().

> +             pmc_to_pmu(pmc)->global_status |= BIT_ULL(pmc->idx);
> +             if (pmc_pmi_enabled(pmc))
> +                     kvm_make_request(KVM_REQ_PMI, vcpu);
> +     }
>  }
>  
>  static inline bool cpl_is_matched(struct kvm_pmc *pmc)
> -- 
> 2.49.0.395.g12beb8f557-goog
> 

Reply via email to