Re: [PATCH V5 04/25] perf/x86/intel: Hybrid PMU support for perf capabilities
On 4/8/2021 9:40 AM, Peter Zijlstra wrote: @@ -4330,7 +4347,7 @@ static int intel_pmu_check_period(struct perf_event *event, u64 value) static int intel_pmu_aux_output_match(struct perf_event *event) { - if (!x86_pmu.intel_cap.pebs_output_pt_available) + if (!intel_pmu_has_cap(event, PERF_CAP_PT_IDX)) return 0; return is_intel_pt_event(event); these sites can have !event->pmu ? I don't think the event->pmu can be NULL, but it could be pt_pmu.pmu. If so, it should be a problem. I think I will still use the x86_pmu.intel_cap.pebs_output_pt_available here in V6. The worst case is that we lost the PEBS via PT support on the small core for now. I guess Alexander may provide a separate patch later to enable the PEBS via PT support on the ADL small core. Thanks, Kan
Re: [PATCH V5 04/25] perf/x86/intel: Hybrid PMU support for perf capabilities
On 4/8/2021 1:00 PM, Peter Zijlstra wrote: On Mon, Apr 05, 2021 at 08:10:46AM -0700, kan.li...@linux.intel.com wrote: +#define is_hybrid()(!!x86_pmu.num_hybrid_pmus) Given this is sprinkled all over the place, can you make this a static_key_false + static_branch_unlikely() such that the hybrid case is out-of-line? Sure, I will add a new static_key_false "perf_is_hybrid" to indicate a hybrid system as below (not test yet). diff --git a/arch/x86/events/core.c b/arch/x86/events/core.c index f8d1222..bd6412e 100644 --- a/arch/x86/events/core.c +++ b/arch/x86/events/core.c @@ -54,6 +54,7 @@ DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { DEFINE_STATIC_KEY_FALSE(rdpmc_never_available_key); DEFINE_STATIC_KEY_FALSE(rdpmc_always_available_key); +DEFINE_STATIC_KEY_FALSE(perf_is_hybrid); /* * This here uses DEFINE_STATIC_CALL_NULL() to get a static_call defined diff --git a/arch/x86/events/intel/core.c b/arch/x86/events/intel/core.c index 2b553d9..7cef3cd 100644 --- a/arch/x86/events/intel/core.c +++ b/arch/x86/events/intel/core.c @@ -6119,6 +6119,7 @@ __init int intel_pmu_init(void) GFP_KERNEL); if (!x86_pmu.hybrid_pmu) return -ENOMEM; + static_branch_enable(_is_hybrid); x86_pmu.num_hybrid_pmus = X86_HYBRID_NUM_PMUS; x86_pmu.late_ack = true; diff --git a/arch/x86/events/perf_event.h b/arch/x86/events/perf_event.h index bfbecde..d6383d1 100644 --- a/arch/x86/events/perf_event.h +++ b/arch/x86/events/perf_event.h @@ -663,8 +663,8 @@ static __always_inline struct x86_hybrid_pmu *hybrid_pmu(struct pmu *pmu) return container_of(pmu, struct x86_hybrid_pmu, pmu); } -/* The number of hybrid PMUs implies whether it's a hybrid system */ -#define is_hybrid()(!!x86_pmu.num_hybrid_pmus) +extern struct static_key_false perf_is_hybrid; +#define is_hybrid()static_branch_unlikely(_is_hybrid) #define hybrid(_pmu, _field) \ ({ \ Thanks, Kan
Re: [PATCH V5 04/25] perf/x86/intel: Hybrid PMU support for perf capabilities
On Mon, Apr 05, 2021 at 08:10:46AM -0700, kan.li...@linux.intel.com wrote: > +#define is_hybrid() (!!x86_pmu.num_hybrid_pmus) Given this is sprinkled all over the place, can you make this a static_key_false + static_branch_unlikely() such that the hybrid case is out-of-line?
Re: [PATCH V5 04/25] perf/x86/intel: Hybrid PMU support for perf capabilities
On Mon, Apr 05, 2021 at 08:10:46AM -0700, kan.li...@linux.intel.com wrote: > +static inline bool intel_pmu_has_cap(struct perf_event *event, int idx) > +{ > + union perf_capabilities *intel_cap; > + > + intel_cap = is_hybrid() ? _pmu(event->pmu)->intel_cap : > + _pmu.intel_cap; This isn't: intel_cap = _pmu(event->pmu)->intel_cap; because.. > + > + return test_bit(idx, (unsigned long *)_cap->capabilities); > +} > @@ -3712,7 +3721,7 @@ static int intel_pmu_hw_config(struct perf_event *event) >* with a slots event as group leader. When the slots event >* is used in a metrics group, it too cannot support sampling. >*/ > - if (x86_pmu.intel_cap.perf_metrics && is_topdown_event(event)) { > + if (intel_pmu_has_cap(event, PERF_CAP_METRICS_IDX) && > is_topdown_event(event)) { > if (event->attr.config1 || event->attr.config2) > return -EINVAL; > > @@ -4330,7 +4347,7 @@ static int intel_pmu_check_period(struct perf_event > *event, u64 value) > > static int intel_pmu_aux_output_match(struct perf_event *event) > { > - if (!x86_pmu.intel_cap.pebs_output_pt_available) > + if (!intel_pmu_has_cap(event, PERF_CAP_PT_IDX)) > return 0; > > return is_intel_pt_event(event); these sites can have !event->pmu ?
Re: [PATCH V5 04/25] perf/x86/intel: Hybrid PMU support for perf capabilities
On Thu, Apr 08, 2021 at 03:40:56PM +0200, Peter Zijlstra wrote: > On Mon, Apr 05, 2021 at 08:10:46AM -0700, kan.li...@linux.intel.com wrote: > > +static inline bool intel_pmu_has_cap(struct perf_event *event, int idx) > > +{ > > + union perf_capabilities *intel_cap; > > + > > + intel_cap = is_hybrid() ? _pmu(event->pmu)->intel_cap : > > + _pmu.intel_cap; > > This isn't: > > intel_cap = _pmu(event->pmu)->intel_cap; Ah no, its because you want a pointer and GCC is being silly about that. I have something for that, hold on ;-)