To support multiple PMUs, each PMU will need its own accounting data. As we don't know how (in general) many PMUs we'll have to support at compile-time, we must allocate the data at runtime dynamically
Signed-off-by: Mark Rutland <mark.rutl...@arm.com> Reviewed-by: Will Deacon <will.dea...@arm.com> --- arch/arm/kernel/perf_event_cpu.c | 73 ++++++++++++++++++++++++++++------------ 1 file changed, 51 insertions(+), 22 deletions(-) diff --git a/arch/arm/kernel/perf_event_cpu.c b/arch/arm/kernel/perf_event_cpu.c index 80f1a26..ae5be75 100644 --- a/arch/arm/kernel/perf_event_cpu.c +++ b/arch/arm/kernel/perf_event_cpu.c @@ -33,9 +33,25 @@ /* Set at runtime when we know what CPU type we are. */ static struct arm_pmu *cpu_pmu; -static DEFINE_PER_CPU(struct perf_event * [ARMPMU_MAX_HWEVENTS], hw_events); -static DEFINE_PER_CPU(unsigned long [BITS_TO_LONGS(ARMPMU_MAX_HWEVENTS)], used_mask); -static DEFINE_PER_CPU(struct pmu_hw_events, cpu_hw_events); +/* + * All of the dynamically sized pmu_hw data for the number of events supported + * by CPU PMUs, aggregated together for easier allocation / freeing. + */ +struct cpu_pmu_hw { + struct pmu_hw_events cpu_hw_events; + struct perf_event *hw_events[ARMPMU_MAX_HWEVENTS]; + unsigned long used_mask[BITS_TO_LONGS(ARMPMU_MAX_HWEVENTS)]; +}; + +/* + * For mapping between an arm_pmu for a CPU and its CPU-affine data. + */ +struct cpu_pmu { + struct arm_pmu armpmu; + struct cpu_pmu_hw __percpu *cpu_hw; +}; + +#define to_cpu_pmu(p) (container_of(p, struct cpu_pmu, armpmu)) /* * Despite the names, these two functions are CPU-specific and are used @@ -68,7 +84,9 @@ EXPORT_SYMBOL_GPL(perf_num_counters); static struct pmu_hw_events *cpu_pmu_get_cpu_events(struct arm_pmu *pmu) { - return &__get_cpu_var(cpu_hw_events); + struct cpu_pmu *cpu_pmu = to_cpu_pmu(pmu); + struct cpu_pmu_hw *hw = this_cpu_ptr(cpu_pmu->cpu_hw); + return &hw->cpu_hw_events; } static void cpu_pmu_free_irq(struct arm_pmu *cpu_pmu) @@ -132,23 +150,25 @@ static int cpu_pmu_request_irq(struct arm_pmu *cpu_pmu, irq_handler_t handler) return 0; } -static void cpu_pmu_init(struct arm_pmu *cpu_pmu) +static void cpu_pmu_init(struct cpu_pmu *cpu_pmu) { int cpu; + struct arm_pmu *arm_pmu = &cpu_pmu->armpmu; + for_each_possible_cpu(cpu) { - struct pmu_hw_events *events = &per_cpu(cpu_hw_events, cpu); - events->events = per_cpu(hw_events, cpu); - events->used_mask = per_cpu(used_mask, cpu); - raw_spin_lock_init(&events->pmu_lock); + struct cpu_pmu_hw *cpu_hw = per_cpu_ptr(cpu_pmu->cpu_hw, cpu); + cpu_hw->cpu_hw_events.events = cpu_hw->hw_events; + cpu_hw->cpu_hw_events.used_mask = cpu_hw->used_mask; + raw_spin_lock_init(&cpu_hw->cpu_hw_events.pmu_lock); } - cpu_pmu->get_hw_events = cpu_pmu_get_cpu_events; - cpu_pmu->request_irq = cpu_pmu_request_irq; - cpu_pmu->free_irq = cpu_pmu_free_irq; + arm_pmu->get_hw_events = cpu_pmu_get_cpu_events; + arm_pmu->request_irq = cpu_pmu_request_irq; + arm_pmu->free_irq = cpu_pmu_free_irq; /* Ensure the PMU has sane values out of reset. */ - if (cpu_pmu->reset) - on_each_cpu(cpu_pmu->reset, cpu_pmu, 1); + if (arm_pmu->reset) + on_each_cpu(arm_pmu->reset, arm_pmu, 1); } /* @@ -255,7 +275,7 @@ static int cpu_pmu_device_probe(struct platform_device *pdev) const struct of_device_id *of_id; int (*init_fn)(struct arm_pmu *); struct device_node *node = pdev->dev.of_node; - struct arm_pmu *pmu; + struct cpu_pmu *pmu; int ret = -ENODEV; if (cpu_pmu) { @@ -263,34 +283,43 @@ static int cpu_pmu_device_probe(struct platform_device *pdev) return -ENOSPC; } - pmu = kzalloc(sizeof(struct arm_pmu), GFP_KERNEL); + pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); if (!pmu) { pr_info("failed to allocate PMU device!"); return -ENOMEM; } + pmu->cpu_hw = alloc_percpu(struct cpu_pmu_hw); + if (!pmu->cpu_hw) { + pr_info("failed to allocate PMU hw data!\n"); + ret = -ENOMEM; + goto out_pmu; + } + if (node && (of_id = of_match_node(cpu_pmu_of_device_ids, pdev->dev.of_node))) { init_fn = of_id->data; - ret = init_fn(pmu); + ret = init_fn(&pmu->armpmu); } else { - ret = probe_current_pmu(pmu); + ret = probe_current_pmu(&pmu->armpmu); } if (ret) { pr_info("failed to probe PMU!"); - goto out_free; + goto out_hw; } - cpu_pmu = pmu; + cpu_pmu = &pmu->armpmu; cpu_pmu->plat_device = pdev; - cpu_pmu_init(cpu_pmu); + cpu_pmu_init(pmu); ret = armpmu_register(cpu_pmu, -1); if (!ret) return 0; -out_free: +out_hw: + free_percpu(pmu->cpu_hw); pr_info("failed to register PMU devices!"); +out_pmu: kfree(pmu); return ret; } -- 1.8.1.1 _______________________________________________ devicetree-discuss mailing list devicetree-discuss@lists.ozlabs.org https://lists.ozlabs.org/listinfo/devicetree-discuss