Re: [PATCH] ppc/pnv: Introduce pnv_chip_foreach_cpu()
On Wed Apr 24, 2024 at 7:30 PM AEST, Cédric Le Goater wrote: > This helper routine uses the machine definition, sockets, cores and > threads, to loop on all CPUs of the machine. Replace CPU_FOREACH() > with it. > > Signed-off-by: Cédric Le Goater Reviewed-by: Nicholas Piggin > --- > hw/ppc/pnv.c | 48 > 1 file changed, 36 insertions(+), 12 deletions(-) > > diff --git a/hw/ppc/pnv.c b/hw/ppc/pnv.c > index > 6e3a5ccdec764c8f6cbd076e27f59c7082e64876..5f400ed127921c4c3a45bc54863b2cafa53e7030 > 100644 > --- a/hw/ppc/pnv.c > +++ b/hw/ppc/pnv.c > @@ -2264,6 +2264,21 @@ PowerPCCPU *pnv_chip_find_cpu(PnvChip *chip, uint32_t > pir) > return NULL; > } > > +static void pnv_chip_foreach_cpu(PnvChip *chip, > + void (*fn)(PnvChip *chip, PowerPCCPU *cpu, void *opaque), > + void *opaque) > +{ > +int i, j; > + > +for (i = 0; i < chip->nr_cores; i++) { > +PnvCore *pc = chip->cores[i]; > + > +for (j = 0; j < CPU_CORE(pc)->nr_threads; j++) { > +fn(chip, pc->threads[j], opaque); > +} > +} > +} > + > static ICSState *pnv_ics_get(XICSFabric *xi, int irq) > { > PnvMachineState *pnv = PNV_MACHINE(xi); > @@ -2332,23 +2347,26 @@ static ICPState *pnv_icp_get(XICSFabric *xi, int pir) > return cpu ? ICP(pnv_cpu_state(cpu)->intc) : NULL; > } > > +static void pnv_pic_intc_print_info(PnvChip *chip, PowerPCCPU *cpu, > +void *opaque) > +{ > +PNV_CHIP_GET_CLASS(chip)->intc_print_info(chip, cpu, opaque); > +} > + > static void pnv_pic_print_info(InterruptStatsProvider *obj, > Monitor *mon) > { > PnvMachineState *pnv = PNV_MACHINE(obj); > int i; > -CPUState *cs; > > -CPU_FOREACH(cs) { > -PowerPCCPU *cpu = POWERPC_CPU(cs); > +for (i = 0; i < pnv->num_chips; i++) { > +PnvChip *chip = pnv->chips[i]; > > -/* XXX: loop on each chip/core/thread instead of CPU_FOREACH() */ > -PNV_CHIP_GET_CLASS(pnv->chips[0])->intc_print_info(pnv->chips[0], > cpu, > - mon); > -} > +/* First CPU presenters */ > +pnv_chip_foreach_cpu(chip, pnv_pic_intc_print_info, mon); > > -for (i = 0; i < pnv->num_chips; i++) { > -PNV_CHIP_GET_CLASS(pnv->chips[i])->pic_print_info(pnv->chips[i], > mon); > +/* Then other devices, PHB, PSI, XIVE */ > +PNV_CHIP_GET_CLASS(chip)->pic_print_info(chip, mon); > } > } > > @@ -2549,12 +2567,18 @@ static void pnv_cpu_do_nmi_on_cpu(CPUState *cs, > run_on_cpu_data arg) > } > } > > +static void pnv_cpu_do_nmi(PnvChip *chip, PowerPCCPU *cpu, void *opaque) > +{ > +async_run_on_cpu(CPU(cpu), pnv_cpu_do_nmi_on_cpu, RUN_ON_CPU_NULL); > +} > + > static void pnv_nmi(NMIState *n, int cpu_index, Error **errp) > { > -CPUState *cs; > +PnvMachineState *pnv = PNV_MACHINE(qdev_get_machine()); > +int i; > > -CPU_FOREACH(cs) { > -async_run_on_cpu(cs, pnv_cpu_do_nmi_on_cpu, RUN_ON_CPU_NULL); > +for (i = 0; i < pnv->num_chips; i++) { > +pnv_chip_foreach_cpu(pnv->chips[i], pnv_cpu_do_nmi, NULL); > } > } >
Re: [PATCH] ppc/pnv: Introduce pnv_chip_foreach_cpu()
On 24/4/24 11:30, Cédric Le Goater wrote: This helper routine uses the machine definition, sockets, cores and threads, to loop on all CPUs of the machine. Replace CPU_FOREACH() with it. Signed-off-by: Cédric Le Goater --- hw/ppc/pnv.c | 48 1 file changed, 36 insertions(+), 12 deletions(-) Reviewed-by: Philippe Mathieu-Daudé Thanks!
[PATCH] ppc/pnv: Introduce pnv_chip_foreach_cpu()
This helper routine uses the machine definition, sockets, cores and threads, to loop on all CPUs of the machine. Replace CPU_FOREACH() with it. Signed-off-by: Cédric Le Goater --- hw/ppc/pnv.c | 48 1 file changed, 36 insertions(+), 12 deletions(-) diff --git a/hw/ppc/pnv.c b/hw/ppc/pnv.c index 6e3a5ccdec764c8f6cbd076e27f59c7082e64876..5f400ed127921c4c3a45bc54863b2cafa53e7030 100644 --- a/hw/ppc/pnv.c +++ b/hw/ppc/pnv.c @@ -2264,6 +2264,21 @@ PowerPCCPU *pnv_chip_find_cpu(PnvChip *chip, uint32_t pir) return NULL; } +static void pnv_chip_foreach_cpu(PnvChip *chip, + void (*fn)(PnvChip *chip, PowerPCCPU *cpu, void *opaque), + void *opaque) +{ +int i, j; + +for (i = 0; i < chip->nr_cores; i++) { +PnvCore *pc = chip->cores[i]; + +for (j = 0; j < CPU_CORE(pc)->nr_threads; j++) { +fn(chip, pc->threads[j], opaque); +} +} +} + static ICSState *pnv_ics_get(XICSFabric *xi, int irq) { PnvMachineState *pnv = PNV_MACHINE(xi); @@ -2332,23 +2347,26 @@ static ICPState *pnv_icp_get(XICSFabric *xi, int pir) return cpu ? ICP(pnv_cpu_state(cpu)->intc) : NULL; } +static void pnv_pic_intc_print_info(PnvChip *chip, PowerPCCPU *cpu, +void *opaque) +{ +PNV_CHIP_GET_CLASS(chip)->intc_print_info(chip, cpu, opaque); +} + static void pnv_pic_print_info(InterruptStatsProvider *obj, Monitor *mon) { PnvMachineState *pnv = PNV_MACHINE(obj); int i; -CPUState *cs; -CPU_FOREACH(cs) { -PowerPCCPU *cpu = POWERPC_CPU(cs); +for (i = 0; i < pnv->num_chips; i++) { +PnvChip *chip = pnv->chips[i]; -/* XXX: loop on each chip/core/thread instead of CPU_FOREACH() */ -PNV_CHIP_GET_CLASS(pnv->chips[0])->intc_print_info(pnv->chips[0], cpu, - mon); -} +/* First CPU presenters */ +pnv_chip_foreach_cpu(chip, pnv_pic_intc_print_info, mon); -for (i = 0; i < pnv->num_chips; i++) { -PNV_CHIP_GET_CLASS(pnv->chips[i])->pic_print_info(pnv->chips[i], mon); +/* Then other devices, PHB, PSI, XIVE */ +PNV_CHIP_GET_CLASS(chip)->pic_print_info(chip, mon); } } @@ -2549,12 +2567,18 @@ static void pnv_cpu_do_nmi_on_cpu(CPUState *cs, run_on_cpu_data arg) } } +static void pnv_cpu_do_nmi(PnvChip *chip, PowerPCCPU *cpu, void *opaque) +{ +async_run_on_cpu(CPU(cpu), pnv_cpu_do_nmi_on_cpu, RUN_ON_CPU_NULL); +} + static void pnv_nmi(NMIState *n, int cpu_index, Error **errp) { -CPUState *cs; +PnvMachineState *pnv = PNV_MACHINE(qdev_get_machine()); +int i; -CPU_FOREACH(cs) { -async_run_on_cpu(cs, pnv_cpu_do_nmi_on_cpu, RUN_ON_CPU_NULL); +for (i = 0; i < pnv->num_chips; i++) { +pnv_chip_foreach_cpu(pnv->chips[i], pnv_cpu_do_nmi, NULL); } } -- 2.44.0