Re: [PATCH] ppc/pnv: Introduce pnv_chip_foreach_cpu()

2024-05-02 Thread Nicholas Piggin
On Wed Apr 24, 2024 at 7:30 PM AEST, Cédric Le Goater wrote:
> This helper routine uses the machine definition, sockets, cores and
> threads, to loop on all CPUs of the machine. Replace CPU_FOREACH()
> with it.
>
> Signed-off-by: Cédric Le Goater 

Reviewed-by: Nicholas Piggin 

> ---
>  hw/ppc/pnv.c | 48 
>  1 file changed, 36 insertions(+), 12 deletions(-)
>
> diff --git a/hw/ppc/pnv.c b/hw/ppc/pnv.c
> index 
> 6e3a5ccdec764c8f6cbd076e27f59c7082e64876..5f400ed127921c4c3a45bc54863b2cafa53e7030
>  100644
> --- a/hw/ppc/pnv.c
> +++ b/hw/ppc/pnv.c
> @@ -2264,6 +2264,21 @@ PowerPCCPU *pnv_chip_find_cpu(PnvChip *chip, uint32_t 
> pir)
>  return NULL;
>  }
>  
> +static void pnv_chip_foreach_cpu(PnvChip *chip,
> +   void (*fn)(PnvChip *chip, PowerPCCPU *cpu, void *opaque),
> +   void *opaque)
> +{
> +int i, j;
> +
> +for (i = 0; i < chip->nr_cores; i++) {
> +PnvCore *pc = chip->cores[i];
> +
> +for (j = 0; j < CPU_CORE(pc)->nr_threads; j++) {
> +fn(chip, pc->threads[j], opaque);
> +}
> +}
> +}
> +
>  static ICSState *pnv_ics_get(XICSFabric *xi, int irq)
>  {
>  PnvMachineState *pnv = PNV_MACHINE(xi);
> @@ -2332,23 +2347,26 @@ static ICPState *pnv_icp_get(XICSFabric *xi, int pir)
>  return cpu ? ICP(pnv_cpu_state(cpu)->intc) : NULL;
>  }
>  
> +static void pnv_pic_intc_print_info(PnvChip *chip, PowerPCCPU *cpu,
> +void *opaque)
> +{
> +PNV_CHIP_GET_CLASS(chip)->intc_print_info(chip, cpu, opaque);
> +}
> +
>  static void pnv_pic_print_info(InterruptStatsProvider *obj,
> Monitor *mon)
>  {
>  PnvMachineState *pnv = PNV_MACHINE(obj);
>  int i;
> -CPUState *cs;
>  
> -CPU_FOREACH(cs) {
> -PowerPCCPU *cpu = POWERPC_CPU(cs);
> +for (i = 0; i < pnv->num_chips; i++) {
> +PnvChip *chip = pnv->chips[i];
>  
> -/* XXX: loop on each chip/core/thread instead of CPU_FOREACH() */
> -PNV_CHIP_GET_CLASS(pnv->chips[0])->intc_print_info(pnv->chips[0], 
> cpu,
> -   mon);
> -}
> +/* First CPU presenters */
> +pnv_chip_foreach_cpu(chip, pnv_pic_intc_print_info, mon);
>  
> -for (i = 0; i < pnv->num_chips; i++) {
> -PNV_CHIP_GET_CLASS(pnv->chips[i])->pic_print_info(pnv->chips[i], 
> mon);
> +/* Then other devices, PHB, PSI, XIVE */
> +PNV_CHIP_GET_CLASS(chip)->pic_print_info(chip, mon);
>  }
>  }
>  
> @@ -2549,12 +2567,18 @@ static void pnv_cpu_do_nmi_on_cpu(CPUState *cs, 
> run_on_cpu_data arg)
>  }
>  }
>  
> +static void pnv_cpu_do_nmi(PnvChip *chip, PowerPCCPU *cpu, void *opaque)
> +{
> +async_run_on_cpu(CPU(cpu), pnv_cpu_do_nmi_on_cpu, RUN_ON_CPU_NULL);
> +}
> +
>  static void pnv_nmi(NMIState *n, int cpu_index, Error **errp)
>  {
> -CPUState *cs;
> +PnvMachineState *pnv = PNV_MACHINE(qdev_get_machine());
> +int i;
>  
> -CPU_FOREACH(cs) {
> -async_run_on_cpu(cs, pnv_cpu_do_nmi_on_cpu, RUN_ON_CPU_NULL);
> +for (i = 0; i < pnv->num_chips; i++) {
> +pnv_chip_foreach_cpu(pnv->chips[i], pnv_cpu_do_nmi, NULL);
>  }
>  }
>  




Re: [PATCH] ppc/pnv: Introduce pnv_chip_foreach_cpu()

2024-04-24 Thread Philippe Mathieu-Daudé

On 24/4/24 11:30, Cédric Le Goater wrote:

This helper routine uses the machine definition, sockets, cores and
threads, to loop on all CPUs of the machine. Replace CPU_FOREACH()
with it.

Signed-off-by: Cédric Le Goater 
---
  hw/ppc/pnv.c | 48 
  1 file changed, 36 insertions(+), 12 deletions(-)


Reviewed-by: Philippe Mathieu-Daudé 

Thanks!



[PATCH] ppc/pnv: Introduce pnv_chip_foreach_cpu()

2024-04-24 Thread Cédric Le Goater
This helper routine uses the machine definition, sockets, cores and
threads, to loop on all CPUs of the machine. Replace CPU_FOREACH()
with it.

Signed-off-by: Cédric Le Goater 
---
 hw/ppc/pnv.c | 48 
 1 file changed, 36 insertions(+), 12 deletions(-)

diff --git a/hw/ppc/pnv.c b/hw/ppc/pnv.c
index 
6e3a5ccdec764c8f6cbd076e27f59c7082e64876..5f400ed127921c4c3a45bc54863b2cafa53e7030
 100644
--- a/hw/ppc/pnv.c
+++ b/hw/ppc/pnv.c
@@ -2264,6 +2264,21 @@ PowerPCCPU *pnv_chip_find_cpu(PnvChip *chip, uint32_t 
pir)
 return NULL;
 }
 
+static void pnv_chip_foreach_cpu(PnvChip *chip,
+   void (*fn)(PnvChip *chip, PowerPCCPU *cpu, void *opaque),
+   void *opaque)
+{
+int i, j;
+
+for (i = 0; i < chip->nr_cores; i++) {
+PnvCore *pc = chip->cores[i];
+
+for (j = 0; j < CPU_CORE(pc)->nr_threads; j++) {
+fn(chip, pc->threads[j], opaque);
+}
+}
+}
+
 static ICSState *pnv_ics_get(XICSFabric *xi, int irq)
 {
 PnvMachineState *pnv = PNV_MACHINE(xi);
@@ -2332,23 +2347,26 @@ static ICPState *pnv_icp_get(XICSFabric *xi, int pir)
 return cpu ? ICP(pnv_cpu_state(cpu)->intc) : NULL;
 }
 
+static void pnv_pic_intc_print_info(PnvChip *chip, PowerPCCPU *cpu,
+void *opaque)
+{
+PNV_CHIP_GET_CLASS(chip)->intc_print_info(chip, cpu, opaque);
+}
+
 static void pnv_pic_print_info(InterruptStatsProvider *obj,
Monitor *mon)
 {
 PnvMachineState *pnv = PNV_MACHINE(obj);
 int i;
-CPUState *cs;
 
-CPU_FOREACH(cs) {
-PowerPCCPU *cpu = POWERPC_CPU(cs);
+for (i = 0; i < pnv->num_chips; i++) {
+PnvChip *chip = pnv->chips[i];
 
-/* XXX: loop on each chip/core/thread instead of CPU_FOREACH() */
-PNV_CHIP_GET_CLASS(pnv->chips[0])->intc_print_info(pnv->chips[0], cpu,
-   mon);
-}
+/* First CPU presenters */
+pnv_chip_foreach_cpu(chip, pnv_pic_intc_print_info, mon);
 
-for (i = 0; i < pnv->num_chips; i++) {
-PNV_CHIP_GET_CLASS(pnv->chips[i])->pic_print_info(pnv->chips[i], mon);
+/* Then other devices, PHB, PSI, XIVE */
+PNV_CHIP_GET_CLASS(chip)->pic_print_info(chip, mon);
 }
 }
 
@@ -2549,12 +2567,18 @@ static void pnv_cpu_do_nmi_on_cpu(CPUState *cs, 
run_on_cpu_data arg)
 }
 }
 
+static void pnv_cpu_do_nmi(PnvChip *chip, PowerPCCPU *cpu, void *opaque)
+{
+async_run_on_cpu(CPU(cpu), pnv_cpu_do_nmi_on_cpu, RUN_ON_CPU_NULL);
+}
+
 static void pnv_nmi(NMIState *n, int cpu_index, Error **errp)
 {
-CPUState *cs;
+PnvMachineState *pnv = PNV_MACHINE(qdev_get_machine());
+int i;
 
-CPU_FOREACH(cs) {
-async_run_on_cpu(cs, pnv_cpu_do_nmi_on_cpu, RUN_ON_CPU_NULL);
+for (i = 0; i < pnv->num_chips; i++) {
+pnv_chip_foreach_cpu(pnv->chips[i], pnv_cpu_do_nmi, NULL);
 }
 }
 
-- 
2.44.0