Re: [Qemu-devel] [RFC for-4.1 17/25] target/ppc: Style fixes for mmu-hash64.[ch]

2019-03-25 Thread Cédric Le Goater
On 3/22/19 1:15 AM, David Gibson wrote:
> Signed-off-by: David Gibson 


Reviewed-by: Cédric Le Goater 

Thanks,

C.

> ---
>  target/ppc/mmu-hash64.c | 62 +
>  1 file changed, 38 insertions(+), 24 deletions(-)
> 
> diff --git a/target/ppc/mmu-hash64.c b/target/ppc/mmu-hash64.c
> index a2b1ec5040..90f4b306b2 100644
> --- a/target/ppc/mmu-hash64.c
> +++ b/target/ppc/mmu-hash64.c
> @@ -29,7 +29,7 @@
>  #include "hw/hw.h"
>  #include "mmu-book3s-v3.h"
>  
> -//#define DEBUG_SLB
> +/* #define DEBUG_SLB */
>  
>  #ifdef DEBUG_SLB
>  #  define LOG_SLB(...) qemu_log_mask(CPU_LOG_MMU, __VA_ARGS__)
> @@ -57,9 +57,11 @@ static ppc_slb_t *slb_lookup(PowerPCCPU *cpu, target_ulong 
> eaddr)
>  
>  LOG_SLB("%s: slot %d %016" PRIx64 " %016"
>  PRIx64 "\n", __func__, n, slb->esid, slb->vsid);
> -/* We check for 1T matches on all MMUs here - if the MMU
> +/*
> + * We check for 1T matches on all MMUs here - if the MMU
>   * doesn't have 1T segment support, we will have prevented 1T
> - * entries from being inserted in the slbmte code. */
> + * entries from being inserted in the slbmte code.
> + */
>  if (((slb->esid == esid_256M) &&
>   ((slb->vsid & SLB_VSID_B) == SLB_VSID_B_256M))
>  || ((slb->esid == esid_1T) &&
> @@ -102,7 +104,8 @@ void helper_slbia(CPUPPCState *env)
>  
>  if (slb->esid & SLB_ESID_V) {
>  slb->esid &= ~SLB_ESID_V;
> -/* XXX: given the fact that segment size is 256 MB or 1TB,
> +/*
> + * XXX: given the fact that segment size is 256 MB or 1TB,
>   *  and we still don't have a tlb_flush_mask(env, n, mask)
>   *  in QEMU, we just invalidate all TLBs
>   */
> @@ -125,7 +128,8 @@ static void __helper_slbie(CPUPPCState *env, target_ulong 
> addr,
>  if (slb->esid & SLB_ESID_V) {
>  slb->esid &= ~SLB_ESID_V;
>  
> -/* XXX: given the fact that segment size is 256 MB or 1TB,
> +/*
> + * XXX: given the fact that segment size is 256 MB or 1TB,
>   *  and we still don't have a tlb_flush_mask(env, n, mask)
>   *  in QEMU, we just invalidate all TLBs
>   */
> @@ -305,8 +309,10 @@ static int ppc_hash64_pte_prot(PowerPCCPU *cpu,
>  {
>  CPUPPCState *env = >env;
>  unsigned pp, key;
> -/* Some pp bit combinations have undefined behaviour, so default
> - * to no access in those cases */
> +/*
> + * Some pp bit combinations have undefined behaviour, so default
> + * to no access in those cases
> + */
>  int prot = 0;
>  
>  key = !!(msr_pr ? (slb->vsid & SLB_VSID_KP)
> @@ -375,7 +381,7 @@ static int ppc_hash64_amr_prot(PowerPCCPU *cpu, 
> ppc_hash_pte64_t pte)
>  }
>  
>  key = HPTE64_R_KEY(pte.pte1);
> -amrbits = (env->spr[SPR_AMR] >> 2*(31 - key)) & 0x3;
> +amrbits = (env->spr[SPR_AMR] >> 2 * (31 - key)) & 0x3;
>  
>  /* fprintf(stderr, "AMR protection: key=%d AMR=0x%" PRIx64 "\n", key, */
>  /* env->spr[SPR_AMR]); */
> @@ -546,8 +552,9 @@ static hwaddr ppc_hash64_pteg_search(PowerPCCPU *cpu, 
> hwaddr hash,
>  if (*pshift == 0) {
>  continue;
>  }
> -/* We don't do anything with pshift yet as qemu TLB only deals
> - * with 4K pages anyway
> +/*
> + * We don't do anything with pshift yet as qemu TLB only
> + * deals with 4K pages anyway
>   */
>  pte->pte0 = pte0;
>  pte->pte1 = pte1;
> @@ -571,8 +578,10 @@ static hwaddr ppc_hash64_htab_lookup(PowerPCCPU *cpu,
>  uint64_t vsid, epnmask, epn, ptem;
>  const PPCHash64SegmentPageSizes *sps = slb->sps;
>  
> -/* The SLB store path should prevent any bad page size encodings
> - * getting in there, so: */
> +/*
> + * The SLB store path should prevent any bad page size encodings
> + * getting in there, so:
> + */
>  assert(sps);
>  
>  /* If ISL is set in LPCR we need to clamp the page size to 4K */
> @@ -731,11 +740,12 @@ int ppc_hash64_handle_mmu_fault(PowerPCCPU *cpu, vaddr 
> eaddr,
>  
>  assert((rwx == 0) || (rwx == 1) || (rwx == 2));
>  
> -/* Note on LPCR usage: 970 uses HID4, but our special variant
> - * of store_spr copies relevant fields into env->spr[SPR_LPCR].
> - * Similarily we filter unimplemented bits when storing into
> - * LPCR depending on the MMU version. This code can thus just
> - * use the LPCR "as-is".
> +/*
> + * Note on LPCR usage: 970 uses HID4, but our special variant of
> + * store_spr copies relevant fields into env->spr[SPR_LPCR].
> + * Similarily we filter unimplemented bits when storing into LPCR
> + * depending on the MMU version. This code can thus just use the
> + * LPCR "as-is".
>   */
>  
>  /* 1. Handle real mode 

[Qemu-devel] [RFC for-4.1 17/25] target/ppc: Style fixes for mmu-hash64.[ch]

2019-03-21 Thread David Gibson
Signed-off-by: David Gibson 
---
 target/ppc/mmu-hash64.c | 62 +
 1 file changed, 38 insertions(+), 24 deletions(-)

diff --git a/target/ppc/mmu-hash64.c b/target/ppc/mmu-hash64.c
index a2b1ec5040..90f4b306b2 100644
--- a/target/ppc/mmu-hash64.c
+++ b/target/ppc/mmu-hash64.c
@@ -29,7 +29,7 @@
 #include "hw/hw.h"
 #include "mmu-book3s-v3.h"
 
-//#define DEBUG_SLB
+/* #define DEBUG_SLB */
 
 #ifdef DEBUG_SLB
 #  define LOG_SLB(...) qemu_log_mask(CPU_LOG_MMU, __VA_ARGS__)
@@ -57,9 +57,11 @@ static ppc_slb_t *slb_lookup(PowerPCCPU *cpu, target_ulong 
eaddr)
 
 LOG_SLB("%s: slot %d %016" PRIx64 " %016"
 PRIx64 "\n", __func__, n, slb->esid, slb->vsid);
-/* We check for 1T matches on all MMUs here - if the MMU
+/*
+ * We check for 1T matches on all MMUs here - if the MMU
  * doesn't have 1T segment support, we will have prevented 1T
- * entries from being inserted in the slbmte code. */
+ * entries from being inserted in the slbmte code.
+ */
 if (((slb->esid == esid_256M) &&
  ((slb->vsid & SLB_VSID_B) == SLB_VSID_B_256M))
 || ((slb->esid == esid_1T) &&
@@ -102,7 +104,8 @@ void helper_slbia(CPUPPCState *env)
 
 if (slb->esid & SLB_ESID_V) {
 slb->esid &= ~SLB_ESID_V;
-/* XXX: given the fact that segment size is 256 MB or 1TB,
+/*
+ * XXX: given the fact that segment size is 256 MB or 1TB,
  *  and we still don't have a tlb_flush_mask(env, n, mask)
  *  in QEMU, we just invalidate all TLBs
  */
@@ -125,7 +128,8 @@ static void __helper_slbie(CPUPPCState *env, target_ulong 
addr,
 if (slb->esid & SLB_ESID_V) {
 slb->esid &= ~SLB_ESID_V;
 
-/* XXX: given the fact that segment size is 256 MB or 1TB,
+/*
+ * XXX: given the fact that segment size is 256 MB or 1TB,
  *  and we still don't have a tlb_flush_mask(env, n, mask)
  *  in QEMU, we just invalidate all TLBs
  */
@@ -305,8 +309,10 @@ static int ppc_hash64_pte_prot(PowerPCCPU *cpu,
 {
 CPUPPCState *env = >env;
 unsigned pp, key;
-/* Some pp bit combinations have undefined behaviour, so default
- * to no access in those cases */
+/*
+ * Some pp bit combinations have undefined behaviour, so default
+ * to no access in those cases
+ */
 int prot = 0;
 
 key = !!(msr_pr ? (slb->vsid & SLB_VSID_KP)
@@ -375,7 +381,7 @@ static int ppc_hash64_amr_prot(PowerPCCPU *cpu, 
ppc_hash_pte64_t pte)
 }
 
 key = HPTE64_R_KEY(pte.pte1);
-amrbits = (env->spr[SPR_AMR] >> 2*(31 - key)) & 0x3;
+amrbits = (env->spr[SPR_AMR] >> 2 * (31 - key)) & 0x3;
 
 /* fprintf(stderr, "AMR protection: key=%d AMR=0x%" PRIx64 "\n", key, */
 /* env->spr[SPR_AMR]); */
@@ -546,8 +552,9 @@ static hwaddr ppc_hash64_pteg_search(PowerPCCPU *cpu, 
hwaddr hash,
 if (*pshift == 0) {
 continue;
 }
-/* We don't do anything with pshift yet as qemu TLB only deals
- * with 4K pages anyway
+/*
+ * We don't do anything with pshift yet as qemu TLB only
+ * deals with 4K pages anyway
  */
 pte->pte0 = pte0;
 pte->pte1 = pte1;
@@ -571,8 +578,10 @@ static hwaddr ppc_hash64_htab_lookup(PowerPCCPU *cpu,
 uint64_t vsid, epnmask, epn, ptem;
 const PPCHash64SegmentPageSizes *sps = slb->sps;
 
-/* The SLB store path should prevent any bad page size encodings
- * getting in there, so: */
+/*
+ * The SLB store path should prevent any bad page size encodings
+ * getting in there, so:
+ */
 assert(sps);
 
 /* If ISL is set in LPCR we need to clamp the page size to 4K */
@@ -731,11 +740,12 @@ int ppc_hash64_handle_mmu_fault(PowerPCCPU *cpu, vaddr 
eaddr,
 
 assert((rwx == 0) || (rwx == 1) || (rwx == 2));
 
-/* Note on LPCR usage: 970 uses HID4, but our special variant
- * of store_spr copies relevant fields into env->spr[SPR_LPCR].
- * Similarily we filter unimplemented bits when storing into
- * LPCR depending on the MMU version. This code can thus just
- * use the LPCR "as-is".
+/*
+ * Note on LPCR usage: 970 uses HID4, but our special variant of
+ * store_spr copies relevant fields into env->spr[SPR_LPCR].
+ * Similarily we filter unimplemented bits when storing into LPCR
+ * depending on the MMU version. This code can thus just use the
+ * LPCR "as-is".
  */
 
 /* 1. Handle real mode accesses */
@@ -874,8 +884,10 @@ skip_slb_search:
 if (rwx == 1) {
 new_pte1 |= HPTE64_R_C; /* set changed (dirty) bit */
 } else {
-/* Treat the page as read-only for now, so that a later write
- * will pass through this function again to set the C bit */
+/*
+ * Treat the page as