On Thu, 10 Aug 2023, juzhe.zh...@rivai.ai wrote:

> Hi, Richi.
> 
> >> What inserts the required LC SSA PHI in that case?
> 
> Here is the flow how GCC insert LC SSA PHI flow for ARM SVE.
> You can see this following 'vect' dump details:
> https://godbolt.org/z/564o87oz3 
> 
> You can see this following information:
> 
> ;; Created LCSSA PHI: loop_mask_36 = PHI <loop_mask_22(3)>
>   # loop_mask_36 = PHI <loop_mask_22(3)>
>   _25 = .EXTRACT_LAST (loop_mask_36, vect_last_12.8_24);
>   last_17 = _25;
> 
> The '# loop_mask_36 = PHI <loop_mask_22(3)>' is inserted as follows:
> 
> Step 1 - Enter file tree-vectorizer.cc
> In the function pass_vectorize::execute (function *fun): 1358
> 'rewrite_into_loop_closed_ssa' is the key function insert the LC SSA PHI for 
> ARM SVE in mask loop case.
> 
> Step 2 - Investigate more into 'rewrite_into_loop_closed_ssa':
> In file tree-ssa-loop-manip.cc:628, 'rewrite_into_loop_closed_ssa' is 
> directly calling 'rewrite_into_loop_closed_ssa_1'.
> Step 3 - Investigate 'rewrite_into_loop_closed_ssa_1':
> In file tree-ssa-loop-manip.cc:588 which is the function 
> 'find_uses_to_rename' that:
> /* Marks names matching USE_FLAGS that are used outside of the loop they are
>    defined in for rewrite.  Records the set of blocks in which the ssa names 
> are
>    used to USE_BLOCKS.  Record the SSA names that will need exit PHIs in
>    NEED_PHIS.  If CHANGED_BBS is not NULL, scan only blocks in this set.  */
> 
> static void
> find_uses_to_rename (bitmap changed_bbs, bitmap *use_blocks, bitmap need_phis,
>          int use_flags)
> {
>   basic_block bb;
>   unsigned index;
>   bitmap_iterator bi;
> 
>   if (changed_bbs)
>     EXECUTE_IF_SET_IN_BITMAP (changed_bbs, 0, index, bi)
>       {
>   bb = BASIC_BLOCK_FOR_FN (cfun, index);
>   if (bb)
>     find_uses_to_rename_bb (bb, use_blocks, need_phis, use_flags);
>       }
>   else
>     FOR_EACH_BB_FN (bb, cfun)
>       find_uses_to_rename_bb (bb, use_blocks, need_phis, use_flags);
> }
> 
> This function is iterating all blocks of the function to set the BITMAP which 
> SSA need to be renamed then the later function will insert LC SSA for it.
> 
> In file tree-ssa-loop-manip.cc:606 which is the function 'add_exit_phis' that 
> is the real function that is adding LC SSA by calling
> this eventually:
> /* Add a loop-closing PHI for VAR in basic block EXIT.  */
> 
> static void
> add_exit_phi (basic_block exit, tree var)
> {
>   gphi *phi;
>   edge e;
>   edge_iterator ei;
> 
>   /* Check that at least one of the edges entering the EXIT block exits
>      the loop, or a superloop of that loop, that VAR is defined in.  */
>   if (flag_checking)
>     {
>       gimple *def_stmt = SSA_NAME_DEF_STMT (var);
>       basic_block def_bb = gimple_bb (def_stmt);
>       FOR_EACH_EDGE (e, ei, exit->preds)
>   {
>     class loop *aloop = find_common_loop (def_bb->loop_father,
>              e->src->loop_father);
>     if (!flow_bb_inside_loop_p (aloop, e->dest))
>       break;
>   }
>       gcc_assert (e);
>     }
> 
>   phi = create_phi_node (NULL_TREE, exit);
>   create_new_def_for (var, phi, gimple_phi_result_ptr (phi));
>   FOR_EACH_EDGE (e, ei, exit->preds)
>     add_phi_arg (phi, var, e, UNKNOWN_LOCATION);
> 
>   if (dump_file && (dump_flags & TDF_DETAILS))
>     {
>       fprintf (dump_file, ";; Created LCSSA PHI: ");
>       print_gimple_stmt (dump_file, phi, 0, dump_flags);
>     }
> }
> 
> 
> This is how it works for ARM SVE in EXTRACT_LAST. Such flow 
> (rewrite_into_loop_closed_ssa) can always insert LC SSA for RVV which is 
> using length loop.
> 
> However,
> 
> >> I want to know why we don't need this for SVE fully masked loops. 
> 
> Before entering 'rewrite_into_loop_closed_ssa', there is a check here that 
> RVV assertion failed but ARM SVE passed:
> 
>   /* We should not have to update virtual SSA form here but some
>      transforms involve creating new virtual definitions which makes
>      updating difficult.
>      We delay the actual update to the end of the pass but avoid
>      confusing ourselves by forcing need_ssa_update_p () to false.  */
>   unsigned todo = 0;
>   if (need_ssa_update_p (cfun))
>     {
>       gcc_assert (loop_vinfo->any_known_not_updated_vssa);
>       fun->gimple_df->ssa_renaming_needed = false;
>       todo |= TODO_update_ssa_only_virtuals;
>     }
> 
> in tree-vectorizer.cc, function 'vect_transform_loops'
> The assertion (gcc_assert (loop_vinfo->any_known_not_updated_vssa);)
> failed for RVV since it is false.
> 
> The reason why ARM SVE can pass is that the STMT1 before 
> 'vectorizable_live_operation' and STMT2 after vectorization of 
> 'vectorizable_live_operation'
> are both CONST or PURE since ARM SVE is using EXTRACT_LAST, here is the 
> define of 'EXTRACT_LAST' internal function:
> /* Extract the last active element from a vector.  */
> DEF_INTERNAL_OPTAB_FN (EXTRACT_LAST, ECF_CONST | ECF_NOTHROW,
>            extract_last, fold_left)
> 
> You can see 'EXTRACT_LAST' is ECF_CONST.
> 
> Wheras, RVV will fail since it is 'VEC_EXTRACT' which is not ECF_CONST:
> DEF_INTERNAL_OPTAB_FN (VEC_EXTRACT, 0, vec_extract, vec_extract)
> 
> When I changed VEC_EXTRACT into ECF_CONST, we don't need 
> 'vinfo->any_known_not_updated_vssa = true'
> The flow can perfectly work and no different from ARM SVE.
> 
> However, I found we can't make 'VEC_EXTRACT' as ECF_CONST since I found some 
> targets use VEC_EXTRACT, extract element into a memory.
> 
> So.... I use 'vinfo->any_known_not_updated_vssa = true'
> 
> The alternative approach I think is adding IFN_EXTRACT_LAST_LEN as Richard 
> said, and make IFN_EXTRACT_LAST_LEN as ECF_CONST, it can definitely work
> but such pattern is redundant since we can reuse 'VEC_EXTRACT' pattern which 
> is suitable for us.

Hmm, I think VEC_EXTRACT and VEC_SET should be ECF_CONST.  Maybe the 
GIMPLE ISEL
comments do not match the implementation, but then that should be fixed?

/* Expand all ARRAY_REF(VIEW_CONVERT_EXPR) gimple assignments into calls 
to
   internal function based on vector type of selected expansion.

   For vec_set:

     VIEW_CONVERT_EXPR<int[4]>(u)[_1] = i_4(D);
   =>
     _7 = u;
     _8 = .VEC_SET (_7, i_4(D), _1);
     u = _8;
  
   For vec_extract:

      _3 = VIEW_CONVERT_EXPR<intD.1[4]>(vD.2208)[idx_2(D)];
   =>
      _4 = vD.2208;
      _3 = .VEC_EXTRACT (_4, idx_2(D));  */

Richard.

> Thanks.
> 
> 
> juzhe.zh...@rivai.ai
>  
> From: Richard Biener
> Date: 2023-08-10 19:09
> To: juzhe.zh...@rivai.ai
> CC: gcc-patches; richard.sandiford
> Subject: Re: Re: [PATCH V2] VECT: Support loop len control on EXTRACT_LAST 
> vectorization
> On Thu, 10 Aug 2023, juzhe.zh...@rivai.ai wrote:
>  
> > >> I guess as a temporary thing your approach is OK but we shouldn't
> > >> add these as part of new code - it's supposed to handle legacy
> > >> cases that we didn't fixup yet.
> > 
> > Do you mean we need to fix LC SSA PHI flow so that we don't need to 
> > set vinfo->any_known_not_updated_vssa = true ?
> > 
> > After it's fixed then this patch with removing 
> > 'vinfo->any_known_not_updated_vssa = true' is ok for trunk, am I right?
>  
> I want to know why we don't need this for SVE fully masked loops.  What
> inserts the required LC SSA PHI in that case?
>  
> > Thanks.
> > 
> > 
> > juzhe.zh...@rivai.ai
> >  
> > From: Richard Biener
> > Date: 2023-08-10 15:58
> > To: Ju-Zhe Zhong
> > CC: gcc-patches; richard.sandiford
> > Subject: Re: [PATCH V2] VECT: Support loop len control on EXTRACT_LAST 
> > vectorization
> > On Thu, 10 Aug 2023, juzhe.zh...@rivai.ai wrote:
> >  
> > > From: Ju-Zhe Zhong <juzhe.zh...@rivai.ai>
> > > 
> > > Hi, Richard and Richi.
> > > 
> > > This patch add support live vectorization by VEC_EXTRACT for LEN loop 
> > > control.
> > > 
> > > Consider this following case:
> > > 
> > > #include <stdint.h>
> > > 
> > > #define EXTRACT_LAST(TYPE) \
> > >   TYPE __attribute__ ((noinline, noclone)) \
> > >   test_##TYPE (TYPE *x, int n, TYPE value) \
> > >   { \
> > >     TYPE last; \
> > >     for (int j = 0; j < n; ++j) \
> > >       { \
> > > last = x[j]; \
> > > x[j] = last * value; \
> > >       } \
> > >     return last; \
> > >   }
> > > 
> > > #define TEST_ALL(T) \
> > >   T (uint8_t) \
> > > 
> > > TEST_ALL (EXTRACT_LAST)
> > > 
> > > ARM SVE IR:
> > > 
> > > Preheader:
> > >   max_mask_34 = .WHILE_ULT (0, bnd.5_6, { 0, ... });
> > > 
> > > Loop:
> > >   ...
> > >   # loop_mask_22 = PHI <next_mask_35(4), max_mask_34(3)>
> > >   ...
> > >   vect_last_12.8_23 = .MASK_LOAD (_7, 8B, loop_mask_22);
> > >   vect__4.9_27 = vect_last_12.8_23 * vect_cst__26;
> > >   .MASK_STORE (_7, 8B, loop_mask_22, vect__4.9_27);
> > >   ...
> > >   next_mask_35 = .WHILE_ULT (_1, bnd.5_6, { 0, ... });
> > >   ...
> > > 
> > > Epilogue:
> > >   _25 = .EXTRACT_LAST (loop_mask_22, vect_last_12.8_23);
> > > 
> > > For RVV since we prefer len in loop control, after this patch for RVV:
> > > 
> > > Loop:
> > >   ...
> > >   loop_len_22 = SELECT_VL;
> > >   vect_last_12.8_23 = .MASK_LOAD (_7, 8B, loop_len_22);
> > >   vect__4.9_27 = vect_last_12.8_23 * vect_cst__26;
> > >   .MASK_STORE (_7, 8B, loop_len_22, vect__4.9_27);
> > >   ...
> > > 
> > > Epilogue:
> > >   _25 = .VEC_EXTRACT (loop_len_22 - 1 - bias, vect_last_12.8_23);
> > > 
> > > Details of this approach:
> > > 
> > > 1. Step 1 - Add 'vect_can_vectorize_extract_last_with_len_p'  to enable 
> > > live vectorization
> > >             for LEN loop control.
> > >    
> > >    This function we check whether target support:
> > >     - Use LEN as the loop control.
> > >     - Support VEC_EXTRACT optab.
> > > 
> > > 2. Step 2 - Record LEN for loop control if 
> > > 'vect_can_vectorize_extract_last_with_len_p' is true.
> > > 
> > > 3. Step 3 - Gerenate VEC_EXTRACT (v, LEN - 1 - BIAS).
> > > 
> > > NOTE: This patch set 'vinfo->any_known_not_updated_vssa = true;' since 
> > > the original STMT is a simple
> > >       assignment wheras VEC_EXTRACT is neither pure nor const function 
> > > according to internal-fn.def:
> > > 
> > >       DEF_INTERNAL_OPTAB_FN (VEC_EXTRACT, 0, vec_extract, vec_extract)
> > > 
> > >       If we don't set 'vinfo->any_known_not_updated_vssa' as true, it 
> > > will cause ICE in:
> > > 
> > >         if (need_ssa_update_p (cfun))
> > >           {
> > >             gcc_assert (loop_vinfo->any_known_not_updated_vssa);  ----> 
> > > Report assertion fail here.
> > >             fun->gimple_df->ssa_renaming_needed = false;
> > >             todo |= TODO_update_ssa_only_virtuals;
> > >           }
> > >       
> > >       I saw there are 2 places set 'vinfo->any_known_not_updated_vssa' as 
> > > true:
> > >         
> > > - The one is in 'vectorizable_simd_clone_call':
> > > 
> > > /* When the original call is pure or const but the SIMD ABI dictates
> > > an aggregate return we will have to use a virtual definition and
> > > in a loop eventually even need to add a virtual PHI.  That's
> > > not straight-forward so allow to fix this up via renaming.  */
> > >       if (gimple_call_lhs (stmt)
> > >   && !gimple_vdef (stmt)
> > >   && TREE_CODE (TREE_TYPE (TREE_TYPE (bestn->decl))) == ARRAY_TYPE)
> > > vinfo->any_known_not_updated_vssa = true;
> > >        
> > >        - The other is in 'vectorizable_load':
> > >        
> > >         if (memory_access_type == VMAT_LOAD_STORE_LANES)
> > >   vinfo->any_known_not_updated_vssa = true;
> > > 
> > >       It seems that they are the same reason as me doing in 
> > > 'vectorizable_live_operation'.
> > >       Feel free to correct me if I am wrong.
> >  
> > You should always manually update things.  Did you verify the mask
> > case is handled by this?
> >  
> > There's the odd
> >  
> >       if (stmts)
> >         {
> >           gimple_stmt_iterator exit_gsi = gsi_after_labels (exit_bb);
> >           gsi_insert_seq_before (&exit_gsi, stmts, GSI_SAME_STMT);
> >  
> >           /* Remove existing phi from lhs and create one copy from 
> > new_tree.  */
> >           tree lhs_phi = NULL_TREE;
> >           gimple_stmt_iterator gsi;
> >           for (gsi = gsi_start_phis (exit_bb);
> >                !gsi_end_p (gsi); gsi_next (&gsi))
> >             {
> >               gimple *phi = gsi_stmt (gsi);
> >               if ((gimple_phi_arg_def (phi, 0) == lhs))
> >                 {
> >                   remove_phi_node (&gsi, false);
> >                   lhs_phi = gimple_phi_result (phi);
> >                   gimple *copy = gimple_build_assign (lhs_phi, new_tree);
> >                   gsi_insert_before (&exit_gsi, copy, GSI_SAME_STMT);
> >                   break;
> >                 }
> >             }
> >  
> > code but I don't think it will create new LC PHIs for the mask, instead
> > it will break LC SSA as well by removing a PHI?
> >  
> > I guess as a temporary thing your approach is OK but we shouldn't
> > add these as part of new code - it's supposed to handle legacy
> > cases that we didn't fixup yet.
> >  
> > Richard.
> >  
> >  
> >  
> > >       Bootstrap and Regression on X86 passed.
> > > 
> > > gcc/ChangeLog:
> > > 
> > >         * tree-vect-loop.cc (vect_can_vectorize_extract_last_with_len_p): 
> > > New function.
> > >         (vectorizable_live_operation): Add loop LEN control.
> > > 
> > > ---
> > >  gcc/tree-vect-loop.cc | 74 +++++++++++++++++++++++++++++++++++++++----
> > >  1 file changed, 68 insertions(+), 6 deletions(-)
> > > 
> > > diff --git a/gcc/tree-vect-loop.cc b/gcc/tree-vect-loop.cc
> > > index 00058c3c13e..208918f53fb 100644
> > > --- a/gcc/tree-vect-loop.cc
> > > +++ b/gcc/tree-vect-loop.cc
> > > @@ -8964,6 +8964,24 @@ vect_can_vectorize_without_simd_p (code_helper 
> > > code)
> > >    && vect_can_vectorize_without_simd_p (tree_code (code)));
> > >  }
> > >  
> > > +/* Return true if target supports extract last vectorization with LEN.  
> > > */
> > > +
> > > +static bool
> > > +vect_can_vectorize_extract_last_with_len_p (tree vectype)
> > > +{
> > > +  /* Return false if target doesn't support LEN in loop control.  */
> > > +  machine_mode vmode;
> > > +  if (!get_len_load_store_mode (TYPE_MODE (vectype), true).exists 
> > > (&vmode)
> > > +      || !get_len_load_store_mode (TYPE_MODE (vectype), false).exists 
> > > (&vmode))
> > > +    return false;
> > > +
> > > +  /* Target need to support VEC_EXTRACT to extract the last active 
> > > element.  */
> > > +  return convert_optab_handler (vec_extract_optab,
> > > + TYPE_MODE (vectype),
> > > + TYPE_MODE (TREE_TYPE (vectype)))
> > > + != CODE_FOR_nothing;
> > > +}
> > > +
> > >  /* Create vector init for vectorized iv.  */
> > >  static tree
> > >  vect_create_nonlinear_iv_init (gimple_seq* stmts, tree init_expr,
> > > @@ -10282,7 +10300,8 @@ vectorizable_live_operation (vec_info *vinfo,
> > >        if (loop_vinfo && LOOP_VINFO_CAN_USE_PARTIAL_VECTORS_P 
> > > (loop_vinfo))
> > >  {
> > >    if (!direct_internal_fn_supported_p (IFN_EXTRACT_LAST, vectype,
> > > -        OPTIMIZE_FOR_SPEED))
> > > +        OPTIMIZE_FOR_SPEED)
> > > +       && !vect_can_vectorize_extract_last_with_len_p (vectype))
> > >      {
> > >        if (dump_enabled_p ())
> > >  dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
> > > @@ -10311,9 +10330,14 @@ vectorizable_live_operation (vec_info *vinfo,
> > >    else
> > >      {
> > >        gcc_assert (ncopies == 1 && !slp_node);
> > > -       vect_record_loop_mask (loop_vinfo,
> > > -      &LOOP_VINFO_MASKS (loop_vinfo),
> > > -      1, vectype, NULL);
> > > +       if (vect_can_vectorize_extract_last_with_len_p (vectype))
> > > + vect_record_loop_len (loop_vinfo,
> > > +       &LOOP_VINFO_LENS (loop_vinfo),
> > > +       1, vectype, 1);
> > > +       else
> > > + vect_record_loop_mask (loop_vinfo,
> > > +        &LOOP_VINFO_MASKS (loop_vinfo),
> > > +        1, vectype, NULL);
> > >      }
> > >  }
> > >        /* ???  Enable for loop costing as well.  */
> > > @@ -10339,7 +10363,9 @@ vectorizable_live_operation (vec_info *vinfo,
> > >    gimple *vec_stmt;
> > >    if (slp_node)
> > >      {
> > > -      gcc_assert (!loop_vinfo || !LOOP_VINFO_FULLY_MASKED_P 
> > > (loop_vinfo));
> > > +      gcc_assert (!loop_vinfo
> > > +   || (!LOOP_VINFO_FULLY_MASKED_P (loop_vinfo)
> > > +       && !LOOP_VINFO_FULLY_WITH_LENGTH_P (loop_vinfo)));
> > >  
> > >        /* Get the correct slp vectorized stmt.  */
> > >        vec_lhs = SLP_TREE_VEC_DEFS (slp_node)[vec_entry];
> > > @@ -10383,7 +10409,43 @@ vectorizable_live_operation (vec_info *vinfo,
> > >  
> > >        gimple_seq stmts = NULL;
> > >        tree new_tree;
> > > -      if (LOOP_VINFO_FULLY_MASKED_P (loop_vinfo))
> > > +      if (LOOP_VINFO_FULLY_WITH_LENGTH_P (loop_vinfo))
> > > + {
> > > +   /* Emit:
> > > +
> > > +        SCALAR_RES = VEC_EXTRACT <VEC_LHS, LEN - BIAS - 1>
> > > +
> > > +      where VEC_LHS is the vectorized live-out result and MASK is
> > > +      the loop mask for the final iteration.  */
> > > +   gcc_assert (ncopies == 1 && !slp_node);
> > > +   tree scalar_type = TREE_TYPE (STMT_VINFO_VECTYPE (stmt_info));
> > > +   tree len
> > > +     = vect_get_loop_len (loop_vinfo, gsi, &LOOP_VINFO_LENS (loop_vinfo),
> > > + 1, vectype, 0, 0);
> > > +
> > > +   /* BIAS + 1.  */
> > > +   signed char biasval = LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS (loop_vinfo);
> > > +   tree bias_one
> > > +     = size_binop (PLUS_EXPR, build_int_cst (TREE_TYPE (len), biasval),
> > > +   build_one_cst (TREE_TYPE (len)));
> > > +
> > > +   /* LAST_INDEX = LEN - (BIAS + 1).  */
> > > +   tree last_index
> > > +     = gimple_build (&stmts, MINUS_EXPR, TREE_TYPE (len), len, bias_one);
> > > +
> > > +   tree scalar_res = gimple_build (&stmts, CFN_VEC_EXTRACT, scalar_type,
> > > +   vec_lhs_phi, last_index);
> > > +
> > > +   /* Convert the extracted vector element to the scalar type.  */
> > > +   new_tree = gimple_convert (&stmts, lhs_type, scalar_res);
> > > +   /* When the original stmt is an assignment but VEC_EXTRACT is not pure
> > > +      or const since it may return a memory result.  We will have to use
> > > +      a virtual definition and in a loop eventually even need to add a
> > > +      virtual PHI. That's not straight-forward so allow to fix this up
> > > +      via renaming.  */
> > > +   vinfo->any_known_not_updated_vssa = true;
> > > + }
> > > +      else if (LOOP_VINFO_FULLY_MASKED_P (loop_vinfo))
> > >  {
> > >    /* Emit:
> > >  
> > > 
> >  
> > 
>  
> 

-- 
Richard Biener <rguent...@suse.de>
SUSE Software Solutions Germany GmbH,
Frankenstrasse 146, 90461 Nuernberg, Germany;
GF: Ivo Totev, Andrew McDonald, Werner Knoblich; (HRB 36809, AG Nuernberg)

Reply via email to