gcc/
        * explow.c (force_reg): Strengthen local "insn" from rtx to
        rtx_insn *.
        (adjust_stack_1): Likewise.
        (allocate_dynamic_stack_space): Likewise.  Strengthen locals
        "final_label", "available_label", "space_available" from rtx to
        rtx_code_label *.
        (probe_stack_range): Likewise for locals "loop_lab", "end_lab".
        (anti_adjust_stack_and_probe): Likewise.
---
 gcc/explow.c | 30 +++++++++++++++++-------------
 1 file changed, 17 insertions(+), 13 deletions(-)

diff --git a/gcc/explow.c b/gcc/explow.c
index e39db05..88beb3b 100644
--- a/gcc/explow.c
+++ b/gcc/explow.c
@@ -649,7 +649,8 @@ copy_to_mode_reg (enum machine_mode mode, rtx x)
 rtx
 force_reg (enum machine_mode mode, rtx x)
 {
-  rtx temp, insn, set;
+  rtx temp, set;
+  rtx_insn *insn;
 
   if (REG_P (x))
     return x;
@@ -880,7 +881,8 @@ static bool suppress_reg_args_size;
 static void
 adjust_stack_1 (rtx adjust, bool anti_p)
 {
-  rtx temp, insn;
+  rtx temp;
+  rtx_insn *insn;
 
 #ifndef STACK_GROWS_DOWNWARD
   /* Hereafter anti_p means subtract_p.  */
@@ -1160,7 +1162,8 @@ allocate_dynamic_stack_space (rtx size, unsigned 
size_align,
                              unsigned required_align, bool cannot_accumulate)
 {
   HOST_WIDE_INT stack_usage_size = -1;
-  rtx final_label, final_target, target;
+  rtx_code_label *final_label;
+  rtx final_target, target;
   unsigned extra_align = 0;
   bool must_align;
 
@@ -1184,7 +1187,8 @@ allocate_dynamic_stack_space (rtx size, unsigned 
size_align,
         {
          /* Look into the last emitted insn and see if we can deduce
             something for the register.  */
-         rtx insn, set, note;
+         rtx_insn *insn;
+         rtx set, note;
          insn = get_last_insn ();
          if ((set = single_set (insn)) && rtx_equal_p (SET_DEST (set), size))
            {
@@ -1313,7 +1317,7 @@ allocate_dynamic_stack_space (rtx size, unsigned 
size_align,
        current_function_has_unbounded_dynamic_stack_size = 1;
     }
 
-  final_label = NULL_RTX;
+  final_label = NULL;
   final_target = NULL_RTX;
 
   /* If we are splitting the stack, we need to ask the backend whether
@@ -1325,9 +1329,10 @@ allocate_dynamic_stack_space (rtx size, unsigned 
size_align,
      least it doesn't cause a stack overflow.  */
   if (flag_split_stack)
     {
-      rtx available_label, ask, space, func;
+      rtx_code_label *available_label;
+      rtx ask, space, func;
 
-      available_label = NULL_RTX;
+      available_label = NULL;
 
 #ifdef HAVE_split_stack_space_check
       if (HAVE_split_stack_space_check)
@@ -1420,7 +1425,7 @@ allocate_dynamic_stack_space (rtx size, unsigned 
size_align,
       if (crtl->limit_stack)
        {
          rtx available;
-         rtx space_available = gen_label_rtx ();
+         rtx_code_label *space_available = gen_label_rtx ();
 #ifdef STACK_GROWS_DOWNWARD
          available = expand_binop (Pmode, sub_optab,
                                    stack_pointer_rtx, stack_limit_rtx,
@@ -1622,9 +1627,8 @@ probe_stack_range (HOST_WIDE_INT first, rtx size)
   else
     {
       rtx rounded_size, rounded_size_op, test_addr, last_addr, temp;
-      rtx loop_lab = gen_label_rtx ();
-      rtx end_lab = gen_label_rtx ();
-
+      rtx_code_label *loop_lab = gen_label_rtx ();
+      rtx_code_label *end_lab = gen_label_rtx ();
 
       /* Step 1: round SIZE to the previous multiple of the interval.  */
 
@@ -1770,8 +1774,8 @@ anti_adjust_stack_and_probe (rtx size, bool adjust_back)
   else
     {
       rtx rounded_size, rounded_size_op, last_addr, temp;
-      rtx loop_lab = gen_label_rtx ();
-      rtx end_lab = gen_label_rtx ();
+      rtx_code_label *loop_lab = gen_label_rtx ();
+      rtx_code_label *end_lab = gen_label_rtx ();
 
 
       /* Step 1: round SIZE to the previous multiple of the interval.  */
-- 
1.8.5.3

Reply via email to