diff options
-rw-r--r-- | gcc/ChangeLog | 35 | ||||
-rw-r--r-- | gcc/lra-assigns.c | 47 | ||||
-rw-r--r-- | gcc/lra-constraints.c | 423 | ||||
-rw-r--r-- | gcc/lra-int.h | 7 | ||||
-rw-r--r-- | gcc/lra-remat.c | 86 | ||||
-rw-r--r-- | gcc/lra-spills.c | 9 | ||||
-rw-r--r-- | gcc/lra.c | 88 |
7 files changed, 533 insertions, 162 deletions
diff --git a/gcc/ChangeLog b/gcc/ChangeLog index a995aa75bea..a5fe1d5cf31 100644 --- a/gcc/ChangeLog +++ b/gcc/ChangeLog @@ -1,3 +1,38 @@ +2016-08-02 Vladimir Makarov <vmakarov@redhat.com> + + PR rtl-optimization/69847 + * lra-int.h (struct lra-reg): Use restore_rtx instead of + restore_regno. + (lra_rtx_hash): New. + * lra.c (initialize_lra_reg_info_element): Use restore_rtx instead + of restore_regno. + (lra_rtx_hash): Rename and move lra-remat.c::rtx_hash. + * lra-remat.c (rtx_hash): Rename and Move to lra.c. + * lra-spills.c (lra_final_code_change): Don't delete insn when the + next insn is USE with the same reg as the current insn source. + * lra-constraints.c (curr_insn_transform): Use restore_rtx instead + of restore_regno. + (lra_constraints_init): Call initiate_invariants. + (lra_constraints_finish): Call finish_invariants. + (struct invariant, invariant_t, invariant_ptr_t): New. + (const_invariant_ptr_t, invariants, invariants_pool): New. + (invariant_table, invariant_hash, invariant_eq_p): New. + (insert_invariant, initiate_invariants, finish_invariants): New. + (clear_invariants, invalid_invariant_regs): New. + (inherit_reload_reg, split_reg, fix_bb_live_info): Use restore_rtx + instead of restore_regno. + (invariant_p, process_invariant_for_inheritance): New. + (inherit_in_ebb): Implement invariant inheritance. + (lra_inheritance): Initialize and finalize invalid_invariant_regs. + (remove_inheritance_pseudos): Implement undoing invariant + inheritance. + (undo_optional_reloads, lra_undo_inheritance): Use restore_rtx + instead of restore_regno. + * lra-assigns.c (regno_live_length): New. + (reload_pseudo_compare_func): Use regno_live_length. + (assign_by_spills): Use restore_rtx instead of restore_regno. + (lra_assign): Ditto. Initiate regno_live_length. + 2016-02-08 James Greenhalgh <james.greenhalgh@arm.com> * config/aarch64/arm_neon.h (vminnm_f64): Add back missing 'f' from diff --git a/gcc/lra-assigns.c b/gcc/lra-assigns.c index fb3de849ec1..7248f89869e 100644 --- a/gcc/lra-assigns.c +++ b/gcc/lra-assigns.c @@ -107,6 +107,10 @@ static bool former_reload_pseudo_spill_p; lra_get_allocno_class. It is used to speed up the code. */ static enum reg_class *regno_allocno_class_array; +/* Array containing lengths of pseudo live ranges. It is used to + speed up the code. */ +static int *regno_live_length; + /* Information about the thread to which a pseudo belongs. Threads are a set of connected reload and inheritance pseudos with the same set of available hard registers. Lone registers belong to their own threads. */ @@ -227,6 +231,11 @@ reload_pseudo_compare_func (const void *v1p, const void *v2p) /* Put pseudos from the thread nearby. */ if ((diff = regno_assign_info[r1].first - regno_assign_info[r2].first) != 0) return diff; + /* Prefer pseudos with longer live ranges. It sets up better + prefered hard registers for the thread pseudos and decreases + register-register moves between the thread pseudos. */ + if ((diff = regno_live_length[r2] - regno_live_length[r1]) != 0) + return diff; /* If regs are equally good, sort by their numbers, so that the results of qsort leave nothing to chance. */ return r1 - r2; @@ -1300,7 +1309,8 @@ find_all_spills_for (int regno) static void assign_by_spills (void) { - int i, n, nfails, iter, regno, hard_regno, cost, restore_regno; + int i, n, nfails, iter, regno, hard_regno, cost; + rtx restore_rtx; rtx_insn *insn; bitmap_head changed_insns, do_not_assign_nonreload_pseudos; unsigned int u, conflict_regno; @@ -1333,6 +1343,8 @@ assign_by_spills (void) for (i = 0; i < n; i++) { regno = sorted_pseudos[i]; + if (reg_renumber[regno] >= 0) + continue; if (lra_dump_file != NULL) fprintf (lra_dump_file, " Assigning to %d " "(cl=%s, orig=%d, freq=%d, tfirst=%d, tfreq=%d)...\n", @@ -1471,21 +1483,25 @@ assign_by_spills (void) such inheritance or split pseudos. */ bitmap_initialize (&do_not_assign_nonreload_pseudos, ®_obstack); EXECUTE_IF_SET_IN_BITMAP (&lra_inheritance_pseudos, 0, u, bi) - if ((restore_regno = lra_reg_info[u].restore_regno) >= 0 + if ((restore_rtx = lra_reg_info[u].restore_rtx) != NULL_RTX + && REG_P (restore_rtx) && reg_renumber[u] < 0 && bitmap_bit_p (&lra_inheritance_pseudos, u)) - bitmap_set_bit (&do_not_assign_nonreload_pseudos, restore_regno); + bitmap_set_bit (&do_not_assign_nonreload_pseudos, REGNO (restore_rtx)); EXECUTE_IF_SET_IN_BITMAP (&lra_split_regs, 0, u, bi) - if ((restore_regno = lra_reg_info[u].restore_regno) >= 0 + if ((restore_rtx = lra_reg_info[u].restore_rtx) != NULL_RTX && reg_renumber[u] >= 0) - bitmap_set_bit (&do_not_assign_nonreload_pseudos, restore_regno); + { + lra_assert (REG_P (restore_rtx)); + bitmap_set_bit (&do_not_assign_nonreload_pseudos, REGNO (restore_rtx)); + } for (n = 0, i = FIRST_PSEUDO_REGISTER; i < max_regno; i++) if (((i < lra_constraint_new_regno_start && ! bitmap_bit_p (&do_not_assign_nonreload_pseudos, i)) || (bitmap_bit_p (&lra_inheritance_pseudos, i) - && lra_reg_info[i].restore_regno >= 0) + && lra_reg_info[i].restore_rtx != NULL_RTX) || (bitmap_bit_p (&lra_split_regs, i) - && lra_reg_info[i].restore_regno >= 0) + && lra_reg_info[i].restore_rtx != NULL_RTX) || bitmap_bit_p (&lra_subreg_reload_pseudos, i) || bitmap_bit_p (&lra_optional_reload_pseudos, i)) && reg_renumber[i] < 0 && lra_reg_info[i].nrefs != 0 @@ -1513,7 +1529,7 @@ assign_by_spills (void) enum reg_class spill_class; if (targetm.spill_class == NULL - || lra_reg_info[regno].restore_regno < 0 + || lra_reg_info[regno].restore_rtx == NULL_RTX || ! bitmap_bit_p (&lra_inheritance_pseudos, regno) || (spill_class = ((enum reg_class) @@ -1570,8 +1586,17 @@ lra_assign (void) sorted_pseudos = XNEWVEC (int, max_regno); sorted_reload_pseudos = XNEWVEC (int, max_regno); regno_allocno_class_array = XNEWVEC (enum reg_class, max_regno); + regno_live_length = XNEWVEC (int, max_regno); for (i = FIRST_PSEUDO_REGISTER; i < max_regno; i++) - regno_allocno_class_array[i] = lra_get_allocno_class (i); + { + int l; + lra_live_range_t r; + + regno_allocno_class_array[i] = lra_get_allocno_class (i); + for (l = 0, r = lra_reg_info[i].live_ranges; r != NULL; r = r->next) + l += r->finish - r->start + 1; + regno_live_length[i] = l; + } former_reload_pseudo_spill_p = false; init_regno_assign_info (); bitmap_initialize (&all_spilled_pseudos, ®_obstack); @@ -1594,7 +1619,7 @@ lra_assign (void) EXECUTE_IF_SET_IN_BITMAP (&all_spilled_pseudos, 0, u, bi) /* We ignore spilled pseudos created on last inheritance pass because they will be removed. */ - if (lra_reg_info[u].restore_regno < 0) + if (lra_reg_info[u].restore_rtx == NULL_RTX) { no_spills_p = false; break; @@ -1613,6 +1638,7 @@ lra_assign (void) } bitmap_clear (&insns_to_process); finish_regno_assign_info (); + free (regno_live_length); free (regno_allocno_class_array); free (sorted_pseudos); free (sorted_reload_pseudos); @@ -1631,3 +1657,4 @@ lra_assign (void) LRA_MAX_ASSIGNMENT_ITERATION_NUMBER); return no_spills_p; } + diff --git a/gcc/lra-constraints.c b/gcc/lra-constraints.c index 45b65069e68..43c05ba06d4 100644 --- a/gcc/lra-constraints.c +++ b/gcc/lra-constraints.c @@ -4105,14 +4105,16 @@ curr_insn_transform (bool check_only_p) if (optional_p) { - lra_assert (REG_P (op)); - regno = REGNO (op); + rtx reg = op; + + lra_assert (REG_P (reg)); + regno = REGNO (reg); op = *curr_id->operand_loc[i]; /* Substitution. */ if (GET_CODE (op) == SUBREG) op = SUBREG_REG (op); gcc_assert (REG_P (op) && (int) REGNO (op) >= new_regno_start); bitmap_set_bit (&lra_optional_reload_pseudos, REGNO (op)); - lra_reg_info[REGNO (op)].restore_regno = regno; + lra_reg_info[REGNO (op)].restore_rtx = reg; if (lra_dump_file != NULL) fprintf (lra_dump_file, " Making reload reg %d for reg %d optional\n", @@ -4646,11 +4648,15 @@ lra_constraints (bool first_p) return changed_p; } +static void initiate_invariants (void); +static void finish_invariants (void); + /* Initiate the LRA constraint pass. It is done once per function. */ void lra_constraints_init (void) { + initiate_invariants (); } /* Finalize the LRA constraint pass. It is done once per @@ -4658,6 +4664,100 @@ lra_constraints_init (void) void lra_constraints_finish (void) { + finish_invariants (); +} + + + +/* Structure describes invariants for ineheritance. */ +struct invariant +{ + /* The order number of the invariant. */ + int num; + /* The invariant RTX. */ + rtx invariant_rtx; + /* The origin insn of the invariant. */ + rtx_insn *insn; +}; + +typedef struct invariant invariant_t; +typedef invariant_t *invariant_ptr_t; +typedef const invariant_t *const_invariant_ptr_t; + +/* Pointer to the inheritance invariants. */ +static vec<invariant_ptr_t> invariants; + +/* Allocation pool for the invariants. */ +static object_allocator<struct invariant> *invariants_pool; + +/* Hash table for the invariants. */ +static htab_t invariant_table; + +/* Hash function for INVARIANT. */ +static hashval_t +invariant_hash (const void *invariant) +{ + rtx inv = ((const_invariant_ptr_t) invariant)->invariant_rtx; + return lra_rtx_hash (inv); +} + +/* Equal function for invariants INVARIANT1 and INVARIANT2. */ +static int +invariant_eq_p (const void *invariant1, const void *invariant2) +{ + rtx inv1 = ((const_invariant_ptr_t) invariant1)->invariant_rtx; + rtx inv2 = ((const_invariant_ptr_t) invariant2)->invariant_rtx; + + return rtx_equal_p (inv1, inv2); +} + +/* Insert INVARIANT_RTX into the table if it is not there yet. Return + invariant which is in the table. */ +static invariant_ptr_t +insert_invariant (rtx invariant_rtx) +{ + void **entry_ptr; + invariant_t invariant; + invariant_ptr_t invariant_ptr; + + invariant.invariant_rtx = invariant_rtx; + entry_ptr = htab_find_slot (invariant_table, &invariant, INSERT); + if (*entry_ptr == NULL) + { + invariant_ptr = invariants_pool->allocate (); + invariant_ptr->invariant_rtx = invariant_rtx; + invariant_ptr->insn = NULL; + invariants.safe_push (invariant_ptr); + *entry_ptr = (void *) invariant_ptr; + } + return (invariant_ptr_t) *entry_ptr; +} + +/* Initiate the invariant table. */ +static void +initiate_invariants (void) +{ + invariants.create (100); + invariants_pool = new object_allocator<struct invariant> ("Inheritance invariants"); + invariant_table = htab_create (100, invariant_hash, invariant_eq_p, NULL); +} + +/* Finish the invariant table. */ +static void +finish_invariants (void) +{ + htab_delete (invariant_table); + delete invariants_pool; + invariants.release (); +} + +/* Make the invariant table empty. */ +static void +clear_invariants (void) +{ + htab_empty (invariant_table); + invariants_pool->release (); + invariants.truncate (0); } @@ -4789,6 +4889,10 @@ check_secondary_memory_needed_p (enum reg_class inher_cl ATTRIBUTE_UNUSED, (inheritance/split pseudos and original registers). */ static bitmap_head check_only_regs; +/* Reload pseudos can not be involded in invariant inheritance in the + current EBB. */ +static bitmap_head invalid_invariant_regs; + /* Do inheritance transformations for insn INSN, which defines (if DEF_P) or uses ORIGINAL_REGNO. NEXT_USAGE_INSNS specifies which instruction in the EBB next uses ORIGINAL_REGNO; it has the same @@ -4910,7 +5014,7 @@ inherit_reload_reg (bool def_p, int original_regno, if (lra_dump_file != NULL) fprintf (lra_dump_file, " Original reg change %d->%d (bb%d):\n", original_regno, REGNO (new_reg), BLOCK_FOR_INSN (insn)->index); - lra_reg_info[REGNO (new_reg)].restore_regno = original_regno; + lra_reg_info[REGNO (new_reg)].restore_rtx = regno_reg_rtx[original_regno]; bitmap_set_bit (&check_only_regs, REGNO (new_reg)); bitmap_set_bit (&check_only_regs, original_regno); bitmap_set_bit (&lra_inheritance_pseudos, REGNO (new_reg)); @@ -5194,7 +5298,7 @@ split_reg (bool before_p, int original_regno, rtx_insn *insn, return false; } after_p = usage_insns[original_regno].after_p; - lra_reg_info[REGNO (new_reg)].restore_regno = original_regno; + lra_reg_info[REGNO (new_reg)].restore_rtx = regno_reg_rtx[original_regno]; bitmap_set_bit (&check_only_regs, REGNO (new_reg)); bitmap_set_bit (&check_only_regs, original_regno); bitmap_set_bit (&lra_split_regs, REGNO (new_reg)); @@ -5273,6 +5377,134 @@ split_if_necessary (int regno, machine_mode mode, return res; } +/* Return TRUE if rtx X is considered as an invariant for + inheritance. */ +static bool +invariant_p (const_rtx x) +{ + machine_mode mode; + const char *fmt; + enum rtx_code code; + int i, j; + + code = GET_CODE (x); + mode = GET_MODE (x); + if (code == SUBREG) + { + x = SUBREG_REG (x); + code = GET_CODE (x); + if (GET_MODE_SIZE (GET_MODE (x)) > GET_MODE_SIZE (mode)) + mode = GET_MODE (x); + } + + if (MEM_P (x)) + return false; + + if (REG_P (x)) + { + int i, nregs, regno = REGNO (x); + + if (regno >= FIRST_PSEUDO_REGISTER || regno == STACK_POINTER_REGNUM + || TEST_HARD_REG_BIT (eliminable_regset, regno) + || GET_MODE_CLASS (GET_MODE (x)) == MODE_CC) + return false; + nregs = hard_regno_nregs[regno][mode]; + for (i = 0; i < nregs; i++) + if (! fixed_regs[regno + i] + /* A hard register may be clobbered in the current insn + but we can ignore this case because if the hard + register is used it should be set somewhere after the + clobber. */ + || bitmap_bit_p (&invalid_invariant_regs, regno + i)) + return false; + } + fmt = GET_RTX_FORMAT (code); + for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--) + { + if (fmt[i] == 'e') + { + if (! invariant_p (XEXP (x, i))) + return false; + } + else if (fmt[i] == 'E') + { + for (j = XVECLEN (x, i) - 1; j >= 0; j--) + if (! invariant_p (XVECEXP (x, i, j))) + return false; + } + } + return true; +} + +/* We have 'dest_reg <- invariant'. Let us try to make an invariant + inheritance transformation (using dest_reg instead invariant in a + subsequent insn). */ +static bool +process_invariant_for_inheritance (rtx dst_reg, rtx invariant_rtx) +{ + invariant_ptr_t invariant_ptr; + rtx_insn *insn, *new_insns; + rtx insn_set, insn_reg, new_reg; + int insn_regno; + bool succ_p = false; + int dst_regno = REGNO (dst_reg); + enum machine_mode dst_mode = GET_MODE (dst_reg); + enum reg_class cl = lra_get_allocno_class (dst_regno), insn_reg_cl; + + invariant_ptr = insert_invariant (invariant_rtx); + if ((insn = invariant_ptr->insn) != NULL_RTX) + { + /* We have a subsequent insn using the invariant. */ + insn_set = single_set (insn); + lra_assert (insn_set != NULL); + insn_reg = SET_DEST (insn_set); + lra_assert (REG_P (insn_reg)); + insn_regno = REGNO (insn_reg); + insn_reg_cl = lra_get_allocno_class (insn_regno); + + if (dst_mode == GET_MODE (insn_reg) + /* We should consider only result move reg insns which are + cheap. */ + && targetm.register_move_cost (dst_mode, cl, insn_reg_cl) == 2 + && targetm.register_move_cost (dst_mode, cl, cl) == 2) + { + if (lra_dump_file != NULL) + fprintf (lra_dump_file, + " [[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[[\n"); + new_reg = lra_create_new_reg (dst_mode, dst_reg, + cl, "invariant inheritance"); + bitmap_set_bit (&lra_inheritance_pseudos, REGNO (new_reg)); + bitmap_set_bit (&check_only_regs, REGNO (new_reg)); + lra_reg_info[REGNO (new_reg)].restore_rtx = invariant_rtx; + start_sequence (); + lra_emit_move (new_reg, dst_reg); + new_insns = get_insns (); + end_sequence (); + lra_process_new_insns (curr_insn, NULL, new_insns, + "Add invariant inheritance<-original"); + start_sequence (); + lra_emit_move (SET_DEST (insn_set), new_reg); + new_insns = get_insns (); + end_sequence (); + lra_process_new_insns (insn, NULL, new_insns, + "Changing reload<-inheritance"); + lra_set_insn_deleted (insn); + succ_p = true; + if (lra_dump_file != NULL) + { + fprintf (lra_dump_file, + " Invariant inheritance reuse change %d (bb%d):\n", + REGNO (new_reg), BLOCK_FOR_INSN (insn)->index); + dump_insn_slim (lra_dump_file, insn); + fprintf (lra_dump_file, + " ]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]\n"); + } + } + } + invariant_ptr->insn = curr_insn; + return succ_p; +} + /* Check only registers living at the current program point in the current EBB. */ static bitmap_head live_regs; @@ -5495,7 +5727,7 @@ inherit_in_ebb (rtx_insn *head, rtx_insn *tail) int i, src_regno, dst_regno, nregs; bool change_p, succ_p, update_reloads_num_p; rtx_insn *prev_insn, *last_insn; - rtx next_usage_insns, set; + rtx next_usage_insns, curr_set; enum reg_class cl; struct lra_insn_reg *reg; basic_block last_processed_bb, curr_bb = NULL; @@ -5507,8 +5739,10 @@ inherit_in_ebb (rtx_insn *head, rtx_insn *tail) change_p = false; curr_usage_insns_check++; + clear_invariants (); reloads_num = calls_num = 0; bitmap_clear (&check_only_regs); + bitmap_clear (&invalid_invariant_regs); last_processed_bb = NULL; CLEAR_HARD_REG_SET (potential_reload_hard_regs); COPY_HARD_REG_SET (live_hard_regs, eliminable_regset); @@ -5556,13 +5790,11 @@ inherit_in_ebb (rtx_insn *head, rtx_insn *tail) } } src_regno = dst_regno = -1; - if (NONDEBUG_INSN_P (curr_insn) - && (set = single_set (curr_insn)) != NULL_RTX - && REG_P (SET_DEST (set)) && REG_P (SET_SRC (set))) - { - src_regno = REGNO (SET_SRC (set)); - dst_regno = REGNO (SET_DEST (set)); - } + curr_set = single_set (curr_insn); + if (curr_set != NULL_RTX && REG_P (SET_DEST (curr_set))) + dst_regno = REGNO (SET_DEST (curr_set)); + if (curr_set != NULL_RTX && REG_P (SET_SRC (curr_set))) + src_regno = REGNO (SET_SRC (curr_set)); update_reloads_num_p = true; if (src_regno < lra_constraint_new_regno_start && src_regno >= FIRST_PSEUDO_REGISTER @@ -5587,6 +5819,22 @@ inherit_in_ebb (rtx_insn *head, rtx_insn *tail) IOR_HARD_REG_SET (potential_reload_hard_regs, reg_class_contents[cl]); } + else if (src_regno < 0 + && dst_regno >= lra_constraint_new_regno_start + && invariant_p (SET_SRC (curr_set)) + && (cl = lra_get_allocno_class (dst_regno)) != NO_REGS + && ! bitmap_bit_p (&invalid_invariant_regs, dst_regno)) + { + /* 'reload_pseudo <- invariant'. */ + if (ira_class_hard_regs_num[cl] <= max_small_class_regs_num) + reloads_num++; + update_reloads_num_p = false; + if (process_invariant_for_inheritance (SET_DEST (curr_set), SET_SRC (curr_set))) + change_p = true; + if (hard_reg_set_subset_p (reg_class_contents[cl], live_hard_regs)) + IOR_HARD_REG_SET (potential_reload_hard_regs, + reg_class_contents[cl]); + } else if (src_regno >= lra_constraint_new_regno_start && dst_regno < lra_constraint_new_regno_start && dst_regno >= FIRST_PSEUDO_REGISTER @@ -5819,16 +6067,15 @@ inherit_in_ebb (rtx_insn *head, rtx_insn *tail) } } if (update_reloads_num_p - && NONDEBUG_INSN_P (curr_insn) - && (set = single_set (curr_insn)) != NULL_RTX) + && NONDEBUG_INSN_P (curr_insn) && curr_set != NULL_RTX) { int regno = -1; - if ((REG_P (SET_DEST (set)) - && (regno = REGNO (SET_DEST (set))) >= lra_constraint_new_regno_start + if ((REG_P (SET_DEST (curr_set)) + && (regno = REGNO (SET_DEST (curr_set))) >= lra_constraint_new_regno_start && reg_renumber[regno] < 0 && (cl = lra_get_allocno_class (regno)) != NO_REGS) - || (REG_P (SET_SRC (set)) - && (regno = REGNO (SET_SRC (set))) >= lra_constraint_new_regno_start + || (REG_P (SET_SRC (curr_set)) + && (regno = REGNO (SET_SRC (curr_set))) >= lra_constraint_new_regno_start && reg_renumber[regno] < 0 && (cl = lra_get_allocno_class (regno)) != NO_REGS)) { @@ -5839,6 +6086,25 @@ inherit_in_ebb (rtx_insn *head, rtx_insn *tail) reg_class_contents[cl]); } } + if (NONDEBUG_INSN_P (curr_insn)) + { + int regno; + + /* Invalidate invariants with changed regs. */ + curr_id = lra_get_insn_recog_data (curr_insn); + for (reg = curr_id->regs; reg != NULL; reg = reg->next) + if (reg->type != OP_IN) + bitmap_set_bit (&invalid_invariant_regs, reg->regno); + curr_static_id = curr_id->insn_static_data; + for (reg = curr_static_id->hard_regs; reg != NULL; reg = reg->next) + if (reg->type != OP_IN) + bitmap_set_bit (&invalid_invariant_regs, reg->regno); + if (curr_id->arg_hard_regs != NULL) + for (i = 0; (regno = curr_id->arg_hard_regs[i]) >= 0; i++) + bitmap_set_bit (&invalid_invariant_regs, + regno >= FIRST_PSEUDO_REGISTER + ? regno : regno - FIRST_PSEUDO_REGISTER); + } /* We reached the start of the current basic block. */ if (prev_insn == NULL_RTX || prev_insn == PREV_INSN (head) || BLOCK_FOR_INSN (prev_insn) != curr_bb) @@ -5911,6 +6177,7 @@ lra_inheritance (void) for (i = 0; i < lra_constraint_new_regno_start; i++) usage_insns[i].check = 0; bitmap_initialize (&check_only_regs, ®_obstack); + bitmap_initialize (&invalid_invariant_regs, ®_obstack); bitmap_initialize (&live_regs, ®_obstack); bitmap_initialize (&temp_bitmap, ®_obstack); bitmap_initialize (&ebb_global_regs, ®_obstack); @@ -5947,6 +6214,7 @@ lra_inheritance (void) bitmap_clear (&ebb_global_regs); bitmap_clear (&temp_bitmap); bitmap_clear (&live_regs); + bitmap_clear (&invalid_invariant_regs); bitmap_clear (&check_only_regs); free (usage_insns); @@ -5970,8 +6238,9 @@ fix_bb_live_info (bitmap live, bitmap removed_pseudos) bitmap_iterator bi; EXECUTE_IF_SET_IN_BITMAP (removed_pseudos, 0, regno, bi) - if (bitmap_clear_bit (live, regno)) - bitmap_set_bit (live, lra_reg_info[regno].restore_regno); + if (bitmap_clear_bit (live, regno) + && REG_P (lra_reg_info[regno].restore_rtx)) + bitmap_set_bit (live, REGNO (lra_reg_info[regno].restore_rtx)); } /* Return regno of the (subreg of) REG. Otherwise, return a negative @@ -6024,7 +6293,8 @@ static bool remove_inheritance_pseudos (bitmap remove_pseudos) { basic_block bb; - int regno, sregno, prev_sregno, dregno, restore_regno; + int regno, sregno, prev_sregno, dregno; + rtx restore_rtx; rtx set, prev_set; rtx_insn *prev_insn; bool change_p, done_p; @@ -6053,13 +6323,42 @@ remove_inheritance_pseudos (bitmap remove_pseudos) if (sregno >= 0 && dregno >= 0) { - if ((bitmap_bit_p (remove_pseudos, sregno) - && (lra_reg_info[sregno].restore_regno == dregno + if (bitmap_bit_p (remove_pseudos, dregno) + && ! REG_P (lra_reg_info[dregno].restore_rtx)) + { + /* invariant inheritance pseudo <- original pseudo */ + if (lra_dump_file != NULL) + { + fprintf (lra_dump_file, " Removing invariant inheritance:\n"); + dump_insn_slim (lra_dump_file, curr_insn); + fprintf (lra_dump_file, "\n"); + } + delete_move_and_clobber (curr_insn, dregno); + done_p = true; + } + else if (bitmap_bit_p (remove_pseudos, sregno) + && ! REG_P (lra_reg_info[sregno].restore_rtx)) + { + /* reload pseudo <- invariant inheritance pseudo */ + start_sequence (); + /* We can not just change the source. It might be + an insn different from the move. */ + lra_emit_move (SET_DEST (set), lra_reg_info[sregno].restore_rtx); + rtx_insn *new_insns = get_insns (); + end_sequence (); + lra_process_new_insns (curr_insn, NULL, new_insns, + "Changing reload<-invariant inheritance"); + delete_move_and_clobber (curr_insn, dregno); + done_p = true; + } + else if ((bitmap_bit_p (remove_pseudos, sregno) + && (get_regno (lra_reg_info[sregno].restore_rtx) == dregno + || (bitmap_bit_p (remove_pseudos, dregno) + && get_regno (lra_reg_info[sregno].restore_rtx) >= 0 + && (get_regno (lra_reg_info[sregno].restore_rtx) + == get_regno (lra_reg_info[dregno].restore_rtx))))) || (bitmap_bit_p (remove_pseudos, dregno) - && (lra_reg_info[sregno].restore_regno - == lra_reg_info[dregno].restore_regno)))) - || (bitmap_bit_p (remove_pseudos, dregno) - && lra_reg_info[dregno].restore_regno == sregno)) + && get_regno (lra_reg_info[dregno].restore_rtx) == sregno)) /* One of the following cases: original <- removed inheritance pseudo removed inherit pseudo <- another removed inherit pseudo @@ -6104,13 +6403,16 @@ remove_inheritance_pseudos (bitmap remove_pseudos) && (int) REGNO (SET_DEST (prev_set)) == sregno && ((prev_sregno = REGNO (SET_SRC (prev_set))) >= FIRST_PSEUDO_REGISTER) - /* As we consider chain of inheritance or - splitting described in above comment we should - check that sregno and prev_sregno were - inheritance/split pseudos created from the - same original regno. */ - && (lra_reg_info[sregno].restore_regno - == lra_reg_info[prev_sregno].restore_regno) + && (lra_reg_info[prev_sregno].restore_rtx == NULL_RTX + || + /* As we consider chain of inheritance or + splitting described in above comment we should + check that sregno and prev_sregno were + inheritance/split pseudos created from the + same original regno. */ + (get_regno (lra_reg_info[sregno].restore_rtx) >= 0 + && (get_regno (lra_reg_info[sregno].restore_rtx) + == get_regno (lra_reg_info[prev_sregno].restore_rtx)))) && ! bitmap_bit_p (remove_pseudos, prev_sregno)) { lra_assert (GET_MODE (SET_SRC (prev_set)) @@ -6124,14 +6426,13 @@ remove_inheritance_pseudos (bitmap remove_pseudos) inheritance pseudo for another pseudo. */ if (bitmap_bit_p (remove_pseudos, dregno) && bitmap_bit_p (&lra_inheritance_pseudos, dregno) - && (restore_regno - = lra_reg_info[dregno].restore_regno) >= 0) + && (restore_rtx + = lra_reg_info[dregno].restore_rtx) != NULL_RTX) { if (GET_CODE (SET_DEST (set)) == SUBREG) - SUBREG_REG (SET_DEST (set)) - = regno_reg_rtx[restore_regno]; + SUBREG_REG (SET_DEST (set)) = restore_rtx; else - SET_DEST (set) = regno_reg_rtx[restore_regno]; + SET_DEST (set) = restore_rtx; } lra_push_insn_and_update_insn_regno_info (curr_insn); lra_set_used_insn_alternative_by_uid @@ -6155,14 +6456,13 @@ remove_inheritance_pseudos (bitmap remove_pseudos) for (reg = curr_id->regs; reg != NULL; reg = reg->next) { regno = reg->regno; - restore_regno = lra_reg_info[regno].restore_regno; - if (restore_regno >= 0) + restore_rtx = lra_reg_info[regno].restore_rtx; + if (restore_rtx != NULL_RTX) { if (change_p && bitmap_bit_p (remove_pseudos, regno)) { lra_substitute_pseudo_within_insn - (curr_insn, regno, regno_reg_rtx[restore_regno], - false); + (curr_insn, regno, restore_rtx, false); restored_regs_p = true; } else @@ -6213,11 +6513,11 @@ undo_optional_reloads (void) { keep_p = false; /* Keep optional reloads from previous subpasses. */ - if (lra_reg_info[regno].restore_regno < 0 + if (lra_reg_info[regno].restore_rtx == NULL_RTX /* If the original pseudo changed its allocation, just removing the optional pseudo is dangerous as the original pseudo will have longer live range. */ - || reg_renumber[lra_reg_info[regno].restore_regno] >= 0) + || reg_renumber[REGNO (lra_reg_info[regno].restore_rtx)] >= 0) keep_p = true; else if (reg_renumber[regno] >= 0) EXECUTE_IF_SET_IN_BITMAP (&lra_reg_info[regno].insn_bitmap, 0, uid, bi2) @@ -6231,7 +6531,7 @@ undo_optional_reloads (void) continue; if (REGNO (dest) == regno /* Ignore insn for optional reloads itself. */ - && lra_reg_info[regno].restore_regno != (int) REGNO (src) + && REGNO (lra_reg_info[regno].restore_rtx) != REGNO (src) /* Check only inheritance on last inheritance pass. */ && (int) REGNO (src) >= new_regno_start /* Check that the optional reload was inherited. */ @@ -6264,11 +6564,11 @@ undo_optional_reloads (void) dest = SET_DEST (set); if (REG_P (src) && REG_P (dest) && ((REGNO (src) == regno - && (lra_reg_info[regno].restore_regno - == (int) REGNO (dest))) + && (REGNO (lra_reg_info[regno].restore_rtx) + == REGNO (dest))) || (REGNO (dest) == regno - && (lra_reg_info[regno].restore_regno - == (int) REGNO (src))))) + && (REGNO (lra_reg_info[regno].restore_rtx) + == REGNO (src))))) { if (lra_dump_file != NULL) { @@ -6286,8 +6586,7 @@ undo_optional_reloads (void) reload. */ } lra_substitute_pseudo_within_insn - (insn, regno, regno_reg_rtx[lra_reg_info[regno].restore_regno], - false); + (insn, regno, lra_reg_info[regno].restore_rtx, false); lra_update_insn_regno_info (insn); if (lra_dump_file != NULL) { @@ -6299,7 +6598,7 @@ undo_optional_reloads (void) } /* Clear restore_regnos. */ EXECUTE_IF_SET_IN_BITMAP (&lra_optional_reload_pseudos, 0, regno, bi) - lra_reg_info[regno].restore_regno = -1; + lra_reg_info[regno].restore_rtx = NULL_RTX; bitmap_clear (&insn_bitmap); bitmap_clear (&removed_optional_reload_pseudos); return change_p; @@ -6311,8 +6610,9 @@ bool lra_undo_inheritance (void) { unsigned int regno; - int restore_regno, hard_regno; + int hard_regno; int n_all_inherit, n_inherit, n_all_split, n_split; + rtx restore_rtx; bitmap_head remove_pseudos; bitmap_iterator bi; bool change_p; @@ -6327,14 +6627,15 @@ lra_undo_inheritance (void) bitmap_initialize (&remove_pseudos, ®_obstack); n_inherit = n_all_inherit = 0; EXECUTE_IF_SET_IN_BITMAP (&lra_inheritance_pseudos, 0, regno, bi) - if (lra_reg_info[regno].restore_regno >= 0) + if (lra_reg_info[regno].restore_rtx != NULL_RTX) { n_all_inherit++; if (reg_renumber[regno] < 0 /* If the original pseudo changed its allocation, just removing inheritance is dangerous as for changing allocation we used shorter live-ranges. */ - && reg_renumber[lra_reg_info[regno].restore_regno] < 0) + && (! REG_P (lra_reg_info[regno].restore_rtx) + || reg_renumber[REGNO (lra_reg_info[regno].restore_rtx)] < 0)) bitmap_set_bit (&remove_pseudos, regno); else n_inherit++; @@ -6345,8 +6646,10 @@ lra_undo_inheritance (void) (double) n_inherit / n_all_inherit * 100); n_split = n_all_split = 0; EXECUTE_IF_SET_IN_BITMAP (&lra_split_regs, 0, regno, bi) - if ((restore_regno = lra_reg_info[regno].restore_regno) >= 0) + if ((restore_rtx = lra_reg_info[regno].restore_rtx) != NULL_RTX) { + int restore_regno = REGNO (restore_rtx); + n_all_split++; hard_regno = (restore_regno >= FIRST_PSEUDO_REGISTER ? reg_renumber[restore_regno] : restore_regno); @@ -6368,9 +6671,9 @@ lra_undo_inheritance (void) bitmap_clear (&remove_pseudos); /* Clear restore_regnos. */ EXECUTE_IF_SET_IN_BITMAP (&lra_inheritance_pseudos, 0, regno, bi) - lra_reg_info[regno].restore_regno = -1; + lra_reg_info[regno].restore_rtx = NULL_RTX; EXECUTE_IF_SET_IN_BITMAP (&lra_split_regs, 0, regno, bi) - lra_reg_info[regno].restore_regno = -1; + lra_reg_info[regno].restore_rtx = NULL_RTX; change_p = undo_optional_reloads () || change_p; return change_p; } diff --git a/gcc/lra-int.h b/gcc/lra-int.h index 8ba0bacd774..cfab209cef1 100644 --- a/gcc/lra-int.h +++ b/gcc/lra-int.h @@ -99,9 +99,9 @@ struct lra_reg *non-debug* insns. */ int nrefs, freq; int last_reload; - /* Regno used to undo the inheritance. It can be non-zero only - between couple of inheritance and undo inheritance passes. */ - int restore_regno; + /* rtx used to undo the inheritance. It can be non-null only + between subsequent inheritance and undo inheritance passes. */ + rtx restore_rtx; /* Value holding by register. If the pseudos have the same value they do not conflict. */ int val; @@ -285,6 +285,7 @@ extern lra_insn_recog_data_t *lra_insn_recog_data; extern int lra_curr_reload_num; extern void lra_dump_bitmap_with_title (const char *, bitmap, int); +extern hashval_t lra_rtx_hash (rtx x); extern void lra_push_insn (rtx_insn *); extern void lra_push_insn_by_uid (unsigned int); extern void lra_push_insn_and_update_insn_regno_info (rtx_insn *); diff --git a/gcc/lra-remat.c b/gcc/lra-remat.c index 187ee3e7752..c124d5ba2e1 100644 --- a/gcc/lra-remat.c +++ b/gcc/lra-remat.c @@ -167,92 +167,6 @@ get_remat_bb_data_by_index (int index) -/* Recursive hash function for RTL X. */ -static hashval_t -rtx_hash (rtx x) -{ - int i, j; - enum rtx_code code; - const char *fmt; - hashval_t val = 0; - - if (x == 0) - return val; - - code = GET_CODE (x); - val += (int) code + 4095; - - /* Some RTL can be compared nonrecursively. */ - switch (code) - { - case REG: - return val + REGNO (x); - - case LABEL_REF: - return iterative_hash_object (XEXP (x, 0), val); - - case SYMBOL_REF: - return iterative_hash_object (XSTR (x, 0), val); - - case SCRATCH: - case CONST_DOUBLE: - case CONST_INT: - case CONST_VECTOR: - return val; - - default: - break; - } - - /* Hash the elements. */ - fmt = GET_RTX_FORMAT (code); - for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--) - { - switch (fmt[i]) - { - case 'w': - val += XWINT (x, i); - break; - - case 'n': - case 'i': - val += XINT (x, i); - break; - - case 'V': - case 'E': - val += XVECLEN (x, i); - - for (j = 0; j < XVECLEN (x, i); j++) - val += rtx_hash (XVECEXP (x, i, j)); - break; - - case 'e': - val += rtx_hash (XEXP (x, i)); - break; - - case 'S': - case 's': - val += htab_hash_string (XSTR (x, i)); - break; - - case 'u': - case '0': - case 't': - break; - - /* It is believed that rtx's at this level will never - contain anything but integers and other rtx's, except for - within LABEL_REFs and SYMBOL_REFs. */ - default: - abort (); - } - } - return val; -} - - - /* Hash table for the candidates. Different insns (e.g. structurally the same insns or even insns with different unused output regs) can be represented by the same candidate in the table. */ diff --git a/gcc/lra-spills.c b/gcc/lra-spills.c index 7e5f63785a0..1801b7bb391 100644 --- a/gcc/lra-spills.c +++ b/gcc/lra-spills.c @@ -693,7 +693,7 @@ lra_final_code_change (void) { int i, hard_regno; basic_block bb; - rtx_insn *insn, *curr; + rtx_insn *insn, *curr, *next_insn; int max_regno = max_reg_num (); for (i = FIRST_PSEUDO_REGISTER; i < max_regno; i++) @@ -727,7 +727,12 @@ lra_final_code_change (void) if (NONJUMP_INSN_P (insn) && GET_CODE (pat) == SET && REG_P (SET_SRC (pat)) && REG_P (SET_DEST (pat)) && REGNO (SET_SRC (pat)) == REGNO (SET_DEST (pat)) - && ! return_regno_p (REGNO (SET_SRC (pat)))) + && (! return_regno_p (REGNO (SET_SRC (pat))) + || (next_insn = next_nondebug_insn (insn)) == NULL_RTX + || ! INSN_P (next_insn) + || GET_CODE (PATTERN (next_insn)) != USE + || ! REG_P (XEXP (PATTERN (next_insn), 0)) + || REGNO (SET_SRC (pat)) != REGNO (XEXP (PATTERN (next_insn), 0)))) { lra_invalidate_insn_data (insn); delete_insn (insn); diff --git a/gcc/lra.c b/gcc/lra.c index a63e3933ed1..cb2bd35cffd 100644 --- a/gcc/lra.c +++ b/gcc/lra.c @@ -1286,7 +1286,7 @@ initialize_lra_reg_info_element (int i) lra_reg_info[i].live_ranges = NULL; lra_reg_info[i].nrefs = lra_reg_info[i].freq = 0; lra_reg_info[i].last_reload = 0; - lra_reg_info[i].restore_regno = -1; + lra_reg_info[i].restore_rtx = NULL_RTX; lra_reg_info[i].val = get_new_reg_value (); lra_reg_info[i].offset = 0; lra_reg_info[i].copies = NULL; @@ -1621,6 +1621,92 @@ lra_get_insn_regs (int uid) +/* Recursive hash function for RTL X. */ +hashval_t +lra_rtx_hash (rtx x) +{ + int i, j; + enum rtx_code code; + const char *fmt; + hashval_t val = 0; + + if (x == 0) + return val; + + code = GET_CODE (x); + val += (int) code + 4095; + + /* Some RTL can be compared nonrecursively. */ + switch (code) + { + case REG: + return val + REGNO (x); + + case LABEL_REF: + return iterative_hash_object (XEXP (x, 0), val); + + case SYMBOL_REF: + return iterative_hash_object (XSTR (x, 0), val); + + case SCRATCH: + case CONST_DOUBLE: + case CONST_INT: + case CONST_VECTOR: + return val; + + default: + break; + } + + /* Hash the elements. */ + fmt = GET_RTX_FORMAT (code); + for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--) + { + switch (fmt[i]) + { + case 'w': + val += XWINT (x, i); + break; + + case 'n': + case 'i': + val += XINT (x, i); + break; + + case 'V': + case 'E': + val += XVECLEN (x, i); + + for (j = 0; j < XVECLEN (x, i); j++) + val += lra_rtx_hash (XVECEXP (x, i, j)); + break; + + case 'e': + val += lra_rtx_hash (XEXP (x, i)); + break; + + case 'S': + case 's': + val += htab_hash_string (XSTR (x, i)); + break; + + case 'u': + case '0': + case 't': + break; + + /* It is believed that rtx's at this level will never + contain anything but integers and other rtx's, except for + within LABEL_REFs and SYMBOL_REFs. */ + default: + abort (); + } + } + return val; +} + + + /* This page contains code dealing with stack of the insns which should be processed by the next constraint pass. */ |