/* Rematerialize pseudos values. Copyright (C) 2014 Free Software Foundation, Inc. Contributed by Vladimir Makarov . This file is part of GCC. GCC is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3, or (at your option) any later version. GCC is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with GCC; see the file COPYING3. If not see . */ /* This code objective is to rematerialize spilled pseudo values. To do this we calculate available insn candidates. The candidate is available at some point if there is dominated set of insns with the same pattern, the insn inputs are not dying or modified on any path from the set, the outputs are not modified. The insns containing memory or spilled pseudos (except for the rematerialized pseudo) are not considered as such insns are not profitable in comparison with regular loads of spilled pseudo values. That simplifies the implementation as we don't need to deal with memory aliasing. To speed up available candidate calculation, we calculate partially available candidates first and use them for initialization of the availability. That is because (partial) availability sets are sparse. The rematerialization sub-pass could be improved further in the following ways: o We could make longer live ranges of inputs in the rematerialization candidates if their hard registers are not used for other purposes. This could be complicated if we need to update BB live info information as LRA does not use DF-infrastructure for compile-time reasons. This problem could be overcome if constrain making live ranges longer only in BB/EBB scope. o We could use cost-based decision to choose rematerialization insn (currently all insns without memory is can be used). o We could use other free hard regs for unused output pseudos in rematerialization candidates although such cases probably will be very rare. */ #include "config.h" #include "system.h" #include "coretypes.h" #include "tm.h" #include "hard-reg-set.h" #include "rtl.h" #include "rtl-error.h" #include "tm_p.h" #include "target.h" #include "insn-config.h" #include "recog.h" #include "output.h" #include "regs.h" #include "hashtab.h" #include "hash-set.h" #include "vec.h" #include "machmode.h" #include "input.h" #include "function.h" #include "expr.h" #include "predict.h" #include "dominance.h" #include "cfg.h" #include "basic-block.h" #include "except.h" #include "df.h" #include "ira.h" #include "sparseset.h" #include "params.h" #include "df.h" #include "lra-int.h" /* Number of candidates for rematerialization. */ static unsigned int cands_num; /* The following is used for representation of call_used_reg_set in form array whose elements are hard register numbers with nonzero bit in CALL_USED_REG_SET. */ static int call_used_regs_arr_len; static int call_used_regs_arr[FIRST_PSEUDO_REGISTER]; /* Bitmap used for different calculations. */ static bitmap_head temp_bitmap; typedef struct cand *cand_t; typedef const struct cand *const_cand_t; /* Insn candidates for rematerialization. The candidate insn should have the following properies: o no any memory (as access to memory is non-profitable) o no INOUT regs (it means no non-paradoxical subreg of output reg) o one output spilled pseudo (or reload pseudo of a spilled pseudo) o all other pseudos are with assigned hard regs. */ struct cand { /* Index of the candidates in all_cands. */ int index; /* The candidate insn. */ rtx_insn *insn; /* Insn pseudo regno for rematerialization. */ int regno; /* Non-negative if a reload pseudo is in the insn instead of the pseudo for rematerialization. */ int reload_regno; /* Number of the operand containing the regno or its reload regno. */ int nop; /* Next candidate for the same regno. */ cand_t next_regno_cand; }; /* Vector containing all candidates. */ static vec all_cands; /* Map: insn -> candidate representing it. It is null if the insn can not be used for rematerialization. */ static cand_t *insn_to_cand; /* Map regno -> candidates can be used for the regno rematerialization. */ static cand_t *regno_cands; /* Data about basic blocks used for the rematerialization sub-pass. */ struct remat_bb_data { /* Basic block about which the below data are. */ basic_block bb; /* Registers changed in the basic block: */ bitmap_head changed_regs; /* Registers becoming dead in the BB. */ bitmap_head dead_regs; /* Cands present in the BB whose in/out regs are not changed after the cands occurence and are not dead (except the reload regno). */ bitmap_head gen_cands; bitmap_head livein_cands; /* cands whose inputs live at the BB start. */ bitmap_head pavin_cands; /* cands partially available at BB entry. */ bitmap_head pavout_cands; /* cands partially available at BB exit. */ bitmap_head avin_cands; /* cands available at the entry of the BB. */ bitmap_head avout_cands; /* cands available at the exit of the BB. */ }; /* Array for all BB data. Indexed by the corresponding BB index. */ typedef struct remat_bb_data *remat_bb_data_t; /* Basic blocks for data flow problems -- all bocks except the special ones. */ static bitmap_head all_blocks; /* All basic block data are referred through the following array. */ static remat_bb_data_t remat_bb_data; /* Two small functions for access to the bb data. */ static inline remat_bb_data_t get_remat_bb_data (basic_block bb) { return &remat_bb_data[(bb)->index]; } static inline remat_bb_data_t get_remat_bb_data_by_index (int index) { return &remat_bb_data[index]; } /* Recursive hash function for RTL X. */ static hashval_t rtx_hash (rtx x) { int i, j; enum rtx_code code; const char *fmt; hashval_t val = 0; if (x == 0) return val; code = GET_CODE (x); val += (int) code + 4095; /* Some RTL can be compared nonrecursively. */ switch (code) { case REG: return val + REGNO (x); case LABEL_REF: return iterative_hash_object (XEXP (x, 0), val); case SYMBOL_REF: return iterative_hash_object (XSTR (x, 0), val); case SCRATCH: case CONST_DOUBLE: case CONST_INT: case CONST_VECTOR: return val; default: break; } /* Hash the elements. */ fmt = GET_RTX_FORMAT (code); for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--) { switch (fmt[i]) { case 'w': val += XWINT (x, i); break; case 'n': case 'i': val += XINT (x, i); break; case 'V': case 'E': val += XVECLEN (x, i); for (j = 0; j < XVECLEN (x, i); j++) val += rtx_hash (XVECEXP (x, i, j)); break; case 'e': val += rtx_hash (XEXP (x, i)); break; case 'S': case 's': val += htab_hash_string (XSTR (x, i)); break; case 'u': case '0': case 't': break; /* It is believed that rtx's at this level will never contain anything but integers and other rtx's, except for within LABEL_REFs and SYMBOL_REFs. */ default: abort (); } } return val; } /* Hash table for the candidates. Different insns (e.g. structurally the same insns or even insns with different unused output regs) can be represented by the same candidate in the table. */ static htab_t cand_table; /* Hash function for candidate CAND. */ static hashval_t cand_hash (const void *cand) { const_cand_t c = (const_cand_t) cand; lra_insn_recog_data_t id = lra_get_insn_recog_data (c->insn); struct lra_static_insn_data *static_id = id->insn_static_data; int nops = static_id->n_operands; hashval_t hash = 0; for (int i = 0; i < nops; i++) if (i == c->nop) hash = iterative_hash_object (c->regno, hash); else if (static_id->operand[i].type == OP_IN) hash = iterative_hash_object (*id->operand_loc[i], hash); return hash; } /* Equal function for candidates CAND1 and CAND2. They are equal if the corresponding candidate insns have the same code, the same regno for rematerialization, the same input operands. */ static int cand_eq_p (const void *cand1, const void *cand2) { const_cand_t c1 = (const_cand_t) cand1; const_cand_t c2 = (const_cand_t) cand2; lra_insn_recog_data_t id1 = lra_get_insn_recog_data (c1->insn); lra_insn_recog_data_t id2 = lra_get_insn_recog_data (c2->insn); struct lra_static_insn_data *static_id1 = id1->insn_static_data; int nops = static_id1->n_operands; if (c1->regno != c2->regno || INSN_CODE (c1->insn) < 0 || INSN_CODE (c1->insn) != INSN_CODE (c2->insn)) return false; gcc_assert (c1->nop == c2->nop); for (int i = 0; i < nops; i++) if (i != c1->nop && static_id1->operand[i].type == OP_IN && *id1->operand_loc[i] != *id2->operand_loc[i]) return false; return true; } /* Insert candidate CAND into the table if it is not there yet. Return candidate which is in the table. */ static cand_t insert_cand (cand_t cand) { void **entry_ptr; entry_ptr = htab_find_slot (cand_table, cand, INSERT); if (*entry_ptr == NULL) *entry_ptr = (void *) cand; return (cand_t) *entry_ptr; } /* Free candidate CAND memory. */ static void free_cand (void *cand) { free (cand); } /* Initiate the candidate table. */ static void initiate_cand_table (void) { cand_table = htab_create (8000, cand_hash, cand_eq_p, (htab_del) free_cand); } /* Finish the candidate table. */ static void finish_cand_table (void) { htab_delete (cand_table); } /* Return true if X contains memory or UNSPEC. We can not just check insn operands as memory or unspec might be not an operand itself but contain an operand. Insn with memory access is not profitable for rematerialization. Rematerialization of UNSPEC might result in wrong code generation as the UNPEC effect is unknown (e.g. generating a label). */ static bool bad_for_rematerialization_p (rtx x) { int i, j; const char *fmt; enum rtx_code code; if (MEM_P (x) || GET_CODE (x) == UNSPEC) return true; code = GET_CODE (x); fmt = GET_RTX_FORMAT (code); for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--) { if (fmt[i] == 'e') { if (bad_for_rematerialization_p (XEXP (x, i))) return true; } else if (fmt[i] == 'E') { for (j = XVECLEN (x, i) - 1; j >= 0; j--) if (bad_for_rematerialization_p (XVECEXP (x, i, j))) return true; } } return false; } /* If INSN can not be used for rematerialization, return negative value. If INSN can be considered as a candidate for rematerialization, return value which is the operand number of the pseudo for which the insn can be used for rematerialization. Here we consider the insns without any memory, spilled pseudo (except for the rematerialization pseudo), or dying or unused regs. */ static int operand_to_remat (rtx_insn *insn) { lra_insn_recog_data_t id = lra_get_insn_recog_data (insn); struct lra_static_insn_data *static_id = id->insn_static_data; struct lra_insn_reg *reg, *found_reg = NULL; /* First find a pseudo which can be rematerialized. */ for (reg = id->regs; reg != NULL; reg = reg->next) /* True FRAME_POINTER_NEEDED might be because we can not follow changing sp offsets, e.g. alloca is used. If the insn contains stack pointer in such case, we can not rematerialize it as we can not know sp offset at a rematerialization place. */ if (reg->regno == STACK_POINTER_REGNUM && frame_pointer_needed) return -1; else if (reg->type == OP_OUT && ! reg->subreg_p && find_regno_note (insn, REG_UNUSED, reg->regno) == NULL) { /* We permits only one spilled reg. */ if (found_reg != NULL) return -1; found_reg = reg; } if (found_reg == NULL) return -1; if (found_reg->regno < FIRST_PSEUDO_REGISTER) return -1; if (bad_for_rematerialization_p (PATTERN (insn))) return -1; /* Check the other regs are not spilled. */ for (reg = id->regs; reg != NULL; reg = reg->next) if (found_reg == reg) continue; else if (reg->type == OP_INOUT) return -1; else if (reg->regno >= FIRST_PSEUDO_REGISTER && reg_renumber[reg->regno] < 0) /* Another spilled reg. */ return -1; else if (reg->type == OP_IN) { if (find_regno_note (insn, REG_DEAD, reg->regno) != NULL) /* We don't want to make live ranges longer. */ return -1; /* Check that there is no output reg as the input one. */ for (struct lra_insn_reg *reg2 = id->regs; reg2 != NULL; reg2 = reg2->next) if (reg2->type == OP_OUT && reg->regno == reg2->regno) return -1; } /* Find the rematerialization operand. */ int nop = static_id->n_operands; for (int i = 0; i < nop; i++) if (REG_P (*id->operand_loc[i]) && (int) REGNO (*id->operand_loc[i]) == found_reg->regno) return i; return -1; } /* Create candidate for INSN with rematerialization operand NOP and REGNO. Insert the candidate into the table and set up the corresponding INSN_TO_CAND element. */ static void create_cand (rtx_insn *insn, int nop, int regno) { lra_insn_recog_data_t id = lra_get_insn_recog_data (insn); rtx reg = *id->operand_loc[nop]; gcc_assert (REG_P (reg)); int op_regno = REGNO (reg); gcc_assert (op_regno >= FIRST_PSEUDO_REGISTER); cand_t cand = XNEW (struct cand); cand->insn = insn; cand->nop = nop; cand->regno = regno; cand->reload_regno = op_regno == regno ? -1 : op_regno; gcc_assert (cand->regno >= 0); cand_t cand_in_table = insert_cand (cand); insn_to_cand[INSN_UID (insn)] = cand_in_table; if (cand != cand_in_table) free (cand); else { /* A new cand. */ cand->index = all_cands.length (); all_cands.safe_push (cand); cand->next_regno_cand = regno_cands[cand->regno]; regno_cands[cand->regno] = cand; } } /* Create rematerialization candidates (inserting them into the table). */ static void create_cands (void) { rtx_insn *insn; struct potential_cand { rtx_insn *insn; int nop; }; struct potential_cand *regno_potential_cand; /* Create candidates. */ regno_potential_cand = XCNEWVEC (struct potential_cand, max_reg_num ()); for (insn = get_insns (); insn; insn = NEXT_INSN (insn)) if (INSN_P (insn)) { rtx set; int src_regno, dst_regno; rtx_insn *insn2; lra_insn_recog_data_t id = lra_get_insn_recog_data (insn); int nop = operand_to_remat (insn); int regno = -1; if ((set = single_set (insn)) != NULL && REG_P (SET_SRC (set)) && REG_P (SET_DEST (set)) && (src_regno = REGNO (SET_SRC (set))) >= FIRST_PSEUDO_REGISTER && (dst_regno = REGNO (SET_DEST (set))) >= FIRST_PSEUDO_REGISTER && reg_renumber[dst_regno] < 0 && (insn2 = regno_potential_cand[src_regno].insn) != NULL && BLOCK_FOR_INSN (insn2) == BLOCK_FOR_INSN (insn)) create_cand (insn2, regno_potential_cand[src_regno].nop, dst_regno); if (nop < 0) goto fail; gcc_assert (REG_P (*id->operand_loc[nop])); regno = REGNO (*id->operand_loc[nop]); gcc_assert (regno >= FIRST_PSEUDO_REGISTER); if (reg_renumber[regno] < 0) create_cand (insn, nop, regno); else { regno_potential_cand[regno].insn = insn; regno_potential_cand[regno].nop = nop; goto fail; } regno = -1; fail: for (struct lra_insn_reg *reg = id->regs; reg != NULL; reg = reg->next) if (reg->type != OP_IN && reg->regno != regno && reg->regno >= FIRST_PSEUDO_REGISTER) regno_potential_cand[reg->regno].insn = NULL; } cands_num = all_cands.length (); free (regno_potential_cand); } /* Create and initialize BB data. */ static void create_remat_bb_data (void) { basic_block bb; remat_bb_data_t bb_info; remat_bb_data = XNEWVEC (struct remat_bb_data, last_basic_block_for_fn (cfun)); FOR_ALL_BB_FN (bb, cfun) { #ifdef ENABLE_CHECKING if (bb->index < 0 || bb->index >= last_basic_block_for_fn (cfun)) abort (); #endif bb_info = get_remat_bb_data (bb); bb_info->bb = bb; bitmap_initialize (&bb_info->changed_regs, ®_obstack); bitmap_initialize (&bb_info->dead_regs, ®_obstack); bitmap_initialize (&bb_info->gen_cands, ®_obstack); bitmap_initialize (&bb_info->livein_cands, ®_obstack); bitmap_initialize (&bb_info->pavin_cands, ®_obstack); bitmap_initialize (&bb_info->pavout_cands, ®_obstack); bitmap_initialize (&bb_info->avin_cands, ®_obstack); bitmap_initialize (&bb_info->avout_cands, ®_obstack); } } /* Dump all candidates to DUMP_FILE. */ static void dump_cands (FILE *dump_file) { int i; cand_t cand; fprintf (dump_file, "\nCands:\n"); for (i = 0; i < (int) cands_num; i++) { cand = all_cands[i]; fprintf (dump_file, "%d (nop=%d, remat_regno=%d, reload_regno=%d):\n", i, cand->nop, cand->regno, cand->reload_regno); print_inline_rtx (dump_file, cand->insn, 6); fprintf (dump_file, "\n"); } } /* Dump all candidates and BB data. */ static void dump_candidates_and_remat_bb_data (void) { basic_block bb; if (lra_dump_file == NULL) return; dump_cands (lra_dump_file); FOR_EACH_BB_FN (bb, cfun) { fprintf (lra_dump_file, "\nBB %d:\n", bb->index); /* Livein */ fprintf (lra_dump_file, " register live in:"); dump_regset (df_get_live_in (bb), lra_dump_file); putc ('\n', lra_dump_file); /* Liveout */ fprintf (lra_dump_file, " register live out:"); dump_regset (df_get_live_out (bb), lra_dump_file); putc ('\n', lra_dump_file); /* Changed/dead regs: */ fprintf (lra_dump_file, " changed regs:"); dump_regset (&get_remat_bb_data (bb)->changed_regs, lra_dump_file); putc ('\n', lra_dump_file); fprintf (lra_dump_file, " dead regs:"); dump_regset (&get_remat_bb_data (bb)->dead_regs, lra_dump_file); putc ('\n', lra_dump_file); lra_dump_bitmap_with_title ("cands generated in BB", &get_remat_bb_data (bb)->gen_cands, bb->index); lra_dump_bitmap_with_title ("livein cands in BB", &get_remat_bb_data (bb)->livein_cands, bb->index); lra_dump_bitmap_with_title ("pavin cands in BB", &get_remat_bb_data (bb)->pavin_cands, bb->index); lra_dump_bitmap_with_title ("pavout cands in BB", &get_remat_bb_data (bb)->pavout_cands, bb->index); lra_dump_bitmap_with_title ("avin cands in BB", &get_remat_bb_data (bb)->avin_cands, bb->index); lra_dump_bitmap_with_title ("avout cands in BB", &get_remat_bb_data (bb)->avout_cands, bb->index); } } /* Free all BB data. */ static void finish_remat_bb_data (void) { basic_block bb; FOR_EACH_BB_FN (bb, cfun) { bitmap_clear (&get_remat_bb_data (bb)->avout_cands); bitmap_clear (&get_remat_bb_data (bb)->avin_cands); bitmap_clear (&get_remat_bb_data (bb)->pavout_cands); bitmap_clear (&get_remat_bb_data (bb)->pavin_cands); bitmap_clear (&get_remat_bb_data (bb)->livein_cands); bitmap_clear (&get_remat_bb_data (bb)->gen_cands); bitmap_clear (&get_remat_bb_data (bb)->dead_regs); bitmap_clear (&get_remat_bb_data (bb)->changed_regs); } free (remat_bb_data); } /* Update changed_regs and dead_regs of BB from INSN. */ static void set_bb_regs (basic_block bb, rtx_insn *insn) { lra_insn_recog_data_t id = lra_get_insn_recog_data (insn); struct lra_insn_reg *reg; for (reg = id->regs; reg != NULL; reg = reg->next) if (reg->type != OP_IN) bitmap_set_bit (&get_remat_bb_data (bb)->changed_regs, reg->regno); else { if (find_regno_note (insn, REG_DEAD, (unsigned) reg->regno) != NULL) bitmap_set_bit (&get_remat_bb_data (bb)->dead_regs, reg->regno); } if (CALL_P (insn)) for (int i = 0; i < call_used_regs_arr_len; i++) bitmap_set_bit (&get_remat_bb_data (bb)->dead_regs, call_used_regs_arr[i]); } /* Calculate changed_regs and dead_regs for each BB. */ static void calculate_local_reg_remat_bb_data (void) { basic_block bb; rtx_insn *insn; FOR_EACH_BB_FN (bb, cfun) FOR_BB_INSNS (bb, insn) if (INSN_P (insn)) set_bb_regs (bb, insn); } /* Return true if REGNO is an input operand of INSN. */ static bool input_regno_present_p (rtx_insn *insn, int regno) { lra_insn_recog_data_t id = lra_get_insn_recog_data (insn); struct lra_insn_reg *reg; for (reg = id->regs; reg != NULL; reg = reg->next) if (reg->type == OP_IN && reg->regno == regno) return true; return false; } /* Return true if a call used register is an input operand of INSN. */ static bool call_used_input_regno_present_p (rtx_insn *insn) { lra_insn_recog_data_t id = lra_get_insn_recog_data (insn); struct lra_insn_reg *reg; for (reg = id->regs; reg != NULL; reg = reg->next) if (reg->type == OP_IN && reg->regno <= FIRST_PSEUDO_REGISTER && TEST_HARD_REG_BIT (call_used_reg_set, reg->regno)) return true; return false; } /* Calculate livein_cands for each BB. */ static void calculate_livein_cands (void) { basic_block bb; FOR_EACH_BB_FN (bb, cfun) { bitmap livein_regs = df_get_live_in (bb); bitmap livein_cands = &get_remat_bb_data (bb)->livein_cands; for (unsigned int i = 0; i < cands_num; i++) { cand_t cand = all_cands[i]; lra_insn_recog_data_t id = lra_get_insn_recog_data (cand->insn); struct lra_insn_reg *reg; for (reg = id->regs; reg != NULL; reg = reg->next) if (reg->type == OP_IN && ! bitmap_bit_p (livein_regs, reg->regno)) break; if (reg == NULL) bitmap_set_bit (livein_cands, i); } } } /* Calculate gen_cands for each BB. */ static void calculate_gen_cands (void) { basic_block bb; bitmap gen_cands; bitmap_head gen_insns; rtx_insn *insn; bitmap_initialize (&gen_insns, ®_obstack); FOR_EACH_BB_FN (bb, cfun) { gen_cands = &get_remat_bb_data (bb)->gen_cands; bitmap_clear (&gen_insns); FOR_BB_INSNS (bb, insn) if (INSN_P (insn)) { lra_insn_recog_data_t id = lra_get_insn_recog_data (insn); struct lra_insn_reg *reg; unsigned int uid; bitmap_iterator bi; cand_t cand; rtx set; int src_regno = -1, dst_regno = -1; if ((set = single_set (insn)) != NULL && REG_P (SET_SRC (set)) && REG_P (SET_DEST (set))) { src_regno = REGNO (SET_SRC (set)); dst_regno = REGNO (SET_DEST (set)); } /* Update gen_cands: */ bitmap_clear (&temp_bitmap); for (reg = id->regs; reg != NULL; reg = reg->next) if (reg->type != OP_IN || find_regno_note (insn, REG_DEAD, reg->regno) != NULL) EXECUTE_IF_SET_IN_BITMAP (&gen_insns, 0, uid, bi) { rtx_insn *insn2 = lra_insn_recog_data[uid]->insn; cand = insn_to_cand[INSN_UID (insn2)]; gcc_assert (cand != NULL); /* Ignore the reload insn. */ if (src_regno == cand->reload_regno && dst_regno == cand->regno) continue; if (cand->regno == reg->regno || input_regno_present_p (insn2, reg->regno)) { bitmap_clear_bit (gen_cands, cand->index); bitmap_set_bit (&temp_bitmap, uid); } } if (CALL_P (insn)) EXECUTE_IF_SET_IN_BITMAP (&gen_insns, 0, uid, bi) { rtx_insn *insn2 = lra_insn_recog_data[uid]->insn; cand = insn_to_cand[INSN_UID (insn2)]; gcc_assert (cand != NULL); if (call_used_input_regno_present_p (insn2)) { bitmap_clear_bit (gen_cands, cand->index); bitmap_set_bit (&temp_bitmap, uid); } } bitmap_and_compl_into (&gen_insns, &temp_bitmap); cand = insn_to_cand[INSN_UID (insn)]; if (cand != NULL) { bitmap_set_bit (gen_cands, cand->index); bitmap_set_bit (&gen_insns, INSN_UID (insn)); } } } bitmap_clear (&gen_insns); } /* The common transfer function used by the DF equation solver to propagate (partial) availability info BB_IN to BB_OUT through block with BB_INDEX according to the following equation: bb.out = ((bb.in & bb.livein) - bb.killed) OR bb.gen */ static bool cand_trans_fun (int bb_index, bitmap bb_in, bitmap bb_out) { remat_bb_data_t bb_info; bitmap bb_livein, bb_changed_regs, bb_dead_regs; unsigned int cid; bitmap_iterator bi; bb_info = get_remat_bb_data_by_index (bb_index); bb_livein = &bb_info->livein_cands; bb_changed_regs = &bb_info->changed_regs; bb_dead_regs = &bb_info->dead_regs; /* Calculate killed avin cands -- cands whose regs are changed or becoming dead in the BB. We calculate it here as we hope that repeated calculations are compensated by smaller size of BB_IN in comparison with all candidates number. */ bitmap_clear (&temp_bitmap); EXECUTE_IF_SET_IN_BITMAP (bb_in, 0, cid, bi) { cand_t cand = all_cands[cid]; lra_insn_recog_data_t id = lra_get_insn_recog_data (cand->insn); struct lra_insn_reg *reg; if (! bitmap_bit_p (bb_livein, cid)) { bitmap_set_bit (&temp_bitmap, cid); continue; } for (reg = id->regs; reg != NULL; reg = reg->next) /* Ignore all outputs which are not the regno for rematerialization. */ if (reg->type == OP_OUT && reg->regno != cand->regno) continue; else if (bitmap_bit_p (bb_changed_regs, reg->regno) || bitmap_bit_p (bb_dead_regs, reg->regno)) { bitmap_set_bit (&temp_bitmap, cid); break; } /* Check regno for rematerialization. */ if (bitmap_bit_p (bb_changed_regs, cand->regno) || bitmap_bit_p (bb_dead_regs, cand->regno)) bitmap_set_bit (&temp_bitmap, cid); } return bitmap_ior_and_compl (bb_out, &bb_info->gen_cands, bb_in, &temp_bitmap); } /* The transfer function used by the DF equation solver to propagate partial candidate availability info through block with BB_INDEX according to the following equation: bb.pavout = ((bb.pavin & bb.livein) - bb.killed) OR bb.gen */ static bool cand_pav_trans_fun (int bb_index) { remat_bb_data_t bb_info; bb_info = get_remat_bb_data_by_index (bb_index); return cand_trans_fun (bb_index, &bb_info->pavin_cands, &bb_info->pavout_cands); } /* The confluence function used by the DF equation solver to set up cand_pav info for a block BB without predecessor. */ static void cand_pav_con_fun_0 (basic_block bb) { bitmap_clear (&get_remat_bb_data (bb)->pavin_cands); } /* The confluence function used by the DF equation solver to propagate partial candidate availability info from predecessor to successor on edge E (pred->bb) according to the following equation: bb.pavin_cands = 0 for entry block | OR (pavout_cands of predecessors) */ static bool cand_pav_con_fun_n (edge e) { basic_block pred = e->src; basic_block bb = e->dest; remat_bb_data_t bb_info; bitmap bb_pavin, pred_pavout; bb_info = get_remat_bb_data (bb); bb_pavin = &bb_info->pavin_cands; pred_pavout = &get_remat_bb_data (pred)->pavout_cands; return bitmap_ior_into (bb_pavin, pred_pavout); } /* The transfer function used by the DF equation solver to propagate candidate availability info through block with BB_INDEX according to the following equation: bb.avout = ((bb.avin & bb.livein) - bb.killed) OR bb.gen */ static bool cand_av_trans_fun (int bb_index) { remat_bb_data_t bb_info; bb_info = get_remat_bb_data_by_index (bb_index); return cand_trans_fun (bb_index, &bb_info->avin_cands, &bb_info->avout_cands); } /* The confluence function used by the DF equation solver to set up cand_av info for a block BB without predecessor. */ static void cand_av_con_fun_0 (basic_block bb) { bitmap_clear (&get_remat_bb_data (bb)->avin_cands); } /* The confluence function used by the DF equation solver to propagate cand_av info from predecessor to successor on edge E (pred->bb) according to the following equation: bb.avin_cands = 0 for entry block | AND (avout_cands of predecessors) */ static bool cand_av_con_fun_n (edge e) { basic_block pred = e->src; basic_block bb = e->dest; remat_bb_data_t bb_info; bitmap bb_avin, pred_avout; bb_info = get_remat_bb_data (bb); bb_avin = &bb_info->avin_cands; pred_avout = &get_remat_bb_data (pred)->avout_cands; return bitmap_and_into (bb_avin, pred_avout); } /* Calculate available candidates for each BB. */ static void calculate_global_remat_bb_data (void) { basic_block bb; df_simple_dataflow (DF_FORWARD, NULL, cand_pav_con_fun_0, cand_pav_con_fun_n, cand_pav_trans_fun, &all_blocks, df_get_postorder (DF_FORWARD), df_get_n_blocks (DF_FORWARD)); /* Initialize avin by pavin. */ FOR_EACH_BB_FN (bb, cfun) bitmap_copy (&get_remat_bb_data (bb)->avin_cands, &get_remat_bb_data (bb)->pavin_cands); df_simple_dataflow (DF_FORWARD, NULL, cand_av_con_fun_0, cand_av_con_fun_n, cand_av_trans_fun, &all_blocks, df_get_postorder (DF_FORWARD), df_get_n_blocks (DF_FORWARD)); } /* Setup sp offset attribute to SP_OFFSET for all INSNS. */ static void change_sp_offset (rtx_insn *insns, HOST_WIDE_INT sp_offset) { for (rtx_insn *insn = insns; insn != NULL; insn = NEXT_INSN (insn)) eliminate_regs_in_insn (insn, false, false, sp_offset); } /* Return start hard register of REG (can be a hard or a pseudo reg) or -1 (if it is a spilled pseudo). Return number of hard registers occupied by REG through parameter NREGS if the start hard reg is not negative. */ static int get_hard_regs (struct lra_insn_reg *reg, int &nregs) { int regno = reg->regno; int hard_regno = regno < FIRST_PSEUDO_REGISTER ? regno : reg_renumber[regno]; if (hard_regno >= 0) nregs = hard_regno_nregs[hard_regno][reg->biggest_mode]; return hard_regno; } /* Insert rematerialization insns using the data-flow data calculated earlier. */ static bool do_remat (void) { rtx_insn *insn; basic_block bb; bitmap_head avail_cands; bool changed_p = false; /* Living hard regs and hard registers of living pseudos. */ HARD_REG_SET live_hard_regs; bitmap_initialize (&avail_cands, ®_obstack); FOR_EACH_BB_FN (bb, cfun) { REG_SET_TO_HARD_REG_SET (live_hard_regs, df_get_live_out (bb)); bitmap_and (&avail_cands, &get_remat_bb_data (bb)->avin_cands, &get_remat_bb_data (bb)->livein_cands); FOR_BB_INSNS (bb, insn) { if (!NONDEBUG_INSN_P (insn)) continue; lra_insn_recog_data_t id = lra_get_insn_recog_data (insn); struct lra_static_insn_data *static_id = id->insn_static_data; struct lra_insn_reg *reg; cand_t cand; unsigned int cid; bitmap_iterator bi; rtx set; int src_regno = -1, dst_regno = -1; if ((set = single_set (insn)) != NULL && REG_P (SET_SRC (set)) && REG_P (SET_DEST (set))) { src_regno = REGNO (SET_SRC (set)); dst_regno = REGNO (SET_DEST (set)); } cand = NULL; /* Check possibility of rematerialization (hard reg or unpsilled pseudo <- spilled pseudo): */ if (dst_regno >= 0 && src_regno >= FIRST_PSEUDO_REGISTER && reg_renumber[src_regno] < 0 && (dst_regno < FIRST_PSEUDO_REGISTER || reg_renumber[dst_regno] >= 0)) { for (cand = regno_cands[src_regno]; cand != NULL; cand = cand->next_regno_cand) if (bitmap_bit_p (&avail_cands, cand->index)) break; } int i, hard_regno, nregs; rtx_insn *remat_insn = NULL; HOST_WIDE_INT cand_sp_offset = 0; if (cand != NULL) { lra_insn_recog_data_t cand_id = lra_get_insn_recog_data (cand->insn); struct lra_static_insn_data *static_cand_id = cand_id->insn_static_data; rtx saved_op = *cand_id->operand_loc[cand->nop]; /* Check clobbers do not kill something living. */ gcc_assert (REG_P (saved_op)); int ignore_regno = REGNO (saved_op); for (reg = cand_id->regs; reg != NULL; reg = reg->next) if (reg->type != OP_IN && reg->regno != ignore_regno) { hard_regno = get_hard_regs (reg, nregs); gcc_assert (hard_regno >= 0); for (i = 0; i < nregs; i++) if (TEST_HARD_REG_BIT (live_hard_regs, hard_regno + i)) break; if (i < nregs) break; } if (reg == NULL) { for (reg = static_cand_id->hard_regs; reg != NULL; reg = reg->next) if (reg->type != OP_IN && TEST_HARD_REG_BIT (live_hard_regs, reg->regno)) break; } if (reg == NULL) { *cand_id->operand_loc[cand->nop] = SET_DEST (set); lra_update_insn_regno_info (cand->insn); bool ok_p = lra_constrain_insn (cand->insn); if (ok_p) { rtx remat_pat = copy_insn (PATTERN (cand->insn)); start_sequence (); emit_insn (remat_pat); remat_insn = get_insns (); end_sequence (); if (recog_memoized (remat_insn) < 0) remat_insn = NULL; cand_sp_offset = cand_id->sp_offset; } *cand_id->operand_loc[cand->nop] = saved_op; lra_update_insn_regno_info (cand->insn); } } bitmap_clear (&temp_bitmap); /* Update avail_cands (see analogous code for calculate_gen_cands). */ for (reg = id->regs; reg != NULL; reg = reg->next) if (reg->type != OP_IN || find_regno_note (insn, REG_DEAD, reg->regno) != NULL) EXECUTE_IF_SET_IN_BITMAP (&avail_cands, 0, cid, bi) { cand = all_cands[cid]; /* Ignore the reload insn. */ if (src_regno == cand->reload_regno && dst_regno == cand->regno) continue; if (cand->regno == reg->regno || input_regno_present_p (cand->insn, reg->regno)) bitmap_set_bit (&temp_bitmap, cand->index); } if (CALL_P (insn)) EXECUTE_IF_SET_IN_BITMAP (&avail_cands, 0, cid, bi) { cand = all_cands[cid]; if (call_used_input_regno_present_p (cand->insn)) bitmap_set_bit (&temp_bitmap, cand->index); } bitmap_and_compl_into (&avail_cands, &temp_bitmap); if ((cand = insn_to_cand[INSN_UID (insn)]) != NULL) bitmap_set_bit (&avail_cands, cand->index); if (remat_insn != NULL) { HOST_WIDE_INT sp_offset_change = cand_sp_offset - id->sp_offset; if (sp_offset_change != 0) change_sp_offset (remat_insn, sp_offset_change); lra_process_new_insns (insn, remat_insn, NULL, "Inserting rematerialization insn"); lra_set_insn_deleted (insn); changed_p = true; continue; } /* Update live hard regs: */ for (reg = id->regs; reg != NULL; reg = reg->next) if (reg->type == OP_IN && find_regno_note (insn, REG_DEAD, reg->regno) != NULL) { if ((hard_regno = get_hard_regs (reg, nregs)) < 0) continue; for (i = 0; i < nregs; i++) CLEAR_HARD_REG_BIT (live_hard_regs, hard_regno + i); } else if (reg->type != OP_IN && find_regno_note (insn, REG_UNUSED, reg->regno) == NULL) { if ((hard_regno = get_hard_regs (reg, nregs)) < 0) continue; for (i = 0; i < nregs; i++) SET_HARD_REG_BIT (live_hard_regs, hard_regno + i); } /* Process also hard regs (e.g. CC register) which are part of insn definition. */ for (reg = static_id->hard_regs; reg != NULL; reg = reg->next) if (reg->type == OP_IN && find_regno_note (insn, REG_DEAD, reg->regno) != NULL) CLEAR_HARD_REG_BIT (live_hard_regs, reg->regno); else if (reg->type != OP_IN && find_regno_note (insn, REG_UNUSED, reg->regno) == NULL) SET_HARD_REG_BIT (live_hard_regs, reg->regno); } } bitmap_clear (&avail_cands); return changed_p; } /* Entry point of the rematerialization sub-pass. Return true if we did any rematerialization. */ bool lra_remat (void) { basic_block bb; bool result; int max_regno = max_reg_num (); if (! flag_lra_remat) return false; timevar_push (TV_LRA_REMAT); insn_to_cand = XCNEWVEC (cand_t, get_max_uid ()); regno_cands = XCNEWVEC (cand_t, max_regno); all_cands.create (8000); call_used_regs_arr_len = 0; for (int i = 0; i < FIRST_PSEUDO_REGISTER; i++) if (call_used_regs[i]) call_used_regs_arr[call_used_regs_arr_len++] = i; initiate_cand_table (); create_cands (); create_remat_bb_data (); bitmap_initialize (&temp_bitmap, ®_obstack); calculate_local_reg_remat_bb_data (); calculate_livein_cands (); calculate_gen_cands (); bitmap_initialize (&all_blocks, ®_obstack); FOR_ALL_BB_FN (bb, cfun) bitmap_set_bit (&all_blocks, bb->index); calculate_global_remat_bb_data (); dump_candidates_and_remat_bb_data (); result = do_remat (); all_cands.release (); bitmap_clear (&temp_bitmap); finish_remat_bb_data (); finish_cand_table (); bitmap_clear (&all_blocks); free (regno_cands); free (insn_to_cand); timevar_pop (TV_LRA_REMAT); return result; }