diff options
-rw-r--r-- | gcc/ChangeLog | 14 | ||||
-rw-r--r-- | gcc/basic-block.h | 14 | ||||
-rw-r--r-- | gcc/gimple-streamer-in.c | 4 | ||||
-rw-r--r-- | gcc/lto-cgraph.c | 16 | ||||
-rw-r--r-- | gcc/lto-streamer-in.c | 4 | ||||
-rw-r--r-- | gcc/tree-inline.c | 6 | ||||
-rw-r--r-- | gcc/tree-optimize.c | 10 |
7 files changed, 46 insertions, 22 deletions
diff --git a/gcc/ChangeLog b/gcc/ChangeLog index 34dcd586384..1529dd37f14 100644 --- a/gcc/ChangeLog +++ b/gcc/ChangeLog @@ -1,3 +1,17 @@ +2013-04-29 Teresa Johnson <tejohnson@google.com> + + PR bootstrap/57077 + * basic-block.h (apply_scale): New function. + (apply_probability): Use apply_scale. + * gimple-streamer-in.c (input_bb): Ditto. + * lto-streamer-in.c (input_cfg): Ditto. + * lto-cgraph.c (merge_profile_summaries): Ditto. + * tree-optimize.c (execute_fixup_cfg): Ditto. + * tree-inline.c (copy_bb): Update comment to use + apply_scale. + (copy_edges_for_bb): Ditto. + (copy_cfg_body): Ditto. + 2013-04-29 Tom de Vries <tom@codesourcery.com> * tree-ssa-tail-merge.c (find_same_succ_bb): Skip loop latch bbs. diff --git a/gcc/basic-block.h b/gcc/basic-block.h index 9b5192eb877..eed320c9ea0 100644 --- a/gcc/basic-block.h +++ b/gcc/basic-block.h @@ -500,7 +500,7 @@ struct edge_list REG_BR_PROB_BASE) /* Compute a scale factor (or probability) suitable for scaling of - gcov_type values via apply_probability(). */ + gcov_type values via apply_probability() and apply_scale(). */ #define GCOV_COMPUTE_SCALE(num,den) \ ((den) ? RDIV ((num) * REG_BR_PROB_BASE, (den)) : REG_BR_PROB_BASE) @@ -952,13 +952,23 @@ combine_probabilities (int prob1, int prob2) return RDIV (prob1 * prob2, REG_BR_PROB_BASE); } +/* Apply scale factor SCALE on frequency or count FREQ. Use this + interface when potentially scaling up, so that SCALE is not + constrained to be < REG_BR_PROB_BASE. */ + +static inline gcov_type +apply_scale (gcov_type freq, int scale) +{ + return RDIV (freq * scale, REG_BR_PROB_BASE); +} + /* Apply probability PROB on frequency or count FREQ. */ static inline gcov_type apply_probability (gcov_type freq, int prob) { check_probability (prob); - return RDIV (freq * prob, REG_BR_PROB_BASE); + return apply_scale (freq, prob); } /* Return inverse probability for PROB. */ diff --git a/gcc/gimple-streamer-in.c b/gcc/gimple-streamer-in.c index a27f0d6d62f..03fbe91bbe2 100644 --- a/gcc/gimple-streamer-in.c +++ b/gcc/gimple-streamer-in.c @@ -329,8 +329,8 @@ input_bb (struct lto_input_block *ib, enum LTO_tags tag, index = streamer_read_uhwi (ib); bb = BASIC_BLOCK_FOR_FUNCTION (fn, index); - bb->count = apply_probability (streamer_read_gcov_count (ib), - count_materialization_scale); + bb->count = apply_scale (streamer_read_gcov_count (ib), + count_materialization_scale); bb->frequency = streamer_read_hwi (ib); bb->flags = streamer_read_hwi (ib); diff --git a/gcc/lto-cgraph.c b/gcc/lto-cgraph.c index 69f5e3a659f..cead76b0cd2 100644 --- a/gcc/lto-cgraph.c +++ b/gcc/lto-cgraph.c @@ -1347,10 +1347,10 @@ merge_profile_summaries (struct lto_file_decl_data **file_data_vec) file_data->profile_info.runs); lto_gcov_summary.sum_max = MAX (lto_gcov_summary.sum_max, - apply_probability (file_data->profile_info.sum_max, scale)); + apply_scale (file_data->profile_info.sum_max, scale)); lto_gcov_summary.sum_all = MAX (lto_gcov_summary.sum_all, - apply_probability (file_data->profile_info.sum_all, scale)); + apply_scale (file_data->profile_info.sum_all, scale)); /* Save a pointer to the profile_info with the largest scaled sum_all and the scale for use in merging the histogram. */ @@ -1372,8 +1372,8 @@ merge_profile_summaries (struct lto_file_decl_data **file_data_vec) /* Scale up the min value as we did the corresponding sum_all above. Use that to find the new histogram index. */ gcov_type scaled_min - = apply_probability (saved_profile_info->histogram[h_ix].min_value, - saved_scale); + = apply_scale (saved_profile_info->histogram[h_ix].min_value, + saved_scale); /* The new index may be shared with another scaled histogram entry, so we need to account for a non-zero histogram entry at new_ix. */ unsigned new_ix = gcov_histo_index (scaled_min); @@ -1386,8 +1386,8 @@ merge_profile_summaries (struct lto_file_decl_data **file_data_vec) here and place the scaled cumulative counter value in the bucket corresponding to the scaled minimum counter value. */ lto_gcov_summary.histogram[new_ix].cum_value - += apply_probability (saved_profile_info->histogram[h_ix].cum_value, - saved_scale); + += apply_scale (saved_profile_info->histogram[h_ix].cum_value, + saved_scale); lto_gcov_summary.histogram[new_ix].num_counters += saved_profile_info->histogram[h_ix].num_counters; } @@ -1419,8 +1419,8 @@ merge_profile_summaries (struct lto_file_decl_data **file_data_vec) if (scale == REG_BR_PROB_BASE) continue; for (edge = node->callees; edge; edge = edge->next_callee) - edge->count = apply_probability (edge->count, scale); - node->count = apply_probability (node->count, scale); + edge->count = apply_scale (edge->count, scale); + node->count = apply_scale (node->count, scale); } } diff --git a/gcc/lto-streamer-in.c b/gcc/lto-streamer-in.c index f5789c01277..87158bf0b33 100644 --- a/gcc/lto-streamer-in.c +++ b/gcc/lto-streamer-in.c @@ -635,8 +635,8 @@ input_cfg (struct lto_input_block *ib, struct function *fn, dest_index = streamer_read_uhwi (ib); probability = (int) streamer_read_hwi (ib); - count = apply_probability ((gcov_type) streamer_read_gcov_count (ib), - count_materialization_scale); + count = apply_scale ((gcov_type) streamer_read_gcov_count (ib), + count_materialization_scale); edge_flags = streamer_read_uhwi (ib); dest = BASIC_BLOCK_FOR_FUNCTION (fn, dest_index); diff --git a/gcc/tree-inline.c b/gcc/tree-inline.c index c358cbef2a7..69455cebf79 100644 --- a/gcc/tree-inline.c +++ b/gcc/tree-inline.c @@ -1519,7 +1519,7 @@ copy_bb (copy_body_data *id, basic_block bb, int frequency_scale, basic_block_info automatically. */ copy_basic_block = create_basic_block (NULL, (void *) 0, (basic_block) prev->aux); - /* Update to use apply_probability(). */ + /* Update to use apply_scale(). */ copy_basic_block->count = bb->count * count_scale / REG_BR_PROB_BASE; /* We are going to rebuild frequencies from scratch. These values @@ -1891,7 +1891,7 @@ copy_edges_for_bb (basic_block bb, gcov_type count_scale, basic_block ret_bb, && old_edge->dest->aux != EXIT_BLOCK_PTR) flags |= EDGE_FALLTHRU; new_edge = make_edge (new_bb, (basic_block) old_edge->dest->aux, flags); - /* Update to use apply_probability(). */ + /* Update to use apply_scale(). */ new_edge->count = old_edge->count * count_scale / REG_BR_PROB_BASE; new_edge->probability = old_edge->probability; } @@ -2278,7 +2278,7 @@ copy_cfg_body (copy_body_data * id, gcov_type count, int frequency_scale, incoming_frequency += EDGE_FREQUENCY (e); incoming_count += e->count; } - /* Update to use apply_probability(). */ + /* Update to use apply_scale(). */ incoming_count = incoming_count * count_scale / REG_BR_PROB_BASE; /* Update to use EDGE_FREQUENCY. */ incoming_frequency diff --git a/gcc/tree-optimize.c b/gcc/tree-optimize.c index d399d8a98c6..0948d2e6977 100644 --- a/gcc/tree-optimize.c +++ b/gcc/tree-optimize.c @@ -131,15 +131,15 @@ execute_fixup_cfg (void) ENTRY_BLOCK_PTR->count); ENTRY_BLOCK_PTR->count = cgraph_get_node (current_function_decl)->count; - EXIT_BLOCK_PTR->count = apply_probability (EXIT_BLOCK_PTR->count, - count_scale); + EXIT_BLOCK_PTR->count = apply_scale (EXIT_BLOCK_PTR->count, + count_scale); FOR_EACH_EDGE (e, ei, ENTRY_BLOCK_PTR->succs) - e->count = apply_probability (e->count, count_scale); + e->count = apply_scale (e->count, count_scale); FOR_EACH_BB (bb) { - bb->count = apply_probability (bb->count, count_scale); + bb->count = apply_scale (bb->count, count_scale); for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) { gimple stmt = gsi_stmt (gsi); @@ -172,7 +172,7 @@ execute_fixup_cfg (void) } FOR_EACH_EDGE (e, ei, bb->succs) - e->count = apply_probability (e->count, count_scale); + e->count = apply_scale (e->count, count_scale); /* If we have a basic block with no successors that does not end with a control statement or a noreturn call end it with |