diff options
-rw-r--r-- | gcc/ChangeLog | 40 | ||||
-rw-r--r-- | gcc/cgraph.c | 17 | ||||
-rw-r--r-- | gcc/cgraph.h | 13 | ||||
-rw-r--r-- | gcc/doc/invoke.texi | 15 | ||||
-rw-r--r-- | gcc/except.c | 37 | ||||
-rw-r--r-- | gcc/except.h | 1 | ||||
-rw-r--r-- | gcc/ipa-cp.c | 13 | ||||
-rw-r--r-- | gcc/ipa-inline.c | 375 | ||||
-rw-r--r-- | gcc/params.def | 10 | ||||
-rw-r--r-- | gcc/tree-inline.c | 6 | ||||
-rw-r--r-- | gcc/tree-ssa-structalias.c | 2 |
11 files changed, 399 insertions, 130 deletions
diff --git a/gcc/ChangeLog b/gcc/ChangeLog index 8caac7e09b2..39cd78d0947 100644 --- a/gcc/ChangeLog +++ b/gcc/ChangeLog @@ -1,3 +1,43 @@ +2009-05-15 Jan Hubicka <jh@suse.cz> + + * cgraph.c (dump_cgraph_node): Dump size/time/benefit. + * cgraph.h (struct inline_summary): New filed self_wize, + size_inlining_benefit, self_time and time_inlining_benefit. + (struct cgraph_global_info): Replace insns by time ans size fields. + * ipa-cp (ipcp_cloning_candidate_p): Base estimate on size + (ipcp_estimate_growth, ipcp_insert_stage): Likewise. + (ipcp_update_callgraph): Do not touch function bodies. + * ipa-inline.c: Include except.h + MAX_TIME: New constant. + (overall_insns): Remove + (overall_size, max_benefit): New static variables. + (cgraph_estimate_time_after_inlining): New function. + (cgraph_estimate_size_after_inlining): Rewrite using benefits. + (cgraph_clone_inlined_nodes): Update size. + (cgraph_mark_inline_edge): Update size. + (cgraph_estimate_growth): Use size info. + (cgraph_check_inline_limits): Check size. + (cgraph_default_inline_p): Likewise. + (cgraph_edge_badness): Compute badness based on benefit and size cost. + (cgraph_decide_recursive_inlining): Check size. + (cgraph_decide_inlining_of_small_function): Update size; dump sizes and times. + (cgraph_decide_inlining): Likewise. + (cgraph_decide_inlining_incrementally): Likewise; honor PARAM_EARLY_INLINING_INSNS. + (likely_eliminated_by_inlining_p): New predicate. + (estimate_function_body_sizes): New function. + (compute_inline_parameters): Use it. + * except.c (must_not_throw_labels): New function. + * except.h (must_not_throw_labels): Declare. + * tree-inline.c (init_inline_once): Kill inlining_weigths + * tree-ssa-structalias.c: Avoid uninitialized warning. + * params.def (PARAM_MAX_INLINE_INSNS_SINGLE): Reduce to 300. + (PARAM_MAX_INLINE_INSNS_AUTO): Reduce to 60. + (PARAM_INLINE_CALL_COST): Remove. + (PARAM_EARLY_INLINING_INSNS): New. + doc/invoke.texi (max-inline-insns-auto, early-inlining-insns): Update. + (inline-call-cost): Remove. + (early-inlining-insns): New. + 2009-05-15 Eric Botcazou <ebotcazou@adacore.com> * dbxout.c (dbxout_range_type): Add LOW and HIGH parameters. Use them diff --git a/gcc/cgraph.c b/gcc/cgraph.c index 60ff1685dc9..c7939b9d2c4 100644 --- a/gcc/cgraph.c +++ b/gcc/cgraph.c @@ -1393,11 +1393,18 @@ dump_cgraph_node (FILE *f, struct cgraph_node *node) if (node->count) fprintf (f, " executed "HOST_WIDEST_INT_PRINT_DEC"x", (HOST_WIDEST_INT)node->count); - if (node->local.inline_summary.self_insns) - fprintf (f, " %i insns", node->local.inline_summary.self_insns); - if (node->global.insns && node->global.insns - != node->local.inline_summary.self_insns) - fprintf (f, " (%i after inlining)", node->global.insns); + if (node->local.inline_summary.self_time) + fprintf (f, " %i time, %i benefit", node->local.inline_summary.self_time, + node->local.inline_summary.time_inlining_benefit); + if (node->global.time && node->global.time + != node->local.inline_summary.self_time) + fprintf (f, " (%i after inlining)", node->global.time); + if (node->local.inline_summary.self_size) + fprintf (f, " %i size, %i benefit", node->local.inline_summary.self_size, + node->local.inline_summary.size_inlining_benefit); + if (node->global.size && node->global.size + != node->local.inline_summary.self_size) + fprintf (f, " (%i after inlining)", node->global.size); if (node->local.inline_summary.estimated_self_stack_size) fprintf (f, " %i bytes stack usage", (int)node->local.inline_summary.estimated_self_stack_size); if (node->global.estimated_stack_size != node->local.inline_summary.estimated_self_stack_size) diff --git a/gcc/cgraph.h b/gcc/cgraph.h index 58ae06852cf..d4a099539a0 100644 --- a/gcc/cgraph.h +++ b/gcc/cgraph.h @@ -55,8 +55,14 @@ struct GTY(()) inline_summary /* Estimated stack frame consumption by the function. */ HOST_WIDE_INT estimated_self_stack_size; - /* Size of the function before inlining. */ - int self_insns; + /* Size of the function body. */ + int self_size; + /* How many instructions are likely going to disappear after inlining. */ + int size_inlining_benefit; + /* Estimated time spent executing the function body. */ + int self_time; + /* How much time is going to be saved by inlining. */ + int time_inlining_benefit; }; /* Information about the function collected locally. @@ -108,7 +114,8 @@ struct GTY(()) cgraph_global_info { struct cgraph_node *inlined_to; /* Estimated size of the function after inlining. */ - int insns; + int time; + int size; /* Estimated growth after inlining. INT_MIN if not computed. */ int estimated_growth; diff --git a/gcc/doc/invoke.texi b/gcc/doc/invoke.texi index da5ea9076c6..3dbb8c570ae 100644 --- a/gcc/doc/invoke.texi +++ b/gcc/doc/invoke.texi @@ -7375,7 +7375,7 @@ This number sets the maximum number of instructions (counted in GCC's internal representation) in a single function that the tree inliner will consider for inlining. This only affects functions declared inline and methods implemented in a class declaration (C++). -The default value is 450. +The default value is 300. @item max-inline-insns-auto When you use @option{-finline-functions} (included in @option{-O3}), @@ -7383,7 +7383,7 @@ a lot of functions that would otherwise not be considered for inlining by the compiler will be investigated. To those functions, a different (more restrictive) limit compared to functions declared inline can be applied. -The default value is 90. +The default value is 60. @item large-function-insns The limit specifying really large functions. For functions larger than this @@ -7461,14 +7461,9 @@ given call expression. This parameter limits inlining only to call expression whose probability exceeds given threshold (in percents). The default value is 10. -@item inline-call-cost -Specify cost of call instruction relative to simple arithmetics operations -(having cost of 1). Increasing this cost disqualifies inlining of non-leaf -functions and at the same time increases size of leaf function that is believed to -reduce function size by being inlined. In effect it increases amount of -inlining for code having large abstraction penalty (many functions that just -pass the arguments to other functions) and decrease inlining for code with low -abstraction penalty. The default value is 12. +@item early-inlining-insns +Specify growth that early inliner can make. In effect it increases amount of +inlining for code having large abstraction penalty. The default value is 12. @item min-vect-loop-bound The minimum number of iterations under which a loop will not get vectorized diff --git a/gcc/except.c b/gcc/except.c index a33f3096fbf..68276e0c922 100644 --- a/gcc/except.c +++ b/gcc/except.c @@ -1039,6 +1039,43 @@ get_next_region_sharing_label (int region) return r->next_region_sharing_label->region_number; } +/* Return bitmap of all labels that are handlers of must not throw regions. */ + +bitmap +must_not_throw_labels (void) +{ + struct eh_region *i; + bitmap labels = BITMAP_ALLOC (NULL); + + i = cfun->eh->region_tree; + if (! i) + return labels; + + while (1) + { + if (i->type == ERT_MUST_NOT_THROW && i->tree_label + && LABEL_DECL_UID (i->tree_label) >= 0) + bitmap_set_bit (labels, LABEL_DECL_UID (i->tree_label)); + + /* If there are sub-regions, process them. */ + if (i->inner) + i = i->inner; + /* If there are peers, process them. */ + else if (i->next_peer) + i = i->next_peer; + /* Otherwise, step back up the tree to the next peer. */ + else + { + do { + i = i->outer; + if (i == NULL) + return labels; + } while (i->next_peer == NULL); + i = i->next_peer; + } + } +} + /* Set up EH labels for RTL. */ void diff --git a/gcc/except.h b/gcc/except.h index 70506d3a8c5..cf0bba91728 100644 --- a/gcc/except.h +++ b/gcc/except.h @@ -274,5 +274,6 @@ extern void set_eh_throw_stmt_table (struct function *, struct htab *); extern void remove_unreachable_regions (sbitmap, sbitmap); extern VEC(int,heap) * label_to_region_map (void); extern int num_eh_regions (void); +extern bitmap must_not_throw_labels (void); extern struct eh_region *redirect_eh_edge_to_label (struct edge_def *, tree, bool, bool, int); extern int get_next_region_sharing_label (int); diff --git a/gcc/ipa-cp.c b/gcc/ipa-cp.c index 7761aab92d6..11084dc5ec0 100644 --- a/gcc/ipa-cp.c +++ b/gcc/ipa-cp.c @@ -396,7 +396,7 @@ ipcp_cloning_candidate_p (struct cgraph_node *node) cgraph_node_name (node)); return false; } - if (node->local.inline_summary.self_insns < n_calls) + if (node->local.inline_summary.self_size < n_calls) { if (dump_file) fprintf (dump_file, "Considering %s for cloning; code would shrink.\n", @@ -837,10 +837,7 @@ ipcp_update_callgraph (void) { next = cs->next_caller; if (!ipcp_node_is_clone (cs->caller) && ipcp_need_redirect_p (cs)) - { - cgraph_redirect_edge_callee (cs, orig_node); - gimple_call_set_fndecl (cs->call_stmt, orig_node->decl); - } + cgraph_redirect_edge_callee (cs, orig_node); } } } @@ -916,7 +913,7 @@ ipcp_estimate_growth (struct cgraph_node *node) call site. Precise cost is dificult to get, as our size metric counts constants and moves as free. Generally we are looking for cases that small function is called very many times. */ - growth = node->local.inline_summary.self_insns + growth = node->local.inline_summary.self_size - removable_args * redirectable_node_callers; if (growth < 0) return 0; @@ -956,7 +953,7 @@ ipcp_estimate_cloning_cost (struct cgraph_node *node) cost /= freq_sum * 1000 / REG_BR_PROB_BASE + 1; if (dump_file) fprintf (dump_file, "Cost of versioning %s is %i, (size: %i, freq: %i)\n", - cgraph_node_name (node), cost, node->local.inline_summary.self_insns, + cgraph_node_name (node), cost, node->local.inline_summary.self_size, freq_sum); return cost + 1; } @@ -1012,7 +1009,7 @@ ipcp_insert_stage (void) { if (node->count > max_count) max_count = node->count; - overall_size += node->local.inline_summary.self_insns; + overall_size += node->local.inline_summary.self_size; } max_new_size = overall_size; diff --git a/gcc/ipa-inline.c b/gcc/ipa-inline.c index 99640bfdb55..17e050197a6 100644 --- a/gcc/ipa-inline.c +++ b/gcc/ipa-inline.c @@ -138,6 +138,9 @@ along with GCC; see the file COPYING3. If not see #include "tree-flow.h" #include "rtl.h" #include "ipa-prop.h" +#include "except.h" + +#define MAX_TIME 1000000000 /* Mode incremental inliner operate on: @@ -163,8 +166,8 @@ cgraph_decide_inlining_incrementally (struct cgraph_node *, enum inlining_mode, /* Statistics we collect about inlining algorithm. */ static int ncalls_inlined; static int nfunctions_inlined; -static int overall_insns; -static gcov_type max_count; +static int overall_size; +static gcov_type max_count, max_benefit; /* Holders of ipa cgraph hooks: */ static struct cgraph_node_hook_list *function_insertion_hook_holder; @@ -175,19 +178,30 @@ inline_summary (struct cgraph_node *node) return &node->local.inline_summary; } -/* Estimate size of the function after inlining WHAT into TO. */ +/* Estimate self time of the function after inlining WHAT into TO. */ static int -cgraph_estimate_size_after_inlining (int times, struct cgraph_node *to, +cgraph_estimate_time_after_inlining (int frequency, struct cgraph_node *to, struct cgraph_node *what) { - int size; - tree fndecl = what->decl, arg; - int call_insns = PARAM_VALUE (PARAM_INLINE_CALL_COST); + gcov_type time = (((gcov_type)what->global.time - inline_summary + (what)->time_inlining_benefit) + * frequency + CGRAPH_FREQ_BASE / 2) / CGRAPH_FREQ_BASE + + to->global.time; + if (time < 0) + time = 0; + if (time > MAX_TIME) + time = MAX_TIME; + return time; +} + +/* Estimate self time of the function after inlining WHAT into TO. */ - for (arg = DECL_ARGUMENTS (fndecl); arg; arg = TREE_CHAIN (arg)) - call_insns += estimate_move_cost (TREE_TYPE (arg)); - size = (what->global.insns - call_insns) * times + to->global.insns; +static int +cgraph_estimate_size_after_inlining (int times, struct cgraph_node *to, + struct cgraph_node *what) +{ + int size = (what->global.size - inline_summary (what)->size_inlining_benefit) * times + to->global.size; gcc_assert (size >= 0); return size; } @@ -213,7 +227,10 @@ cgraph_clone_inlined_nodes (struct cgraph_edge *e, bool duplicate, { gcc_assert (!e->callee->global.inlined_to); if (e->callee->analyzed) - overall_insns -= e->callee->global.insns, nfunctions_inlined++; + { + overall_size -= e->callee->global.size; + nfunctions_inlined++; + } duplicate = false; } else @@ -253,7 +270,7 @@ static bool cgraph_mark_inline_edge (struct cgraph_edge *e, bool update_original, VEC (cgraph_edge_p, heap) **new_edges) { - int old_insns = 0, new_insns = 0; + int old_size = 0, new_size = 0; struct cgraph_node *to = NULL, *what; struct cgraph_edge *curr = e; @@ -274,16 +291,15 @@ cgraph_mark_inline_edge (struct cgraph_edge *e, bool update_original, /* Now update size of caller and all functions caller is inlined into. */ for (;e && !e->inline_failed; e = e->caller->callers) { - old_insns = e->caller->global.insns; - new_insns = cgraph_estimate_size_after_inlining (1, e->caller, - what); - gcc_assert (new_insns >= 0); to = e->caller; - to->global.insns = new_insns; + old_size = e->caller->global.size; + new_size = cgraph_estimate_size_after_inlining (1, to, what); + to->global.size = new_size; + to->global.time = cgraph_estimate_time_after_inlining (e->frequency, to, what); } gcc_assert (what->global.inlined_to == to); - if (new_insns > old_insns) - overall_insns += new_insns - old_insns; + if (new_size > old_size) + overall_size += new_size - old_size; ncalls_inlined++; if (flag_indirect_inlining) @@ -338,7 +354,7 @@ cgraph_estimate_growth (struct cgraph_node *node) self_recursive = true; if (e->inline_failed) growth += (cgraph_estimate_size_after_inlining (1, e->caller, node) - - e->caller->global.insns); + - e->caller->global.size); } /* ??? Wrong for non-trivially self recursive functions or cases where @@ -346,7 +362,7 @@ cgraph_estimate_growth (struct cgraph_node *node) as in that case we will keep the body around, but we will also avoid some inlining. */ if (!node->needed && !DECL_EXTERNAL (node->decl) && !self_recursive) - growth -= node->global.insns; + growth -= node->global.size; node->global.estimated_growth = growth; return growth; @@ -381,17 +397,17 @@ cgraph_check_inline_limits (struct cgraph_node *to, struct cgraph_node *what, /* When inlining large function body called once into small function, take the inlined function as base for limiting the growth. */ - if (inline_summary (to)->self_insns > inline_summary(what)->self_insns) - limit = inline_summary (to)->self_insns; + if (inline_summary (to)->self_size > inline_summary(what)->self_size) + limit = inline_summary (to)->self_size; else - limit = inline_summary (what)->self_insns; + limit = inline_summary (what)->self_size; limit += limit * PARAM_VALUE (PARAM_LARGE_FUNCTION_GROWTH) / 100; /* Check the size after inlining against the function limits. But allow the function to shrink if it went over the limits by forced inlining. */ newsize = cgraph_estimate_size_after_inlining (times, to, what); - if (newsize >= to->global.insns + if (newsize >= to->global.size && newsize > PARAM_VALUE (PARAM_LARGE_FUNCTION_INSNS) && newsize > limit) { @@ -442,7 +458,7 @@ cgraph_default_inline_p (struct cgraph_node *n, cgraph_inline_failed_t *reason) if (DECL_DECLARED_INLINE_P (decl)) { - if (n->global.insns >= MAX_INLINE_INSNS_SINGLE) + if (n->global.size >= MAX_INLINE_INSNS_SINGLE) { if (reason) *reason = CIF_MAX_INLINE_INSNS_SINGLE_LIMIT; @@ -451,7 +467,7 @@ cgraph_default_inline_p (struct cgraph_node *n, cgraph_inline_failed_t *reason) } else { - if (n->global.insns >= MAX_INLINE_INSNS_AUTO) + if (n->global.size >= MAX_INLINE_INSNS_AUTO) { if (reason) *reason = CIF_MAX_INLINE_INSNS_AUTO_LIMIT; @@ -497,7 +513,7 @@ cgraph_edge_badness (struct cgraph_edge *edge) int growth = cgraph_estimate_size_after_inlining (1, edge->caller, edge->callee); - growth -= edge->caller->global.insns; + growth -= edge->caller->global.size; /* Always prefer inlining saving code size. */ if (growth <= 0) @@ -506,7 +522,8 @@ cgraph_edge_badness (struct cgraph_edge *edge) /* When profiling is available, base priorities -(#calls / growth). So we optimize for overall number of "executed" inlined calls. */ else if (max_count) - badness = ((int)((double)edge->count * INT_MIN / max_count)) / growth; + badness = ((int)((double)edge->count * INT_MIN / max_count / (max_benefit + 1)) + * (inline_summary (edge->callee)->time_inlining_benefit + 1)) / growth; /* When function local profile is available, base priorities on growth / frequency, so we optimize for overall frequency of inlined @@ -519,11 +536,11 @@ cgraph_edge_badness (struct cgraph_edge *edge) of the same size gets priority). */ else if (flag_guess_branch_prob) { - int div = edge->frequency * 100 / CGRAPH_FREQ_BASE; - int growth = - cgraph_estimate_size_after_inlining (1, edge->caller, edge->callee); - growth -= edge->caller->global.insns; + int div = edge->frequency * 100 / CGRAPH_FREQ_BASE + 1; badness = growth * 256; + div *= MIN (100 * inline_summary (edge->callee)->time_inlining_benefit + / (edge->callee->global.time + 1) + 1, 100); + /* Decrease badness if call is nested. */ /* Compress the range so we don't overflow. */ @@ -766,8 +783,9 @@ cgraph_decide_recursive_inlining (struct cgraph_node *node, fibheap_delete (heap); if (dump_file) fprintf (dump_file, - "\n Inlined %i times, body grown from %i to %i insns\n", n, - master_clone->global.insns, node->global.insns); + "\n Inlined %i times, body grown from size %i to %i, time %i to %i\n", n, + master_clone->global.size, node->global.size, + master_clone->global.time, node->global.time); /* Remove master clone we used for inlining. We rely that clones inlined into master clone gets queued just before master clone so we don't @@ -845,7 +863,7 @@ cgraph_decide_inlining_of_small_functions (void) cgraph_inline_failed_t failed_reason; fibheap_t heap = fibheap_new (); bitmap updated_nodes = BITMAP_ALLOC (NULL); - int min_insns, max_insns; + int min_size, max_size; VEC (cgraph_edge_p, heap) *new_indirect_edges = NULL; if (flag_indirect_inlining) @@ -879,26 +897,26 @@ cgraph_decide_inlining_of_small_functions (void) } } - max_insns = compute_max_insns (overall_insns); - min_insns = overall_insns; + max_size = compute_max_insns (overall_size); + min_size = overall_size; - while (overall_insns <= max_insns + while (overall_size <= max_size && (edge = (struct cgraph_edge *) fibheap_extract_min (heap))) { - int old_insns = overall_insns; + int old_size = overall_size; struct cgraph_node *where; int growth = cgraph_estimate_size_after_inlining (1, edge->caller, edge->callee); cgraph_inline_failed_t not_good = CIF_OK; - growth -= edge->caller->global.insns; + growth -= edge->caller->global.size; if (dump_file) { fprintf (dump_file, - "\nConsidering %s with %i insns\n", + "\nConsidering %s with %i size\n", cgraph_node_name (edge->callee), - edge->callee->global.insns); + edge->callee->global.size); fprintf (dump_file, " to be inlined into %s in %s:%i\n" " Estimated growth after inlined into all callees is %+i insns.\n" @@ -1040,19 +1058,20 @@ cgraph_decide_inlining_of_small_functions (void) if (dump_file) { fprintf (dump_file, - " Inlined into %s which now has %i insns," - "net change of %+i insns.\n", + " Inlined into %s which now has sie %i and self time %i," + "net change of %+i.\n", cgraph_node_name (edge->caller), - edge->caller->global.insns, - overall_insns - old_insns); + edge->caller->global.time, + edge->caller->global.size, + overall_size - old_size); } - if (min_insns > overall_insns) + if (min_size > overall_size) { - min_insns = overall_insns; - max_insns = compute_max_insns (min_insns); + min_size = overall_size; + max_size = compute_max_insns (min_size); if (dump_file) - fprintf (dump_file, "New minimal insns reached: %i\n", min_insns); + fprintf (dump_file, "New minimal size reached: %i\n", min_size); } } while ((edge = (struct cgraph_edge *) fibheap_extract_min (heap)) != NULL) @@ -1081,34 +1100,38 @@ cgraph_decide_inlining (void) int nnodes; struct cgraph_node **order = XCNEWVEC (struct cgraph_node *, cgraph_n_nodes); - int old_insns = 0; + int old_size = 0; int i; - int initial_insns = 0; bool redo_always_inline = true; + int initial_size = 0; cgraph_remove_function_insertion_hook (function_insertion_hook_holder); max_count = 0; + max_benefit = 0; for (node = cgraph_nodes; node; node = node->next) - if (node->analyzed && (node->needed || node->reachable)) + if (node->analyzed) { struct cgraph_edge *e; - initial_insns += inline_summary (node)->self_insns; - gcc_assert (inline_summary (node)->self_insns == node->global.insns); + gcc_assert (inline_summary (node)->self_size == node->global.size); + gcc_assert (node->needed || node->reachable); + initial_size += node->global.size; for (e = node->callees; e; e = e->next_callee) if (max_count < e->count) max_count = e->count; + if (max_benefit < inline_summary (node)->time_inlining_benefit) + max_benefit = inline_summary (node)->time_inlining_benefit; } - overall_insns = initial_insns; gcc_assert (!max_count || (profile_info && flag_branch_probabilities)); + overall_size = initial_size; nnodes = cgraph_postorder (order); if (dump_file) fprintf (dump_file, - "\nDeciding on inlining. Starting with %i insns.\n", - initial_insns); + "\nDeciding on inlining. Starting with size %i.\n", + initial_size); for (node = cgraph_nodes; node; node = node->next) node->aux = 0; @@ -1142,9 +1165,9 @@ cgraph_decide_inlining (void) continue; if (dump_file) fprintf (dump_file, - "\nConsidering %s %i insns (always inline)\n", - cgraph_node_name (node), node->global.insns); - old_insns = overall_insns; + "\nConsidering %s size:%i (always inline)\n", + cgraph_node_name (node), node->global.size); + old_size = overall_size; for (e = node->callers; e; e = next) { next = e->next_caller; @@ -1163,9 +1186,9 @@ cgraph_decide_inlining (void) redo_always_inline = true; if (dump_file) fprintf (dump_file, - " Inlined into %s which now has %i insns.\n", + " Inlined into %s which now has size %i.\n", cgraph_node_name (e->caller), - e->caller->global.insns); + e->caller->global.size); } /* Inlining self recursive function might introduce new calls to themselves we didn't see in the loop above. Fill in the proper @@ -1175,8 +1198,8 @@ cgraph_decide_inlining (void) e->inline_failed = CIF_RECURSIVE_INLINING; if (dump_file) fprintf (dump_file, - " Inlined for a net change of %+i insns.\n", - overall_insns - old_insns); + " Inlined for a net change of %+i size.\n", + overall_size - old_size); } } @@ -1204,27 +1227,25 @@ cgraph_decide_inlining (void) if (dump_file) { fprintf (dump_file, - "\nConsidering %s %i insns.\n", - cgraph_node_name (node), node->global.insns); + "\nConsidering %s size %i.\n", + cgraph_node_name (node), node->global.size); fprintf (dump_file, " Called once from %s %i insns.\n", cgraph_node_name (node->callers->caller), - node->callers->caller->global.insns); + node->callers->caller->global.size); } - old_insns = overall_insns; - if (cgraph_check_inline_limits (node->callers->caller, node, NULL, false)) { cgraph_mark_inline (node->callers); if (dump_file) fprintf (dump_file, - " Inlined into %s which now has %i insns" - " for a net change of %+i insns.\n", + " Inlined into %s which now has %i size" + " for a net change of %+i size.\n", cgraph_node_name (node->callers->caller), - node->callers->caller->global.insns, - overall_insns - old_insns); + node->callers->caller->global.size, + overall_size - old_size); } else { @@ -1243,9 +1264,9 @@ cgraph_decide_inlining (void) if (dump_file) fprintf (dump_file, "\nInlined %i calls, eliminated %i functions, " - "%i insns turned to %i insns.\n\n", - ncalls_inlined, nfunctions_inlined, initial_insns, - overall_insns); + "size %i turned to %i size.\n\n", + ncalls_inlined, nfunctions_inlined, initial_size, + overall_size); free (order); return 0; } @@ -1430,6 +1451,7 @@ cgraph_decide_inlining_incrementally (struct cgraph_node *node, if (mode != INLINE_ALL && mode != INLINE_ALWAYS_INLINE) for (e = node->callees; e; e = e->next_callee) { + int allowed_growth = 0; if (!e->callee->local.inlinable || !e->inline_failed || e->callee->local.disregard_inline_limits) @@ -1456,6 +1478,10 @@ cgraph_decide_inlining_incrementally (struct cgraph_node *node, } continue; } + + if (cgraph_maybe_hot_edge_p (e)) + allowed_growth = PARAM_VALUE (PARAM_EARLY_INLINING_INSNS); + /* When the function body would grow and inlining the function won't eliminate the need for offline copy of the function, don't inline. */ @@ -1463,17 +1489,17 @@ cgraph_decide_inlining_incrementally (struct cgraph_node *node, || (!flag_inline_functions && !DECL_DECLARED_INLINE_P (e->callee->decl))) && (cgraph_estimate_size_after_inlining (1, e->caller, e->callee) - > e->caller->global.insns) - && cgraph_estimate_growth (e->callee) > 0) + >= e->caller->global.size + allowed_growth) + && cgraph_estimate_growth (e->callee) >= allowed_growth) { if (dump_file) { indent_to (dump_file, depth); fprintf (dump_file, - "Not inlining: code size would grow by %i insns.\n", + "Not inlining: code size would grow by %i.\n", cgraph_estimate_size_after_inlining (1, e->caller, e->callee) - - e->caller->global.insns); + - e->caller->global.size); } continue; } @@ -1600,6 +1626,177 @@ struct simple_ipa_opt_pass pass_ipa_early_inline = } }; +/* See if statement might disappear after inlining. We are not terribly + sophisficated, basically looking for simple abstraction penalty wrappers. */ + +static bool +likely_eliminated_by_inlining_p (gimple stmt) +{ + enum gimple_code code = gimple_code (stmt); + switch (code) + { + case GIMPLE_RETURN: + return true; + case GIMPLE_ASSIGN: + if (gimple_num_ops (stmt) != 2) + return false; + + /* Casts of parameters, loads from parameters passed by reference + and stores to return value or parameters are probably free after + inlining. */ + if (gimple_assign_rhs_code (stmt) == CONVERT_EXPR + || gimple_assign_rhs_code (stmt) == NOP_EXPR + || gimple_assign_rhs_code (stmt) == VIEW_CONVERT_EXPR + || gimple_assign_rhs_class (stmt) == GIMPLE_SINGLE_RHS) + { + tree rhs = gimple_assign_rhs1 (stmt); + tree lhs = gimple_assign_lhs (stmt); + tree inner_rhs = rhs; + tree inner_lhs = lhs; + bool rhs_free = false; + bool lhs_free = false; + + while (handled_component_p (inner_lhs) || TREE_CODE (inner_lhs) == INDIRECT_REF) + inner_lhs = TREE_OPERAND (inner_lhs, 0); + while (handled_component_p (inner_rhs) + || TREE_CODE (inner_rhs) == ADDR_EXPR || TREE_CODE (inner_rhs) == INDIRECT_REF) + inner_rhs = TREE_OPERAND (inner_rhs, 0); + + + if (TREE_CODE (inner_rhs) == PARM_DECL + || (TREE_CODE (inner_rhs) == SSA_NAME + && SSA_NAME_IS_DEFAULT_DEF (inner_rhs) + && TREE_CODE (SSA_NAME_VAR (inner_rhs)) == PARM_DECL)) + rhs_free = true; + if (rhs_free && is_gimple_reg (lhs)) + lhs_free = true; + if (((TREE_CODE (inner_lhs) == PARM_DECL + || (TREE_CODE (inner_lhs) == SSA_NAME + && SSA_NAME_IS_DEFAULT_DEF (inner_lhs) + && TREE_CODE (SSA_NAME_VAR (inner_lhs)) == PARM_DECL)) + && inner_lhs != lhs) + || TREE_CODE (inner_lhs) == RESULT_DECL + || (TREE_CODE (inner_lhs) == SSA_NAME + && TREE_CODE (SSA_NAME_VAR (inner_lhs)) == RESULT_DECL)) + lhs_free = true; + if (lhs_free && (is_gimple_reg (rhs) || is_gimple_min_invariant (rhs))) + rhs_free = true; + if (lhs_free && rhs_free) + return true; + } + return false; + default: + return false; + } +} + +/* Compute function body size parameters for NODE. */ + +static void +estimate_function_body_sizes (struct cgraph_node *node) +{ + gcov_type time = 0; + gcov_type time_inlining_benefit = 0; + int size = 0; + int size_inlining_benefit = 0; + basic_block bb; + gimple_stmt_iterator bsi; + struct function *my_function = DECL_STRUCT_FUNCTION (node->decl); + tree arg; + int freq; + tree funtype = TREE_TYPE (node->decl); + bitmap must_not_throw = must_not_throw_labels (); + + if (dump_file) + { + fprintf (dump_file, "Analyzing function body size: %s\n", cgraph_node_name (node)); + } + + gcc_assert (my_function && my_function->cfg); + FOR_EACH_BB_FN (bb, my_function) + { + freq = compute_call_stmt_bb_frequency (node->decl, bb); + for (bsi = gsi_start_bb (bb); !gsi_end_p (bsi); gsi_next (&bsi)) + { + int this_size = estimate_num_insns (gsi_stmt (bsi), &eni_size_weights); + int this_time = estimate_num_insns (gsi_stmt (bsi), &eni_time_weights); + + /* MUST_NOT_THROW is usually handled by runtime calling terminate and stopping + stacking unwinding. However when there is local cleanup that can resume + to MUST_NOT_THROW then we generate explicit handler containing + std::terminate () call. + + Because inlining of function can introduce new cleanup region, prior + inlining we keep std::terinate () calls for every MUST_NOT_THROW containing + function call. Wast majority of these will be eliminated after inlining + and crossjumping will inify possible duplicated calls. So ignore + the handlers for function body estimates. */ + if (gimple_code (gsi_stmt (bsi)) == GIMPLE_LABEL + && bitmap_bit_p (must_not_throw, + LABEL_DECL_UID (gimple_label_label (gsi_stmt (bsi))))) + { + if (dump_file) + fprintf (dump_file, " MUST_NOT_THROW landing pad. Ignoring whole BB.\n"); + } + if (dump_file) + { + fprintf (dump_file, " freq:%6i size:%3i time:%3i ", freq, this_size, this_time); + print_gimple_stmt (dump_file, gsi_stmt (bsi), 0, 0); + } + this_time *= freq; + time += this_time; + size += this_size; + if (likely_eliminated_by_inlining_p (gsi_stmt (bsi))) + { + size_inlining_benefit += this_size; + time_inlining_benefit += this_time; + if (dump_file) + fprintf (dump_file, " Likely eliminated\n"); + } + gcc_assert (time >= 0); + gcc_assert (size >= 0); + } + } + time = (time + CGRAPH_FREQ_BASE / 2) / CGRAPH_FREQ_BASE; + time_inlining_benefit = ((time_inlining_benefit + CGRAPH_FREQ_BASE / 2) + / CGRAPH_FREQ_BASE); + if (dump_file) + { + fprintf (dump_file, "Overall function body time: %i-%i size: %i-%i\n", + (int)time, (int)time_inlining_benefit, + size, size_inlining_benefit); + } + time_inlining_benefit += eni_time_weights.call_cost; + size_inlining_benefit += eni_size_weights.call_cost; + if (!VOID_TYPE_P (TREE_TYPE (funtype))) + { + int cost = estimate_move_cost (TREE_TYPE (funtype)); + time_inlining_benefit += cost; + size_inlining_benefit += cost; + } + for (arg = DECL_ARGUMENTS (node->decl); arg; arg = TREE_CHAIN (arg)) + { + int cost = estimate_move_cost (TREE_TYPE (arg)); + time_inlining_benefit += cost; + size_inlining_benefit += cost; + } + if (time_inlining_benefit > MAX_TIME) + time_inlining_benefit = MAX_TIME; + if (time > MAX_TIME) + time = MAX_TIME; + inline_summary (node)->self_time = time; + inline_summary (node)->self_size = size; + if (dump_file) + { + fprintf (dump_file, "With function call overhead time: %i-%i size: %i-%i\n", + (int)time, (int)time_inlining_benefit, + size, size_inlining_benefit); + } + inline_summary (node)->time_inlining_benefit = time_inlining_benefit; + inline_summary (node)->size_inlining_benefit = size_inlining_benefit; + BITMAP_FREE (must_not_throw); +} + /* Compute parameters of functions used by inliner. */ unsigned int compute_inline_parameters (struct cgraph_node *node) @@ -1617,19 +1814,13 @@ compute_inline_parameters (struct cgraph_node *node) /* Can this function be inlined at all? */ node->local.inlinable = tree_inlinable_function_p (current_function_decl); - - /* Estimate the number of instructions for this function. - ??? At -O0 we don't use this information except for the dumps, and - even then only for always_inline functions. But disabling this - causes ICEs in the inline heuristics... */ - inline_summary (node)->self_insns - = estimate_num_insns_fn (current_function_decl, &eni_inlining_weights); if (node->local.inlinable && !node->local.disregard_inline_limits) node->local.disregard_inline_limits = DECL_DISREGARD_INLINE_LIMITS (current_function_decl); - + estimate_function_body_sizes (node); /* Inlining characteristics are maintained by the cgraph_mark_inline. */ - node->global.insns = inline_summary (node)->self_insns; + node->global.time = inline_summary (node)->self_time; + node->global.size = inline_summary (node)->self_size; return 0; } @@ -1647,7 +1838,7 @@ struct gimple_opt_pass pass_inline_parameters = { { GIMPLE_PASS, - NULL, /* name */ + "inline_param", /* name */ NULL, /* gate */ compute_inline_parameters_for_current,/* execute */ NULL, /* sub */ diff --git a/gcc/params.def b/gcc/params.def index 684de5ed3ca..bb74b701486 100644 --- a/gcc/params.def +++ b/gcc/params.def @@ -100,7 +100,7 @@ DEFPARAM (PARAM_PREDICTABLE_BRANCH_OUTCOME, DEFPARAM (PARAM_MAX_INLINE_INSNS_SINGLE, "max-inline-insns-single", "The maximum number of instructions in a single function eligible for inlining", - 450, 0, 0) + 300, 0, 0) /* The single function inlining limit for functions that are inlined by virtue of -finline-functions (-O3). @@ -112,7 +112,7 @@ DEFPARAM (PARAM_MAX_INLINE_INSNS_SINGLE, DEFPARAM (PARAM_MAX_INLINE_INSNS_AUTO, "max-inline-insns-auto", "The maximum number of instructions when automatically inlining", - 90, 0, 0) + 60, 0, 0) DEFPARAM (PARAM_MAX_INLINE_INSNS_RECURSIVE, "max-inline-insns-recursive", @@ -204,9 +204,9 @@ DEFPARAM(PARAM_IPCP_UNIT_GROWTH, "ipcp-unit-growth", "how much can given compilation unit grow because of the interprocedural constant propagation (in percent)", 10, 0, 0) -DEFPARAM(PARAM_INLINE_CALL_COST, - "inline-call-cost", - "expense of call operation relative to ordinary arithmetic operations", +DEFPARAM(PARAM_EARLY_INLINING_INSNS, + "early-inlining-insns", + "maximal estimated growth of function body caused by early inlining of single call", 12, 0, 0) DEFPARAM(PARAM_LARGE_STACK_FRAME, "large-stack-frame", diff --git a/gcc/tree-inline.c b/gcc/tree-inline.c index 78cb716d7ad..1e75f7c0717 100644 --- a/gcc/tree-inline.c +++ b/gcc/tree-inline.c @@ -3156,12 +3156,6 @@ estimate_num_insns_fn (tree fndecl, eni_weights *weights) void init_inline_once (void) { - eni_inlining_weights.call_cost = PARAM_VALUE (PARAM_INLINE_CALL_COST); - eni_inlining_weights.target_builtin_call_cost = 1; - eni_inlining_weights.div_mod_cost = 10; - eni_inlining_weights.omp_cost = 40; - eni_inlining_weights.time_based = true; - eni_size_weights.call_cost = 1; eni_size_weights.target_builtin_call_cost = 1; eni_size_weights.div_mod_cost = 1; diff --git a/gcc/tree-ssa-structalias.c b/gcc/tree-ssa-structalias.c index eda053f2430..2f9a8e0c915 100644 --- a/gcc/tree-ssa-structalias.c +++ b/gcc/tree-ssa-structalias.c @@ -3425,7 +3425,7 @@ handle_lhs_call (tree lhs, int flags, VEC(ce_s, heap) *rhsc) static void handle_const_call (gimple stmt, VEC(ce_s, heap) **results) { - struct constraint_expr rhsc, tmpc; + struct constraint_expr rhsc, tmpc = {SCALAR, 0, 0}; tree tmpvar = NULL_TREE; unsigned int k; |