| /* Callgraph handling code. |
| Copyright (C) 2003-2014 Free Software Foundation, Inc. |
| Contributed by Jan Hubicka |
| |
| This file is part of GCC. |
| |
| GCC is free software; you can redistribute it and/or modify it under |
| the terms of the GNU General Public License as published by the Free |
| Software Foundation; either version 3, or (at your option) any later |
| version. |
| |
| GCC is distributed in the hope that it will be useful, but WITHOUT ANY |
| WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
| for more details. |
| |
| You should have received a copy of the GNU General Public License |
| along with GCC; see the file COPYING3. If not see |
| <http://www.gnu.org/licenses/>. */ |
| |
| /* This file contains basic routines manipulating call graph |
| |
| The call-graph is a data structure designed for intra-procedural optimization. |
| It represents a multi-graph where nodes are functions and edges are call sites. */ |
| |
| #include "config.h" |
| #include "system.h" |
| #include "coretypes.h" |
| #include "tm.h" |
| #include "tree.h" |
| #include "varasm.h" |
| #include "calls.h" |
| #include "print-tree.h" |
| #include "tree-inline.h" |
| #include "langhooks.h" |
| #include "hashtab.h" |
| #include "toplev.h" |
| #include "flags.h" |
| #include "debug.h" |
| #include "target.h" |
| #include "cgraph.h" |
| #include "intl.h" |
| #include "tree-ssa-alias.h" |
| #include "internal-fn.h" |
| #include "tree-eh.h" |
| #include "gimple-expr.h" |
| #include "gimple.h" |
| #include "gimple-iterator.h" |
| #include "timevar.h" |
| #include "dumpfile.h" |
| #include "gimple-ssa.h" |
| #include "cgraph.h" |
| #include "tree-cfg.h" |
| #include "tree-ssa.h" |
| #include "value-prof.h" |
| #include "except.h" |
| #include "diagnostic-core.h" |
| #include "rtl.h" |
| #include "ipa-utils.h" |
| #include "lto-streamer.h" |
| #include "l-ipo.h" |
| #include "ipa-inline.h" |
| #include "cfgloop.h" |
| #include "gimple-pretty-print.h" |
| #include "expr.h" |
| #include "tree-dfa.h" |
| |
| /* FIXME: Only for PROP_loops, but cgraph shouldn't have to know about this. */ |
| #include "tree-pass.h" |
| |
| static void cgraph_node_remove_callers (struct cgraph_node *node); |
| static inline void cgraph_edge_remove_caller (struct cgraph_edge *e); |
| static inline void cgraph_edge_remove_callee (struct cgraph_edge *e); |
| |
| /* Queue of cgraph nodes scheduled to be lowered. */ |
| symtab_node *x_cgraph_nodes_queue; |
| #define cgraph_nodes_queue ((struct cgraph_node *)x_cgraph_nodes_queue) |
| |
| /* Number of nodes in existence. */ |
| int cgraph_n_nodes; |
| |
| /* Maximal uid used in cgraph nodes. */ |
| int cgraph_max_uid; |
| |
| /* Maximal uid used in cgraph edges. */ |
| int cgraph_edge_max_uid; |
| |
| /* Set when whole unit has been analyzed so we can access global info. */ |
| bool cgraph_global_info_ready = false; |
| |
| /* What state callgraph is in right now. */ |
| enum cgraph_state cgraph_state = CGRAPH_STATE_PARSING; |
| |
| /* Set when the cgraph is fully build and the basic flags are computed. */ |
| bool cgraph_function_flags_ready = false; |
| |
| /* List of hooks triggered on cgraph_edge events. */ |
| struct cgraph_edge_hook_list { |
| cgraph_edge_hook hook; |
| void *data; |
| struct cgraph_edge_hook_list *next; |
| }; |
| |
| /* List of hooks triggered on cgraph_node events. */ |
| struct cgraph_node_hook_list { |
| cgraph_node_hook hook; |
| void *data; |
| struct cgraph_node_hook_list *next; |
| }; |
| |
| /* List of hooks triggered on events involving two cgraph_edges. */ |
| struct cgraph_2edge_hook_list { |
| cgraph_2edge_hook hook; |
| void *data; |
| struct cgraph_2edge_hook_list *next; |
| }; |
| |
| /* List of hooks triggered on events involving two cgraph_nodes. */ |
| struct cgraph_2node_hook_list { |
| cgraph_2node_hook hook; |
| void *data; |
| struct cgraph_2node_hook_list *next; |
| }; |
| |
| /* List of hooks triggered when an edge is removed. */ |
| struct cgraph_edge_hook_list *first_cgraph_edge_removal_hook; |
| /* List of hooks triggered when a node is removed. */ |
| struct cgraph_node_hook_list *first_cgraph_node_removal_hook; |
| /* List of hooks triggered when an edge is duplicated. */ |
| struct cgraph_2edge_hook_list *first_cgraph_edge_duplicated_hook; |
| /* List of hooks triggered when a node is duplicated. */ |
| struct cgraph_2node_hook_list *first_cgraph_node_duplicated_hook; |
| /* List of hooks triggered when an function is inserted. */ |
| struct cgraph_node_hook_list *first_cgraph_function_insertion_hook; |
| |
| /* Head of a linked list of unused (freed) call graph nodes. |
| Do not GTY((delete)) this list so UIDs gets reliably recycled. */ |
| static GTY(()) struct cgraph_node *free_nodes; |
| /* Head of a linked list of unused (freed) call graph edges. |
| Do not GTY((delete)) this list so UIDs gets reliably recycled. */ |
| static GTY(()) struct cgraph_edge *free_edges; |
| |
| /* Did procss_same_body_aliases run? */ |
| bool cpp_implicit_aliases_done; |
| |
| /* Map a cgraph_node to cgraph_function_version_info using this htab. |
| The cgraph_function_version_info has a THIS_NODE field that is the |
| corresponding cgraph_node.. */ |
| |
| static GTY((param_is (struct cgraph_function_version_info))) htab_t |
| cgraph_fnver_htab = NULL; |
| |
| /* Hash function for cgraph_fnver_htab. */ |
| static hashval_t |
| cgraph_fnver_htab_hash (const void *ptr) |
| { |
| int uid = ((const struct cgraph_function_version_info *)ptr)->this_node->uid; |
| return (hashval_t)(uid); |
| } |
| |
| /* eq function for cgraph_fnver_htab. */ |
| static int |
| cgraph_fnver_htab_eq (const void *p1, const void *p2) |
| { |
| const struct cgraph_function_version_info *n1 |
| = (const struct cgraph_function_version_info *)p1; |
| const struct cgraph_function_version_info *n2 |
| = (const struct cgraph_function_version_info *)p2; |
| |
| return n1->this_node->uid == n2->this_node->uid; |
| } |
| |
| /* Mark as GC root all allocated nodes. */ |
| static GTY(()) struct cgraph_function_version_info * |
| version_info_node = NULL; |
| |
| /* Get the cgraph_function_version_info node corresponding to node. */ |
| struct cgraph_function_version_info * |
| get_cgraph_node_version (struct cgraph_node *node) |
| { |
| struct cgraph_function_version_info *ret; |
| struct cgraph_function_version_info key; |
| key.this_node = node; |
| |
| if (cgraph_fnver_htab == NULL) |
| return NULL; |
| |
| ret = (struct cgraph_function_version_info *) |
| htab_find (cgraph_fnver_htab, &key); |
| |
| return ret; |
| } |
| |
| /* Insert a new cgraph_function_version_info node into cgraph_fnver_htab |
| corresponding to cgraph_node NODE. */ |
| struct cgraph_function_version_info * |
| insert_new_cgraph_node_version (struct cgraph_node *node) |
| { |
| void **slot; |
| |
| version_info_node = NULL; |
| version_info_node = ggc_alloc_cleared_cgraph_function_version_info (); |
| version_info_node->this_node = node; |
| |
| if (cgraph_fnver_htab == NULL) |
| cgraph_fnver_htab = htab_create_ggc (2, cgraph_fnver_htab_hash, |
| cgraph_fnver_htab_eq, NULL); |
| |
| slot = htab_find_slot (cgraph_fnver_htab, version_info_node, INSERT); |
| gcc_assert (slot != NULL); |
| *slot = version_info_node; |
| return version_info_node; |
| } |
| |
| /* Remove the cgraph_function_version_info and cgraph_node for DECL. This |
| DECL is a duplicate declaration. */ |
| void |
| delete_function_version (tree decl) |
| { |
| struct cgraph_node *decl_node = cgraph_get_node (decl); |
| struct cgraph_function_version_info *decl_v = NULL; |
| |
| if (decl_node == NULL) |
| return; |
| |
| decl_v = get_cgraph_node_version (decl_node); |
| |
| if (decl_v == NULL) |
| return; |
| |
| if (decl_v->prev != NULL) |
| decl_v->prev->next = decl_v->next; |
| |
| if (decl_v->next != NULL) |
| decl_v->next->prev = decl_v->prev; |
| |
| if (cgraph_fnver_htab != NULL) |
| htab_remove_elt (cgraph_fnver_htab, decl_v); |
| |
| cgraph_remove_node (decl_node); |
| } |
| |
| /* Record that DECL1 and DECL2 are semantically identical function |
| versions. */ |
| void |
| record_function_versions (tree decl1, tree decl2) |
| { |
| struct cgraph_node *decl1_node = cgraph_get_create_node (decl1); |
| struct cgraph_node *decl2_node = cgraph_get_create_node (decl2); |
| struct cgraph_function_version_info *decl1_v = NULL; |
| struct cgraph_function_version_info *decl2_v = NULL; |
| struct cgraph_function_version_info *before; |
| struct cgraph_function_version_info *after; |
| |
| gcc_assert (decl1_node != NULL && decl2_node != NULL); |
| decl1_v = get_cgraph_node_version (decl1_node); |
| decl2_v = get_cgraph_node_version (decl2_node); |
| |
| if (decl1_v != NULL && decl2_v != NULL) |
| return; |
| |
| if (decl1_v == NULL) |
| decl1_v = insert_new_cgraph_node_version (decl1_node); |
| |
| if (decl2_v == NULL) |
| decl2_v = insert_new_cgraph_node_version (decl2_node); |
| |
| /* Chain decl2_v and decl1_v. All semantically identical versions |
| will be chained together. */ |
| |
| before = decl1_v; |
| after = decl2_v; |
| |
| while (before->next != NULL) |
| before = before->next; |
| |
| while (after->prev != NULL) |
| after= after->prev; |
| |
| before->next = after; |
| after->prev = before; |
| } |
| |
| /* Macros to access the next item in the list of free cgraph nodes and |
| edges. */ |
| #define NEXT_FREE_NODE(NODE) cgraph ((NODE)->next) |
| #define SET_NEXT_FREE_NODE(NODE,NODE2) ((NODE))->next = NODE2 |
| #define NEXT_FREE_EDGE(EDGE) (EDGE)->prev_caller |
| |
| /* Register HOOK to be called with DATA on each removed edge. */ |
| struct cgraph_edge_hook_list * |
| cgraph_add_edge_removal_hook (cgraph_edge_hook hook, void *data) |
| { |
| struct cgraph_edge_hook_list *entry; |
| struct cgraph_edge_hook_list **ptr = &first_cgraph_edge_removal_hook; |
| |
| entry = (struct cgraph_edge_hook_list *) xmalloc (sizeof (*entry)); |
| entry->hook = hook; |
| entry->data = data; |
| entry->next = NULL; |
| while (*ptr) |
| ptr = &(*ptr)->next; |
| *ptr = entry; |
| return entry; |
| } |
| |
| /* Remove ENTRY from the list of hooks called on removing edges. */ |
| void |
| cgraph_remove_edge_removal_hook (struct cgraph_edge_hook_list *entry) |
| { |
| struct cgraph_edge_hook_list **ptr = &first_cgraph_edge_removal_hook; |
| |
| while (*ptr != entry) |
| ptr = &(*ptr)->next; |
| *ptr = entry->next; |
| free (entry); |
| } |
| |
| /* Call all edge removal hooks. */ |
| static void |
| cgraph_call_edge_removal_hooks (struct cgraph_edge *e) |
| { |
| struct cgraph_edge_hook_list *entry = first_cgraph_edge_removal_hook; |
| while (entry) |
| { |
| entry->hook (e, entry->data); |
| entry = entry->next; |
| } |
| } |
| |
| /* Register HOOK to be called with DATA on each removed node. */ |
| struct cgraph_node_hook_list * |
| cgraph_add_node_removal_hook (cgraph_node_hook hook, void *data) |
| { |
| struct cgraph_node_hook_list *entry; |
| struct cgraph_node_hook_list **ptr = &first_cgraph_node_removal_hook; |
| |
| entry = (struct cgraph_node_hook_list *) xmalloc (sizeof (*entry)); |
| entry->hook = hook; |
| entry->data = data; |
| entry->next = NULL; |
| while (*ptr) |
| ptr = &(*ptr)->next; |
| *ptr = entry; |
| return entry; |
| } |
| |
| /* Remove ENTRY from the list of hooks called on removing nodes. */ |
| void |
| cgraph_remove_node_removal_hook (struct cgraph_node_hook_list *entry) |
| { |
| struct cgraph_node_hook_list **ptr = &first_cgraph_node_removal_hook; |
| |
| while (*ptr != entry) |
| ptr = &(*ptr)->next; |
| *ptr = entry->next; |
| free (entry); |
| } |
| |
| /* Call all node removal hooks. */ |
| static void |
| cgraph_call_node_removal_hooks (struct cgraph_node *node) |
| { |
| struct cgraph_node_hook_list *entry = first_cgraph_node_removal_hook; |
| while (entry) |
| { |
| entry->hook (node, entry->data); |
| entry = entry->next; |
| } |
| } |
| |
| /* Register HOOK to be called with DATA on each inserted node. */ |
| struct cgraph_node_hook_list * |
| cgraph_add_function_insertion_hook (cgraph_node_hook hook, void *data) |
| { |
| struct cgraph_node_hook_list *entry; |
| struct cgraph_node_hook_list **ptr = &first_cgraph_function_insertion_hook; |
| |
| entry = (struct cgraph_node_hook_list *) xmalloc (sizeof (*entry)); |
| entry->hook = hook; |
| entry->data = data; |
| entry->next = NULL; |
| while (*ptr) |
| ptr = &(*ptr)->next; |
| *ptr = entry; |
| return entry; |
| } |
| |
| /* Remove ENTRY from the list of hooks called on inserted nodes. */ |
| void |
| cgraph_remove_function_insertion_hook (struct cgraph_node_hook_list *entry) |
| { |
| struct cgraph_node_hook_list **ptr = &first_cgraph_function_insertion_hook; |
| |
| while (*ptr != entry) |
| ptr = &(*ptr)->next; |
| *ptr = entry->next; |
| free (entry); |
| } |
| |
| /* Call all node insertion hooks. */ |
| void |
| cgraph_call_function_insertion_hooks (struct cgraph_node *node) |
| { |
| struct cgraph_node_hook_list *entry = first_cgraph_function_insertion_hook; |
| while (entry) |
| { |
| entry->hook (node, entry->data); |
| entry = entry->next; |
| } |
| } |
| |
| /* Register HOOK to be called with DATA on each duplicated edge. */ |
| struct cgraph_2edge_hook_list * |
| cgraph_add_edge_duplication_hook (cgraph_2edge_hook hook, void *data) |
| { |
| struct cgraph_2edge_hook_list *entry; |
| struct cgraph_2edge_hook_list **ptr = &first_cgraph_edge_duplicated_hook; |
| |
| entry = (struct cgraph_2edge_hook_list *) xmalloc (sizeof (*entry)); |
| entry->hook = hook; |
| entry->data = data; |
| entry->next = NULL; |
| while (*ptr) |
| ptr = &(*ptr)->next; |
| *ptr = entry; |
| return entry; |
| } |
| |
| /* Remove ENTRY from the list of hooks called on duplicating edges. */ |
| void |
| cgraph_remove_edge_duplication_hook (struct cgraph_2edge_hook_list *entry) |
| { |
| struct cgraph_2edge_hook_list **ptr = &first_cgraph_edge_duplicated_hook; |
| |
| while (*ptr != entry) |
| ptr = &(*ptr)->next; |
| *ptr = entry->next; |
| free (entry); |
| } |
| |
| /* Call all edge duplication hooks. */ |
| void |
| cgraph_call_edge_duplication_hooks (struct cgraph_edge *cs1, |
| struct cgraph_edge *cs2) |
| { |
| struct cgraph_2edge_hook_list *entry = first_cgraph_edge_duplicated_hook; |
| while (entry) |
| { |
| entry->hook (cs1, cs2, entry->data); |
| entry = entry->next; |
| } |
| } |
| |
| /* Register HOOK to be called with DATA on each duplicated node. */ |
| struct cgraph_2node_hook_list * |
| cgraph_add_node_duplication_hook (cgraph_2node_hook hook, void *data) |
| { |
| struct cgraph_2node_hook_list *entry; |
| struct cgraph_2node_hook_list **ptr = &first_cgraph_node_duplicated_hook; |
| |
| entry = (struct cgraph_2node_hook_list *) xmalloc (sizeof (*entry)); |
| entry->hook = hook; |
| entry->data = data; |
| entry->next = NULL; |
| while (*ptr) |
| ptr = &(*ptr)->next; |
| *ptr = entry; |
| return entry; |
| } |
| |
| /* Remove ENTRY from the list of hooks called on duplicating nodes. */ |
| void |
| cgraph_remove_node_duplication_hook (struct cgraph_2node_hook_list *entry) |
| { |
| struct cgraph_2node_hook_list **ptr = &first_cgraph_node_duplicated_hook; |
| |
| while (*ptr != entry) |
| ptr = &(*ptr)->next; |
| *ptr = entry->next; |
| free (entry); |
| } |
| |
| /* Call all node duplication hooks. */ |
| void |
| cgraph_call_node_duplication_hooks (struct cgraph_node *node1, |
| struct cgraph_node *node2) |
| { |
| struct cgraph_2node_hook_list *entry = first_cgraph_node_duplicated_hook; |
| while (entry) |
| { |
| entry->hook (node1, node2, entry->data); |
| entry = entry->next; |
| } |
| } |
| |
| /* Allocate new callgraph node. */ |
| |
| static inline struct cgraph_node * |
| cgraph_allocate_node (void) |
| { |
| struct cgraph_node *node; |
| |
| if (free_nodes) |
| { |
| node = free_nodes; |
| free_nodes = NEXT_FREE_NODE (node); |
| } |
| else |
| { |
| node = ggc_alloc_cleared_cgraph_node (); |
| node->uid = cgraph_max_uid++; |
| } |
| |
| return node; |
| } |
| |
| /* Allocate new callgraph node and insert it into basic data structures. */ |
| |
| struct cgraph_node * |
| cgraph_create_empty_node (void) |
| { |
| struct cgraph_node *node = cgraph_allocate_node (); |
| |
| node->type = SYMTAB_FUNCTION; |
| node->frequency = NODE_FREQUENCY_NORMAL; |
| node->count_materialization_scale = REG_BR_PROB_BASE; |
| cgraph_n_nodes++; |
| return node; |
| } |
| |
| /* Return cgraph node assigned to DECL. Create new one when needed. */ |
| |
| struct cgraph_node * |
| cgraph_create_node (tree decl) |
| { |
| struct cgraph_node *node = cgraph_create_empty_node (); |
| gcc_assert (TREE_CODE (decl) == FUNCTION_DECL); |
| |
| node->decl = decl; |
| symtab_register_node (node); |
| |
| if (DECL_CONTEXT (decl) && TREE_CODE (DECL_CONTEXT (decl)) == FUNCTION_DECL) |
| { |
| node->origin = cgraph_get_create_node (DECL_CONTEXT (decl)); |
| node->next_nested = node->origin->nested; |
| node->origin->nested = node; |
| } |
| return node; |
| } |
| |
| /* Try to find a call graph node for declaration DECL and if it does not exist |
| or if it corresponds to an inline clone, create a new one. */ |
| |
| struct cgraph_node * |
| cgraph_get_create_node (tree decl) |
| { |
| struct cgraph_node *first_clone = cgraph_get_node (decl); |
| |
| if (first_clone && !first_clone->global.inlined_to) |
| return first_clone; |
| |
| struct cgraph_node *node = cgraph_create_node (decl); |
| if (first_clone) |
| { |
| first_clone->clone_of = node; |
| node->clones = first_clone; |
| symtab_prevail_in_asm_name_hash (node); |
| symtab_insert_node_to_hashtable (node); |
| if (dump_file) |
| fprintf (dump_file, "Introduced new external node " |
| "(%s/%i) and turned into root of the clone tree.\n", |
| xstrdup (node->name ()), node->order); |
| } |
| else if (dump_file) |
| fprintf (dump_file, "Introduced new external node " |
| "(%s/%i).\n", xstrdup (node->name ()), |
| node->order); |
| return node; |
| } |
| |
| /* Mark ALIAS as an alias to DECL. DECL_NODE is cgraph node representing |
| the function body is associated with (not necessarily cgraph_node (DECL). */ |
| |
| struct cgraph_node * |
| cgraph_create_function_alias (tree alias, tree target) |
| { |
| struct cgraph_node *alias_node; |
| |
| gcc_assert (TREE_CODE (target) == FUNCTION_DECL |
| || TREE_CODE (target) == IDENTIFIER_NODE); |
| gcc_assert (TREE_CODE (alias) == FUNCTION_DECL); |
| alias_node = cgraph_get_create_node (alias); |
| gcc_assert (!alias_node->definition); |
| alias_node->alias_target = target; |
| alias_node->definition = true; |
| alias_node->alias = true; |
| if (lookup_attribute ("weakref", DECL_ATTRIBUTES (alias)) != NULL) |
| alias_node->weakref = true; |
| return alias_node; |
| } |
| |
| /* Attempt to mark ALIAS as an alias to DECL. Return alias node if successful |
| and NULL otherwise. |
| Same body aliases are output whenever the body of DECL is output, |
| and cgraph_get_node (ALIAS) transparently returns cgraph_get_node (DECL). */ |
| |
| struct cgraph_node * |
| cgraph_same_body_alias (struct cgraph_node *decl_node ATTRIBUTE_UNUSED, tree alias, tree decl) |
| { |
| struct cgraph_node *n; |
| #ifndef ASM_OUTPUT_DEF |
| /* If aliases aren't supported by the assembler, fail. */ |
| return NULL; |
| #endif |
| /* Langhooks can create same body aliases of symbols not defined. |
| Those are useless. Drop them on the floor. */ |
| if (cgraph_global_info_ready) |
| return NULL; |
| |
| n = cgraph_create_function_alias (alias, decl); |
| n->cpp_implicit_alias = true; |
| if (cpp_implicit_aliases_done) |
| symtab_resolve_alias (n, |
| cgraph_get_node (decl)); |
| return n; |
| } |
| |
| /* Add thunk alias into callgraph. The alias declaration is ALIAS and it |
| aliases DECL with an adjustments made into the first parameter. |
| See comments in thunk_adjust for detail on the parameters. */ |
| |
| struct cgraph_node * |
| cgraph_add_thunk (struct cgraph_node *decl_node ATTRIBUTE_UNUSED, |
| tree alias, tree decl ATTRIBUTE_UNUSED, |
| bool this_adjusting, |
| HOST_WIDE_INT fixed_offset, HOST_WIDE_INT virtual_value, |
| tree virtual_offset, |
| tree real_alias) |
| { |
| struct cgraph_node *node; |
| |
| node = cgraph_get_node (alias); |
| if (node) |
| { |
| gcc_assert (node->definition); |
| gcc_assert (!node->alias); |
| gcc_assert (!node->thunk.thunk_p); |
| cgraph_remove_node (node); |
| } |
| |
| node = cgraph_create_node (alias); |
| gcc_checking_assert (!virtual_offset |
| || tree_to_double_int (virtual_offset) == |
| double_int::from_shwi (virtual_value)); |
| node->thunk.fixed_offset = fixed_offset; |
| node->thunk.this_adjusting = this_adjusting; |
| node->thunk.virtual_value = virtual_value; |
| node->thunk.virtual_offset_p = virtual_offset != NULL; |
| node->thunk.alias = real_alias; |
| node->thunk.thunk_p = true; |
| node->definition = true; |
| |
| return node; |
| } |
| |
| /* Return the cgraph node that has ASMNAME for its DECL_ASSEMBLER_NAME. |
| Return NULL if there's no such node. */ |
| |
| struct cgraph_node * |
| cgraph_node_for_asm (tree asmname) |
| { |
| /* We do not want to look at inline clones. */ |
| for (symtab_node *node = symtab_node_for_asm (asmname); |
| node; |
| node = node->next_sharing_asm_name) |
| { |
| cgraph_node *cn = dyn_cast <cgraph_node> (node); |
| if (cn && !cn->global.inlined_to) |
| { |
| if (L_IPO_COMP_MODE && cgraph_pre_profiling_inlining_done) |
| return cgraph_lipo_get_resolved_node (cn->decl); |
| return cn; |
| } |
| } |
| return NULL; |
| } |
| |
| /* Returns a hash value for X (which really is a cgraph_edge). */ |
| |
| static hashval_t |
| edge_hash (const void *x) |
| { |
| return htab_hash_pointer (((const struct cgraph_edge *) x)->call_stmt); |
| } |
| |
| /* Return nonzero if the call_stmt of of cgraph_edge X is stmt *Y. */ |
| |
| static int |
| edge_eq (const void *x, const void *y) |
| { |
| return ((const struct cgraph_edge *) x)->call_stmt == y; |
| } |
| |
| /* Add call graph edge E to call site hash of its caller. */ |
| |
| static inline void |
| cgraph_update_edge_in_call_site_hash (struct cgraph_edge *e) |
| { |
| void **slot; |
| slot = htab_find_slot_with_hash (e->caller->call_site_hash, |
| e->call_stmt, |
| htab_hash_pointer (e->call_stmt), |
| INSERT); |
| *slot = e; |
| } |
| |
| /* Add call graph edge E to call site hash of its caller. */ |
| |
| static inline void |
| cgraph_add_edge_to_call_site_hash (struct cgraph_edge *e) |
| { |
| void **slot; |
| /* There are two speculative edges for every statement (one direct, |
| one indirect); always hash the direct one. */ |
| if (e->speculative && e->indirect_unknown_callee) |
| return; |
| slot = htab_find_slot_with_hash (e->caller->call_site_hash, |
| e->call_stmt, |
| htab_hash_pointer (e->call_stmt), |
| INSERT); |
| if (*slot) |
| { |
| gcc_assert (((struct cgraph_edge *)*slot)->speculative); |
| if (e->callee) |
| *slot = e; |
| return; |
| } |
| gcc_assert (!*slot || e->speculative); |
| *slot = e; |
| } |
| |
| /* Return the callgraph edge representing the GIMPLE_CALL statement |
| CALL_STMT. */ |
| |
| struct cgraph_edge * |
| cgraph_edge (struct cgraph_node *node, gimple call_stmt) |
| { |
| struct cgraph_edge *e, *e2; |
| int n = 0; |
| |
| if (node->call_site_hash) |
| return (struct cgraph_edge *) |
| htab_find_with_hash (node->call_site_hash, call_stmt, |
| htab_hash_pointer (call_stmt)); |
| |
| /* This loop may turn out to be performance problem. In such case adding |
| hashtables into call nodes with very many edges is probably best |
| solution. It is not good idea to add pointer into CALL_EXPR itself |
| because we want to make possible having multiple cgraph nodes representing |
| different clones of the same body before the body is actually cloned. */ |
| for (e = node->callees; e; e = e->next_callee) |
| { |
| if (e->call_stmt == call_stmt) |
| break; |
| n++; |
| } |
| |
| if (!e) |
| for (e = node->indirect_calls; e; e = e->next_callee) |
| { |
| if (e->call_stmt == call_stmt) |
| break; |
| n++; |
| } |
| |
| if (n > 100) |
| { |
| node->call_site_hash = htab_create_ggc (120, edge_hash, edge_eq, NULL); |
| for (e2 = node->callees; e2; e2 = e2->next_callee) |
| /* Skip fake edges. */ |
| if (e2->call_stmt) |
| cgraph_add_edge_to_call_site_hash (e2); |
| for (e2 = node->indirect_calls; e2; e2 = e2->next_callee) |
| cgraph_add_edge_to_call_site_hash (e2); |
| } |
| |
| return e; |
| } |
| |
| |
| /* Change field call_stmt of edge E to NEW_STMT. |
| If UPDATE_SPECULATIVE and E is any component of speculative |
| edge, then update all components. */ |
| |
| void |
| cgraph_set_call_stmt (struct cgraph_edge *e, gimple new_stmt, |
| bool update_speculative) |
| { |
| tree decl; |
| |
| /* Speculative edges has three component, update all of them |
| when asked to. */ |
| if (update_speculative && e->speculative) |
| { |
| struct cgraph_edge *direct, *indirect; |
| struct ipa_ref *ref; |
| |
| cgraph_speculative_call_info (e, direct, indirect, ref); |
| cgraph_set_call_stmt (direct, new_stmt, false); |
| cgraph_set_call_stmt (indirect, new_stmt, false); |
| ref->stmt = new_stmt; |
| return; |
| } |
| |
| /* Only direct speculative edges go to call_site_hash. */ |
| if (e->caller->call_site_hash |
| && (!e->speculative || !e->indirect_unknown_callee)) |
| { |
| htab_remove_elt_with_hash (e->caller->call_site_hash, |
| e->call_stmt, |
| htab_hash_pointer (e->call_stmt)); |
| } |
| |
| e->call_stmt = new_stmt; |
| if (e->indirect_unknown_callee |
| && (decl = gimple_call_fndecl (new_stmt))) |
| { |
| /* Constant propagation (and possibly also inlining?) can turn an |
| indirect call into a direct one. */ |
| struct cgraph_node *new_callee = cgraph_get_node (decl); |
| if (L_IPO_COMP_MODE && cgraph_pre_profiling_inlining_done) |
| new_callee = cgraph_lipo_get_resolved_node (decl); |
| |
| gcc_checking_assert (new_callee); |
| e = cgraph_make_edge_direct (e, new_callee); |
| } |
| |
| push_cfun (DECL_STRUCT_FUNCTION (e->caller->decl)); |
| e->can_throw_external = stmt_can_throw_external (new_stmt); |
| pop_cfun (); |
| if (e->caller->call_site_hash) |
| cgraph_add_edge_to_call_site_hash (e); |
| } |
| |
| /* Allocate a cgraph_edge structure and fill it with data according to the |
| parameters of which only CALLEE can be NULL (when creating an indirect call |
| edge). */ |
| |
| static struct cgraph_edge * |
| cgraph_create_edge_1 (struct cgraph_node *caller, struct cgraph_node *callee, |
| gimple call_stmt, gcov_type count, int freq, |
| bool indir_unknown_callee) |
| { |
| struct cgraph_edge *edge; |
| |
| /* LTO does not actually have access to the call_stmt since these |
| have not been loaded yet. */ |
| if (call_stmt) |
| { |
| /* This is a rather expensive check possibly triggering |
| construction of call stmt hashtable. */ |
| #ifdef ENABLE_CHECKING |
| struct cgraph_edge *e; |
| gcc_checking_assert (!(e=cgraph_edge (caller, call_stmt)) || e->speculative); |
| #endif |
| |
| gcc_assert (is_gimple_call (call_stmt)); |
| } |
| |
| if (free_edges) |
| { |
| edge = free_edges; |
| free_edges = NEXT_FREE_EDGE (edge); |
| } |
| else |
| { |
| edge = ggc_alloc_cgraph_edge (); |
| edge->uid = cgraph_edge_max_uid++; |
| } |
| |
| edge->aux = NULL; |
| edge->caller = caller; |
| edge->callee = callee; |
| edge->prev_caller = NULL; |
| edge->next_caller = NULL; |
| edge->prev_callee = NULL; |
| edge->next_callee = NULL; |
| edge->lto_stmt_uid = 0; |
| |
| edge->count = count; |
| gcc_assert (count >= 0); |
| edge->frequency = freq; |
| gcc_assert (freq >= 0); |
| gcc_assert (freq <= CGRAPH_FREQ_MAX); |
| |
| edge->call_stmt = call_stmt; |
| push_cfun (DECL_STRUCT_FUNCTION (caller->decl)); |
| edge->can_throw_external |
| = call_stmt ? stmt_can_throw_external (call_stmt) : false; |
| pop_cfun (); |
| if (call_stmt |
| && callee && callee->decl |
| && !gimple_check_call_matching_types (call_stmt, callee->decl, |
| false)) |
| edge->call_stmt_cannot_inline_p = true; |
| else |
| edge->call_stmt_cannot_inline_p = false; |
| |
| edge->indirect_info = NULL; |
| edge->indirect_inlining_edge = 0; |
| edge->speculative = false; |
| edge->indirect_unknown_callee = indir_unknown_callee; |
| if (call_stmt && caller->call_site_hash) |
| cgraph_add_edge_to_call_site_hash (edge); |
| |
| return edge; |
| } |
| |
| /* Create edge from CALLER to CALLEE in the cgraph. */ |
| |
| struct cgraph_edge * |
| cgraph_create_edge (struct cgraph_node *caller, struct cgraph_node *callee, |
| gimple call_stmt, gcov_type count, int freq) |
| { |
| struct cgraph_edge *edge = cgraph_create_edge_1 (caller, callee, call_stmt, |
| count, freq, false); |
| |
| initialize_inline_failed (edge); |
| |
| edge->next_caller = callee->callers; |
| if (callee->callers) |
| callee->callers->prev_caller = edge; |
| edge->next_callee = caller->callees; |
| if (caller->callees) |
| caller->callees->prev_callee = edge; |
| caller->callees = edge; |
| callee->callers = edge; |
| |
| return edge; |
| } |
| |
| /* Allocate cgraph_indirect_call_info and set its fields to default values. */ |
| |
| struct cgraph_indirect_call_info * |
| cgraph_allocate_init_indirect_info (void) |
| { |
| struct cgraph_indirect_call_info *ii; |
| |
| ii = ggc_alloc_cleared_cgraph_indirect_call_info (); |
| ii->param_index = -1; |
| return ii; |
| } |
| |
| /* Create an indirect edge with a yet-undetermined callee where the call |
| statement destination is a formal parameter of the caller with index |
| PARAM_INDEX. */ |
| |
| struct cgraph_edge * |
| cgraph_create_indirect_edge (struct cgraph_node *caller, gimple call_stmt, |
| int ecf_flags, |
| gcov_type count, int freq) |
| { |
| struct cgraph_edge *edge = cgraph_create_edge_1 (caller, NULL, call_stmt, |
| count, freq, true); |
| tree target; |
| |
| initialize_inline_failed (edge); |
| |
| edge->indirect_info = cgraph_allocate_init_indirect_info (); |
| edge->indirect_info->ecf_flags = ecf_flags; |
| |
| /* Record polymorphic call info. */ |
| if (call_stmt |
| && (target = gimple_call_fn (call_stmt)) |
| && virtual_method_call_p (target)) |
| { |
| tree otr_type; |
| HOST_WIDE_INT otr_token; |
| ipa_polymorphic_call_context context; |
| |
| get_polymorphic_call_info (caller->decl, |
| target, |
| &otr_type, &otr_token, |
| &context); |
| |
| /* Only record types can have virtual calls. */ |
| gcc_assert (TREE_CODE (otr_type) == RECORD_TYPE); |
| edge->indirect_info->polymorphic = true; |
| edge->indirect_info->param_index = -1; |
| edge->indirect_info->otr_token = otr_token; |
| edge->indirect_info->otr_type = otr_type; |
| edge->indirect_info->outer_type = context.outer_type; |
| edge->indirect_info->offset = context.offset; |
| edge->indirect_info->maybe_in_construction |
| = context.maybe_in_construction; |
| edge->indirect_info->maybe_derived_type = context.maybe_derived_type; |
| } |
| |
| edge->next_callee = caller->indirect_calls; |
| if (caller->indirect_calls) |
| caller->indirect_calls->prev_callee = edge; |
| caller->indirect_calls = edge; |
| |
| return edge; |
| } |
| |
| /* Remove the edge E from the list of the callers of the callee. */ |
| |
| static inline void |
| cgraph_edge_remove_callee (struct cgraph_edge *e) |
| { |
| gcc_assert (!e->indirect_unknown_callee); |
| if (e->prev_caller) |
| e->prev_caller->next_caller = e->next_caller; |
| if (e->next_caller) |
| e->next_caller->prev_caller = e->prev_caller; |
| if (!e->prev_caller) |
| e->callee->callers = e->next_caller; |
| } |
| |
| /* Remove the edge E from the list of the callees of the caller. */ |
| |
| static inline void |
| cgraph_edge_remove_caller (struct cgraph_edge *e) |
| { |
| if (e->prev_callee) |
| e->prev_callee->next_callee = e->next_callee; |
| if (e->next_callee) |
| e->next_callee->prev_callee = e->prev_callee; |
| if (!e->prev_callee) |
| { |
| if (e->indirect_unknown_callee) |
| e->caller->indirect_calls = e->next_callee; |
| else |
| e->caller->callees = e->next_callee; |
| } |
| if (e->caller->call_site_hash && e->call_stmt) |
| htab_remove_elt_with_hash (e->caller->call_site_hash, |
| e->call_stmt, |
| htab_hash_pointer (e->call_stmt)); |
| } |
| |
| /* Put the edge onto the free list. */ |
| |
| static void |
| cgraph_free_edge (struct cgraph_edge *e) |
| { |
| int uid = e->uid; |
| |
| if (e->indirect_info) |
| ggc_free (e->indirect_info); |
| |
| /* Clear out the edge so we do not dangle pointers. */ |
| memset (e, 0, sizeof (*e)); |
| e->uid = uid; |
| NEXT_FREE_EDGE (e) = free_edges; |
| free_edges = e; |
| } |
| |
| /* Remove the edge E in the cgraph. */ |
| |
| void |
| cgraph_remove_edge (struct cgraph_edge *e) |
| { |
| /* Call all edge removal hooks. */ |
| cgraph_call_edge_removal_hooks (e); |
| |
| if (!e->indirect_unknown_callee) |
| /* Remove from callers list of the callee. */ |
| cgraph_edge_remove_callee (e); |
| |
| /* Remove from callees list of the callers. */ |
| cgraph_edge_remove_caller (e); |
| |
| /* Put the edge onto the free list. */ |
| cgraph_free_edge (e); |
| } |
| |
| /* Remove fake cgraph edges for indirect calls. NODE is the callee |
| of the edges. */ |
| |
| void |
| cgraph_remove_fake_indirect_call_in_edges (struct cgraph_node *node) |
| { |
| struct cgraph_edge *f, *e; |
| |
| if (!L_IPO_COMP_MODE) |
| return; |
| |
| for (e = node->callers; e; e = f) |
| { |
| f = e->next_caller; |
| if (!e->call_stmt) |
| cgraph_remove_edge (e); |
| } |
| } |
| |
| |
| /* Set callee of call graph edge E and add it to the corresponding set of |
| callers. */ |
| |
| static void |
| cgraph_set_edge_callee (struct cgraph_edge *e, struct cgraph_node *n) |
| { |
| e->prev_caller = NULL; |
| if (n->callers) |
| n->callers->prev_caller = e; |
| e->next_caller = n->callers; |
| n->callers = e; |
| e->callee = n; |
| } |
| |
| /* Turn edge E into speculative call calling N2. Update |
| the profile so the direct call is taken COUNT times |
| with FREQUENCY. |
| |
| At clone materialization time, the indirect call E will |
| be expanded as: |
| |
| if (call_dest == N2) |
| n2 (); |
| else |
| call call_dest |
| |
| At this time the function just creates the direct call, |
| the referencd representing the if conditional and attaches |
| them all to the orginal indirect call statement. |
| |
| Return direct edge created. */ |
| |
| struct cgraph_edge * |
| cgraph_turn_edge_to_speculative (struct cgraph_edge *e, |
| struct cgraph_node *n2, |
| gcov_type direct_count, |
| int direct_frequency) |
| { |
| struct cgraph_node *n = e->caller; |
| struct ipa_ref *ref; |
| struct cgraph_edge *e2; |
| |
| if (dump_file) |
| { |
| fprintf (dump_file, "Indirect call -> speculative call" |
| " %s/%i => %s/%i\n", |
| xstrdup (n->name ()), n->order, |
| xstrdup (n2->name ()), n2->order); |
| } |
| e->speculative = true; |
| e2 = cgraph_create_edge (n, n2, e->call_stmt, direct_count, direct_frequency); |
| initialize_inline_failed (e2); |
| e2->speculative = true; |
| if (TREE_NOTHROW (n2->decl)) |
| e2->can_throw_external = false; |
| else |
| e2->can_throw_external = e->can_throw_external; |
| e2->lto_stmt_uid = e->lto_stmt_uid; |
| e->count -= e2->count; |
| e->frequency -= e2->frequency; |
| cgraph_call_edge_duplication_hooks (e, e2); |
| ref = ipa_record_reference (n, n2, |
| IPA_REF_ADDR, e->call_stmt); |
| ref->lto_stmt_uid = e->lto_stmt_uid; |
| ref->speculative = e->speculative; |
| cgraph_mark_address_taken_node (n2); |
| return e2; |
| } |
| |
| /* Speculative call consist of three components: |
| 1) an indirect edge representing the original call |
| 2) an direct edge representing the new call |
| 3) ADDR_EXPR reference representing the speculative check. |
| All three components are attached to single statement (the indirect |
| call) and if one of them exists, all of them must exist. |
| |
| Given speculative call edge E, return all three components. |
| */ |
| |
| void |
| cgraph_speculative_call_info (struct cgraph_edge *e, |
| struct cgraph_edge *&direct, |
| struct cgraph_edge *&indirect, |
| struct ipa_ref *&reference) |
| { |
| struct ipa_ref *ref; |
| int i; |
| struct cgraph_edge *e2; |
| |
| if (!e->indirect_unknown_callee) |
| for (e2 = e->caller->indirect_calls; |
| e2->call_stmt != e->call_stmt || e2->lto_stmt_uid != e->lto_stmt_uid; |
| e2 = e2->next_callee) |
| ; |
| else |
| { |
| e2 = e; |
| /* We can take advantage of the call stmt hash. */ |
| if (e2->call_stmt) |
| { |
| e = cgraph_edge (e->caller, e2->call_stmt); |
| gcc_assert (e->speculative && !e->indirect_unknown_callee); |
| } |
| else |
| for (e = e->caller->callees; |
| e2->call_stmt != e->call_stmt |
| || e2->lto_stmt_uid != e->lto_stmt_uid; |
| e = e->next_callee) |
| ; |
| } |
| gcc_assert (e->speculative && e2->speculative); |
| direct = e; |
| indirect = e2; |
| |
| reference = NULL; |
| for (i = 0; ipa_ref_list_reference_iterate (&e->caller->ref_list, |
| i, ref); i++) |
| if (ref->speculative |
| && ((ref->stmt && ref->stmt == e->call_stmt) |
| || (!ref->stmt && ref->lto_stmt_uid == e->lto_stmt_uid))) |
| { |
| reference = ref; |
| break; |
| } |
| |
| /* Speculative edge always consist of all three components - direct edge, |
| indirect and reference. */ |
| |
| gcc_assert (e && e2 && ref); |
| } |
| |
| /* Redirect callee of E to N. The function does not update underlying |
| call expression. */ |
| |
| void |
| cgraph_redirect_edge_callee (struct cgraph_edge *e, struct cgraph_node *n) |
| { |
| /* Remove from callers list of the current callee. */ |
| cgraph_edge_remove_callee (e); |
| |
| /* Insert to callers list of the new callee. */ |
| cgraph_set_edge_callee (e, n); |
| } |
| |
| /* Speculative call EDGE turned out to be direct call to CALLE_DECL. |
| Remove the speculative call sequence and return edge representing the call. |
| It is up to caller to redirect the call as appropriate. */ |
| |
| struct cgraph_edge * |
| cgraph_resolve_speculation (struct cgraph_edge *edge, tree callee_decl) |
| { |
| struct cgraph_edge *e2; |
| struct ipa_ref *ref; |
| |
| gcc_assert (edge->speculative); |
| cgraph_speculative_call_info (edge, e2, edge, ref); |
| if (!callee_decl |
| || !symtab_semantically_equivalent_p (ref->referred, |
| symtab_get_node (callee_decl))) |
| { |
| if (dump_file) |
| { |
| if (callee_decl) |
| { |
| fprintf (dump_file, "Speculative indirect call %s/%i => %s/%i has " |
| "turned out to have contradicting known target ", |
| xstrdup (edge->caller->name ()), edge->caller->order, |
| xstrdup (e2->callee->name ()), e2->callee->order); |
| print_generic_expr (dump_file, callee_decl, 0); |
| fprintf (dump_file, "\n"); |
| } |
| else |
| { |
| fprintf (dump_file, "Removing speculative call %s/%i => %s/%i\n", |
| xstrdup (edge->caller->name ()), edge->caller->order, |
| xstrdup (e2->callee->name ()), e2->callee->order); |
| } |
| } |
| } |
| else |
| { |
| struct cgraph_edge *tmp = edge; |
| if (dump_file) |
| fprintf (dump_file, "Speculative call turned into direct call.\n"); |
| edge = e2; |
| e2 = tmp; |
| /* FIXME: If EDGE is inlined, we should scale up the frequencies and counts |
| in the functions inlined through it. */ |
| } |
| edge->count += e2->count; |
| edge->frequency += e2->frequency; |
| if (edge->frequency > CGRAPH_FREQ_MAX) |
| edge->frequency = CGRAPH_FREQ_MAX; |
| edge->speculative = false; |
| e2->speculative = false; |
| ipa_remove_reference (ref); |
| if (e2->indirect_unknown_callee || e2->inline_failed) |
| cgraph_remove_edge (e2); |
| else |
| cgraph_remove_node_and_inline_clones (e2->callee, NULL); |
| if (edge->caller->call_site_hash) |
| cgraph_update_edge_in_call_site_hash (edge); |
| return edge; |
| } |
| |
| /* Make an indirect EDGE with an unknown callee an ordinary edge leading to |
| CALLEE. DELTA is an integer constant that is to be added to the this |
| pointer (first parameter) to compensate for skipping a thunk adjustment. */ |
| |
| struct cgraph_edge * |
| cgraph_make_edge_direct (struct cgraph_edge *edge, struct cgraph_node *callee) |
| { |
| gcc_assert (edge->indirect_unknown_callee); |
| |
| /* If we are redirecting speculative call, make it non-speculative. */ |
| if (edge->indirect_unknown_callee && edge->speculative) |
| { |
| edge = cgraph_resolve_speculation (edge, callee->decl); |
| |
| /* On successful speculation just return the pre existing direct edge. */ |
| if (!edge->indirect_unknown_callee) |
| return edge; |
| } |
| |
| edge->indirect_unknown_callee = 0; |
| ggc_free (edge->indirect_info); |
| edge->indirect_info = NULL; |
| |
| /* Get the edge out of the indirect edge list. */ |
| if (edge->prev_callee) |
| edge->prev_callee->next_callee = edge->next_callee; |
| if (edge->next_callee) |
| edge->next_callee->prev_callee = edge->prev_callee; |
| if (!edge->prev_callee) |
| edge->caller->indirect_calls = edge->next_callee; |
| |
| /* Put it into the normal callee list */ |
| edge->prev_callee = NULL; |
| edge->next_callee = edge->caller->callees; |
| if (edge->caller->callees) |
| edge->caller->callees->prev_callee = edge; |
| edge->caller->callees = edge; |
| |
| /* Insert to callers list of the new callee. */ |
| cgraph_set_edge_callee (edge, callee); |
| |
| if (edge->call_stmt) |
| edge->call_stmt_cannot_inline_p |
| = !gimple_check_call_matching_types (edge->call_stmt, callee->decl, |
| false); |
| |
| /* We need to re-determine the inlining status of the edge. */ |
| initialize_inline_failed (edge); |
| return edge; |
| } |
| |
| /* If necessary, change the function declaration in the call statement |
| associated with E so that it corresponds to the edge callee. */ |
| |
| gimple |
| cgraph_redirect_edge_call_stmt_to_callee (struct cgraph_edge *e) |
| { |
| tree decl = gimple_call_fndecl (e->call_stmt); |
| tree lhs = gimple_call_lhs (e->call_stmt); |
| gimple new_stmt; |
| gimple_stmt_iterator gsi; |
| #ifdef ENABLE_CHECKING |
| struct cgraph_node *node; |
| #endif |
| |
| if (e->speculative) |
| { |
| struct cgraph_edge *e2; |
| gimple new_stmt; |
| struct ipa_ref *ref; |
| |
| cgraph_speculative_call_info (e, e, e2, ref); |
| /* If there already is an direct call (i.e. as a result of inliner's |
| substitution), forget about speculating. */ |
| if (decl) |
| e = cgraph_resolve_speculation (e, decl); |
| /* If types do not match, speculation was likely wrong. |
| The direct edge was posisbly redirected to the clone with a different |
| signature. We did not update the call statement yet, so compare it |
| with the reference that still points to the proper type. */ |
| else if (!gimple_check_call_matching_types (e->call_stmt, |
| ref->referred->decl, |
| true)) |
| { |
| if (dump_file) |
| fprintf (dump_file, "Not expanding speculative call of %s/%i -> %s/%i\n" |
| "Type mismatch.\n", |
| xstrdup (e->caller->name ()), |
| e->caller->order, |
| xstrdup (e->callee->name ()), |
| e->callee->order); |
| e = cgraph_resolve_speculation (e, NULL); |
| /* We are producing the final function body and will throw away the |
| callgraph edges really soon. Reset the counts/frequencies to |
| keep verifier happy in the case of roundoff errors. */ |
| e->count = gimple_bb (e->call_stmt)->count; |
| e->frequency = compute_call_stmt_bb_frequency |
| (e->caller->decl, gimple_bb (e->call_stmt)); |
| } |
| /* Expand speculation into GIMPLE code. */ |
| else |
| { |
| if (dump_file) |
| fprintf (dump_file, |
| "Expanding speculative call of %s/%i -> %s/%i count:" |
| HOST_WIDEST_INT_PRINT_DEC"\n", |
| xstrdup (e->caller->name ()), |
| e->caller->order, |
| xstrdup (e->callee->name ()), |
| e->callee->order, |
| (HOST_WIDEST_INT)e->count); |
| gcc_assert (e2->speculative); |
| push_cfun (DECL_STRUCT_FUNCTION (e->caller->decl)); |
| new_stmt = gimple_ic (e->call_stmt, cgraph (ref->referred), |
| e->count || e2->count |
| ? RDIV (e->count * REG_BR_PROB_BASE, |
| e->count + e2->count) |
| : e->frequency || e2->frequency |
| ? RDIV (e->frequency * REG_BR_PROB_BASE, |
| e->frequency + e2->frequency) |
| : REG_BR_PROB_BASE / 2, |
| e->count, e->count + e2->count); |
| e->speculative = false; |
| cgraph_set_call_stmt_including_clones (e->caller, e->call_stmt, |
| new_stmt, false); |
| e->frequency = compute_call_stmt_bb_frequency |
| (e->caller->decl, gimple_bb (e->call_stmt)); |
| e2->frequency = compute_call_stmt_bb_frequency |
| (e2->caller->decl, gimple_bb (e2->call_stmt)); |
| e2->speculative = false; |
| ref->speculative = false; |
| ref->stmt = NULL; |
| /* Indirect edges are not both in the call site hash. |
| get it updated. */ |
| if (e->caller->call_site_hash) |
| cgraph_update_edge_in_call_site_hash (e2); |
| pop_cfun (); |
| /* Continue redirecting E to proper target. */ |
| } |
| } |
| |
| if (e->indirect_unknown_callee |
| || decl == e->callee->decl) |
| return e->call_stmt; |
| |
| #ifdef ENABLE_CHECKING |
| if (decl) |
| { |
| node = cgraph_get_node (decl); |
| gcc_assert (!node || !node->clone.combined_args_to_skip); |
| } |
| #endif |
| |
| if (cgraph_dump_file) |
| { |
| fprintf (cgraph_dump_file, "updating call of %s/%i -> %s/%i: ", |
| xstrdup (e->caller->name ()), e->caller->order, |
| xstrdup (e->callee->name ()), e->callee->order); |
| print_gimple_stmt (cgraph_dump_file, e->call_stmt, 0, dump_flags); |
| if (e->callee->clone.combined_args_to_skip) |
| { |
| fprintf (cgraph_dump_file, " combined args to skip: "); |
| dump_bitmap (cgraph_dump_file, |
| e->callee->clone.combined_args_to_skip); |
| } |
| } |
| |
| if (e->callee->clone.combined_args_to_skip) |
| { |
| int lp_nr; |
| |
| new_stmt |
| = gimple_call_copy_skip_args (e->call_stmt, |
| e->callee->clone.combined_args_to_skip); |
| gimple_call_set_fndecl (new_stmt, e->callee->decl); |
| gimple_call_set_fntype (new_stmt, gimple_call_fntype (e->call_stmt)); |
| |
| if (gimple_vdef (new_stmt) |
| && TREE_CODE (gimple_vdef (new_stmt)) == SSA_NAME) |
| SSA_NAME_DEF_STMT (gimple_vdef (new_stmt)) = new_stmt; |
| |
| gsi = gsi_for_stmt (e->call_stmt); |
| gsi_replace (&gsi, new_stmt, false); |
| /* We need to defer cleaning EH info on the new statement to |
| fixup-cfg. We may not have dominator information at this point |
| and thus would end up with unreachable blocks and have no way |
| to communicate that we need to run CFG cleanup then. */ |
| lp_nr = lookup_stmt_eh_lp (e->call_stmt); |
| if (lp_nr != 0) |
| { |
| remove_stmt_from_eh_lp (e->call_stmt); |
| add_stmt_to_eh_lp (new_stmt, lp_nr); |
| } |
| } |
| else |
| { |
| new_stmt = e->call_stmt; |
| gimple_call_set_fndecl (new_stmt, e->callee->decl); |
| update_stmt_fn (DECL_STRUCT_FUNCTION (e->caller->decl), new_stmt); |
| if (L_IPO_COMP_MODE) |
| { |
| int lp_nr = lookup_stmt_eh_lp (e->call_stmt); |
| if (lp_nr != 0 && !stmt_could_throw_p (e->call_stmt)) |
| remove_stmt_from_eh_lp (e->call_stmt); |
| } |
| } |
| |
| /* If the call becomes noreturn, remove the lhs. */ |
| if (lhs && (gimple_call_flags (new_stmt) & ECF_NORETURN)) |
| { |
| if (TREE_CODE (lhs) == SSA_NAME) |
| { |
| tree var = create_tmp_reg_fn (DECL_STRUCT_FUNCTION (e->caller->decl), |
| TREE_TYPE (lhs), NULL); |
| var = get_or_create_ssa_default_def |
| (DECL_STRUCT_FUNCTION (e->caller->decl), var); |
| gimple set_stmt = gimple_build_assign (lhs, var); |
| gsi = gsi_for_stmt (new_stmt); |
| gsi_insert_before_without_update (&gsi, set_stmt, GSI_SAME_STMT); |
| update_stmt_fn (DECL_STRUCT_FUNCTION (e->caller->decl), set_stmt); |
| } |
| gimple_call_set_lhs (new_stmt, NULL_TREE); |
| update_stmt_fn (DECL_STRUCT_FUNCTION (e->caller->decl), new_stmt); |
| } |
| |
| /* If new callee has no static chain, remove it. */ |
| if (gimple_call_chain (new_stmt) && !DECL_STATIC_CHAIN (e->callee->decl)) |
| { |
| gimple_call_set_chain (new_stmt, NULL); |
| update_stmt_fn (DECL_STRUCT_FUNCTION (e->caller->decl), new_stmt); |
| } |
| |
| cgraph_set_call_stmt_including_clones (e->caller, e->call_stmt, new_stmt, false); |
| |
| if (cgraph_dump_file) |
| { |
| fprintf (cgraph_dump_file, " updated to:"); |
| print_gimple_stmt (cgraph_dump_file, e->call_stmt, 0, dump_flags); |
| } |
| return new_stmt; |
| } |
| |
| /* Update or remove the corresponding cgraph edge if a GIMPLE_CALL |
| OLD_STMT changed into NEW_STMT. OLD_CALL is gimple_call_fndecl |
| of OLD_STMT if it was previously call statement. |
| If NEW_STMT is NULL, the call has been dropped without any |
| replacement. */ |
| |
| static void |
| cgraph_update_edges_for_call_stmt_node (struct cgraph_node *node, |
| gimple old_stmt, tree old_call, |
| gimple new_stmt) |
| { |
| tree new_call = (new_stmt && is_gimple_call (new_stmt)) |
| ? gimple_call_fndecl (new_stmt) : 0; |
| |
| /* We are seeing indirect calls, then there is nothing to update. */ |
| if (!new_call && !old_call) |
| return; |
| /* See if we turned indirect call into direct call or folded call to one builtin |
| into different builtin. */ |
| if (old_call != new_call) |
| { |
| struct cgraph_edge *e = cgraph_edge (node, old_stmt); |
| struct cgraph_edge *ne = NULL; |
| gcov_type count; |
| int frequency; |
| |
| if (e) |
| { |
| /* See if the edge is already there and has the correct callee. It |
| might be so because of indirect inlining has already updated |
| it. We also might've cloned and redirected the edge. */ |
| if (new_call && e->callee) |
| { |
| struct cgraph_node *callee = e->callee; |
| while (callee) |
| { |
| if (callee->decl == new_call |
| || callee->former_clone_of == new_call) |
| { |
| cgraph_set_call_stmt (e, new_stmt); |
| return; |
| } |
| callee = callee->clone_of; |
| } |
| } |
| |
| /* Otherwise remove edge and create new one; we can't simply redirect |
| since function has changed, so inline plan and other information |
| attached to edge is invalid. */ |
| count = e->count; |
| frequency = e->frequency; |
| if (e->indirect_unknown_callee || e->inline_failed) |
| cgraph_remove_edge (e); |
| else |
| cgraph_remove_node_and_inline_clones (e->callee, NULL); |
| } |
| else if (new_call) |
| { |
| /* We are seeing new direct call; compute profile info based on BB. */ |
| basic_block bb = gimple_bb (new_stmt); |
| count = bb->count; |
| frequency = compute_call_stmt_bb_frequency (current_function_decl, |
| bb); |
| } |
| |
| if (new_call) |
| { |
| ne = cgraph_create_edge (node, cgraph_get_create_node (new_call), |
| new_stmt, count, frequency); |
| gcc_assert (ne->inline_failed); |
| } |
| } |
| /* We only updated the call stmt; update pointer in cgraph edge.. */ |
| else if (old_stmt != new_stmt) |
| cgraph_set_call_stmt (cgraph_edge (node, old_stmt), new_stmt); |
| } |
| |
| /* Update or remove the corresponding cgraph edge if a GIMPLE_CALL |
| OLD_STMT changed into NEW_STMT. OLD_DECL is gimple_call_fndecl |
| of OLD_STMT before it was updated (updating can happen inplace). */ |
| |
| void |
| cgraph_update_edges_for_call_stmt (gimple old_stmt, tree old_decl, gimple new_stmt) |
| { |
| struct cgraph_node *orig = cgraph_get_node (cfun->decl); |
| struct cgraph_node *node; |
| |
| gcc_checking_assert (orig); |
| cgraph_update_edges_for_call_stmt_node (orig, old_stmt, old_decl, new_stmt); |
| if (orig->clones) |
| for (node = orig->clones; node != orig;) |
| { |
| cgraph_update_edges_for_call_stmt_node (node, old_stmt, old_decl, new_stmt); |
| if (node->clones) |
| node = node->clones; |
| else if (node->next_sibling_clone) |
| node = node->next_sibling_clone; |
| else |
| { |
| while (node != orig && !node->next_sibling_clone) |
| node = node->clone_of; |
| if (node != orig) |
| node = node->next_sibling_clone; |
| } |
| } |
| } |
| |
| |
| /* Remove all callees from the node. */ |
| |
| void |
| cgraph_node_remove_callees (struct cgraph_node *node) |
| { |
| struct cgraph_edge *e, *f; |
| |
| /* It is sufficient to remove the edges from the lists of callers of |
| the callees. The callee list of the node can be zapped with one |
| assignment. */ |
| for (e = node->callees; e; e = f) |
| { |
| f = e->next_callee; |
| cgraph_call_edge_removal_hooks (e); |
| if (!e->indirect_unknown_callee) |
| cgraph_edge_remove_callee (e); |
| cgraph_free_edge (e); |
| } |
| for (e = node->indirect_calls; e; e = f) |
| { |
| f = e->next_callee; |
| cgraph_call_edge_removal_hooks (e); |
| if (!e->indirect_unknown_callee) |
| cgraph_edge_remove_callee (e); |
| cgraph_free_edge (e); |
| } |
| node->indirect_calls = NULL; |
| node->callees = NULL; |
| if (node->call_site_hash) |
| { |
| htab_delete (node->call_site_hash); |
| node->call_site_hash = NULL; |
| } |
| } |
| |
| /* Remove all callers from the node. */ |
| |
| static void |
| cgraph_node_remove_callers (struct cgraph_node *node) |
| { |
| struct cgraph_edge *e, *f; |
| |
| /* It is sufficient to remove the edges from the lists of callees of |
| the callers. The caller list of the node can be zapped with one |
| assignment. */ |
| for (e = node->callers; e; e = f) |
| { |
| f = e->next_caller; |
| cgraph_call_edge_removal_hooks (e); |
| cgraph_edge_remove_caller (e); |
| cgraph_free_edge (e); |
| } |
| node->callers = NULL; |
| } |
| |
| /* Helper function for cgraph_release_function_body and free_lang_data. |
| It releases body from function DECL without having to inspect its |
| possibly non-existent symtab node. */ |
| |
| void |
| release_function_body (tree decl) |
| { |
| if (cgraph_get_node (decl) |
| && cgraph_is_aux_decl_external (cgraph_get_node (decl))) |
| DECL_EXTERNAL (decl) = 1; |
| |
| if (DECL_STRUCT_FUNCTION (decl)) |
| { |
| push_cfun (DECL_STRUCT_FUNCTION (decl)); |
| if (cfun->cfg |
| && current_loops) |
| { |
| cfun->curr_properties &= ~PROP_loops; |
| loop_optimizer_finalize (); |
| } |
| if (cfun->gimple_df) |
| { |
| delete_tree_ssa (); |
| delete_tree_cfg_annotations (); |
| cfun->eh = NULL; |
| } |
| if (cfun->cfg) |
| { |
| gcc_assert (dom_computed[0] == DOM_NONE); |
| gcc_assert (dom_computed[1] == DOM_NONE); |
| clear_edges (); |
| cfun->cfg = NULL; |
| } |
| if (cfun->value_histograms) |
| free_histograms (); |
| pop_cfun (); |
| gimple_set_body (decl, NULL); |
| /* Struct function hangs a lot of data that would leak if we didn't |
| removed all pointers to it. */ |
| ggc_free (DECL_STRUCT_FUNCTION (decl)); |
| DECL_STRUCT_FUNCTION (decl) = NULL; |
| } |
| DECL_SAVED_TREE (decl) = NULL; |
| } |
| |
| /* Release memory used to represent body of function NODE. |
| Use this only for functions that are released before being translated to |
| target code (i.e. RTL). Functions that are compiled to RTL and beyond |
| are free'd in final.c via free_after_compilation(). */ |
| |
| void |
| cgraph_release_function_body (struct cgraph_node *node) |
| { |
| node->ipa_transforms_to_apply.release (); |
| if (!node->used_as_abstract_origin && cgraph_state != CGRAPH_STATE_PARSING) |
| { |
| DECL_RESULT (node->decl) = NULL; |
| DECL_ARGUMENTS (node->decl) = NULL; |
| } |
| /* If the node is abstract and needed, then do not clear DECL_INITIAL |
| of its associated function function declaration because it's |
| needed to emit debug info later. */ |
| if (!node->used_as_abstract_origin && DECL_INITIAL (node->decl)) |
| DECL_INITIAL (node->decl) = error_mark_node; |
| release_function_body (node->decl); |
| if (node->lto_file_data) |
| lto_free_function_in_decl_state_for_node (node); |
| } |
| |
| /* Remove the node from cgraph. */ |
| |
| void |
| cgraph_remove_node (struct cgraph_node *node) |
| { |
| struct cgraph_node *n; |
| int uid = node->uid; |
| |
| cgraph_call_node_removal_hooks (node); |
| cgraph_node_remove_callers (node); |
| cgraph_node_remove_callees (node); |
| node->ipa_transforms_to_apply.release (); |
| |
| /* Incremental inlining access removed nodes stored in the postorder list. |
| */ |
| node->force_output = false; |
| node->forced_by_abi = false; |
| for (n = node->nested; n; n = n->next_nested) |
| n->origin = NULL; |
| node->nested = NULL; |
| if (node->origin) |
| { |
| struct cgraph_node **node2 = &node->origin->nested; |
| |
| while (*node2 != node) |
| node2 = &(*node2)->next_nested; |
| *node2 = node->next_nested; |
| } |
| symtab_unregister_node (node); |
| if (node->prev_sibling_clone) |
| node->prev_sibling_clone->next_sibling_clone = node->next_sibling_clone; |
| else if (node->clone_of) |
| node->clone_of->clones = node->next_sibling_clone; |
| if (node->next_sibling_clone) |
| node->next_sibling_clone->prev_sibling_clone = node->prev_sibling_clone; |
| if (node->clones) |
| { |
| struct cgraph_node *n, *next; |
| |
| if (node->clone_of) |
| { |
| for (n = node->clones; n->next_sibling_clone; n = n->next_sibling_clone) |
| n->clone_of = node->clone_of; |
| n->clone_of = node->clone_of; |
| n->next_sibling_clone = node->clone_of->clones; |
| if (node->clone_of->clones) |
| node->clone_of->clones->prev_sibling_clone = n; |
| node->clone_of->clones = node->clones; |
| } |
| else |
| { |
| /* We are removing node with clones. This makes clones inconsistent, |
| but assume they will be removed subsequently and just keep clone |
| tree intact. This can happen in unreachable function removal since |
| we remove unreachable functions in random order, not by bottom-up |
| walk of clone trees. */ |
| for (n = node->clones; n; n = next) |
| { |
| next = n->next_sibling_clone; |
| n->next_sibling_clone = NULL; |
| n->prev_sibling_clone = NULL; |
| n->clone_of = NULL; |
| } |
| } |
| } |
| |
| /* While all the clones are removed after being proceeded, the function |
| itself is kept in the cgraph even after it is compiled. Check whether |
| we are done with this body and reclaim it proactively if this is the case. |
| */ |
| if (cgraph_state != CGRAPH_LTO_STREAMING) |
| { |
| n = cgraph_get_node (node->decl); |
| if (!n |
| || (!n->clones && !n->clone_of && !n->global.inlined_to |
| && (cgraph_global_info_ready |
| && (TREE_ASM_WRITTEN (n->decl) |
| || DECL_EXTERNAL (n->decl) |
| || !n->analyzed |
| || (!flag_wpa && n->in_other_partition))))) |
| cgraph_release_function_body (node); |
| } |
| |
| cgraph_remove_link_node (node); |
| node->decl = NULL; |
| |
| if (node->call_site_hash) |
| { |
| htab_delete (node->call_site_hash); |
| node->call_site_hash = NULL; |
| } |
| cgraph_n_nodes--; |
| |
| /* Clear out the node to NULL all pointers and add the node to the free |
| list. */ |
| memset (node, 0, sizeof (*node)); |
| node->type = SYMTAB_FUNCTION; |
| node->uid = uid; |
| SET_NEXT_FREE_NODE (node, free_nodes); |
| free_nodes = node; |
| } |
| |
| |
| /* Likewise indicate that a node is having address taken. */ |
| |
| void |
| cgraph_mark_address_taken_node (struct cgraph_node *node) |
| { |
| /* Indirect inlining can figure out that all uses of the address are |
| inlined. */ |
| if (node->global.inlined_to) |
| { |
| gcc_assert (cfun->after_inlining); |
| gcc_assert (node->callers->indirect_inlining_edge); |
| return; |
| } |
| /* FIXME: address_taken flag is used both as a shortcut for testing whether |
| IPA_REF_ADDR reference exists (and thus it should be set on node |
| representing alias we take address of) and as a test whether address |
| of the object was taken (and thus it should be set on node alias is |
| referring to). We should remove the first use and the remove the |
| following set. */ |
| node->address_taken = 1; |
| node = cgraph_function_or_thunk_node (node, NULL); |
| node->address_taken = 1; |
| } |
| |
| /* Return local info for the compiled function. */ |
| |
| struct cgraph_local_info * |
| cgraph_local_info (tree decl) |
| { |
| struct cgraph_node *node; |
| |
| gcc_assert (TREE_CODE (decl) == FUNCTION_DECL); |
| node = cgraph_get_node (decl); |
| if (!node) |
| return NULL; |
| return &node->local; |
| } |
| |
| /* Return local info for the compiled function. */ |
| |
| struct cgraph_global_info * |
| cgraph_global_info (tree decl) |
| { |
| struct cgraph_node *node; |
| |
| gcc_assert (TREE_CODE (decl) == FUNCTION_DECL && cgraph_global_info_ready); |
| node = cgraph_get_node (decl); |
| if (!node) |
| return NULL; |
| return &node->global; |
| } |
| |
| /* Return local info for the compiled function. */ |
| |
| struct cgraph_rtl_info * |
| cgraph_rtl_info (tree decl) |
| { |
| struct cgraph_node *node; |
| |
| gcc_assert (TREE_CODE (decl) == FUNCTION_DECL); |
| node = cgraph_get_node (decl); |
| if (!node |
| || (decl != current_function_decl |
| && !TREE_ASM_WRITTEN (node->decl))) |
| return NULL; |
| return &node->rtl; |
| } |
| |
| /* Return a string describing the failure REASON. */ |
| |
| const char* |
| cgraph_inline_failed_string (cgraph_inline_failed_t reason) |
| { |
| #undef DEFCIFCODE |
| #define DEFCIFCODE(code, type, string) string, |
| |
| static const char *cif_string_table[CIF_N_REASONS] = { |
| #include "cif-code.def" |
| }; |
| |
| /* Signedness of an enum type is implementation defined, so cast it |
| to unsigned before testing. */ |
| gcc_assert ((unsigned) reason < CIF_N_REASONS); |
| return cif_string_table[reason]; |
| } |
| |
| /* Return a type describing the failure REASON. */ |
| |
| cgraph_inline_failed_type_t |
| cgraph_inline_failed_type (cgraph_inline_failed_t reason) |
| { |
| #undef DEFCIFCODE |
| #define DEFCIFCODE(code, type, string) type, |
| |
| static cgraph_inline_failed_type_t cif_type_table[CIF_N_REASONS] = { |
| #include "cif-code.def" |
| }; |
| |
| /* Signedness of an enum type is implementation defined, so cast it |
| to unsigned before testing. */ |
| gcc_assert ((unsigned) reason < CIF_N_REASONS); |
| return cif_type_table[reason]; |
| } |
| |
| /* Names used to print out the availability enum. */ |
| const char * const cgraph_availability_names[] = |
| {"unset", "not_available", "overwritable", "available", "local"}; |
| |
| |
| /* Dump call graph node NODE to file F. */ |
| |
| void |
| dump_cgraph_node (FILE *f, struct cgraph_node *node) |
| { |
| struct cgraph_edge *edge; |
| int indirect_calls_count = 0; |
| |
| dump_symtab_base (f, node); |
| |
| if (node->global.inlined_to) |
| fprintf (f, " Function %s/%i is inline copy in %s/%i\n", |
| xstrdup (node->name ()), |
| node->order, |
| xstrdup (node->global.inlined_to->name ()), |
| node->global.inlined_to->order); |
| if (node->clone_of) |
| fprintf (f, " Clone of %s/%i\n", |
| node->clone_of->asm_name (), |
| node->clone_of->order); |
| if (cgraph_function_flags_ready) |
| fprintf (f, " Availability: %s\n", |
| cgraph_availability_names [cgraph_function_body_availability (node)]); |
| |
| if (node->profile_id) |
| fprintf (f, " Profile id: %i\n", |
| node->profile_id); |
| fprintf (f, " First run: %i\n", node->tp_first_run); |
| fprintf (f, " Function flags:"); |
| if (node->count) |
| fprintf (f, " executed "HOST_WIDEST_INT_PRINT_DEC"x", |
| (HOST_WIDEST_INT)node->count); |
| if (node->max_bb_count) |
| fprintf (f, " hottest bb executed "HOST_WIDEST_INT_PRINT_DEC"x", |
| (HOST_WIDEST_INT)node->max_bb_count); |
| if (node->origin) |
| fprintf (f, " nested in: %s", node->origin->asm_name ()); |
| if (gimple_has_body_p (node->decl)) |
| fprintf (f, " body"); |
| if (node->process) |
| fprintf (f, " process"); |
| if (node->local.local) |
| fprintf (f, " local"); |
| if (node->local.redefined_extern_inline) |
| fprintf (f, " redefined_extern_inline"); |
| if (node->only_called_at_startup) |
| fprintf (f, " only_called_at_startup"); |
| if (node->only_called_at_exit) |
| fprintf (f, " only_called_at_exit"); |
| if (node->tm_clone) |
| fprintf (f, " tm_clone"); |
| |
| fprintf (f, "\n"); |
| |
| if (node->thunk.thunk_p) |
| { |
| fprintf (f, " Thunk"); |
| if (node->thunk.alias) |
| fprintf (f, " of %s (asm: %s)", |
| lang_hooks.decl_printable_name (node->thunk.alias, 2), |
| IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (node->thunk.alias))); |
| fprintf (f, " fixed offset %i virtual value %i has " |
| "virtual offset %i)\n", |
| (int)node->thunk.fixed_offset, |
| (int)node->thunk.virtual_value, |
| (int)node->thunk.virtual_offset_p); |
| } |
| if (node->alias && node->thunk.alias |
| && DECL_P (node->thunk.alias)) |
| { |
| fprintf (f, " Alias of %s", |
| lang_hooks.decl_printable_name (node->thunk.alias, 2)); |
| if (DECL_ASSEMBLER_NAME_SET_P (node->thunk.alias)) |
| fprintf (f, " (asm: %s)", |
| IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (node->thunk.alias))); |
| fprintf (f, "\n"); |
| } |
| |
| fprintf (f, " Called by: "); |
| |
| for (edge = node->callers; edge; edge = edge->next_caller) |
| { |
| fprintf (f, "%s/%i ", edge->caller->asm_name (), |
| edge->caller->order); |
| if (edge->count) |
| fprintf (f, "("HOST_WIDEST_INT_PRINT_DEC"x) ", |
| (HOST_WIDEST_INT)edge->count); |
| if (edge->frequency) |
| fprintf (f, "(%.2f per call) ", |
| edge->frequency / (double)CGRAPH_FREQ_BASE); |
| if (edge->speculative) |
| fprintf (f, "(speculative) "); |
| if (!edge->inline_failed) |
| fprintf (f, "(inlined) "); |
| if (edge->indirect_inlining_edge) |
| fprintf (f, "(indirect_inlining) "); |
| if (edge->can_throw_external) |
| fprintf (f, "(can throw external) "); |
| } |
| |
| fprintf (f, "\n Calls: "); |
| for (edge = node->callees; edge; edge = edge->next_callee) |
| { |
| fprintf (f, "%s/%i ", edge->callee->asm_name (), |
| edge->callee->order); |
| if (edge->speculative) |
| fprintf (f, "(speculative) "); |
| if (!edge->inline_failed) |
| fprintf (f, "(inlined) "); |
| if (edge->indirect_inlining_edge) |
| fprintf (f, "(indirect_inlining) "); |
| if (edge->count) |
| fprintf (f, "("HOST_WIDEST_INT_PRINT_DEC"x) ", |
| (HOST_WIDEST_INT)edge->count); |
| if (edge->frequency) |
| fprintf (f, "(%.2f per call) ", |
| edge->frequency / (double)CGRAPH_FREQ_BASE); |
| if (edge->can_throw_external) |
| fprintf (f, "(can throw external) "); |
| } |
| fprintf (f, "\n"); |
| |
| for (edge = node->indirect_calls; edge; edge = edge->next_callee) |
| indirect_calls_count++; |
| if (indirect_calls_count) |
| fprintf (f, " Has %i outgoing edges for indirect calls.\n", |
| indirect_calls_count); |
| } |
| |
| |
| /* Dump call graph node NODE to stderr. */ |
| |
| DEBUG_FUNCTION void |
| debug_cgraph_node (struct cgraph_node *node) |
| { |
| dump_cgraph_node (stderr, node); |
| } |
| |
| |
| /* Dump the callgraph to file F. */ |
| |
| void |
| dump_cgraph (FILE *f) |
| { |
| struct cgraph_node *node; |
| |
| fprintf (f, "callgraph:\n\n"); |
| FOR_EACH_FUNCTION (node) |
| dump_cgraph_node (f, node); |
| } |
| |
| |
| /* Dump the call graph to stderr. */ |
| |
| DEBUG_FUNCTION void |
| debug_cgraph (void) |
| { |
| dump_cgraph (stderr); |
| } |
| |
| /* Return true when the DECL can possibly be inlined. */ |
| bool |
| cgraph_function_possibly_inlined_p (tree decl) |
| { |
| if (!cgraph_global_info_ready) |
| return !DECL_UNINLINABLE (decl); |
| return DECL_POSSIBLY_INLINED (decl); |
| } |
| |
| /* NODE is no longer nested function; update cgraph accordingly. */ |
| void |
| cgraph_unnest_node (struct cgraph_node *node) |
| { |
| struct cgraph_node **node2 = &node->origin->nested; |
| gcc_assert (node->origin); |
| |
| while (*node2 != node) |
| node2 = &(*node2)->next_nested; |
| *node2 = node->next_nested; |
| node->origin = NULL; |
| } |
| |
| /* Return function availability. See cgraph.h for description of individual |
| return values. */ |
| enum availability |
| cgraph_function_body_availability (struct cgraph_node *node) |
| { |
| enum availability avail; |
| if (!node->analyzed) |
| avail = AVAIL_NOT_AVAILABLE; |
| else if (node->local.local) |
| avail = AVAIL_LOCAL; |
| else if (node->alias && node->weakref) |
| cgraph_function_or_thunk_node (node, &avail); |
| else if (lookup_attribute ("ifunc", DECL_ATTRIBUTES (node->decl))) |
| avail = AVAIL_OVERWRITABLE; |
| else if (!node->externally_visible) |
| avail = AVAIL_AVAILABLE; |
| /* Inline functions are safe to be analyzed even if their symbol can |
| be overwritten at runtime. It is not meaningful to enforce any sane |
| behaviour on replacing inline function by different body. */ |
| else if (DECL_DECLARED_INLINE_P (node->decl)) |
| avail = AVAIL_AVAILABLE; |
| |
| /* If the function can be overwritten, return OVERWRITABLE. Take |
| care at least of two notable extensions - the COMDAT functions |
| used to share template instantiations in C++ (this is symmetric |
| to code cp_cannot_inline_tree_fn and probably shall be shared and |
| the inlinability hooks completely eliminated). |
| |
| ??? Does the C++ one definition rule allow us to always return |
| AVAIL_AVAILABLE here? That would be good reason to preserve this |
| bit. */ |
| |
| else if (decl_replaceable_p (node->decl) |
| && !DECL_EXTERNAL (node->decl)) |
| avail = AVAIL_OVERWRITABLE; |
| else avail = AVAIL_AVAILABLE; |
| |
| return avail; |
| } |
| |
| /* Worker for cgraph_node_can_be_local_p. */ |
| static bool |
| cgraph_node_cannot_be_local_p_1 (struct cgraph_node *node, |
| void *data ATTRIBUTE_UNUSED) |
| { |
| return !(!node->force_output |
| && ((DECL_COMDAT (node->decl) |
| && !node->forced_by_abi |
| && !symtab_used_from_object_file_p (node) |
| && !node->same_comdat_group) |
| || !node->externally_visible)); |
| } |
| |
| /* Return true if NODE can be made local for API change. |
| Extern inline functions and C++ COMDAT functions can be made local |
| at the expense of possible code size growth if function is used in multiple |
| compilation units. */ |
| bool |
| cgraph_node_can_be_local_p (struct cgraph_node *node) |
| { |
| return (!node->address_taken |
| && !cgraph_for_node_and_aliases (node, |
| cgraph_node_cannot_be_local_p_1, |
| NULL, true)); |
| } |
| |
| /* Call calback on NODE, thunks and aliases associated to NODE. |
| When INCLUDE_OVERWRITABLE is false, overwritable aliases and thunks are |
| skipped. */ |
| |
| bool |
| cgraph_for_node_thunks_and_aliases (struct cgraph_node *node, |
| bool (*callback) (struct cgraph_node *, void *), |
| void *data, |
| bool include_overwritable) |
| { |
| struct cgraph_edge *e; |
| int i; |
| struct ipa_ref *ref; |
| |
| if (callback (node, data)) |
| return true; |
| for (e = node->callers; e; e = e->next_caller) |
| if (e->caller->thunk.thunk_p |
| && (include_overwritable |
| || cgraph_function_body_availability (e->caller) > AVAIL_OVERWRITABLE)) |
| if (cgraph_for_node_thunks_and_aliases (e->caller, callback, data, |
| include_overwritable)) |
| return true; |
| for (i = 0; ipa_ref_list_referring_iterate (&node->ref_list, i, ref); i++) |
| if (ref->use == IPA_REF_ALIAS) |
| { |
| struct cgraph_node *alias = ipa_ref_referring_node (ref); |
| if (include_overwritable |
| || cgraph_function_body_availability (alias) > AVAIL_OVERWRITABLE) |
| if (cgraph_for_node_thunks_and_aliases (alias, callback, data, |
| include_overwritable)) |
| return true; |
| } |
| return false; |
| } |
| |
| /* Call calback on NODE and aliases associated to NODE. |
| When INCLUDE_OVERWRITABLE is false, overwritable aliases and thunks are |
| skipped. */ |
| |
| bool |
| cgraph_for_node_and_aliases (struct cgraph_node *node, |
| bool (*callback) (struct cgraph_node *, void *), |
| void *data, |
| bool include_overwritable) |
| { |
| int i; |
| struct ipa_ref *ref; |
| |
| if (callback (node, data)) |
| return true; |
| for (i = 0; ipa_ref_list_referring_iterate (&node->ref_list, i, ref); i++) |
| if (ref->use == IPA_REF_ALIAS) |
| { |
| struct cgraph_node *alias = ipa_ref_referring_node (ref); |
| if (include_overwritable |
| || cgraph_function_body_availability (alias) > AVAIL_OVERWRITABLE) |
| if (cgraph_for_node_and_aliases (alias, callback, data, |
| include_overwritable)) |
| return true; |
| } |
| return false; |
| } |
| |
| /* Worker to bring NODE local. */ |
| |
| static bool |
| cgraph_make_node_local_1 (struct cgraph_node *node, void *data ATTRIBUTE_UNUSED) |
| { |
| gcc_checking_assert (cgraph_node_can_be_local_p (node)); |
| if (DECL_COMDAT (node->decl) || DECL_EXTERNAL (node->decl)) |
| { |
| symtab_make_decl_local (node->decl); |
| |
| node->externally_visible = false; |
| node->forced_by_abi = false; |
| node->local.local = true; |
| node->unique_name = (node->resolution == LDPR_PREVAILING_DEF_IRONLY |
| || node->resolution == LDPR_PREVAILING_DEF_IRONLY_EXP); |
| node->resolution = LDPR_PREVAILING_DEF_IRONLY; |
| gcc_assert (cgraph_function_body_availability (node) == AVAIL_LOCAL); |
| } |
| return false; |
| } |
| |
| /* Bring NODE local. */ |
| |
| void |
| cgraph_make_node_local (struct cgraph_node *node) |
| { |
| cgraph_for_node_thunks_and_aliases (node, cgraph_make_node_local_1, |
| NULL, true); |
| } |
| |
| /* Worker to set nothrow flag. */ |
| |
| static bool |
| cgraph_set_nothrow_flag_1 (struct cgraph_node *node, void *data) |
| { |
| struct cgraph_edge *e; |
| |
| TREE_NOTHROW (node->decl) = data != NULL; |
| |
| if (data != NULL) |
| for (e = node->callers; e; e = e->next_caller) |
| e->can_throw_external = false; |
| return false; |
| } |
| |
| /* Set TREE_NOTHROW on NODE's decl and on aliases of NODE |
| if any to NOTHROW. */ |
| |
| void |
| cgraph_set_nothrow_flag (struct cgraph_node *node, bool nothrow) |
| { |
| cgraph_for_node_thunks_and_aliases (node, cgraph_set_nothrow_flag_1, |
| (void *)(size_t)nothrow, false); |
| } |
| |
| /* Worker to set const flag. */ |
| |
| static bool |
| cgraph_set_const_flag_1 (struct cgraph_node *node, void *data) |
| { |
| /* Static constructors and destructors without a side effect can be |
| optimized out. */ |
| if (data && !((size_t)data & 2)) |
| { |
| if (DECL_STATIC_CONSTRUCTOR (node->decl)) |
| DECL_STATIC_CONSTRUCTOR (node->decl) = 0; |
| if (DECL_STATIC_DESTRUCTOR (node->decl)) |
| DECL_STATIC_DESTRUCTOR (node->decl) = 0; |
| } |
| TREE_READONLY (node->decl) = data != NULL; |
| DECL_LOOPING_CONST_OR_PURE_P (node->decl) = ((size_t)data & 2) != 0; |
| return false; |
| } |
| |
| /* Set TREE_READONLY on NODE's decl and on aliases of NODE |
| if any to READONLY. */ |
| |
| void |
| cgraph_set_const_flag (struct cgraph_node *node, bool readonly, bool looping) |
| { |
| cgraph_for_node_thunks_and_aliases (node, cgraph_set_const_flag_1, |
| (void *)(size_t)(readonly + (int)looping * 2), |
| false); |
| } |
| |
| /* Worker to set pure flag. */ |
| |
| static bool |
| cgraph_set_pure_flag_1 (struct cgraph_node *node, void *data) |
| { |
| /* Static constructors and destructors without a side effect can be |
| optimized out. */ |
| if (data && !((size_t)data & 2)) |
| { |
| if (DECL_STATIC_CONSTRUCTOR (node->decl)) |
| DECL_STATIC_CONSTRUCTOR (node->decl) = 0; |
| if (DECL_STATIC_DESTRUCTOR (node->decl)) |
| DECL_STATIC_DESTRUCTOR (node->decl) = 0; |
| } |
| DECL_PURE_P (node->decl) = data != NULL; |
| DECL_LOOPING_CONST_OR_PURE_P (node->decl) = ((size_t)data & 2) != 0; |
| return false; |
| } |
| |
| /* Set DECL_PURE_P on NODE's decl and on aliases of NODE |
| if any to PURE. */ |
| |
| void |
| cgraph_set_pure_flag (struct cgraph_node *node, bool pure, bool looping) |
| { |
| cgraph_for_node_thunks_and_aliases (node, cgraph_set_pure_flag_1, |
| (void *)(size_t)(pure + (int)looping * 2), |
| false); |
| } |
| |
| /* Return true when NODE can not return or throw and thus |
| it is safe to ignore its side effects for IPA analysis. */ |
| |
| bool |
| cgraph_node_cannot_return (struct cgraph_node *node) |
| { |
| int flags = flags_from_decl_or_type (node->decl); |
| if (!flag_exceptions) |
| return (flags & ECF_NORETURN) != 0; |
| else |
| return ((flags & (ECF_NORETURN | ECF_NOTHROW)) |
| == (ECF_NORETURN | ECF_NOTHROW)); |
| } |
| |
| /* Return true when call of E can not lead to return from caller |
| and thus it is safe to ignore its side effects for IPA analysis |
| when computing side effects of the caller. |
| FIXME: We could actually mark all edges that have no reaching |
| patch to the exit block or throw to get better results. */ |
| bool |
| cgraph_edge_cannot_lead_to_return (struct cgraph_edge *e) |
| { |
| if (cgraph_node_cannot_return (e->caller)) |
| return true; |
| if (e->indirect_unknown_callee) |
| { |
| int flags = e->indirect_info->ecf_flags; |
| if (!flag_exceptions) |
| return (flags & ECF_NORETURN) != 0; |
| else |
| return ((flags & (ECF_NORETURN | ECF_NOTHROW)) |
| == (ECF_NORETURN | ECF_NOTHROW)); |
| } |
| else |
| return cgraph_node_cannot_return (e->callee); |
| } |
| |
| /* Return true when function NODE can be removed from callgraph |
| if all direct calls are eliminated. */ |
| |
| bool |
| cgraph_can_remove_if_no_direct_calls_and_refs_p (struct cgraph_node *node) |
| { |
| gcc_assert (!node->global.inlined_to); |
| /* Extern inlines can always go, we will use the external definition. */ |
| if (DECL_EXTERNAL (node->decl)) |
| return true; |
| /* When function is needed, we can not remove it. */ |
| if (node->force_output || node->used_from_other_partition) |
| return false; |
| if (DECL_STATIC_CONSTRUCTOR (node->decl) |
| || DECL_STATIC_DESTRUCTOR (node->decl)) |
| return false; |
| /* Only COMDAT functions can be removed if externally visible. */ |
| if (node->externally_visible |
| && (!DECL_COMDAT (node->decl) |
| || node->forced_by_abi |
| || symtab_used_from_object_file_p (node))) |
| return false; |
| return true; |
| } |
| |
| /* Worker for cgraph_can_remove_if_no_direct_calls_p. */ |
| |
| static bool |
| nonremovable_p (struct cgraph_node *node, void *data ATTRIBUTE_UNUSED) |
| { |
| return !cgraph_can_remove_if_no_direct_calls_and_refs_p (node); |
| } |
| |
| /* Return true when function NODE and its aliases can be removed from callgraph |
| if all direct calls are eliminated. */ |
| |
| bool |
| cgraph_can_remove_if_no_direct_calls_p (struct cgraph_node *node) |
| { |
| /* Extern inlines can always go, we will use the external definition. */ |
| if (DECL_EXTERNAL (node->decl)) |
| return true; |
| if (node->address_taken) |
| return false; |
| return !cgraph_for_node_and_aliases (node, nonremovable_p, NULL, true); |
| } |
| |
| /* Worker for cgraph_can_remove_if_no_direct_calls_p. */ |
| |
| static bool |
| used_from_object_file_p (struct cgraph_node *node, void *data ATTRIBUTE_UNUSED) |
| { |
| return symtab_used_from_object_file_p (node); |
| } |
| |
| /* Return true when function NODE can be expected to be removed |
| from program when direct calls in this compilation unit are removed. |
| |
| As a special case COMDAT functions are |
| cgraph_can_remove_if_no_direct_calls_p while the are not |
| cgraph_only_called_directly_p (it is possible they are called from other |
| unit) |
| |
| This function behaves as cgraph_only_called_directly_p because eliminating |
| all uses of COMDAT function does not make it necessarily disappear from |
| the program unless we are compiling whole program or we do LTO. In this |
| case we know we win since dynamic linking will not really discard the |
| linkonce section. */ |
| |
| bool |
| cgraph_will_be_removed_from_program_if_no_direct_calls (struct cgraph_node *node) |
| { |
| gcc_assert (!node->global.inlined_to); |
| if (cgraph_for_node_and_aliases (node, used_from_object_file_p, NULL, true)) |
| return false; |
| if (!in_lto_p && !flag_whole_program) |
| return cgraph_only_called_directly_p (node); |
| else |
| { |
| if (DECL_EXTERNAL (node->decl)) |
| return true; |
| return cgraph_can_remove_if_no_direct_calls_p (node); |
| } |
| } |
| |
| |
| /* Worker for cgraph_only_called_directly_p. */ |
| |
| static bool |
| cgraph_not_only_called_directly_p_1 (struct cgraph_node *node, void *data ATTRIBUTE_UNUSED) |
| { |
| return !cgraph_only_called_directly_or_aliased_p (node); |
| } |
| |
| /* Return true when function NODE and all its aliases are only called |
| directly. |
| i.e. it is not externally visible, address was not taken and |
| it is not used in any other non-standard way. */ |
| |
| bool |
| cgraph_only_called_directly_p (struct cgraph_node *node) |
| { |
| gcc_assert (cgraph_function_or_thunk_node (node, NULL) == node); |
| return !cgraph_for_node_and_aliases (node, cgraph_not_only_called_directly_p_1, |
| NULL, true); |
| } |
| |
| |
| /* Collect all callers of NODE. Worker for collect_callers_of_node. */ |
| |
| static bool |
| collect_callers_of_node_1 (struct cgraph_node *node, void *data) |
| { |
| vec<cgraph_edge_p> *redirect_callers = (vec<cgraph_edge_p> *)data; |
| struct cgraph_edge *cs; |
| enum availability avail; |
| cgraph_function_or_thunk_node (node, &avail); |
| |
| if (avail > AVAIL_OVERWRITABLE) |
| for (cs = node->callers; cs != NULL; cs = cs->next_caller) |
| if (!cs->indirect_inlining_edge) |
| redirect_callers->safe_push (cs); |
| return false; |
| } |
| |
| /* Collect all callers of NODE and its aliases that are known to lead to NODE |
| (i.e. are not overwritable). */ |
| |
| vec<cgraph_edge_p> |
| collect_callers_of_node (struct cgraph_node *node) |
| { |
| vec<cgraph_edge_p> redirect_callers = vNULL; |
| cgraph_for_node_and_aliases (node, collect_callers_of_node_1, |
| &redirect_callers, false); |
| return redirect_callers; |
| } |
| |
| /* Return TRUE if NODE2 a clone of NODE or is equivalent to it. */ |
| |
| static bool |
| clone_of_p (struct cgraph_node *node, struct cgraph_node *node2) |
| { |
| bool skipped_thunk = false; |
| node = cgraph_function_or_thunk_node (node, NULL); |
| node2 = cgraph_function_or_thunk_node (node2, NULL); |
| |
| /* There are no virtual clones of thunks so check former_clone_of or if we |
| might have skipped thunks because this adjustments are no longer |
| necessary. */ |
| while (node->thunk.thunk_p) |
| { |
| if (node2->former_clone_of == node->decl) |
| return true; |
| if (!node->thunk.this_adjusting) |
| return false; |
| node = cgraph_function_or_thunk_node (node->callees->callee, NULL); |
| skipped_thunk = true; |
| } |
| |
| if (skipped_thunk) |
| { |
| if (!node2->clone.args_to_skip |
| || !bitmap_bit_p (node2->clone.args_to_skip, 0)) |
| return false; |
| if (node2->former_clone_of == node->decl) |
| return true; |
| else if (!node2->clone_of) |
| return false; |
| } |
| |
| while (node != node2 && node2) |
| node2 = node2->clone_of; |
| return node2 != NULL; |
| } |
| |
| /* Verify edge E count and frequency. */ |
| |
| static bool |
| verify_edge_count_and_frequency (struct cgraph_edge *e) |
| { |
| bool error_found = false; |
| if (e->count < 0) |
| { |
| error ("caller edge count is negative"); |
| error_found = true; |
| } |
| if (e->frequency < 0) |
| { |
| error ("caller edge frequency is negative"); |
| error_found = true; |
| } |
| if (e->frequency > CGRAPH_FREQ_MAX) |
| { |
| error ("caller edge frequency is too large"); |
| error_found = true; |
| } |
| if (gimple_has_body_p (e->caller->decl) |
| && e->call_stmt |
| && !e->caller->global.inlined_to |
| && !e->speculative |
| /* FIXME: Inline-analysis sets frequency to 0 when edge is optimized out. |
| Remove this once edges are actually removed from the function at that time. */ |
| && (e->frequency |
| || (inline_edge_summary_vec.exists () |
| && ((inline_edge_summary_vec.length () <= (unsigned) e->uid) |
| || !inline_edge_summary (e)->predicate))) |
| && (e->frequency |
| != compute_call_stmt_bb_frequency (e->caller->decl, |
| gimple_bb (e->call_stmt)))) |
| { |
| error ("caller edge frequency %i does not match BB frequency %i", |
| e->frequency, |
| compute_call_stmt_bb_frequency (e->caller->decl, |
| gimple_bb (e->call_stmt))); |
| error_found = true; |
| } |
| return error_found; |
| } |
| |
| /* Switch to THIS_CFUN if needed and print STMT to stderr. */ |
| static void |
| cgraph_debug_gimple_stmt (struct function *this_cfun, gimple stmt) |
| { |
| bool fndecl_was_null = false; |
| /* debug_gimple_stmt needs correct cfun */ |
| if (cfun != this_cfun) |
| set_cfun (this_cfun); |
| /* ...and an actual current_function_decl */ |
| if (!current_function_decl) |
| { |
| current_function_decl = this_cfun->decl; |
| fndecl_was_null = true; |
| } |
| debug_gimple_stmt (stmt); |
| if (fndecl_was_null) |
| current_function_decl = NULL; |
| } |
| |
| /* Verify that call graph edge E corresponds to DECL from the associated |
| statement. Return true if the verification should fail. */ |
| |
| static bool |
| verify_edge_corresponds_to_fndecl (struct cgraph_edge *e, tree decl) |
| { |
| struct cgraph_node *node; |
| |
| if (!decl || e->callee->global.inlined_to) |
| return false; |
| if (cgraph_state == CGRAPH_LTO_STREAMING) |
| return false; |
| node = cgraph_get_node (decl); |
| |
| /* We do not know if a node from a different partition is an alias or what it |
| aliases and therefore cannot do the former_clone_of check reliably. When |
| body_removed is set, we have lost all information about what was alias or |
| thunk of and also cannot proceed. */ |
| if (!node |
| || node->body_removed |
| || node->in_other_partition |
| || e->callee->in_other_partition) |
| return false; |
| |
| /* Optimizers can redirect unreachable calls or calls triggering undefined |
| behaviour to builtin_unreachable. */ |
| if (DECL_BUILT_IN_CLASS (e->callee->decl) == BUILT_IN_NORMAL |
| && DECL_FUNCTION_CODE (e->callee->decl) == BUILT_IN_UNREACHABLE) |
| return false; |
| node = cgraph_function_or_thunk_node (node, NULL); |
| |
| if (e->callee->former_clone_of != node->decl |
| && (node != cgraph_function_or_thunk_node (e->callee, NULL)) |
| && !clone_of_p (node, e->callee)) |
| return true; |
| else |
| return false; |
| } |
| |
| /* Verify cgraph nodes of given cgraph node. */ |
| DEBUG_FUNCTION void |
| verify_cgraph_node (struct cgraph_node *node) |
| { |
| struct cgraph_edge *e; |
| struct function *this_cfun = DECL_STRUCT_FUNCTION (node->decl); |
| basic_block this_block; |
| gimple_stmt_iterator gsi; |
| bool error_found = false; |
| |
| if (seen_error ()) |
| return; |
| |
| timevar_push (TV_CGRAPH_VERIFY); |
| error_found |= verify_symtab_base (node); |
| for (e = node->callees; e; e = e->next_callee) |
| if (e->aux) |
| { |
| error ("aux field set for edge %s->%s", |
| identifier_to_locale (e->caller->name ()), |
| identifier_to_locale (e->callee->name ())); |
| error_found = true; |
| } |
| if (node->count < 0) |
| { |
| error ("execution count is negative"); |
| error_found = true; |
| } |
| if (node->global.inlined_to && node->same_comdat_group) |
| { |
| error ("inline clone in same comdat group list"); |
| error_found = true; |
| } |
| if (!node->definition && !node->in_other_partition && node->local.local) |
| { |
| error ("local symbols must be defined"); |
| error_found = true; |
| } |
| if (node->global.inlined_to && node->externally_visible) |
| { |
| error ("externally visible inline clone"); |
| error_found = true; |
| } |
| if (node->global.inlined_to && node->address_taken) |
| { |
| error ("inline clone with address taken"); |
| error_found = true; |
| } |
| if (node->global.inlined_to && node->force_output) |
| { |
| error ("inline clone is forced to output"); |
| error_found = true; |
| } |
| for (e = node->indirect_calls; e; e = e->next_callee) |
| { |
| if (e->aux) |
| { |
| error ("aux field set for indirect edge from %s", |
| identifier_to_locale (e->caller->name ())); |
| error_found = true; |
| } |
| if (!e->indirect_unknown_callee |
| || !e->indirect_info) |
| { |
| error ("An indirect edge from %s is not marked as indirect or has " |
| "associated indirect_info, the corresponding statement is: ", |
| identifier_to_locale (e->caller->name ())); |
| cgraph_debug_gimple_stmt (this_cfun, e->call_stmt); |
| error_found = true; |
| } |
| } |
| bool check_comdat = symtab_comdat_local_p (node); |
| for (e = node->callers; e; e = e->next_caller) |
| { |
| if (verify_edge_count_and_frequency (e)) |
| error_found = true; |
| if (check_comdat |
| && !symtab_in_same_comdat_p (e->caller, node)) |
| { |
| error ("comdat-local function called by %s outside its comdat", |
| identifier_to_locale (e->caller->name ())); |
| error_found = true; |
| } |
| if (!e->inline_failed) |
| { |
| if (node->global.inlined_to |
| != (e->caller->global.inlined_to |
| ? e->caller->global.inlined_to : e->caller)) |
| { |
| error ("inlined_to pointer is wrong"); |
| error_found = true; |
| } |
| if (node->callers->next_caller) |
| { |
| error ("multiple inline callers"); |
| error_found = true; |
| } |
| } |
| else |
| if (node->global.inlined_to) |
| { |
| error ("inlined_to pointer set for noninline callers"); |
| error_found = true; |
| } |
| } |
| for (e = node->indirect_calls; e; e = e->next_callee) |
| if (verify_edge_count_and_frequency (e)) |
| error_found = true; |
| if (!node->callers && node->global.inlined_to) |
| { |
| error ("inlined_to pointer is set but no predecessors found"); |
| error_found = true; |
| } |
| if (node->global.inlined_to == node) |
| { |
| error ("inlined_to pointer refers to itself"); |
| error_found = true; |
| } |
| |
| if (node->clone_of) |
| { |
| struct cgraph_node *n; |
| for (n = node->clone_of->clones; n; n = n->next_sibling_clone) |
| if (n == node) |
| break; |
| if (!n) |
| { |
| error ("node has wrong clone_of"); |
| error_found = true; |
| } |
| } |
| if (node->clones) |
| { |
| struct cgraph_node *n; |
| for (n = node->clones; n; n = n->next_sibling_clone) |
| if (n->clone_of != node) |
| break; |
| if (n) |
| { |
| error ("node has wrong clone list"); |
| error_found = true; |
| } |
| } |
| if ((node->prev_sibling_clone || node->next_sibling_clone) && !node->clone_of) |
| { |
| error ("node is in clone list but it is not clone"); |
| error_found = true; |
| } |
| if (!node->prev_sibling_clone && node->clone_of && node->clone_of->clones != node) |
| { |
| error ("node has wrong prev_clone pointer"); |
| error_found = true; |
| } |
| if (node->prev_sibling_clone && node->prev_sibling_clone->next_sibling_clone != node) |
| { |
| error ("double linked list of clones corrupted"); |
| error_found = true; |
| } |
| |
| if (node->analyzed && node->alias) |
| { |
| bool ref_found = false; |
| int i; |
| struct ipa_ref *ref; |
| |
| if (node->callees) |
| { |
| error ("Alias has call edges"); |
| error_found = true; |
| } |
| for (i = 0; ipa_ref_list_reference_iterate (&node->ref_list, |
| i, ref); i++) |
| if (ref->use != IPA_REF_ALIAS) |
| { |
| error ("Alias has non-alias reference"); |
| error_found = true; |
| } |
| else if (ref_found |
| /* in LIPO mode, the alias can refer to the real target also */ |
| && !L_IPO_COMP_MODE) |
| { |
| error ("Alias has more than one alias reference"); |
| error_found = true; |
| } |
| else |
| ref_found = true; |
| if (!ref_found) |
| { |
| error ("Analyzed alias has no reference"); |
| error_found = true; |
| } |
| } |
| if (node->analyzed && node->thunk.thunk_p) |
| { |
| if (!node->callees) |
| { |
| error ("No edge out of thunk node"); |
| error_found = true; |
| } |
| else if (node->callees->next_callee) |
| { |
| error ("More than one edge out of thunk node"); |
| error_found = true; |
| } |
| if (gimple_has_body_p (node->decl)) |
| { |
| error ("Thunk is not supposed to have body"); |
| error_found = true; |
| } |
| } |
| else if (node->analyzed && gimple_has_body_p (node->decl) |
| && !TREE_ASM_WRITTEN (node->decl) |
| && (!DECL_EXTERNAL (node->decl) || node->global.inlined_to) |
| && !flag_wpa) |
| { |
| if (this_cfun->cfg) |
| { |
| pointer_set_t *stmts = pointer_set_create (); |
| int i; |
| struct ipa_ref *ref; |
| |
| /* Reach the trees by walking over the CFG, and note the |
| enclosing basic-blocks in the call edges. */ |
| FOR_EACH_BB_FN (this_block, this_cfun) |
| { |
| for (gsi = gsi_start_phis (this_block); |
| !gsi_end_p (gsi); gsi_next (&gsi)) |
| pointer_set_insert (stmts, gsi_stmt (gsi)); |
| for (gsi = gsi_start_bb (this_block); |
| !gsi_end_p (gsi); |
| gsi_next (&gsi)) |
| { |
| gimple stmt = gsi_stmt (gsi); |
| pointer_set_insert (stmts, stmt); |
| if (is_gimple_call (stmt)) |
| { |
| struct cgraph_edge *e = cgraph_edge (node, stmt); |
| tree decl = gimple_call_fndecl (stmt); |
| if (e) |
| { |
| if (e->aux) |
| { |
| error ("shared call_stmt:"); |
| cgraph_debug_gimple_stmt (this_cfun, stmt); |
| error_found = true; |
| } |
| if (!e->indirect_unknown_callee) |
| { |
| if (verify_edge_corresponds_to_fndecl (e, decl)) |
| { |
| error ("edge points to wrong declaration:"); |
| debug_tree (e->callee->decl); |
| fprintf (stderr," Instead of:"); |
| debug_tree (decl); |
| error_found = true; |
| } |
| } |
| else if (decl) |
| { |
| error ("an indirect edge with unknown callee " |
| "corresponding to a call_stmt with " |
| "a known declaration:"); |
| error_found = true; |
| cgraph_debug_gimple_stmt (this_cfun, e->call_stmt); |
| } |
| e->aux = (void *)1; |
| } |
| else if (decl) |
| { |
| error ("missing callgraph edge for call stmt:"); |
| cgraph_debug_gimple_stmt (this_cfun, stmt); |
| error_found = true; |
| } |
| } |
| } |
| } |
| for (i = 0; |
| ipa_ref_list_reference_iterate (&node->ref_list, i, ref); |
| i++) |
| if (ref->stmt && !pointer_set_contains (stmts, ref->stmt)) |
| { |
| error ("reference to dead statement"); |
| cgraph_debug_gimple_stmt (this_cfun, ref->stmt); |
| error_found = true; |
| } |
| pointer_set_destroy (stmts); |
| } |
| else |
| /* No CFG available?! */ |
| gcc_unreachable (); |
| |
| for (e = node->callees; e; e = e->next_callee) |
| { |
| if (!e->aux && e->call_stmt) |
| { |
| error ("edge %s->%s has no corresponding call_stmt", |
| identifier_to_locale (e->caller->name ()), |
| identifier_to_locale (e->callee->name ())); |
| cgraph_debug_gimple_stmt (this_cfun, e->call_stmt); |
| error_found = true; |
| } |
| e->aux = 0; |
| } |
| for (e = node->indirect_calls; e; e = e->next_callee) |
| { |
| if (!e->aux && !e->speculative) |
| { |
| error ("an indirect edge from %s has no corresponding call_stmt", |
| identifier_to_locale (e->caller->name ())); |
| cgraph_debug_gimple_stmt (this_cfun, e->call_stmt); |
| error_found = true; |
| } |
| e->aux = 0; |
| } |
| } |
| if (error_found) |
| { |
| dump_cgraph_node (stderr, node); |
| internal_error ("verify_cgraph_node failed"); |
| } |
| timevar_pop (TV_CGRAPH_VERIFY); |
| } |
| |
| /* Verify whole cgraph structure. */ |
| DEBUG_FUNCTION void |
| verify_cgraph (void) |
| { |
| struct cgraph_node *node; |
| |
| if (seen_error ()) |
| return; |
| |
| FOR_EACH_FUNCTION (node) |
| verify_cgraph_node (node); |
| } |
| |
| /* Given NODE, walk the alias chain to return the function NODE is alias of. |
| Walk through thunk, too. |
| When AVAILABILITY is non-NULL, get minimal availability in the chain. */ |
| |
| struct cgraph_node * |
| cgraph_function_node (struct cgraph_node *node, enum availability *availability) |
| { |
| do |
| { |
| node = cgraph_function_or_thunk_node (node, availability); |
| if (node->thunk.thunk_p) |
| { |
| node = node->callees->callee; |
| if (availability) |
| { |
| enum availability a; |
| a = cgraph_function_body_availability (node); |
| if (a < *availability) |
| *availability = a; |
| } |
| node = cgraph_function_or_thunk_node (node, availability); |
| } |
| } while (node && node->thunk.thunk_p); |
| return node; |
| } |
| |
| /* When doing LTO, read NODE's body from disk if it is not already present. */ |
| |
| bool |
| cgraph_get_body (struct cgraph_node *node) |
| { |
| struct lto_file_decl_data *file_data; |
| const char *data, *name; |
| size_t len; |
| tree decl = node->decl; |
| |
| if (DECL_RESULT (decl)) |
| return false; |
| |
| gcc_assert (in_lto_p); |
| |
| file_data = node->lto_file_data; |
| name = IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl)); |
| |
| /* We may have renamed the declaration, e.g., a static function. */ |
| name = lto_get_decl_name_mapping (file_data, name); |
| |
| data = lto_get_section_data (file_data, LTO_section_function_body, |
| name, &len); |
| if (!data) |
| { |
| dump_cgraph_node (stderr, node); |
| fatal_error ("%s: section %s is missing", |
| file_data->file_name, |
| name); |
| } |
| |
| gcc_assert (DECL_STRUCT_FUNCTION (decl) == NULL); |
| |
| lto_input_function_body (file_data, node, data); |
| lto_stats.num_function_bodies++; |
| lto_free_section_data (file_data, LTO_section_function_body, name, |
| data, len); |
| lto_free_function_in_decl_state_for_node (node); |
| return true; |
| } |
| |
| /* Verify if the type of the argument matches that of the function |
| declaration. If we cannot verify this or there is a mismatch, |
| return false. */ |
| |
| static bool |
| gimple_check_call_args (gimple stmt, tree fndecl, bool args_count_match) |
| { |
| tree parms, p; |
| unsigned int i, nargs; |
| |
| /* Calls to internal functions always match their signature. */ |
| if (gimple_call_internal_p (stmt)) |
| return true; |
| |
| nargs = gimple_call_num_args (stmt); |
| |
| /* Get argument types for verification. */ |
| if (fndecl) |
| parms = TYPE_ARG_TYPES (TREE_TYPE (fndecl)); |
| else |
| parms = TYPE_ARG_TYPES (gimple_call_fntype (stmt)); |
| |
| /* Verify if the type of the argument matches that of the function |
| declaration. If we cannot verify this or there is a mismatch, |
| return false. */ |
| if (fndecl && DECL_ARGUMENTS (fndecl)) |
| { |
| for (i = 0, p = DECL_ARGUMENTS (fndecl); |
| i < nargs; |
| i++, p = DECL_CHAIN (p)) |
| { |
| tree arg; |
| /* We cannot distinguish a varargs function from the case |
| of excess parameters, still deferring the inlining decision |
| to the callee is possible. */ |
| if (!p) |
| break; |
| arg = gimple_call_arg (stmt, i); |
| if (p == error_mark_node |
| || DECL_ARG_TYPE (p) == error_mark_node |
| || arg == error_mark_node |
| || (!types_compatible_p (DECL_ARG_TYPE (p), TREE_TYPE (arg)) |
| && !fold_convertible_p (DECL_ARG_TYPE (p), arg))) |
| return false; |
| } |
| if (args_count_match && p) |
| return false; |
| } |
| else if (parms) |
| { |
| for (i = 0, p = parms; i < nargs; i++, p = TREE_CHAIN (p)) |
| { |
| tree arg; |
| /* If this is a varargs function defer inlining decision |
| to callee. */ |
| if (!p) |
| break; |
| arg = gimple_call_arg (stmt, i); |
| if (TREE_VALUE (p) == error_mark_node |
| || arg == error_mark_node |
| || TREE_CODE (TREE_VALUE (p)) == VOID_TYPE |
| || (!types_compatible_p (TREE_VALUE (p), TREE_TYPE (arg)) |
| && !fold_convertible_p (TREE_VALUE (p), arg))) |
| return false; |
| } |
| } |
| else |
| { |
| if (nargs != 0) |
| return false; |
| } |
| return true; |
| } |
| |
| /* Verify if the type of the argument and lhs of CALL_STMT matches |
| that of the function declaration CALLEE. If ARGS_COUNT_MATCH is |
| true, the arg count needs to be the same. |
| If we cannot verify this or there is a mismatch, return false. */ |
| |
| bool |
| gimple_check_call_matching_types (gimple call_stmt, tree callee, |
| bool args_count_match) |
| { |
| tree lhs; |
| |
| if ((DECL_RESULT (callee) |
| && !DECL_BY_REFERENCE (DECL_RESULT (callee)) |
| && (lhs = gimple_call_lhs (call_stmt)) != NULL_TREE |
| && !useless_type_conversion_p (TREE_TYPE (DECL_RESULT (callee)), |
| TREE_TYPE (lhs)) |
| && !fold_convertible_p (TREE_TYPE (DECL_RESULT (callee)), lhs)) |
| || !gimple_check_call_args (call_stmt, callee, args_count_match)) |
| return false; |
| return true; |
| } |
| |
| #include "gt-cgraph.h" |