1 /* Callgraph based interprocedural optimizations. 2 Copyright (C) 2003, 2004, 2005 Free Software Foundation, Inc. 3 Contributed by Jan Hubicka 4 5 This file is part of GCC. 6 7 GCC is free software; you can redistribute it and/or modify it under 8 the terms of the GNU General Public License as published by the Free 9 Software Foundation; either version 2, or (at your option) any later 10 version. 11 12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY 13 WARRANTY; without even the implied warranty of MERCHANTABILITY or 14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 15 for more details. 16 17 You should have received a copy of the GNU General Public License 18 along with GCC; see the file COPYING. If not, write to the Free 19 Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 20 02110-1301, USA. */ 21 22 /* This module implements main driver of compilation process as well as 23 few basic interprocedural optimizers. 24 25 The main scope of this file is to act as an interface in between 26 tree based frontends and the backend (and middle end) 27 28 The front-end is supposed to use following functionality: 29 30 - cgraph_finalize_function 31 32 This function is called once front-end has parsed whole body of function 33 and it is certain that the function body nor the declaration will change. 34 35 (There is one exception needed for implementing GCC extern inline function.) 36 37 - cgraph_varpool_finalize_variable 38 39 This function has same behavior as the above but is used for static 40 variables. 41 42 - cgraph_finalize_compilation_unit 43 44 This function is called once compilation unit is finalized and it will 45 no longer change. 46 47 In the unit-at-a-time the call-graph construction and local function 48 analysis takes place here. Bodies of unreachable functions are released 49 to conserve memory usage. 50 51 ??? The compilation unit in this point of view should be compilation 52 unit as defined by the language - for instance C frontend allows multiple 53 compilation units to be parsed at once and it should call function each 54 time parsing is done so we save memory. 55 56 - cgraph_optimize 57 58 In this unit-at-a-time compilation the intra procedural analysis takes 59 place here. In particular the static functions whose address is never 60 taken are marked as local. Backend can then use this information to 61 modify calling conventions, do better inlining or similar optimizations. 62 63 - cgraph_assemble_pending_functions 64 - cgraph_varpool_assemble_pending_variables 65 66 In non-unit-at-a-time mode these functions can be used to force compilation 67 of functions or variables that are known to be needed at given stage 68 of compilation 69 70 - cgraph_mark_needed_node 71 - cgraph_varpool_mark_needed_node 72 73 When function or variable is referenced by some hidden way (for instance 74 via assembly code and marked by attribute "used"), the call-graph data structure 75 must be updated accordingly by this function. 76 77 - analyze_expr callback 78 79 This function is responsible for lowering tree nodes not understood by 80 generic code into understandable ones or alternatively marking 81 callgraph and varpool nodes referenced by the as needed. 82 83 ??? On the tree-ssa genericizing should take place here and we will avoid 84 need for these hooks (replacing them by genericizing hook) 85 86 - expand_function callback 87 88 This function is used to expand function and pass it into RTL back-end. 89 Front-end should not make any assumptions about when this function can be 90 called. In particular cgraph_assemble_pending_functions, 91 cgraph_varpool_assemble_pending_variables, cgraph_finalize_function, 92 cgraph_varpool_finalize_function, cgraph_optimize can cause arbitrarily 93 previously finalized functions to be expanded. 94 95 We implement two compilation modes. 96 97 - unit-at-a-time: In this mode analyzing of all functions is deferred 98 to cgraph_finalize_compilation_unit and expansion into cgraph_optimize. 99 100 In cgraph_finalize_compilation_unit the reachable functions are 101 analyzed. During analysis the call-graph edges from reachable 102 functions are constructed and their destinations are marked as 103 reachable. References to functions and variables are discovered too 104 and variables found to be needed output to the assembly file. Via 105 mark_referenced call in assemble_variable functions referenced by 106 static variables are noticed too. 107 108 The intra-procedural information is produced and its existence 109 indicated by global_info_ready. Once this flag is set it is impossible 110 to change function from !reachable to reachable and thus 111 assemble_variable no longer call mark_referenced. 112 113 Finally the call-graph is topologically sorted and all reachable functions 114 that has not been completely inlined or are not external are output. 115 116 ??? It is possible that reference to function or variable is optimized 117 out. We can not deal with this nicely because topological order is not 118 suitable for it. For tree-ssa we may consider another pass doing 119 optimization and re-discovering reachable functions. 120 121 ??? Reorganize code so variables are output very last and only if they 122 really has been referenced by produced code, so we catch more cases 123 where reference has been optimized out. 124 125 - non-unit-at-a-time 126 127 All functions are variables are output as early as possible to conserve 128 memory consumption. This may or may not result in less memory used but 129 it is still needed for some legacy code that rely on particular ordering 130 of things output from the compiler. 131 132 Varpool data structures are not used and variables are output directly. 133 134 Functions are output early using call of 135 cgraph_assemble_pending_function from cgraph_finalize_function. The 136 decision on whether function is needed is made more conservative so 137 uninlininable static functions are needed too. During the call-graph 138 construction the edge destinations are not marked as reachable and it 139 is completely relied upn assemble_variable to mark them. */ 140 141 142 #include "config.h" 143 #include "system.h" 144 #include "coretypes.h" 145 #include "tm.h" 146 #include "tree.h" 147 #include "rtl.h" 148 #include "tree-flow.h" 149 #include "tree-inline.h" 150 #include "langhooks.h" 151 #include "pointer-set.h" 152 #include "toplev.h" 153 #include "flags.h" 154 #include "ggc.h" 155 #include "debug.h" 156 #include "target.h" 157 #include "cgraph.h" 158 #include "diagnostic.h" 159 #include "timevar.h" 160 #include "params.h" 161 #include "fibheap.h" 162 #include "c-common.h" 163 #include "intl.h" 164 #include "function.h" 165 #include "ipa-prop.h" 166 #include "tree-gimple.h" 167 #include "tree-pass.h" 168 #include "output.h" 169 170 static void cgraph_expand_all_functions (void); 171 static void cgraph_mark_functions_to_output (void); 172 static void cgraph_expand_function (struct cgraph_node *); 173 static tree record_reference (tree *, int *, void *); 174 static void cgraph_output_pending_asms (void); 175 static void cgraph_increase_alignment (void); 176 177 /* Lists all assembled variables to be sent to debugger output later on. */ 178 static GTY(()) struct cgraph_varpool_node *cgraph_varpool_assembled_nodes_queue; 179 180 /* Records tree nodes seen in record_reference. Simply using 181 walk_tree_without_duplicates doesn't guarantee each node is visited 182 once because it gets a new htab upon each recursive call from 183 record_reference itself. */ 184 static struct pointer_set_t *visited_nodes; 185 186 static FILE *cgraph_dump_file; 187 188 /* Determine if function DECL is needed. That is, visible to something 189 either outside this translation unit, something magic in the system 190 configury, or (if not doing unit-at-a-time) to something we havn't 191 seen yet. */ 192 193 static bool 194 decide_is_function_needed (struct cgraph_node *node, tree decl) 195 { 196 tree origin; 197 if (MAIN_NAME_P (DECL_NAME (decl)) 198 && TREE_PUBLIC (decl)) 199 { 200 node->local.externally_visible = true; 201 return true; 202 } 203 204 /* If the user told us it is used, then it must be so. */ 205 if (node->local.externally_visible) 206 return true; 207 208 if (!flag_unit_at_a_time && lookup_attribute ("used", DECL_ATTRIBUTES (decl))) 209 return true; 210 211 /* ??? If the assembler name is set by hand, it is possible to assemble 212 the name later after finalizing the function and the fact is noticed 213 in assemble_name then. This is arguably a bug. */ 214 if (DECL_ASSEMBLER_NAME_SET_P (decl) 215 && TREE_SYMBOL_REFERENCED (DECL_ASSEMBLER_NAME (decl))) 216 return true; 217 218 /* If we decided it was needed before, but at the time we didn't have 219 the body of the function available, then it's still needed. We have 220 to go back and re-check its dependencies now. */ 221 if (node->needed) 222 return true; 223 224 /* Externally visible functions must be output. The exception is 225 COMDAT functions that must be output only when they are needed. 226 227 When not optimizing, also output the static functions. (see 228 PR24561), but don't do so for always_inline functions, functions 229 declared inline and nested functions. These was optimized out 230 in the original implementation and it is unclear whether we want 231 to change the behavior here. */ 232 if (((TREE_PUBLIC (decl) 233 || (!optimize && !node->local.disregard_inline_limits 234 && !DECL_DECLARED_INLINE_P (decl) 235 && !node->origin)) 236 && !flag_whole_program) 237 && !DECL_COMDAT (decl) && !DECL_EXTERNAL (decl)) 238 return true; 239 240 /* Constructors and destructors are reachable from the runtime by 241 some mechanism. */ 242 if (DECL_STATIC_CONSTRUCTOR (decl) || DECL_STATIC_DESTRUCTOR (decl)) 243 return true; 244 245 if (flag_unit_at_a_time) 246 return false; 247 248 /* If not doing unit at a time, then we'll only defer this function 249 if its marked for inlining. Otherwise we want to emit it now. */ 250 251 /* "extern inline" functions are never output locally. */ 252 if (DECL_EXTERNAL (decl)) 253 return false; 254 /* Nested functions of extern inline function shall not be emit unless 255 we inlined the origin. */ 256 for (origin = decl_function_context (decl); origin; 257 origin = decl_function_context (origin)) 258 if (DECL_EXTERNAL (origin)) 259 return false; 260 /* We want to emit COMDAT functions only when absolutely necessary. */ 261 if (DECL_COMDAT (decl)) 262 return false; 263 if (!DECL_INLINE (decl) 264 || (!node->local.disregard_inline_limits 265 /* When declared inline, defer even the uninlinable functions. 266 This allows them to be eliminated when unused. */ 267 && !DECL_DECLARED_INLINE_P (decl) 268 && (!node->local.inlinable || !cgraph_default_inline_p (node, NULL)))) 269 return true; 270 271 return false; 272 } 273 274 /* Walk the decls we marked as necessary and see if they reference new 275 variables or functions and add them into the worklists. */ 276 static bool 277 cgraph_varpool_analyze_pending_decls (void) 278 { 279 bool changed = false; 280 timevar_push (TV_CGRAPH); 281 282 while (cgraph_varpool_first_unanalyzed_node) 283 { 284 tree decl = cgraph_varpool_first_unanalyzed_node->decl; 285 286 cgraph_varpool_first_unanalyzed_node->analyzed = true; 287 288 cgraph_varpool_first_unanalyzed_node = cgraph_varpool_first_unanalyzed_node->next_needed; 289 290 /* Compute the alignment early so function body expanders are 291 already informed about increased alignment. */ 292 align_variable (decl, 0); 293 294 if (DECL_INITIAL (decl)) 295 { 296 visited_nodes = pointer_set_create (); 297 walk_tree (&DECL_INITIAL (decl), record_reference, NULL, visited_nodes); 298 pointer_set_destroy (visited_nodes); 299 visited_nodes = NULL; 300 } 301 changed = true; 302 } 303 timevar_pop (TV_CGRAPH); 304 return changed; 305 } 306 307 /* Optimization of function bodies might've rendered some variables as 308 unnecessary so we want to avoid these from being compiled. 309 310 This is done by pruning the queue and keeping only the variables that 311 really appear needed (ie they are either externally visible or referenced 312 by compiled function). Re-doing the reachability analysis on variables 313 brings back the remaining variables referenced by these. */ 314 static void 315 cgraph_varpool_remove_unreferenced_decls (void) 316 { 317 struct cgraph_varpool_node *next, *node = cgraph_varpool_nodes_queue; 318 319 cgraph_varpool_reset_queue (); 320 321 if (errorcount || sorrycount) 322 return; 323 324 while (node) 325 { 326 tree decl = node->decl; 327 next = node->next_needed; 328 node->needed = 0; 329 330 if (node->finalized 331 && ((DECL_ASSEMBLER_NAME_SET_P (decl) 332 && TREE_SYMBOL_REFERENCED (DECL_ASSEMBLER_NAME (decl))) 333 || node->force_output 334 || decide_is_variable_needed (node, decl) 335 /* ??? Cgraph does not yet rule the world with an iron hand, 336 and does not control the emission of debug information. 337 After a variable has its DECL_RTL set, we must assume that 338 it may be referenced by the debug information, and we can 339 no longer elide it. */ 340 || DECL_RTL_SET_P (decl))) 341 cgraph_varpool_mark_needed_node (node); 342 343 node = next; 344 } 345 /* Make sure we mark alias targets as used targets. */ 346 finish_aliases_1 (); 347 cgraph_varpool_analyze_pending_decls (); 348 } 349 350 351 /* When not doing unit-at-a-time, output all functions enqueued. 352 Return true when such a functions were found. */ 353 354 bool 355 cgraph_assemble_pending_functions (void) 356 { 357 bool output = false; 358 359 if (flag_unit_at_a_time) 360 return false; 361 362 cgraph_output_pending_asms (); 363 364 while (cgraph_nodes_queue) 365 { 366 struct cgraph_node *n = cgraph_nodes_queue; 367 368 cgraph_nodes_queue = cgraph_nodes_queue->next_needed; 369 n->next_needed = NULL; 370 if (!n->global.inlined_to 371 && !n->alias 372 && !DECL_EXTERNAL (n->decl)) 373 { 374 cgraph_expand_function (n); 375 output = true; 376 } 377 } 378 379 /* Process CGRAPH_EXPAND_QUEUE, these are functions created during 380 the expansion process. Note that this queue may grow as its 381 being processed, as the new functions may generate new ones. */ 382 while (cgraph_expand_queue) 383 { 384 struct cgraph_node *n = cgraph_expand_queue; 385 cgraph_expand_queue = cgraph_expand_queue->next_needed; 386 n->next_needed = NULL; 387 cgraph_finalize_function (n->decl, false); 388 output = true; 389 } 390 391 return output; 392 } 393 394 395 /* As an GCC extension we allow redefinition of the function. The 396 semantics when both copies of bodies differ is not well defined. 397 We replace the old body with new body so in unit at a time mode 398 we always use new body, while in normal mode we may end up with 399 old body inlined into some functions and new body expanded and 400 inlined in others. 401 402 ??? It may make more sense to use one body for inlining and other 403 body for expanding the function but this is difficult to do. */ 404 405 static void 406 cgraph_reset_node (struct cgraph_node *node) 407 { 408 /* If node->output is set, then this is a unit-at-a-time compilation 409 and we have already begun whole-unit analysis. This is *not* 410 testing for whether we've already emitted the function. That 411 case can be sort-of legitimately seen with real function 412 redefinition errors. I would argue that the front end should 413 never present us with such a case, but don't enforce that for now. */ 414 gcc_assert (!node->output); 415 416 /* Reset our data structures so we can analyze the function again. */ 417 memset (&node->local, 0, sizeof (node->local)); 418 memset (&node->global, 0, sizeof (node->global)); 419 memset (&node->rtl, 0, sizeof (node->rtl)); 420 node->analyzed = false; 421 node->local.redefined_extern_inline = true; 422 node->local.finalized = false; 423 424 if (!flag_unit_at_a_time) 425 { 426 struct cgraph_node *n, *next; 427 428 for (n = cgraph_nodes; n; n = next) 429 { 430 next = n->next; 431 if (n->global.inlined_to == node) 432 cgraph_remove_node (n); 433 } 434 } 435 436 cgraph_node_remove_callees (node); 437 438 /* We may need to re-queue the node for assembling in case 439 we already proceeded it and ignored as not needed. */ 440 if (node->reachable && !flag_unit_at_a_time) 441 { 442 struct cgraph_node *n; 443 444 for (n = cgraph_nodes_queue; n; n = n->next_needed) 445 if (n == node) 446 break; 447 if (!n) 448 node->reachable = 0; 449 } 450 } 451 452 static void 453 cgraph_lower_function (struct cgraph_node *node) 454 { 455 if (node->lowered) 456 return; 457 tree_lowering_passes (node->decl); 458 node->lowered = true; 459 } 460 461 /* DECL has been parsed. Take it, queue it, compile it at the whim of the 462 logic in effect. If NESTED is true, then our caller cannot stand to have 463 the garbage collector run at the moment. We would need to either create 464 a new GC context, or just not compile right now. */ 465 466 void 467 cgraph_finalize_function (tree decl, bool nested) 468 { 469 struct cgraph_node *node = cgraph_node (decl); 470 471 if (node->local.finalized) 472 cgraph_reset_node (node); 473 474 notice_global_symbol (decl); 475 node->decl = decl; 476 node->local.finalized = true; 477 node->lowered = DECL_STRUCT_FUNCTION (decl)->cfg != NULL; 478 if (node->nested) 479 lower_nested_functions (decl); 480 gcc_assert (!node->nested); 481 482 /* If not unit at a time, then we need to create the call graph 483 now, so that called functions can be queued and emitted now. */ 484 if (!flag_unit_at_a_time) 485 { 486 cgraph_analyze_function (node); 487 cgraph_decide_inlining_incrementally (node, false); 488 } 489 490 if (decide_is_function_needed (node, decl)) 491 cgraph_mark_needed_node (node); 492 493 /* Since we reclaim unreachable nodes at the end of every language 494 level unit, we need to be conservative about possible entry points 495 there. */ 496 if ((TREE_PUBLIC (decl) && !DECL_COMDAT (decl) && !DECL_EXTERNAL (decl))) 497 cgraph_mark_reachable_node (node); 498 499 /* If not unit at a time, go ahead and emit everything we've found 500 to be reachable at this time. */ 501 if (!nested) 502 { 503 if (!cgraph_assemble_pending_functions ()) 504 ggc_collect (); 505 } 506 507 /* If we've not yet emitted decl, tell the debug info about it. */ 508 if (!TREE_ASM_WRITTEN (decl)) 509 (*debug_hooks->deferred_inline_function) (decl); 510 511 /* Possibly warn about unused parameters. */ 512 if (warn_unused_parameter) 513 do_warn_unused_parameter (decl); 514 } 515 516 /* Walk tree and record all calls. Called via walk_tree. */ 517 static tree 518 record_reference (tree *tp, int *walk_subtrees, void *data) 519 { 520 tree t = *tp; 521 522 switch (TREE_CODE (t)) 523 { 524 case VAR_DECL: 525 /* ??? Really, we should mark this decl as *potentially* referenced 526 by this function and re-examine whether the decl is actually used 527 after rtl has been generated. */ 528 if (TREE_STATIC (t) || DECL_EXTERNAL (t)) 529 { 530 cgraph_varpool_mark_needed_node (cgraph_varpool_node (t)); 531 if (lang_hooks.callgraph.analyze_expr) 532 return lang_hooks.callgraph.analyze_expr (tp, walk_subtrees, 533 data); 534 } 535 break; 536 537 case FDESC_EXPR: 538 case ADDR_EXPR: 539 if (flag_unit_at_a_time) 540 { 541 /* Record dereferences to the functions. This makes the 542 functions reachable unconditionally. */ 543 tree decl = TREE_OPERAND (*tp, 0); 544 if (TREE_CODE (decl) == FUNCTION_DECL) 545 cgraph_mark_needed_node (cgraph_node (decl)); 546 } 547 break; 548 549 default: 550 /* Save some cycles by not walking types and declaration as we 551 won't find anything useful there anyway. */ 552 if (IS_TYPE_OR_DECL_P (*tp)) 553 { 554 *walk_subtrees = 0; 555 break; 556 } 557 558 if ((unsigned int) TREE_CODE (t) >= LAST_AND_UNUSED_TREE_CODE) 559 return lang_hooks.callgraph.analyze_expr (tp, walk_subtrees, data); 560 break; 561 } 562 563 return NULL; 564 } 565 566 /* Create cgraph edges for function calls inside BODY from NODE. */ 567 568 static void 569 cgraph_create_edges (struct cgraph_node *node, tree body) 570 { 571 basic_block bb; 572 573 struct function *this_cfun = DECL_STRUCT_FUNCTION (body); 574 block_stmt_iterator bsi; 575 tree step; 576 visited_nodes = pointer_set_create (); 577 578 /* Reach the trees by walking over the CFG, and note the 579 enclosing basic-blocks in the call edges. */ 580 FOR_EACH_BB_FN (bb, this_cfun) 581 for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi)) 582 { 583 tree stmt = bsi_stmt (bsi); 584 tree call = get_call_expr_in (stmt); 585 tree decl; 586 587 if (call && (decl = get_callee_fndecl (call))) 588 { 589 cgraph_create_edge (node, cgraph_node (decl), stmt, 590 bb->count, 591 bb->loop_depth); 592 walk_tree (&TREE_OPERAND (call, 1), 593 record_reference, node, visited_nodes); 594 if (TREE_CODE (stmt) == MODIFY_EXPR) 595 walk_tree (&TREE_OPERAND (stmt, 0), 596 record_reference, node, visited_nodes); 597 } 598 else 599 walk_tree (bsi_stmt_ptr (bsi), record_reference, node, visited_nodes); 600 } 601 602 /* Look for initializers of constant variables and private statics. */ 603 for (step = DECL_STRUCT_FUNCTION (body)->unexpanded_var_list; 604 step; 605 step = TREE_CHAIN (step)) 606 { 607 tree decl = TREE_VALUE (step); 608 if (TREE_CODE (decl) == VAR_DECL 609 && (TREE_STATIC (decl) && !DECL_EXTERNAL (decl)) 610 && flag_unit_at_a_time) 611 cgraph_varpool_finalize_decl (decl); 612 else if (TREE_CODE (decl) == VAR_DECL && DECL_INITIAL (decl)) 613 walk_tree (&DECL_INITIAL (decl), record_reference, node, visited_nodes); 614 } 615 616 pointer_set_destroy (visited_nodes); 617 visited_nodes = NULL; 618 } 619 620 /* Give initial reasons why inlining would fail. Those gets 621 either NULLified or usually overwritten by more precise reason 622 later. */ 623 static void 624 initialize_inline_failed (struct cgraph_node *node) 625 { 626 struct cgraph_edge *e; 627 628 for (e = node->callers; e; e = e->next_caller) 629 { 630 gcc_assert (!e->callee->global.inlined_to); 631 gcc_assert (e->inline_failed); 632 if (node->local.redefined_extern_inline) 633 e->inline_failed = N_("redefined extern inline functions are not " 634 "considered for inlining"); 635 else if (!node->local.inlinable) 636 e->inline_failed = N_("function not inlinable"); 637 else 638 e->inline_failed = N_("function not considered for inlining"); 639 } 640 } 641 642 /* Rebuild call edges from current function after a passes not aware 643 of cgraph updating. */ 644 static unsigned int 645 rebuild_cgraph_edges (void) 646 { 647 basic_block bb; 648 struct cgraph_node *node = cgraph_node (current_function_decl); 649 block_stmt_iterator bsi; 650 651 cgraph_node_remove_callees (node); 652 653 node->count = ENTRY_BLOCK_PTR->count; 654 655 FOR_EACH_BB (bb) 656 for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi)) 657 { 658 tree stmt = bsi_stmt (bsi); 659 tree call = get_call_expr_in (stmt); 660 tree decl; 661 662 if (call && (decl = get_callee_fndecl (call))) 663 cgraph_create_edge (node, cgraph_node (decl), stmt, 664 bb->count, 665 bb->loop_depth); 666 } 667 initialize_inline_failed (node); 668 gcc_assert (!node->global.inlined_to); 669 return 0; 670 } 671 672 struct tree_opt_pass pass_rebuild_cgraph_edges = 673 { 674 NULL, /* name */ 675 NULL, /* gate */ 676 rebuild_cgraph_edges, /* execute */ 677 NULL, /* sub */ 678 NULL, /* next */ 679 0, /* static_pass_number */ 680 0, /* tv_id */ 681 PROP_cfg, /* properties_required */ 682 0, /* properties_provided */ 683 0, /* properties_destroyed */ 684 0, /* todo_flags_start */ 685 0, /* todo_flags_finish */ 686 0 /* letter */ 687 }; 688 689 /* Verify cgraph nodes of given cgraph node. */ 690 void 691 verify_cgraph_node (struct cgraph_node *node) 692 { 693 struct cgraph_edge *e; 694 struct cgraph_node *main_clone; 695 struct function *this_cfun = DECL_STRUCT_FUNCTION (node->decl); 696 basic_block this_block; 697 block_stmt_iterator bsi; 698 bool error_found = false; 699 700 if (errorcount || sorrycount) 701 return; 702 703 timevar_push (TV_CGRAPH_VERIFY); 704 for (e = node->callees; e; e = e->next_callee) 705 if (e->aux) 706 { 707 error ("aux field set for edge %s->%s", 708 cgraph_node_name (e->caller), cgraph_node_name (e->callee)); 709 error_found = true; 710 } 711 if (node->count < 0) 712 { 713 error ("Execution count is negative"); 714 error_found = true; 715 } 716 for (e = node->callers; e; e = e->next_caller) 717 { 718 if (e->count < 0) 719 { 720 error ("caller edge count is negative"); 721 error_found = true; 722 } 723 if (!e->inline_failed) 724 { 725 if (node->global.inlined_to 726 != (e->caller->global.inlined_to 727 ? e->caller->global.inlined_to : e->caller)) 728 { 729 error ("inlined_to pointer is wrong"); 730 error_found = true; 731 } 732 if (node->callers->next_caller) 733 { 734 error ("multiple inline callers"); 735 error_found = true; 736 } 737 } 738 else 739 if (node->global.inlined_to) 740 { 741 error ("inlined_to pointer set for noninline callers"); 742 error_found = true; 743 } 744 } 745 if (!node->callers && node->global.inlined_to) 746 { 747 error ("inlined_to pointer is set but no predecessors found"); 748 error_found = true; 749 } 750 if (node->global.inlined_to == node) 751 { 752 error ("inlined_to pointer refers to itself"); 753 error_found = true; 754 } 755 756 for (main_clone = cgraph_node (node->decl); main_clone; 757 main_clone = main_clone->next_clone) 758 if (main_clone == node) 759 break; 760 if (!cgraph_node (node->decl)) 761 { 762 error ("node not found in cgraph_hash"); 763 error_found = true; 764 } 765 766 if (node->analyzed 767 && DECL_SAVED_TREE (node->decl) && !TREE_ASM_WRITTEN (node->decl) 768 && (!DECL_EXTERNAL (node->decl) || node->global.inlined_to)) 769 { 770 if (this_cfun->cfg) 771 { 772 /* The nodes we're interested in are never shared, so walk 773 the tree ignoring duplicates. */ 774 visited_nodes = pointer_set_create (); 775 /* Reach the trees by walking over the CFG, and note the 776 enclosing basic-blocks in the call edges. */ 777 FOR_EACH_BB_FN (this_block, this_cfun) 778 for (bsi = bsi_start (this_block); !bsi_end_p (bsi); bsi_next (&bsi)) 779 { 780 tree stmt = bsi_stmt (bsi); 781 tree call = get_call_expr_in (stmt); 782 tree decl; 783 if (call && (decl = get_callee_fndecl (call))) 784 { 785 struct cgraph_edge *e = cgraph_edge (node, stmt); 786 if (e) 787 { 788 if (e->aux) 789 { 790 error ("shared call_stmt:"); 791 debug_generic_stmt (stmt); 792 error_found = true; 793 } 794 if (e->callee->decl != cgraph_node (decl)->decl 795 && e->inline_failed) 796 { 797 error ("edge points to wrong declaration:"); 798 debug_tree (e->callee->decl); 799 fprintf (stderr," Instead of:"); 800 debug_tree (decl); 801 } 802 e->aux = (void *)1; 803 } 804 else 805 { 806 error ("missing callgraph edge for call stmt:"); 807 debug_generic_stmt (stmt); 808 error_found = true; 809 } 810 } 811 } 812 pointer_set_destroy (visited_nodes); 813 visited_nodes = NULL; 814 } 815 else 816 /* No CFG available?! */ 817 gcc_unreachable (); 818 819 for (e = node->callees; e; e = e->next_callee) 820 { 821 if (!e->aux) 822 { 823 error ("edge %s->%s has no corresponding call_stmt", 824 cgraph_node_name (e->caller), 825 cgraph_node_name (e->callee)); 826 debug_generic_stmt (e->call_stmt); 827 error_found = true; 828 } 829 e->aux = 0; 830 } 831 } 832 if (error_found) 833 { 834 dump_cgraph_node (stderr, node); 835 internal_error ("verify_cgraph_node failed"); 836 } 837 timevar_pop (TV_CGRAPH_VERIFY); 838 } 839 840 /* Verify whole cgraph structure. */ 841 void 842 verify_cgraph (void) 843 { 844 struct cgraph_node *node; 845 846 if (sorrycount || errorcount) 847 return; 848 849 for (node = cgraph_nodes; node; node = node->next) 850 verify_cgraph_node (node); 851 } 852 853 /* Output one variable, if necessary. Return whether we output it. */ 854 static bool 855 cgraph_varpool_assemble_decl (struct cgraph_varpool_node *node) 856 { 857 tree decl = node->decl; 858 859 if (!TREE_ASM_WRITTEN (decl) 860 && !node->alias 861 && !DECL_EXTERNAL (decl) 862 && (TREE_CODE (decl) != VAR_DECL || !DECL_HAS_VALUE_EXPR_P (decl))) 863 { 864 assemble_variable (decl, 0, 1, 0); 865 return TREE_ASM_WRITTEN (decl); 866 } 867 868 return false; 869 } 870 871 /* Output all variables enqueued to be assembled. */ 872 bool 873 cgraph_varpool_assemble_pending_decls (void) 874 { 875 bool changed = false; 876 877 if (errorcount || sorrycount) 878 return false; 879 880 /* EH might mark decls as needed during expansion. This should be safe since 881 we don't create references to new function, but it should not be used 882 elsewhere. */ 883 cgraph_varpool_analyze_pending_decls (); 884 885 while (cgraph_varpool_nodes_queue) 886 { 887 struct cgraph_varpool_node *node = cgraph_varpool_nodes_queue; 888 889 cgraph_varpool_nodes_queue = cgraph_varpool_nodes_queue->next_needed; 890 if (cgraph_varpool_assemble_decl (node)) 891 { 892 changed = true; 893 node->next_needed = cgraph_varpool_assembled_nodes_queue; 894 cgraph_varpool_assembled_nodes_queue = node; 895 node->finalized = 1; 896 } 897 else 898 node->next_needed = NULL; 899 } 900 /* cgraph_varpool_nodes_queue is now empty, clear the pointer to the last 901 element in the queue. */ 902 cgraph_varpool_last_needed_node = NULL; 903 return changed; 904 } 905 /* Output all variables enqueued to be assembled. */ 906 static void 907 cgraph_varpool_output_debug_info (void) 908 { 909 timevar_push (TV_SYMOUT); 910 if (errorcount == 0 && sorrycount == 0) 911 while (cgraph_varpool_assembled_nodes_queue) 912 { 913 struct cgraph_varpool_node *node = cgraph_varpool_assembled_nodes_queue; 914 915 /* Local static variables are never seen by check_global_declarations 916 so we need to output debug info by hand. */ 917 if (DECL_CONTEXT (node->decl) 918 && (TREE_CODE (DECL_CONTEXT (node->decl)) == BLOCK 919 || TREE_CODE (DECL_CONTEXT (node->decl)) == FUNCTION_DECL) 920 && errorcount == 0 && sorrycount == 0) 921 (*debug_hooks->global_decl) (node->decl); 922 cgraph_varpool_assembled_nodes_queue = node->next_needed; 923 node->next_needed = 0; 924 } 925 timevar_pop (TV_SYMOUT); 926 } 927 928 /* Output all asm statements we have stored up to be output. */ 929 930 static void 931 cgraph_output_pending_asms (void) 932 { 933 struct cgraph_asm_node *can; 934 935 if (errorcount || sorrycount) 936 return; 937 938 for (can = cgraph_asm_nodes; can; can = can->next) 939 assemble_asm (can->asm_str); 940 cgraph_asm_nodes = NULL; 941 } 942 943 /* Analyze the function scheduled to be output. */ 944 void 945 cgraph_analyze_function (struct cgraph_node *node) 946 { 947 tree decl = node->decl; 948 949 current_function_decl = decl; 950 push_cfun (DECL_STRUCT_FUNCTION (decl)); 951 cgraph_lower_function (node); 952 953 /* First kill forward declaration so reverse inlining works properly. */ 954 cgraph_create_edges (node, decl); 955 956 node->local.inlinable = tree_inlinable_function_p (decl); 957 if (!flag_unit_at_a_time) 958 node->local.self_insns = estimate_num_insns (decl); 959 if (node->local.inlinable) 960 node->local.disregard_inline_limits 961 = lang_hooks.tree_inlining.disregard_inline_limits (decl); 962 initialize_inline_failed (node); 963 if (flag_really_no_inline && !node->local.disregard_inline_limits) 964 node->local.inlinable = 0; 965 /* Inlining characteristics are maintained by the cgraph_mark_inline. */ 966 node->global.insns = node->local.self_insns; 967 968 node->analyzed = true; 969 pop_cfun (); 970 current_function_decl = NULL; 971 } 972 973 /* Look for externally_visible and used attributes and mark cgraph nodes 974 accordingly. 975 976 We cannot mark the nodes at the point the attributes are processed (in 977 handle_*_attribute) because the copy of the declarations available at that 978 point may not be canonical. For example, in: 979 980 void f(); 981 void f() __attribute__((used)); 982 983 the declaration we see in handle_used_attribute will be the second 984 declaration -- but the front end will subsequently merge that declaration 985 with the original declaration and discard the second declaration. 986 987 Furthermore, we can't mark these nodes in cgraph_finalize_function because: 988 989 void f() {} 990 void f() __attribute__((externally_visible)); 991 992 is valid. 993 994 So, we walk the nodes at the end of the translation unit, applying the 995 attributes at that point. */ 996 997 static void 998 process_function_and_variable_attributes (struct cgraph_node *first, 999 struct cgraph_varpool_node *first_var) 1000 { 1001 struct cgraph_node *node; 1002 struct cgraph_varpool_node *vnode; 1003 1004 for (node = cgraph_nodes; node != first; node = node->next) 1005 { 1006 tree decl = node->decl; 1007 if (lookup_attribute ("used", DECL_ATTRIBUTES (decl))) 1008 { 1009 mark_decl_referenced (decl); 1010 if (node->local.finalized) 1011 cgraph_mark_needed_node (node); 1012 } 1013 if (lookup_attribute ("externally_visible", DECL_ATTRIBUTES (decl))) 1014 { 1015 if (! TREE_PUBLIC (node->decl)) 1016 warning (OPT_Wattributes, 1017 "%J%<externally_visible%> attribute have effect only on public objects", 1018 node->decl); 1019 else 1020 { 1021 if (node->local.finalized) 1022 cgraph_mark_needed_node (node); 1023 node->local.externally_visible = true; 1024 } 1025 } 1026 } 1027 for (vnode = cgraph_varpool_nodes; vnode != first_var; vnode = vnode->next) 1028 { 1029 tree decl = vnode->decl; 1030 if (lookup_attribute ("used", DECL_ATTRIBUTES (decl))) 1031 { 1032 mark_decl_referenced (decl); 1033 if (vnode->finalized) 1034 cgraph_varpool_mark_needed_node (vnode); 1035 } 1036 if (lookup_attribute ("externally_visible", DECL_ATTRIBUTES (decl))) 1037 { 1038 if (! TREE_PUBLIC (vnode->decl)) 1039 warning (OPT_Wattributes, 1040 "%J%<externally_visible%> attribute have effect only on public objects", 1041 vnode->decl); 1042 else 1043 { 1044 if (vnode->finalized) 1045 cgraph_varpool_mark_needed_node (vnode); 1046 vnode->externally_visible = true; 1047 } 1048 } 1049 } 1050 } 1051 1052 /* Analyze the whole compilation unit once it is parsed completely. */ 1053 1054 void 1055 cgraph_finalize_compilation_unit (void) 1056 { 1057 struct cgraph_node *node, *next; 1058 /* Keep track of already processed nodes when called multiple times for 1059 intermodule optimization. */ 1060 static struct cgraph_node *first_analyzed; 1061 struct cgraph_node *first_processed = first_analyzed; 1062 static struct cgraph_varpool_node *first_analyzed_var; 1063 1064 if (errorcount || sorrycount) 1065 return; 1066 1067 finish_aliases_1 (); 1068 1069 if (!flag_unit_at_a_time) 1070 { 1071 cgraph_output_pending_asms (); 1072 cgraph_assemble_pending_functions (); 1073 cgraph_varpool_output_debug_info (); 1074 return; 1075 } 1076 1077 if (!quiet_flag) 1078 { 1079 fprintf (stderr, "\nAnalyzing compilation unit"); 1080 fflush (stderr); 1081 } 1082 1083 timevar_push (TV_CGRAPH); 1084 process_function_and_variable_attributes (first_processed, 1085 first_analyzed_var); 1086 first_processed = cgraph_nodes; 1087 first_analyzed_var = cgraph_varpool_nodes; 1088 cgraph_varpool_analyze_pending_decls (); 1089 if (cgraph_dump_file) 1090 { 1091 fprintf (cgraph_dump_file, "Initial entry points:"); 1092 for (node = cgraph_nodes; node != first_analyzed; node = node->next) 1093 if (node->needed && DECL_SAVED_TREE (node->decl)) 1094 fprintf (cgraph_dump_file, " %s", cgraph_node_name (node)); 1095 fprintf (cgraph_dump_file, "\n"); 1096 } 1097 1098 /* Propagate reachability flag and lower representation of all reachable 1099 functions. In the future, lowering will introduce new functions and 1100 new entry points on the way (by template instantiation and virtual 1101 method table generation for instance). */ 1102 while (cgraph_nodes_queue) 1103 { 1104 struct cgraph_edge *edge; 1105 tree decl = cgraph_nodes_queue->decl; 1106 1107 node = cgraph_nodes_queue; 1108 cgraph_nodes_queue = cgraph_nodes_queue->next_needed; 1109 node->next_needed = NULL; 1110 1111 /* ??? It is possible to create extern inline function and later using 1112 weak alias attribute to kill its body. See 1113 gcc.c-torture/compile/20011119-1.c */ 1114 if (!DECL_SAVED_TREE (decl)) 1115 { 1116 cgraph_reset_node (node); 1117 continue; 1118 } 1119 1120 gcc_assert (!node->analyzed && node->reachable); 1121 gcc_assert (DECL_SAVED_TREE (decl)); 1122 1123 cgraph_analyze_function (node); 1124 1125 for (edge = node->callees; edge; edge = edge->next_callee) 1126 if (!edge->callee->reachable) 1127 cgraph_mark_reachable_node (edge->callee); 1128 1129 /* We finalize local static variables during constructing callgraph 1130 edges. Process their attributes too. */ 1131 process_function_and_variable_attributes (first_processed, 1132 first_analyzed_var); 1133 first_processed = cgraph_nodes; 1134 first_analyzed_var = cgraph_varpool_nodes; 1135 cgraph_varpool_analyze_pending_decls (); 1136 } 1137 1138 /* Collect entry points to the unit. */ 1139 if (cgraph_dump_file) 1140 { 1141 fprintf (cgraph_dump_file, "Unit entry points:"); 1142 for (node = cgraph_nodes; node != first_analyzed; node = node->next) 1143 if (node->needed && DECL_SAVED_TREE (node->decl)) 1144 fprintf (cgraph_dump_file, " %s", cgraph_node_name (node)); 1145 fprintf (cgraph_dump_file, "\n\nInitial "); 1146 dump_cgraph (cgraph_dump_file); 1147 } 1148 1149 if (cgraph_dump_file) 1150 fprintf (cgraph_dump_file, "\nReclaiming functions:"); 1151 1152 for (node = cgraph_nodes; node != first_analyzed; node = next) 1153 { 1154 tree decl = node->decl; 1155 next = node->next; 1156 1157 if (node->local.finalized && !DECL_SAVED_TREE (decl)) 1158 cgraph_reset_node (node); 1159 1160 if (!node->reachable && DECL_SAVED_TREE (decl)) 1161 { 1162 if (cgraph_dump_file) 1163 fprintf (cgraph_dump_file, " %s", cgraph_node_name (node)); 1164 cgraph_remove_node (node); 1165 continue; 1166 } 1167 else 1168 node->next_needed = NULL; 1169 gcc_assert (!node->local.finalized || DECL_SAVED_TREE (decl)); 1170 gcc_assert (node->analyzed == node->local.finalized); 1171 } 1172 if (cgraph_dump_file) 1173 { 1174 fprintf (cgraph_dump_file, "\n\nReclaimed "); 1175 dump_cgraph (cgraph_dump_file); 1176 } 1177 first_analyzed = cgraph_nodes; 1178 ggc_collect (); 1179 timevar_pop (TV_CGRAPH); 1180 } 1181 /* Figure out what functions we want to assemble. */ 1182 1183 static void 1184 cgraph_mark_functions_to_output (void) 1185 { 1186 struct cgraph_node *node; 1187 1188 for (node = cgraph_nodes; node; node = node->next) 1189 { 1190 tree decl = node->decl; 1191 struct cgraph_edge *e; 1192 1193 gcc_assert (!node->output); 1194 1195 for (e = node->callers; e; e = e->next_caller) 1196 if (e->inline_failed) 1197 break; 1198 1199 /* We need to output all local functions that are used and not 1200 always inlined, as well as those that are reachable from 1201 outside the current compilation unit. */ 1202 if (DECL_SAVED_TREE (decl) 1203 && !node->global.inlined_to 1204 && (node->needed 1205 || (e && node->reachable)) 1206 && !TREE_ASM_WRITTEN (decl) 1207 && !DECL_EXTERNAL (decl)) 1208 node->output = 1; 1209 else 1210 { 1211 /* We should've reclaimed all functions that are not needed. */ 1212 #ifdef ENABLE_CHECKING 1213 if (!node->global.inlined_to && DECL_SAVED_TREE (decl) 1214 && !DECL_EXTERNAL (decl)) 1215 { 1216 dump_cgraph_node (stderr, node); 1217 internal_error ("failed to reclaim unneeded function"); 1218 } 1219 #endif 1220 gcc_assert (node->global.inlined_to || !DECL_SAVED_TREE (decl) 1221 || DECL_EXTERNAL (decl)); 1222 1223 } 1224 1225 } 1226 } 1227 1228 /* Expand function specified by NODE. */ 1229 1230 static void 1231 cgraph_expand_function (struct cgraph_node *node) 1232 { 1233 tree decl = node->decl; 1234 1235 /* We ought to not compile any inline clones. */ 1236 gcc_assert (!node->global.inlined_to); 1237 1238 if (flag_unit_at_a_time) 1239 announce_function (decl); 1240 1241 cgraph_lower_function (node); 1242 1243 /* Generate RTL for the body of DECL. */ 1244 lang_hooks.callgraph.expand_function (decl); 1245 1246 /* Make sure that BE didn't give up on compiling. */ 1247 /* ??? Can happen with nested function of extern inline. */ 1248 gcc_assert (TREE_ASM_WRITTEN (node->decl)); 1249 1250 current_function_decl = NULL; 1251 if (!cgraph_preserve_function_body_p (node->decl)) 1252 { 1253 DECL_SAVED_TREE (node->decl) = NULL; 1254 DECL_STRUCT_FUNCTION (node->decl) = NULL; 1255 DECL_INITIAL (node->decl) = error_mark_node; 1256 /* Eliminate all call edges. This is important so the call_expr no longer 1257 points to the dead function body. */ 1258 cgraph_node_remove_callees (node); 1259 } 1260 1261 cgraph_function_flags_ready = true; 1262 } 1263 1264 /* Return true when CALLER_DECL should be inlined into CALLEE_DECL. */ 1265 1266 bool 1267 cgraph_inline_p (struct cgraph_edge *e, const char **reason) 1268 { 1269 *reason = e->inline_failed; 1270 return !e->inline_failed; 1271 } 1272 1273 1274 1275 /* Expand all functions that must be output. 1276 1277 Attempt to topologically sort the nodes so function is output when 1278 all called functions are already assembled to allow data to be 1279 propagated across the callgraph. Use a stack to get smaller distance 1280 between a function and its callees (later we may choose to use a more 1281 sophisticated algorithm for function reordering; we will likely want 1282 to use subsections to make the output functions appear in top-down 1283 order). */ 1284 1285 static void 1286 cgraph_expand_all_functions (void) 1287 { 1288 struct cgraph_node *node; 1289 struct cgraph_node **order = XCNEWVEC (struct cgraph_node *, cgraph_n_nodes); 1290 int order_pos = 0, new_order_pos = 0; 1291 int i; 1292 1293 order_pos = cgraph_postorder (order); 1294 gcc_assert (order_pos == cgraph_n_nodes); 1295 1296 /* Garbage collector may remove inline clones we eliminate during 1297 optimization. So we must be sure to not reference them. */ 1298 for (i = 0; i < order_pos; i++) 1299 if (order[i]->output) 1300 order[new_order_pos++] = order[i]; 1301 1302 for (i = new_order_pos - 1; i >= 0; i--) 1303 { 1304 node = order[i]; 1305 if (node->output) 1306 { 1307 gcc_assert (node->reachable); 1308 node->output = 0; 1309 cgraph_expand_function (node); 1310 } 1311 } 1312 1313 free (order); 1314 1315 /* Process CGRAPH_EXPAND_QUEUE, these are functions created during 1316 the expansion process. Note that this queue may grow as its 1317 being processed, as the new functions may generate new ones. */ 1318 while (cgraph_expand_queue) 1319 { 1320 node = cgraph_expand_queue; 1321 cgraph_expand_queue = cgraph_expand_queue->next_needed; 1322 node->next_needed = NULL; 1323 node->output = 0; 1324 node->lowered = DECL_STRUCT_FUNCTION (node->decl)->cfg != NULL; 1325 cgraph_expand_function (node); 1326 } 1327 } 1328 1329 /* This is used to sort the node types by the cgraph order number. */ 1330 1331 struct cgraph_order_sort 1332 { 1333 enum { ORDER_UNDEFINED = 0, ORDER_FUNCTION, ORDER_VAR, ORDER_ASM } kind; 1334 union 1335 { 1336 struct cgraph_node *f; 1337 struct cgraph_varpool_node *v; 1338 struct cgraph_asm_node *a; 1339 } u; 1340 }; 1341 1342 /* Output all functions, variables, and asm statements in the order 1343 according to their order fields, which is the order in which they 1344 appeared in the file. This implements -fno-toplevel-reorder. In 1345 this mode we may output functions and variables which don't really 1346 need to be output. */ 1347 1348 static void 1349 cgraph_output_in_order (void) 1350 { 1351 int max; 1352 size_t size; 1353 struct cgraph_order_sort *nodes; 1354 int i; 1355 struct cgraph_node *pf; 1356 struct cgraph_varpool_node *pv; 1357 struct cgraph_asm_node *pa; 1358 1359 max = cgraph_order; 1360 size = max * sizeof (struct cgraph_order_sort); 1361 nodes = (struct cgraph_order_sort *) alloca (size); 1362 memset (nodes, 0, size); 1363 1364 cgraph_varpool_analyze_pending_decls (); 1365 1366 for (pf = cgraph_nodes; pf; pf = pf->next) 1367 { 1368 if (pf->output) 1369 { 1370 i = pf->order; 1371 gcc_assert (nodes[i].kind == ORDER_UNDEFINED); 1372 nodes[i].kind = ORDER_FUNCTION; 1373 nodes[i].u.f = pf; 1374 } 1375 } 1376 1377 for (pv = cgraph_varpool_nodes_queue; pv; pv = pv->next_needed) 1378 { 1379 i = pv->order; 1380 gcc_assert (nodes[i].kind == ORDER_UNDEFINED); 1381 nodes[i].kind = ORDER_VAR; 1382 nodes[i].u.v = pv; 1383 } 1384 1385 for (pa = cgraph_asm_nodes; pa; pa = pa->next) 1386 { 1387 i = pa->order; 1388 gcc_assert (nodes[i].kind == ORDER_UNDEFINED); 1389 nodes[i].kind = ORDER_ASM; 1390 nodes[i].u.a = pa; 1391 } 1392 1393 for (i = 0; i < max; ++i) 1394 { 1395 switch (nodes[i].kind) 1396 { 1397 case ORDER_FUNCTION: 1398 nodes[i].u.f->output = 0; 1399 cgraph_expand_function (nodes[i].u.f); 1400 break; 1401 1402 case ORDER_VAR: 1403 cgraph_varpool_assemble_decl (nodes[i].u.v); 1404 break; 1405 1406 case ORDER_ASM: 1407 assemble_asm (nodes[i].u.a->asm_str); 1408 break; 1409 1410 case ORDER_UNDEFINED: 1411 break; 1412 1413 default: 1414 gcc_unreachable (); 1415 } 1416 } 1417 1418 cgraph_asm_nodes = NULL; 1419 } 1420 1421 /* Mark visibility of all functions. 1422 1423 A local function is one whose calls can occur only in the current 1424 compilation unit and all its calls are explicit, so we can change 1425 its calling convention. We simply mark all static functions whose 1426 address is not taken as local. 1427 1428 We also change the TREE_PUBLIC flag of all declarations that are public 1429 in language point of view but we want to overwrite this default 1430 via visibilities for the backend point of view. */ 1431 1432 static void 1433 cgraph_function_and_variable_visibility (void) 1434 { 1435 struct cgraph_node *node; 1436 struct cgraph_varpool_node *vnode; 1437 1438 for (node = cgraph_nodes; node; node = node->next) 1439 { 1440 if (node->reachable 1441 && (DECL_COMDAT (node->decl) 1442 || (!flag_whole_program 1443 && TREE_PUBLIC (node->decl) && !DECL_EXTERNAL (node->decl)))) 1444 node->local.externally_visible = true; 1445 if (!node->local.externally_visible && node->analyzed 1446 && !DECL_EXTERNAL (node->decl)) 1447 { 1448 gcc_assert (flag_whole_program || !TREE_PUBLIC (node->decl)); 1449 TREE_PUBLIC (node->decl) = 0; 1450 } 1451 node->local.local = (!node->needed 1452 && node->analyzed 1453 && !DECL_EXTERNAL (node->decl) 1454 && !node->local.externally_visible); 1455 } 1456 for (vnode = cgraph_varpool_nodes_queue; vnode; vnode = vnode->next_needed) 1457 { 1458 if (vnode->needed 1459 && !flag_whole_program 1460 && (DECL_COMDAT (vnode->decl) || TREE_PUBLIC (vnode->decl))) 1461 vnode->externally_visible = 1; 1462 if (!vnode->externally_visible) 1463 { 1464 gcc_assert (flag_whole_program || !TREE_PUBLIC (vnode->decl)); 1465 TREE_PUBLIC (vnode->decl) = 0; 1466 } 1467 gcc_assert (TREE_STATIC (vnode->decl)); 1468 } 1469 1470 /* Because we have to be conservative on the boundaries of source 1471 level units, it is possible that we marked some functions in 1472 reachable just because they might be used later via external 1473 linkage, but after making them local they are really unreachable 1474 now. */ 1475 cgraph_remove_unreachable_nodes (true, cgraph_dump_file); 1476 1477 if (cgraph_dump_file) 1478 { 1479 fprintf (cgraph_dump_file, "\nMarking local functions:"); 1480 for (node = cgraph_nodes; node; node = node->next) 1481 if (node->local.local) 1482 fprintf (cgraph_dump_file, " %s", cgraph_node_name (node)); 1483 fprintf (cgraph_dump_file, "\n\n"); 1484 fprintf (cgraph_dump_file, "\nMarking externally visible functions:"); 1485 for (node = cgraph_nodes; node; node = node->next) 1486 if (node->local.externally_visible) 1487 fprintf (cgraph_dump_file, " %s", cgraph_node_name (node)); 1488 fprintf (cgraph_dump_file, "\n\n"); 1489 } 1490 cgraph_function_flags_ready = true; 1491 } 1492 1493 /* Return true when function body of DECL still needs to be kept around 1494 for later re-use. */ 1495 bool 1496 cgraph_preserve_function_body_p (tree decl) 1497 { 1498 struct cgraph_node *node; 1499 if (!cgraph_global_info_ready) 1500 return (flag_really_no_inline 1501 ? lang_hooks.tree_inlining.disregard_inline_limits (decl) 1502 : DECL_INLINE (decl)); 1503 /* Look if there is any clone around. */ 1504 for (node = cgraph_node (decl); node; node = node->next_clone) 1505 if (node->global.inlined_to) 1506 return true; 1507 return false; 1508 } 1509 1510 static void 1511 ipa_passes (void) 1512 { 1513 cfun = NULL; 1514 tree_register_cfg_hooks (); 1515 bitmap_obstack_initialize (NULL); 1516 execute_ipa_pass_list (all_ipa_passes); 1517 bitmap_obstack_release (NULL); 1518 } 1519 1520 /* Perform simple optimizations based on callgraph. */ 1521 1522 void 1523 cgraph_optimize (void) 1524 { 1525 if (errorcount || sorrycount) 1526 return; 1527 1528 #ifdef ENABLE_CHECKING 1529 verify_cgraph (); 1530 #endif 1531 if (!flag_unit_at_a_time) 1532 { 1533 cgraph_output_pending_asms (); 1534 cgraph_varpool_assemble_pending_decls (); 1535 cgraph_varpool_output_debug_info (); 1536 return; 1537 } 1538 1539 /* Frontend may output common variables after the unit has been finalized. 1540 It is safe to deal with them here as they are always zero initialized. */ 1541 cgraph_varpool_analyze_pending_decls (); 1542 1543 timevar_push (TV_CGRAPHOPT); 1544 if (!quiet_flag) 1545 fprintf (stderr, "Performing interprocedural optimizations\n"); 1546 1547 cgraph_function_and_variable_visibility (); 1548 if (cgraph_dump_file) 1549 { 1550 fprintf (cgraph_dump_file, "Marked "); 1551 dump_cgraph (cgraph_dump_file); 1552 } 1553 1554 /* Don't run the IPA passes if there was any error or sorry messages. */ 1555 if (errorcount == 0 && sorrycount == 0) 1556 ipa_passes (); 1557 1558 /* This pass remove bodies of extern inline functions we never inlined. 1559 Do this later so other IPA passes see what is really going on. */ 1560 cgraph_remove_unreachable_nodes (false, dump_file); 1561 cgraph_increase_alignment (); 1562 cgraph_global_info_ready = true; 1563 if (cgraph_dump_file) 1564 { 1565 fprintf (cgraph_dump_file, "Optimized "); 1566 dump_cgraph (cgraph_dump_file); 1567 dump_varpool (cgraph_dump_file); 1568 } 1569 timevar_pop (TV_CGRAPHOPT); 1570 1571 /* Output everything. */ 1572 if (!quiet_flag) 1573 fprintf (stderr, "Assembling functions:\n"); 1574 #ifdef ENABLE_CHECKING 1575 verify_cgraph (); 1576 #endif 1577 1578 cgraph_mark_functions_to_output (); 1579 1580 if (!flag_toplevel_reorder) 1581 cgraph_output_in_order (); 1582 else 1583 { 1584 cgraph_output_pending_asms (); 1585 1586 cgraph_expand_all_functions (); 1587 cgraph_varpool_remove_unreferenced_decls (); 1588 1589 cgraph_varpool_assemble_pending_decls (); 1590 cgraph_varpool_output_debug_info (); 1591 } 1592 1593 if (cgraph_dump_file) 1594 { 1595 fprintf (cgraph_dump_file, "\nFinal "); 1596 dump_cgraph (cgraph_dump_file); 1597 } 1598 #ifdef ENABLE_CHECKING 1599 verify_cgraph (); 1600 /* Double check that all inline clones are gone and that all 1601 function bodies have been released from memory. */ 1602 if (flag_unit_at_a_time 1603 && !(sorrycount || errorcount)) 1604 { 1605 struct cgraph_node *node; 1606 bool error_found = false; 1607 1608 for (node = cgraph_nodes; node; node = node->next) 1609 if (node->analyzed 1610 && (node->global.inlined_to 1611 || DECL_SAVED_TREE (node->decl))) 1612 { 1613 error_found = true; 1614 dump_cgraph_node (stderr, node); 1615 } 1616 if (error_found) 1617 internal_error ("nodes with no released memory found"); 1618 } 1619 #endif 1620 } 1621 1622 /* Increase alignment of global arrays to improve vectorization potential. 1623 TODO: 1624 - Consider also structs that have an array field. 1625 - Use ipa analysis to prune arrays that can't be vectorized? 1626 This should involve global alignment analysis and in the future also 1627 array padding. */ 1628 1629 static void 1630 cgraph_increase_alignment (void) 1631 { 1632 if (flag_section_anchors && flag_tree_vectorize) 1633 { 1634 struct cgraph_varpool_node *vnode; 1635 1636 /* Increase the alignment of all global arrays for vectorization. */ 1637 for (vnode = cgraph_varpool_nodes_queue; 1638 vnode; 1639 vnode = vnode->next_needed) 1640 { 1641 tree vectype, decl = vnode->decl; 1642 unsigned int alignment; 1643 1644 if (TREE_CODE (TREE_TYPE (decl)) != ARRAY_TYPE) 1645 continue; 1646 vectype = get_vectype_for_scalar_type (TREE_TYPE (TREE_TYPE (decl))); 1647 if (!vectype) 1648 continue; 1649 alignment = TYPE_ALIGN (vectype); 1650 if (DECL_ALIGN (decl) >= alignment) 1651 continue; 1652 1653 if (vect_can_force_dr_alignment_p (decl, alignment)) 1654 { 1655 DECL_ALIGN (decl) = TYPE_ALIGN (vectype); 1656 DECL_USER_ALIGN (decl) = 1; 1657 if (cgraph_dump_file) 1658 { 1659 fprintf (cgraph_dump_file, "Increasing alignment of decl: "); 1660 print_generic_expr (cgraph_dump_file, decl, TDF_SLIM); 1661 } 1662 } 1663 } 1664 } 1665 } 1666 1667 /* Generate and emit a static constructor or destructor. WHICH must be 1668 one of 'I' or 'D'. BODY should be a STATEMENT_LIST containing 1669 GENERIC statements. */ 1670 1671 void 1672 cgraph_build_static_cdtor (char which, tree body, int priority) 1673 { 1674 static int counter = 0; 1675 char which_buf[16]; 1676 tree decl, name, resdecl; 1677 1678 sprintf (which_buf, "%c_%d", which, counter++); 1679 name = get_file_function_name_long (which_buf); 1680 1681 decl = build_decl (FUNCTION_DECL, name, 1682 build_function_type (void_type_node, void_list_node)); 1683 current_function_decl = decl; 1684 1685 resdecl = build_decl (RESULT_DECL, NULL_TREE, void_type_node); 1686 DECL_ARTIFICIAL (resdecl) = 1; 1687 DECL_IGNORED_P (resdecl) = 1; 1688 DECL_RESULT (decl) = resdecl; 1689 1690 allocate_struct_function (decl); 1691 1692 TREE_STATIC (decl) = 1; 1693 TREE_USED (decl) = 1; 1694 DECL_ARTIFICIAL (decl) = 1; 1695 DECL_IGNORED_P (decl) = 1; 1696 DECL_NO_INSTRUMENT_FUNCTION_ENTRY_EXIT (decl) = 1; 1697 DECL_SAVED_TREE (decl) = body; 1698 TREE_PUBLIC (decl) = ! targetm.have_ctors_dtors; 1699 DECL_UNINLINABLE (decl) = 1; 1700 1701 DECL_INITIAL (decl) = make_node (BLOCK); 1702 TREE_USED (DECL_INITIAL (decl)) = 1; 1703 1704 DECL_SOURCE_LOCATION (decl) = input_location; 1705 cfun->function_end_locus = input_location; 1706 1707 switch (which) 1708 { 1709 case 'I': 1710 DECL_STATIC_CONSTRUCTOR (decl) = 1; 1711 break; 1712 case 'D': 1713 DECL_STATIC_DESTRUCTOR (decl) = 1; 1714 break; 1715 default: 1716 gcc_unreachable (); 1717 } 1718 1719 gimplify_function_tree (decl); 1720 1721 /* ??? We will get called LATE in the compilation process. */ 1722 if (cgraph_global_info_ready) 1723 { 1724 tree_lowering_passes (decl); 1725 tree_rest_of_compilation (decl); 1726 } 1727 else 1728 cgraph_finalize_function (decl, 0); 1729 1730 if (targetm.have_ctors_dtors) 1731 { 1732 void (*fn) (rtx, int); 1733 1734 if (which == 'I') 1735 fn = targetm.asm_out.constructor; 1736 else 1737 fn = targetm.asm_out.destructor; 1738 fn (XEXP (DECL_RTL (decl), 0), priority); 1739 } 1740 } 1741 1742 void 1743 init_cgraph (void) 1744 { 1745 cgraph_dump_file = dump_begin (TDI_cgraph, NULL); 1746 } 1747 1748 /* The edges representing the callers of the NEW_VERSION node were 1749 fixed by cgraph_function_versioning (), now the call_expr in their 1750 respective tree code should be updated to call the NEW_VERSION. */ 1751 1752 static void 1753 update_call_expr (struct cgraph_node *new_version) 1754 { 1755 struct cgraph_edge *e; 1756 1757 gcc_assert (new_version); 1758 for (e = new_version->callers; e; e = e->next_caller) 1759 /* Update the call expr on the edges 1760 to call the new version. */ 1761 TREE_OPERAND (TREE_OPERAND (get_call_expr_in (e->call_stmt), 0), 0) = new_version->decl; 1762 } 1763 1764 1765 /* Create a new cgraph node which is the new version of 1766 OLD_VERSION node. REDIRECT_CALLERS holds the callers 1767 edges which should be redirected to point to 1768 NEW_VERSION. ALL the callees edges of OLD_VERSION 1769 are cloned to the new version node. Return the new 1770 version node. */ 1771 1772 static struct cgraph_node * 1773 cgraph_copy_node_for_versioning (struct cgraph_node *old_version, 1774 tree new_decl, 1775 VEC(cgraph_edge_p,heap) *redirect_callers) 1776 { 1777 struct cgraph_node *new_version; 1778 struct cgraph_edge *e, *new_e; 1779 struct cgraph_edge *next_callee; 1780 unsigned i; 1781 1782 gcc_assert (old_version); 1783 1784 new_version = cgraph_node (new_decl); 1785 1786 new_version->analyzed = true; 1787 new_version->local = old_version->local; 1788 new_version->global = old_version->global; 1789 new_version->rtl = new_version->rtl; 1790 new_version->reachable = true; 1791 new_version->count = old_version->count; 1792 1793 /* Clone the old node callees. Recursive calls are 1794 also cloned. */ 1795 for (e = old_version->callees;e; e=e->next_callee) 1796 { 1797 new_e = cgraph_clone_edge (e, new_version, e->call_stmt, 0, e->loop_nest, true); 1798 new_e->count = e->count; 1799 } 1800 /* Fix recursive calls. 1801 If OLD_VERSION has a recursive call after the 1802 previous edge cloning, the new version will have an edge 1803 pointing to the old version, which is wrong; 1804 Redirect it to point to the new version. */ 1805 for (e = new_version->callees ; e; e = next_callee) 1806 { 1807 next_callee = e->next_callee; 1808 if (e->callee == old_version) 1809 cgraph_redirect_edge_callee (e, new_version); 1810 1811 if (!next_callee) 1812 break; 1813 } 1814 for (i = 0; VEC_iterate (cgraph_edge_p, redirect_callers, i, e); i++) 1815 { 1816 /* Redirect calls to the old version node to point to its new 1817 version. */ 1818 cgraph_redirect_edge_callee (e, new_version); 1819 } 1820 1821 return new_version; 1822 } 1823 1824 /* Perform function versioning. 1825 Function versioning includes copying of the tree and 1826 a callgraph update (creating a new cgraph node and updating 1827 its callees and callers). 1828 1829 REDIRECT_CALLERS varray includes the edges to be redirected 1830 to the new version. 1831 1832 TREE_MAP is a mapping of tree nodes we want to replace with 1833 new ones (according to results of prior analysis). 1834 OLD_VERSION_NODE is the node that is versioned. 1835 It returns the new version's cgraph node. */ 1836 1837 struct cgraph_node * 1838 cgraph_function_versioning (struct cgraph_node *old_version_node, 1839 VEC(cgraph_edge_p,heap) *redirect_callers, 1840 varray_type tree_map) 1841 { 1842 tree old_decl = old_version_node->decl; 1843 struct cgraph_node *new_version_node = NULL; 1844 tree new_decl; 1845 1846 if (!tree_versionable_function_p (old_decl)) 1847 return NULL; 1848 1849 /* Make a new FUNCTION_DECL tree node for the 1850 new version. */ 1851 new_decl = copy_node (old_decl); 1852 1853 /* Create the new version's call-graph node. 1854 and update the edges of the new node. */ 1855 new_version_node = 1856 cgraph_copy_node_for_versioning (old_version_node, new_decl, 1857 redirect_callers); 1858 1859 /* Copy the OLD_VERSION_NODE function tree to the new version. */ 1860 tree_function_versioning (old_decl, new_decl, tree_map, false); 1861 /* Update the call_expr on the edges to call the new version node. */ 1862 update_call_expr (new_version_node); 1863 1864 /* Update the new version's properties. 1865 Make The new version visible only within this translation unit. 1866 ??? We cannot use COMDAT linkage because there is no 1867 ABI support for this. */ 1868 DECL_EXTERNAL (new_version_node->decl) = 0; 1869 DECL_ONE_ONLY (new_version_node->decl) = 0; 1870 TREE_PUBLIC (new_version_node->decl) = 0; 1871 DECL_COMDAT (new_version_node->decl) = 0; 1872 new_version_node->local.externally_visible = 0; 1873 new_version_node->local.local = 1; 1874 new_version_node->lowered = true; 1875 return new_version_node; 1876 } 1877 1878 /* Produce separate function body for inline clones so the offline copy can be 1879 modified without affecting them. */ 1880 struct cgraph_node * 1881 save_inline_function_body (struct cgraph_node *node) 1882 { 1883 struct cgraph_node *first_clone; 1884 1885 gcc_assert (node == cgraph_node (node->decl)); 1886 1887 cgraph_lower_function (node); 1888 1889 /* In non-unit-at-a-time we construct full fledged clone we never output to 1890 assembly file. This clone is pointed out by inline_decl of original function 1891 and inlining infrastructure knows how to deal with this. */ 1892 if (!flag_unit_at_a_time) 1893 { 1894 struct cgraph_edge *e; 1895 1896 first_clone = cgraph_clone_node (node, node->count, 0, false); 1897 first_clone->needed = 0; 1898 first_clone->reachable = 1; 1899 /* Recursively clone all bodies. */ 1900 for (e = first_clone->callees; e; e = e->next_callee) 1901 if (!e->inline_failed) 1902 cgraph_clone_inlined_nodes (e, true, false); 1903 } 1904 else 1905 first_clone = node->next_clone; 1906 1907 first_clone->decl = copy_node (node->decl); 1908 node->next_clone = NULL; 1909 if (!flag_unit_at_a_time) 1910 node->inline_decl = first_clone->decl; 1911 first_clone->prev_clone = NULL; 1912 cgraph_insert_node_to_hashtable (first_clone); 1913 gcc_assert (first_clone == cgraph_node (first_clone->decl)); 1914 1915 /* Copy the OLD_VERSION_NODE function tree to the new version. */ 1916 tree_function_versioning (node->decl, first_clone->decl, NULL, true); 1917 1918 DECL_EXTERNAL (first_clone->decl) = 0; 1919 DECL_ONE_ONLY (first_clone->decl) = 0; 1920 TREE_PUBLIC (first_clone->decl) = 0; 1921 DECL_COMDAT (first_clone->decl) = 0; 1922 1923 for (node = first_clone->next_clone; node; node = node->next_clone) 1924 node->decl = first_clone->decl; 1925 #ifdef ENABLE_CHECKING 1926 verify_cgraph_node (first_clone); 1927 #endif 1928 return first_clone; 1929 } 1930 1931 #include "gt-cgraphunit.h" 1932