/openbsd/gnu/gcc/gcc/ |
H A D | lcm.c | 125 qend = &worklist[n_basic_blocks - NUM_FIXED_BLOCKS]; in compute_antinout_edge() 126 qlen = n_basic_blocks - NUM_FIXED_BLOCKS; in compute_antinout_edge() 299 qend = &worklist[n_basic_blocks - NUM_FIXED_BLOCKS]; in compute_laterin() 300 qlen = n_basic_blocks - NUM_FIXED_BLOCKS; in compute_laterin() 488 XNEWVEC (basic_block, n_basic_blocks - NUM_FIXED_BLOCKS); in compute_available() 502 qend = &worklist[n_basic_blocks - NUM_FIXED_BLOCKS]; in compute_available() 503 qlen = n_basic_blocks - NUM_FIXED_BLOCKS; in compute_available()
|
H A D | cfgbuild.c | 146 int count = NUM_FIXED_BLOCKS; in count_basic_blocks() 172 if (count == NUM_FIXED_BLOCKS) in count_basic_blocks() 175 count = NUM_FIXED_BLOCKS + 1; in count_basic_blocks() 537 last_basic_block = NUM_FIXED_BLOCKS; in find_basic_blocks()
|
H A D | bt-load.c | 465 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in compute_defs_uses_and_gen() 626 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in compute_kill() 649 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in compute_out() 656 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in compute_out() 675 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in link_btr_uses() 1411 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in migrate_btr_defs()
|
H A D | tracer.c | 75 if (bb->index < NUM_FIXED_BLOCKS) in ignore_bb_p() 366 if (n_basic_blocks <= NUM_FIXED_BLOCKS + 1) in tracer()
|
H A D | cfgloop.c | 76 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in flow_loops_cfg_dump() 86 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in flow_loops_cfg_dump() 612 if (n_basic_blocks == NUM_FIXED_BLOCKS) in flow_loops_find() 706 for (b = 0; b < n_basic_blocks - NUM_FIXED_BLOCKS; b++) in flow_loops_find()
|
H A D | df-core.c | 891 i = NUM_FIXED_BLOCKS; in df_compact_blocks() 903 for (i = NUM_FIXED_BLOCKS; i < last_basic_block; i++) in df_compact_blocks() 915 i = NUM_FIXED_BLOCKS; in df_compact_blocks()
|
H A D | bb-reorder.c | 1897 if (n_basic_blocks <= NUM_FIXED_BLOCKS + 1) in reorder_basic_blocks() 1989 if (n_basic_blocks <= NUM_FIXED_BLOCKS + 1) in duplicate_computed_gotos() 2173 if (n_basic_blocks <= NUM_FIXED_BLOCKS + 1) in partition_hot_cold_basic_blocks() 2181 if (cur_bb->index >= NUM_FIXED_BLOCKS in partition_hot_cold_basic_blocks() 2182 && cur_bb->next_bb->index >= NUM_FIXED_BLOCKS) in partition_hot_cold_basic_blocks()
|
H A D | cfglayout.c | 601 for (bb = ENTRY_BLOCK_PTR->next_bb, index = NUM_FIXED_BLOCKS; in fixup_reorder_chain() 790 for (bb = ENTRY_BLOCK_PTR->next_bb, index = NUM_FIXED_BLOCKS; in fixup_reorder_chain() 809 index = NUM_FIXED_BLOCKS; in fixup_reorder_chain()
|
H A D | tree-ssa-phiopt.c | 152 n = n_basic_blocks - NUM_FIXED_BLOCKS; in tree_ssa_phiopt() 255 unsigned n = n_basic_blocks - NUM_FIXED_BLOCKS; in blocks_in_phiopt_order()
|
H A D | cfganal.c | 753 rev_post_order_num -= NUM_FIXED_BLOCKS; in pre_and_rev_post_order_compute() 826 gcc_assert (pre_order_num == n_basic_blocks - NUM_FIXED_BLOCKS); in pre_and_rev_post_order_compute()
|
H A D | profile.c | 526 if (bb->index >= NUM_FIXED_BLOCKS in compute_branch_probabilities() 576 if (bb->index >= NUM_FIXED_BLOCKS in compute_branch_probabilities()
|
H A D | cfgcleanup.c | 447 counter = NUM_FIXED_BLOCKS; in try_forward_edges() 508 gcc_assert (nthreaded_edges < n_basic_blocks - NUM_FIXED_BLOCKS); in try_forward_edges() 2051 && n_basic_blocks > NUM_FIXED_BLOCKS + 1) in try_optimize_cfg()
|
H A D | tree-cfg.c | 132 n_basic_blocks = NUM_FIXED_BLOCKS; in init_empty_tree_cfg() 133 last_basic_block = NUM_FIXED_BLOCKS; in init_empty_tree_cfg() 180 if (n_basic_blocks == NUM_FIXED_BLOCKS) in build_tree_cfg() 464 make_edge (ENTRY_BLOCK_PTR, BASIC_BLOCK (NUM_FIXED_BLOCKS), EDGE_FALLTHRU); in make_edges() 833 bsi_start (BASIC_BLOCK (NUM_FIXED_BLOCKS)); in label_to_block_fn() 5219 bb = BASIC_BLOCK (NUM_FIXED_BLOCKS); in print_loop_ir() 5301 if (n_basic_blocks == NUM_FIXED_BLOCKS) in tree_flow_call_edges_add()
|
H A D | cfgrtl.c | 441 return (n_basic_blocks > NUM_FIXED_BLOCKS ? in entry_of_function() 2102 if (num_bb_notes != n_basic_blocks - NUM_FIXED_BLOCKS) in rtl_verify_flow_info() 2757 if (n_basic_blocks == NUM_FIXED_BLOCKS) in rtl_flow_call_edges_add() 2804 for (i = NUM_FIXED_BLOCKS; i < last_bb; i++) in rtl_flow_call_edges_add()
|
H A D | sched-ebb.c | 557 if (n_basic_blocks == NUM_FIXED_BLOCKS) in schedule_ebbs()
|
H A D | global.c | 623 if (n_basic_blocks > NUM_FIXED_BLOCKS) in global_alloc() 2305 rts_order = XNEWVEC (int, n_basic_blocks - NUM_FIXED_BLOCKS); in set_up_bb_rts_numbers() 2307 for (i = 0; i < n_basic_blocks - NUM_FIXED_BLOCKS; i++) in set_up_bb_rts_numbers()
|
H A D | tree-flow-inline.h | 653 gcc_assert (bb->index < NUM_FIXED_BLOCKS); in bsi_start() 685 gcc_assert (bb->index < NUM_FIXED_BLOCKS); in bsi_last()
|
H A D | sched-rgn.c | 1037 int nblocks = n_basic_blocks - NUM_FIXED_BLOCKS; in extend_rgns() 2887 || n_basic_blocks == NUM_FIXED_BLOCKS + 1 in init_regions() 2938 if (n_basic_blocks == NUM_FIXED_BLOCKS) in schedule_insns()
|
H A D | predict.c | 769 if (e->dest->index < NUM_FIXED_BLOCKS in predict_loops() 1515 if (e->dest->index >= NUM_FIXED_BLOCKS in predict_paths_leading_to()
|
H A D | tree-dfa.c | 524 for (i = bsi_start (BASIC_BLOCK (NUM_FIXED_BLOCKS)); in collect_dfa_stats()
|
H A D | basic-block.h | 476 #define NUM_FIXED_BLOCKS (2) macro
|
H A D | cfg.c | 170 i = NUM_FIXED_BLOCKS; in compact_blocks()
|
H A D | tree-ssa-reassoc.c | 1456 for (i = 0; i < n_basic_blocks - NUM_FIXED_BLOCKS; i++) in init_reassoc()
|
H A D | var-tracking.c | 1814 rc_order = XNEWVEC (int, n_basic_blocks - NUM_FIXED_BLOCKS); in vt_find_locations() 1817 for (i = 0; i < n_basic_blocks - NUM_FIXED_BLOCKS; i++) in vt_find_locations()
|
H A D | ifcvt.c | 3498 if (then_bb->index < NUM_FIXED_BLOCKS) in find_if_case_2() 3505 else if (else_succ->dest->index < NUM_FIXED_BLOCKS in find_if_case_2()
|