/openbsd-src/gnu/gcc/gcc/ |
H A D | lcm.c | 110 qin = qout = worklist = XNEWVEC (basic_block, n_basic_blocks); in compute_antinout_edge() 125 qend = &worklist[n_basic_blocks - NUM_FIXED_BLOCKS]; in compute_antinout_edge() 126 qlen = n_basic_blocks - NUM_FIXED_BLOCKS; in compute_antinout_edge() 263 = XNEWVEC (basic_block, n_basic_blocks); in compute_laterin() 299 qend = &worklist[n_basic_blocks - NUM_FIXED_BLOCKS]; in compute_laterin() 300 qlen = n_basic_blocks - NUM_FIXED_BLOCKS; in compute_laterin() 488 XNEWVEC (basic_block, n_basic_blocks - NUM_FIXED_BLOCKS); in compute_available() 502 qend = &worklist[n_basic_blocks - NUM_FIXED_BLOCKS]; in compute_available() 503 qlen = n_basic_blocks - NUM_FIXED_BLOCKS; in compute_available() 617 tos = worklist = XNEWVEC (basic_block, n_basic_blocks + 1); in compute_nearerout()
|
H A D | bt-load.c | 464 sbitmap_vector_zero (bb_gen, n_basic_blocks); in compute_defs_uses_and_gen() 465 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in compute_defs_uses_and_gen() 625 sbitmap_vector_zero (bb_kill, n_basic_blocks); in compute_kill() 626 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in compute_kill() 649 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in compute_out() 656 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in compute_out() 675 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in link_btr_uses() 787 sbitmap *bb_gen = sbitmap_vector_alloc (n_basic_blocks, max_uid); in build_btr_def_use_webs() 788 HARD_REG_SET *btrs_written = XCNEWVEC (HARD_REG_SET, n_basic_blocks); in build_btr_def_use_webs() 797 bb_kill = sbitmap_vector_alloc (n_basic_blocks, max_uid); in build_btr_def_use_webs() [all …]
|
H A D | cfganal.c | 174 stack = XNEWVEC (edge_iterator, n_basic_blocks + 1); in mark_dfs_back_edges() 285 tos = worklist = XNEWVEC (basic_block, n_basic_blocks); in find_unreachable_blocks() 348 block_count = n_basic_blocks; /* Include the entry and exit blocks. */ in create_edge_list() 663 stack = XNEWVEC (edge_iterator, n_basic_blocks + 1); in post_order_compute() 737 int rev_post_order_num = n_basic_blocks - 1; in pre_and_rev_post_order_compute() 741 stack = XNEWVEC (edge_iterator, n_basic_blocks + 1); in pre_and_rev_post_order_compute() 821 gcc_assert (pre_order_num == n_basic_blocks); in pre_and_rev_post_order_compute() 826 gcc_assert (pre_order_num == n_basic_blocks - NUM_FIXED_BLOCKS); in pre_and_rev_post_order_compute() 865 data->stack = XNEWVEC (basic_block, n_basic_blocks); in flow_dfs_compute_reverse_init()
|
H A D | tree-ssa-loop-ch.c | 144 bbs = XNEWVEC (basic_block, n_basic_blocks); in copy_loop_headers() 145 copied_bbs = XNEWVEC (basic_block, n_basic_blocks); in copy_loop_headers() 146 bbs_size = n_basic_blocks; in copy_loop_headers()
|
H A D | cfgloop.c | 76 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in flow_loops_cfg_dump() 86 for (i = NUM_FIXED_BLOCKS; i < n_basic_blocks; i++) in flow_loops_cfg_dump() 235 stack = XNEWVEC (basic_block, n_basic_blocks); in flow_loop_nodes_find() 612 if (n_basic_blocks == NUM_FIXED_BLOCKS) in flow_loops_find() 678 loops->parray[0]->num_nodes = n_basic_blocks; in flow_loops_find() 696 dfs_order = XNEWVEC (int, n_basic_blocks); in flow_loops_find() 697 rc_order = XNEWVEC (int, n_basic_blocks); in flow_loops_find() 706 for (b = 0; b < n_basic_blocks - NUM_FIXED_BLOCKS; b++) in flow_loops_find() 797 gcc_assert (loop->num_nodes == (unsigned) n_basic_blocks); in get_loop_body()
|
H A D | cfgbuild.c | 506 gcc_assert (last_basic_block == n_basic_blocks); in find_basic_blocks_1() 536 n_basic_blocks = count_basic_blocks (f); in find_basic_blocks() 550 basic_block_info = VEC_alloc (basic_block, gc, n_basic_blocks); in find_basic_blocks() 551 VEC_safe_grow (basic_block, gc, basic_block_info, n_basic_blocks); in find_basic_blocks() 553 sizeof (basic_block) * n_basic_blocks); in find_basic_blocks()
|
H A D | cfgloopmanip.c | 87 *bbs = XCNEWVEC (basic_block, n_basic_blocks); in find_path() 89 n_basic_blocks, e->dest); in find_path() 285 bord_bbs = XCNEWVEC (basic_block, n_basic_blocks); in remove_path() 311 dom_bbs = XCNEWVEC (basic_block, n_basic_blocks); in remove_path() 379 bbs = XCNEWVEC (basic_block, n_basic_blocks); in add_loop() 381 bbs, n_basic_blocks, loop->header); in add_loop() 471 dom_bbs = XCNEWVEC (basic_block, n_basic_blocks); in loopify() 1415 loops->tree_root->num_nodes = n_basic_blocks; in fix_loop_structure()
|
H A D | tracer.c | 203 basic_block *trace = XNEWVEC (basic_block, n_basic_blocks); in tail_duplicate() 366 if (n_basic_blocks <= NUM_FIXED_BLOCKS + 1) in tracer()
|
H A D | dominance.c | 152 unsigned int num = n_basic_blocks; in init_dom_info() 217 stack = XNEWVEC (edge_iterator, n_basic_blocks + 1); in calc_dfs_tree_nonrec() 374 gcc_assert (di->nodes == (unsigned int) n_basic_blocks - 1); in calc_dfs_tree() 629 n_bbs_in_dom_tree[dir] = n_basic_blocks; in calculate_dominance_info()
|
H A D | cfgcleanup.c | 463 while (counter < n_basic_blocks) in try_forward_edges() 476 counter = n_basic_blocks; in try_forward_edges() 487 threaded_edges = XNEWVEC (edge, n_basic_blocks); in try_forward_edges() 499 counter = n_basic_blocks; in try_forward_edges() 508 gcc_assert (nthreaded_edges < n_basic_blocks - NUM_FIXED_BLOCKS); in try_forward_edges() 524 if (counter >= n_basic_blocks) in try_forward_edges() 2051 && n_basic_blocks > NUM_FIXED_BLOCKS + 1) in try_optimize_cfg()
|
H A D | cfg.c | 178 gcc_assert (i == n_basic_blocks); in compact_blocks() 183 last_basic_block = n_basic_blocks; in compact_blocks() 193 n_basic_blocks--; in expunge_block() 585 fprintf (file, "\n%d basic blocks, %d edges.\n", n_basic_blocks, n_edges); in dump_flow_info()
|
H A D | tree-ssa-uncprop.c | 197 tree *info = XCNEWVEC (tree, n_basic_blocks); in associate_equivalences_with_edges() 218 for (i = 0; i < n_basic_blocks; i++) in associate_equivalences_with_edges()
|
H A D | domwalk.c | 149 basic_block *worklist = XNEWVEC (basic_block, n_basic_blocks * 2); in walk_dominator_tree()
|
H A D | tree-ssa-phiopt.c | 152 n = n_basic_blocks - NUM_FIXED_BLOCKS; in tree_ssa_phiopt() 254 basic_block *order = XNEWVEC (basic_block, n_basic_blocks); in blocks_in_phiopt_order() 255 unsigned n = n_basic_blocks - NUM_FIXED_BLOCKS; in blocks_in_phiopt_order()
|
H A D | tree-cfgcleanup.c | 518 basic_block *worklist = XNEWVEC (basic_block, n_basic_blocks); in cleanup_forwarder_blocks() 762 basic_block *worklist = XNEWVEC (basic_block, n_basic_blocks); in merge_phi_nodes()
|
H A D | global.c | 623 if (n_basic_blocks > NUM_FIXED_BLOCKS) in global_alloc() 2305 rts_order = XNEWVEC (int, n_basic_blocks - NUM_FIXED_BLOCKS); in set_up_bb_rts_numbers() 2307 for (i = 0; i < n_basic_blocks - NUM_FIXED_BLOCKS; i++) in set_up_bb_rts_numbers() 2344 bbs = VEC_alloc (basic_block, heap, n_basic_blocks); in calculate_reg_pav() 2345 new_bbs = VEC_alloc (basic_block, heap, n_basic_blocks); in calculate_reg_pav() 2351 wset = sbitmap_alloc (n_basic_blocks + 1); in calculate_reg_pav()
|
/openbsd-src/gnu/usr.bin/gcc/gcc/ |
H A D | cfganal.c | 150 stack = (edge *) xmalloc ((n_basic_blocks + 1) * sizeof (edge)); in mark_dfs_back_edges() 291 if (n_basic_blocks == 0) 412 (basic_block *) xmalloc (sizeof (basic_block) * n_basic_blocks); in find_unreachable_blocks() 470 block_count = n_basic_blocks + 2; /* Include the entry and exit blocks. */ in create_edge_list() 760 stack = (edge *) xmalloc ((n_basic_blocks + 1) * sizeof (edge)); 827 int rcnum = n_basic_blocks - 1; 831 stack = (edge *) xmalloc ((n_basic_blocks + 1) * sizeof (edge)); 894 if (dfsnum > n_basic_blocks) 898 if (dfsnum < n_basic_blocks) 937 stack = (edge *) xmalloc ((n_basic_blocks + 1) * sizeof (edge)); [all …]
|
H A D | cfgloop.c | 76 for (i = 0; i < n_basic_blocks; i++) 86 for (i = 0; i < n_basic_blocks; i++) 342 stack = (basic_block *) xmalloc (n_basic_blocks * sizeof (basic_block)); 783 if (n_basic_blocks == 0) 849 loops->parray[0]->num_nodes = n_basic_blocks + 2; 867 dfs_order = (int *) xmalloc (n_basic_blocks * sizeof (int)); 868 rc_order = (int *) xmalloc (n_basic_blocks * sizeof (int)); 878 for (b = 0; b < n_basic_blocks; b++) 1006 if (loop->num_nodes != n_basic_blocks + 2)
|
H A D | lcm.c | 119 = (basic_block *) xmalloc (sizeof (basic_block) * n_basic_blocks); 134 qend = &worklist[n_basic_blocks]; 135 qlen = n_basic_blocks; 273 = (basic_block *) xmalloc (sizeof (basic_block) * (n_basic_blocks + 1)); 309 qend = &worklist[n_basic_blocks]; 310 qlen = n_basic_blocks; 502 = (basic_block *) xmalloc (sizeof (basic_block) * n_basic_blocks); 516 qend = &worklist[n_basic_blocks]; 517 qlen = n_basic_blocks; 633 = (basic_block *) xmalloc (sizeof (basic_block) * (n_basic_blocks + 1));
|
H A D | cfg.c | 66 int n_basic_blocks; variable 267 if (i != n_basic_blocks) in compact_blocks() 270 last_basic_block = n_basic_blocks; in compact_blocks() 282 n_basic_blocks--; 568 fprintf (file, "\n%d basic blocks, %d edges.\n", n_basic_blocks, n_edges);
|
H A D | dominance.c | 156 unsigned int num = n_basic_blocks + 1 + 1; 221 stack = (edge *) xmalloc ((n_basic_blocks + 3) * sizeof (edge)); 355 if (di->nodes != (unsigned int) n_basic_blocks + 1) 657 *bbs = xmalloc (n_basic_blocks * sizeof (basic_block)); 671 et_forest_node_t *bbs = xmalloc (n_basic_blocks * sizeof (basic_block));
|
H A D | cfgcleanup.c | 463 while (counter < n_basic_blocks) 473 counter = n_basic_blocks; 486 * n_basic_blocks); 498 counter = n_basic_blocks; 507 if (nthreaded_edges >= n_basic_blocks) 557 if (counter >= n_basic_blocks) 1682 && n_basic_blocks > 1)
|
H A D | tracer.c | 203 basic_block *trace = xmalloc (sizeof (basic_block) * n_basic_blocks); in tail_duplicate() 365 if (n_basic_blocks <= 1) in tracer()
|
H A D | cfgbuild.c | 594 if (last_basic_block != n_basic_blocks) 631 n_basic_blocks = count_basic_blocks (f); 644 VARRAY_BB_INIT (basic_block_info, n_basic_blocks, "basic_block_info");
|
H A D | bb-reorder.c | 308 if (n_basic_blocks <= 1) in reorder_basic_blocks()
|