1 /* Gimple walk support. 2 3 Copyright (C) 2007-2020 Free Software Foundation, Inc. 4 Contributed by Aldy Hernandez <aldyh@redhat.com> 5 6 This file is part of GCC. 7 8 GCC is free software; you can redistribute it and/or modify it under 9 the terms of the GNU General Public License as published by the Free 10 Software Foundation; either version 3, or (at your option) any later 11 version. 12 13 GCC is distributed in the hope that it will be useful, but WITHOUT ANY 14 WARRANTY; without even the implied warranty of MERCHANTABILITY or 15 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 16 for more details. 17 18 You should have received a copy of the GNU General Public License 19 along with GCC; see the file COPYING3. If not see 20 <http://www.gnu.org/licenses/>. */ 21 22 #include "config.h" 23 #include "system.h" 24 #include "coretypes.h" 25 #include "backend.h" 26 #include "tree.h" 27 #include "gimple.h" 28 #include "gimple-iterator.h" 29 #include "gimple-walk.h" 30 #include "stmt.h" 31 32 /* Walk all the statements in the sequence *PSEQ calling walk_gimple_stmt 33 on each one. WI is as in walk_gimple_stmt. 34 35 If walk_gimple_stmt returns non-NULL, the walk is stopped, and the 36 value is stored in WI->CALLBACK_RESULT. Also, the statement that 37 produced the value is returned if this statement has not been 38 removed by a callback (wi->removed_stmt). If the statement has 39 been removed, NULL is returned. 40 41 Otherwise, all the statements are walked and NULL returned. */ 42 43 gimple * 44 walk_gimple_seq_mod (gimple_seq *pseq, walk_stmt_fn callback_stmt, 45 walk_tree_fn callback_op, struct walk_stmt_info *wi) 46 { 47 gimple_stmt_iterator gsi; 48 49 for (gsi = gsi_start (*pseq); !gsi_end_p (gsi); ) 50 { 51 tree ret = walk_gimple_stmt (&gsi, callback_stmt, callback_op, wi); 52 if (ret) 53 { 54 /* If CALLBACK_STMT or CALLBACK_OP return a value, WI must exist 55 to hold it. */ 56 gcc_assert (wi); 57 wi->callback_result = ret; 58 59 return wi->removed_stmt ? NULL : gsi_stmt (gsi); 60 } 61 62 if (!wi->removed_stmt) 63 gsi_next (&gsi); 64 } 65 66 if (wi) 67 wi->callback_result = NULL_TREE; 68 69 return NULL; 70 } 71 72 73 /* Like walk_gimple_seq_mod, but ensure that the head of SEQ isn't 74 changed by the callbacks. */ 75 76 gimple * 77 walk_gimple_seq (gimple_seq seq, walk_stmt_fn callback_stmt, 78 walk_tree_fn callback_op, struct walk_stmt_info *wi) 79 { 80 gimple_seq seq2 = seq; 81 gimple *ret = walk_gimple_seq_mod (&seq2, callback_stmt, callback_op, wi); 82 gcc_assert (seq2 == seq); 83 return ret; 84 } 85 86 87 /* Helper function for walk_gimple_stmt. Walk operands of a GIMPLE_ASM. */ 88 89 static tree 90 walk_gimple_asm (gasm *stmt, walk_tree_fn callback_op, 91 struct walk_stmt_info *wi) 92 { 93 tree ret, op; 94 unsigned noutputs; 95 const char **oconstraints; 96 unsigned i, n; 97 const char *constraint; 98 bool allows_mem, allows_reg, is_inout; 99 100 noutputs = gimple_asm_noutputs (stmt); 101 oconstraints = (const char **) alloca ((noutputs) * sizeof (const char *)); 102 103 for (i = 0; i < noutputs; i++) 104 { 105 op = gimple_asm_output_op (stmt, i); 106 constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op))); 107 oconstraints[i] = constraint; 108 if (wi) 109 { 110 if (parse_output_constraint (&constraint, i, 0, 0, &allows_mem, 111 &allows_reg, &is_inout)) 112 wi->val_only = (allows_reg || !allows_mem); 113 } 114 if (wi) 115 wi->is_lhs = true; 116 ret = walk_tree (&TREE_VALUE (op), callback_op, wi, NULL); 117 if (ret) 118 return ret; 119 } 120 121 n = gimple_asm_ninputs (stmt); 122 for (i = 0; i < n; i++) 123 { 124 op = gimple_asm_input_op (stmt, i); 125 constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op))); 126 127 if (wi) 128 { 129 if (parse_input_constraint (&constraint, 0, 0, noutputs, 0, 130 oconstraints, &allows_mem, &allows_reg)) 131 { 132 wi->val_only = (allows_reg || !allows_mem); 133 /* Although input "m" is not really a LHS, we need a lvalue. */ 134 wi->is_lhs = !wi->val_only; 135 } 136 } 137 ret = walk_tree (&TREE_VALUE (op), callback_op, wi, NULL); 138 if (ret) 139 return ret; 140 } 141 142 if (wi) 143 { 144 wi->is_lhs = false; 145 wi->val_only = true; 146 } 147 148 n = gimple_asm_nlabels (stmt); 149 for (i = 0; i < n; i++) 150 { 151 op = gimple_asm_label_op (stmt, i); 152 ret = walk_tree (&TREE_VALUE (op), callback_op, wi, NULL); 153 if (ret) 154 return ret; 155 } 156 157 return NULL_TREE; 158 } 159 160 161 /* Helper function of WALK_GIMPLE_STMT. Walk every tree operand in 162 STMT. CALLBACK_OP and WI are as in WALK_GIMPLE_STMT. 163 164 CALLBACK_OP is called on each operand of STMT via walk_tree. 165 Additional parameters to walk_tree must be stored in WI. For each operand 166 OP, walk_tree is called as: 167 168 walk_tree (&OP, CALLBACK_OP, WI, WI->PSET) 169 170 If CALLBACK_OP returns non-NULL for an operand, the remaining 171 operands are not scanned. 172 173 The return value is that returned by the last call to walk_tree, or 174 NULL_TREE if no CALLBACK_OP is specified. */ 175 176 tree 177 walk_gimple_op (gimple *stmt, walk_tree_fn callback_op, 178 struct walk_stmt_info *wi) 179 { 180 hash_set<tree> *pset = (wi) ? wi->pset : NULL; 181 unsigned i; 182 tree ret = NULL_TREE; 183 184 if (wi) 185 wi->stmt = stmt; 186 187 switch (gimple_code (stmt)) 188 { 189 case GIMPLE_ASSIGN: 190 /* Walk the RHS operands. If the LHS is of a non-renamable type or 191 is a register variable, we may use a COMPONENT_REF on the RHS. */ 192 if (wi) 193 { 194 tree lhs = gimple_assign_lhs (stmt); 195 wi->val_only 196 = (is_gimple_reg_type (TREE_TYPE (lhs)) && !is_gimple_reg (lhs)) 197 || gimple_assign_rhs_class (stmt) != GIMPLE_SINGLE_RHS; 198 } 199 200 for (i = 1; i < gimple_num_ops (stmt); i++) 201 { 202 ret = walk_tree (gimple_op_ptr (stmt, i), callback_op, wi, 203 pset); 204 if (ret) 205 return ret; 206 } 207 208 /* Walk the LHS. If the RHS is appropriate for a memory, we 209 may use a COMPONENT_REF on the LHS. */ 210 if (wi) 211 { 212 /* If the RHS is of a non-renamable type or is a register variable, 213 we may use a COMPONENT_REF on the LHS. */ 214 tree rhs1 = gimple_assign_rhs1 (stmt); 215 wi->val_only 216 = (is_gimple_reg_type (TREE_TYPE (rhs1)) && !is_gimple_reg (rhs1)) 217 || gimple_assign_rhs_class (stmt) != GIMPLE_SINGLE_RHS; 218 wi->is_lhs = true; 219 } 220 221 ret = walk_tree (gimple_op_ptr (stmt, 0), callback_op, wi, pset); 222 if (ret) 223 return ret; 224 225 if (wi) 226 { 227 wi->val_only = true; 228 wi->is_lhs = false; 229 } 230 break; 231 232 case GIMPLE_CALL: 233 if (wi) 234 { 235 wi->is_lhs = false; 236 wi->val_only = true; 237 } 238 239 ret = walk_tree (gimple_call_chain_ptr (as_a <gcall *> (stmt)), 240 callback_op, wi, pset); 241 if (ret) 242 return ret; 243 244 ret = walk_tree (gimple_call_fn_ptr (stmt), callback_op, wi, pset); 245 if (ret) 246 return ret; 247 248 for (i = 0; i < gimple_call_num_args (stmt); i++) 249 { 250 if (wi) 251 wi->val_only 252 = is_gimple_reg_type (TREE_TYPE (gimple_call_arg (stmt, i))); 253 ret = walk_tree (gimple_call_arg_ptr (stmt, i), callback_op, wi, 254 pset); 255 if (ret) 256 return ret; 257 } 258 259 if (gimple_call_lhs (stmt)) 260 { 261 if (wi) 262 { 263 wi->is_lhs = true; 264 wi->val_only 265 = is_gimple_reg_type (TREE_TYPE (gimple_call_lhs (stmt))); 266 } 267 268 ret = walk_tree (gimple_call_lhs_ptr (stmt), callback_op, wi, pset); 269 if (ret) 270 return ret; 271 } 272 273 if (wi) 274 { 275 wi->is_lhs = false; 276 wi->val_only = true; 277 } 278 break; 279 280 case GIMPLE_CATCH: 281 ret = walk_tree (gimple_catch_types_ptr (as_a <gcatch *> (stmt)), 282 callback_op, wi, pset); 283 if (ret) 284 return ret; 285 break; 286 287 case GIMPLE_EH_FILTER: 288 ret = walk_tree (gimple_eh_filter_types_ptr (stmt), callback_op, wi, 289 pset); 290 if (ret) 291 return ret; 292 break; 293 294 case GIMPLE_ASM: 295 ret = walk_gimple_asm (as_a <gasm *> (stmt), callback_op, wi); 296 if (ret) 297 return ret; 298 break; 299 300 case GIMPLE_OMP_CONTINUE: 301 { 302 gomp_continue *cont_stmt = as_a <gomp_continue *> (stmt); 303 ret = walk_tree (gimple_omp_continue_control_def_ptr (cont_stmt), 304 callback_op, wi, pset); 305 if (ret) 306 return ret; 307 308 ret = walk_tree (gimple_omp_continue_control_use_ptr (cont_stmt), 309 callback_op, wi, pset); 310 if (ret) 311 return ret; 312 } 313 break; 314 315 case GIMPLE_OMP_CRITICAL: 316 { 317 gomp_critical *omp_stmt = as_a <gomp_critical *> (stmt); 318 ret = walk_tree (gimple_omp_critical_name_ptr (omp_stmt), 319 callback_op, wi, pset); 320 if (ret) 321 return ret; 322 ret = walk_tree (gimple_omp_critical_clauses_ptr (omp_stmt), 323 callback_op, wi, pset); 324 if (ret) 325 return ret; 326 } 327 break; 328 329 case GIMPLE_OMP_ORDERED: 330 { 331 gomp_ordered *omp_stmt = as_a <gomp_ordered *> (stmt); 332 ret = walk_tree (gimple_omp_ordered_clauses_ptr (omp_stmt), 333 callback_op, wi, pset); 334 if (ret) 335 return ret; 336 } 337 break; 338 339 case GIMPLE_OMP_SCAN: 340 { 341 gomp_scan *scan_stmt = as_a <gomp_scan *> (stmt); 342 ret = walk_tree (gimple_omp_scan_clauses_ptr (scan_stmt), 343 callback_op, wi, pset); 344 if (ret) 345 return ret; 346 } 347 break; 348 349 case GIMPLE_OMP_FOR: 350 ret = walk_tree (gimple_omp_for_clauses_ptr (stmt), callback_op, wi, 351 pset); 352 if (ret) 353 return ret; 354 for (i = 0; i < gimple_omp_for_collapse (stmt); i++) 355 { 356 ret = walk_tree (gimple_omp_for_index_ptr (stmt, i), callback_op, 357 wi, pset); 358 if (ret) 359 return ret; 360 ret = walk_tree (gimple_omp_for_initial_ptr (stmt, i), callback_op, 361 wi, pset); 362 if (ret) 363 return ret; 364 ret = walk_tree (gimple_omp_for_final_ptr (stmt, i), callback_op, 365 wi, pset); 366 if (ret) 367 return ret; 368 ret = walk_tree (gimple_omp_for_incr_ptr (stmt, i), callback_op, 369 wi, pset); 370 if (ret) 371 return ret; 372 } 373 break; 374 375 case GIMPLE_OMP_PARALLEL: 376 { 377 gomp_parallel *omp_par_stmt = as_a <gomp_parallel *> (stmt); 378 ret = walk_tree (gimple_omp_parallel_clauses_ptr (omp_par_stmt), 379 callback_op, wi, pset); 380 if (ret) 381 return ret; 382 ret = walk_tree (gimple_omp_parallel_child_fn_ptr (omp_par_stmt), 383 callback_op, wi, pset); 384 if (ret) 385 return ret; 386 ret = walk_tree (gimple_omp_parallel_data_arg_ptr (omp_par_stmt), 387 callback_op, wi, pset); 388 if (ret) 389 return ret; 390 } 391 break; 392 393 case GIMPLE_OMP_TASK: 394 ret = walk_tree (gimple_omp_task_clauses_ptr (stmt), callback_op, 395 wi, pset); 396 if (ret) 397 return ret; 398 ret = walk_tree (gimple_omp_task_child_fn_ptr (stmt), callback_op, 399 wi, pset); 400 if (ret) 401 return ret; 402 ret = walk_tree (gimple_omp_task_data_arg_ptr (stmt), callback_op, 403 wi, pset); 404 if (ret) 405 return ret; 406 ret = walk_tree (gimple_omp_task_copy_fn_ptr (stmt), callback_op, 407 wi, pset); 408 if (ret) 409 return ret; 410 ret = walk_tree (gimple_omp_task_arg_size_ptr (stmt), callback_op, 411 wi, pset); 412 if (ret) 413 return ret; 414 ret = walk_tree (gimple_omp_task_arg_align_ptr (stmt), callback_op, 415 wi, pset); 416 if (ret) 417 return ret; 418 break; 419 420 case GIMPLE_OMP_SECTIONS: 421 ret = walk_tree (gimple_omp_sections_clauses_ptr (stmt), callback_op, 422 wi, pset); 423 if (ret) 424 return ret; 425 ret = walk_tree (gimple_omp_sections_control_ptr (stmt), callback_op, 426 wi, pset); 427 if (ret) 428 return ret; 429 430 break; 431 432 case GIMPLE_OMP_SINGLE: 433 ret = walk_tree (gimple_omp_single_clauses_ptr (stmt), callback_op, wi, 434 pset); 435 if (ret) 436 return ret; 437 break; 438 439 case GIMPLE_OMP_TARGET: 440 { 441 gomp_target *omp_stmt = as_a <gomp_target *> (stmt); 442 ret = walk_tree (gimple_omp_target_clauses_ptr (omp_stmt), 443 callback_op, wi, pset); 444 if (ret) 445 return ret; 446 ret = walk_tree (gimple_omp_target_child_fn_ptr (omp_stmt), 447 callback_op, wi, pset); 448 if (ret) 449 return ret; 450 ret = walk_tree (gimple_omp_target_data_arg_ptr (omp_stmt), 451 callback_op, wi, pset); 452 if (ret) 453 return ret; 454 } 455 break; 456 457 case GIMPLE_OMP_TEAMS: 458 ret = walk_tree (gimple_omp_teams_clauses_ptr (stmt), callback_op, wi, 459 pset); 460 if (ret) 461 return ret; 462 break; 463 464 case GIMPLE_OMP_ATOMIC_LOAD: 465 { 466 gomp_atomic_load *omp_stmt = as_a <gomp_atomic_load *> (stmt); 467 ret = walk_tree (gimple_omp_atomic_load_lhs_ptr (omp_stmt), 468 callback_op, wi, pset); 469 if (ret) 470 return ret; 471 ret = walk_tree (gimple_omp_atomic_load_rhs_ptr (omp_stmt), 472 callback_op, wi, pset); 473 if (ret) 474 return ret; 475 } 476 break; 477 478 case GIMPLE_OMP_ATOMIC_STORE: 479 { 480 gomp_atomic_store *omp_stmt = as_a <gomp_atomic_store *> (stmt); 481 ret = walk_tree (gimple_omp_atomic_store_val_ptr (omp_stmt), 482 callback_op, wi, pset); 483 if (ret) 484 return ret; 485 } 486 break; 487 488 case GIMPLE_TRANSACTION: 489 { 490 gtransaction *txn = as_a <gtransaction *> (stmt); 491 492 ret = walk_tree (gimple_transaction_label_norm_ptr (txn), 493 callback_op, wi, pset); 494 if (ret) 495 return ret; 496 ret = walk_tree (gimple_transaction_label_uninst_ptr (txn), 497 callback_op, wi, pset); 498 if (ret) 499 return ret; 500 ret = walk_tree (gimple_transaction_label_over_ptr (txn), 501 callback_op, wi, pset); 502 if (ret) 503 return ret; 504 } 505 break; 506 507 case GIMPLE_OMP_RETURN: 508 ret = walk_tree (gimple_omp_return_lhs_ptr (stmt), callback_op, wi, 509 pset); 510 if (ret) 511 return ret; 512 break; 513 514 /* Tuples that do not have operands. */ 515 case GIMPLE_NOP: 516 case GIMPLE_RESX: 517 case GIMPLE_PREDICT: 518 break; 519 520 default: 521 { 522 enum gimple_statement_structure_enum gss; 523 gss = gimple_statement_structure (stmt); 524 if (gss == GSS_WITH_OPS || gss == GSS_WITH_MEM_OPS) 525 for (i = 0; i < gimple_num_ops (stmt); i++) 526 { 527 ret = walk_tree (gimple_op_ptr (stmt, i), callback_op, wi, pset); 528 if (ret) 529 return ret; 530 } 531 } 532 break; 533 } 534 535 return NULL_TREE; 536 } 537 538 539 /* Walk the current statement in GSI (optionally using traversal state 540 stored in WI). If WI is NULL, no state is kept during traversal. 541 The callback CALLBACK_STMT is called. If CALLBACK_STMT indicates 542 that it has handled all the operands of the statement, its return 543 value is returned. Otherwise, the return value from CALLBACK_STMT 544 is discarded and its operands are scanned. 545 546 If CALLBACK_STMT is NULL or it didn't handle the operands, 547 CALLBACK_OP is called on each operand of the statement via 548 walk_gimple_op. If walk_gimple_op returns non-NULL for any 549 operand, the remaining operands are not scanned. In this case, the 550 return value from CALLBACK_OP is returned. 551 552 In any other case, NULL_TREE is returned. */ 553 554 tree 555 walk_gimple_stmt (gimple_stmt_iterator *gsi, walk_stmt_fn callback_stmt, 556 walk_tree_fn callback_op, struct walk_stmt_info *wi) 557 { 558 gimple *ret; 559 tree tree_ret; 560 gimple *stmt = gsi_stmt (*gsi); 561 562 if (wi) 563 { 564 wi->gsi = *gsi; 565 wi->removed_stmt = false; 566 567 if (wi->want_locations && gimple_has_location (stmt)) 568 input_location = gimple_location (stmt); 569 } 570 571 ret = NULL; 572 573 /* Invoke the statement callback. Return if the callback handled 574 all of STMT operands by itself. */ 575 if (callback_stmt) 576 { 577 bool handled_ops = false; 578 tree_ret = callback_stmt (gsi, &handled_ops, wi); 579 if (handled_ops) 580 return tree_ret; 581 582 /* If CALLBACK_STMT did not handle operands, it should not have 583 a value to return. */ 584 gcc_assert (tree_ret == NULL); 585 586 if (wi && wi->removed_stmt) 587 return NULL; 588 589 /* Re-read stmt in case the callback changed it. */ 590 stmt = gsi_stmt (*gsi); 591 } 592 593 /* If CALLBACK_OP is defined, invoke it on every operand of STMT. */ 594 if (callback_op) 595 { 596 tree_ret = walk_gimple_op (stmt, callback_op, wi); 597 if (tree_ret) 598 return tree_ret; 599 } 600 601 /* If STMT can have statements inside (e.g. GIMPLE_BIND), walk them. */ 602 switch (gimple_code (stmt)) 603 { 604 case GIMPLE_BIND: 605 ret = walk_gimple_seq_mod (gimple_bind_body_ptr (as_a <gbind *> (stmt)), 606 callback_stmt, callback_op, wi); 607 if (ret) 608 return wi->callback_result; 609 break; 610 611 case GIMPLE_CATCH: 612 ret = walk_gimple_seq_mod (gimple_catch_handler_ptr ( 613 as_a <gcatch *> (stmt)), 614 callback_stmt, callback_op, wi); 615 if (ret) 616 return wi->callback_result; 617 break; 618 619 case GIMPLE_EH_FILTER: 620 ret = walk_gimple_seq_mod (gimple_eh_filter_failure_ptr (stmt), callback_stmt, 621 callback_op, wi); 622 if (ret) 623 return wi->callback_result; 624 break; 625 626 case GIMPLE_EH_ELSE: 627 { 628 geh_else *eh_else_stmt = as_a <geh_else *> (stmt); 629 ret = walk_gimple_seq_mod (gimple_eh_else_n_body_ptr (eh_else_stmt), 630 callback_stmt, callback_op, wi); 631 if (ret) 632 return wi->callback_result; 633 ret = walk_gimple_seq_mod (gimple_eh_else_e_body_ptr (eh_else_stmt), 634 callback_stmt, callback_op, wi); 635 if (ret) 636 return wi->callback_result; 637 } 638 break; 639 640 case GIMPLE_TRY: 641 ret = walk_gimple_seq_mod (gimple_try_eval_ptr (stmt), callback_stmt, callback_op, 642 wi); 643 if (ret) 644 return wi->callback_result; 645 646 ret = walk_gimple_seq_mod (gimple_try_cleanup_ptr (stmt), callback_stmt, 647 callback_op, wi); 648 if (ret) 649 return wi->callback_result; 650 break; 651 652 case GIMPLE_OMP_FOR: 653 ret = walk_gimple_seq_mod (gimple_omp_for_pre_body_ptr (stmt), callback_stmt, 654 callback_op, wi); 655 if (ret) 656 return wi->callback_result; 657 658 /* FALL THROUGH. */ 659 case GIMPLE_OMP_CRITICAL: 660 case GIMPLE_OMP_MASTER: 661 case GIMPLE_OMP_TASKGROUP: 662 case GIMPLE_OMP_ORDERED: 663 case GIMPLE_OMP_SCAN: 664 case GIMPLE_OMP_SECTION: 665 case GIMPLE_OMP_PARALLEL: 666 case GIMPLE_OMP_TASK: 667 case GIMPLE_OMP_SECTIONS: 668 case GIMPLE_OMP_SINGLE: 669 case GIMPLE_OMP_TARGET: 670 case GIMPLE_OMP_TEAMS: 671 case GIMPLE_OMP_GRID_BODY: 672 ret = walk_gimple_seq_mod (gimple_omp_body_ptr (stmt), callback_stmt, 673 callback_op, wi); 674 if (ret) 675 return wi->callback_result; 676 break; 677 678 case GIMPLE_WITH_CLEANUP_EXPR: 679 ret = walk_gimple_seq_mod (gimple_wce_cleanup_ptr (stmt), callback_stmt, 680 callback_op, wi); 681 if (ret) 682 return wi->callback_result; 683 break; 684 685 case GIMPLE_TRANSACTION: 686 ret = walk_gimple_seq_mod (gimple_transaction_body_ptr ( 687 as_a <gtransaction *> (stmt)), 688 callback_stmt, callback_op, wi); 689 if (ret) 690 return wi->callback_result; 691 break; 692 693 default: 694 gcc_assert (!gimple_has_substatements (stmt)); 695 break; 696 } 697 698 return NULL; 699 } 700 701 /* From a tree operand OP return the base of a load or store operation 702 or NULL_TREE if OP is not a load or a store. */ 703 704 static tree 705 get_base_loadstore (tree op) 706 { 707 while (handled_component_p (op)) 708 op = TREE_OPERAND (op, 0); 709 if (DECL_P (op) 710 || INDIRECT_REF_P (op) 711 || TREE_CODE (op) == MEM_REF 712 || TREE_CODE (op) == TARGET_MEM_REF) 713 return op; 714 return NULL_TREE; 715 } 716 717 718 /* For the statement STMT call the callbacks VISIT_LOAD, VISIT_STORE and 719 VISIT_ADDR if non-NULL on loads, store and address-taken operands 720 passing the STMT, the base of the operand, the operand itself containing 721 the base and DATA to it. The base will be either a decl, an indirect 722 reference (including TARGET_MEM_REF) or the argument of an address 723 expression. 724 Returns the results of these callbacks or'ed. */ 725 726 bool 727 walk_stmt_load_store_addr_ops (gimple *stmt, void *data, 728 walk_stmt_load_store_addr_fn visit_load, 729 walk_stmt_load_store_addr_fn visit_store, 730 walk_stmt_load_store_addr_fn visit_addr) 731 { 732 bool ret = false; 733 unsigned i; 734 if (gimple_assign_single_p (stmt)) 735 { 736 tree lhs, rhs, arg; 737 if (visit_store) 738 { 739 arg = gimple_assign_lhs (stmt); 740 lhs = get_base_loadstore (arg); 741 if (lhs) 742 ret |= visit_store (stmt, lhs, arg, data); 743 } 744 arg = gimple_assign_rhs1 (stmt); 745 rhs = arg; 746 while (handled_component_p (rhs)) 747 rhs = TREE_OPERAND (rhs, 0); 748 if (visit_addr) 749 { 750 if (TREE_CODE (rhs) == ADDR_EXPR) 751 ret |= visit_addr (stmt, TREE_OPERAND (rhs, 0), arg, data); 752 else if (TREE_CODE (rhs) == TARGET_MEM_REF 753 && TREE_CODE (TMR_BASE (rhs)) == ADDR_EXPR) 754 ret |= visit_addr (stmt, TREE_OPERAND (TMR_BASE (rhs), 0), arg, 755 data); 756 else if (TREE_CODE (rhs) == OBJ_TYPE_REF 757 && TREE_CODE (OBJ_TYPE_REF_OBJECT (rhs)) == ADDR_EXPR) 758 ret |= visit_addr (stmt, TREE_OPERAND (OBJ_TYPE_REF_OBJECT (rhs), 759 0), arg, data); 760 else if (TREE_CODE (rhs) == CONSTRUCTOR) 761 { 762 unsigned int ix; 763 tree val; 764 765 FOR_EACH_CONSTRUCTOR_VALUE (CONSTRUCTOR_ELTS (rhs), ix, val) 766 if (TREE_CODE (val) == ADDR_EXPR) 767 ret |= visit_addr (stmt, TREE_OPERAND (val, 0), arg, data); 768 else if (TREE_CODE (val) == OBJ_TYPE_REF 769 && TREE_CODE (OBJ_TYPE_REF_OBJECT (val)) == ADDR_EXPR) 770 ret |= visit_addr (stmt, 771 TREE_OPERAND (OBJ_TYPE_REF_OBJECT (val), 772 0), arg, data); 773 } 774 lhs = gimple_assign_lhs (stmt); 775 if (TREE_CODE (lhs) == TARGET_MEM_REF 776 && TREE_CODE (TMR_BASE (lhs)) == ADDR_EXPR) 777 ret |= visit_addr (stmt, TREE_OPERAND (TMR_BASE (lhs), 0), lhs, data); 778 } 779 if (visit_load) 780 { 781 rhs = get_base_loadstore (rhs); 782 if (rhs) 783 ret |= visit_load (stmt, rhs, arg, data); 784 } 785 } 786 else if (visit_addr 787 && (is_gimple_assign (stmt) 788 || gimple_code (stmt) == GIMPLE_COND)) 789 { 790 for (i = 0; i < gimple_num_ops (stmt); ++i) 791 { 792 tree op = gimple_op (stmt, i); 793 if (op == NULL_TREE) 794 ; 795 else if (TREE_CODE (op) == ADDR_EXPR) 796 ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data); 797 /* COND_EXPR and VCOND_EXPR rhs1 argument is a comparison 798 tree with two operands. */ 799 else if (i == 1 && COMPARISON_CLASS_P (op)) 800 { 801 if (TREE_CODE (TREE_OPERAND (op, 0)) == ADDR_EXPR) 802 ret |= visit_addr (stmt, TREE_OPERAND (TREE_OPERAND (op, 0), 803 0), op, data); 804 if (TREE_CODE (TREE_OPERAND (op, 1)) == ADDR_EXPR) 805 ret |= visit_addr (stmt, TREE_OPERAND (TREE_OPERAND (op, 1), 806 0), op, data); 807 } 808 } 809 } 810 else if (gcall *call_stmt = dyn_cast <gcall *> (stmt)) 811 { 812 if (visit_store) 813 { 814 tree arg = gimple_call_lhs (call_stmt); 815 if (arg) 816 { 817 tree lhs = get_base_loadstore (arg); 818 if (lhs) 819 ret |= visit_store (stmt, lhs, arg, data); 820 } 821 } 822 if (visit_load || visit_addr) 823 for (i = 0; i < gimple_call_num_args (call_stmt); ++i) 824 { 825 tree arg = gimple_call_arg (call_stmt, i); 826 if (visit_addr 827 && TREE_CODE (arg) == ADDR_EXPR) 828 ret |= visit_addr (stmt, TREE_OPERAND (arg, 0), arg, data); 829 else if (visit_load) 830 { 831 tree rhs = get_base_loadstore (arg); 832 if (rhs) 833 ret |= visit_load (stmt, rhs, arg, data); 834 } 835 } 836 if (visit_addr 837 && gimple_call_chain (call_stmt) 838 && TREE_CODE (gimple_call_chain (call_stmt)) == ADDR_EXPR) 839 ret |= visit_addr (stmt, TREE_OPERAND (gimple_call_chain (call_stmt), 0), 840 gimple_call_chain (call_stmt), data); 841 if (visit_addr 842 && gimple_call_return_slot_opt_p (call_stmt) 843 && gimple_call_lhs (call_stmt) != NULL_TREE 844 && TREE_ADDRESSABLE (TREE_TYPE (gimple_call_lhs (call_stmt)))) 845 ret |= visit_addr (stmt, gimple_call_lhs (call_stmt), 846 gimple_call_lhs (call_stmt), data); 847 } 848 else if (gasm *asm_stmt = dyn_cast <gasm *> (stmt)) 849 { 850 unsigned noutputs; 851 const char *constraint; 852 const char **oconstraints; 853 bool allows_mem, allows_reg, is_inout; 854 noutputs = gimple_asm_noutputs (asm_stmt); 855 oconstraints = XALLOCAVEC (const char *, noutputs); 856 if (visit_store || visit_addr) 857 for (i = 0; i < gimple_asm_noutputs (asm_stmt); ++i) 858 { 859 tree link = gimple_asm_output_op (asm_stmt, i); 860 tree op = get_base_loadstore (TREE_VALUE (link)); 861 if (op && visit_store) 862 ret |= visit_store (stmt, op, TREE_VALUE (link), data); 863 if (visit_addr) 864 { 865 constraint = TREE_STRING_POINTER 866 (TREE_VALUE (TREE_PURPOSE (link))); 867 oconstraints[i] = constraint; 868 parse_output_constraint (&constraint, i, 0, 0, &allows_mem, 869 &allows_reg, &is_inout); 870 if (op && !allows_reg && allows_mem) 871 ret |= visit_addr (stmt, op, TREE_VALUE (link), data); 872 } 873 } 874 if (visit_load || visit_addr) 875 for (i = 0; i < gimple_asm_ninputs (asm_stmt); ++i) 876 { 877 tree link = gimple_asm_input_op (asm_stmt, i); 878 tree op = TREE_VALUE (link); 879 if (visit_addr 880 && TREE_CODE (op) == ADDR_EXPR) 881 ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data); 882 else if (visit_load || visit_addr) 883 { 884 op = get_base_loadstore (op); 885 if (op) 886 { 887 if (visit_load) 888 ret |= visit_load (stmt, op, TREE_VALUE (link), data); 889 if (visit_addr) 890 { 891 constraint = TREE_STRING_POINTER 892 (TREE_VALUE (TREE_PURPOSE (link))); 893 parse_input_constraint (&constraint, 0, 0, noutputs, 894 0, oconstraints, 895 &allows_mem, &allows_reg); 896 if (!allows_reg && allows_mem) 897 ret |= visit_addr (stmt, op, TREE_VALUE (link), 898 data); 899 } 900 } 901 } 902 } 903 } 904 else if (greturn *return_stmt = dyn_cast <greturn *> (stmt)) 905 { 906 tree op = gimple_return_retval (return_stmt); 907 if (op) 908 { 909 if (visit_addr 910 && TREE_CODE (op) == ADDR_EXPR) 911 ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data); 912 else if (visit_load) 913 { 914 tree base = get_base_loadstore (op); 915 if (base) 916 ret |= visit_load (stmt, base, op, data); 917 } 918 } 919 } 920 else if (visit_addr 921 && gimple_code (stmt) == GIMPLE_PHI) 922 { 923 for (i = 0; i < gimple_phi_num_args (stmt); ++i) 924 { 925 tree op = gimple_phi_arg_def (stmt, i); 926 if (TREE_CODE (op) == ADDR_EXPR) 927 ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data); 928 } 929 } 930 else if (visit_addr 931 && gimple_code (stmt) == GIMPLE_GOTO) 932 { 933 tree op = gimple_goto_dest (stmt); 934 if (TREE_CODE (op) == ADDR_EXPR) 935 ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data); 936 } 937 938 return ret; 939 } 940 941 /* Like walk_stmt_load_store_addr_ops but with NULL visit_addr. IPA-CP 942 should make a faster clone for this case. */ 943 944 bool 945 walk_stmt_load_store_ops (gimple *stmt, void *data, 946 walk_stmt_load_store_addr_fn visit_load, 947 walk_stmt_load_store_addr_fn visit_store) 948 { 949 return walk_stmt_load_store_addr_ops (stmt, data, 950 visit_load, visit_store, NULL); 951 } 952