xref: /dflybsd-src/contrib/gcc-4.7/gcc/tree-ssa-threadedge.c (revision 81fc95a5293ee307c688a350a3feb4734aaddbb4)
1e4b17023SJohn Marino /* SSA Jump Threading
2e4b17023SJohn Marino    Copyright (C) 2005, 2006, 2007, 2008, 2009, 2010, 2011
3e4b17023SJohn Marino    Free Software Foundation, Inc.
4e4b17023SJohn Marino    Contributed by Jeff Law  <law@redhat.com>
5e4b17023SJohn Marino 
6e4b17023SJohn Marino This file is part of GCC.
7e4b17023SJohn Marino 
8e4b17023SJohn Marino GCC is free software; you can redistribute it and/or modify
9e4b17023SJohn Marino it under the terms of the GNU General Public License as published by
10e4b17023SJohn Marino the Free Software Foundation; either version 3, or (at your option)
11e4b17023SJohn Marino any later version.
12e4b17023SJohn Marino 
13e4b17023SJohn Marino GCC is distributed in the hope that it will be useful,
14e4b17023SJohn Marino but WITHOUT ANY WARRANTY; without even the implied warranty of
15e4b17023SJohn Marino MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
16e4b17023SJohn Marino GNU General Public License for more details.
17e4b17023SJohn Marino 
18e4b17023SJohn Marino You should have received a copy of the GNU General Public License
19e4b17023SJohn Marino along with GCC; see the file COPYING3.  If not see
20e4b17023SJohn Marino <http://www.gnu.org/licenses/>.  */
21e4b17023SJohn Marino 
22e4b17023SJohn Marino #include "config.h"
23e4b17023SJohn Marino #include "system.h"
24e4b17023SJohn Marino #include "coretypes.h"
25e4b17023SJohn Marino #include "tm.h"
26e4b17023SJohn Marino #include "tree.h"
27e4b17023SJohn Marino #include "flags.h"
28e4b17023SJohn Marino #include "tm_p.h"
29e4b17023SJohn Marino #include "basic-block.h"
30e4b17023SJohn Marino #include "cfgloop.h"
31e4b17023SJohn Marino #include "output.h"
32e4b17023SJohn Marino #include "function.h"
33e4b17023SJohn Marino #include "timevar.h"
34e4b17023SJohn Marino #include "tree-dump.h"
35e4b17023SJohn Marino #include "tree-flow.h"
36e4b17023SJohn Marino #include "tree-pass.h"
37e4b17023SJohn Marino #include "tree-ssa-propagate.h"
38e4b17023SJohn Marino #include "langhooks.h"
39e4b17023SJohn Marino #include "params.h"
40e4b17023SJohn Marino 
41e4b17023SJohn Marino /* To avoid code explosion due to jump threading, we limit the
42e4b17023SJohn Marino    number of statements we are going to copy.  This variable
43e4b17023SJohn Marino    holds the number of statements currently seen that we'll have
44e4b17023SJohn Marino    to copy as part of the jump threading process.  */
45e4b17023SJohn Marino static int stmt_count;
46e4b17023SJohn Marino 
47e4b17023SJohn Marino /* Array to record value-handles per SSA_NAME.  */
VEC(tree,heap)48e4b17023SJohn Marino VEC(tree,heap) *ssa_name_values;
49e4b17023SJohn Marino 
50e4b17023SJohn Marino /* Set the value for the SSA name NAME to VALUE.  */
51e4b17023SJohn Marino 
52e4b17023SJohn Marino void
53e4b17023SJohn Marino set_ssa_name_value (tree name, tree value)
54e4b17023SJohn Marino {
55e4b17023SJohn Marino   if (SSA_NAME_VERSION (name) >= VEC_length (tree, ssa_name_values))
56e4b17023SJohn Marino     VEC_safe_grow_cleared (tree, heap, ssa_name_values,
57e4b17023SJohn Marino 			   SSA_NAME_VERSION (name) + 1);
58e4b17023SJohn Marino   VEC_replace (tree, ssa_name_values, SSA_NAME_VERSION (name), value);
59e4b17023SJohn Marino }
60e4b17023SJohn Marino 
61e4b17023SJohn Marino /* Initialize the per SSA_NAME value-handles array.  Returns it.  */
62e4b17023SJohn Marino void
threadedge_initialize_values(void)63e4b17023SJohn Marino threadedge_initialize_values (void)
64e4b17023SJohn Marino {
65e4b17023SJohn Marino   gcc_assert (ssa_name_values == NULL);
66e4b17023SJohn Marino   ssa_name_values = VEC_alloc(tree, heap, num_ssa_names);
67e4b17023SJohn Marino }
68e4b17023SJohn Marino 
69e4b17023SJohn Marino /* Free the per SSA_NAME value-handle array.  */
70e4b17023SJohn Marino void
threadedge_finalize_values(void)71e4b17023SJohn Marino threadedge_finalize_values (void)
72e4b17023SJohn Marino {
73e4b17023SJohn Marino   VEC_free(tree, heap, ssa_name_values);
74e4b17023SJohn Marino }
75e4b17023SJohn Marino 
76e4b17023SJohn Marino /* Return TRUE if we may be able to thread an incoming edge into
77e4b17023SJohn Marino    BB to an outgoing edge from BB.  Return FALSE otherwise.  */
78e4b17023SJohn Marino 
79e4b17023SJohn Marino bool
potentially_threadable_block(basic_block bb)80e4b17023SJohn Marino potentially_threadable_block (basic_block bb)
81e4b17023SJohn Marino {
82e4b17023SJohn Marino   gimple_stmt_iterator gsi;
83e4b17023SJohn Marino 
84e4b17023SJohn Marino   /* If BB has a single successor or a single predecessor, then
85e4b17023SJohn Marino      there is no threading opportunity.  */
86e4b17023SJohn Marino   if (single_succ_p (bb) || single_pred_p (bb))
87e4b17023SJohn Marino     return false;
88e4b17023SJohn Marino 
89e4b17023SJohn Marino   /* If BB does not end with a conditional, switch or computed goto,
90e4b17023SJohn Marino      then there is no threading opportunity.  */
91e4b17023SJohn Marino   gsi = gsi_last_bb (bb);
92e4b17023SJohn Marino   if (gsi_end_p (gsi)
93e4b17023SJohn Marino       || ! gsi_stmt (gsi)
94e4b17023SJohn Marino       || (gimple_code (gsi_stmt (gsi)) != GIMPLE_COND
95e4b17023SJohn Marino 	  && gimple_code (gsi_stmt (gsi)) != GIMPLE_GOTO
96e4b17023SJohn Marino 	  && gimple_code (gsi_stmt (gsi)) != GIMPLE_SWITCH))
97e4b17023SJohn Marino     return false;
98e4b17023SJohn Marino 
99e4b17023SJohn Marino   return true;
100e4b17023SJohn Marino }
101e4b17023SJohn Marino 
102e4b17023SJohn Marino /* Return the LHS of any ASSERT_EXPR where OP appears as the first
103e4b17023SJohn Marino    argument to the ASSERT_EXPR and in which the ASSERT_EXPR dominates
104e4b17023SJohn Marino    BB.  If no such ASSERT_EXPR is found, return OP.  */
105e4b17023SJohn Marino 
106e4b17023SJohn Marino static tree
lhs_of_dominating_assert(tree op,basic_block bb,gimple stmt)107e4b17023SJohn Marino lhs_of_dominating_assert (tree op, basic_block bb, gimple stmt)
108e4b17023SJohn Marino {
109e4b17023SJohn Marino   imm_use_iterator imm_iter;
110e4b17023SJohn Marino   gimple use_stmt;
111e4b17023SJohn Marino   use_operand_p use_p;
112e4b17023SJohn Marino 
113e4b17023SJohn Marino   FOR_EACH_IMM_USE_FAST (use_p, imm_iter, op)
114e4b17023SJohn Marino     {
115e4b17023SJohn Marino       use_stmt = USE_STMT (use_p);
116e4b17023SJohn Marino       if (use_stmt != stmt
117e4b17023SJohn Marino           && gimple_assign_single_p (use_stmt)
118e4b17023SJohn Marino           && TREE_CODE (gimple_assign_rhs1 (use_stmt)) == ASSERT_EXPR
119e4b17023SJohn Marino           && TREE_OPERAND (gimple_assign_rhs1 (use_stmt), 0) == op
120e4b17023SJohn Marino 	  && dominated_by_p (CDI_DOMINATORS, bb, gimple_bb (use_stmt)))
121e4b17023SJohn Marino 	{
122e4b17023SJohn Marino 	  return gimple_assign_lhs (use_stmt);
123e4b17023SJohn Marino 	}
124e4b17023SJohn Marino     }
125e4b17023SJohn Marino   return op;
126e4b17023SJohn Marino }
127e4b17023SJohn Marino 
128e4b17023SJohn Marino /* We record temporary equivalences created by PHI nodes or
129e4b17023SJohn Marino    statements within the target block.  Doing so allows us to
130e4b17023SJohn Marino    identify more jump threading opportunities, even in blocks
131e4b17023SJohn Marino    with side effects.
132e4b17023SJohn Marino 
133e4b17023SJohn Marino    We keep track of those temporary equivalences in a stack
134e4b17023SJohn Marino    structure so that we can unwind them when we're done processing
135e4b17023SJohn Marino    a particular edge.  This routine handles unwinding the data
136e4b17023SJohn Marino    structures.  */
137e4b17023SJohn Marino 
138e4b17023SJohn Marino static void
remove_temporary_equivalences(VEC (tree,heap)** stack)139e4b17023SJohn Marino remove_temporary_equivalences (VEC(tree, heap) **stack)
140e4b17023SJohn Marino {
141e4b17023SJohn Marino   while (VEC_length (tree, *stack) > 0)
142e4b17023SJohn Marino     {
143e4b17023SJohn Marino       tree prev_value, dest;
144e4b17023SJohn Marino 
145e4b17023SJohn Marino       dest = VEC_pop (tree, *stack);
146e4b17023SJohn Marino 
147e4b17023SJohn Marino       /* A NULL value indicates we should stop unwinding, otherwise
148e4b17023SJohn Marino 	 pop off the next entry as they're recorded in pairs.  */
149e4b17023SJohn Marino       if (dest == NULL)
150e4b17023SJohn Marino 	break;
151e4b17023SJohn Marino 
152e4b17023SJohn Marino       prev_value = VEC_pop (tree, *stack);
153e4b17023SJohn Marino       set_ssa_name_value (dest, prev_value);
154e4b17023SJohn Marino     }
155e4b17023SJohn Marino }
156e4b17023SJohn Marino 
157e4b17023SJohn Marino /* Record a temporary equivalence, saving enough information so that
158e4b17023SJohn Marino    we can restore the state of recorded equivalences when we're
159e4b17023SJohn Marino    done processing the current edge.  */
160e4b17023SJohn Marino 
161e4b17023SJohn Marino static void
record_temporary_equivalence(tree x,tree y,VEC (tree,heap)** stack)162e4b17023SJohn Marino record_temporary_equivalence (tree x, tree y, VEC(tree, heap) **stack)
163e4b17023SJohn Marino {
164e4b17023SJohn Marino   tree prev_x = SSA_NAME_VALUE (x);
165e4b17023SJohn Marino 
166e4b17023SJohn Marino   if (TREE_CODE (y) == SSA_NAME)
167e4b17023SJohn Marino     {
168e4b17023SJohn Marino       tree tmp = SSA_NAME_VALUE (y);
169e4b17023SJohn Marino       y = tmp ? tmp : y;
170e4b17023SJohn Marino     }
171e4b17023SJohn Marino 
172e4b17023SJohn Marino   set_ssa_name_value (x, y);
173e4b17023SJohn Marino   VEC_reserve (tree, heap, *stack, 2);
174e4b17023SJohn Marino   VEC_quick_push (tree, *stack, prev_x);
175e4b17023SJohn Marino   VEC_quick_push (tree, *stack, x);
176e4b17023SJohn Marino }
177e4b17023SJohn Marino 
178e4b17023SJohn Marino /* Record temporary equivalences created by PHIs at the target of the
179e4b17023SJohn Marino    edge E.  Record unwind information for the equivalences onto STACK.
180e4b17023SJohn Marino 
181e4b17023SJohn Marino    If a PHI which prevents threading is encountered, then return FALSE
182e4b17023SJohn Marino    indicating we should not thread this edge, else return TRUE.  */
183e4b17023SJohn Marino 
184e4b17023SJohn Marino static bool
record_temporary_equivalences_from_phis(edge e,VEC (tree,heap)** stack)185e4b17023SJohn Marino record_temporary_equivalences_from_phis (edge e, VEC(tree, heap) **stack)
186e4b17023SJohn Marino {
187e4b17023SJohn Marino   gimple_stmt_iterator gsi;
188e4b17023SJohn Marino 
189e4b17023SJohn Marino   /* Each PHI creates a temporary equivalence, record them.
190e4b17023SJohn Marino      These are context sensitive equivalences and will be removed
191e4b17023SJohn Marino      later.  */
192e4b17023SJohn Marino   for (gsi = gsi_start_phis (e->dest); !gsi_end_p (gsi); gsi_next (&gsi))
193e4b17023SJohn Marino     {
194e4b17023SJohn Marino       gimple phi = gsi_stmt (gsi);
195e4b17023SJohn Marino       tree src = PHI_ARG_DEF_FROM_EDGE (phi, e);
196e4b17023SJohn Marino       tree dst = gimple_phi_result (phi);
197e4b17023SJohn Marino 
198e4b17023SJohn Marino       /* If the desired argument is not the same as this PHI's result
199e4b17023SJohn Marino 	 and it is set by a PHI in E->dest, then we can not thread
200e4b17023SJohn Marino 	 through E->dest.  */
201e4b17023SJohn Marino       if (src != dst
202e4b17023SJohn Marino 	  && TREE_CODE (src) == SSA_NAME
203e4b17023SJohn Marino 	  && gimple_code (SSA_NAME_DEF_STMT (src)) == GIMPLE_PHI
204e4b17023SJohn Marino 	  && gimple_bb (SSA_NAME_DEF_STMT (src)) == e->dest)
205e4b17023SJohn Marino 	return false;
206e4b17023SJohn Marino 
207e4b17023SJohn Marino       /* We consider any non-virtual PHI as a statement since it
208e4b17023SJohn Marino 	 count result in a constant assignment or copy operation.  */
209e4b17023SJohn Marino       if (is_gimple_reg (dst))
210e4b17023SJohn Marino 	stmt_count++;
211e4b17023SJohn Marino 
212e4b17023SJohn Marino       record_temporary_equivalence (dst, src, stack);
213e4b17023SJohn Marino     }
214e4b17023SJohn Marino   return true;
215e4b17023SJohn Marino }
216e4b17023SJohn Marino 
217e4b17023SJohn Marino /* Fold the RHS of an assignment statement and return it as a tree.
218e4b17023SJohn Marino    May return NULL_TREE if no simplification is possible.  */
219e4b17023SJohn Marino 
220e4b17023SJohn Marino static tree
fold_assignment_stmt(gimple stmt)221e4b17023SJohn Marino fold_assignment_stmt (gimple stmt)
222e4b17023SJohn Marino {
223e4b17023SJohn Marino   enum tree_code subcode = gimple_assign_rhs_code (stmt);
224e4b17023SJohn Marino 
225e4b17023SJohn Marino   switch (get_gimple_rhs_class (subcode))
226e4b17023SJohn Marino     {
227e4b17023SJohn Marino     case GIMPLE_SINGLE_RHS:
228e4b17023SJohn Marino       return fold (gimple_assign_rhs1 (stmt));
229e4b17023SJohn Marino 
230e4b17023SJohn Marino     case GIMPLE_UNARY_RHS:
231e4b17023SJohn Marino       {
232e4b17023SJohn Marino         tree lhs = gimple_assign_lhs (stmt);
233e4b17023SJohn Marino         tree op0 = gimple_assign_rhs1 (stmt);
234e4b17023SJohn Marino         return fold_unary (subcode, TREE_TYPE (lhs), op0);
235e4b17023SJohn Marino       }
236e4b17023SJohn Marino 
237e4b17023SJohn Marino     case GIMPLE_BINARY_RHS:
238e4b17023SJohn Marino       {
239e4b17023SJohn Marino         tree lhs = gimple_assign_lhs (stmt);
240e4b17023SJohn Marino         tree op0 = gimple_assign_rhs1 (stmt);
241e4b17023SJohn Marino         tree op1 = gimple_assign_rhs2 (stmt);
242e4b17023SJohn Marino         return fold_binary (subcode, TREE_TYPE (lhs), op0, op1);
243e4b17023SJohn Marino       }
244e4b17023SJohn Marino 
245e4b17023SJohn Marino     case GIMPLE_TERNARY_RHS:
246e4b17023SJohn Marino       {
247e4b17023SJohn Marino         tree lhs = gimple_assign_lhs (stmt);
248e4b17023SJohn Marino         tree op0 = gimple_assign_rhs1 (stmt);
249e4b17023SJohn Marino         tree op1 = gimple_assign_rhs2 (stmt);
250e4b17023SJohn Marino         tree op2 = gimple_assign_rhs3 (stmt);
251e4b17023SJohn Marino 
252e4b17023SJohn Marino 	/* Sadly, we have to handle conditional assignments specially
253e4b17023SJohn Marino 	   here, because fold expects all the operands of an expression
254e4b17023SJohn Marino 	   to be folded before the expression itself is folded, but we
255e4b17023SJohn Marino 	   can't just substitute the folded condition here.  */
256e4b17023SJohn Marino         if (gimple_assign_rhs_code (stmt) == COND_EXPR)
257e4b17023SJohn Marino 	  op0 = fold (op0);
258e4b17023SJohn Marino 
259e4b17023SJohn Marino         return fold_ternary (subcode, TREE_TYPE (lhs), op0, op1, op2);
260e4b17023SJohn Marino       }
261e4b17023SJohn Marino 
262e4b17023SJohn Marino     default:
263e4b17023SJohn Marino       gcc_unreachable ();
264e4b17023SJohn Marino     }
265e4b17023SJohn Marino }
266e4b17023SJohn Marino 
267e4b17023SJohn Marino /* Try to simplify each statement in E->dest, ultimately leading to
268e4b17023SJohn Marino    a simplification of the COND_EXPR at the end of E->dest.
269e4b17023SJohn Marino 
270e4b17023SJohn Marino    Record unwind information for temporary equivalences onto STACK.
271e4b17023SJohn Marino 
272e4b17023SJohn Marino    Use SIMPLIFY (a pointer to a callback function) to further simplify
273e4b17023SJohn Marino    statements using pass specific information.
274e4b17023SJohn Marino 
275e4b17023SJohn Marino    We might consider marking just those statements which ultimately
276e4b17023SJohn Marino    feed the COND_EXPR.  It's not clear if the overhead of bookkeeping
277e4b17023SJohn Marino    would be recovered by trying to simplify fewer statements.
278e4b17023SJohn Marino 
279e4b17023SJohn Marino    If we are able to simplify a statement into the form
280e4b17023SJohn Marino    SSA_NAME = (SSA_NAME | gimple invariant), then we can record
281e4b17023SJohn Marino    a context sensitive equivalence which may help us simplify
282e4b17023SJohn Marino    later statements in E->dest.  */
283e4b17023SJohn Marino 
284e4b17023SJohn Marino static gimple
record_temporary_equivalences_from_stmts_at_dest(edge e,VEC (tree,heap)** stack,tree (* simplify)(gimple,gimple))285e4b17023SJohn Marino record_temporary_equivalences_from_stmts_at_dest (edge e,
286e4b17023SJohn Marino 						  VEC(tree, heap) **stack,
287e4b17023SJohn Marino 						  tree (*simplify) (gimple,
288e4b17023SJohn Marino 								    gimple))
289e4b17023SJohn Marino {
290e4b17023SJohn Marino   gimple stmt = NULL;
291e4b17023SJohn Marino   gimple_stmt_iterator gsi;
292e4b17023SJohn Marino   int max_stmt_count;
293e4b17023SJohn Marino 
294e4b17023SJohn Marino   max_stmt_count = PARAM_VALUE (PARAM_MAX_JUMP_THREAD_DUPLICATION_STMTS);
295e4b17023SJohn Marino 
296e4b17023SJohn Marino   /* Walk through each statement in the block recording equivalences
297e4b17023SJohn Marino      we discover.  Note any equivalences we discover are context
298e4b17023SJohn Marino      sensitive (ie, are dependent on traversing E) and must be unwound
299e4b17023SJohn Marino      when we're finished processing E.  */
300e4b17023SJohn Marino   for (gsi = gsi_start_bb (e->dest); !gsi_end_p (gsi); gsi_next (&gsi))
301e4b17023SJohn Marino     {
302e4b17023SJohn Marino       tree cached_lhs = NULL;
303e4b17023SJohn Marino 
304e4b17023SJohn Marino       stmt = gsi_stmt (gsi);
305e4b17023SJohn Marino 
306e4b17023SJohn Marino       /* Ignore empty statements and labels.  */
307e4b17023SJohn Marino       if (gimple_code (stmt) == GIMPLE_NOP
308e4b17023SJohn Marino 	  || gimple_code (stmt) == GIMPLE_LABEL
309e4b17023SJohn Marino 	  || is_gimple_debug (stmt))
310e4b17023SJohn Marino 	continue;
311e4b17023SJohn Marino 
312e4b17023SJohn Marino       /* If the statement has volatile operands, then we assume we
313e4b17023SJohn Marino 	 can not thread through this block.  This is overly
314e4b17023SJohn Marino 	 conservative in some ways.  */
315e4b17023SJohn Marino       if (gimple_code (stmt) == GIMPLE_ASM && gimple_asm_volatile_p (stmt))
316e4b17023SJohn Marino 	return NULL;
317e4b17023SJohn Marino 
318e4b17023SJohn Marino       /* If duplicating this block is going to cause too much code
319e4b17023SJohn Marino 	 expansion, then do not thread through this block.  */
320e4b17023SJohn Marino       stmt_count++;
321e4b17023SJohn Marino       if (stmt_count > max_stmt_count)
322e4b17023SJohn Marino 	return NULL;
323e4b17023SJohn Marino 
324e4b17023SJohn Marino       /* If this is not a statement that sets an SSA_NAME to a new
325e4b17023SJohn Marino 	 value, then do not try to simplify this statement as it will
326e4b17023SJohn Marino 	 not simplify in any way that is helpful for jump threading.  */
327e4b17023SJohn Marino       if ((gimple_code (stmt) != GIMPLE_ASSIGN
328e4b17023SJohn Marino            || TREE_CODE (gimple_assign_lhs (stmt)) != SSA_NAME)
329e4b17023SJohn Marino           && (gimple_code (stmt) != GIMPLE_CALL
330e4b17023SJohn Marino               || gimple_call_lhs (stmt) == NULL_TREE
331e4b17023SJohn Marino               || TREE_CODE (gimple_call_lhs (stmt)) != SSA_NAME))
332e4b17023SJohn Marino 	continue;
333e4b17023SJohn Marino 
334e4b17023SJohn Marino       /* The result of __builtin_object_size depends on all the arguments
335e4b17023SJohn Marino 	 of a phi node. Temporarily using only one edge produces invalid
336e4b17023SJohn Marino 	 results. For example
337e4b17023SJohn Marino 
338e4b17023SJohn Marino 	 if (x < 6)
339e4b17023SJohn Marino 	   goto l;
340e4b17023SJohn Marino 	 else
341e4b17023SJohn Marino 	   goto l;
342e4b17023SJohn Marino 
343e4b17023SJohn Marino 	 l:
344e4b17023SJohn Marino 	 r = PHI <&w[2].a[1](2), &a.a[6](3)>
345e4b17023SJohn Marino 	 __builtin_object_size (r, 0)
346e4b17023SJohn Marino 
347e4b17023SJohn Marino 	 The result of __builtin_object_size is defined to be the maximum of
348e4b17023SJohn Marino 	 remaining bytes. If we use only one edge on the phi, the result will
349e4b17023SJohn Marino 	 change to be the remaining bytes for the corresponding phi argument.
350e4b17023SJohn Marino 
351e4b17023SJohn Marino 	 Similarly for __builtin_constant_p:
352e4b17023SJohn Marino 
353e4b17023SJohn Marino 	 r = PHI <1(2), 2(3)>
354e4b17023SJohn Marino 	 __builtin_constant_p (r)
355e4b17023SJohn Marino 
356e4b17023SJohn Marino 	 Both PHI arguments are constant, but x ? 1 : 2 is still not
357e4b17023SJohn Marino 	 constant.  */
358e4b17023SJohn Marino 
359e4b17023SJohn Marino       if (is_gimple_call (stmt))
360e4b17023SJohn Marino 	{
361e4b17023SJohn Marino 	  tree fndecl = gimple_call_fndecl (stmt);
362e4b17023SJohn Marino 	  if (fndecl
363e4b17023SJohn Marino 	      && (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_OBJECT_SIZE
364e4b17023SJohn Marino 		  || DECL_FUNCTION_CODE (fndecl) == BUILT_IN_CONSTANT_P))
365e4b17023SJohn Marino 	    continue;
366e4b17023SJohn Marino 	}
367e4b17023SJohn Marino 
368e4b17023SJohn Marino       /* At this point we have a statement which assigns an RHS to an
369e4b17023SJohn Marino 	 SSA_VAR on the LHS.  We want to try and simplify this statement
370e4b17023SJohn Marino 	 to expose more context sensitive equivalences which in turn may
371e4b17023SJohn Marino 	 allow us to simplify the condition at the end of the loop.
372e4b17023SJohn Marino 
373e4b17023SJohn Marino 	 Handle simple copy operations as well as implied copies from
374e4b17023SJohn Marino 	 ASSERT_EXPRs.  */
375e4b17023SJohn Marino       if (gimple_assign_single_p (stmt)
376e4b17023SJohn Marino           && TREE_CODE (gimple_assign_rhs1 (stmt)) == SSA_NAME)
377e4b17023SJohn Marino 	cached_lhs = gimple_assign_rhs1 (stmt);
378e4b17023SJohn Marino       else if (gimple_assign_single_p (stmt)
379e4b17023SJohn Marino                && TREE_CODE (gimple_assign_rhs1 (stmt)) == ASSERT_EXPR)
380e4b17023SJohn Marino 	cached_lhs = TREE_OPERAND (gimple_assign_rhs1 (stmt), 0);
381e4b17023SJohn Marino       else
382e4b17023SJohn Marino 	{
383e4b17023SJohn Marino 	  /* A statement that is not a trivial copy or ASSERT_EXPR.
384e4b17023SJohn Marino 	     We're going to temporarily copy propagate the operands
385e4b17023SJohn Marino 	     and see if that allows us to simplify this statement.  */
386e4b17023SJohn Marino 	  tree *copy;
387e4b17023SJohn Marino 	  ssa_op_iter iter;
388e4b17023SJohn Marino 	  use_operand_p use_p;
389e4b17023SJohn Marino 	  unsigned int num, i = 0;
390e4b17023SJohn Marino 
391e4b17023SJohn Marino 	  num = NUM_SSA_OPERANDS (stmt, (SSA_OP_USE | SSA_OP_VUSE));
392e4b17023SJohn Marino 	  copy = XCNEWVEC (tree, num);
393e4b17023SJohn Marino 
394e4b17023SJohn Marino 	  /* Make a copy of the uses & vuses into USES_COPY, then cprop into
395e4b17023SJohn Marino 	     the operands.  */
396e4b17023SJohn Marino 	  FOR_EACH_SSA_USE_OPERAND (use_p, stmt, iter, SSA_OP_USE | SSA_OP_VUSE)
397e4b17023SJohn Marino 	    {
398e4b17023SJohn Marino 	      tree tmp = NULL;
399e4b17023SJohn Marino 	      tree use = USE_FROM_PTR (use_p);
400e4b17023SJohn Marino 
401e4b17023SJohn Marino 	      copy[i++] = use;
402e4b17023SJohn Marino 	      if (TREE_CODE (use) == SSA_NAME)
403e4b17023SJohn Marino 		tmp = SSA_NAME_VALUE (use);
404e4b17023SJohn Marino 	      if (tmp)
405e4b17023SJohn Marino 		SET_USE (use_p, tmp);
406e4b17023SJohn Marino 	    }
407e4b17023SJohn Marino 
408e4b17023SJohn Marino 	  /* Try to fold/lookup the new expression.  Inserting the
409e4b17023SJohn Marino 	     expression into the hash table is unlikely to help.  */
410e4b17023SJohn Marino           if (is_gimple_call (stmt))
411e4b17023SJohn Marino             cached_lhs = fold_call_stmt (stmt, false);
412e4b17023SJohn Marino 	  else
413e4b17023SJohn Marino             cached_lhs = fold_assignment_stmt (stmt);
414e4b17023SJohn Marino 
415e4b17023SJohn Marino           if (!cached_lhs
416e4b17023SJohn Marino               || (TREE_CODE (cached_lhs) != SSA_NAME
417e4b17023SJohn Marino                   && !is_gimple_min_invariant (cached_lhs)))
418e4b17023SJohn Marino             cached_lhs = (*simplify) (stmt, stmt);
419e4b17023SJohn Marino 
420e4b17023SJohn Marino 	  /* Restore the statement's original uses/defs.  */
421e4b17023SJohn Marino 	  i = 0;
422e4b17023SJohn Marino 	  FOR_EACH_SSA_USE_OPERAND (use_p, stmt, iter, SSA_OP_USE | SSA_OP_VUSE)
423e4b17023SJohn Marino 	    SET_USE (use_p, copy[i++]);
424e4b17023SJohn Marino 
425e4b17023SJohn Marino 	  free (copy);
426e4b17023SJohn Marino 	}
427e4b17023SJohn Marino 
428e4b17023SJohn Marino       /* Record the context sensitive equivalence if we were able
429e4b17023SJohn Marino 	 to simplify this statement.  */
430e4b17023SJohn Marino       if (cached_lhs
431e4b17023SJohn Marino 	  && (TREE_CODE (cached_lhs) == SSA_NAME
432e4b17023SJohn Marino 	      || is_gimple_min_invariant (cached_lhs)))
433e4b17023SJohn Marino 	record_temporary_equivalence (gimple_get_lhs (stmt), cached_lhs, stack);
434e4b17023SJohn Marino     }
435e4b17023SJohn Marino   return stmt;
436e4b17023SJohn Marino }
437e4b17023SJohn Marino 
438e4b17023SJohn Marino /* Simplify the control statement at the end of the block E->dest.
439e4b17023SJohn Marino 
440e4b17023SJohn Marino    To avoid allocating memory unnecessarily, a scratch GIMPLE_COND
441e4b17023SJohn Marino    is available to use/clobber in DUMMY_COND.
442e4b17023SJohn Marino 
443e4b17023SJohn Marino    Use SIMPLIFY (a pointer to a callback function) to further simplify
444e4b17023SJohn Marino    a condition using pass specific information.
445e4b17023SJohn Marino 
446e4b17023SJohn Marino    Return the simplified condition or NULL if simplification could
447e4b17023SJohn Marino    not be performed.  */
448e4b17023SJohn Marino 
449e4b17023SJohn Marino static tree
simplify_control_stmt_condition(edge e,gimple stmt,gimple dummy_cond,tree (* simplify)(gimple,gimple),bool handle_dominating_asserts)450e4b17023SJohn Marino simplify_control_stmt_condition (edge e,
451e4b17023SJohn Marino 				 gimple stmt,
452e4b17023SJohn Marino 				 gimple dummy_cond,
453e4b17023SJohn Marino 				 tree (*simplify) (gimple, gimple),
454e4b17023SJohn Marino 				 bool handle_dominating_asserts)
455e4b17023SJohn Marino {
456e4b17023SJohn Marino   tree cond, cached_lhs;
457e4b17023SJohn Marino   enum gimple_code code = gimple_code (stmt);
458e4b17023SJohn Marino 
459e4b17023SJohn Marino   /* For comparisons, we have to update both operands, then try
460e4b17023SJohn Marino      to simplify the comparison.  */
461e4b17023SJohn Marino   if (code == GIMPLE_COND)
462e4b17023SJohn Marino     {
463e4b17023SJohn Marino       tree op0, op1;
464e4b17023SJohn Marino       enum tree_code cond_code;
465e4b17023SJohn Marino 
466e4b17023SJohn Marino       op0 = gimple_cond_lhs (stmt);
467e4b17023SJohn Marino       op1 = gimple_cond_rhs (stmt);
468e4b17023SJohn Marino       cond_code = gimple_cond_code (stmt);
469e4b17023SJohn Marino 
470e4b17023SJohn Marino       /* Get the current value of both operands.  */
471e4b17023SJohn Marino       if (TREE_CODE (op0) == SSA_NAME)
472e4b17023SJohn Marino 	{
473e4b17023SJohn Marino           tree tmp = SSA_NAME_VALUE (op0);
474e4b17023SJohn Marino 	  if (tmp)
475e4b17023SJohn Marino 	    op0 = tmp;
476e4b17023SJohn Marino 	}
477e4b17023SJohn Marino 
478e4b17023SJohn Marino       if (TREE_CODE (op1) == SSA_NAME)
479e4b17023SJohn Marino 	{
480e4b17023SJohn Marino 	  tree tmp = SSA_NAME_VALUE (op1);
481e4b17023SJohn Marino 	  if (tmp)
482e4b17023SJohn Marino 	    op1 = tmp;
483e4b17023SJohn Marino 	}
484e4b17023SJohn Marino 
485e4b17023SJohn Marino       if (handle_dominating_asserts)
486e4b17023SJohn Marino 	{
487e4b17023SJohn Marino 	  /* Now see if the operand was consumed by an ASSERT_EXPR
488e4b17023SJohn Marino 	     which dominates E->src.  If so, we want to replace the
489e4b17023SJohn Marino 	     operand with the LHS of the ASSERT_EXPR.  */
490e4b17023SJohn Marino 	  if (TREE_CODE (op0) == SSA_NAME)
491e4b17023SJohn Marino 	    op0 = lhs_of_dominating_assert (op0, e->src, stmt);
492e4b17023SJohn Marino 
493e4b17023SJohn Marino 	  if (TREE_CODE (op1) == SSA_NAME)
494e4b17023SJohn Marino 	    op1 = lhs_of_dominating_assert (op1, e->src, stmt);
495e4b17023SJohn Marino 	}
496e4b17023SJohn Marino 
497e4b17023SJohn Marino       /* We may need to canonicalize the comparison.  For
498e4b17023SJohn Marino 	 example, op0 might be a constant while op1 is an
499e4b17023SJohn Marino 	 SSA_NAME.  Failure to canonicalize will cause us to
500e4b17023SJohn Marino 	 miss threading opportunities.  */
501e4b17023SJohn Marino       if (tree_swap_operands_p (op0, op1, false))
502e4b17023SJohn Marino 	{
503e4b17023SJohn Marino 	  tree tmp;
504e4b17023SJohn Marino 	  cond_code = swap_tree_comparison (cond_code);
505e4b17023SJohn Marino 	  tmp = op0;
506e4b17023SJohn Marino 	  op0 = op1;
507e4b17023SJohn Marino 	  op1 = tmp;
508e4b17023SJohn Marino 	}
509e4b17023SJohn Marino 
510e4b17023SJohn Marino       /* Stuff the operator and operands into our dummy conditional
511e4b17023SJohn Marino 	 expression.  */
512e4b17023SJohn Marino       gimple_cond_set_code (dummy_cond, cond_code);
513e4b17023SJohn Marino       gimple_cond_set_lhs (dummy_cond, op0);
514e4b17023SJohn Marino       gimple_cond_set_rhs (dummy_cond, op1);
515e4b17023SJohn Marino 
516e4b17023SJohn Marino       /* We absolutely do not care about any type conversions
517e4b17023SJohn Marino          we only care about a zero/nonzero value.  */
518e4b17023SJohn Marino       fold_defer_overflow_warnings ();
519e4b17023SJohn Marino 
520e4b17023SJohn Marino       cached_lhs = fold_binary (cond_code, boolean_type_node, op0, op1);
521e4b17023SJohn Marino       if (cached_lhs)
522e4b17023SJohn Marino 	while (CONVERT_EXPR_P (cached_lhs))
523e4b17023SJohn Marino           cached_lhs = TREE_OPERAND (cached_lhs, 0);
524e4b17023SJohn Marino 
525e4b17023SJohn Marino       fold_undefer_overflow_warnings ((cached_lhs
526e4b17023SJohn Marino                                        && is_gimple_min_invariant (cached_lhs)),
527e4b17023SJohn Marino 				      stmt, WARN_STRICT_OVERFLOW_CONDITIONAL);
528e4b17023SJohn Marino 
529e4b17023SJohn Marino       /* If we have not simplified the condition down to an invariant,
530e4b17023SJohn Marino 	 then use the pass specific callback to simplify the condition.  */
531e4b17023SJohn Marino       if (!cached_lhs
532e4b17023SJohn Marino           || !is_gimple_min_invariant (cached_lhs))
533e4b17023SJohn Marino         cached_lhs = (*simplify) (dummy_cond, stmt);
534e4b17023SJohn Marino 
535e4b17023SJohn Marino       return cached_lhs;
536e4b17023SJohn Marino     }
537e4b17023SJohn Marino 
538e4b17023SJohn Marino   if (code == GIMPLE_SWITCH)
539e4b17023SJohn Marino     cond = gimple_switch_index (stmt);
540e4b17023SJohn Marino   else if (code == GIMPLE_GOTO)
541e4b17023SJohn Marino     cond = gimple_goto_dest (stmt);
542e4b17023SJohn Marino   else
543e4b17023SJohn Marino     gcc_unreachable ();
544e4b17023SJohn Marino 
545e4b17023SJohn Marino   /* We can have conditionals which just test the state of a variable
546e4b17023SJohn Marino      rather than use a relational operator.  These are simpler to handle.  */
547e4b17023SJohn Marino   if (TREE_CODE (cond) == SSA_NAME)
548e4b17023SJohn Marino     {
549e4b17023SJohn Marino       cached_lhs = cond;
550e4b17023SJohn Marino 
551e4b17023SJohn Marino       /* Get the variable's current value from the equivalence chains.
552e4b17023SJohn Marino 
553e4b17023SJohn Marino 	 It is possible to get loops in the SSA_NAME_VALUE chains
554e4b17023SJohn Marino 	 (consider threading the backedge of a loop where we have
555e4b17023SJohn Marino 	 a loop invariant SSA_NAME used in the condition.  */
556e4b17023SJohn Marino       if (cached_lhs
557e4b17023SJohn Marino 	  && TREE_CODE (cached_lhs) == SSA_NAME
558e4b17023SJohn Marino 	  && SSA_NAME_VALUE (cached_lhs))
559e4b17023SJohn Marino 	cached_lhs = SSA_NAME_VALUE (cached_lhs);
560e4b17023SJohn Marino 
561e4b17023SJohn Marino       /* If we're dominated by a suitable ASSERT_EXPR, then
562e4b17023SJohn Marino 	 update CACHED_LHS appropriately.  */
563e4b17023SJohn Marino       if (handle_dominating_asserts && TREE_CODE (cached_lhs) == SSA_NAME)
564e4b17023SJohn Marino 	cached_lhs = lhs_of_dominating_assert (cached_lhs, e->src, stmt);
565e4b17023SJohn Marino 
566e4b17023SJohn Marino       /* If we haven't simplified to an invariant yet, then use the
567e4b17023SJohn Marino 	 pass specific callback to try and simplify it further.  */
568e4b17023SJohn Marino       if (cached_lhs && ! is_gimple_min_invariant (cached_lhs))
569e4b17023SJohn Marino         cached_lhs = (*simplify) (stmt, stmt);
570e4b17023SJohn Marino     }
571e4b17023SJohn Marino   else
572e4b17023SJohn Marino     cached_lhs = NULL;
573e4b17023SJohn Marino 
574e4b17023SJohn Marino   return cached_lhs;
575e4b17023SJohn Marino }
576e4b17023SJohn Marino 
577*5ce9237cSJohn Marino /* Return TRUE if the statement at the end of e->dest depends on
578*5ce9237cSJohn Marino    the output of any statement in BB.   Otherwise return FALSE.
579*5ce9237cSJohn Marino 
580*5ce9237cSJohn Marino    This is used when we are threading a backedge and need to ensure
581*5ce9237cSJohn Marino    that temporary equivalences from BB do not affect the condition
582*5ce9237cSJohn Marino    in e->dest.  */
583*5ce9237cSJohn Marino 
584*5ce9237cSJohn Marino static bool
cond_arg_set_in_bb(edge e,basic_block bb)585*5ce9237cSJohn Marino cond_arg_set_in_bb (edge e, basic_block bb)
586*5ce9237cSJohn Marino {
587*5ce9237cSJohn Marino   ssa_op_iter iter;
588*5ce9237cSJohn Marino   use_operand_p use_p;
589*5ce9237cSJohn Marino   gimple last = last_stmt (e->dest);
590*5ce9237cSJohn Marino 
591*5ce9237cSJohn Marino   /* E->dest does not have to end with a control transferring
592*5ce9237cSJohn Marino      instruction.  This can occurr when we try to extend a jump
593*5ce9237cSJohn Marino      threading opportunity deeper into the CFG.  In that case
594*5ce9237cSJohn Marino      it is safe for this check to return false.  */
595*5ce9237cSJohn Marino   if (!last)
596*5ce9237cSJohn Marino     return false;
597*5ce9237cSJohn Marino 
598*5ce9237cSJohn Marino   if (gimple_code (last) != GIMPLE_COND
599*5ce9237cSJohn Marino       && gimple_code (last) != GIMPLE_GOTO
600*5ce9237cSJohn Marino       && gimple_code (last) != GIMPLE_SWITCH)
601*5ce9237cSJohn Marino     return false;
602*5ce9237cSJohn Marino 
603*5ce9237cSJohn Marino   FOR_EACH_SSA_USE_OPERAND (use_p, last, iter, SSA_OP_USE | SSA_OP_VUSE)
604*5ce9237cSJohn Marino     {
605*5ce9237cSJohn Marino       tree use = USE_FROM_PTR (use_p);
606*5ce9237cSJohn Marino 
607*5ce9237cSJohn Marino       if (TREE_CODE (use) == SSA_NAME
608*5ce9237cSJohn Marino 	  && gimple_code (SSA_NAME_DEF_STMT (use)) != GIMPLE_PHI
609*5ce9237cSJohn Marino 	  && gimple_bb (SSA_NAME_DEF_STMT (use)) == bb)
610*5ce9237cSJohn Marino 	return true;
611*5ce9237cSJohn Marino     }
612*5ce9237cSJohn Marino   return false;
613*5ce9237cSJohn Marino }
614*5ce9237cSJohn Marino 
615e4b17023SJohn Marino /* TAKEN_EDGE represents the an edge taken as a result of jump threading.
616e4b17023SJohn Marino    See if we can thread around TAKEN_EDGE->dest as well.  If so, return
617e4b17023SJohn Marino    the edge out of TAKEN_EDGE->dest that we can statically compute will be
618e4b17023SJohn Marino    traversed.
619e4b17023SJohn Marino 
620e4b17023SJohn Marino    We are much more restrictive as to the contents of TAKEN_EDGE->dest
621e4b17023SJohn Marino    as the path isolation code in tree-ssa-threadupdate.c isn't prepared
622e4b17023SJohn Marino    to handle copying intermediate blocks on a threaded path.
623e4b17023SJohn Marino 
624e4b17023SJohn Marino    Long term a more consistent and structured approach to path isolation
625e4b17023SJohn Marino    would be a huge help.   */
626e4b17023SJohn Marino static edge
thread_around_empty_block(edge taken_edge,gimple dummy_cond,bool handle_dominating_asserts,tree (* simplify)(gimple,gimple),bitmap visited)627e4b17023SJohn Marino thread_around_empty_block (edge taken_edge,
628e4b17023SJohn Marino 			   gimple dummy_cond,
629e4b17023SJohn Marino 			   bool handle_dominating_asserts,
630e4b17023SJohn Marino 			   tree (*simplify) (gimple, gimple),
631e4b17023SJohn Marino 			   bitmap visited)
632e4b17023SJohn Marino {
633e4b17023SJohn Marino   basic_block bb = taken_edge->dest;
634e4b17023SJohn Marino   gimple_stmt_iterator gsi;
635e4b17023SJohn Marino   gimple stmt;
636e4b17023SJohn Marino   tree cond;
637e4b17023SJohn Marino 
638e4b17023SJohn Marino   /* This block must have a single predecessor (E->dest).  */
639e4b17023SJohn Marino   if (!single_pred_p (bb))
640e4b17023SJohn Marino     return NULL;
641e4b17023SJohn Marino 
642e4b17023SJohn Marino   /* This block must have more than one successor.  */
643e4b17023SJohn Marino   if (single_succ_p (bb))
644e4b17023SJohn Marino     return NULL;
645e4b17023SJohn Marino 
646e4b17023SJohn Marino   /* This block can have no PHI nodes.  This is overly conservative.  */
647e4b17023SJohn Marino   if (!gsi_end_p (gsi_start_phis (bb)))
648e4b17023SJohn Marino     return NULL;
649e4b17023SJohn Marino 
650e4b17023SJohn Marino   /* Skip over DEBUG statements at the start of the block.  */
651e4b17023SJohn Marino   gsi = gsi_start_nondebug_bb (bb);
652e4b17023SJohn Marino 
653e4b17023SJohn Marino   if (gsi_end_p (gsi))
654e4b17023SJohn Marino     return NULL;
655e4b17023SJohn Marino 
656e4b17023SJohn Marino   /* This block can have no statements other than its control altering
657e4b17023SJohn Marino      statement.  This is overly conservative.  */
658e4b17023SJohn Marino   stmt = gsi_stmt (gsi);
659e4b17023SJohn Marino   if (gimple_code (stmt) != GIMPLE_COND
660e4b17023SJohn Marino       && gimple_code (stmt) != GIMPLE_GOTO
661e4b17023SJohn Marino       && gimple_code (stmt) != GIMPLE_SWITCH)
662e4b17023SJohn Marino     return NULL;
663e4b17023SJohn Marino 
664e4b17023SJohn Marino   /* Extract and simplify the condition.  */
665e4b17023SJohn Marino   cond = simplify_control_stmt_condition (taken_edge, stmt, dummy_cond,
666e4b17023SJohn Marino 					  simplify, handle_dominating_asserts);
667e4b17023SJohn Marino 
668e4b17023SJohn Marino   /* If the condition can be statically computed and we have not already
669e4b17023SJohn Marino      visited the destination edge, then add the taken edge to our thread
670e4b17023SJohn Marino      path.  */
671e4b17023SJohn Marino   if (cond && is_gimple_min_invariant (cond))
672e4b17023SJohn Marino     {
673e4b17023SJohn Marino       edge taken_edge = find_taken_edge (bb, cond);
674e4b17023SJohn Marino 
675e4b17023SJohn Marino       if (bitmap_bit_p (visited, taken_edge->dest->index))
676e4b17023SJohn Marino 	return NULL;
677e4b17023SJohn Marino       bitmap_set_bit (visited, taken_edge->dest->index);
678e4b17023SJohn Marino       return taken_edge;
679e4b17023SJohn Marino     }
680e4b17023SJohn Marino 
681e4b17023SJohn Marino   return NULL;
682e4b17023SJohn Marino }
683e4b17023SJohn Marino 
684e4b17023SJohn Marino /* E1 and E2 are edges into the same basic block.  Return TRUE if the
685e4b17023SJohn Marino    PHI arguments associated with those edges are equal or there are no
686e4b17023SJohn Marino    PHI arguments, otherwise return FALSE.  */
687e4b17023SJohn Marino 
688e4b17023SJohn Marino static bool
phi_args_equal_on_edges(edge e1,edge e2)689e4b17023SJohn Marino phi_args_equal_on_edges (edge e1, edge e2)
690e4b17023SJohn Marino {
691e4b17023SJohn Marino   gimple_stmt_iterator gsi;
692e4b17023SJohn Marino   int indx1 = e1->dest_idx;
693e4b17023SJohn Marino   int indx2 = e2->dest_idx;
694e4b17023SJohn Marino 
695e4b17023SJohn Marino   for (gsi = gsi_start_phis (e1->dest); !gsi_end_p (gsi); gsi_next (&gsi))
696e4b17023SJohn Marino     {
697e4b17023SJohn Marino       gimple phi = gsi_stmt (gsi);
698e4b17023SJohn Marino 
699e4b17023SJohn Marino       if (!operand_equal_p (gimple_phi_arg_def (phi, indx1),
700e4b17023SJohn Marino 			    gimple_phi_arg_def (phi, indx2), 0))
701e4b17023SJohn Marino 	return false;
702e4b17023SJohn Marino     }
703e4b17023SJohn Marino   return true;
704e4b17023SJohn Marino }
705e4b17023SJohn Marino 
706e4b17023SJohn Marino /* We are exiting E->src, see if E->dest ends with a conditional
707e4b17023SJohn Marino    jump which has a known value when reached via E.
708e4b17023SJohn Marino 
709e4b17023SJohn Marino    Special care is necessary if E is a back edge in the CFG as we
710e4b17023SJohn Marino    may have already recorded equivalences for E->dest into our
711e4b17023SJohn Marino    various tables, including the result of the conditional at
712e4b17023SJohn Marino    the end of E->dest.  Threading opportunities are severely
713e4b17023SJohn Marino    limited in that case to avoid short-circuiting the loop
714e4b17023SJohn Marino    incorrectly.
715e4b17023SJohn Marino 
716e4b17023SJohn Marino    Note it is quite common for the first block inside a loop to
717e4b17023SJohn Marino    end with a conditional which is either always true or always
718e4b17023SJohn Marino    false when reached via the loop backedge.  Thus we do not want
719e4b17023SJohn Marino    to blindly disable threading across a loop backedge.
720e4b17023SJohn Marino 
721e4b17023SJohn Marino    DUMMY_COND is a shared cond_expr used by condition simplification as scratch,
722e4b17023SJohn Marino    to avoid allocating memory.
723e4b17023SJohn Marino 
724e4b17023SJohn Marino    HANDLE_DOMINATING_ASSERTS is true if we should try to replace operands of
725e4b17023SJohn Marino    the simplified condition with left-hand sides of ASSERT_EXPRs they are
726e4b17023SJohn Marino    used in.
727e4b17023SJohn Marino 
728e4b17023SJohn Marino    STACK is used to undo temporary equivalences created during the walk of
729e4b17023SJohn Marino    E->dest.
730e4b17023SJohn Marino 
731e4b17023SJohn Marino    SIMPLIFY is a pass-specific function used to simplify statements.  */
732e4b17023SJohn Marino 
733e4b17023SJohn Marino void
thread_across_edge(gimple dummy_cond,edge e,bool handle_dominating_asserts,VEC (tree,heap)** stack,tree (* simplify)(gimple,gimple))734e4b17023SJohn Marino thread_across_edge (gimple dummy_cond,
735e4b17023SJohn Marino 		    edge e,
736e4b17023SJohn Marino 		    bool handle_dominating_asserts,
737e4b17023SJohn Marino 		    VEC(tree, heap) **stack,
738e4b17023SJohn Marino 		    tree (*simplify) (gimple, gimple))
739e4b17023SJohn Marino {
740e4b17023SJohn Marino   gimple stmt;
741e4b17023SJohn Marino 
742e4b17023SJohn Marino   /* If E is a backedge, then we want to verify that the COND_EXPR,
743e4b17023SJohn Marino      SWITCH_EXPR or GOTO_EXPR at the end of e->dest is not affected
744e4b17023SJohn Marino      by any statements in e->dest.  If it is affected, then it is not
745e4b17023SJohn Marino      safe to thread this edge.  */
746e4b17023SJohn Marino   if (e->flags & EDGE_DFS_BACK)
747e4b17023SJohn Marino     {
748*5ce9237cSJohn Marino       if (cond_arg_set_in_bb (e, e->dest))
749e4b17023SJohn Marino 	goto fail;
750e4b17023SJohn Marino     }
751e4b17023SJohn Marino 
752e4b17023SJohn Marino   stmt_count = 0;
753e4b17023SJohn Marino 
754e4b17023SJohn Marino   /* PHIs create temporary equivalences.  */
755e4b17023SJohn Marino   if (!record_temporary_equivalences_from_phis (e, stack))
756e4b17023SJohn Marino     goto fail;
757e4b17023SJohn Marino 
758e4b17023SJohn Marino   /* Now walk each statement recording any context sensitive
759e4b17023SJohn Marino      temporary equivalences we can detect.  */
760e4b17023SJohn Marino   stmt = record_temporary_equivalences_from_stmts_at_dest (e, stack, simplify);
761e4b17023SJohn Marino   if (!stmt)
762e4b17023SJohn Marino     goto fail;
763e4b17023SJohn Marino 
764e4b17023SJohn Marino   /* If we stopped at a COND_EXPR or SWITCH_EXPR, see if we know which arm
765e4b17023SJohn Marino      will be taken.  */
766e4b17023SJohn Marino   if (gimple_code (stmt) == GIMPLE_COND
767e4b17023SJohn Marino       || gimple_code (stmt) == GIMPLE_GOTO
768e4b17023SJohn Marino       || gimple_code (stmt) == GIMPLE_SWITCH)
769e4b17023SJohn Marino     {
770e4b17023SJohn Marino       tree cond;
771e4b17023SJohn Marino 
772e4b17023SJohn Marino       /* Extract and simplify the condition.  */
773e4b17023SJohn Marino       cond = simplify_control_stmt_condition (e, stmt, dummy_cond, simplify,
774e4b17023SJohn Marino 					      handle_dominating_asserts);
775e4b17023SJohn Marino 
776e4b17023SJohn Marino       if (cond && is_gimple_min_invariant (cond))
777e4b17023SJohn Marino 	{
778e4b17023SJohn Marino 	  edge taken_edge = find_taken_edge (e->dest, cond);
779e4b17023SJohn Marino 	  basic_block dest = (taken_edge ? taken_edge->dest : NULL);
780e4b17023SJohn Marino 	  bitmap visited;
781e4b17023SJohn Marino 	  edge e2;
782e4b17023SJohn Marino 
783e4b17023SJohn Marino 	  if (dest == e->dest)
784e4b17023SJohn Marino 	    goto fail;
785e4b17023SJohn Marino 
786e4b17023SJohn Marino 	  /* DEST could be null for a computed jump to an absolute
787e4b17023SJohn Marino 	     address.  If DEST is not null, then see if we can thread
788e4b17023SJohn Marino 	     through it as well, this helps capture secondary effects
789e4b17023SJohn Marino 	     of threading without having to re-run DOM or VRP.  */
790*5ce9237cSJohn Marino 	  if (dest
791*5ce9237cSJohn Marino 	      && ((e->flags & EDGE_DFS_BACK) == 0
792*5ce9237cSJohn Marino 		  || ! cond_arg_set_in_bb (taken_edge, e->dest)))
793e4b17023SJohn Marino 	    {
794e4b17023SJohn Marino 	      /* We don't want to thread back to a block we have already
795e4b17023SJohn Marino  		 visited.  This may be overly conservative.  */
796e4b17023SJohn Marino 	      visited = BITMAP_ALLOC (NULL);
797e4b17023SJohn Marino 	      bitmap_set_bit (visited, dest->index);
798e4b17023SJohn Marino 	      bitmap_set_bit (visited, e->dest->index);
799e4b17023SJohn Marino 	      do
800e4b17023SJohn Marino 		{
801e4b17023SJohn Marino 		  e2 = thread_around_empty_block (taken_edge,
802e4b17023SJohn Marino 						  dummy_cond,
803e4b17023SJohn Marino 						  handle_dominating_asserts,
804e4b17023SJohn Marino 						  simplify,
805e4b17023SJohn Marino 						  visited);
806e4b17023SJohn Marino 		  if (e2)
807e4b17023SJohn Marino 		    taken_edge = e2;
808e4b17023SJohn Marino 		}
809e4b17023SJohn Marino 	      while (e2);
810e4b17023SJohn Marino 	      BITMAP_FREE (visited);
811e4b17023SJohn Marino 	    }
812e4b17023SJohn Marino 
813e4b17023SJohn Marino 	  remove_temporary_equivalences (stack);
814e4b17023SJohn Marino 	  register_jump_thread (e, taken_edge, NULL);
815e4b17023SJohn Marino 	  return;
816e4b17023SJohn Marino 	}
817e4b17023SJohn Marino     }
818e4b17023SJohn Marino 
819e4b17023SJohn Marino  /* We were unable to determine what out edge from E->dest is taken.  However,
820e4b17023SJohn Marino     we might still be able to thread through successors of E->dest.  This
821e4b17023SJohn Marino     often occurs when E->dest is a joiner block which then fans back out
822e4b17023SJohn Marino     based on redundant tests.
823e4b17023SJohn Marino 
824e4b17023SJohn Marino     If so, we'll copy E->dest and redirect the appropriate predecessor to
825e4b17023SJohn Marino     the copy.  Within the copy of E->dest, we'll thread one or more edges
826e4b17023SJohn Marino     to points deeper in the CFG.
827e4b17023SJohn Marino 
828e4b17023SJohn Marino     This is a stopgap until we have a more structured approach to path
829e4b17023SJohn Marino     isolation.  */
830e4b17023SJohn Marino   {
831e4b17023SJohn Marino     edge e2, e3, taken_edge;
832e4b17023SJohn Marino     edge_iterator ei;
833e4b17023SJohn Marino     bool found = false;
834e4b17023SJohn Marino     bitmap visited = BITMAP_ALLOC (NULL);
835e4b17023SJohn Marino 
836e4b17023SJohn Marino     /* Look at each successor of E->dest to see if we can thread through it.  */
837e4b17023SJohn Marino     FOR_EACH_EDGE (taken_edge, ei, e->dest->succs)
838e4b17023SJohn Marino       {
839e4b17023SJohn Marino 	/* Avoid threading to any block we have already visited.  */
840e4b17023SJohn Marino 	bitmap_clear (visited);
841e4b17023SJohn Marino 	bitmap_set_bit (visited, taken_edge->dest->index);
842e4b17023SJohn Marino 	bitmap_set_bit (visited, e->dest->index);
843e4b17023SJohn Marino 
844e4b17023SJohn Marino 	/* Record whether or not we were able to thread through a successor
845e4b17023SJohn Marino 	   of E->dest.  */
846e4b17023SJohn Marino 	found = false;
847e4b17023SJohn Marino 	e3 = taken_edge;
848e4b17023SJohn Marino 	do
849e4b17023SJohn Marino 	  {
850*5ce9237cSJohn Marino 	    if ((e->flags & EDGE_DFS_BACK) == 0
851*5ce9237cSJohn Marino 		|| ! cond_arg_set_in_bb (e3, e->dest))
852e4b17023SJohn Marino 	      e2 = thread_around_empty_block (e3,
853e4b17023SJohn Marino 					      dummy_cond,
854e4b17023SJohn Marino 					      handle_dominating_asserts,
855e4b17023SJohn Marino 					      simplify,
856e4b17023SJohn Marino 					      visited);
857*5ce9237cSJohn Marino 	    else
858*5ce9237cSJohn Marino 	      e2 = NULL;
859*5ce9237cSJohn Marino 
860e4b17023SJohn Marino 	    if (e2)
861e4b17023SJohn Marino 	      {
862e4b17023SJohn Marino 	        e3 = e2;
863e4b17023SJohn Marino 		found = true;
864e4b17023SJohn Marino 	      }
865e4b17023SJohn Marino 	  }
866e4b17023SJohn Marino         while (e2);
867e4b17023SJohn Marino 
868e4b17023SJohn Marino 	/* If we were able to thread through a successor of E->dest, then
869e4b17023SJohn Marino 	   record the jump threading opportunity.  */
870e4b17023SJohn Marino 	if (found)
871e4b17023SJohn Marino 	  {
872e4b17023SJohn Marino 	    edge tmp;
873e4b17023SJohn Marino 	    /* If there is already an edge from the block to be duplicated
874e4b17023SJohn Marino 	       (E2->src) to the final target (E3->dest), then make sure that
875e4b17023SJohn Marino 	       the PHI args associated with the edges E2 and E3 are the
876e4b17023SJohn Marino 	       same.  */
877e4b17023SJohn Marino 	    tmp = find_edge (taken_edge->src, e3->dest);
878e4b17023SJohn Marino 	    if (!tmp || phi_args_equal_on_edges (tmp, e3))
879e4b17023SJohn Marino 	      register_jump_thread (e, taken_edge, e3);
880e4b17023SJohn Marino 	  }
881e4b17023SJohn Marino 
882e4b17023SJohn Marino       }
883e4b17023SJohn Marino     BITMAP_FREE (visited);
884e4b17023SJohn Marino   }
885e4b17023SJohn Marino 
886e4b17023SJohn Marino  fail:
887e4b17023SJohn Marino   remove_temporary_equivalences (stack);
888e4b17023SJohn Marino }
889