1*38fd1498Szrj /* Loop unswitching.
2*38fd1498Szrj Copyright (C) 2004-2018 Free Software Foundation, Inc.
3*38fd1498Szrj
4*38fd1498Szrj This file is part of GCC.
5*38fd1498Szrj
6*38fd1498Szrj GCC is free software; you can redistribute it and/or modify it
7*38fd1498Szrj under the terms of the GNU General Public License as published by the
8*38fd1498Szrj Free Software Foundation; either version 3, or (at your option) any
9*38fd1498Szrj later version.
10*38fd1498Szrj
11*38fd1498Szrj GCC is distributed in the hope that it will be useful, but WITHOUT
12*38fd1498Szrj ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
13*38fd1498Szrj FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
14*38fd1498Szrj for more details.
15*38fd1498Szrj
16*38fd1498Szrj You should have received a copy of the GNU General Public License
17*38fd1498Szrj along with GCC; see the file COPYING3. If not see
18*38fd1498Szrj <http://www.gnu.org/licenses/>. */
19*38fd1498Szrj
20*38fd1498Szrj #include "config.h"
21*38fd1498Szrj #include "system.h"
22*38fd1498Szrj #include "coretypes.h"
23*38fd1498Szrj #include "backend.h"
24*38fd1498Szrj #include "tree.h"
25*38fd1498Szrj #include "gimple.h"
26*38fd1498Szrj #include "tree-pass.h"
27*38fd1498Szrj #include "ssa.h"
28*38fd1498Szrj #include "fold-const.h"
29*38fd1498Szrj #include "gimplify.h"
30*38fd1498Szrj #include "tree-cfg.h"
31*38fd1498Szrj #include "tree-ssa.h"
32*38fd1498Szrj #include "tree-ssa-loop-niter.h"
33*38fd1498Szrj #include "tree-ssa-loop.h"
34*38fd1498Szrj #include "tree-into-ssa.h"
35*38fd1498Szrj #include "cfgloop.h"
36*38fd1498Szrj #include "params.h"
37*38fd1498Szrj #include "tree-inline.h"
38*38fd1498Szrj #include "gimple-iterator.h"
39*38fd1498Szrj #include "cfghooks.h"
40*38fd1498Szrj #include "tree-ssa-loop-manip.h"
41*38fd1498Szrj
42*38fd1498Szrj /* This file implements the loop unswitching, i.e. transformation of loops like
43*38fd1498Szrj
44*38fd1498Szrj while (A)
45*38fd1498Szrj {
46*38fd1498Szrj if (inv)
47*38fd1498Szrj B;
48*38fd1498Szrj
49*38fd1498Szrj X;
50*38fd1498Szrj
51*38fd1498Szrj if (!inv)
52*38fd1498Szrj C;
53*38fd1498Szrj }
54*38fd1498Szrj
55*38fd1498Szrj where inv is the loop invariant, into
56*38fd1498Szrj
57*38fd1498Szrj if (inv)
58*38fd1498Szrj {
59*38fd1498Szrj while (A)
60*38fd1498Szrj {
61*38fd1498Szrj B;
62*38fd1498Szrj X;
63*38fd1498Szrj }
64*38fd1498Szrj }
65*38fd1498Szrj else
66*38fd1498Szrj {
67*38fd1498Szrj while (A)
68*38fd1498Szrj {
69*38fd1498Szrj X;
70*38fd1498Szrj C;
71*38fd1498Szrj }
72*38fd1498Szrj }
73*38fd1498Szrj
74*38fd1498Szrj Inv is considered invariant iff the values it compares are both invariant;
75*38fd1498Szrj tree-ssa-loop-im.c ensures that all the suitable conditions are in this
76*38fd1498Szrj shape. */
77*38fd1498Szrj
78*38fd1498Szrj static struct loop *tree_unswitch_loop (struct loop *, basic_block, tree);
79*38fd1498Szrj static bool tree_unswitch_single_loop (struct loop *, int);
80*38fd1498Szrj static tree tree_may_unswitch_on (basic_block, struct loop *);
81*38fd1498Szrj static bool tree_unswitch_outer_loop (struct loop *);
82*38fd1498Szrj static edge find_loop_guard (struct loop *);
83*38fd1498Szrj static bool empty_bb_without_guard_p (struct loop *, basic_block);
84*38fd1498Szrj static bool used_outside_loop_p (struct loop *, tree);
85*38fd1498Szrj static void hoist_guard (struct loop *, edge);
86*38fd1498Szrj static bool check_exit_phi (struct loop *);
87*38fd1498Szrj static tree get_vop_from_header (struct loop *);
88*38fd1498Szrj
89*38fd1498Szrj /* Main entry point. Perform loop unswitching on all suitable loops. */
90*38fd1498Szrj
91*38fd1498Szrj unsigned int
tree_ssa_unswitch_loops(void)92*38fd1498Szrj tree_ssa_unswitch_loops (void)
93*38fd1498Szrj {
94*38fd1498Szrj struct loop *loop;
95*38fd1498Szrj bool changed = false;
96*38fd1498Szrj
97*38fd1498Szrj /* Go through all loops starting from innermost. */
98*38fd1498Szrj FOR_EACH_LOOP (loop, LI_FROM_INNERMOST)
99*38fd1498Szrj {
100*38fd1498Szrj if (!loop->inner)
101*38fd1498Szrj /* Unswitch innermost loop. */
102*38fd1498Szrj changed |= tree_unswitch_single_loop (loop, 0);
103*38fd1498Szrj else
104*38fd1498Szrj changed |= tree_unswitch_outer_loop (loop);
105*38fd1498Szrj }
106*38fd1498Szrj
107*38fd1498Szrj if (changed)
108*38fd1498Szrj return TODO_cleanup_cfg;
109*38fd1498Szrj return 0;
110*38fd1498Szrj }
111*38fd1498Szrj
112*38fd1498Szrj /* Return TRUE if an SSA_NAME maybe undefined and is therefore
113*38fd1498Szrj unsuitable for unswitching. STMT is the statement we are
114*38fd1498Szrj considering for unswitching and LOOP is the loop it appears in. */
115*38fd1498Szrj
116*38fd1498Szrj static bool
is_maybe_undefined(const tree name,gimple * stmt,struct loop * loop)117*38fd1498Szrj is_maybe_undefined (const tree name, gimple *stmt, struct loop *loop)
118*38fd1498Szrj {
119*38fd1498Szrj /* The loop header is the only block we can trivially determine that
120*38fd1498Szrj will always be executed. If the comparison is in the loop
121*38fd1498Szrj header, we know it's OK to unswitch on it. */
122*38fd1498Szrj if (gimple_bb (stmt) == loop->header)
123*38fd1498Szrj return false;
124*38fd1498Szrj
125*38fd1498Szrj auto_bitmap visited_ssa;
126*38fd1498Szrj auto_vec<tree> worklist;
127*38fd1498Szrj worklist.safe_push (name);
128*38fd1498Szrj bitmap_set_bit (visited_ssa, SSA_NAME_VERSION (name));
129*38fd1498Szrj while (!worklist.is_empty ())
130*38fd1498Szrj {
131*38fd1498Szrj tree t = worklist.pop ();
132*38fd1498Szrj
133*38fd1498Szrj /* If it's obviously undefined, avoid further computations. */
134*38fd1498Szrj if (ssa_undefined_value_p (t, true))
135*38fd1498Szrj return true;
136*38fd1498Szrj
137*38fd1498Szrj if (ssa_defined_default_def_p (t))
138*38fd1498Szrj continue;
139*38fd1498Szrj
140*38fd1498Szrj gimple *def = SSA_NAME_DEF_STMT (t);
141*38fd1498Szrj
142*38fd1498Szrj /* Check that all the PHI args are fully defined. */
143*38fd1498Szrj if (gphi *phi = dyn_cast <gphi *> (def))
144*38fd1498Szrj {
145*38fd1498Szrj for (unsigned i = 0; i < gimple_phi_num_args (phi); ++i)
146*38fd1498Szrj {
147*38fd1498Szrj tree t = gimple_phi_arg_def (phi, i);
148*38fd1498Szrj /* If an SSA has already been seen, it may be a loop,
149*38fd1498Szrj but we can continue and ignore this use. Otherwise,
150*38fd1498Szrj add the SSA_NAME to the queue and visit it later. */
151*38fd1498Szrj if (TREE_CODE (t) == SSA_NAME
152*38fd1498Szrj && bitmap_set_bit (visited_ssa, SSA_NAME_VERSION (t)))
153*38fd1498Szrj worklist.safe_push (t);
154*38fd1498Szrj }
155*38fd1498Szrj continue;
156*38fd1498Szrj }
157*38fd1498Szrj
158*38fd1498Szrj /* Uses in stmts always executed when the region header executes
159*38fd1498Szrj are fine. */
160*38fd1498Szrj if (dominated_by_p (CDI_DOMINATORS, loop->header, gimple_bb (def)))
161*38fd1498Szrj continue;
162*38fd1498Szrj
163*38fd1498Szrj /* Handle calls and memory loads conservatively. */
164*38fd1498Szrj if (!is_gimple_assign (def)
165*38fd1498Szrj || (gimple_assign_single_p (def)
166*38fd1498Szrj && gimple_vuse (def)))
167*38fd1498Szrj return true;
168*38fd1498Szrj
169*38fd1498Szrj /* Check that any SSA names used to define NAME are also fully
170*38fd1498Szrj defined. */
171*38fd1498Szrj use_operand_p use_p;
172*38fd1498Szrj ssa_op_iter iter;
173*38fd1498Szrj FOR_EACH_SSA_USE_OPERAND (use_p, def, iter, SSA_OP_USE)
174*38fd1498Szrj {
175*38fd1498Szrj tree t = USE_FROM_PTR (use_p);
176*38fd1498Szrj /* If an SSA has already been seen, it may be a loop,
177*38fd1498Szrj but we can continue and ignore this use. Otherwise,
178*38fd1498Szrj add the SSA_NAME to the queue and visit it later. */
179*38fd1498Szrj if (bitmap_set_bit (visited_ssa, SSA_NAME_VERSION (t)))
180*38fd1498Szrj worklist.safe_push (t);
181*38fd1498Szrj }
182*38fd1498Szrj }
183*38fd1498Szrj return false;
184*38fd1498Szrj }
185*38fd1498Szrj
186*38fd1498Szrj /* Checks whether we can unswitch LOOP on condition at end of BB -- one of its
187*38fd1498Szrj basic blocks (for what it means see comments below). */
188*38fd1498Szrj
189*38fd1498Szrj static tree
tree_may_unswitch_on(basic_block bb,struct loop * loop)190*38fd1498Szrj tree_may_unswitch_on (basic_block bb, struct loop *loop)
191*38fd1498Szrj {
192*38fd1498Szrj gimple *last, *def;
193*38fd1498Szrj gcond *stmt;
194*38fd1498Szrj tree cond, use;
195*38fd1498Szrj basic_block def_bb;
196*38fd1498Szrj ssa_op_iter iter;
197*38fd1498Szrj
198*38fd1498Szrj /* BB must end in a simple conditional jump. */
199*38fd1498Szrj last = last_stmt (bb);
200*38fd1498Szrj if (!last || gimple_code (last) != GIMPLE_COND)
201*38fd1498Szrj return NULL_TREE;
202*38fd1498Szrj stmt = as_a <gcond *> (last);
203*38fd1498Szrj
204*38fd1498Szrj /* To keep the things simple, we do not directly remove the conditions,
205*38fd1498Szrj but just replace tests with 0 != 0 resp. 1 != 0. Prevent the infinite
206*38fd1498Szrj loop where we would unswitch again on such a condition. */
207*38fd1498Szrj if (gimple_cond_true_p (stmt) || gimple_cond_false_p (stmt))
208*38fd1498Szrj return NULL_TREE;
209*38fd1498Szrj
210*38fd1498Szrj /* Condition must be invariant. */
211*38fd1498Szrj FOR_EACH_SSA_TREE_OPERAND (use, stmt, iter, SSA_OP_USE)
212*38fd1498Szrj {
213*38fd1498Szrj def = SSA_NAME_DEF_STMT (use);
214*38fd1498Szrj def_bb = gimple_bb (def);
215*38fd1498Szrj if (def_bb
216*38fd1498Szrj && flow_bb_inside_loop_p (loop, def_bb))
217*38fd1498Szrj return NULL_TREE;
218*38fd1498Szrj /* Unswitching on undefined values would introduce undefined
219*38fd1498Szrj behavior that the original program might never exercise. */
220*38fd1498Szrj if (is_maybe_undefined (use, stmt, loop))
221*38fd1498Szrj return NULL_TREE;
222*38fd1498Szrj }
223*38fd1498Szrj
224*38fd1498Szrj cond = build2 (gimple_cond_code (stmt), boolean_type_node,
225*38fd1498Szrj gimple_cond_lhs (stmt), gimple_cond_rhs (stmt));
226*38fd1498Szrj
227*38fd1498Szrj return cond;
228*38fd1498Szrj }
229*38fd1498Szrj
230*38fd1498Szrj /* Simplifies COND using checks in front of the entry of the LOOP. Just very
231*38fd1498Szrj simplish (sufficient to prevent us from duplicating loop in unswitching
232*38fd1498Szrj unnecessarily). */
233*38fd1498Szrj
234*38fd1498Szrj static tree
simplify_using_entry_checks(struct loop * loop,tree cond)235*38fd1498Szrj simplify_using_entry_checks (struct loop *loop, tree cond)
236*38fd1498Szrj {
237*38fd1498Szrj edge e = loop_preheader_edge (loop);
238*38fd1498Szrj gimple *stmt;
239*38fd1498Szrj
240*38fd1498Szrj while (1)
241*38fd1498Szrj {
242*38fd1498Szrj stmt = last_stmt (e->src);
243*38fd1498Szrj if (stmt
244*38fd1498Szrj && gimple_code (stmt) == GIMPLE_COND
245*38fd1498Szrj && gimple_cond_code (stmt) == TREE_CODE (cond)
246*38fd1498Szrj && operand_equal_p (gimple_cond_lhs (stmt),
247*38fd1498Szrj TREE_OPERAND (cond, 0), 0)
248*38fd1498Szrj && operand_equal_p (gimple_cond_rhs (stmt),
249*38fd1498Szrj TREE_OPERAND (cond, 1), 0))
250*38fd1498Szrj return (e->flags & EDGE_TRUE_VALUE
251*38fd1498Szrj ? boolean_true_node
252*38fd1498Szrj : boolean_false_node);
253*38fd1498Szrj
254*38fd1498Szrj if (!single_pred_p (e->src))
255*38fd1498Szrj return cond;
256*38fd1498Szrj
257*38fd1498Szrj e = single_pred_edge (e->src);
258*38fd1498Szrj if (e->src == ENTRY_BLOCK_PTR_FOR_FN (cfun))
259*38fd1498Szrj return cond;
260*38fd1498Szrj }
261*38fd1498Szrj }
262*38fd1498Szrj
263*38fd1498Szrj /* Unswitch single LOOP. NUM is number of unswitchings done; we do not allow
264*38fd1498Szrj it to grow too much, it is too easy to create example on that the code would
265*38fd1498Szrj grow exponentially. */
266*38fd1498Szrj
267*38fd1498Szrj static bool
tree_unswitch_single_loop(struct loop * loop,int num)268*38fd1498Szrj tree_unswitch_single_loop (struct loop *loop, int num)
269*38fd1498Szrj {
270*38fd1498Szrj basic_block *bbs;
271*38fd1498Szrj struct loop *nloop;
272*38fd1498Szrj unsigned i, found;
273*38fd1498Szrj tree cond = NULL_TREE;
274*38fd1498Szrj gimple *stmt;
275*38fd1498Szrj bool changed = false;
276*38fd1498Szrj HOST_WIDE_INT iterations;
277*38fd1498Szrj
278*38fd1498Szrj /* Perform initial tests if unswitch is eligible. */
279*38fd1498Szrj if (num == 0)
280*38fd1498Szrj {
281*38fd1498Szrj /* Do not unswitch in cold regions. */
282*38fd1498Szrj if (optimize_loop_for_size_p (loop))
283*38fd1498Szrj {
284*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
285*38fd1498Szrj fprintf (dump_file, ";; Not unswitching cold loops\n");
286*38fd1498Szrj return false;
287*38fd1498Szrj }
288*38fd1498Szrj
289*38fd1498Szrj /* The loop should not be too large, to limit code growth. */
290*38fd1498Szrj if (tree_num_loop_insns (loop, &eni_size_weights)
291*38fd1498Szrj > (unsigned) PARAM_VALUE (PARAM_MAX_UNSWITCH_INSNS))
292*38fd1498Szrj {
293*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
294*38fd1498Szrj fprintf (dump_file, ";; Not unswitching, loop too big\n");
295*38fd1498Szrj return false;
296*38fd1498Szrj }
297*38fd1498Szrj
298*38fd1498Szrj /* If the loop is not expected to iterate, there is no need
299*38fd1498Szrj for unswitching. */
300*38fd1498Szrj iterations = estimated_loop_iterations_int (loop);
301*38fd1498Szrj if (iterations < 0)
302*38fd1498Szrj iterations = likely_max_loop_iterations_int (loop);
303*38fd1498Szrj if (iterations >= 0 && iterations <= 1)
304*38fd1498Szrj {
305*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
306*38fd1498Szrj fprintf (dump_file, ";; Not unswitching, loop is not expected"
307*38fd1498Szrj " to iterate\n");
308*38fd1498Szrj return false;
309*38fd1498Szrj }
310*38fd1498Szrj }
311*38fd1498Szrj
312*38fd1498Szrj i = 0;
313*38fd1498Szrj bbs = get_loop_body (loop);
314*38fd1498Szrj found = loop->num_nodes;
315*38fd1498Szrj
316*38fd1498Szrj while (1)
317*38fd1498Szrj {
318*38fd1498Szrj /* Find a bb to unswitch on. */
319*38fd1498Szrj for (; i < loop->num_nodes; i++)
320*38fd1498Szrj if ((cond = tree_may_unswitch_on (bbs[i], loop)))
321*38fd1498Szrj break;
322*38fd1498Szrj
323*38fd1498Szrj if (i == loop->num_nodes)
324*38fd1498Szrj {
325*38fd1498Szrj if (dump_file
326*38fd1498Szrj && num > PARAM_VALUE (PARAM_MAX_UNSWITCH_LEVEL)
327*38fd1498Szrj && (dump_flags & TDF_DETAILS))
328*38fd1498Szrj fprintf (dump_file, ";; Not unswitching anymore, hit max level\n");
329*38fd1498Szrj
330*38fd1498Szrj if (found == loop->num_nodes)
331*38fd1498Szrj {
332*38fd1498Szrj free (bbs);
333*38fd1498Szrj return changed;
334*38fd1498Szrj }
335*38fd1498Szrj break;
336*38fd1498Szrj }
337*38fd1498Szrj
338*38fd1498Szrj cond = simplify_using_entry_checks (loop, cond);
339*38fd1498Szrj stmt = last_stmt (bbs[i]);
340*38fd1498Szrj if (integer_nonzerop (cond))
341*38fd1498Szrj {
342*38fd1498Szrj /* Remove false path. */
343*38fd1498Szrj gimple_cond_set_condition_from_tree (as_a <gcond *> (stmt),
344*38fd1498Szrj boolean_true_node);
345*38fd1498Szrj changed = true;
346*38fd1498Szrj }
347*38fd1498Szrj else if (integer_zerop (cond))
348*38fd1498Szrj {
349*38fd1498Szrj /* Remove true path. */
350*38fd1498Szrj gimple_cond_set_condition_from_tree (as_a <gcond *> (stmt),
351*38fd1498Szrj boolean_false_node);
352*38fd1498Szrj changed = true;
353*38fd1498Szrj }
354*38fd1498Szrj /* Do not unswitch too much. */
355*38fd1498Szrj else if (num > PARAM_VALUE (PARAM_MAX_UNSWITCH_LEVEL))
356*38fd1498Szrj {
357*38fd1498Szrj i++;
358*38fd1498Szrj continue;
359*38fd1498Szrj }
360*38fd1498Szrj /* In nested tree_unswitch_single_loop first optimize all conditions
361*38fd1498Szrj using entry checks, then discover still reachable blocks in the
362*38fd1498Szrj loop and find the condition only among those still reachable bbs. */
363*38fd1498Szrj else if (num != 0)
364*38fd1498Szrj {
365*38fd1498Szrj if (found == loop->num_nodes)
366*38fd1498Szrj found = i;
367*38fd1498Szrj i++;
368*38fd1498Szrj continue;
369*38fd1498Szrj }
370*38fd1498Szrj else
371*38fd1498Szrj {
372*38fd1498Szrj found = i;
373*38fd1498Szrj break;
374*38fd1498Szrj }
375*38fd1498Szrj
376*38fd1498Szrj update_stmt (stmt);
377*38fd1498Szrj i++;
378*38fd1498Szrj }
379*38fd1498Szrj
380*38fd1498Szrj if (num != 0)
381*38fd1498Szrj {
382*38fd1498Szrj basic_block *tos, *worklist;
383*38fd1498Szrj
384*38fd1498Szrj /* When called recursively, first do a quick discovery
385*38fd1498Szrj of reachable bbs after the above changes and only
386*38fd1498Szrj consider conditions in still reachable bbs. */
387*38fd1498Szrj tos = worklist = XNEWVEC (basic_block, loop->num_nodes);
388*38fd1498Szrj
389*38fd1498Szrj for (i = 0; i < loop->num_nodes; i++)
390*38fd1498Szrj bbs[i]->flags &= ~BB_REACHABLE;
391*38fd1498Szrj
392*38fd1498Szrj /* Start with marking header. */
393*38fd1498Szrj *tos++ = bbs[0];
394*38fd1498Szrj bbs[0]->flags |= BB_REACHABLE;
395*38fd1498Szrj
396*38fd1498Szrj /* Iterate: find everything reachable from what we've already seen
397*38fd1498Szrj within the same innermost loop. Don't look through false edges
398*38fd1498Szrj if condition is always true or true edges if condition is
399*38fd1498Szrj always false. */
400*38fd1498Szrj while (tos != worklist)
401*38fd1498Szrj {
402*38fd1498Szrj basic_block b = *--tos;
403*38fd1498Szrj edge e;
404*38fd1498Szrj edge_iterator ei;
405*38fd1498Szrj int flags = 0;
406*38fd1498Szrj
407*38fd1498Szrj if (EDGE_COUNT (b->succs) == 2)
408*38fd1498Szrj {
409*38fd1498Szrj gimple *stmt = last_stmt (b);
410*38fd1498Szrj if (stmt
411*38fd1498Szrj && gimple_code (stmt) == GIMPLE_COND)
412*38fd1498Szrj {
413*38fd1498Szrj gcond *cond_stmt = as_a <gcond *> (stmt);
414*38fd1498Szrj if (gimple_cond_true_p (cond_stmt))
415*38fd1498Szrj flags = EDGE_FALSE_VALUE;
416*38fd1498Szrj else if (gimple_cond_false_p (cond_stmt))
417*38fd1498Szrj flags = EDGE_TRUE_VALUE;
418*38fd1498Szrj }
419*38fd1498Szrj }
420*38fd1498Szrj
421*38fd1498Szrj FOR_EACH_EDGE (e, ei, b->succs)
422*38fd1498Szrj {
423*38fd1498Szrj basic_block dest = e->dest;
424*38fd1498Szrj
425*38fd1498Szrj if (dest->loop_father == loop
426*38fd1498Szrj && !(dest->flags & BB_REACHABLE)
427*38fd1498Szrj && !(e->flags & flags))
428*38fd1498Szrj {
429*38fd1498Szrj *tos++ = dest;
430*38fd1498Szrj dest->flags |= BB_REACHABLE;
431*38fd1498Szrj }
432*38fd1498Szrj }
433*38fd1498Szrj }
434*38fd1498Szrj
435*38fd1498Szrj free (worklist);
436*38fd1498Szrj
437*38fd1498Szrj /* Find a bb to unswitch on. */
438*38fd1498Szrj for (; found < loop->num_nodes; found++)
439*38fd1498Szrj if ((bbs[found]->flags & BB_REACHABLE)
440*38fd1498Szrj && (cond = tree_may_unswitch_on (bbs[found], loop)))
441*38fd1498Szrj break;
442*38fd1498Szrj
443*38fd1498Szrj if (found == loop->num_nodes)
444*38fd1498Szrj {
445*38fd1498Szrj free (bbs);
446*38fd1498Szrj return changed;
447*38fd1498Szrj }
448*38fd1498Szrj }
449*38fd1498Szrj
450*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
451*38fd1498Szrj fprintf (dump_file, ";; Unswitching loop\n");
452*38fd1498Szrj
453*38fd1498Szrj initialize_original_copy_tables ();
454*38fd1498Szrj /* Unswitch the loop on this condition. */
455*38fd1498Szrj nloop = tree_unswitch_loop (loop, bbs[found], cond);
456*38fd1498Szrj if (!nloop)
457*38fd1498Szrj {
458*38fd1498Szrj free_original_copy_tables ();
459*38fd1498Szrj free (bbs);
460*38fd1498Szrj return changed;
461*38fd1498Szrj }
462*38fd1498Szrj
463*38fd1498Szrj /* Update the SSA form after unswitching. */
464*38fd1498Szrj update_ssa (TODO_update_ssa);
465*38fd1498Szrj free_original_copy_tables ();
466*38fd1498Szrj
467*38fd1498Szrj /* Invoke itself on modified loops. */
468*38fd1498Szrj tree_unswitch_single_loop (nloop, num + 1);
469*38fd1498Szrj tree_unswitch_single_loop (loop, num + 1);
470*38fd1498Szrj free (bbs);
471*38fd1498Szrj return true;
472*38fd1498Szrj }
473*38fd1498Szrj
474*38fd1498Szrj /* Unswitch a LOOP w.r. to given basic block UNSWITCH_ON. We only support
475*38fd1498Szrj unswitching of innermost loops. COND is the condition determining which
476*38fd1498Szrj loop is entered -- the new loop is entered if COND is true. Returns NULL
477*38fd1498Szrj if impossible, new loop otherwise. */
478*38fd1498Szrj
479*38fd1498Szrj static struct loop *
tree_unswitch_loop(struct loop * loop,basic_block unswitch_on,tree cond)480*38fd1498Szrj tree_unswitch_loop (struct loop *loop,
481*38fd1498Szrj basic_block unswitch_on, tree cond)
482*38fd1498Szrj {
483*38fd1498Szrj profile_probability prob_true;
484*38fd1498Szrj edge edge_true, edge_false;
485*38fd1498Szrj
486*38fd1498Szrj /* Some sanity checking. */
487*38fd1498Szrj gcc_assert (flow_bb_inside_loop_p (loop, unswitch_on));
488*38fd1498Szrj gcc_assert (EDGE_COUNT (unswitch_on->succs) == 2);
489*38fd1498Szrj gcc_assert (loop->inner == NULL);
490*38fd1498Szrj
491*38fd1498Szrj extract_true_false_edges_from_block (unswitch_on, &edge_true, &edge_false);
492*38fd1498Szrj prob_true = edge_true->probability;
493*38fd1498Szrj return loop_version (loop, unshare_expr (cond),
494*38fd1498Szrj NULL, prob_true,
495*38fd1498Szrj prob_true.invert (),
496*38fd1498Szrj prob_true, prob_true.invert (),
497*38fd1498Szrj false);
498*38fd1498Szrj }
499*38fd1498Szrj
500*38fd1498Szrj /* Unswitch outer loops by hoisting invariant guard on
501*38fd1498Szrj inner loop without code duplication. */
502*38fd1498Szrj static bool
tree_unswitch_outer_loop(struct loop * loop)503*38fd1498Szrj tree_unswitch_outer_loop (struct loop *loop)
504*38fd1498Szrj {
505*38fd1498Szrj edge exit, guard;
506*38fd1498Szrj HOST_WIDE_INT iterations;
507*38fd1498Szrj
508*38fd1498Szrj gcc_assert (loop->inner);
509*38fd1498Szrj if (loop->inner->next)
510*38fd1498Szrj return false;
511*38fd1498Szrj /* Accept loops with single exit only which is not from inner loop. */
512*38fd1498Szrj exit = single_exit (loop);
513*38fd1498Szrj if (!exit || exit->src->loop_father != loop)
514*38fd1498Szrj return false;
515*38fd1498Szrj /* Check that phi argument of exit edge is not defined inside loop. */
516*38fd1498Szrj if (!check_exit_phi (loop))
517*38fd1498Szrj return false;
518*38fd1498Szrj /* If the loop is not expected to iterate, there is no need
519*38fd1498Szrj for unswitching. */
520*38fd1498Szrj iterations = estimated_loop_iterations_int (loop);
521*38fd1498Szrj if (iterations < 0)
522*38fd1498Szrj iterations = likely_max_loop_iterations_int (loop);
523*38fd1498Szrj if (iterations >= 0 && iterations <= 1)
524*38fd1498Szrj {
525*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
526*38fd1498Szrj fprintf (dump_file, ";; Not unswitching, loop is not expected"
527*38fd1498Szrj " to iterate\n");
528*38fd1498Szrj return false;
529*38fd1498Szrj }
530*38fd1498Szrj
531*38fd1498Szrj bool changed = false;
532*38fd1498Szrj while ((guard = find_loop_guard (loop)))
533*38fd1498Szrj {
534*38fd1498Szrj if (! changed)
535*38fd1498Szrj rewrite_virtuals_into_loop_closed_ssa (loop);
536*38fd1498Szrj hoist_guard (loop, guard);
537*38fd1498Szrj changed = true;
538*38fd1498Szrj }
539*38fd1498Szrj return changed;
540*38fd1498Szrj }
541*38fd1498Szrj
542*38fd1498Szrj /* Checks if the body of the LOOP is within an invariant guard. If this
543*38fd1498Szrj is the case, returns the edge that jumps over the real body of the loop,
544*38fd1498Szrj otherwise returns NULL. */
545*38fd1498Szrj
546*38fd1498Szrj static edge
find_loop_guard(struct loop * loop)547*38fd1498Szrj find_loop_guard (struct loop *loop)
548*38fd1498Szrj {
549*38fd1498Szrj basic_block header = loop->header;
550*38fd1498Szrj edge guard_edge, te, fe;
551*38fd1498Szrj basic_block *body = NULL;
552*38fd1498Szrj unsigned i;
553*38fd1498Szrj tree use;
554*38fd1498Szrj ssa_op_iter iter;
555*38fd1498Szrj
556*38fd1498Szrj /* We check for the following situation:
557*38fd1498Szrj
558*38fd1498Szrj while (1)
559*38fd1498Szrj {
560*38fd1498Szrj [header]]
561*38fd1498Szrj loop_phi_nodes;
562*38fd1498Szrj something1;
563*38fd1498Szrj if (cond1)
564*38fd1498Szrj body;
565*38fd1498Szrj nvar = phi(orig, bvar) ... for all variables changed in body;
566*38fd1498Szrj [guard_end]
567*38fd1498Szrj something2;
568*38fd1498Szrj if (cond2)
569*38fd1498Szrj break;
570*38fd1498Szrj something3;
571*38fd1498Szrj }
572*38fd1498Szrj
573*38fd1498Szrj where:
574*38fd1498Szrj
575*38fd1498Szrj 1) cond1 is loop invariant
576*38fd1498Szrj 2) If cond1 is false, then the loop is essentially empty; i.e.,
577*38fd1498Szrj a) nothing in something1, something2 and something3 has side
578*38fd1498Szrj effects
579*38fd1498Szrj b) anything defined in something1, something2 and something3
580*38fd1498Szrj is not used outside of the loop. */
581*38fd1498Szrj
582*38fd1498Szrj gcond *cond;
583*38fd1498Szrj do
584*38fd1498Szrj {
585*38fd1498Szrj basic_block next = NULL;
586*38fd1498Szrj if (single_succ_p (header))
587*38fd1498Szrj next = single_succ (header);
588*38fd1498Szrj else
589*38fd1498Szrj {
590*38fd1498Szrj cond = dyn_cast <gcond *> (last_stmt (header));
591*38fd1498Szrj if (! cond)
592*38fd1498Szrj return NULL;
593*38fd1498Szrj extract_true_false_edges_from_block (header, &te, &fe);
594*38fd1498Szrj /* Make sure to skip earlier hoisted guards that are left
595*38fd1498Szrj in place as if (true). */
596*38fd1498Szrj if (gimple_cond_true_p (cond))
597*38fd1498Szrj next = te->dest;
598*38fd1498Szrj else if (gimple_cond_false_p (cond))
599*38fd1498Szrj next = fe->dest;
600*38fd1498Szrj else
601*38fd1498Szrj break;
602*38fd1498Szrj }
603*38fd1498Szrj /* Never traverse a backedge. */
604*38fd1498Szrj if (header->loop_father->header == next)
605*38fd1498Szrj return NULL;
606*38fd1498Szrj header = next;
607*38fd1498Szrj }
608*38fd1498Szrj while (1);
609*38fd1498Szrj if (!flow_bb_inside_loop_p (loop, te->dest)
610*38fd1498Szrj || !flow_bb_inside_loop_p (loop, fe->dest))
611*38fd1498Szrj return NULL;
612*38fd1498Szrj
613*38fd1498Szrj if (just_once_each_iteration_p (loop, te->dest)
614*38fd1498Szrj || (single_succ_p (te->dest)
615*38fd1498Szrj && just_once_each_iteration_p (loop, single_succ (te->dest))))
616*38fd1498Szrj {
617*38fd1498Szrj if (just_once_each_iteration_p (loop, fe->dest))
618*38fd1498Szrj return NULL;
619*38fd1498Szrj guard_edge = te;
620*38fd1498Szrj }
621*38fd1498Szrj else if (just_once_each_iteration_p (loop, fe->dest)
622*38fd1498Szrj || (single_succ_p (fe->dest)
623*38fd1498Szrj && just_once_each_iteration_p (loop, single_succ (fe->dest))))
624*38fd1498Szrj guard_edge = fe;
625*38fd1498Szrj else
626*38fd1498Szrj return NULL;
627*38fd1498Szrj
628*38fd1498Szrj /* Guard edge must skip inner loop. */
629*38fd1498Szrj if (!dominated_by_p (CDI_DOMINATORS, loop->inner->header,
630*38fd1498Szrj guard_edge == fe ? te->dest : fe->dest))
631*38fd1498Szrj {
632*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
633*38fd1498Szrj fprintf (dump_file, "Guard edge %d --> %d is not around the loop!\n",
634*38fd1498Szrj guard_edge->src->index, guard_edge->dest->index);
635*38fd1498Szrj return NULL;
636*38fd1498Szrj }
637*38fd1498Szrj if (guard_edge->dest == loop->latch)
638*38fd1498Szrj {
639*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
640*38fd1498Szrj fprintf (dump_file, "Guard edge destination is loop latch.\n");
641*38fd1498Szrj return NULL;
642*38fd1498Szrj }
643*38fd1498Szrj
644*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
645*38fd1498Szrj fprintf (dump_file,
646*38fd1498Szrj "Considering guard %d -> %d in loop %d\n",
647*38fd1498Szrj guard_edge->src->index, guard_edge->dest->index, loop->num);
648*38fd1498Szrj /* Check if condition operands do not have definitions inside loop since
649*38fd1498Szrj any bb copying is not performed. */
650*38fd1498Szrj FOR_EACH_SSA_TREE_OPERAND (use, cond, iter, SSA_OP_USE)
651*38fd1498Szrj {
652*38fd1498Szrj gimple *def = SSA_NAME_DEF_STMT (use);
653*38fd1498Szrj basic_block def_bb = gimple_bb (def);
654*38fd1498Szrj if (def_bb
655*38fd1498Szrj && flow_bb_inside_loop_p (loop, def_bb))
656*38fd1498Szrj {
657*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
658*38fd1498Szrj fprintf (dump_file, " guard operands have definitions"
659*38fd1498Szrj " inside loop\n");
660*38fd1498Szrj return NULL;
661*38fd1498Szrj }
662*38fd1498Szrj }
663*38fd1498Szrj
664*38fd1498Szrj body = get_loop_body (loop);
665*38fd1498Szrj for (i = 0; i < loop->num_nodes; i++)
666*38fd1498Szrj {
667*38fd1498Szrj basic_block bb = body[i];
668*38fd1498Szrj if (bb->loop_father != loop)
669*38fd1498Szrj continue;
670*38fd1498Szrj if (bb->flags & BB_IRREDUCIBLE_LOOP)
671*38fd1498Szrj {
672*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
673*38fd1498Szrj fprintf (dump_file, "Block %d is marked as irreducible in loop\n",
674*38fd1498Szrj bb->index);
675*38fd1498Szrj guard_edge = NULL;
676*38fd1498Szrj goto end;
677*38fd1498Szrj }
678*38fd1498Szrj if (!empty_bb_without_guard_p (loop, bb))
679*38fd1498Szrj {
680*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
681*38fd1498Szrj fprintf (dump_file, " block %d has side effects\n", bb->index);
682*38fd1498Szrj guard_edge = NULL;
683*38fd1498Szrj goto end;
684*38fd1498Szrj }
685*38fd1498Szrj }
686*38fd1498Szrj
687*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
688*38fd1498Szrj fprintf (dump_file, " suitable to hoist\n");
689*38fd1498Szrj end:
690*38fd1498Szrj if (body)
691*38fd1498Szrj free (body);
692*38fd1498Szrj return guard_edge;
693*38fd1498Szrj }
694*38fd1498Szrj
695*38fd1498Szrj /* Returns true if
696*38fd1498Szrj 1) no statement in BB has side effects
697*38fd1498Szrj 2) assuming that edge GUARD is always taken, all definitions in BB
698*38fd1498Szrj are noy used outside of the loop.
699*38fd1498Szrj KNOWN_INVARIANTS is a set of ssa names we know to be invariant, and
700*38fd1498Szrj PROCESSED is a set of ssa names for that we already tested whether they
701*38fd1498Szrj are invariant or not. */
702*38fd1498Szrj
703*38fd1498Szrj static bool
empty_bb_without_guard_p(struct loop * loop,basic_block bb)704*38fd1498Szrj empty_bb_without_guard_p (struct loop *loop, basic_block bb)
705*38fd1498Szrj {
706*38fd1498Szrj basic_block exit_bb = single_exit (loop)->src;
707*38fd1498Szrj bool may_be_used_outside = (bb == exit_bb
708*38fd1498Szrj || !dominated_by_p (CDI_DOMINATORS, bb, exit_bb));
709*38fd1498Szrj tree name;
710*38fd1498Szrj ssa_op_iter op_iter;
711*38fd1498Szrj
712*38fd1498Szrj /* Phi nodes do not have side effects, but their results might be used
713*38fd1498Szrj outside of the loop. */
714*38fd1498Szrj if (may_be_used_outside)
715*38fd1498Szrj {
716*38fd1498Szrj for (gphi_iterator gsi = gsi_start_phis (bb);
717*38fd1498Szrj !gsi_end_p (gsi); gsi_next (&gsi))
718*38fd1498Szrj {
719*38fd1498Szrj gphi *phi = gsi.phi ();
720*38fd1498Szrj name = PHI_RESULT (phi);
721*38fd1498Szrj if (virtual_operand_p (name))
722*38fd1498Szrj continue;
723*38fd1498Szrj
724*38fd1498Szrj if (used_outside_loop_p (loop, name))
725*38fd1498Szrj return false;
726*38fd1498Szrj }
727*38fd1498Szrj }
728*38fd1498Szrj
729*38fd1498Szrj for (gimple_stmt_iterator gsi = gsi_start_bb (bb);
730*38fd1498Szrj !gsi_end_p (gsi); gsi_next (&gsi))
731*38fd1498Szrj {
732*38fd1498Szrj gimple *stmt = gsi_stmt (gsi);
733*38fd1498Szrj if (gimple_has_side_effects (stmt))
734*38fd1498Szrj return false;
735*38fd1498Szrj
736*38fd1498Szrj if (gimple_vdef(stmt))
737*38fd1498Szrj return false;
738*38fd1498Szrj
739*38fd1498Szrj FOR_EACH_SSA_TREE_OPERAND (name, stmt, op_iter, SSA_OP_DEF)
740*38fd1498Szrj {
741*38fd1498Szrj if (may_be_used_outside
742*38fd1498Szrj && used_outside_loop_p (loop, name))
743*38fd1498Szrj return false;
744*38fd1498Szrj }
745*38fd1498Szrj }
746*38fd1498Szrj return true;
747*38fd1498Szrj }
748*38fd1498Szrj
749*38fd1498Szrj /* Return true if NAME is used outside of LOOP. */
750*38fd1498Szrj
751*38fd1498Szrj static bool
used_outside_loop_p(struct loop * loop,tree name)752*38fd1498Szrj used_outside_loop_p (struct loop *loop, tree name)
753*38fd1498Szrj {
754*38fd1498Szrj imm_use_iterator it;
755*38fd1498Szrj use_operand_p use;
756*38fd1498Szrj
757*38fd1498Szrj FOR_EACH_IMM_USE_FAST (use, it, name)
758*38fd1498Szrj {
759*38fd1498Szrj gimple *stmt = USE_STMT (use);
760*38fd1498Szrj if (!flow_bb_inside_loop_p (loop, gimple_bb (stmt)))
761*38fd1498Szrj return true;
762*38fd1498Szrj }
763*38fd1498Szrj
764*38fd1498Szrj return false;
765*38fd1498Szrj }
766*38fd1498Szrj
767*38fd1498Szrj /* Return argument for loop preheader edge in header virtual phi if any. */
768*38fd1498Szrj
769*38fd1498Szrj static tree
get_vop_from_header(struct loop * loop)770*38fd1498Szrj get_vop_from_header (struct loop *loop)
771*38fd1498Szrj {
772*38fd1498Szrj for (gphi_iterator gsi = gsi_start_phis (loop->header);
773*38fd1498Szrj !gsi_end_p (gsi); gsi_next (&gsi))
774*38fd1498Szrj {
775*38fd1498Szrj gphi *phi = gsi.phi ();
776*38fd1498Szrj if (!virtual_operand_p (gimple_phi_result (phi)))
777*38fd1498Szrj continue;
778*38fd1498Szrj return PHI_ARG_DEF_FROM_EDGE (phi, loop_preheader_edge (loop));
779*38fd1498Szrj }
780*38fd1498Szrj return NULL_TREE;
781*38fd1498Szrj }
782*38fd1498Szrj
783*38fd1498Szrj /* Move the check of GUARD outside of LOOP. */
784*38fd1498Szrj
785*38fd1498Szrj static void
hoist_guard(struct loop * loop,edge guard)786*38fd1498Szrj hoist_guard (struct loop *loop, edge guard)
787*38fd1498Szrj {
788*38fd1498Szrj edge exit = single_exit (loop);
789*38fd1498Szrj edge preh = loop_preheader_edge (loop);
790*38fd1498Szrj basic_block pre_header = preh->src;
791*38fd1498Szrj basic_block bb;
792*38fd1498Szrj edge te, fe, e, new_edge;
793*38fd1498Szrj gimple *stmt;
794*38fd1498Szrj basic_block guard_bb = guard->src;
795*38fd1498Szrj edge not_guard;
796*38fd1498Szrj gimple_stmt_iterator gsi;
797*38fd1498Szrj int flags = 0;
798*38fd1498Szrj bool fix_dom_of_exit;
799*38fd1498Szrj gcond *cond_stmt, *new_cond_stmt;
800*38fd1498Szrj
801*38fd1498Szrj bb = get_immediate_dominator (CDI_DOMINATORS, exit->dest);
802*38fd1498Szrj fix_dom_of_exit = flow_bb_inside_loop_p (loop, bb);
803*38fd1498Szrj gsi = gsi_last_bb (guard_bb);
804*38fd1498Szrj stmt = gsi_stmt (gsi);
805*38fd1498Szrj gcc_assert (gimple_code (stmt) == GIMPLE_COND);
806*38fd1498Szrj cond_stmt = as_a <gcond *> (stmt);
807*38fd1498Szrj extract_true_false_edges_from_block (guard_bb, &te, &fe);
808*38fd1498Szrj /* Insert guard to PRE_HEADER. */
809*38fd1498Szrj if (!empty_block_p (pre_header))
810*38fd1498Szrj gsi = gsi_last_bb (pre_header);
811*38fd1498Szrj else
812*38fd1498Szrj gsi = gsi_start_bb (pre_header);
813*38fd1498Szrj /* Create copy of COND_STMT. */
814*38fd1498Szrj new_cond_stmt = gimple_build_cond (gimple_cond_code (cond_stmt),
815*38fd1498Szrj gimple_cond_lhs (cond_stmt),
816*38fd1498Szrj gimple_cond_rhs (cond_stmt),
817*38fd1498Szrj NULL_TREE, NULL_TREE);
818*38fd1498Szrj gsi_insert_after (&gsi, new_cond_stmt, GSI_NEW_STMT);
819*38fd1498Szrj /* Convert COND_STMT to true/false conditional. */
820*38fd1498Szrj if (guard == te)
821*38fd1498Szrj gimple_cond_make_false (cond_stmt);
822*38fd1498Szrj else
823*38fd1498Szrj gimple_cond_make_true (cond_stmt);
824*38fd1498Szrj update_stmt (cond_stmt);
825*38fd1498Szrj /* Create new loop pre-header. */
826*38fd1498Szrj e = split_block (pre_header, last_stmt (pre_header));
827*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
828*38fd1498Szrj {
829*38fd1498Szrj fprintf (dump_file, " Moving guard %i->%i (prob ",
830*38fd1498Szrj guard->src->index, guard->dest->index);
831*38fd1498Szrj guard->probability.dump (dump_file);
832*38fd1498Szrj fprintf (dump_file, ") to bb %i, new preheader is %i\n",
833*38fd1498Szrj e->src->index, e->dest->index);
834*38fd1498Szrj }
835*38fd1498Szrj
836*38fd1498Szrj gcc_assert (loop_preheader_edge (loop)->src == e->dest);
837*38fd1498Szrj
838*38fd1498Szrj if (guard == fe)
839*38fd1498Szrj {
840*38fd1498Szrj e->flags = EDGE_TRUE_VALUE;
841*38fd1498Szrj flags |= EDGE_FALSE_VALUE;
842*38fd1498Szrj not_guard = te;
843*38fd1498Szrj }
844*38fd1498Szrj else
845*38fd1498Szrj {
846*38fd1498Szrj e->flags = EDGE_FALSE_VALUE;
847*38fd1498Szrj flags |= EDGE_TRUE_VALUE;
848*38fd1498Szrj not_guard = fe;
849*38fd1498Szrj }
850*38fd1498Szrj new_edge = make_edge (pre_header, exit->dest, flags);
851*38fd1498Szrj
852*38fd1498Szrj /* Determine the probability that we skip the loop. Assume that loop has
853*38fd1498Szrj same average number of iterations regardless outcome of guard. */
854*38fd1498Szrj new_edge->probability = guard->probability;
855*38fd1498Szrj profile_count skip_count = guard->src->count.nonzero_p ()
856*38fd1498Szrj ? guard->count ().apply_scale (pre_header->count,
857*38fd1498Szrj guard->src->count)
858*38fd1498Szrj : guard->count ().apply_probability (new_edge->probability);
859*38fd1498Szrj
860*38fd1498Szrj if (skip_count > e->count ())
861*38fd1498Szrj {
862*38fd1498Szrj fprintf (dump_file, " Capping count; expect profile inconsistency\n");
863*38fd1498Szrj skip_count = e->count ();
864*38fd1498Szrj }
865*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
866*38fd1498Szrj {
867*38fd1498Szrj fprintf (dump_file, " Estimated probability of skipping loop is ");
868*38fd1498Szrj new_edge->probability.dump (dump_file);
869*38fd1498Szrj fprintf (dump_file, "\n");
870*38fd1498Szrj }
871*38fd1498Szrj
872*38fd1498Szrj /* Update profile after the transform:
873*38fd1498Szrj
874*38fd1498Szrj First decrease count of path from newly hoisted loop guard
875*38fd1498Szrj to loop header... */
876*38fd1498Szrj e->probability = new_edge->probability.invert ();
877*38fd1498Szrj e->dest->count = e->count ();
878*38fd1498Szrj
879*38fd1498Szrj /* ... now update profile to represent that original guard will be optimized
880*38fd1498Szrj away ... */
881*38fd1498Szrj guard->probability = profile_probability::never ();
882*38fd1498Szrj not_guard->probability = profile_probability::always ();
883*38fd1498Szrj
884*38fd1498Szrj /* ... finally scale everything in the loop except for guarded basic blocks
885*38fd1498Szrj where profile does not change. */
886*38fd1498Szrj basic_block *body = get_loop_body (loop);
887*38fd1498Szrj
888*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
889*38fd1498Szrj fprintf (dump_file, " Scaling nonguarded BBs in loop:");
890*38fd1498Szrj for (unsigned int i = 0; i < loop->num_nodes; i++)
891*38fd1498Szrj {
892*38fd1498Szrj basic_block bb = body[i];
893*38fd1498Szrj if (!dominated_by_p (CDI_DOMINATORS, bb, not_guard->dest))
894*38fd1498Szrj {
895*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
896*38fd1498Szrj fprintf (dump_file, " %i", bb->index);
897*38fd1498Szrj if (e->probability.initialized_p ())
898*38fd1498Szrj scale_bbs_frequencies (&bb, 1, e->probability);
899*38fd1498Szrj }
900*38fd1498Szrj }
901*38fd1498Szrj
902*38fd1498Szrj if (fix_dom_of_exit)
903*38fd1498Szrj set_immediate_dominator (CDI_DOMINATORS, exit->dest, pre_header);
904*38fd1498Szrj /* Add NEW_ADGE argument for all phi in post-header block. */
905*38fd1498Szrj bb = exit->dest;
906*38fd1498Szrj for (gphi_iterator gsi = gsi_start_phis (bb);
907*38fd1498Szrj !gsi_end_p (gsi); gsi_next (&gsi))
908*38fd1498Szrj {
909*38fd1498Szrj gphi *phi = gsi.phi ();
910*38fd1498Szrj tree arg;
911*38fd1498Szrj if (virtual_operand_p (gimple_phi_result (phi)))
912*38fd1498Szrj {
913*38fd1498Szrj arg = get_vop_from_header (loop);
914*38fd1498Szrj if (arg == NULL_TREE)
915*38fd1498Szrj /* Use exit edge argument. */
916*38fd1498Szrj arg = PHI_ARG_DEF_FROM_EDGE (phi, exit);
917*38fd1498Szrj add_phi_arg (phi, arg, new_edge, UNKNOWN_LOCATION);
918*38fd1498Szrj }
919*38fd1498Szrj else
920*38fd1498Szrj {
921*38fd1498Szrj /* Use exit edge argument. */
922*38fd1498Szrj arg = PHI_ARG_DEF_FROM_EDGE (phi, exit);
923*38fd1498Szrj add_phi_arg (phi, arg, new_edge, UNKNOWN_LOCATION);
924*38fd1498Szrj }
925*38fd1498Szrj }
926*38fd1498Szrj
927*38fd1498Szrj if (dump_file && (dump_flags & TDF_DETAILS))
928*38fd1498Szrj fprintf (dump_file, "\n guard hoisted.\n");
929*38fd1498Szrj
930*38fd1498Szrj free (body);
931*38fd1498Szrj }
932*38fd1498Szrj
933*38fd1498Szrj /* Return true if phi argument for exit edge can be used
934*38fd1498Szrj for edge around loop. */
935*38fd1498Szrj
936*38fd1498Szrj static bool
check_exit_phi(struct loop * loop)937*38fd1498Szrj check_exit_phi (struct loop *loop)
938*38fd1498Szrj {
939*38fd1498Szrj edge exit = single_exit (loop);
940*38fd1498Szrj basic_block pre_header = loop_preheader_edge (loop)->src;
941*38fd1498Szrj
942*38fd1498Szrj for (gphi_iterator gsi = gsi_start_phis (exit->dest);
943*38fd1498Szrj !gsi_end_p (gsi); gsi_next (&gsi))
944*38fd1498Szrj {
945*38fd1498Szrj gphi *phi = gsi.phi ();
946*38fd1498Szrj tree arg;
947*38fd1498Szrj gimple *def;
948*38fd1498Szrj basic_block def_bb;
949*38fd1498Szrj if (virtual_operand_p (gimple_phi_result (phi)))
950*38fd1498Szrj continue;
951*38fd1498Szrj arg = PHI_ARG_DEF_FROM_EDGE (phi, exit);
952*38fd1498Szrj if (TREE_CODE (arg) != SSA_NAME)
953*38fd1498Szrj continue;
954*38fd1498Szrj def = SSA_NAME_DEF_STMT (arg);
955*38fd1498Szrj if (!def)
956*38fd1498Szrj continue;
957*38fd1498Szrj def_bb = gimple_bb (def);
958*38fd1498Szrj if (!def_bb)
959*38fd1498Szrj continue;
960*38fd1498Szrj if (!dominated_by_p (CDI_DOMINATORS, pre_header, def_bb))
961*38fd1498Szrj /* Definition inside loop! */
962*38fd1498Szrj return false;
963*38fd1498Szrj /* Check loop closed phi invariant. */
964*38fd1498Szrj if (!flow_bb_inside_loop_p (def_bb->loop_father, pre_header))
965*38fd1498Szrj return false;
966*38fd1498Szrj }
967*38fd1498Szrj return true;
968*38fd1498Szrj }
969*38fd1498Szrj
970*38fd1498Szrj /* Loop unswitching pass. */
971*38fd1498Szrj
972*38fd1498Szrj namespace {
973*38fd1498Szrj
974*38fd1498Szrj const pass_data pass_data_tree_unswitch =
975*38fd1498Szrj {
976*38fd1498Szrj GIMPLE_PASS, /* type */
977*38fd1498Szrj "unswitch", /* name */
978*38fd1498Szrj OPTGROUP_LOOP, /* optinfo_flags */
979*38fd1498Szrj TV_TREE_LOOP_UNSWITCH, /* tv_id */
980*38fd1498Szrj PROP_cfg, /* properties_required */
981*38fd1498Szrj 0, /* properties_provided */
982*38fd1498Szrj 0, /* properties_destroyed */
983*38fd1498Szrj 0, /* todo_flags_start */
984*38fd1498Szrj 0, /* todo_flags_finish */
985*38fd1498Szrj };
986*38fd1498Szrj
987*38fd1498Szrj class pass_tree_unswitch : public gimple_opt_pass
988*38fd1498Szrj {
989*38fd1498Szrj public:
pass_tree_unswitch(gcc::context * ctxt)990*38fd1498Szrj pass_tree_unswitch (gcc::context *ctxt)
991*38fd1498Szrj : gimple_opt_pass (pass_data_tree_unswitch, ctxt)
992*38fd1498Szrj {}
993*38fd1498Szrj
994*38fd1498Szrj /* opt_pass methods: */
gate(function *)995*38fd1498Szrj virtual bool gate (function *) { return flag_unswitch_loops != 0; }
996*38fd1498Szrj virtual unsigned int execute (function *);
997*38fd1498Szrj
998*38fd1498Szrj }; // class pass_tree_unswitch
999*38fd1498Szrj
1000*38fd1498Szrj unsigned int
execute(function * fun)1001*38fd1498Szrj pass_tree_unswitch::execute (function *fun)
1002*38fd1498Szrj {
1003*38fd1498Szrj if (number_of_loops (fun) <= 1)
1004*38fd1498Szrj return 0;
1005*38fd1498Szrj
1006*38fd1498Szrj return tree_ssa_unswitch_loops ();
1007*38fd1498Szrj }
1008*38fd1498Szrj
1009*38fd1498Szrj } // anon namespace
1010*38fd1498Szrj
1011*38fd1498Szrj gimple_opt_pass *
make_pass_tree_unswitch(gcc::context * ctxt)1012*38fd1498Szrj make_pass_tree_unswitch (gcc::context *ctxt)
1013*38fd1498Szrj {
1014*38fd1498Szrj return new pass_tree_unswitch (ctxt);
1015*38fd1498Szrj }
1016*38fd1498Szrj
1017*38fd1498Szrj
1018