]>
Commit | Line | Data |
---|---|---|
fbdec14e | 1 | /* Loop interchange. |
85ec4feb | 2 | Copyright (C) 2017-2018 Free Software Foundation, Inc. |
fbdec14e BC |
3 | Contributed by ARM Ltd. |
4 | ||
5 | This file is part of GCC. | |
6 | ||
7 | GCC is free software; you can redistribute it and/or modify it | |
8 | under the terms of the GNU General Public License as published by the | |
9 | Free Software Foundation; either version 3, or (at your option) any | |
10 | later version. | |
11 | ||
12 | GCC is distributed in the hope that it will be useful, but WITHOUT | |
13 | ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | |
14 | FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License | |
15 | for more details. | |
16 | ||
17 | You should have received a copy of the GNU General Public License | |
18 | along with GCC; see the file COPYING3. If not see | |
19 | <http://www.gnu.org/licenses/>. */ | |
20 | ||
21 | #include "config.h" | |
22 | #include "system.h" | |
23 | #include "coretypes.h" | |
24 | #include "backend.h" | |
25 | #include "is-a.h" | |
26 | #include "tree.h" | |
27 | #include "gimple.h" | |
28 | #include "tree-pass.h" | |
29 | #include "ssa.h" | |
30 | #include "gimple-pretty-print.h" | |
31 | #include "fold-const.h" | |
32 | #include "gimplify.h" | |
33 | #include "gimple-iterator.h" | |
34 | #include "gimplify-me.h" | |
35 | #include "cfgloop.h" | |
36 | #include "params.h" | |
37 | #include "tree-ssa.h" | |
38 | #include "tree-scalar-evolution.h" | |
39 | #include "tree-ssa-loop-manip.h" | |
40 | #include "tree-ssa-loop-niter.h" | |
41 | #include "tree-ssa-loop-ivopts.h" | |
42 | #include "tree-ssa-dce.h" | |
43 | #include "tree-data-ref.h" | |
44 | #include "tree-vectorizer.h" | |
45 | ||
46 | /* This pass performs loop interchange: for example, the loop nest | |
47 | ||
48 | for (int j = 0; j < N; j++) | |
49 | for (int k = 0; k < N; k++) | |
50 | for (int i = 0; i < N; i++) | |
51 | c[i][j] = c[i][j] + a[i][k]*b[k][j]; | |
52 | ||
53 | is transformed to | |
54 | ||
55 | for (int i = 0; i < N; i++) | |
56 | for (int j = 0; j < N; j++) | |
57 | for (int k = 0; k < N; k++) | |
58 | c[i][j] = c[i][j] + a[i][k]*b[k][j]; | |
59 | ||
60 | This pass implements loop interchange in the following steps: | |
61 | ||
62 | 1) Find perfect loop nest for each innermost loop and compute data | |
63 | dependence relations for it. For above example, loop nest is | |
64 | <loop_j, loop_k, loop_i>. | |
65 | 2) From innermost to outermost loop, this pass tries to interchange | |
66 | each loop pair. For above case, it firstly tries to interchange | |
67 | <loop_k, loop_i> and loop nest becomes <loop_j, loop_i, loop_k>. | |
68 | Then it tries to interchange <loop_j, loop_i> and loop nest becomes | |
69 | <loop_i, loop_j, loop_k>. The overall effect is to move innermost | |
70 | loop to the outermost position. For loop pair <loop_i, loop_j> | |
71 | to be interchanged, we: | |
72 | 3) Check if data dependence relations are valid for loop interchange. | |
73 | 4) Check if both loops can be interchanged in terms of transformation. | |
74 | 5) Check if interchanging the two loops is profitable. | |
75 | 6) Interchange the two loops by mapping induction variables. | |
76 | ||
77 | This pass also handles reductions in loop nest. So far we only support | |
78 | simple reduction of inner loop and double reduction of the loop nest. */ | |
79 | ||
80 | /* Maximum number of stmts in each loop that should be interchanged. */ | |
81 | #define MAX_NUM_STMT (PARAM_VALUE (PARAM_LOOP_INTERCHANGE_MAX_NUM_STMTS)) | |
82 | /* Maximum number of data references in loop nest. */ | |
83 | #define MAX_DATAREFS (PARAM_VALUE (PARAM_LOOP_MAX_DATAREFS_FOR_DATADEPS)) | |
84 | ||
85 | /* Comparison ratio of access stride between inner/outer loops to be | |
86 | interchanged. This is the minimum stride ratio for loop interchange | |
87 | to be profitable. */ | |
88 | #define OUTER_STRIDE_RATIO (PARAM_VALUE (PARAM_LOOP_INTERCHANGE_STRIDE_RATIO)) | |
89 | /* The same as above, but we require higher ratio for interchanging the | |
90 | innermost two loops. */ | |
91 | #define INNER_STRIDE_RATIO ((OUTER_STRIDE_RATIO) + 1) | |
92 | ||
1eeeda47 BC |
93 | /* Comparison ratio of stmt cost between inner/outer loops. Loops won't |
94 | be interchanged if outer loop has too many stmts. */ | |
95 | #define STMT_COST_RATIO (3) | |
96 | ||
fbdec14e BC |
97 | /* Vector of strides that DR accesses in each level loop of a loop nest. */ |
98 | #define DR_ACCESS_STRIDE(dr) ((vec<tree> *) dr->aux) | |
99 | ||
100 | /* Structure recording loop induction variable. */ | |
101 | typedef struct induction | |
102 | { | |
103 | /* IV itself. */ | |
104 | tree var; | |
105 | /* IV's initializing value, which is the init arg of the IV PHI node. */ | |
106 | tree init_val; | |
107 | /* IV's initializing expr, which is (the expanded result of) init_val. */ | |
108 | tree init_expr; | |
109 | /* IV's step. */ | |
110 | tree step; | |
111 | } *induction_p; | |
112 | ||
113 | /* Enum type for loop reduction variable. */ | |
114 | enum reduction_type | |
115 | { | |
116 | UNKNOWN_RTYPE = 0, | |
117 | SIMPLE_RTYPE, | |
118 | DOUBLE_RTYPE | |
119 | }; | |
120 | ||
121 | /* Structure recording loop reduction variable. */ | |
122 | typedef struct reduction | |
123 | { | |
124 | /* Reduction itself. */ | |
125 | tree var; | |
126 | /* PHI node defining reduction variable. */ | |
127 | gphi *phi; | |
128 | /* Init and next variables of the reduction. */ | |
129 | tree init; | |
130 | tree next; | |
131 | /* Lcssa PHI node if reduction is used outside of its definition loop. */ | |
132 | gphi *lcssa_phi; | |
133 | /* Stmts defining init and next. */ | |
134 | gimple *producer; | |
135 | gimple *consumer; | |
136 | /* If init is loaded from memory, this is the loading memory reference. */ | |
137 | tree init_ref; | |
138 | /* If reduction is finally stored to memory, this is the stored memory | |
139 | reference. */ | |
140 | tree fini_ref; | |
141 | enum reduction_type type; | |
142 | } *reduction_p; | |
143 | ||
144 | ||
145 | /* Dump reduction RE. */ | |
146 | ||
147 | static void | |
148 | dump_reduction (reduction_p re) | |
149 | { | |
150 | if (re->type == SIMPLE_RTYPE) | |
151 | fprintf (dump_file, " Simple reduction: "); | |
152 | else if (re->type == DOUBLE_RTYPE) | |
153 | fprintf (dump_file, " Double reduction: "); | |
154 | else | |
155 | fprintf (dump_file, " Unknown reduction: "); | |
156 | ||
157 | print_gimple_stmt (dump_file, re->phi, 0); | |
158 | } | |
159 | ||
160 | /* Dump LOOP's induction IV. */ | |
161 | static void | |
162 | dump_induction (struct loop *loop, induction_p iv) | |
163 | { | |
164 | fprintf (dump_file, " Induction: "); | |
165 | print_generic_expr (dump_file, iv->var, TDF_SLIM); | |
166 | fprintf (dump_file, " = {"); | |
167 | print_generic_expr (dump_file, iv->init_expr, TDF_SLIM); | |
168 | fprintf (dump_file, ", "); | |
169 | print_generic_expr (dump_file, iv->step, TDF_SLIM); | |
170 | fprintf (dump_file, "}_%d\n", loop->num); | |
171 | } | |
172 | ||
173 | /* Loop candidate for interchange. */ | |
174 | ||
175 | struct loop_cand | |
176 | { | |
177 | loop_cand (struct loop *, struct loop *); | |
178 | ~loop_cand (); | |
179 | ||
180 | reduction_p find_reduction_by_stmt (gimple *); | |
181 | void classify_simple_reduction (reduction_p); | |
182 | bool analyze_iloop_reduction_var (tree); | |
183 | bool analyze_oloop_reduction_var (loop_cand *, tree); | |
184 | bool analyze_induction_var (tree, tree); | |
185 | bool analyze_carried_vars (loop_cand *); | |
186 | bool analyze_lcssa_phis (void); | |
187 | bool can_interchange_p (loop_cand *); | |
fbdec14e BC |
188 | void undo_simple_reduction (reduction_p, bitmap); |
189 | ||
190 | /* The loop itself. */ | |
191 | struct loop *m_loop; | |
192 | /* The outer loop for interchange. It equals to loop if this loop cand | |
193 | itself represents the outer loop. */ | |
194 | struct loop *m_outer; | |
195 | /* Vector of induction variables in loop. */ | |
196 | vec<induction_p> m_inductions; | |
197 | /* Vector of reduction variables in loop. */ | |
198 | vec<reduction_p> m_reductions; | |
199 | /* Lcssa PHI nodes of this loop. */ | |
200 | vec<gphi *> m_lcssa_nodes; | |
201 | /* Single exit edge of this loop. */ | |
202 | edge m_exit; | |
203 | /* Basic blocks of this loop. */ | |
204 | basic_block *m_bbs; | |
1eeeda47 BC |
205 | /* Number of stmts of this loop. Inner loops' stmts are not included. */ |
206 | int m_num_stmts; | |
207 | /* Number of constant initialized simple reduction. */ | |
208 | int m_const_init_reduc; | |
fbdec14e BC |
209 | }; |
210 | ||
211 | /* Constructor. */ | |
212 | ||
213 | loop_cand::loop_cand (struct loop *loop, struct loop *outer) | |
1eeeda47 BC |
214 | : m_loop (loop), m_outer (outer), m_exit (single_exit (loop)), |
215 | m_bbs (get_loop_body (loop)), m_num_stmts (0), m_const_init_reduc (0) | |
fbdec14e BC |
216 | { |
217 | m_inductions.create (3); | |
218 | m_reductions.create (3); | |
219 | m_lcssa_nodes.create (3); | |
220 | } | |
221 | ||
222 | /* Destructor. */ | |
223 | ||
224 | loop_cand::~loop_cand () | |
225 | { | |
226 | induction_p iv; | |
227 | for (unsigned i = 0; m_inductions.iterate (i, &iv); ++i) | |
228 | free (iv); | |
229 | ||
230 | reduction_p re; | |
231 | for (unsigned i = 0; m_reductions.iterate (i, &re); ++i) | |
232 | free (re); | |
233 | ||
234 | m_inductions.release (); | |
235 | m_reductions.release (); | |
236 | m_lcssa_nodes.release (); | |
237 | free (m_bbs); | |
238 | } | |
239 | ||
240 | /* Return single use stmt of VAR in LOOP, otherwise return NULL. */ | |
241 | ||
242 | static gimple * | |
243 | single_use_in_loop (tree var, struct loop *loop) | |
244 | { | |
245 | gimple *stmt, *res = NULL; | |
246 | use_operand_p use_p; | |
247 | imm_use_iterator iterator; | |
248 | ||
249 | FOR_EACH_IMM_USE_FAST (use_p, iterator, var) | |
250 | { | |
251 | stmt = USE_STMT (use_p); | |
252 | if (is_gimple_debug (stmt)) | |
253 | continue; | |
254 | ||
255 | if (!flow_bb_inside_loop_p (loop, gimple_bb (stmt))) | |
256 | continue; | |
257 | ||
258 | if (res) | |
259 | return NULL; | |
260 | ||
261 | res = stmt; | |
262 | } | |
263 | return res; | |
264 | } | |
265 | ||
266 | /* Return true if E is unsupported in loop interchange, i.e, E is a complex | |
267 | edge or part of irreducible loop. */ | |
268 | ||
269 | static inline bool | |
270 | unsupported_edge (edge e) | |
271 | { | |
272 | return (e->flags & (EDGE_COMPLEX | EDGE_IRREDUCIBLE_LOOP)); | |
273 | } | |
274 | ||
275 | /* Return the reduction if STMT is one of its lcssa PHI, producer or consumer | |
276 | stmt. */ | |
277 | ||
278 | reduction_p | |
279 | loop_cand::find_reduction_by_stmt (gimple *stmt) | |
280 | { | |
281 | gphi *phi = dyn_cast <gphi *> (stmt); | |
282 | reduction_p re; | |
283 | ||
284 | for (unsigned i = 0; m_reductions.iterate (i, &re); ++i) | |
285 | if ((phi != NULL && phi == re->lcssa_phi) | |
286 | || (stmt == re->producer || stmt == re->consumer)) | |
287 | return re; | |
288 | ||
289 | return NULL; | |
290 | } | |
291 | ||
fbdec14e BC |
292 | /* Return true if current loop_cand be interchanged. ILOOP is not NULL if |
293 | current loop_cand is outer loop in loop nest. */ | |
294 | ||
295 | bool | |
296 | loop_cand::can_interchange_p (loop_cand *iloop) | |
297 | { | |
298 | /* For now we only support at most one reduction. */ | |
299 | unsigned allowed_reduction_num = 1; | |
300 | ||
301 | /* Only support reduction if the loop nest to be interchanged is the | |
302 | innermostin two loops. */ | |
303 | if ((iloop == NULL && m_loop->inner != NULL) | |
304 | || (iloop != NULL && iloop->m_loop->inner != NULL)) | |
305 | allowed_reduction_num = 0; | |
306 | ||
307 | if (m_reductions.length () > allowed_reduction_num | |
308 | || (m_reductions.length () == 1 | |
309 | && m_reductions[0]->type == UNKNOWN_RTYPE)) | |
310 | return false; | |
311 | ||
312 | /* Only support lcssa PHI node which is for reduction. */ | |
313 | if (m_lcssa_nodes.length () > allowed_reduction_num) | |
314 | return false; | |
315 | ||
1eeeda47 BC |
316 | /* Check if basic block has any unsupported operation. Note basic blocks |
317 | of inner loops are not checked here. */ | |
fbdec14e BC |
318 | for (unsigned i = 0; i < m_loop->num_nodes; i++) |
319 | { | |
320 | basic_block bb = m_bbs[i]; | |
1eeeda47 BC |
321 | gphi_iterator psi; |
322 | gimple_stmt_iterator gsi; | |
fbdec14e BC |
323 | |
324 | /* Skip basic blocks of inner loops. */ | |
325 | if (bb->loop_father != m_loop) | |
1eeeda47 | 326 | continue; |
fbdec14e | 327 | |
1eeeda47 BC |
328 | for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) |
329 | { | |
330 | gimple *stmt = gsi_stmt (gsi); | |
331 | if (is_gimple_debug (stmt)) | |
332 | continue; | |
333 | ||
334 | if (gimple_has_side_effects (stmt)) | |
335 | return false; | |
336 | ||
337 | m_num_stmts++; | |
338 | if (gcall *call = dyn_cast <gcall *> (stmt)) | |
339 | { | |
340 | /* In basic block of outer loop, the call should be cheap since | |
341 | it will be moved to inner loop. */ | |
342 | if (iloop != NULL | |
343 | && !gimple_inexpensive_call_p (call)) | |
344 | return false; | |
345 | continue; | |
346 | } | |
347 | ||
348 | if (!iloop || !gimple_vuse (stmt)) | |
349 | continue; | |
fbdec14e | 350 | |
1eeeda47 BC |
351 | /* Support stmt accessing memory in outer loop only if it is for |
352 | inner loop's reduction. */ | |
353 | if (iloop->find_reduction_by_stmt (stmt)) | |
354 | continue; | |
355 | ||
356 | tree lhs; | |
357 | /* Support loop invariant memory reference if it's only used once by | |
358 | inner loop. */ | |
359 | /* ??? How's this checking for invariantness? */ | |
360 | if (gimple_assign_single_p (stmt) | |
361 | && (lhs = gimple_assign_lhs (stmt)) != NULL_TREE | |
362 | && TREE_CODE (lhs) == SSA_NAME | |
363 | && single_use_in_loop (lhs, iloop->m_loop)) | |
364 | continue; | |
365 | ||
366 | return false; | |
367 | } | |
fbdec14e | 368 | /* Check if loop has too many stmts. */ |
1eeeda47 | 369 | if (m_num_stmts > MAX_NUM_STMT) |
fbdec14e | 370 | return false; |
1eeeda47 BC |
371 | |
372 | /* Allow PHI nodes in any basic block of inner loop, PHI nodes in outer | |
373 | loop's header, or PHI nodes in dest bb of inner loop's exit edge. */ | |
374 | if (!iloop || bb == m_loop->header | |
375 | || bb == iloop->m_exit->dest) | |
376 | continue; | |
377 | ||
378 | /* Don't allow any other PHI nodes. */ | |
379 | for (psi = gsi_start_phis (bb); !gsi_end_p (psi); gsi_next (&psi)) | |
380 | if (!virtual_operand_p (PHI_RESULT (psi.phi ()))) | |
381 | return false; | |
fbdec14e BC |
382 | } |
383 | ||
384 | return true; | |
385 | } | |
386 | ||
387 | /* Programmers and optimizers (like loop store motion) may optimize code: | |
388 | ||
389 | for (int i = 0; i < N; i++) | |
390 | for (int j = 0; j < N; j++) | |
391 | a[i] += b[j][i] * c[j][i]; | |
392 | ||
393 | into reduction: | |
394 | ||
395 | for (int i = 0; i < N; i++) | |
396 | { | |
397 | // producer. Note sum can be intitialized to a constant. | |
398 | int sum = a[i]; | |
399 | for (int j = 0; j < N; j++) | |
400 | { | |
401 | sum += b[j][i] * c[j][i]; | |
402 | } | |
403 | // consumer. | |
404 | a[i] = sum; | |
405 | } | |
406 | ||
407 | The result code can't be interchanged without undoing the optimization. | |
408 | This function classifies this kind reduction and records information so | |
409 | that we can undo the store motion during interchange. */ | |
410 | ||
411 | void | |
412 | loop_cand::classify_simple_reduction (reduction_p re) | |
413 | { | |
414 | gimple *producer, *consumer; | |
415 | ||
416 | /* Check init variable of reduction and how it is initialized. */ | |
417 | if (TREE_CODE (re->init) == SSA_NAME) | |
418 | { | |
419 | producer = SSA_NAME_DEF_STMT (re->init); | |
420 | re->producer = producer; | |
421 | basic_block bb = gimple_bb (producer); | |
422 | if (!bb || bb->loop_father != m_outer) | |
423 | return; | |
424 | ||
425 | if (!gimple_assign_load_p (producer)) | |
426 | return; | |
427 | ||
428 | re->init_ref = gimple_assign_rhs1 (producer); | |
429 | } | |
1eeeda47 BC |
430 | else if (CONSTANT_CLASS_P (re->init)) |
431 | m_const_init_reduc++; | |
432 | else | |
fbdec14e BC |
433 | return; |
434 | ||
435 | /* Check how reduction variable is used. */ | |
436 | consumer = single_use_in_loop (PHI_RESULT (re->lcssa_phi), m_outer); | |
437 | if (!consumer | |
438 | || !gimple_store_p (consumer)) | |
439 | return; | |
440 | ||
441 | re->fini_ref = gimple_get_lhs (consumer); | |
442 | re->consumer = consumer; | |
443 | ||
444 | /* Simple reduction with constant initializer. */ | |
445 | if (!re->init_ref) | |
446 | { | |
447 | gcc_assert (CONSTANT_CLASS_P (re->init)); | |
448 | re->init_ref = unshare_expr (re->fini_ref); | |
449 | } | |
450 | ||
451 | /* Require memory references in producer and consumer are the same so | |
452 | that we can undo reduction during interchange. */ | |
453 | if (re->init_ref && !operand_equal_p (re->init_ref, re->fini_ref, 0)) | |
454 | return; | |
455 | ||
456 | re->type = SIMPLE_RTYPE; | |
457 | } | |
458 | ||
459 | /* Analyze reduction variable VAR for inner loop of the loop nest to be | |
460 | interchanged. Return true if analysis succeeds. */ | |
461 | ||
462 | bool | |
463 | loop_cand::analyze_iloop_reduction_var (tree var) | |
464 | { | |
465 | gphi *phi = as_a <gphi *> (SSA_NAME_DEF_STMT (var)); | |
466 | gphi *lcssa_phi = NULL, *use_phi; | |
467 | tree init = PHI_ARG_DEF_FROM_EDGE (phi, loop_preheader_edge (m_loop)); | |
468 | tree next = PHI_ARG_DEF_FROM_EDGE (phi, loop_latch_edge (m_loop)); | |
469 | reduction_p re; | |
470 | gimple *stmt, *next_def, *single_use = NULL; | |
471 | use_operand_p use_p; | |
472 | imm_use_iterator iterator; | |
473 | ||
474 | if (TREE_CODE (next) != SSA_NAME) | |
475 | return false; | |
476 | ||
477 | next_def = SSA_NAME_DEF_STMT (next); | |
478 | basic_block bb = gimple_bb (next_def); | |
479 | if (!bb || !flow_bb_inside_loop_p (m_loop, bb)) | |
480 | return false; | |
481 | ||
482 | /* In restricted reduction, the var is (and must be) used in defining | |
483 | the updated var. The process can be depicted as below: | |
484 | ||
485 | var ;; = PHI<init, next> | |
486 | | | |
487 | | | |
488 | v | |
489 | +---------------------+ | |
490 | | reduction operators | <-- other operands | |
491 | +---------------------+ | |
492 | | | |
493 | | | |
494 | v | |
495 | next | |
496 | ||
497 | In terms loop interchange, we don't change how NEXT is computed based | |
498 | on VAR and OTHER OPERANDS. In case of double reduction in loop nest | |
499 | to be interchanged, we don't changed it at all. In the case of simple | |
500 | reduction in inner loop, we only make change how VAR/NEXT is loaded or | |
501 | stored. With these conditions, we can relax restrictions on reduction | |
502 | in a way that reduction operation is seen as black box. In general, | |
503 | we can ignore reassociation of reduction operator; we can handle fake | |
504 | reductions in which VAR is not even used to compute NEXT. */ | |
505 | if (! single_imm_use (var, &use_p, &single_use) | |
506 | || ! flow_bb_inside_loop_p (m_loop, gimple_bb (single_use))) | |
507 | return false; | |
508 | ||
509 | /* Check the reduction operation. We require a left-associative operation. | |
510 | For FP math we also need to be allowed to associate operations. */ | |
511 | if (gassign *ass = dyn_cast <gassign *> (single_use)) | |
512 | { | |
513 | enum tree_code code = gimple_assign_rhs_code (ass); | |
514 | if (! (associative_tree_code (code) | |
515 | || (code == MINUS_EXPR | |
516 | && use_p->use == gimple_assign_rhs1_ptr (ass))) | |
517 | || (FLOAT_TYPE_P (TREE_TYPE (var)) | |
518 | && ! flag_associative_math)) | |
519 | return false; | |
520 | } | |
521 | else | |
522 | return false; | |
523 | ||
524 | /* Handle and verify a series of stmts feeding the reduction op. */ | |
525 | if (single_use != next_def | |
4f5b9c80 | 526 | && !check_reduction_path (dump_user_location_t (), m_loop, phi, next, |
fbdec14e BC |
527 | gimple_assign_rhs_code (single_use))) |
528 | return false; | |
529 | ||
530 | /* Only support cases in which INIT is used in inner loop. */ | |
531 | if (TREE_CODE (init) == SSA_NAME) | |
532 | FOR_EACH_IMM_USE_FAST (use_p, iterator, init) | |
533 | { | |
534 | stmt = USE_STMT (use_p); | |
535 | if (is_gimple_debug (stmt)) | |
536 | continue; | |
537 | ||
538 | if (!flow_bb_inside_loop_p (m_loop, gimple_bb (stmt))) | |
539 | return false; | |
540 | } | |
541 | ||
542 | FOR_EACH_IMM_USE_FAST (use_p, iterator, next) | |
543 | { | |
544 | stmt = USE_STMT (use_p); | |
545 | if (is_gimple_debug (stmt)) | |
546 | continue; | |
547 | ||
548 | /* Or else it's used in PHI itself. */ | |
549 | use_phi = dyn_cast <gphi *> (stmt); | |
550 | if (use_phi == phi) | |
551 | continue; | |
552 | ||
553 | if (use_phi != NULL | |
554 | && lcssa_phi == NULL | |
555 | && gimple_bb (stmt) == m_exit->dest | |
556 | && PHI_ARG_DEF_FROM_EDGE (use_phi, m_exit) == next) | |
557 | lcssa_phi = use_phi; | |
558 | else | |
559 | return false; | |
560 | } | |
561 | if (!lcssa_phi) | |
562 | return false; | |
563 | ||
564 | re = XCNEW (struct reduction); | |
565 | re->var = var; | |
566 | re->init = init; | |
567 | re->next = next; | |
568 | re->phi = phi; | |
569 | re->lcssa_phi = lcssa_phi; | |
570 | ||
571 | classify_simple_reduction (re); | |
572 | ||
573 | if (dump_file && (dump_flags & TDF_DETAILS)) | |
574 | dump_reduction (re); | |
575 | ||
576 | m_reductions.safe_push (re); | |
577 | return true; | |
578 | } | |
579 | ||
580 | /* Analyze reduction variable VAR for outer loop of the loop nest to be | |
581 | interchanged. ILOOP is not NULL and points to inner loop. For the | |
582 | moment, we only support double reduction for outer loop, like: | |
583 | ||
584 | for (int i = 0; i < n; i++) | |
585 | { | |
586 | int sum = 0; | |
587 | ||
588 | for (int j = 0; j < n; j++) // outer loop | |
589 | for (int k = 0; k < n; k++) // inner loop | |
590 | sum += a[i][k]*b[k][j]; | |
591 | ||
592 | s[i] = sum; | |
593 | } | |
594 | ||
595 | Note the innermost two loops are the loop nest to be interchanged. | |
596 | Return true if analysis succeeds. */ | |
597 | ||
598 | bool | |
599 | loop_cand::analyze_oloop_reduction_var (loop_cand *iloop, tree var) | |
600 | { | |
601 | gphi *phi = as_a <gphi *> (SSA_NAME_DEF_STMT (var)); | |
602 | gphi *lcssa_phi = NULL, *use_phi; | |
603 | tree init = PHI_ARG_DEF_FROM_EDGE (phi, loop_preheader_edge (m_loop)); | |
604 | tree next = PHI_ARG_DEF_FROM_EDGE (phi, loop_latch_edge (m_loop)); | |
605 | reduction_p re; | |
606 | gimple *stmt, *next_def; | |
607 | use_operand_p use_p; | |
608 | imm_use_iterator iterator; | |
609 | ||
610 | if (TREE_CODE (next) != SSA_NAME) | |
611 | return false; | |
612 | ||
613 | next_def = SSA_NAME_DEF_STMT (next); | |
614 | basic_block bb = gimple_bb (next_def); | |
615 | if (!bb || !flow_bb_inside_loop_p (m_loop, bb)) | |
616 | return false; | |
617 | ||
618 | /* Find inner loop's simple reduction that uses var as initializer. */ | |
619 | reduction_p inner_re = NULL; | |
620 | for (unsigned i = 0; iloop->m_reductions.iterate (i, &inner_re); ++i) | |
621 | if (inner_re->init == var || operand_equal_p (inner_re->init, var, 0)) | |
622 | break; | |
623 | ||
624 | if (inner_re == NULL | |
625 | || inner_re->type != UNKNOWN_RTYPE | |
626 | || inner_re->producer != phi) | |
627 | return false; | |
628 | ||
629 | /* In case of double reduction, outer loop's reduction should be updated | |
630 | by inner loop's simple reduction. */ | |
631 | if (next_def != inner_re->lcssa_phi) | |
632 | return false; | |
633 | ||
634 | /* Outer loop's reduction should only be used to initialize inner loop's | |
635 | simple reduction. */ | |
636 | if (! single_imm_use (var, &use_p, &stmt) | |
637 | || stmt != inner_re->phi) | |
638 | return false; | |
639 | ||
640 | /* Check this reduction is correctly used outside of loop via lcssa phi. */ | |
641 | FOR_EACH_IMM_USE_FAST (use_p, iterator, next) | |
642 | { | |
643 | stmt = USE_STMT (use_p); | |
644 | if (is_gimple_debug (stmt)) | |
645 | continue; | |
646 | ||
647 | /* Or else it's used in PHI itself. */ | |
648 | use_phi = dyn_cast <gphi *> (stmt); | |
649 | if (use_phi == phi) | |
650 | continue; | |
651 | ||
652 | if (lcssa_phi == NULL | |
653 | && use_phi != NULL | |
654 | && gimple_bb (stmt) == m_exit->dest | |
655 | && PHI_ARG_DEF_FROM_EDGE (use_phi, m_exit) == next) | |
656 | lcssa_phi = use_phi; | |
657 | else | |
658 | return false; | |
659 | } | |
660 | if (!lcssa_phi) | |
661 | return false; | |
662 | ||
663 | re = XCNEW (struct reduction); | |
664 | re->var = var; | |
665 | re->init = init; | |
666 | re->next = next; | |
667 | re->phi = phi; | |
668 | re->lcssa_phi = lcssa_phi; | |
669 | re->type = DOUBLE_RTYPE; | |
670 | inner_re->type = DOUBLE_RTYPE; | |
671 | ||
672 | if (dump_file && (dump_flags & TDF_DETAILS)) | |
673 | dump_reduction (re); | |
674 | ||
675 | m_reductions.safe_push (re); | |
676 | return true; | |
677 | } | |
678 | ||
679 | /* Return true if VAR is induction variable of current loop whose scev is | |
680 | specified by CHREC. */ | |
681 | ||
682 | bool | |
683 | loop_cand::analyze_induction_var (tree var, tree chrec) | |
684 | { | |
685 | gphi *phi = as_a <gphi *> (SSA_NAME_DEF_STMT (var)); | |
686 | tree init = PHI_ARG_DEF_FROM_EDGE (phi, loop_preheader_edge (m_loop)); | |
687 | ||
688 | /* Var is loop invariant, though it's unlikely to happen. */ | |
689 | if (tree_does_not_contain_chrecs (chrec)) | |
690 | { | |
691 | struct induction *iv = XCNEW (struct induction); | |
692 | iv->var = var; | |
693 | iv->init_val = init; | |
694 | iv->init_expr = chrec; | |
695 | iv->step = build_int_cst (TREE_TYPE (chrec), 0); | |
696 | m_inductions.safe_push (iv); | |
697 | return true; | |
698 | } | |
699 | ||
700 | if (TREE_CODE (chrec) != POLYNOMIAL_CHREC | |
701 | || CHREC_VARIABLE (chrec) != (unsigned) m_loop->num | |
702 | || tree_contains_chrecs (CHREC_LEFT (chrec), NULL) | |
703 | || tree_contains_chrecs (CHREC_RIGHT (chrec), NULL)) | |
704 | return false; | |
705 | ||
706 | struct induction *iv = XCNEW (struct induction); | |
707 | iv->var = var; | |
708 | iv->init_val = init; | |
709 | iv->init_expr = CHREC_LEFT (chrec); | |
710 | iv->step = CHREC_RIGHT (chrec); | |
711 | ||
712 | if (dump_file && (dump_flags & TDF_DETAILS)) | |
713 | dump_induction (m_loop, iv); | |
714 | ||
715 | m_inductions.safe_push (iv); | |
716 | return true; | |
717 | } | |
718 | ||
719 | /* Return true if all loop carried variables defined in loop header can | |
720 | be successfully analyzed. */ | |
721 | ||
722 | bool | |
723 | loop_cand::analyze_carried_vars (loop_cand *iloop) | |
724 | { | |
725 | edge e = loop_preheader_edge (m_outer); | |
726 | gphi_iterator gsi; | |
727 | ||
728 | if (dump_file && (dump_flags & TDF_DETAILS)) | |
729 | fprintf (dump_file, "\nLoop(%d) carried vars:\n", m_loop->num); | |
730 | ||
731 | for (gsi = gsi_start_phis (m_loop->header); !gsi_end_p (gsi); gsi_next (&gsi)) | |
732 | { | |
733 | gphi *phi = gsi.phi (); | |
734 | ||
735 | tree var = PHI_RESULT (phi); | |
736 | if (virtual_operand_p (var)) | |
737 | continue; | |
738 | ||
739 | tree chrec = analyze_scalar_evolution (m_loop, var); | |
740 | chrec = instantiate_scev (e, m_loop, chrec); | |
741 | ||
742 | /* Analyze var as reduction variable. */ | |
743 | if (chrec_contains_undetermined (chrec) | |
744 | || chrec_contains_symbols_defined_in_loop (chrec, m_outer->num)) | |
745 | { | |
746 | if (iloop && !analyze_oloop_reduction_var (iloop, var)) | |
747 | return false; | |
748 | if (!iloop && !analyze_iloop_reduction_var (var)) | |
749 | return false; | |
750 | } | |
751 | /* Analyze var as induction variable. */ | |
752 | else if (!analyze_induction_var (var, chrec)) | |
753 | return false; | |
754 | } | |
755 | ||
756 | return true; | |
757 | } | |
758 | ||
759 | /* Return TRUE if loop closed PHI nodes can be analyzed successfully. */ | |
760 | ||
761 | bool | |
762 | loop_cand::analyze_lcssa_phis (void) | |
763 | { | |
764 | gphi_iterator gsi; | |
765 | for (gsi = gsi_start_phis (m_exit->dest); !gsi_end_p (gsi); gsi_next (&gsi)) | |
766 | { | |
767 | gphi *phi = gsi.phi (); | |
768 | ||
769 | if (virtual_operand_p (PHI_RESULT (phi))) | |
770 | continue; | |
771 | ||
772 | /* TODO: We only support lcssa phi for reduction for now. */ | |
773 | if (!find_reduction_by_stmt (phi)) | |
774 | return false; | |
775 | } | |
776 | ||
777 | return true; | |
778 | } | |
779 | ||
780 | /* CONSUMER is a stmt in BB storing reduction result into memory object. | |
781 | When the reduction is intialized from constant value, we need to add | |
782 | a stmt loading from the memory object to target basic block in inner | |
783 | loop during undoing the reduction. Problem is that memory reference | |
784 | may use ssa variables not dominating the target basic block. This | |
785 | function finds all stmts on which CONSUMER depends in basic block BB, | |
786 | records and returns them via STMTS. */ | |
787 | ||
788 | static void | |
789 | find_deps_in_bb_for_stmt (gimple_seq *stmts, basic_block bb, gimple *consumer) | |
790 | { | |
791 | auto_vec<gimple *, 4> worklist; | |
792 | use_operand_p use_p; | |
793 | ssa_op_iter iter; | |
794 | gimple *stmt, *def_stmt; | |
795 | gimple_stmt_iterator gsi; | |
796 | ||
797 | /* First clear flag for stmts in bb. */ | |
798 | for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) | |
799 | gimple_set_plf (gsi_stmt (gsi), GF_PLF_1, false); | |
800 | ||
801 | /* DFS search all depended stmts in bb and mark flag for these stmts. */ | |
802 | worklist.safe_push (consumer); | |
803 | while (!worklist.is_empty ()) | |
804 | { | |
805 | stmt = worklist.pop (); | |
806 | FOR_EACH_SSA_USE_OPERAND (use_p, stmt, iter, SSA_OP_USE) | |
807 | { | |
808 | def_stmt = SSA_NAME_DEF_STMT (USE_FROM_PTR (use_p)); | |
809 | ||
810 | if (is_a <gphi *> (def_stmt) | |
811 | || gimple_bb (def_stmt) != bb | |
812 | || gimple_plf (def_stmt, GF_PLF_1)) | |
813 | continue; | |
814 | ||
815 | worklist.safe_push (def_stmt); | |
816 | } | |
817 | gimple_set_plf (stmt, GF_PLF_1, true); | |
818 | } | |
65f4b875 | 819 | for (gsi = gsi_start_nondebug_bb (bb); |
fbdec14e BC |
820 | !gsi_end_p (gsi) && (stmt = gsi_stmt (gsi)) != consumer;) |
821 | { | |
822 | /* Move dep stmts to sequence STMTS. */ | |
823 | if (gimple_plf (stmt, GF_PLF_1)) | |
824 | { | |
825 | gsi_remove (&gsi, false); | |
826 | gimple_seq_add_stmt_without_update (stmts, stmt); | |
827 | } | |
828 | else | |
829 | gsi_next_nondebug (&gsi); | |
830 | } | |
831 | } | |
832 | ||
833 | /* User can write, optimizers can generate simple reduction RE for inner | |
834 | loop. In order to make interchange valid, we have to undo reduction by | |
835 | moving producer and consumer stmts into the inner loop. For example, | |
836 | below code: | |
837 | ||
838 | init = MEM_REF[idx]; //producer | |
839 | loop: | |
840 | var = phi<init, next> | |
841 | next = var op ... | |
842 | reduc_sum = phi<next> | |
843 | MEM_REF[idx] = reduc_sum //consumer | |
844 | ||
845 | is transformed into: | |
846 | ||
847 | loop: | |
848 | new_var = MEM_REF[idx]; //producer after moving | |
849 | next = new_var op ... | |
850 | MEM_REF[idx] = next; //consumer after moving | |
851 | ||
852 | Note if the reduction variable is initialized to constant, like: | |
853 | ||
854 | var = phi<0.0, next> | |
855 | ||
856 | we compute new_var as below: | |
857 | ||
858 | loop: | |
859 | tmp = MEM_REF[idx]; | |
860 | new_var = !first_iteration ? tmp : 0.0; | |
861 | ||
862 | so that the initial const is used in the first iteration of loop. Also | |
863 | record ssa variables for dead code elimination in DCE_SEEDS. */ | |
864 | ||
865 | void | |
866 | loop_cand::undo_simple_reduction (reduction_p re, bitmap dce_seeds) | |
867 | { | |
868 | gimple *stmt; | |
869 | gimple_stmt_iterator from, to = gsi_after_labels (m_loop->header); | |
870 | gimple_seq stmts = NULL; | |
871 | tree new_var; | |
872 | ||
873 | /* Prepare the initialization stmts and insert it to inner loop. */ | |
874 | if (re->producer != NULL) | |
875 | { | |
876 | gimple_set_vuse (re->producer, NULL_TREE); | |
877 | from = gsi_for_stmt (re->producer); | |
878 | gsi_remove (&from, false); | |
879 | gimple_seq_add_stmt_without_update (&stmts, re->producer); | |
880 | new_var = re->init; | |
881 | } | |
882 | else | |
883 | { | |
884 | /* Find all stmts on which expression "MEM_REF[idx]" depends. */ | |
885 | find_deps_in_bb_for_stmt (&stmts, gimple_bb (re->consumer), re->consumer); | |
886 | /* Because we generate new stmt loading from the MEM_REF to TMP. */ | |
887 | tree cond, tmp = copy_ssa_name (re->var); | |
888 | stmt = gimple_build_assign (tmp, re->init_ref); | |
889 | gimple_seq_add_stmt_without_update (&stmts, stmt); | |
890 | ||
891 | /* Init new_var to MEM_REF or CONST depending on if it is the first | |
892 | iteration. */ | |
893 | induction_p iv = m_inductions[0]; | |
894 | cond = fold_build2 (NE_EXPR, boolean_type_node, iv->var, iv->init_val); | |
895 | new_var = copy_ssa_name (re->var); | |
896 | stmt = gimple_build_assign (new_var, COND_EXPR, cond, tmp, re->init); | |
897 | gimple_seq_add_stmt_without_update (&stmts, stmt); | |
898 | } | |
899 | gsi_insert_seq_before (&to, stmts, GSI_SAME_STMT); | |
900 | ||
901 | /* Replace all uses of reduction var with new variable. */ | |
902 | use_operand_p use_p; | |
903 | imm_use_iterator iterator; | |
904 | FOR_EACH_IMM_USE_STMT (stmt, iterator, re->var) | |
905 | { | |
906 | FOR_EACH_IMM_USE_ON_STMT (use_p, iterator) | |
907 | SET_USE (use_p, new_var); | |
908 | ||
909 | update_stmt (stmt); | |
910 | } | |
911 | ||
912 | /* Move consumer stmt into inner loop, just after reduction next's def. */ | |
913 | unlink_stmt_vdef (re->consumer); | |
914 | release_ssa_name (gimple_vdef (re->consumer)); | |
915 | gimple_set_vdef (re->consumer, NULL_TREE); | |
916 | gimple_set_vuse (re->consumer, NULL_TREE); | |
917 | gimple_assign_set_rhs1 (re->consumer, re->next); | |
918 | from = gsi_for_stmt (re->consumer); | |
919 | to = gsi_for_stmt (SSA_NAME_DEF_STMT (re->next)); | |
920 | gsi_move_after (&from, &to); | |
921 | ||
922 | /* Mark the reduction variables for DCE. */ | |
923 | bitmap_set_bit (dce_seeds, SSA_NAME_VERSION (re->var)); | |
924 | bitmap_set_bit (dce_seeds, SSA_NAME_VERSION (PHI_RESULT (re->lcssa_phi))); | |
925 | } | |
926 | ||
927 | /* Free DATAREFS and its auxiliary memory. */ | |
928 | ||
929 | static void | |
930 | free_data_refs_with_aux (vec<data_reference_p> datarefs) | |
931 | { | |
932 | data_reference_p dr; | |
933 | for (unsigned i = 0; datarefs.iterate (i, &dr); ++i) | |
934 | if (dr->aux != NULL) | |
935 | { | |
936 | DR_ACCESS_STRIDE (dr)->release (); | |
46bb9d29 | 937 | delete (vec<tree> *) dr->aux; |
fbdec14e BC |
938 | } |
939 | ||
940 | free_data_refs (datarefs); | |
941 | } | |
942 | ||
943 | /* Class for loop interchange transformation. */ | |
944 | ||
945 | class tree_loop_interchange | |
946 | { | |
947 | public: | |
948 | tree_loop_interchange (vec<struct loop *> loop_nest) | |
949 | : m_loop_nest (loop_nest), m_niters_iv_var (NULL_TREE), | |
950 | m_dce_seeds (BITMAP_ALLOC (NULL)) { } | |
951 | ~tree_loop_interchange () { BITMAP_FREE (m_dce_seeds); } | |
952 | bool interchange (vec<data_reference_p>, vec<ddr_p>); | |
953 | ||
954 | private: | |
955 | void update_data_info (unsigned, unsigned, vec<data_reference_p>, vec<ddr_p>); | |
956 | bool valid_data_dependences (unsigned, unsigned, vec<ddr_p>); | |
957 | void interchange_loops (loop_cand &, loop_cand &); | |
958 | void map_inductions_to_loop (loop_cand &, loop_cand &); | |
959 | void move_code_to_inner_loop (struct loop *, struct loop *, basic_block *); | |
960 | ||
961 | /* The whole loop nest in which interchange is ongoing. */ | |
962 | vec<struct loop *> m_loop_nest; | |
963 | /* We create new IV which is only used in loop's exit condition check. | |
964 | In case of 3-level loop nest interchange, when we interchange the | |
965 | innermost two loops, new IV created in the middle level loop does | |
966 | not need to be preserved in interchanging the outermost two loops | |
967 | later. We record the IV so that it can be skipped. */ | |
968 | tree m_niters_iv_var; | |
969 | /* Bitmap of seed variables for dead code elimination after interchange. */ | |
970 | bitmap m_dce_seeds; | |
971 | }; | |
972 | ||
973 | /* Update data refs' access stride and dependence information after loop | |
974 | interchange. I_IDX/O_IDX gives indices of interchanged loops in loop | |
975 | nest. DATAREFS are data references. DDRS are data dependences. */ | |
976 | ||
977 | void | |
978 | tree_loop_interchange::update_data_info (unsigned i_idx, unsigned o_idx, | |
979 | vec<data_reference_p> datarefs, | |
980 | vec<ddr_p> ddrs) | |
981 | { | |
982 | struct data_reference *dr; | |
983 | struct data_dependence_relation *ddr; | |
984 | ||
985 | /* Update strides of data references. */ | |
986 | for (unsigned i = 0; datarefs.iterate (i, &dr); ++i) | |
987 | { | |
988 | vec<tree> *stride = DR_ACCESS_STRIDE (dr); | |
989 | gcc_assert (stride->length () > i_idx); | |
990 | std::swap ((*stride)[i_idx], (*stride)[o_idx]); | |
991 | } | |
992 | /* Update data dependences. */ | |
993 | for (unsigned i = 0; ddrs.iterate (i, &ddr); ++i) | |
994 | if (DDR_ARE_DEPENDENT (ddr) != chrec_known) | |
995 | { | |
996 | for (unsigned j = 0; j < DDR_NUM_DIST_VECTS (ddr); ++j) | |
997 | { | |
998 | lambda_vector dist_vect = DDR_DIST_VECT (ddr, j); | |
999 | std::swap (dist_vect[i_idx], dist_vect[o_idx]); | |
1000 | } | |
1001 | } | |
1002 | } | |
1003 | ||
1004 | /* Check data dependence relations, return TRUE if it's valid to interchange | |
1005 | two loops specified by I_IDX/O_IDX. Theoretically, interchanging the two | |
1006 | loops is valid only if dist vector, after interchanging, doesn't have '>' | |
1007 | as the leftmost non-'=' direction. Practically, this function have been | |
1008 | conservative here by not checking some valid cases. */ | |
1009 | ||
1010 | bool | |
1011 | tree_loop_interchange::valid_data_dependences (unsigned i_idx, unsigned o_idx, | |
1012 | vec<ddr_p> ddrs) | |
1013 | { | |
1014 | struct data_dependence_relation *ddr; | |
1015 | ||
1016 | for (unsigned i = 0; ddrs.iterate (i, &ddr); ++i) | |
1017 | { | |
1018 | /* Skip no-dependence case. */ | |
1019 | if (DDR_ARE_DEPENDENT (ddr) == chrec_known) | |
1020 | continue; | |
1021 | ||
1022 | for (unsigned j = 0; j < DDR_NUM_DIST_VECTS (ddr); ++j) | |
1023 | { | |
1024 | lambda_vector dist_vect = DDR_DIST_VECT (ddr, j); | |
1025 | unsigned level = dependence_level (dist_vect, m_loop_nest.length ()); | |
1026 | ||
1027 | /* If there is no carried dependence. */ | |
1028 | if (level == 0) | |
1029 | continue; | |
1030 | ||
1031 | level --; | |
1032 | ||
1033 | /* If dependence is not carried by any loop in between the two | |
1034 | loops [oloop, iloop] to interchange. */ | |
1035 | if (level < o_idx || level > i_idx) | |
1036 | continue; | |
1037 | ||
1038 | /* Be conservative, skip case if either direction at i_idx/o_idx | |
1039 | levels is not '=' or '<'. */ | |
1040 | if (dist_vect[i_idx] < 0 || dist_vect[o_idx] < 0) | |
1041 | return false; | |
1042 | } | |
1043 | } | |
1044 | ||
1045 | return true; | |
1046 | } | |
1047 | ||
1048 | /* Interchange two loops specified by ILOOP and OLOOP. */ | |
1049 | ||
1050 | void | |
1051 | tree_loop_interchange::interchange_loops (loop_cand &iloop, loop_cand &oloop) | |
1052 | { | |
1053 | reduction_p re; | |
1054 | gimple_stmt_iterator gsi; | |
1055 | tree i_niters, o_niters, var_after; | |
1056 | ||
1057 | /* Undo inner loop's simple reduction. */ | |
1058 | for (unsigned i = 0; iloop.m_reductions.iterate (i, &re); ++i) | |
1059 | if (re->type != DOUBLE_RTYPE) | |
1060 | { | |
1061 | if (re->producer) | |
1062 | reset_debug_uses (re->producer); | |
1063 | ||
1064 | iloop.undo_simple_reduction (re, m_dce_seeds); | |
1065 | } | |
1066 | ||
1067 | /* Only need to reset debug uses for double reduction. */ | |
1068 | for (unsigned i = 0; oloop.m_reductions.iterate (i, &re); ++i) | |
1069 | { | |
1070 | gcc_assert (re->type == DOUBLE_RTYPE); | |
1071 | reset_debug_uses (SSA_NAME_DEF_STMT (re->var)); | |
1072 | reset_debug_uses (SSA_NAME_DEF_STMT (re->next)); | |
1073 | } | |
1074 | ||
1075 | /* Prepare niters for both loops. */ | |
1076 | struct loop *loop_nest = m_loop_nest[0]; | |
1077 | edge instantiate_below = loop_preheader_edge (loop_nest); | |
1078 | gsi = gsi_last_bb (loop_preheader_edge (loop_nest)->src); | |
1079 | i_niters = number_of_latch_executions (iloop.m_loop); | |
1080 | i_niters = analyze_scalar_evolution (loop_outer (iloop.m_loop), i_niters); | |
1081 | i_niters = instantiate_scev (instantiate_below, loop_outer (iloop.m_loop), | |
1082 | i_niters); | |
1083 | i_niters = force_gimple_operand_gsi (&gsi, unshare_expr (i_niters), true, | |
1084 | NULL_TREE, false, GSI_CONTINUE_LINKING); | |
1085 | o_niters = number_of_latch_executions (oloop.m_loop); | |
1086 | if (oloop.m_loop != loop_nest) | |
1087 | { | |
1088 | o_niters = analyze_scalar_evolution (loop_outer (oloop.m_loop), o_niters); | |
1089 | o_niters = instantiate_scev (instantiate_below, loop_outer (oloop.m_loop), | |
1090 | o_niters); | |
1091 | } | |
1092 | o_niters = force_gimple_operand_gsi (&gsi, unshare_expr (o_niters), true, | |
1093 | NULL_TREE, false, GSI_CONTINUE_LINKING); | |
1094 | ||
1095 | /* Move src's code to tgt loop. This is necessary when src is the outer | |
1096 | loop and tgt is the inner loop. */ | |
1097 | move_code_to_inner_loop (oloop.m_loop, iloop.m_loop, oloop.m_bbs); | |
1098 | ||
1099 | /* Map outer loop's IV to inner loop, and vice versa. */ | |
1100 | map_inductions_to_loop (oloop, iloop); | |
1101 | map_inductions_to_loop (iloop, oloop); | |
1102 | ||
1103 | /* Create canonical IV for both loops. Note canonical IV for outer/inner | |
1104 | loop is actually from inner/outer loop. Also we record the new IV | |
1105 | created for the outer loop so that it can be skipped in later loop | |
1106 | interchange. */ | |
1107 | create_canonical_iv (oloop.m_loop, oloop.m_exit, | |
1108 | i_niters, &m_niters_iv_var, &var_after); | |
1109 | bitmap_set_bit (m_dce_seeds, SSA_NAME_VERSION (var_after)); | |
1110 | create_canonical_iv (iloop.m_loop, iloop.m_exit, | |
1111 | o_niters, NULL, &var_after); | |
1112 | bitmap_set_bit (m_dce_seeds, SSA_NAME_VERSION (var_after)); | |
1113 | ||
1114 | /* Scrap niters estimation of interchanged loops. */ | |
1115 | iloop.m_loop->any_upper_bound = false; | |
1116 | iloop.m_loop->any_likely_upper_bound = false; | |
1117 | free_numbers_of_iterations_estimates (iloop.m_loop); | |
1118 | oloop.m_loop->any_upper_bound = false; | |
1119 | oloop.m_loop->any_likely_upper_bound = false; | |
1120 | free_numbers_of_iterations_estimates (oloop.m_loop); | |
1121 | ||
4e090bcc BC |
1122 | /* Clear all cached scev information. This is expensive but shouldn't be |
1123 | a problem given we interchange in very limited times. */ | |
1124 | scev_reset_htab (); | |
1125 | ||
fbdec14e BC |
1126 | /* ??? The association between the loop data structure and the |
1127 | CFG changed, so what was loop N at the source level is now | |
1128 | loop M. We should think of retaining the association or breaking | |
1129 | it fully by creating a new loop instead of re-using the "wrong" one. */ | |
1130 | } | |
1131 | ||
1132 | /* Map induction variables of SRC loop to TGT loop. The function firstly | |
1133 | creates the same IV of SRC loop in TGT loop, then deletes the original | |
1134 | IV and re-initialize it using the newly created IV. For example, loop | |
1135 | nest: | |
1136 | ||
1137 | for (i = 0; i < N; i++) | |
1138 | for (j = 0; j < M; j++) | |
1139 | { | |
1140 | //use of i; | |
1141 | //use of j; | |
1142 | } | |
1143 | ||
1144 | will be transformed into: | |
1145 | ||
1146 | for (jj = 0; jj < M; jj++) | |
1147 | for (ii = 0; ii < N; ii++) | |
1148 | { | |
1149 | //use of ii; | |
1150 | //use of jj; | |
1151 | } | |
1152 | ||
1153 | after loop interchange. */ | |
1154 | ||
1155 | void | |
1156 | tree_loop_interchange::map_inductions_to_loop (loop_cand &src, loop_cand &tgt) | |
1157 | { | |
1158 | induction_p iv; | |
1159 | edge e = tgt.m_exit; | |
1160 | gimple_stmt_iterator incr_pos = gsi_last_bb (e->src), gsi; | |
1161 | ||
1162 | /* Map source loop's IV to target loop. */ | |
1163 | for (unsigned i = 0; src.m_inductions.iterate (i, &iv); ++i) | |
1164 | { | |
1165 | gimple *use_stmt, *stmt = SSA_NAME_DEF_STMT (iv->var); | |
1166 | gcc_assert (is_a <gphi *> (stmt)); | |
1167 | ||
1168 | use_operand_p use_p; | |
1169 | /* Only map original IV to target loop. */ | |
1170 | if (m_niters_iv_var != iv->var) | |
1171 | { | |
1172 | /* Map the IV by creating the same one in target loop. */ | |
1173 | tree var_before, var_after; | |
1174 | tree base = unshare_expr (iv->init_expr); | |
1175 | tree step = unshare_expr (iv->step); | |
1176 | create_iv (base, step, SSA_NAME_VAR (iv->var), | |
1177 | tgt.m_loop, &incr_pos, false, &var_before, &var_after); | |
1178 | bitmap_set_bit (m_dce_seeds, SSA_NAME_VERSION (var_before)); | |
1179 | bitmap_set_bit (m_dce_seeds, SSA_NAME_VERSION (var_after)); | |
1180 | ||
1181 | /* Replace uses of the original IV var with newly created IV var. */ | |
1182 | imm_use_iterator imm_iter; | |
1183 | FOR_EACH_IMM_USE_STMT (use_stmt, imm_iter, iv->var) | |
1184 | { | |
1185 | FOR_EACH_IMM_USE_ON_STMT (use_p, imm_iter) | |
1186 | SET_USE (use_p, var_before); | |
1187 | ||
1188 | update_stmt (use_stmt); | |
1189 | } | |
1190 | } | |
1191 | ||
1192 | /* Mark all uses for DCE. */ | |
1193 | ssa_op_iter op_iter; | |
1194 | FOR_EACH_PHI_OR_STMT_USE (use_p, stmt, op_iter, SSA_OP_USE) | |
1195 | { | |
1196 | tree use = USE_FROM_PTR (use_p); | |
1197 | if (TREE_CODE (use) == SSA_NAME | |
1198 | && ! SSA_NAME_IS_DEFAULT_DEF (use)) | |
1199 | bitmap_set_bit (m_dce_seeds, SSA_NAME_VERSION (use)); | |
1200 | } | |
1201 | ||
1202 | /* Delete definition of the original IV in the source loop. */ | |
1203 | gsi = gsi_for_stmt (stmt); | |
1204 | remove_phi_node (&gsi, true); | |
1205 | } | |
1206 | } | |
1207 | ||
1208 | /* Move stmts of outer loop to inner loop. */ | |
1209 | ||
1210 | void | |
1211 | tree_loop_interchange::move_code_to_inner_loop (struct loop *outer, | |
1212 | struct loop *inner, | |
1213 | basic_block *outer_bbs) | |
1214 | { | |
1215 | basic_block oloop_exit_bb = single_exit (outer)->src; | |
1216 | gimple_stmt_iterator gsi, to; | |
1217 | ||
1218 | for (unsigned i = 0; i < outer->num_nodes; i++) | |
1219 | { | |
1220 | basic_block bb = outer_bbs[i]; | |
1221 | ||
1222 | /* Skip basic blocks of inner loop. */ | |
1223 | if (flow_bb_inside_loop_p (inner, bb)) | |
1224 | continue; | |
1225 | ||
1226 | /* Move code from header/latch to header/latch. */ | |
1227 | if (bb == outer->header) | |
1228 | to = gsi_after_labels (inner->header); | |
1229 | else if (bb == outer->latch) | |
1230 | to = gsi_after_labels (inner->latch); | |
1231 | else | |
1232 | /* Otherwise, simply move to exit->src. */ | |
1233 | to = gsi_last_bb (single_exit (inner)->src); | |
1234 | ||
1235 | for (gsi = gsi_after_labels (bb); !gsi_end_p (gsi);) | |
1236 | { | |
1237 | gimple *stmt = gsi_stmt (gsi); | |
1238 | ||
1239 | if (oloop_exit_bb == bb | |
1240 | && stmt == gsi_stmt (gsi_last_bb (oloop_exit_bb))) | |
1241 | { | |
1242 | gsi_next (&gsi); | |
1243 | continue; | |
1244 | } | |
1245 | ||
1246 | if (gimple_vuse (stmt)) | |
1247 | gimple_set_vuse (stmt, NULL_TREE); | |
1248 | if (gimple_vdef (stmt)) | |
1249 | { | |
1250 | unlink_stmt_vdef (stmt); | |
1251 | release_ssa_name (gimple_vdef (stmt)); | |
1252 | gimple_set_vdef (stmt, NULL_TREE); | |
1253 | } | |
1254 | ||
1255 | reset_debug_uses (stmt); | |
1256 | gsi_move_before (&gsi, &to); | |
1257 | } | |
1258 | } | |
1259 | } | |
1260 | ||
1261 | /* Given data reference DR in LOOP_NEST, the function computes DR's access | |
1262 | stride at each level of loop from innermost LOOP to outer. On success, | |
1263 | it saves access stride at each level loop in a vector which is pointed | |
1264 | by DR->aux. For example: | |
1265 | ||
1266 | int arr[100][100][100]; | |
1267 | for (i = 0; i < 100; i++) ;(DR->aux)strides[0] = 40000 | |
1268 | for (j = 100; j > 0; j--) ;(DR->aux)strides[1] = 400 | |
1269 | for (k = 0; k < 100; k++) ;(DR->aux)strides[2] = 4 | |
1270 | arr[i][j - 1][k] = 0; */ | |
1271 | ||
1272 | static void | |
1273 | compute_access_stride (struct loop *loop_nest, struct loop *loop, | |
1274 | data_reference_p dr) | |
1275 | { | |
1276 | vec<tree> *strides = new vec<tree> (); | |
1277 | basic_block bb = gimple_bb (DR_STMT (dr)); | |
1278 | ||
1279 | while (!flow_bb_inside_loop_p (loop, bb)) | |
1280 | { | |
1281 | strides->safe_push (build_int_cst (sizetype, 0)); | |
1282 | loop = loop_outer (loop); | |
1283 | } | |
1284 | gcc_assert (loop == bb->loop_father); | |
1285 | ||
1286 | tree ref = DR_REF (dr); | |
df0f6bbb JJ |
1287 | if (TREE_CODE (ref) == COMPONENT_REF |
1288 | && DECL_BIT_FIELD (TREE_OPERAND (ref, 1))) | |
1289 | { | |
1290 | /* We can't take address of bitfields. If the bitfield is at constant | |
1291 | offset from the start of the struct, just use address of the | |
1292 | struct, for analysis of the strides that shouldn't matter. */ | |
1293 | if (!TREE_OPERAND (ref, 2) | |
1294 | || TREE_CODE (TREE_OPERAND (ref, 2)) == INTEGER_CST) | |
1295 | ref = TREE_OPERAND (ref, 0); | |
1296 | /* Otherwise, if we have a bit field representative, use that. */ | |
1297 | else if (DECL_BIT_FIELD_REPRESENTATIVE (TREE_OPERAND (ref, 1)) | |
1298 | != NULL_TREE) | |
1299 | { | |
1300 | tree repr = DECL_BIT_FIELD_REPRESENTATIVE (TREE_OPERAND (ref, 1)); | |
1301 | ref = build3 (COMPONENT_REF, TREE_TYPE (repr), TREE_OPERAND (ref, 0), | |
1302 | repr, TREE_OPERAND (ref, 2)); | |
1303 | } | |
1304 | /* Otherwise punt. */ | |
1305 | else | |
1306 | { | |
1307 | dr->aux = strides; | |
1308 | return; | |
1309 | } | |
1310 | } | |
fbdec14e BC |
1311 | tree scev_base = build_fold_addr_expr (ref); |
1312 | tree scev = analyze_scalar_evolution (loop, scev_base); | |
1313 | scev = instantiate_scev (loop_preheader_edge (loop_nest), loop, scev); | |
1314 | if (! chrec_contains_undetermined (scev)) | |
1315 | { | |
1316 | tree sl = scev; | |
1317 | struct loop *expected = loop; | |
1318 | while (TREE_CODE (sl) == POLYNOMIAL_CHREC) | |
1319 | { | |
1320 | struct loop *sl_loop = get_chrec_loop (sl); | |
1321 | while (sl_loop != expected) | |
1322 | { | |
1323 | strides->safe_push (size_int (0)); | |
1324 | expected = loop_outer (expected); | |
1325 | } | |
1326 | strides->safe_push (CHREC_RIGHT (sl)); | |
1327 | sl = CHREC_LEFT (sl); | |
1328 | expected = loop_outer (expected); | |
1329 | } | |
1330 | if (! tree_contains_chrecs (sl, NULL)) | |
1331 | while (expected != loop_outer (loop_nest)) | |
1332 | { | |
1333 | strides->safe_push (size_int (0)); | |
1334 | expected = loop_outer (expected); | |
1335 | } | |
1336 | } | |
1337 | ||
1338 | dr->aux = strides; | |
1339 | } | |
1340 | ||
1341 | /* Given loop nest LOOP_NEST with innermost LOOP, the function computes | |
1342 | access strides with respect to each level loop for all data refs in | |
1343 | DATAREFS from inner loop to outer loop. On success, it returns the | |
1344 | outermost loop that access strides can be computed successfully for | |
1345 | all data references. If access strides cannot be computed at least | |
1346 | for two levels of loop for any data reference, it returns NULL. */ | |
1347 | ||
1348 | static struct loop * | |
1349 | compute_access_strides (struct loop *loop_nest, struct loop *loop, | |
1350 | vec<data_reference_p> datarefs) | |
1351 | { | |
1352 | unsigned i, j, num_loops = (unsigned) -1; | |
1353 | data_reference_p dr; | |
1354 | vec<tree> *stride; | |
1355 | ||
1356 | for (i = 0; datarefs.iterate (i, &dr); ++i) | |
1357 | { | |
1358 | compute_access_stride (loop_nest, loop, dr); | |
1359 | stride = DR_ACCESS_STRIDE (dr); | |
1360 | if (stride->length () < num_loops) | |
1361 | { | |
1362 | num_loops = stride->length (); | |
1363 | if (num_loops < 2) | |
1364 | return NULL; | |
1365 | } | |
1366 | } | |
1367 | ||
1368 | for (i = 0; datarefs.iterate (i, &dr); ++i) | |
1369 | { | |
1370 | stride = DR_ACCESS_STRIDE (dr); | |
1371 | if (stride->length () > num_loops) | |
1372 | stride->truncate (num_loops); | |
1373 | ||
1374 | for (j = 0; j < (num_loops >> 1); ++j) | |
1375 | std::swap ((*stride)[j], (*stride)[num_loops - j - 1]); | |
1376 | } | |
1377 | ||
1378 | loop = superloop_at_depth (loop, loop_depth (loop) + 1 - num_loops); | |
1379 | gcc_assert (loop_nest == loop || flow_loop_nested_p (loop_nest, loop)); | |
1380 | return loop; | |
1381 | } | |
1382 | ||
1383 | /* Prune access strides for data references in DATAREFS by removing strides | |
1384 | of loops that isn't in current LOOP_NEST. */ | |
1385 | ||
1386 | static void | |
1387 | prune_access_strides_not_in_loop (struct loop *loop_nest, | |
1388 | struct loop *innermost, | |
1389 | vec<data_reference_p> datarefs) | |
1390 | { | |
1391 | data_reference_p dr; | |
1392 | unsigned num_loops = loop_depth (innermost) - loop_depth (loop_nest) + 1; | |
1393 | gcc_assert (num_loops > 1); | |
1394 | ||
1395 | /* Block remove strides of loops that is not in current loop nest. */ | |
1396 | for (unsigned i = 0; datarefs.iterate (i, &dr); ++i) | |
1397 | { | |
1398 | vec<tree> *stride = DR_ACCESS_STRIDE (dr); | |
1399 | if (stride->length () > num_loops) | |
1400 | stride->block_remove (0, stride->length () - num_loops); | |
1401 | } | |
1402 | } | |
1403 | ||
1404 | /* Dump access strides for all DATAREFS. */ | |
1405 | ||
1406 | static void | |
1407 | dump_access_strides (vec<data_reference_p> datarefs) | |
1408 | { | |
1409 | data_reference_p dr; | |
1410 | fprintf (dump_file, "Access Strides for DRs:\n"); | |
1411 | for (unsigned i = 0; datarefs.iterate (i, &dr); ++i) | |
1412 | { | |
1413 | fprintf (dump_file, " "); | |
1414 | print_generic_expr (dump_file, DR_REF (dr), TDF_SLIM); | |
1415 | fprintf (dump_file, ":\t\t<"); | |
1416 | ||
1417 | vec<tree> *stride = DR_ACCESS_STRIDE (dr); | |
1418 | unsigned num_loops = stride->length (); | |
1419 | for (unsigned j = 0; j < num_loops; ++j) | |
1420 | { | |
1421 | print_generic_expr (dump_file, (*stride)[j], TDF_SLIM); | |
1422 | fprintf (dump_file, "%s", (j < num_loops - 1) ? ",\t" : ">\n"); | |
1423 | } | |
1424 | } | |
1425 | } | |
1426 | ||
1427 | /* Return true if it's profitable to interchange two loops whose index | |
1428 | in whole loop nest vector are I_IDX/O_IDX respectively. The function | |
1429 | computes and compares three types information from all DATAREFS: | |
1430 | 1) Access stride for loop I_IDX and O_IDX. | |
1431 | 2) Number of invariant memory references with respect to I_IDX before | |
1432 | and after loop interchange. | |
1433 | 3) Flags indicating if all memory references access sequential memory | |
1434 | in ILOOP, before and after loop interchange. | |
1435 | If INNMOST_LOOP_P is true, the two loops for interchanging are the two | |
1436 | innermost loops in loop nest. This function also dumps information if | |
1437 | DUMP_INFO_P is true. */ | |
1438 | ||
1439 | static bool | |
1440 | should_interchange_loops (unsigned i_idx, unsigned o_idx, | |
1441 | vec<data_reference_p> datarefs, | |
1eeeda47 | 1442 | unsigned i_stmt_cost, unsigned o_stmt_cost, |
fbdec14e BC |
1443 | bool innermost_loops_p, bool dump_info_p = true) |
1444 | { | |
1445 | unsigned HOST_WIDE_INT ratio; | |
1446 | unsigned i, j, num_old_inv_drs = 0, num_new_inv_drs = 0; | |
1447 | struct data_reference *dr; | |
1448 | bool all_seq_dr_before_p = true, all_seq_dr_after_p = true; | |
1449 | widest_int iloop_strides = 0, oloop_strides = 0; | |
1450 | unsigned num_unresolved_drs = 0; | |
1451 | unsigned num_resolved_ok_drs = 0; | |
1452 | unsigned num_resolved_not_ok_drs = 0; | |
1453 | ||
1454 | if (dump_info_p && dump_file && (dump_flags & TDF_DETAILS)) | |
1455 | fprintf (dump_file, "\nData ref strides:\n\tmem_ref:\t\tiloop\toloop\n"); | |
1456 | ||
1457 | for (i = 0; datarefs.iterate (i, &dr); ++i) | |
1458 | { | |
1459 | vec<tree> *stride = DR_ACCESS_STRIDE (dr); | |
1460 | tree iloop_stride = (*stride)[i_idx], oloop_stride = (*stride)[o_idx]; | |
1461 | ||
1462 | bool subloop_stride_p = false; | |
1463 | /* Data ref can't be invariant or sequential access at current loop if | |
1464 | its address changes with respect to any subloops. */ | |
1465 | for (j = i_idx + 1; j < stride->length (); ++j) | |
1466 | if (!integer_zerop ((*stride)[j])) | |
1467 | { | |
1468 | subloop_stride_p = true; | |
1469 | break; | |
1470 | } | |
1471 | ||
1472 | if (integer_zerop (iloop_stride)) | |
1473 | { | |
1474 | if (!subloop_stride_p) | |
1475 | num_old_inv_drs++; | |
1476 | } | |
1477 | if (integer_zerop (oloop_stride)) | |
1478 | { | |
1479 | if (!subloop_stride_p) | |
1480 | num_new_inv_drs++; | |
1481 | } | |
1482 | ||
1483 | if (TREE_CODE (iloop_stride) == INTEGER_CST | |
1484 | && TREE_CODE (oloop_stride) == INTEGER_CST) | |
1485 | { | |
1486 | iloop_strides = wi::add (iloop_strides, wi::to_widest (iloop_stride)); | |
1487 | oloop_strides = wi::add (oloop_strides, wi::to_widest (oloop_stride)); | |
1488 | } | |
1489 | else if (multiple_of_p (TREE_TYPE (iloop_stride), | |
1490 | iloop_stride, oloop_stride)) | |
1491 | num_resolved_ok_drs++; | |
1492 | else if (multiple_of_p (TREE_TYPE (iloop_stride), | |
1493 | oloop_stride, iloop_stride)) | |
1494 | num_resolved_not_ok_drs++; | |
1495 | else | |
1496 | num_unresolved_drs++; | |
1497 | ||
1498 | /* Data ref can't be sequential access if its address changes in sub | |
1499 | loop. */ | |
1500 | if (subloop_stride_p) | |
1501 | { | |
1502 | all_seq_dr_before_p = false; | |
1503 | all_seq_dr_after_p = false; | |
1504 | continue; | |
1505 | } | |
1506 | /* Track if all data references are sequential accesses before/after loop | |
1507 | interchange. Note invariant is considered sequential here. */ | |
1508 | tree access_size = TYPE_SIZE_UNIT (TREE_TYPE (DR_REF (dr))); | |
1509 | if (all_seq_dr_before_p | |
1510 | && ! (integer_zerop (iloop_stride) | |
1511 | || operand_equal_p (access_size, iloop_stride, 0))) | |
1512 | all_seq_dr_before_p = false; | |
1513 | if (all_seq_dr_after_p | |
1514 | && ! (integer_zerop (oloop_stride) | |
1515 | || operand_equal_p (access_size, oloop_stride, 0))) | |
1516 | all_seq_dr_after_p = false; | |
1517 | } | |
1518 | ||
1519 | if (dump_info_p && dump_file && (dump_flags & TDF_DETAILS)) | |
1520 | { | |
1521 | fprintf (dump_file, "\toverall:\t\t"); | |
1522 | print_decu (iloop_strides, dump_file); | |
1523 | fprintf (dump_file, "\t"); | |
1524 | print_decu (oloop_strides, dump_file); | |
1525 | fprintf (dump_file, "\n"); | |
1526 | ||
1527 | fprintf (dump_file, "Invariant data ref: before(%d), after(%d)\n", | |
1528 | num_old_inv_drs, num_new_inv_drs); | |
1529 | fprintf (dump_file, "All consecutive stride: before(%s), after(%s)\n", | |
1530 | all_seq_dr_before_p ? "true" : "false", | |
1531 | all_seq_dr_after_p ? "true" : "false"); | |
1532 | fprintf (dump_file, "OK to interchage with variable strides: %d\n", | |
1533 | num_resolved_ok_drs); | |
1534 | fprintf (dump_file, "Not OK to interchage with variable strides: %d\n", | |
1535 | num_resolved_not_ok_drs); | |
1536 | fprintf (dump_file, "Variable strides we cannot decide: %d\n", | |
1537 | num_unresolved_drs); | |
1eeeda47 BC |
1538 | fprintf (dump_file, "Stmt cost of inner loop: %d\n", i_stmt_cost); |
1539 | fprintf (dump_file, "Stmt cost of outer loop: %d\n", o_stmt_cost); | |
fbdec14e BC |
1540 | } |
1541 | ||
1542 | if (num_unresolved_drs != 0 || num_resolved_not_ok_drs != 0) | |
1543 | return false; | |
1544 | ||
1eeeda47 BC |
1545 | /* Stmts of outer loop will be moved to inner loop. If there are two many |
1546 | such stmts, it could make inner loop costly. Here we compare stmt cost | |
1547 | between outer and inner loops. */ | |
1548 | if (i_stmt_cost && o_stmt_cost | |
1549 | && num_old_inv_drs + o_stmt_cost > num_new_inv_drs | |
1550 | && o_stmt_cost * STMT_COST_RATIO > i_stmt_cost) | |
1551 | return false; | |
1552 | ||
fbdec14e BC |
1553 | /* We use different stride comparison ratio for interchanging innermost |
1554 | two loops or not. The idea is to be conservative in interchange for | |
1555 | the innermost loops. */ | |
1556 | ratio = innermost_loops_p ? INNER_STRIDE_RATIO : OUTER_STRIDE_RATIO; | |
1557 | /* Do interchange if it gives better data locality behavior. */ | |
1558 | if (wi::gtu_p (iloop_strides, wi::mul (oloop_strides, ratio))) | |
1559 | return true; | |
1560 | if (wi::gtu_p (iloop_strides, oloop_strides)) | |
1561 | { | |
1562 | /* Or it creates more invariant memory references. */ | |
1563 | if ((!all_seq_dr_before_p || all_seq_dr_after_p) | |
1564 | && num_new_inv_drs > num_old_inv_drs) | |
1565 | return true; | |
1566 | /* Or it makes all memory references sequential. */ | |
1567 | if (num_new_inv_drs >= num_old_inv_drs | |
1568 | && !all_seq_dr_before_p && all_seq_dr_after_p) | |
1569 | return true; | |
1570 | } | |
1571 | ||
1572 | return false; | |
1573 | } | |
1574 | ||
1575 | /* Try to interchange inner loop of a loop nest to outer level. */ | |
1576 | ||
1577 | bool | |
1578 | tree_loop_interchange::interchange (vec<data_reference_p> datarefs, | |
1579 | vec<ddr_p> ddrs) | |
1580 | { | |
4f5b9c80 | 1581 | dump_user_location_t loc = find_loop_location (m_loop_nest[0]); |
fbdec14e BC |
1582 | bool changed_p = false; |
1583 | /* In each iteration we try to interchange I-th loop with (I+1)-th loop. | |
1584 | The overall effect is to push inner loop to outermost level in whole | |
1585 | loop nest. */ | |
1586 | for (unsigned i = m_loop_nest.length (); i > 1; --i) | |
1587 | { | |
1588 | unsigned i_idx = i - 1, o_idx = i - 2; | |
1589 | ||
1590 | /* Check validity for loop interchange. */ | |
1591 | if (!valid_data_dependences (i_idx, o_idx, ddrs)) | |
1592 | break; | |
1593 | ||
1594 | loop_cand iloop (m_loop_nest[i_idx], m_loop_nest[o_idx]); | |
1595 | loop_cand oloop (m_loop_nest[o_idx], m_loop_nest[o_idx]); | |
1596 | ||
1597 | /* Check if we can do transformation for loop interchange. */ | |
1598 | if (!iloop.analyze_carried_vars (NULL) | |
1599 | || !iloop.analyze_lcssa_phis () | |
1600 | || !oloop.analyze_carried_vars (&iloop) | |
1601 | || !oloop.analyze_lcssa_phis () | |
1602 | || !iloop.can_interchange_p (NULL) | |
1603 | || !oloop.can_interchange_p (&iloop)) | |
1604 | break; | |
1605 | ||
1eeeda47 BC |
1606 | /* Outer loop's stmts will be moved to inner loop during interchange. |
1607 | If there are many of them, it may make inner loop very costly. We | |
1608 | need to check number of outer loop's stmts in profit cost model of | |
1609 | interchange. */ | |
1610 | int stmt_cost = oloop.m_num_stmts; | |
1611 | /* Count out the exit checking stmt of outer loop. */ | |
1612 | stmt_cost --; | |
1613 | /* Count out IV's increasing stmt, IVOPTs takes care if it. */ | |
1614 | stmt_cost -= oloop.m_inductions.length (); | |
1615 | /* Count in the additional load and cond_expr stmts caused by inner | |
1616 | loop's constant initialized reduction. */ | |
1617 | stmt_cost += iloop.m_const_init_reduc * 2; | |
1618 | if (stmt_cost < 0) | |
1619 | stmt_cost = 0; | |
1620 | ||
fbdec14e BC |
1621 | /* Check profitability for loop interchange. */ |
1622 | if (should_interchange_loops (i_idx, o_idx, datarefs, | |
1eeeda47 BC |
1623 | (unsigned) iloop.m_num_stmts, |
1624 | (unsigned) stmt_cost, | |
fbdec14e BC |
1625 | iloop.m_loop->inner == NULL)) |
1626 | { | |
1627 | if (dump_file && (dump_flags & TDF_DETAILS)) | |
1628 | fprintf (dump_file, | |
1629 | "Loop_pair<outer:%d, inner:%d> is interchanged\n\n", | |
1630 | oloop.m_loop->num, iloop.m_loop->num); | |
1631 | ||
1632 | changed_p = true; | |
1633 | interchange_loops (iloop, oloop); | |
1634 | /* No need to update if there is no further loop interchange. */ | |
1635 | if (o_idx > 0) | |
1636 | update_data_info (i_idx, o_idx, datarefs, ddrs); | |
1637 | } | |
1638 | else | |
1639 | { | |
1640 | if (dump_file && (dump_flags & TDF_DETAILS)) | |
1641 | fprintf (dump_file, | |
1642 | "Loop_pair<outer:%d, inner:%d> is not interchanged\n\n", | |
1643 | oloop.m_loop->num, iloop.m_loop->num); | |
1644 | } | |
1645 | } | |
fbdec14e | 1646 | simple_dce_from_worklist (m_dce_seeds); |
da472c1b RB |
1647 | |
1648 | if (changed_p) | |
1649 | dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, loc, | |
1650 | "loops interchanged in loop nest\n"); | |
1651 | ||
fbdec14e BC |
1652 | return changed_p; |
1653 | } | |
1654 | ||
1655 | ||
1656 | /* Loop interchange pass. */ | |
1657 | ||
1658 | namespace { | |
1659 | ||
1660 | const pass_data pass_data_linterchange = | |
1661 | { | |
1662 | GIMPLE_PASS, /* type */ | |
1663 | "linterchange", /* name */ | |
1664 | OPTGROUP_LOOP, /* optinfo_flags */ | |
1665 | TV_LINTERCHANGE, /* tv_id */ | |
1666 | PROP_cfg, /* properties_required */ | |
1667 | 0, /* properties_provided */ | |
1668 | 0, /* properties_destroyed */ | |
1669 | 0, /* todo_flags_start */ | |
1670 | 0, /* todo_flags_finish */ | |
1671 | }; | |
1672 | ||
1673 | class pass_linterchange : public gimple_opt_pass | |
1674 | { | |
1675 | public: | |
1676 | pass_linterchange (gcc::context *ctxt) | |
1677 | : gimple_opt_pass (pass_data_linterchange, ctxt) | |
1678 | {} | |
1679 | ||
1680 | /* opt_pass methods: */ | |
1681 | opt_pass * clone () { return new pass_linterchange (m_ctxt); } | |
1682 | virtual bool gate (function *) { return flag_loop_interchange; } | |
1683 | virtual unsigned int execute (function *); | |
1684 | ||
1685 | }; // class pass_linterchange | |
1686 | ||
1687 | ||
1688 | /* Return true if LOOP has proper form for interchange. We check three | |
1689 | conditions in the function: | |
1690 | 1) In general, a loop can be interchanged only if it doesn't have | |
1691 | basic blocks other than header, exit and latch besides possible | |
1692 | inner loop nest. This basically restricts loop interchange to | |
1693 | below form loop nests: | |
1694 | ||
1695 | header<---+ | |
1696 | | | | |
1697 | v | | |
1698 | INNER_LOOP | | |
1699 | | | | |
1700 | v | | |
1701 | exit--->latch | |
1702 | ||
1703 | 2) Data reference in basic block that executes in different times | |
1704 | than loop head/exit is not allowed. | |
1705 | 3) Record the innermost outer loop that doesn't form rectangle loop | |
1706 | nest with LOOP. */ | |
1707 | ||
1708 | static bool | |
1709 | proper_loop_form_for_interchange (struct loop *loop, struct loop **min_outer) | |
1710 | { | |
1711 | edge e0, e1, exit; | |
1712 | ||
1713 | /* Don't interchange if loop has unsupported information for the moment. */ | |
1714 | if (loop->safelen > 0 | |
1715 | || loop->constraints != 0 | |
1716 | || loop->can_be_parallel | |
1717 | || loop->dont_vectorize | |
1718 | || loop->force_vectorize | |
1719 | || loop->in_oacc_kernels_region | |
1720 | || loop->orig_loop_num != 0 | |
1721 | || loop->simduid != NULL_TREE) | |
1722 | return false; | |
1723 | ||
1724 | /* Don't interchange if outer loop has basic block other than header, exit | |
1725 | and latch. */ | |
1726 | if (loop->inner != NULL | |
1727 | && loop->num_nodes != loop->inner->num_nodes + 3) | |
1728 | return false; | |
1729 | ||
1730 | if ((exit = single_dom_exit (loop)) == NULL) | |
1731 | return false; | |
1732 | ||
1733 | /* Check control flow on loop header/exit blocks. */ | |
1734 | if (loop->header == exit->src | |
1735 | && (EDGE_COUNT (loop->header->preds) != 2 | |
1736 | || EDGE_COUNT (loop->header->succs) != 2)) | |
1737 | return false; | |
1738 | else if (loop->header != exit->src | |
1739 | && (EDGE_COUNT (loop->header->preds) != 2 | |
1740 | || !single_succ_p (loop->header) | |
1741 | || unsupported_edge (single_succ_edge (loop->header)) | |
1742 | || EDGE_COUNT (exit->src->succs) != 2 | |
1743 | || !single_pred_p (exit->src) | |
1744 | || unsupported_edge (single_pred_edge (exit->src)))) | |
1745 | return false; | |
1746 | ||
1747 | e0 = EDGE_PRED (loop->header, 0); | |
1748 | e1 = EDGE_PRED (loop->header, 1); | |
1749 | if (unsupported_edge (e0) || unsupported_edge (e1) | |
1750 | || (e0->src != loop->latch && e1->src != loop->latch) | |
1751 | || (e0->src->loop_father == loop && e1->src->loop_father == loop)) | |
1752 | return false; | |
1753 | ||
1754 | e0 = EDGE_SUCC (exit->src, 0); | |
1755 | e1 = EDGE_SUCC (exit->src, 1); | |
1756 | if (unsupported_edge (e0) || unsupported_edge (e1) | |
1757 | || (e0->dest != loop->latch && e1->dest != loop->latch) | |
1758 | || (e0->dest->loop_father == loop && e1->dest->loop_father == loop)) | |
1759 | return false; | |
1760 | ||
1761 | /* Don't interchange if any reference is in basic block that doesn't | |
1762 | dominate exit block. */ | |
1763 | basic_block *bbs = get_loop_body (loop); | |
1764 | for (unsigned i = 0; i < loop->num_nodes; i++) | |
1765 | { | |
1766 | basic_block bb = bbs[i]; | |
1767 | ||
1768 | if (bb->loop_father != loop | |
1769 | || bb == loop->header || bb == exit->src | |
1770 | || dominated_by_p (CDI_DOMINATORS, exit->src, bb)) | |
1771 | continue; | |
1772 | ||
65f4b875 | 1773 | for (gimple_stmt_iterator gsi = gsi_start_nondebug_bb (bb); |
fbdec14e BC |
1774 | !gsi_end_p (gsi); gsi_next_nondebug (&gsi)) |
1775 | if (gimple_vuse (gsi_stmt (gsi))) | |
1776 | { | |
1777 | free (bbs); | |
1778 | return false; | |
1779 | } | |
1780 | } | |
1781 | free (bbs); | |
1782 | ||
1783 | tree niters = number_of_latch_executions (loop); | |
1784 | niters = analyze_scalar_evolution (loop_outer (loop), niters); | |
1785 | if (!niters || chrec_contains_undetermined (niters)) | |
1786 | return false; | |
1787 | ||
1788 | /* Record the innermost outer loop that doesn't form rectangle loop nest. */ | |
1789 | for (loop_p loop2 = loop_outer (loop); | |
1790 | loop2 && flow_loop_nested_p (*min_outer, loop2); | |
1791 | loop2 = loop_outer (loop2)) | |
1792 | { | |
1793 | niters = instantiate_scev (loop_preheader_edge (loop2), | |
1794 | loop_outer (loop), niters); | |
1795 | if (!evolution_function_is_invariant_p (niters, loop2->num)) | |
1796 | { | |
1797 | *min_outer = loop2; | |
1798 | break; | |
1799 | } | |
1800 | } | |
1801 | return true; | |
1802 | } | |
1803 | ||
1804 | /* Return true if any two adjacent loops in loop nest [INNERMOST, LOOP_NEST] | |
1805 | should be interchanged by looking into all DATAREFS. */ | |
1806 | ||
1807 | static bool | |
1808 | should_interchange_loop_nest (struct loop *loop_nest, struct loop *innermost, | |
1809 | vec<data_reference_p> datarefs) | |
1810 | { | |
1811 | unsigned idx = loop_depth (innermost) - loop_depth (loop_nest); | |
1812 | gcc_assert (idx > 0); | |
1813 | ||
1814 | /* Check if any two adjacent loops should be interchanged. */ | |
1815 | for (struct loop *loop = innermost; | |
1816 | loop != loop_nest; loop = loop_outer (loop), idx--) | |
1eeeda47 | 1817 | if (should_interchange_loops (idx, idx - 1, datarefs, 0, 0, |
fbdec14e BC |
1818 | loop == innermost, false)) |
1819 | return true; | |
1820 | ||
1821 | return false; | |
1822 | } | |
1823 | ||
1824 | /* Given loop nest LOOP_NEST and data references DATAREFS, compute data | |
1825 | dependences for loop interchange and store it in DDRS. Note we compute | |
1826 | dependences directly rather than call generic interface so that we can | |
1827 | return on unknown dependence instantly. */ | |
1828 | ||
1829 | static bool | |
1830 | tree_loop_interchange_compute_ddrs (vec<loop_p> loop_nest, | |
1831 | vec<data_reference_p> datarefs, | |
1832 | vec<ddr_p> *ddrs) | |
1833 | { | |
1834 | struct data_reference *a, *b; | |
1835 | struct loop *innermost = loop_nest.last (); | |
1836 | ||
1837 | for (unsigned i = 0; datarefs.iterate (i, &a); ++i) | |
1838 | { | |
1839 | bool a_outer_p = gimple_bb (DR_STMT (a))->loop_father != innermost; | |
1840 | for (unsigned j = i + 1; datarefs.iterate (j, &b); ++j) | |
1841 | if (DR_IS_WRITE (a) || DR_IS_WRITE (b)) | |
1842 | { | |
1843 | bool b_outer_p = gimple_bb (DR_STMT (b))->loop_father != innermost; | |
1844 | /* Don't support multiple write references in outer loop. */ | |
1845 | if (a_outer_p && b_outer_p && DR_IS_WRITE (a) && DR_IS_WRITE (b)) | |
1846 | return false; | |
1847 | ||
1848 | ddr_p ddr = initialize_data_dependence_relation (a, b, loop_nest); | |
1849 | ddrs->safe_push (ddr); | |
1850 | compute_affine_dependence (ddr, loop_nest[0]); | |
1851 | ||
1852 | /* Give up if ddr is unknown dependence or classic direct vector | |
1853 | is not available. */ | |
1854 | if (DDR_ARE_DEPENDENT (ddr) == chrec_dont_know | |
1855 | || (DDR_ARE_DEPENDENT (ddr) == NULL_TREE | |
1856 | && DDR_NUM_DIR_VECTS (ddr) == 0)) | |
1857 | return false; | |
1858 | ||
1859 | /* If either data references is in outer loop of nest, we require | |
1860 | no dependence here because the data reference need to be moved | |
1861 | into inner loop during interchange. */ | |
1862 | if (a_outer_p && b_outer_p | |
1863 | && operand_equal_p (DR_REF (a), DR_REF (b), 0)) | |
1864 | continue; | |
1865 | if (DDR_ARE_DEPENDENT (ddr) != chrec_known | |
1866 | && (a_outer_p || b_outer_p)) | |
1867 | return false; | |
1868 | } | |
1869 | } | |
1870 | ||
1871 | return true; | |
1872 | } | |
1873 | ||
1874 | /* Prune DATAREFS by removing any data reference not inside of LOOP. */ | |
1875 | ||
1876 | static inline void | |
1877 | prune_datarefs_not_in_loop (struct loop *loop, vec<data_reference_p> datarefs) | |
1878 | { | |
1879 | unsigned i, j; | |
1880 | struct data_reference *dr; | |
1881 | ||
1882 | for (i = 0, j = 0; datarefs.iterate (i, &dr); ++i) | |
1883 | { | |
1884 | if (flow_bb_inside_loop_p (loop, gimple_bb (DR_STMT (dr)))) | |
1885 | datarefs[j++] = dr; | |
1886 | else | |
1887 | { | |
1888 | if (dr->aux) | |
1889 | { | |
1890 | DR_ACCESS_STRIDE (dr)->release (); | |
46bb9d29 | 1891 | delete (vec<tree> *) dr->aux; |
fbdec14e BC |
1892 | } |
1893 | free_data_ref (dr); | |
1894 | } | |
1895 | } | |
1896 | datarefs.truncate (j); | |
1897 | } | |
1898 | ||
1899 | /* Find and store data references in DATAREFS for LOOP nest. If there's | |
1900 | difficult data reference in a basic block, we shrink the loop nest to | |
1901 | inner loop of that basic block's father loop. On success, return the | |
1902 | outer loop of the result loop nest. */ | |
1903 | ||
1904 | static struct loop * | |
1905 | prepare_data_references (struct loop *loop, vec<data_reference_p> *datarefs) | |
1906 | { | |
1907 | struct loop *loop_nest = loop; | |
1908 | vec<data_reference_p> *bb_refs; | |
1909 | basic_block bb, *bbs = get_loop_body_in_dom_order (loop); | |
1910 | ||
1911 | for (unsigned i = 0; i < loop->num_nodes; i++) | |
1912 | bbs[i]->aux = NULL; | |
1913 | ||
1914 | /* Find data references for all basic blocks. Shrink loop nest on difficult | |
1915 | data reference. */ | |
1916 | for (unsigned i = 0; loop_nest && i < loop->num_nodes; ++i) | |
1917 | { | |
1918 | bb = bbs[i]; | |
1919 | if (!flow_bb_inside_loop_p (loop_nest, bb)) | |
1920 | continue; | |
1921 | ||
1922 | bb_refs = new vec<data_reference_p> (); | |
1923 | if (find_data_references_in_bb (loop, bb, bb_refs) == chrec_dont_know) | |
1924 | { | |
1925 | loop_nest = bb->loop_father->inner; | |
1926 | if (loop_nest && !loop_nest->inner) | |
1927 | loop_nest = NULL; | |
1928 | ||
1929 | free_data_refs (*bb_refs); | |
1930 | delete bb_refs; | |
1931 | } | |
1932 | else if (bb_refs->is_empty ()) | |
1933 | delete bb_refs; | |
1934 | else | |
1935 | bb->aux = bb_refs; | |
1936 | } | |
1937 | ||
1938 | /* Collect all data references in loop nest. */ | |
1939 | for (unsigned i = 0; i < loop->num_nodes; i++) | |
1940 | { | |
1941 | bb = bbs[i]; | |
1942 | if (!bb->aux) | |
1943 | continue; | |
1944 | ||
1945 | bb_refs = (vec<data_reference_p> *) bb->aux; | |
1946 | if (loop_nest && flow_bb_inside_loop_p (loop_nest, bb)) | |
1947 | datarefs->safe_splice (*bb_refs); | |
1948 | else | |
1949 | free_data_refs (*bb_refs); | |
1950 | ||
1951 | delete bb_refs; | |
1952 | bb->aux = NULL; | |
1953 | } | |
1954 | free (bbs); | |
1955 | ||
1956 | return loop_nest; | |
1957 | } | |
1958 | ||
1959 | /* Given innermost LOOP, return true if perfect loop nest can be found and | |
1960 | data dependences can be computed. If succeed, record the perfect loop | |
1961 | nest in LOOP_NEST; record all data references in DATAREFS and record all | |
1962 | data dependence relations in DDRS. | |
1963 | ||
1964 | We do support a restricted form of imperfect loop nest, i.e, loop nest | |
1965 | with load/store in outer loop initializing/finalizing simple reduction | |
1966 | of the innermost loop. For such outer loop reference, we require that | |
1967 | it has no dependence with others sinve it will be moved to inner loop | |
1968 | in interchange. */ | |
1969 | ||
1970 | static bool | |
1971 | prepare_perfect_loop_nest (struct loop *loop, vec<loop_p> *loop_nest, | |
1972 | vec<data_reference_p> *datarefs, vec<ddr_p> *ddrs) | |
1973 | { | |
1974 | struct loop *start_loop = NULL, *innermost = loop; | |
1975 | struct loop *outermost = loops_for_fn (cfun)->tree_root; | |
1976 | ||
1977 | /* Find loop nest from the innermost loop. The outermost is the innermost | |
1978 | outer*/ | |
1979 | while (loop->num != 0 && loop->inner == start_loop | |
1980 | && flow_loop_nested_p (outermost, loop)) | |
1981 | { | |
1982 | if (!proper_loop_form_for_interchange (loop, &outermost)) | |
1983 | break; | |
1984 | ||
1985 | start_loop = loop; | |
1986 | /* If this loop has sibling loop, the father loop won't be in perfect | |
1987 | loop nest. */ | |
1988 | if (loop->next != NULL) | |
1989 | break; | |
1990 | ||
1991 | loop = loop_outer (loop); | |
1992 | } | |
1993 | if (!start_loop || !start_loop->inner) | |
1994 | return false; | |
1995 | ||
1996 | /* Prepare the data reference vector for the loop nest, pruning outer | |
1997 | loops we cannot handle. */ | |
1998 | start_loop = prepare_data_references (start_loop, datarefs); | |
1999 | if (!start_loop | |
2000 | /* Check if there is no data reference. */ | |
2001 | || datarefs->is_empty () | |
2002 | /* Check if there are too many of data references. */ | |
2003 | || (int) datarefs->length () > MAX_DATAREFS) | |
2004 | return false; | |
2005 | ||
2006 | /* Compute access strides for all data references, pruning outer | |
2007 | loops we cannot analyze refs in. */ | |
2008 | start_loop = compute_access_strides (start_loop, innermost, *datarefs); | |
2009 | if (!start_loop) | |
2010 | return false; | |
2011 | ||
2012 | /* Check if any interchange is profitable in the loop nest. */ | |
2013 | if (!should_interchange_loop_nest (start_loop, innermost, *datarefs)) | |
2014 | return false; | |
2015 | ||
2016 | /* Check if data dependences can be computed for loop nest starting from | |
2017 | start_loop. */ | |
2018 | loop = start_loop; | |
2019 | do { | |
2020 | loop_nest->truncate (0); | |
2021 | ||
2022 | if (loop != start_loop) | |
2023 | prune_datarefs_not_in_loop (start_loop, *datarefs); | |
2024 | ||
2025 | if (find_loop_nest (start_loop, loop_nest) | |
2026 | && tree_loop_interchange_compute_ddrs (*loop_nest, *datarefs, ddrs)) | |
2027 | { | |
2028 | if (dump_file && (dump_flags & TDF_DETAILS)) | |
2029 | fprintf (dump_file, | |
2030 | "\nConsider loop interchange for loop_nest<%d - %d>\n", | |
2031 | start_loop->num, innermost->num); | |
2032 | ||
2033 | if (loop != start_loop) | |
2034 | prune_access_strides_not_in_loop (start_loop, innermost, *datarefs); | |
2035 | ||
2036 | if (dump_file && (dump_flags & TDF_DETAILS)) | |
2037 | dump_access_strides (*datarefs); | |
2038 | ||
2039 | return true; | |
2040 | } | |
2041 | ||
2042 | free_dependence_relations (*ddrs); | |
2043 | *ddrs = vNULL; | |
2044 | /* Try to compute data dependences with the outermost loop stripped. */ | |
2045 | loop = start_loop; | |
2046 | start_loop = start_loop->inner; | |
2047 | } while (start_loop && start_loop->inner); | |
2048 | ||
2049 | return false; | |
2050 | } | |
2051 | ||
2052 | /* Main entry for loop interchange pass. */ | |
2053 | ||
2054 | unsigned int | |
2055 | pass_linterchange::execute (function *fun) | |
2056 | { | |
2057 | if (number_of_loops (fun) <= 2) | |
2058 | return 0; | |
2059 | ||
2060 | bool changed_p = false; | |
2061 | struct loop *loop; | |
2062 | FOR_EACH_LOOP (loop, LI_ONLY_INNERMOST) | |
2063 | { | |
2064 | vec<loop_p> loop_nest = vNULL; | |
2065 | vec<data_reference_p> datarefs = vNULL; | |
2066 | vec<ddr_p> ddrs = vNULL; | |
2067 | if (prepare_perfect_loop_nest (loop, &loop_nest, &datarefs, &ddrs)) | |
2068 | { | |
2069 | tree_loop_interchange loop_interchange (loop_nest); | |
2070 | changed_p |= loop_interchange.interchange (datarefs, ddrs); | |
2071 | } | |
2072 | free_dependence_relations (ddrs); | |
2073 | free_data_refs_with_aux (datarefs); | |
2074 | loop_nest.release (); | |
2075 | } | |
2076 | ||
fbdec14e BC |
2077 | return changed_p ? (TODO_update_ssa_only_virtuals) : 0; |
2078 | } | |
2079 | ||
2080 | } // anon namespace | |
2081 | ||
2082 | gimple_opt_pass * | |
2083 | make_pass_linterchange (gcc::context *ctxt) | |
2084 | { | |
2085 | return new pass_linterchange (ctxt); | |
2086 | } |