]> git.ipfire.org Git - thirdparty/gcc.git/blame - gcc/ipa-inline-transform.c
Improve gimple.vim syntax file.
[thirdparty/gcc.git] / gcc / ipa-inline-transform.c
CommitLineData
fee8b6da 1/* Callgraph transformations to handle inlining
85ec4feb 2 Copyright (C) 2003-2018 Free Software Foundation, Inc.
fee8b6da
JH
3 Contributed by Jan Hubicka
4
5This file is part of GCC.
6
7GCC is free software; you can redistribute it and/or modify it under
8the terms of the GNU General Public License as published by the Free
9Software Foundation; either version 3, or (at your option) any later
10version.
11
12GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13WARRANTY; without even the implied warranty of MERCHANTABILITY or
14FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
15for more details.
16
17You should have received a copy of the GNU General Public License
18along with GCC; see the file COPYING3. If not see
19<http://www.gnu.org/licenses/>. */
20
21/* The inline decisions are stored in callgraph in "inline plan" and
22 applied later.
23
24 To mark given call inline, use inline_call function.
25 The function marks the edge inlinable and, if necessary, produces
26 virtual clone in the callgraph representing the new copy of callee's
27 function body.
28
29 The inline plan is applied on given function body by inline_transform. */
30
31#include "config.h"
32#include "system.h"
33#include "coretypes.h"
34#include "tm.h"
957060b5 35#include "function.h"
fee8b6da 36#include "tree.h"
957060b5
AM
37#include "alloc-pool.h"
38#include "tree-pass.h"
39#include "cgraph.h"
442b4905 40#include "tree-cfg.h"
dd912cb8 41#include "symbol-summary.h"
8bc5448f 42#include "tree-vrp.h"
fee8b6da 43#include "ipa-prop.h"
27d020cf 44#include "ipa-fnsummary.h"
fee8b6da
JH
45#include "ipa-inline.h"
46#include "tree-inline.h"
09fcc0c0
JH
47#include "function.h"
48#include "cfg.h"
49#include "basic-block.h"
fee8b6da
JH
50
51int ncalls_inlined;
52int nfunctions_inlined;
53
1bad9c18 54/* Scale counts of NODE edges by NUM/DEN. */
fee8b6da
JH
55
56static void
1bad9c18
JH
57update_noncloned_counts (struct cgraph_node *node,
58 profile_count num, profile_count den)
fee8b6da
JH
59{
60 struct cgraph_edge *e;
8e7d1486 61
e4373d41 62 profile_count::adjust_for_ipa_scaling (&num, &den);
fee8b6da 63
fee8b6da
JH
64 for (e = node->callees; e; e = e->next_callee)
65 {
fee8b6da 66 if (!e->inline_failed)
1bad9c18 67 update_noncloned_counts (e->callee, num, den);
8e7d1486 68 e->count = e->count.apply_scale (num, den);
898b8927
JH
69 }
70 for (e = node->indirect_calls; e; e = e->next_callee)
1bad9c18 71 e->count = e->count.apply_scale (num, den);
8e7d1486 72 node->count = node->count.apply_scale (num, den);
fee8b6da
JH
73}
74
a5b1779f
JH
75/* We removed or are going to remove the last call to NODE.
76 Return true if we can and want proactively remove the NODE now.
77 This is important to do, since we want inliner to know when offline
78 copy of function was removed. */
79
80static bool
d142079a 81can_remove_node_now_p_1 (struct cgraph_node *node, struct cgraph_edge *e)
a5b1779f 82{
d142079a
JH
83 ipa_ref *ref;
84
85 FOR_EACH_ALIAS (node, ref)
86 {
87 cgraph_node *alias = dyn_cast <cgraph_node *> (ref->referring);
88 if ((alias->callers && alias->callers != e)
89 || !can_remove_node_now_p_1 (alias, e))
90 return false;
91 }
a5b1779f
JH
92 /* FIXME: When address is taken of DECL_EXTERNAL function we still
93 can remove its offline copy, but we would need to keep unanalyzed node in
a6a543bf
JH
94 the callgraph so references can point to it.
95
96 Also for comdat group we can ignore references inside a group as we
97 want to prove the group as a whole to be dead. */
67348ccc 98 return (!node->address_taken
a6a543bf 99 && node->can_remove_if_no_direct_calls_and_refs_p ()
a5b1779f
JH
100 /* Inlining might enable more devirtualizing, so we want to remove
101 those only after all devirtualizable virtual calls are processed.
102 Lacking may edges in callgraph we just preserve them post
103 inlining. */
d142079a
JH
104 && (!DECL_VIRTUAL_P (node->decl)
105 || !opt_for_fn (node->decl, flag_devirtualize))
a5b1779f
JH
106 /* During early inlining some unanalyzed cgraph nodes might be in the
107 callgraph and they might reffer the function in question. */
31acf1bb 108 && !cgraph_new_nodes.exists ());
a5b1779f
JH
109}
110
6c69a029
JH
111/* We are going to eliminate last direct call to NODE (or alias of it) via edge E.
112 Verify that the NODE can be removed from unit and if it is contained in comdat
113 group that the whole comdat group is removable. */
114
115static bool
116can_remove_node_now_p (struct cgraph_node *node, struct cgraph_edge *e)
117{
118 struct cgraph_node *next;
d142079a 119 if (!can_remove_node_now_p_1 (node, e))
6c69a029
JH
120 return false;
121
122 /* When we see same comdat group, we need to be sure that all
123 items can be removed. */
8ccc8042 124 if (!node->same_comdat_group || !node->externally_visible)
6c69a029 125 return true;
d52f5295
ML
126 for (next = dyn_cast<cgraph_node *> (node->same_comdat_group);
127 next != node; next = dyn_cast<cgraph_node *> (next->same_comdat_group))
d142079a
JH
128 {
129 if (next->alias)
130 continue;
131 if ((next->callers && next->callers != e)
132 || !can_remove_node_now_p_1 (next, e))
133 return false;
134 }
6c69a029
JH
135 return true;
136}
137
d83fa499
EB
138/* Return true if NODE is a master clone with non-inline clones. */
139
140static bool
141master_clone_with_noninline_clones_p (struct cgraph_node *node)
142{
143 if (node->clone_of)
144 return false;
145
146 for (struct cgraph_node *n = node->clones; n; n = n->next_sibling_clone)
147 if (n->decl != node->decl)
148 return true;
149
150 return false;
151}
fee8b6da
JH
152
153/* E is expected to be an edge being inlined. Clone destination node of
154 the edge and redirect it to the new clone.
155 DUPLICATE is used for bookkeeping on whether we are actually creating new
156 clones or re-using node originally representing out-of-line function call.
bd936951
JH
157 By default the offline copy is removed, when it appears dead after inlining.
158 UPDATE_ORIGINAL prevents this transformation.
159 If OVERALL_SIZE is non-NULL, the size is updated to reflect the
1bad9c18 160 transformation. */
fee8b6da
JH
161
162void
163clone_inlined_nodes (struct cgraph_edge *e, bool duplicate,
1bad9c18 164 bool update_original, int *overall_size)
fee8b6da 165{
44a60244 166 struct cgraph_node *inlining_into;
09ce3660 167 struct cgraph_edge *next;
44a60244
MJ
168
169 if (e->caller->global.inlined_to)
170 inlining_into = e->caller->global.inlined_to;
171 else
172 inlining_into = e->caller;
173
fee8b6da
JH
174 if (duplicate)
175 {
176 /* We may eliminate the need for out-of-line copy to be output.
177 In that case just go ahead and re-use it. This is not just an
178 memory optimization. Making offline copy of fuction disappear
179 from the program will improve future decisions on inlining. */
180 if (!e->callee->callers->next_caller
181 /* Recursive inlining never wants the master clone to
182 be overwritten. */
183 && update_original
d83fa499
EB
184 && can_remove_node_now_p (e->callee, e)
185 /* We cannot overwrite a master clone with non-inline clones
186 until after these clones are materialized. */
187 && !master_clone_with_noninline_clones_p (e->callee))
fee8b6da 188 {
6c69a029
JH
189 /* TODO: When callee is in a comdat group, we could remove all of it,
190 including all inline clones inlined into it. That would however
191 need small function inlining to register edge removal hook to
192 maintain the priority queue.
193
194 For now we keep the ohter functions in the group in program until
195 cgraph_remove_unreachable_functions gets rid of them. */
fee8b6da 196 gcc_assert (!e->callee->global.inlined_to);
b91b562c 197 e->callee->remove_from_same_comdat_group ();
bb1e543c
JH
198 if (e->callee->definition
199 && inline_account_function_p (e->callee))
fee8b6da 200 {
bb1e543c 201 gcc_assert (!e->callee->alias);
fee8b6da 202 if (overall_size)
99353fcf 203 *overall_size -= ipa_fn_summaries->get_create (e->callee)->size;
fee8b6da
JH
204 nfunctions_inlined++;
205 }
206 duplicate = false;
67348ccc 207 e->callee->externally_visible = false;
1bad9c18 208 update_noncloned_counts (e->callee, e->count, e->callee->count);
0bdad123
ML
209
210 dump_callgraph_transformation (e->callee, inlining_into,
211 "inlining to");
fee8b6da
JH
212 }
213 else
214 {
215 struct cgraph_node *n;
bd936951 216
d52f5295 217 n = e->callee->create_clone (e->callee->decl,
1bad9c18 218 e->count,
d52f5295
ML
219 update_original, vNULL, true,
220 inlining_into,
221 NULL);
4ad08ee8 222 n->used_as_abstract_origin = e->callee->used_as_abstract_origin;
3dafb85c 223 e->redirect_callee (n);
fee8b6da
JH
224 }
225 }
65d630d4 226 else
b91b562c 227 e->callee->remove_from_same_comdat_group ();
fee8b6da 228
44a60244 229 e->callee->global.inlined_to = inlining_into;
fee8b6da
JH
230
231 /* Recursively clone all bodies. */
09ce3660
JH
232 for (e = e->callee->callees; e; e = next)
233 {
234 next = e->next_callee;
235 if (!e->inline_failed)
1bad9c18 236 clone_inlined_nodes (e, duplicate, update_original, overall_size);
db66bf68
JH
237 }
238}
239
240/* Check all speculations in N and resolve them if they seems useless. */
241
242static bool
243check_speculations (cgraph_node *n)
244{
245 bool speculation_removed = false;
246 cgraph_edge *next;
247
248 for (cgraph_edge *e = n->callees; e; e = next)
249 {
250 next = e->next_callee;
09ce3660
JH
251 if (e->speculative && !speculation_useful_p (e, true))
252 {
3dafb85c 253 e->resolve_speculation (NULL);
09ce3660
JH
254 speculation_removed = true;
255 }
db66bf68
JH
256 else if (!e->inline_failed)
257 speculation_removed |= check_speculations (e->callee);
09ce3660 258 }
db66bf68 259 return speculation_removed;
fee8b6da
JH
260}
261
4fd94d1e
MJ
262/* Mark all call graph edges coming out of NODE and all nodes that have been
263 inlined to it as in_polymorphic_cdtor. */
264
265static void
266mark_all_inlined_calls_cdtor (cgraph_node *node)
267{
268 for (cgraph_edge *cs = node->callees; cs; cs = cs->next_callee)
269 {
270 cs->in_polymorphic_cdtor = true;
271 if (!cs->inline_failed)
09fcc0c0 272 mark_all_inlined_calls_cdtor (cs->callee);
4fd94d1e
MJ
273 }
274 for (cgraph_edge *cs = node->indirect_calls; cs; cs = cs->next_callee)
275 cs->in_polymorphic_cdtor = true;
276}
277
fee8b6da
JH
278
279/* Mark edge E as inlined and update callgraph accordingly. UPDATE_ORIGINAL
280 specify whether profile of original function should be updated. If any new
281 indirect edges are discovered in the process, add them to NEW_EDGES, unless
c170d40f
JH
282 it is NULL. If UPDATE_OVERALL_SUMMARY is false, do not bother to recompute overall
283 size of caller after inlining. Caller is required to eventually do it via
0bceb671 284 ipa_update_overall_fn_summary.
1bbb87c4 285 If callee_removed is non-NULL, set it to true if we removed callee node.
c170d40f
JH
286
287 Return true iff any new callgraph edges were discovered as a
fee8b6da
JH
288 result of inlining. */
289
290bool
291inline_call (struct cgraph_edge *e, bool update_original,
d52f5295 292 vec<cgraph_edge *> *new_edges,
1bbb87c4
JH
293 int *overall_size, bool update_overall_summary,
294 bool *callee_removed)
fee8b6da
JH
295{
296 int old_size = 0, new_size = 0;
297 struct cgraph_node *to = NULL;
298 struct cgraph_edge *curr = e;
d52f5295 299 struct cgraph_node *callee = e->callee->ultimate_alias_target ();
0f378cb5
JH
300 bool new_edges_found = false;
301
bddead15
RB
302 int estimated_growth = 0;
303 if (! update_overall_summary)
304 estimated_growth = estimate_edge_growth (e);
f107227b
JH
305 /* This is used only for assert bellow. */
306#if 0
0f378cb5
JH
307 bool predicated = inline_edge_summary (e)->predicate != NULL;
308#endif
fee8b6da
JH
309
310 /* Don't inline inlined edges. */
311 gcc_assert (e->inline_failed);
312 /* Don't even think of inlining inline clone. */
a5b1779f 313 gcc_assert (!callee->global.inlined_to);
fee8b6da 314
632b4f8e
JH
315 to = e->caller;
316 if (to->global.inlined_to)
317 to = to->global.inlined_to;
0b9004ed
JH
318 if (to->thunk.thunk_p)
319 {
c7ed8938 320 struct cgraph_node *target = to->callees->callee;
0b9004ed
JH
321 if (in_lto_p)
322 to->get_untransformed_body ();
323 to->expand_thunk (false, true);
c7ed8938
IE
324 /* When thunk is instrumented we may have multiple callees. */
325 for (e = to->callees; e && e->callee != target; e = e->next_callee)
326 ;
327 gcc_assert (e);
0b9004ed
JH
328 }
329
330
331 e->inline_failed = CIF_OK;
332 DECL_POSSIBLY_INLINED (callee->decl) = true;
632b4f8e 333
5058c037
JH
334 if (DECL_FUNCTION_PERSONALITY (callee->decl))
335 DECL_FUNCTION_PERSONALITY (to->decl)
336 = DECL_FUNCTION_PERSONALITY (callee->decl);
77719b06
ML
337
338 bool reload_optimization_node = false;
45285060
JH
339 if (!opt_for_fn (callee->decl, flag_strict_aliasing)
340 && opt_for_fn (to->decl, flag_strict_aliasing))
341 {
342 struct gcc_options opts = global_options;
343
f7f32acd 344 cl_optimization_restore (&opts, opts_for_fn (to->decl));
45285060
JH
345 opts.x_flag_strict_aliasing = false;
346 if (dump_file)
464d0118
ML
347 fprintf (dump_file, "Dropping flag_strict_aliasing on %s\n",
348 to->dump_name ());
45285060
JH
349 DECL_FUNCTION_SPECIFIC_OPTIMIZATION (to->decl)
350 = build_optimization_node (&opts);
77719b06 351 reload_optimization_node = true;
45285060 352 }
5c846a81 353
99353fcf
ML
354 ipa_fn_summary *caller_info = ipa_fn_summaries->get_create (to);
355 ipa_fn_summary *callee_info = ipa_fn_summaries->get_create (callee);
818b88a7
JH
356 if (!caller_info->fp_expressions && callee_info->fp_expressions)
357 {
358 caller_info->fp_expressions = true;
359 if (opt_for_fn (callee->decl, flag_rounding_math)
360 != opt_for_fn (to->decl, flag_rounding_math)
361 || opt_for_fn (callee->decl, flag_trapping_math)
362 != opt_for_fn (to->decl, flag_trapping_math)
363 || opt_for_fn (callee->decl, flag_unsafe_math_optimizations)
364 != opt_for_fn (to->decl, flag_unsafe_math_optimizations)
365 || opt_for_fn (callee->decl, flag_finite_math_only)
366 != opt_for_fn (to->decl, flag_finite_math_only)
367 || opt_for_fn (callee->decl, flag_signaling_nans)
368 != opt_for_fn (to->decl, flag_signaling_nans)
369 || opt_for_fn (callee->decl, flag_cx_limited_range)
370 != opt_for_fn (to->decl, flag_cx_limited_range)
371 || opt_for_fn (callee->decl, flag_signed_zeros)
372 != opt_for_fn (to->decl, flag_signed_zeros)
373 || opt_for_fn (callee->decl, flag_associative_math)
374 != opt_for_fn (to->decl, flag_associative_math)
375 || opt_for_fn (callee->decl, flag_reciprocal_math)
376 != opt_for_fn (to->decl, flag_reciprocal_math)
0d2f700f
JM
377 || opt_for_fn (callee->decl, flag_fp_int_builtin_inexact)
378 != opt_for_fn (to->decl, flag_fp_int_builtin_inexact)
818b88a7
JH
379 || opt_for_fn (callee->decl, flag_errno_math)
380 != opt_for_fn (to->decl, flag_errno_math))
381 {
382 struct gcc_options opts = global_options;
383
384 cl_optimization_restore (&opts, opts_for_fn (to->decl));
385 opts.x_flag_rounding_math
386 = opt_for_fn (callee->decl, flag_rounding_math);
387 opts.x_flag_trapping_math
388 = opt_for_fn (callee->decl, flag_trapping_math);
389 opts.x_flag_unsafe_math_optimizations
390 = opt_for_fn (callee->decl, flag_unsafe_math_optimizations);
391 opts.x_flag_finite_math_only
392 = opt_for_fn (callee->decl, flag_finite_math_only);
393 opts.x_flag_signaling_nans
394 = opt_for_fn (callee->decl, flag_signaling_nans);
395 opts.x_flag_cx_limited_range
396 = opt_for_fn (callee->decl, flag_cx_limited_range);
397 opts.x_flag_signed_zeros
398 = opt_for_fn (callee->decl, flag_signed_zeros);
399 opts.x_flag_associative_math
400 = opt_for_fn (callee->decl, flag_associative_math);
401 opts.x_flag_reciprocal_math
402 = opt_for_fn (callee->decl, flag_reciprocal_math);
0d2f700f
JM
403 opts.x_flag_fp_int_builtin_inexact
404 = opt_for_fn (callee->decl, flag_fp_int_builtin_inexact);
818b88a7
JH
405 opts.x_flag_errno_math
406 = opt_for_fn (callee->decl, flag_errno_math);
407 if (dump_file)
464d0118
ML
408 fprintf (dump_file, "Copying FP flags from %s to %s\n",
409 callee->dump_name (), to->dump_name ());
818b88a7
JH
410 DECL_FUNCTION_SPECIFIC_OPTIMIZATION (to->decl)
411 = build_optimization_node (&opts);
77719b06 412 reload_optimization_node = true;
818b88a7
JH
413 }
414 }
5058c037 415
77719b06
ML
416 /* Reload global optimization flags. */
417 if (reload_optimization_node && DECL_STRUCT_FUNCTION (to->decl) == cfun)
418 set_cfun (cfun, true);
419
a5b1779f
JH
420 /* If aliases are involved, redirect edge to the actual destination and
421 possibly remove the aliases. */
422 if (e->callee != callee)
39e2db00
JH
423 {
424 struct cgraph_node *alias = e->callee, *next_alias;
3dafb85c 425 e->redirect_callee (callee);
39e2db00
JH
426 while (alias && alias != callee)
427 {
428 if (!alias->callers
8ccc8042
JH
429 && can_remove_node_now_p (alias,
430 !e->next_caller && !e->prev_caller ? e : NULL))
39e2db00 431 {
d52f5295
ML
432 next_alias = alias->get_alias_target ();
433 alias->remove ();
1bbb87c4
JH
434 if (callee_removed)
435 *callee_removed = true;
39e2db00
JH
436 alias = next_alias;
437 }
438 else
439 break;
440 }
441 }
a5b1779f 442
1bad9c18 443 clone_inlined_nodes (e, true, update_original, overall_size);
fee8b6da 444
fee8b6da 445 gcc_assert (curr->callee->global.inlined_to == to);
898b8927 446
99353fcf 447 old_size = ipa_fn_summaries->get_create (to)->size;
0bceb671 448 ipa_merge_fn_summary_after_inlining (e);
4fd94d1e
MJ
449 if (e->in_polymorphic_cdtor)
450 mark_all_inlined_calls_cdtor (e->callee);
bb1e543c 451 if (opt_for_fn (e->caller->decl, optimize))
0f378cb5 452 new_edges_found = ipa_propagate_indirect_call_infos (curr, new_edges);
db66bf68 453 check_speculations (e->callee);
c170d40f 454 if (update_overall_summary)
0bceb671 455 ipa_update_overall_fn_summary (to);
bddead15
RB
456 else
457 /* Update self size by the estimate so overall function growth limits
458 work for further inlining into this function. Before inlining
459 the function we inlined to again we expect the caller to update
460 the overall summary. */
99353fcf
ML
461 ipa_fn_summaries->get_create (to)->size += estimated_growth;
462 new_size = ipa_fn_summaries->get_create (to)->size;
d250540a 463
1f26ac87
JM
464 if (callee->calls_comdat_local)
465 to->calls_comdat_local = true;
d52f5295 466 else if (to->calls_comdat_local && callee->comdat_local_p ())
1f26ac87
JM
467 {
468 struct cgraph_edge *se = to->callees;
469 for (; se; se = se->next_callee)
d52f5295 470 if (se->inline_failed && se->callee->comdat_local_p ())
1f26ac87
JM
471 break;
472 if (se == NULL)
473 to->calls_comdat_local = false;
474 }
475
f107227b
JH
476 /* FIXME: This assert suffers from roundoff errors, disable it for GCC 5
477 and revisit it after conversion to sreals in GCC 6.
478 See PR 65654. */
479#if 0
0f378cb5 480 /* Verify that estimated growth match real growth. Allow off-by-one
0bceb671 481 error due to ipa_fn_summary::size_scale roudoff errors. */
48b1474e 482 gcc_assert (!update_overall_summary || !overall_size || new_edges_found
0f378cb5 483 || abs (estimated_growth - (new_size - old_size)) <= 1
09ce3660 484 || speculation_removed
0f378cb5
JH
485 /* FIXME: a hack. Edges with false predicate are accounted
486 wrong, we should remove them from callgraph. */
487 || predicated);
488#endif
d250540a 489
8256d5ca
JH
490 /* Account the change of overall unit size; external functions will be
491 removed and are thus not accounted. */
bb1e543c 492 if (overall_size && inline_account_function_p (to))
fee8b6da
JH
493 *overall_size += new_size - old_size;
494 ncalls_inlined++;
495
0bceb671 496 /* This must happen after ipa_merge_fn_summary_after_inlining that rely on jump
25837a2f 497 functions of callee to not be updated. */
0f378cb5 498 return new_edges_found;
fee8b6da
JH
499}
500
501
502/* Copy function body of NODE and redirect all inline clones to it.
503 This is done before inline plan is applied to NODE when there are
504 still some inline clones if it.
505
073a8998 506 This is necessary because inline decisions are not really transitive
fee8b6da
JH
507 and the other inline clones may have different bodies. */
508
509static struct cgraph_node *
510save_inline_function_body (struct cgraph_node *node)
511{
512 struct cgraph_node *first_clone, *n;
513
514 if (dump_file)
515 fprintf (dump_file, "\nSaving body of %s for later reuse\n",
fec39fa6 516 node->name ());
fee8b6da 517
d52f5295 518 gcc_assert (node == cgraph_node::get (node->decl));
fee8b6da
JH
519
520 /* first_clone will be turned into real function. */
521 first_clone = node->clones;
ec6a1e35
JH
522
523 /* Arrange first clone to not be thunk as those do not have bodies. */
524 if (first_clone->thunk.thunk_p)
525 {
526 while (first_clone->thunk.thunk_p)
527 first_clone = first_clone->next_sibling_clone;
528 first_clone->prev_sibling_clone->next_sibling_clone
529 = first_clone->next_sibling_clone;
530 if (first_clone->next_sibling_clone)
531 first_clone->next_sibling_clone->prev_sibling_clone
532 = first_clone->prev_sibling_clone;
533 first_clone->next_sibling_clone = node->clones;
534 first_clone->prev_sibling_clone = NULL;
535 node->clones->prev_sibling_clone = first_clone;
536 node->clones = first_clone;
537 }
67348ccc 538 first_clone->decl = copy_node (node->decl);
aede2c10 539 first_clone->decl->decl_with_vis.symtab_node = first_clone;
d52f5295 540 gcc_assert (first_clone == cgraph_node::get (first_clone->decl));
fee8b6da
JH
541
542 /* Now reshape the clone tree, so all other clones descends from
543 first_clone. */
544 if (first_clone->next_sibling_clone)
545 {
ec6a1e35
JH
546 for (n = first_clone->next_sibling_clone; n->next_sibling_clone;
547 n = n->next_sibling_clone)
fee8b6da
JH
548 n->clone_of = first_clone;
549 n->clone_of = first_clone;
550 n->next_sibling_clone = first_clone->clones;
551 if (first_clone->clones)
552 first_clone->clones->prev_sibling_clone = n;
553 first_clone->clones = first_clone->next_sibling_clone;
554 first_clone->next_sibling_clone->prev_sibling_clone = NULL;
555 first_clone->next_sibling_clone = NULL;
556 gcc_assert (!first_clone->prev_sibling_clone);
557 }
558 first_clone->clone_of = NULL;
559
560 /* Now node in question has no clones. */
561 node->clones = NULL;
562
1a3118e9
JH
563 /* Inline clones share decl with the function they are cloned
564 from. Walk the whole clone tree and redirect them all to the
565 new decl. */
fee8b6da
JH
566 if (first_clone->clones)
567 for (n = first_clone->clones; n != first_clone;)
568 {
67348ccc
DM
569 gcc_assert (n->decl == node->decl);
570 n->decl = first_clone->decl;
fee8b6da
JH
571 if (n->clones)
572 n = n->clones;
573 else if (n->next_sibling_clone)
574 n = n->next_sibling_clone;
575 else
576 {
577 while (n != first_clone && !n->next_sibling_clone)
578 n = n->clone_of;
579 if (n != first_clone)
580 n = n->next_sibling_clone;
581 }
582 }
583
584 /* Copy the OLD_VERSION_NODE function tree to the new version. */
67348ccc 585 tree_function_versioning (node->decl, first_clone->decl,
9771b263
DN
586 NULL, true, NULL, false,
587 NULL, NULL);
fee8b6da 588
1a3118e9
JH
589 /* The function will be short lived and removed after we inline all the clones,
590 but make it internal so we won't confuse ourself. */
67348ccc 591 DECL_EXTERNAL (first_clone->decl) = 0;
67348ccc
DM
592 TREE_PUBLIC (first_clone->decl) = 0;
593 DECL_COMDAT (first_clone->decl) = 0;
9771b263 594 first_clone->ipa_transforms_to_apply.release ();
fee8b6da 595
b4e93f45
JH
596 /* When doing recursive inlining, the clone may become unnecessary.
597 This is possible i.e. in the case when the recursive function is proved to be
598 non-throwing and the recursion happens only in the EH landing pad.
599 We can not remove the clone until we are done with saving the body.
600 Remove it now. */
601 if (!first_clone->callers)
602 {
d52f5295 603 first_clone->remove_symbol_and_inline_clones ();
b4e93f45
JH
604 first_clone = NULL;
605 }
b2b29377 606 else if (flag_checking)
d52f5295 607 first_clone->verify ();
b2b29377 608
fee8b6da
JH
609 return first_clone;
610}
611
9c8305f8
JH
612/* Return true when function body of DECL still needs to be kept around
613 for later re-use. */
65d630d4 614static bool
9c8305f8
JH
615preserve_function_body_p (struct cgraph_node *node)
616{
3dafb85c 617 gcc_assert (symtab->global_info_ready);
67348ccc 618 gcc_assert (!node->alias && !node->thunk.thunk_p);
9c8305f8 619
ec6a1e35
JH
620 /* Look if there is any non-thunk clone around. */
621 for (node = node->clones; node; node = node->next_sibling_clone)
622 if (!node->thunk.thunk_p)
623 return true;
9c8305f8
JH
624 return false;
625}
fee8b6da
JH
626
627/* Apply inline plan to function. */
628
629unsigned int
630inline_transform (struct cgraph_node *node)
631{
632 unsigned int todo = 0;
e8aec975 633 struct cgraph_edge *e, *next;
2bf86c84 634 bool has_inline = false;
c9fc06dc 635
fee8b6da
JH
636 /* FIXME: Currently the pass manager is adding inline transform more than
637 once to some clones. This needs revisiting after WPA cleanups. */
638 if (cfun->after_inlining)
639 return 0;
640
641 /* We might need the body of this function so that we can expand
642 it inline somewhere else. */
9c8305f8 643 if (preserve_function_body_p (node))
fee8b6da
JH
644 save_inline_function_body (node);
645
e8aec975
JH
646 for (e = node->callees; e; e = next)
647 {
2bf86c84
JH
648 if (!e->inline_failed)
649 has_inline = true;
e8aec975 650 next = e->next_callee;
3dafb85c 651 e->redirect_call_stmt_to_callee ();
e8aec975 652 }
d122681a 653 node->remove_all_references ();
c9fc06dc
CB
654
655 timevar_push (TV_INTEGRATION);
bb1e543c 656 if (node->callees && (opt_for_fn (node->decl, optimize) || has_inline))
09fcc0c0
JH
657 {
658 profile_count num = node->count;
659 profile_count den = ENTRY_BLOCK_PTR_FOR_FN (cfun)->count;
517048ce 660 bool scale = num.initialized_p () && !(num == den);
09fcc0c0
JH
661 if (scale)
662 {
517048ce 663 profile_count::adjust_for_ipa_scaling (&num, &den);
09fcc0c0
JH
664 if (dump_file)
665 {
666 fprintf (dump_file, "Applying count scale ");
667 num.dump (dump_file);
668 fprintf (dump_file, "/");
669 den.dump (dump_file);
670 fprintf (dump_file, "\n");
671 }
672
673 basic_block bb;
517048ce 674 cfun->cfg->count_max = profile_count::uninitialized ();
09fcc0c0 675 FOR_ALL_BB_FN (bb, cfun)
517048ce 676 {
8e7d1486 677 bb->count = bb->count.apply_scale (num, den);
517048ce
JH
678 cfun->cfg->count_max = cfun->cfg->count_max.max (bb->count);
679 }
09fcc0c0
JH
680 ENTRY_BLOCK_PTR_FOR_FN (cfun)->count = node->count;
681 }
682 todo = optimize_inline_calls (current_function_decl);
683 }
c9fc06dc
CB
684 timevar_pop (TV_INTEGRATION);
685
f8698b37
RG
686 cfun->always_inline_functions_inlined = true;
687 cfun->after_inlining = true;
688 todo |= execute_fixup_cfg ();
689
55f01229
RG
690 if (!(todo & TODO_update_ssa_any))
691 /* Redirecting edges might lead to a need for vops to be recomputed. */
692 todo |= TODO_update_ssa_only_virtuals;
693
f8698b37 694 return todo;
fee8b6da 695}