]> git.ipfire.org Git - thirdparty/gcc.git/blob - gcc/cgraphclones.c
Come up with fndecl_built_in_p.
[thirdparty/gcc.git] / gcc / cgraphclones.c
1 /* Callgraph clones
2 Copyright (C) 2003-2018 Free Software Foundation, Inc.
3 Contributed by Jan Hubicka
4
5 This file is part of GCC.
6
7 GCC is free software; you can redistribute it and/or modify it under
8 the terms of the GNU General Public License as published by the Free
9 Software Foundation; either version 3, or (at your option) any later
10 version.
11
12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
15 for more details.
16
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
20
21 /* This module provide facilities for clonning functions. I.e. creating
22 new functions based on existing functions with simple modifications,
23 such as replacement of parameters.
24
25 To allow whole program optimization without actual presence of function
26 bodies, an additional infrastructure is provided for so-called virtual
27 clones
28
29 A virtual clone in the callgraph is a function that has no
30 associated body, just a description of how to create its body based
31 on a different function (which itself may be a virtual clone).
32
33 The description of function modifications includes adjustments to
34 the function's signature (which allows, for example, removing or
35 adding function arguments), substitutions to perform on the
36 function body, and, for inlined functions, a pointer to the
37 function that it will be inlined into.
38
39 It is also possible to redirect any edge of the callgraph from a
40 function to its virtual clone. This implies updating of the call
41 site to adjust for the new function signature.
42
43 Most of the transformations performed by inter-procedural
44 optimizations can be represented via virtual clones. For
45 instance, a constant propagation pass can produce a virtual clone
46 of the function which replaces one of its arguments by a
47 constant. The inliner can represent its decisions by producing a
48 clone of a function whose body will be later integrated into
49 a given function.
50
51 Using virtual clones, the program can be easily updated
52 during the Execute stage, solving most of pass interactions
53 problems that would otherwise occur during Transform.
54
55 Virtual clones are later materialized in the LTRANS stage and
56 turned into real functions. Passes executed after the virtual
57 clone were introduced also perform their Transform stage
58 on new functions, so for a pass there is no significant
59 difference between operating on a real function or a virtual
60 clone introduced before its Execute stage.
61
62 Optimization passes then work on virtual clones introduced before
63 their Execute stage as if they were real functions. The
64 only difference is that clones are not visible during the
65 Generate Summary stage. */
66
67 #include "config.h"
68 #include "system.h"
69 #include "coretypes.h"
70 #include "backend.h"
71 #include "target.h"
72 #include "rtl.h"
73 #include "tree.h"
74 #include "gimple.h"
75 #include "stringpool.h"
76 #include "cgraph.h"
77 #include "lto-streamer.h"
78 #include "tree-eh.h"
79 #include "tree-cfg.h"
80 #include "tree-inline.h"
81 #include "dumpfile.h"
82 #include "gimple-pretty-print.h"
83
84 /* Create clone of edge in the node N represented by CALL_EXPR
85 the callgraph. */
86
87 cgraph_edge *
88 cgraph_edge::clone (cgraph_node *n, gcall *call_stmt, unsigned stmt_uid,
89 profile_count num, profile_count den,
90 bool update_original)
91 {
92 cgraph_edge *new_edge;
93 profile_count::adjust_for_ipa_scaling (&num, &den);
94 profile_count prof_count = count.apply_scale (num, den);
95
96 if (indirect_unknown_callee)
97 {
98 tree decl;
99
100 if (call_stmt && (decl = gimple_call_fndecl (call_stmt))
101 /* When the call is speculative, we need to resolve it
102 via cgraph_resolve_speculation and not here. */
103 && !speculative)
104 {
105 cgraph_node *callee = cgraph_node::get (decl);
106 gcc_checking_assert (callee);
107 new_edge = n->create_edge (callee, call_stmt, prof_count);
108 }
109 else
110 {
111 new_edge = n->create_indirect_edge (call_stmt,
112 indirect_info->ecf_flags,
113 prof_count, false);
114 *new_edge->indirect_info = *indirect_info;
115 }
116 }
117 else
118 {
119 new_edge = n->create_edge (callee, call_stmt, prof_count);
120 if (indirect_info)
121 {
122 new_edge->indirect_info
123 = ggc_cleared_alloc<cgraph_indirect_call_info> ();
124 *new_edge->indirect_info = *indirect_info;
125 }
126 }
127
128 new_edge->inline_failed = inline_failed;
129 new_edge->indirect_inlining_edge = indirect_inlining_edge;
130 new_edge->lto_stmt_uid = stmt_uid;
131 /* Clone flags that depend on call_stmt availability manually. */
132 new_edge->can_throw_external = can_throw_external;
133 new_edge->call_stmt_cannot_inline_p = call_stmt_cannot_inline_p;
134 new_edge->speculative = speculative;
135 new_edge->in_polymorphic_cdtor = in_polymorphic_cdtor;
136
137 /* Update IPA profile. Local profiles need no updating in original. */
138 if (update_original)
139 count = count.combine_with_ipa_count (count.ipa ()
140 - new_edge->count.ipa ());
141 symtab->call_edge_duplication_hooks (this, new_edge);
142 return new_edge;
143 }
144
145 /* Build variant of function type ORIG_TYPE skipping ARGS_TO_SKIP and the
146 return value if SKIP_RETURN is true. */
147
148 tree
149 cgraph_build_function_type_skip_args (tree orig_type, bitmap args_to_skip,
150 bool skip_return)
151 {
152 tree new_type = NULL;
153 tree args, new_args = NULL;
154 tree new_reversed;
155 int i = 0;
156
157 for (args = TYPE_ARG_TYPES (orig_type); args && args != void_list_node;
158 args = TREE_CHAIN (args), i++)
159 if (!args_to_skip || !bitmap_bit_p (args_to_skip, i))
160 new_args = tree_cons (NULL_TREE, TREE_VALUE (args), new_args);
161
162 new_reversed = nreverse (new_args);
163 if (args)
164 {
165 if (new_reversed)
166 TREE_CHAIN (new_args) = void_list_node;
167 else
168 new_reversed = void_list_node;
169 }
170
171 /* Use copy_node to preserve as much as possible from original type
172 (debug info, attribute lists etc.)
173 Exception is METHOD_TYPEs must have THIS argument.
174 When we are asked to remove it, we need to build new FUNCTION_TYPE
175 instead. */
176 if (TREE_CODE (orig_type) != METHOD_TYPE
177 || !args_to_skip
178 || !bitmap_bit_p (args_to_skip, 0))
179 {
180 new_type = build_distinct_type_copy (orig_type);
181 TYPE_ARG_TYPES (new_type) = new_reversed;
182 }
183 else
184 {
185 new_type
186 = build_distinct_type_copy (build_function_type (TREE_TYPE (orig_type),
187 new_reversed));
188 TYPE_CONTEXT (new_type) = TYPE_CONTEXT (orig_type);
189 }
190
191 if (skip_return)
192 TREE_TYPE (new_type) = void_type_node;
193
194 return new_type;
195 }
196
197 /* Build variant of function decl ORIG_DECL skipping ARGS_TO_SKIP and the
198 return value if SKIP_RETURN is true.
199
200 Arguments from DECL_ARGUMENTS list can't be removed now, since they are
201 linked by TREE_CHAIN directly. The caller is responsible for eliminating
202 them when they are being duplicated (i.e. copy_arguments_for_versioning). */
203
204 static tree
205 build_function_decl_skip_args (tree orig_decl, bitmap args_to_skip,
206 bool skip_return)
207 {
208 tree new_decl = copy_node (orig_decl);
209 tree new_type;
210
211 new_type = TREE_TYPE (orig_decl);
212 if (prototype_p (new_type)
213 || (skip_return && !VOID_TYPE_P (TREE_TYPE (new_type))))
214 new_type
215 = cgraph_build_function_type_skip_args (new_type, args_to_skip,
216 skip_return);
217 TREE_TYPE (new_decl) = new_type;
218
219 /* For declarations setting DECL_VINDEX (i.e. methods)
220 we expect first argument to be THIS pointer. */
221 if (args_to_skip && bitmap_bit_p (args_to_skip, 0))
222 DECL_VINDEX (new_decl) = NULL_TREE;
223
224 /* When signature changes, we need to clear builtin info. */
225 if (fndecl_built_in_p (new_decl)
226 && args_to_skip
227 && !bitmap_empty_p (args_to_skip))
228 {
229 DECL_BUILT_IN_CLASS (new_decl) = NOT_BUILT_IN;
230 DECL_FUNCTION_CODE (new_decl) = (enum built_in_function) 0;
231 }
232 /* The FE might have information and assumptions about the other
233 arguments. */
234 DECL_LANG_SPECIFIC (new_decl) = NULL;
235 return new_decl;
236 }
237
238 /* Set flags of NEW_NODE and its decl. NEW_NODE is a newly created private
239 clone or its thunk. */
240
241 static void
242 set_new_clone_decl_and_node_flags (cgraph_node *new_node)
243 {
244 DECL_EXTERNAL (new_node->decl) = 0;
245 TREE_PUBLIC (new_node->decl) = 0;
246 DECL_COMDAT (new_node->decl) = 0;
247 DECL_WEAK (new_node->decl) = 0;
248 DECL_VIRTUAL_P (new_node->decl) = 0;
249 DECL_STATIC_CONSTRUCTOR (new_node->decl) = 0;
250 DECL_STATIC_DESTRUCTOR (new_node->decl) = 0;
251
252 new_node->externally_visible = 0;
253 new_node->local.local = 1;
254 new_node->lowered = true;
255 }
256
257 /* Duplicate thunk THUNK if necessary but make it to refer to NODE.
258 ARGS_TO_SKIP, if non-NULL, determines which parameters should be omitted.
259 Function can return NODE if no thunk is necessary, which can happen when
260 thunk is this_adjusting but we are removing this parameter. */
261
262 static cgraph_node *
263 duplicate_thunk_for_node (cgraph_node *thunk, cgraph_node *node)
264 {
265 cgraph_node *new_thunk, *thunk_of;
266 thunk_of = thunk->callees->callee->ultimate_alias_target ();
267
268 if (thunk_of->thunk.thunk_p)
269 node = duplicate_thunk_for_node (thunk_of, node);
270
271 if (!DECL_ARGUMENTS (thunk->decl))
272 thunk->get_untransformed_body ();
273
274 cgraph_edge *cs;
275 for (cs = node->callers; cs; cs = cs->next_caller)
276 if (cs->caller->thunk.thunk_p
277 && cs->caller->thunk.this_adjusting == thunk->thunk.this_adjusting
278 && cs->caller->thunk.fixed_offset == thunk->thunk.fixed_offset
279 && cs->caller->thunk.virtual_offset_p == thunk->thunk.virtual_offset_p
280 && cs->caller->thunk.virtual_value == thunk->thunk.virtual_value)
281 return cs->caller;
282
283 tree new_decl;
284 if (!node->clone.args_to_skip)
285 new_decl = copy_node (thunk->decl);
286 else
287 {
288 /* We do not need to duplicate this_adjusting thunks if we have removed
289 this. */
290 if (thunk->thunk.this_adjusting
291 && bitmap_bit_p (node->clone.args_to_skip, 0))
292 return node;
293
294 new_decl = build_function_decl_skip_args (thunk->decl,
295 node->clone.args_to_skip,
296 false);
297 }
298
299 tree *link = &DECL_ARGUMENTS (new_decl);
300 int i = 0;
301 for (tree pd = DECL_ARGUMENTS (thunk->decl); pd; pd = DECL_CHAIN (pd), i++)
302 {
303 if (!node->clone.args_to_skip
304 || !bitmap_bit_p (node->clone.args_to_skip, i))
305 {
306 tree nd = copy_node (pd);
307 DECL_CONTEXT (nd) = new_decl;
308 *link = nd;
309 link = &DECL_CHAIN (nd);
310 }
311 }
312 *link = NULL_TREE;
313
314 gcc_checking_assert (!DECL_STRUCT_FUNCTION (new_decl));
315 gcc_checking_assert (!DECL_INITIAL (new_decl));
316 gcc_checking_assert (!DECL_RESULT (new_decl));
317 gcc_checking_assert (!DECL_RTL_SET_P (new_decl));
318
319 DECL_NAME (new_decl) = clone_function_name (thunk->decl, "artificial_thunk");
320 SET_DECL_ASSEMBLER_NAME (new_decl, DECL_NAME (new_decl));
321
322 new_thunk = cgraph_node::create (new_decl);
323 set_new_clone_decl_and_node_flags (new_thunk);
324 new_thunk->definition = true;
325 new_thunk->local.can_change_signature = node->local.can_change_signature;
326 new_thunk->thunk = thunk->thunk;
327 new_thunk->unique_name = in_lto_p;
328 new_thunk->former_clone_of = thunk->decl;
329 new_thunk->clone.args_to_skip = node->clone.args_to_skip;
330 new_thunk->clone.combined_args_to_skip = node->clone.combined_args_to_skip;
331
332 cgraph_edge *e = new_thunk->create_edge (node, NULL, new_thunk->count);
333 symtab->call_edge_duplication_hooks (thunk->callees, e);
334 symtab->call_cgraph_duplication_hooks (thunk, new_thunk);
335 return new_thunk;
336 }
337
338 /* If E does not lead to a thunk, simply redirect it to N. Otherwise create
339 one or more equivalent thunks for N and redirect E to the first in the
340 chain. Note that it is then necessary to call
341 n->expand_all_artificial_thunks once all callers are redirected. */
342
343 void
344 cgraph_edge::redirect_callee_duplicating_thunks (cgraph_node *n)
345 {
346 cgraph_node *orig_to = callee->ultimate_alias_target ();
347 if (orig_to->thunk.thunk_p)
348 n = duplicate_thunk_for_node (orig_to, n);
349
350 redirect_callee (n);
351 }
352
353 /* Call expand_thunk on all callers that are thunks and if analyze those nodes
354 that were expanded. */
355
356 void
357 cgraph_node::expand_all_artificial_thunks ()
358 {
359 cgraph_edge *e;
360 for (e = callers; e;)
361 if (e->caller->thunk.thunk_p)
362 {
363 cgraph_node *thunk = e->caller;
364
365 e = e->next_caller;
366 if (thunk->expand_thunk (false, false))
367 {
368 thunk->thunk.thunk_p = false;
369 thunk->analyze ();
370 }
371 thunk->expand_all_artificial_thunks ();
372 }
373 else
374 e = e->next_caller;
375 }
376
377 void
378 dump_callgraph_transformation (const cgraph_node *original,
379 const cgraph_node *clone,
380 const char *suffix)
381 {
382 if (symtab->ipa_clones_dump_file)
383 {
384 fprintf (symtab->ipa_clones_dump_file,
385 "Callgraph clone;%s;%d;%s;%d;%d;%s;%d;%s;%d;%d;%s\n",
386 original->asm_name (), original->order,
387 DECL_SOURCE_FILE (original->decl),
388 DECL_SOURCE_LINE (original->decl),
389 DECL_SOURCE_COLUMN (original->decl), clone->asm_name (),
390 clone->order, DECL_SOURCE_FILE (clone->decl),
391 DECL_SOURCE_LINE (clone->decl), DECL_SOURCE_COLUMN (clone->decl),
392 suffix);
393
394 symtab->cloned_nodes.add (original);
395 symtab->cloned_nodes.add (clone);
396 }
397 }
398
399 /* Create node representing clone of N executed COUNT times. Decrease
400 the execution counts from original node too.
401 The new clone will have decl set to DECL that may or may not be the same
402 as decl of N.
403
404 When UPDATE_ORIGINAL is true, the counts are subtracted from the original
405 function's profile to reflect the fact that part of execution is handled
406 by node.
407 When CALL_DUPLICATOIN_HOOK is true, the ipa passes are acknowledged about
408 the new clone. Otherwise the caller is responsible for doing so later.
409
410 If the new node is being inlined into another one, NEW_INLINED_TO should be
411 the outline function the new one is (even indirectly) inlined to. All hooks
412 will see this in node's global.inlined_to, when invoked. Can be NULL if the
413 node is not inlined. */
414
415 cgraph_node *
416 cgraph_node::create_clone (tree new_decl, profile_count prof_count,
417 bool update_original,
418 vec<cgraph_edge *> redirect_callers,
419 bool call_duplication_hook,
420 cgraph_node *new_inlined_to,
421 bitmap args_to_skip, const char *suffix)
422 {
423 cgraph_node *new_node = symtab->create_empty ();
424 cgraph_edge *e;
425 unsigned i;
426 profile_count old_count = count;
427
428 if (new_inlined_to)
429 dump_callgraph_transformation (this, new_inlined_to, "inlining to");
430
431 /* When inlining we scale precisely to prof_count, when cloning we can
432 preserve local profile. */
433 if (!new_inlined_to)
434 prof_count = count.combine_with_ipa_count (prof_count);
435 new_node->count = prof_count;
436
437 /* Update IPA profile. Local profiles need no updating in original. */
438 if (update_original)
439 count = count.combine_with_ipa_count (count.ipa () - prof_count.ipa ());
440 new_node->decl = new_decl;
441 new_node->register_symbol ();
442 new_node->origin = origin;
443 new_node->lto_file_data = lto_file_data;
444 if (new_node->origin)
445 {
446 new_node->next_nested = new_node->origin->nested;
447 new_node->origin->nested = new_node;
448 }
449 new_node->analyzed = analyzed;
450 new_node->definition = definition;
451 new_node->local = local;
452 new_node->externally_visible = false;
453 new_node->no_reorder = no_reorder;
454 new_node->local.local = true;
455 new_node->global = global;
456 new_node->global.inlined_to = new_inlined_to;
457 new_node->rtl = rtl;
458 new_node->frequency = frequency;
459 new_node->tp_first_run = tp_first_run;
460 new_node->tm_clone = tm_clone;
461 new_node->icf_merged = icf_merged;
462 new_node->merged_comdat = merged_comdat;
463 new_node->thunk = thunk;
464
465 new_node->clone.tree_map = NULL;
466 new_node->clone.args_to_skip = args_to_skip;
467 new_node->split_part = split_part;
468 if (!args_to_skip)
469 new_node->clone.combined_args_to_skip = clone.combined_args_to_skip;
470 else if (clone.combined_args_to_skip)
471 {
472 new_node->clone.combined_args_to_skip = BITMAP_GGC_ALLOC ();
473 bitmap_ior (new_node->clone.combined_args_to_skip,
474 clone.combined_args_to_skip, args_to_skip);
475 }
476 else
477 new_node->clone.combined_args_to_skip = args_to_skip;
478
479 FOR_EACH_VEC_ELT (redirect_callers, i, e)
480 {
481 /* Redirect calls to the old version node to point to its new
482 version. The only exception is when the edge was proved to
483 be unreachable during the clonning procedure. */
484 if (!e->callee
485 || !fndecl_built_in_p (e->callee->decl, BUILT_IN_UNREACHABLE))
486 e->redirect_callee_duplicating_thunks (new_node);
487 }
488 new_node->expand_all_artificial_thunks ();
489
490 for (e = callees;e; e=e->next_callee)
491 e->clone (new_node, e->call_stmt, e->lto_stmt_uid, new_node->count, old_count,
492 update_original);
493
494 for (e = indirect_calls; e; e = e->next_callee)
495 e->clone (new_node, e->call_stmt, e->lto_stmt_uid,
496 new_node->count, old_count, update_original);
497 new_node->clone_references (this);
498
499 new_node->next_sibling_clone = clones;
500 if (clones)
501 clones->prev_sibling_clone = new_node;
502 clones = new_node;
503 new_node->clone_of = this;
504
505 if (call_duplication_hook)
506 symtab->call_cgraph_duplication_hooks (this, new_node);
507
508 if (!new_inlined_to)
509 dump_callgraph_transformation (this, new_node, suffix);
510
511 return new_node;
512 }
513
514 static GTY(()) unsigned int clone_fn_id_num;
515
516 /* Return a new assembler name for a clone with SUFFIX of a decl named
517 NAME. */
518
519 tree
520 clone_function_name_1 (const char *name, const char *suffix)
521 {
522 size_t len = strlen (name);
523 char *tmp_name, *prefix;
524
525 prefix = XALLOCAVEC (char, len + strlen (suffix) + 2);
526 memcpy (prefix, name, len);
527 strcpy (prefix + len + 1, suffix);
528 prefix[len] = symbol_table::symbol_suffix_separator ();
529 ASM_FORMAT_PRIVATE_NAME (tmp_name, prefix, clone_fn_id_num++);
530 return get_identifier (tmp_name);
531 }
532
533 /* Return a new assembler name for a clone of DECL with SUFFIX. */
534
535 tree
536 clone_function_name (tree decl, const char *suffix)
537 {
538 tree name = DECL_ASSEMBLER_NAME (decl);
539 return clone_function_name_1 (IDENTIFIER_POINTER (name), suffix);
540 }
541
542
543 /* Create callgraph node clone with new declaration. The actual body will
544 be copied later at compilation stage.
545
546 TODO: after merging in ipa-sra use function call notes instead of args_to_skip
547 bitmap interface.
548 */
549 cgraph_node *
550 cgraph_node::create_virtual_clone (vec<cgraph_edge *> redirect_callers,
551 vec<ipa_replace_map *, va_gc> *tree_map,
552 bitmap args_to_skip, const char * suffix)
553 {
554 tree old_decl = decl;
555 cgraph_node *new_node = NULL;
556 tree new_decl;
557 size_t len, i;
558 ipa_replace_map *map;
559 char *name;
560
561 gcc_checking_assert (local.versionable);
562 gcc_assert (local.can_change_signature || !args_to_skip);
563
564 /* Make a new FUNCTION_DECL tree node */
565 if (!args_to_skip)
566 new_decl = copy_node (old_decl);
567 else
568 new_decl = build_function_decl_skip_args (old_decl, args_to_skip, false);
569
570 /* These pointers represent function body and will be populated only when clone
571 is materialized. */
572 gcc_assert (new_decl != old_decl);
573 DECL_STRUCT_FUNCTION (new_decl) = NULL;
574 DECL_ARGUMENTS (new_decl) = NULL;
575 DECL_INITIAL (new_decl) = NULL;
576 DECL_RESULT (new_decl) = NULL;
577 /* We can not do DECL_RESULT (new_decl) = NULL; here because of LTO partitioning
578 sometimes storing only clone decl instead of original. */
579
580 /* Generate a new name for the new version. */
581 len = IDENTIFIER_LENGTH (DECL_NAME (old_decl));
582 name = XALLOCAVEC (char, len + strlen (suffix) + 2);
583 memcpy (name, IDENTIFIER_POINTER (DECL_NAME (old_decl)), len);
584 strcpy (name + len + 1, suffix);
585 name[len] = '.';
586 DECL_NAME (new_decl) = get_identifier (name);
587 SET_DECL_ASSEMBLER_NAME (new_decl, clone_function_name (old_decl, suffix));
588 SET_DECL_RTL (new_decl, NULL);
589
590 new_node = create_clone (new_decl, count, false,
591 redirect_callers, false, NULL, args_to_skip, suffix);
592
593 /* Update the properties.
594 Make clone visible only within this translation unit. Make sure
595 that is not weak also.
596 ??? We cannot use COMDAT linkage because there is no
597 ABI support for this. */
598 set_new_clone_decl_and_node_flags (new_node);
599 new_node->clone.tree_map = tree_map;
600 if (!implicit_section)
601 new_node->set_section (get_section ());
602
603 /* Clones of global symbols or symbols with unique names are unique. */
604 if ((TREE_PUBLIC (old_decl)
605 && !DECL_EXTERNAL (old_decl)
606 && !DECL_WEAK (old_decl)
607 && !DECL_COMDAT (old_decl))
608 || in_lto_p)
609 new_node->unique_name = true;
610 FOR_EACH_VEC_SAFE_ELT (tree_map, i, map)
611 new_node->maybe_create_reference (map->new_tree, NULL);
612
613 if (ipa_transforms_to_apply.exists ())
614 new_node->ipa_transforms_to_apply
615 = ipa_transforms_to_apply.copy ();
616
617 symtab->call_cgraph_duplication_hooks (this, new_node);
618
619 return new_node;
620 }
621
622 /* callgraph node being removed from symbol table; see if its entry can be
623 replaced by other inline clone. */
624 cgraph_node *
625 cgraph_node::find_replacement (void)
626 {
627 cgraph_node *next_inline_clone, *replacement;
628
629 for (next_inline_clone = clones;
630 next_inline_clone
631 && next_inline_clone->decl != decl;
632 next_inline_clone = next_inline_clone->next_sibling_clone)
633 ;
634
635 /* If there is inline clone of the node being removed, we need
636 to put it into the position of removed node and reorganize all
637 other clones to be based on it. */
638 if (next_inline_clone)
639 {
640 cgraph_node *n;
641 cgraph_node *new_clones;
642
643 replacement = next_inline_clone;
644
645 /* Unlink inline clone from the list of clones of removed node. */
646 if (next_inline_clone->next_sibling_clone)
647 next_inline_clone->next_sibling_clone->prev_sibling_clone
648 = next_inline_clone->prev_sibling_clone;
649 if (next_inline_clone->prev_sibling_clone)
650 {
651 gcc_assert (clones != next_inline_clone);
652 next_inline_clone->prev_sibling_clone->next_sibling_clone
653 = next_inline_clone->next_sibling_clone;
654 }
655 else
656 {
657 gcc_assert (clones == next_inline_clone);
658 clones = next_inline_clone->next_sibling_clone;
659 }
660
661 new_clones = clones;
662 clones = NULL;
663
664 /* Copy clone info. */
665 next_inline_clone->clone = clone;
666
667 /* Now place it into clone tree at same level at NODE. */
668 next_inline_clone->clone_of = clone_of;
669 next_inline_clone->prev_sibling_clone = NULL;
670 next_inline_clone->next_sibling_clone = NULL;
671 if (clone_of)
672 {
673 if (clone_of->clones)
674 clone_of->clones->prev_sibling_clone = next_inline_clone;
675 next_inline_clone->next_sibling_clone = clone_of->clones;
676 clone_of->clones = next_inline_clone;
677 }
678
679 /* Merge the clone list. */
680 if (new_clones)
681 {
682 if (!next_inline_clone->clones)
683 next_inline_clone->clones = new_clones;
684 else
685 {
686 n = next_inline_clone->clones;
687 while (n->next_sibling_clone)
688 n = n->next_sibling_clone;
689 n->next_sibling_clone = new_clones;
690 new_clones->prev_sibling_clone = n;
691 }
692 }
693
694 /* Update clone_of pointers. */
695 n = new_clones;
696 while (n)
697 {
698 n->clone_of = next_inline_clone;
699 n = n->next_sibling_clone;
700 }
701 return replacement;
702 }
703 else
704 return NULL;
705 }
706
707 /* Like cgraph_set_call_stmt but walk the clone tree and update all
708 clones sharing the same function body.
709 When WHOLE_SPECULATIVE_EDGES is true, all three components of
710 speculative edge gets updated. Otherwise we update only direct
711 call. */
712
713 void
714 cgraph_node::set_call_stmt_including_clones (gimple *old_stmt,
715 gcall *new_stmt,
716 bool update_speculative)
717 {
718 cgraph_node *node;
719 cgraph_edge *edge = get_edge (old_stmt);
720
721 if (edge)
722 edge->set_call_stmt (new_stmt, update_speculative);
723
724 node = clones;
725 if (node)
726 while (node != this)
727 {
728 cgraph_edge *edge = node->get_edge (old_stmt);
729 if (edge)
730 {
731 edge->set_call_stmt (new_stmt, update_speculative);
732 /* If UPDATE_SPECULATIVE is false, it means that we are turning
733 speculative call into a real code sequence. Update the
734 callgraph edges. */
735 if (edge->speculative && !update_speculative)
736 {
737 cgraph_edge *direct, *indirect;
738 ipa_ref *ref;
739
740 gcc_assert (!edge->indirect_unknown_callee);
741 edge->speculative_call_info (direct, indirect, ref);
742 direct->speculative = false;
743 indirect->speculative = false;
744 ref->speculative = false;
745 }
746 }
747 if (node->clones)
748 node = node->clones;
749 else if (node->next_sibling_clone)
750 node = node->next_sibling_clone;
751 else
752 {
753 while (node != this && !node->next_sibling_clone)
754 node = node->clone_of;
755 if (node != this)
756 node = node->next_sibling_clone;
757 }
758 }
759 }
760
761 /* Like cgraph_create_edge walk the clone tree and update all clones sharing
762 same function body. If clones already have edge for OLD_STMT; only
763 update the edge same way as cgraph_set_call_stmt_including_clones does.
764
765 TODO: COUNT and LOOP_DEPTH should be properly distributed based on relative
766 frequencies of the clones. */
767
768 void
769 cgraph_node::create_edge_including_clones (cgraph_node *callee,
770 gimple *old_stmt, gcall *stmt,
771 profile_count count,
772 cgraph_inline_failed_t reason)
773 {
774 cgraph_node *node;
775 cgraph_edge *edge;
776
777 if (!get_edge (stmt))
778 {
779 edge = create_edge (callee, stmt, count);
780 edge->inline_failed = reason;
781 }
782
783 node = clones;
784 if (node)
785 while (node != this)
786 /* Thunk clones do not get updated while copying inline function body. */
787 if (!node->thunk.thunk_p)
788 {
789 cgraph_edge *edge = node->get_edge (old_stmt);
790
791 /* It is possible that clones already contain the edge while
792 master didn't. Either we promoted indirect call into direct
793 call in the clone or we are processing clones of unreachable
794 master where edges has been removed. */
795 if (edge)
796 edge->set_call_stmt (stmt);
797 else if (! node->get_edge (stmt))
798 {
799 edge = node->create_edge (callee, stmt, count);
800 edge->inline_failed = reason;
801 }
802
803 if (node->clones)
804 node = node->clones;
805 else if (node->next_sibling_clone)
806 node = node->next_sibling_clone;
807 else
808 {
809 while (node != this && !node->next_sibling_clone)
810 node = node->clone_of;
811 if (node != this)
812 node = node->next_sibling_clone;
813 }
814 }
815 }
816
817 /* Remove the node from cgraph and all inline clones inlined into it.
818 Skip however removal of FORBIDDEN_NODE and return true if it needs to be
819 removed. This allows to call the function from outer loop walking clone
820 tree. */
821
822 bool
823 cgraph_node::remove_symbol_and_inline_clones (cgraph_node *forbidden_node)
824 {
825 cgraph_edge *e, *next;
826 bool found = false;
827
828 if (this == forbidden_node)
829 {
830 callers->remove ();
831 return true;
832 }
833 for (e = callees; e; e = next)
834 {
835 next = e->next_callee;
836 if (!e->inline_failed)
837 found |= e->callee->remove_symbol_and_inline_clones (forbidden_node);
838 }
839 remove ();
840 return found;
841 }
842
843 /* The edges representing the callers of the NEW_VERSION node were
844 fixed by cgraph_function_versioning (), now the call_expr in their
845 respective tree code should be updated to call the NEW_VERSION. */
846
847 static void
848 update_call_expr (cgraph_node *new_version)
849 {
850 cgraph_edge *e;
851
852 gcc_assert (new_version);
853
854 /* Update the call expr on the edges to call the new version. */
855 for (e = new_version->callers; e; e = e->next_caller)
856 {
857 function *inner_function = DECL_STRUCT_FUNCTION (e->caller->decl);
858 gimple_call_set_fndecl (e->call_stmt, new_version->decl);
859 maybe_clean_eh_stmt_fn (inner_function, e->call_stmt);
860 }
861 }
862
863
864 /* Create a new cgraph node which is the new version of
865 callgraph node. REDIRECT_CALLERS holds the callers
866 edges which should be redirected to point to
867 NEW_VERSION. ALL the callees edges of the node
868 are cloned to the new version node. Return the new
869 version node.
870
871 If non-NULL BLOCK_TO_COPY determine what basic blocks
872 was copied to prevent duplications of calls that are dead
873 in the clone. */
874
875 cgraph_node *
876 cgraph_node::create_version_clone (tree new_decl,
877 vec<cgraph_edge *> redirect_callers,
878 bitmap bbs_to_copy,
879 const char *suffix)
880 {
881 cgraph_node *new_version;
882 cgraph_edge *e;
883 unsigned i;
884
885 new_version = cgraph_node::create (new_decl);
886
887 new_version->analyzed = analyzed;
888 new_version->definition = definition;
889 new_version->local = local;
890 new_version->externally_visible = false;
891 new_version->no_reorder = no_reorder;
892 new_version->local.local = new_version->definition;
893 new_version->global = global;
894 new_version->rtl = rtl;
895 new_version->count = count;
896
897 for (e = callees; e; e=e->next_callee)
898 if (!bbs_to_copy
899 || bitmap_bit_p (bbs_to_copy, gimple_bb (e->call_stmt)->index))
900 e->clone (new_version, e->call_stmt,
901 e->lto_stmt_uid, count, count,
902 true);
903 for (e = indirect_calls; e; e=e->next_callee)
904 if (!bbs_to_copy
905 || bitmap_bit_p (bbs_to_copy, gimple_bb (e->call_stmt)->index))
906 e->clone (new_version, e->call_stmt,
907 e->lto_stmt_uid, count, count,
908 true);
909 FOR_EACH_VEC_ELT (redirect_callers, i, e)
910 {
911 /* Redirect calls to the old version node to point to its new
912 version. */
913 e->redirect_callee (new_version);
914 }
915
916 symtab->call_cgraph_duplication_hooks (this, new_version);
917
918 dump_callgraph_transformation (this, new_version, suffix);
919
920 return new_version;
921 }
922
923 /* Perform function versioning.
924 Function versioning includes copying of the tree and
925 a callgraph update (creating a new cgraph node and updating
926 its callees and callers).
927
928 REDIRECT_CALLERS varray includes the edges to be redirected
929 to the new version.
930
931 TREE_MAP is a mapping of tree nodes we want to replace with
932 new ones (according to results of prior analysis).
933
934 If non-NULL ARGS_TO_SKIP determine function parameters to remove
935 from new version.
936 If SKIP_RETURN is true, the new version will return void.
937 If non-NULL BLOCK_TO_COPY determine what basic blocks to copy.
938 If non_NULL NEW_ENTRY determine new entry BB of the clone.
939
940 Return the new version's cgraph node. */
941
942 cgraph_node *
943 cgraph_node::create_version_clone_with_body
944 (vec<cgraph_edge *> redirect_callers,
945 vec<ipa_replace_map *, va_gc> *tree_map, bitmap args_to_skip,
946 bool skip_return, bitmap bbs_to_copy, basic_block new_entry_block,
947 const char *suffix)
948 {
949 tree old_decl = decl;
950 cgraph_node *new_version_node = NULL;
951 tree new_decl;
952
953 if (!tree_versionable_function_p (old_decl))
954 return NULL;
955
956 gcc_assert (local.can_change_signature || !args_to_skip);
957
958 /* Make a new FUNCTION_DECL tree node for the new version. */
959 if (!args_to_skip && !skip_return)
960 new_decl = copy_node (old_decl);
961 else
962 new_decl
963 = build_function_decl_skip_args (old_decl, args_to_skip, skip_return);
964
965 /* Generate a new name for the new version. */
966 DECL_NAME (new_decl) = clone_function_name (old_decl, suffix);
967 SET_DECL_ASSEMBLER_NAME (new_decl, DECL_NAME (new_decl));
968 SET_DECL_RTL (new_decl, NULL);
969
970 /* When the old decl was a con-/destructor make sure the clone isn't. */
971 DECL_STATIC_CONSTRUCTOR (new_decl) = 0;
972 DECL_STATIC_DESTRUCTOR (new_decl) = 0;
973
974 /* Create the new version's call-graph node.
975 and update the edges of the new node. */
976 new_version_node = create_version_clone (new_decl, redirect_callers,
977 bbs_to_copy, suffix);
978
979 if (ipa_transforms_to_apply.exists ())
980 new_version_node->ipa_transforms_to_apply
981 = ipa_transforms_to_apply.copy ();
982 /* Copy the OLD_VERSION_NODE function tree to the new version. */
983 tree_function_versioning (old_decl, new_decl, tree_map, false, args_to_skip,
984 skip_return, bbs_to_copy, new_entry_block);
985
986 /* Update the new version's properties.
987 Make The new version visible only within this translation unit. Make sure
988 that is not weak also.
989 ??? We cannot use COMDAT linkage because there is no
990 ABI support for this. */
991 new_version_node->make_decl_local ();
992 DECL_VIRTUAL_P (new_version_node->decl) = 0;
993 new_version_node->externally_visible = 0;
994 new_version_node->local.local = 1;
995 new_version_node->lowered = true;
996 if (!implicit_section)
997 new_version_node->set_section (get_section ());
998 /* Clones of global symbols or symbols with unique names are unique. */
999 if ((TREE_PUBLIC (old_decl)
1000 && !DECL_EXTERNAL (old_decl)
1001 && !DECL_WEAK (old_decl)
1002 && !DECL_COMDAT (old_decl))
1003 || in_lto_p)
1004 new_version_node->unique_name = true;
1005
1006 /* Update the call_expr on the edges to call the new version node. */
1007 update_call_expr (new_version_node);
1008
1009 symtab->call_cgraph_insertion_hooks (new_version_node);
1010 return new_version_node;
1011 }
1012
1013 /* Given virtual clone, turn it into actual clone. */
1014
1015 static void
1016 cgraph_materialize_clone (cgraph_node *node)
1017 {
1018 bitmap_obstack_initialize (NULL);
1019 node->former_clone_of = node->clone_of->decl;
1020 if (node->clone_of->former_clone_of)
1021 node->former_clone_of = node->clone_of->former_clone_of;
1022 /* Copy the OLD_VERSION_NODE function tree to the new version. */
1023 tree_function_versioning (node->clone_of->decl, node->decl,
1024 node->clone.tree_map, true,
1025 node->clone.args_to_skip, false,
1026 NULL, NULL);
1027 if (symtab->dump_file)
1028 {
1029 dump_function_to_file (node->clone_of->decl, symtab->dump_file,
1030 dump_flags);
1031 dump_function_to_file (node->decl, symtab->dump_file, dump_flags);
1032 }
1033
1034 /* Function is no longer clone. */
1035 if (node->next_sibling_clone)
1036 node->next_sibling_clone->prev_sibling_clone = node->prev_sibling_clone;
1037 if (node->prev_sibling_clone)
1038 node->prev_sibling_clone->next_sibling_clone = node->next_sibling_clone;
1039 else
1040 node->clone_of->clones = node->next_sibling_clone;
1041 node->next_sibling_clone = NULL;
1042 node->prev_sibling_clone = NULL;
1043 if (!node->clone_of->analyzed && !node->clone_of->clones)
1044 {
1045 node->clone_of->release_body ();
1046 node->clone_of->remove_callees ();
1047 node->clone_of->remove_all_references ();
1048 }
1049 node->clone_of = NULL;
1050 bitmap_obstack_release (NULL);
1051 }
1052
1053 /* Once all functions from compilation unit are in memory, produce all clones
1054 and update all calls. We might also do this on demand if we don't want to
1055 bring all functions to memory prior compilation, but current WHOPR
1056 implementation does that and it is a bit easier to keep everything right in
1057 this order. */
1058
1059 void
1060 symbol_table::materialize_all_clones (void)
1061 {
1062 cgraph_node *node;
1063 bool stabilized = false;
1064
1065
1066 if (symtab->dump_file)
1067 fprintf (symtab->dump_file, "Materializing clones\n");
1068
1069 cgraph_node::checking_verify_cgraph_nodes ();
1070
1071 /* We can also do topological order, but number of iterations should be
1072 bounded by number of IPA passes since single IPA pass is probably not
1073 going to create clones of clones it created itself. */
1074 while (!stabilized)
1075 {
1076 stabilized = true;
1077 FOR_EACH_FUNCTION (node)
1078 {
1079 if (node->clone_of && node->decl != node->clone_of->decl
1080 && !gimple_has_body_p (node->decl))
1081 {
1082 if (!node->clone_of->clone_of)
1083 node->clone_of->get_untransformed_body ();
1084 if (gimple_has_body_p (node->clone_of->decl))
1085 {
1086 if (symtab->dump_file)
1087 {
1088 fprintf (symtab->dump_file, "cloning %s to %s\n",
1089 xstrdup_for_dump (node->clone_of->name ()),
1090 xstrdup_for_dump (node->name ()));
1091 if (node->clone.tree_map)
1092 {
1093 unsigned int i;
1094 fprintf (symtab->dump_file, " replace map: ");
1095 for (i = 0;
1096 i < vec_safe_length (node->clone.tree_map);
1097 i++)
1098 {
1099 ipa_replace_map *replace_info;
1100 replace_info = (*node->clone.tree_map)[i];
1101 print_generic_expr (symtab->dump_file,
1102 replace_info->old_tree);
1103 fprintf (symtab->dump_file, " -> ");
1104 print_generic_expr (symtab->dump_file,
1105 replace_info->new_tree);
1106 fprintf (symtab->dump_file, "%s%s;",
1107 replace_info->replace_p ? "(replace)":"",
1108 replace_info->ref_p ? "(ref)":"");
1109 }
1110 fprintf (symtab->dump_file, "\n");
1111 }
1112 if (node->clone.args_to_skip)
1113 {
1114 fprintf (symtab->dump_file, " args_to_skip: ");
1115 dump_bitmap (symtab->dump_file,
1116 node->clone.args_to_skip);
1117 }
1118 if (node->clone.args_to_skip)
1119 {
1120 fprintf (symtab->dump_file, " combined_args_to_skip:");
1121 dump_bitmap (symtab->dump_file, node->clone.combined_args_to_skip);
1122 }
1123 }
1124 cgraph_materialize_clone (node);
1125 stabilized = false;
1126 }
1127 }
1128 }
1129 }
1130 FOR_EACH_FUNCTION (node)
1131 if (!node->analyzed && node->callees)
1132 {
1133 node->remove_callees ();
1134 node->remove_all_references ();
1135 }
1136 else
1137 node->clear_stmts_in_references ();
1138 if (symtab->dump_file)
1139 fprintf (symtab->dump_file, "Materialization Call site updates done.\n");
1140
1141 cgraph_node::checking_verify_cgraph_nodes ();
1142
1143 symtab->remove_unreachable_nodes (symtab->dump_file);
1144 }
1145
1146 #include "gt-cgraphclones.h"