]> git.ipfire.org Git - thirdparty/gcc.git/blame - gcc/function.c
Update copyright years.
[thirdparty/gcc.git] / gcc / function.c
CommitLineData
5e6908ea 1/* Expands front end tree to back end RTL for GCC.
a5544970 2 Copyright (C) 1987-2019 Free Software Foundation, Inc.
6f086dfc 3
1322177d 4This file is part of GCC.
6f086dfc 5
1322177d
LB
6GCC is free software; you can redistribute it and/or modify it under
7the terms of the GNU General Public License as published by the Free
9dcd6f09 8Software Foundation; either version 3, or (at your option) any later
1322177d 9version.
6f086dfc 10
1322177d
LB
11GCC is distributed in the hope that it will be useful, but WITHOUT ANY
12WARRANTY; without even the implied warranty of MERCHANTABILITY or
13FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
14for more details.
6f086dfc
RS
15
16You should have received a copy of the GNU General Public License
9dcd6f09
NC
17along with GCC; see the file COPYING3. If not see
18<http://www.gnu.org/licenses/>. */
6f086dfc 19
6f086dfc
RS
20/* This file handles the generation of rtl code from tree structure
21 at the level of the function as a whole.
22 It creates the rtl expressions for parameters and auto variables
23 and has full responsibility for allocating stack slots.
24
25 `expand_function_start' is called at the beginning of a function,
26 before the function body is parsed, and `expand_function_end' is
27 called after parsing the body.
28
29 Call `assign_stack_local' to allocate a stack slot for a local variable.
30 This is usually done during the RTL generation for the function body,
31 but it can also be done in the reload pass when a pseudo-register does
8fff4fc1 32 not get a hard register. */
6f086dfc
RS
33
34#include "config.h"
670ee920 35#include "system.h"
4977bab6 36#include "coretypes.h"
c7131fb2 37#include "backend.h"
957060b5 38#include "target.h"
c7131fb2 39#include "rtl.h"
957060b5
AM
40#include "tree.h"
41#include "gimple-expr.h"
42#include "cfghooks.h"
c7131fb2 43#include "df.h"
4d0cdd0c 44#include "memmodel.h"
957060b5
AM
45#include "tm_p.h"
46#include "stringpool.h"
47#include "expmed.h"
48#include "optabs.h"
49#include "regs.h"
50#include "emit-rtl.h"
51#include "recog.h"
0cbd9993 52#include "rtl-error.h"
40e23961 53#include "alias.h"
40e23961 54#include "fold-const.h"
d8a2d370
DN
55#include "stor-layout.h"
56#include "varasm.h"
1ef08c63 57#include "except.h"
36566b39
PK
58#include "dojump.h"
59#include "explow.h"
60#include "calls.h"
6f086dfc 61#include "expr.h"
385399a8 62#include "optabs-tree.h"
6f086dfc 63#include "output.h"
7afff7cf 64#include "langhooks.h"
677f3fa8 65#include "common/common-target.h"
45b0be94 66#include "gimplify.h"
ef330312 67#include "tree-pass.h"
60393bbc
AM
68#include "cfgrtl.h"
69#include "cfganal.h"
70#include "cfgbuild.h"
71#include "cfgcleanup.h"
1f9ceff1 72#include "cfgexpand.h"
f30e25a3 73#include "shrink-wrap.h"
b9b5f433 74#include "toplev.h"
b8704801 75#include "rtl-iter.h"
f11a7b6d 76#include "tree-dfa.h"
b3e46655 77#include "tree-ssa.h"
314e6352
ML
78#include "stringpool.h"
79#include "attribs.h"
6aee2fd0
JJ
80#include "gimple.h"
81#include "options.h"
7d69de61 82
5576d6f2
TT
83/* So we can assign to cfun in this file. */
84#undef cfun
85
95f3f59e
JDA
86#ifndef STACK_ALIGNMENT_NEEDED
87#define STACK_ALIGNMENT_NEEDED 1
88#endif
89
975f3818
RS
90#define STACK_BYTES (STACK_BOUNDARY / BITS_PER_UNIT)
91
6f086dfc
RS
92/* Round a value to the lowest integer less than it that is a multiple of
93 the required alignment. Avoid using division in case the value is
94 negative. Assume the alignment is a power of two. */
95#define FLOOR_ROUND(VALUE,ALIGN) ((VALUE) & ~((ALIGN) - 1))
96
97/* Similar, but round to the next highest integer that meets the
98 alignment. */
99#define CEIL_ROUND(VALUE,ALIGN) (((VALUE) + (ALIGN) - 1) & ~((ALIGN)- 1))
100
6f086dfc 101/* Nonzero once virtual register instantiation has been done.
c39ada04
DD
102 assign_stack_local uses frame_pointer_rtx when this is nonzero.
103 calls.c:emit_library_call_value_1 uses it to set up
104 post-instantiation libcalls. */
105int virtuals_instantiated;
6f086dfc 106
df696a75 107/* Assign unique numbers to labels generated for profiling, debugging, etc. */
17211ab5 108static GTY(()) int funcdef_no;
f6f315fe 109
414c4dc4
NC
110/* These variables hold pointers to functions to create and destroy
111 target specific, per-function data structures. */
fa8db1f7 112struct machine_function * (*init_machine_status) (void);
46766466 113
b384405b 114/* The currently compiled function. */
01d939e8 115struct function *cfun = 0;
b384405b 116
cd9c1ca8 117/* These hashes record the prologue and epilogue insns. */
d242408f 118
6c907cff 119struct insn_cache_hasher : ggc_cache_ptr_hash<rtx_def>
d242408f
TS
120{
121 static hashval_t hash (rtx x) { return htab_hash_pointer (x); }
122 static bool equal (rtx a, rtx b) { return a == b; }
123};
124
125static GTY((cache))
126 hash_table<insn_cache_hasher> *prologue_insn_hash;
127static GTY((cache))
128 hash_table<insn_cache_hasher> *epilogue_insn_hash;
6f086dfc 129\f
b646ba3f 130
2a22f99c 131hash_table<used_type_hasher> *types_used_by_vars_hash = NULL;
9771b263 132vec<tree, va_gc> *types_used_by_cur_var_decl;
b646ba3f 133
e15679f8
RK
134/* Forward declarations. */
135
fa8db1f7 136static struct temp_slot *find_temp_slot_from_address (rtx);
fa8db1f7 137static void pad_to_arg_alignment (struct args_size *, int, struct args_size *);
ef4bddc2 138static void pad_below (struct args_size *, machine_mode, tree);
691fe203 139static void reorder_blocks_1 (rtx_insn *, tree, vec<tree> *);
fa8db1f7
AJ
140static int all_blocks (tree, tree *);
141static tree *get_block_vector (tree, int *);
142extern tree debug_find_var_in_block_tree (tree, tree);
1f52178b 143/* We always define `record_insns' even if it's not used so that we
ec97b83a 144 can always export `prologue_epilogue_contains'. */
d242408f
TS
145static void record_insns (rtx_insn *, rtx, hash_table<insn_cache_hasher> **)
146 ATTRIBUTE_UNUSED;
87ac59a0 147static bool contains (const rtx_insn *, hash_table<insn_cache_hasher> *);
db2960f4 148static void prepare_function_start (void);
fa8db1f7
AJ
149static void do_clobber_return_reg (rtx, void *);
150static void do_use_return_reg (rtx, void *);
1f9ceff1 151
c20bf1f3 152\f
936fc9ba
JH
153/* Stack of nested functions. */
154/* Keep track of the cfun stack. */
e5e809f4 155
526ceb68 156static vec<function *> function_context_stack;
6f086dfc
RS
157
158/* Save the current context for compilation of a nested function.
d2784db4 159 This is called from language-specific code. */
6f086dfc
RS
160
161void
d2784db4 162push_function_context (void)
6f086dfc 163{
01d939e8 164 if (cfun == 0)
182e0d71 165 allocate_struct_function (NULL, false);
b384405b 166
9771b263 167 function_context_stack.safe_push (cfun);
db2960f4 168 set_cfun (NULL);
6f086dfc
RS
169}
170
171/* Restore the last saved context, at the end of a nested function.
172 This function is called from language-specific code. */
173
174void
d2784db4 175pop_function_context (void)
6f086dfc 176{
9771b263 177 struct function *p = function_context_stack.pop ();
db2960f4 178 set_cfun (p);
6f086dfc 179 current_function_decl = p->decl;
6f086dfc 180
6f086dfc 181 /* Reset variables that have known state during rtx generation. */
6f086dfc 182 virtuals_instantiated = 0;
1b3d8f8a 183 generating_concat_p = 1;
6f086dfc 184}
e4a4639e 185
fa51b01b
RH
186/* Clear out all parts of the state in F that can safely be discarded
187 after the function has been parsed, but not compiled, to let
188 garbage collection reclaim the memory. */
189
190void
fa8db1f7 191free_after_parsing (struct function *f)
fa51b01b 192{
e8924938 193 f->language = 0;
fa51b01b
RH
194}
195
e2ecd91c
BS
196/* Clear out all parts of the state in F that can safely be discarded
197 after the function has been compiled, to let garbage collection
0a8a198c 198 reclaim the memory. */
21cd906e 199
e2ecd91c 200void
fa8db1f7 201free_after_compilation (struct function *f)
e2ecd91c 202{
cd9c1ca8
RH
203 prologue_insn_hash = NULL;
204 epilogue_insn_hash = NULL;
205
04695783 206 free (crtl->emit.regno_pointer_align);
f995dcfe 207
3e029763 208 memset (crtl, 0, sizeof (struct rtl_data));
e2500fed 209 f->eh = NULL;
e2500fed 210 f->machine = NULL;
997de8ed 211 f->cfg = NULL;
b11f11a1 212 f->curr_properties &= ~PROP_cfg;
fa51b01b 213
57b9e367 214 regno_reg_rtx = NULL;
e2ecd91c 215}
6f086dfc 216\f
49ad7cfa
BS
217/* Return size needed for stack frame based on slots so far allocated.
218 This size counts from zero. It is not rounded to PREFERRED_STACK_BOUNDARY;
219 the caller may have to do that. */
9fb798d7 220
f075bd95 221poly_int64
fa8db1f7 222get_frame_size (void)
49ad7cfa 223{
bd60bab2
JH
224 if (FRAME_GROWS_DOWNWARD)
225 return -frame_offset;
226 else
227 return frame_offset;
49ad7cfa
BS
228}
229
9fb798d7
EB
230/* Issue an error message and return TRUE if frame OFFSET overflows in
231 the signed target pointer arithmetics for function FUNC. Otherwise
232 return FALSE. */
233
234bool
f075bd95 235frame_offset_overflow (poly_int64 offset, tree func)
b8698a0f 236{
f075bd95
RS
237 poly_uint64 size = FRAME_GROWS_DOWNWARD ? -offset : offset;
238 unsigned HOST_WIDE_INT limit
239 = ((HOST_WIDE_INT_1U << (GET_MODE_BITSIZE (Pmode) - 1))
240 /* Leave room for the fixed part of the frame. */
241 - 64 * UNITS_PER_WORD);
9fb798d7 242
f075bd95 243 if (!coeffs_in_range_p (size, 0U, limit))
9fb798d7 244 {
00abf86c
MS
245 unsigned HOST_WIDE_INT hwisize;
246 if (size.is_constant (&hwisize))
247 error_at (DECL_SOURCE_LOCATION (func),
248 "total size of local objects %wu exceeds maximum %wu",
249 hwisize, limit);
250 else
251 error_at (DECL_SOURCE_LOCATION (func),
252 "total size of local objects exceeds maximum %wu",
253 limit);
f075bd95 254 return true;
9fb798d7
EB
255 }
256
f075bd95 257 return false;
9fb798d7
EB
258}
259
83d0488b
RS
260/* Return the minimum spill slot alignment for a register of mode MODE. */
261
262unsigned int
263spill_slot_alignment (machine_mode mode ATTRIBUTE_UNUSED)
264{
265 return STACK_SLOT_ALIGNMENT (NULL_TREE, mode, GET_MODE_ALIGNMENT (mode));
266}
267
76fe54f0
L
268/* Return stack slot alignment in bits for TYPE and MODE. */
269
270static unsigned int
ef4bddc2 271get_stack_local_alignment (tree type, machine_mode mode)
76fe54f0
L
272{
273 unsigned int alignment;
274
275 if (mode == BLKmode)
276 alignment = BIGGEST_ALIGNMENT;
277 else
278 alignment = GET_MODE_ALIGNMENT (mode);
279
280 /* Allow the frond-end to (possibly) increase the alignment of this
281 stack slot. */
282 if (! type)
283 type = lang_hooks.types.type_for_mode (mode, 0);
284
285 return STACK_SLOT_ALIGNMENT (type, mode, alignment);
286}
287
56731d64
BS
288/* Determine whether it is possible to fit a stack slot of size SIZE and
289 alignment ALIGNMENT into an area in the stack frame that starts at
290 frame offset START and has a length of LENGTH. If so, store the frame
291 offset to be used for the stack slot in *POFFSET and return true;
292 return false otherwise. This function will extend the frame size when
293 given a start/length pair that lies at the end of the frame. */
294
295static bool
f075bd95
RS
296try_fit_stack_local (poly_int64 start, poly_int64 length,
297 poly_int64 size, unsigned int alignment,
298 poly_int64_pod *poffset)
56731d64 299{
f075bd95 300 poly_int64 this_frame_offset;
56731d64
BS
301 int frame_off, frame_alignment, frame_phase;
302
303 /* Calculate how many bytes the start of local variables is off from
304 stack alignment. */
305 frame_alignment = PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT;
2a31c321 306 frame_off = targetm.starting_frame_offset () % frame_alignment;
56731d64
BS
307 frame_phase = frame_off ? frame_alignment - frame_off : 0;
308
309 /* Round the frame offset to the specified alignment. */
310
56731d64
BS
311 if (FRAME_GROWS_DOWNWARD)
312 this_frame_offset
f075bd95 313 = (aligned_lower_bound (start + length - size - frame_phase, alignment)
56731d64
BS
314 + frame_phase);
315 else
316 this_frame_offset
f075bd95 317 = aligned_upper_bound (start - frame_phase, alignment) + frame_phase;
56731d64
BS
318
319 /* See if it fits. If this space is at the edge of the frame,
320 consider extending the frame to make it fit. Our caller relies on
321 this when allocating a new slot. */
f075bd95
RS
322 if (maybe_lt (this_frame_offset, start))
323 {
324 if (known_eq (frame_offset, start))
325 frame_offset = this_frame_offset;
326 else
327 return false;
328 }
329 else if (maybe_gt (this_frame_offset + size, start + length))
330 {
331 if (known_eq (frame_offset, start + length))
332 frame_offset = this_frame_offset + size;
333 else
334 return false;
335 }
56731d64
BS
336
337 *poffset = this_frame_offset;
338 return true;
339}
340
341/* Create a new frame_space structure describing free space in the stack
342 frame beginning at START and ending at END, and chain it into the
343 function's frame_space_list. */
344
345static void
f075bd95 346add_frame_space (poly_int64 start, poly_int64 end)
56731d64 347{
766090c2 348 struct frame_space *space = ggc_alloc<frame_space> ();
56731d64
BS
349 space->next = crtl->frame_space_list;
350 crtl->frame_space_list = space;
351 space->start = start;
352 space->length = end - start;
353}
354
6f086dfc
RS
355/* Allocate a stack slot of SIZE bytes and return a MEM rtx for it
356 with machine mode MODE.
718fe406 357
6f086dfc
RS
358 ALIGN controls the amount of alignment for the address of the slot:
359 0 means according to MODE,
360 -1 means use BIGGEST_ALIGNMENT and round size to multiple of that,
cfa29a4c 361 -2 means use BITS_PER_UNIT,
6f086dfc
RS
362 positive specifies alignment boundary in bits.
363
80a832cd
JJ
364 KIND has ASLK_REDUCE_ALIGN bit set if it is OK to reduce
365 alignment and ASLK_RECORD_PAD bit set if we should remember
366 extra space we allocated for alignment purposes. When we are
367 called from assign_stack_temp_for_type, it is not set so we don't
368 track the same stack slot in two independent lists.
2e3f842f 369
bd60bab2 370 We do not round to stack_boundary here. */
6f086dfc 371
bd60bab2 372rtx
f075bd95 373assign_stack_local_1 (machine_mode mode, poly_int64 size,
80a832cd 374 int align, int kind)
6f086dfc 375{
b3694847 376 rtx x, addr;
f075bd95
RS
377 poly_int64 bigend_correction = 0;
378 poly_int64 slot_offset = 0, old_frame_offset;
76fe54f0 379 unsigned int alignment, alignment_in_bits;
0358d788 380 bool dynamic_align_addr = false;
6f086dfc
RS
381
382 if (align == 0)
383 {
76fe54f0 384 alignment = get_stack_local_alignment (NULL, mode);
d16790f2 385 alignment /= BITS_PER_UNIT;
6f086dfc
RS
386 }
387 else if (align == -1)
388 {
389 alignment = BIGGEST_ALIGNMENT / BITS_PER_UNIT;
f075bd95 390 size = aligned_upper_bound (size, alignment);
6f086dfc 391 }
cfa29a4c
EB
392 else if (align == -2)
393 alignment = 1; /* BITS_PER_UNIT / BITS_PER_UNIT */
6f086dfc
RS
394 else
395 alignment = align / BITS_PER_UNIT;
396
2e3f842f
L
397 alignment_in_bits = alignment * BITS_PER_UNIT;
398
2e3f842f
L
399 if (alignment_in_bits > MAX_SUPPORTED_STACK_ALIGNMENT)
400 {
0358d788
RL
401 /* If the required alignment exceeds MAX_SUPPORTED_STACK_ALIGNMENT and
402 it is not OK to reduce it. Align the slot dynamically. */
403 if (mode == BLKmode && (kind & ASLK_REDUCE_ALIGN) == 0)
404 dynamic_align_addr = true;
405 else
406 {
407 alignment_in_bits = MAX_SUPPORTED_STACK_ALIGNMENT;
408 alignment = MAX_SUPPORTED_STACK_ALIGNMENT / BITS_PER_UNIT;
409 }
2e3f842f 410 }
a0871656 411
0358d788 412 if (SUPPORTS_STACK_ALIGNMENT && !dynamic_align_addr)
2e3f842f
L
413 {
414 if (crtl->stack_alignment_estimated < alignment_in_bits)
415 {
416 if (!crtl->stack_realign_processed)
417 crtl->stack_alignment_estimated = alignment_in_bits;
418 else
419 {
420 /* If stack is realigned and stack alignment value
421 hasn't been finalized, it is OK not to increase
422 stack_alignment_estimated. The bigger alignment
423 requirement is recorded in stack_alignment_needed
424 below. */
425 gcc_assert (!crtl->stack_realign_finalized);
426 if (!crtl->stack_realign_needed)
427 {
428 /* It is OK to reduce the alignment as long as the
429 requested size is 0 or the estimated stack
430 alignment >= mode alignment. */
80a832cd 431 gcc_assert ((kind & ASLK_REDUCE_ALIGN)
f075bd95 432 || known_eq (size, 0)
2e3f842f
L
433 || (crtl->stack_alignment_estimated
434 >= GET_MODE_ALIGNMENT (mode)));
435 alignment_in_bits = crtl->stack_alignment_estimated;
436 alignment = alignment_in_bits / BITS_PER_UNIT;
437 }
438 }
439 }
440 }
76fe54f0 441
0358d788
RL
442 /* Handle overalignment here for parameter copy on the stack.
443 Reserved enough space for it and dynamically align the address.
444 No free frame_space is added here. */
445 if (dynamic_align_addr)
446 {
447 rtx allocsize = gen_int_mode (size, Pmode);
448 get_dynamic_stack_size (&allocsize, 0, alignment_in_bits, NULL);
449
450 /* This is the size of space needed to accommodate required size of data
451 with given alignment. */
452 poly_int64 len = rtx_to_poly_int64 (allocsize);
453 old_frame_offset = frame_offset;
454
455 if (FRAME_GROWS_DOWNWARD)
456 {
457 frame_offset -= len;
458 try_fit_stack_local (frame_offset, len, len,
459 PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT,
460 &slot_offset);
461 }
462 else
463 {
464 frame_offset += len;
465 try_fit_stack_local (old_frame_offset, len, len,
466 PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT,
467 &slot_offset);
468 }
469 goto found_space;
470 }
471 else
472 {
473 if (crtl->stack_alignment_needed < alignment_in_bits)
474 crtl->stack_alignment_needed = alignment_in_bits;
475 if (crtl->max_used_stack_slot_alignment < alignment_in_bits)
476 crtl->max_used_stack_slot_alignment = alignment_in_bits;
477 }
a0871656 478
f075bd95 479 if (mode != BLKmode || maybe_ne (size, 0))
56731d64 480 {
80a832cd 481 if (kind & ASLK_RECORD_PAD)
56731d64 482 {
80a832cd
JJ
483 struct frame_space **psp;
484
485 for (psp = &crtl->frame_space_list; *psp; psp = &(*psp)->next)
486 {
487 struct frame_space *space = *psp;
488 if (!try_fit_stack_local (space->start, space->length, size,
489 alignment, &slot_offset))
490 continue;
491 *psp = space->next;
f075bd95 492 if (known_gt (slot_offset, space->start))
80a832cd 493 add_frame_space (space->start, slot_offset);
f075bd95 494 if (known_lt (slot_offset + size, space->start + space->length))
80a832cd
JJ
495 add_frame_space (slot_offset + size,
496 space->start + space->length);
497 goto found_space;
498 }
56731d64
BS
499 }
500 }
501 else if (!STACK_ALIGNMENT_NEEDED)
502 {
503 slot_offset = frame_offset;
504 goto found_space;
505 }
506
507 old_frame_offset = frame_offset;
508
509 if (FRAME_GROWS_DOWNWARD)
510 {
511 frame_offset -= size;
512 try_fit_stack_local (frame_offset, size, size, alignment, &slot_offset);
58dbcf05 513
80a832cd
JJ
514 if (kind & ASLK_RECORD_PAD)
515 {
f075bd95 516 if (known_gt (slot_offset, frame_offset))
80a832cd 517 add_frame_space (frame_offset, slot_offset);
f075bd95 518 if (known_lt (slot_offset + size, old_frame_offset))
80a832cd
JJ
519 add_frame_space (slot_offset + size, old_frame_offset);
520 }
56731d64
BS
521 }
522 else
95f3f59e 523 {
56731d64
BS
524 frame_offset += size;
525 try_fit_stack_local (old_frame_offset, size, size, alignment, &slot_offset);
526
80a832cd
JJ
527 if (kind & ASLK_RECORD_PAD)
528 {
f075bd95 529 if (known_gt (slot_offset, old_frame_offset))
80a832cd 530 add_frame_space (old_frame_offset, slot_offset);
f075bd95 531 if (known_lt (slot_offset + size, frame_offset))
80a832cd
JJ
532 add_frame_space (slot_offset + size, frame_offset);
533 }
95f3f59e 534 }
6f086dfc 535
56731d64 536 found_space:
6f086dfc
RS
537 /* On a big-endian machine, if we are allocating more space than we will use,
538 use the least significant bytes of those that are allocated. */
f075bd95
RS
539 if (mode != BLKmode)
540 {
541 /* The slot size can sometimes be smaller than the mode size;
542 e.g. the rs6000 port allocates slots with a vector mode
543 that have the size of only one element. However, the slot
544 size must always be ordered wrt to the mode size, in the
545 same way as for a subreg. */
546 gcc_checking_assert (ordered_p (GET_MODE_SIZE (mode), size));
547 if (BYTES_BIG_ENDIAN && maybe_lt (GET_MODE_SIZE (mode), size))
548 bigend_correction = size - GET_MODE_SIZE (mode);
549 }
6f086dfc 550
6f086dfc
RS
551 /* If we have already instantiated virtual registers, return the actual
552 address relative to the frame pointer. */
bd60bab2 553 if (virtuals_instantiated)
0a81f074 554 addr = plus_constant (Pmode, frame_pointer_rtx,
c41536f5 555 trunc_int_for_mode
56731d64 556 (slot_offset + bigend_correction
2a31c321 557 + targetm.starting_frame_offset (), Pmode));
6f086dfc 558 else
0a81f074 559 addr = plus_constant (Pmode, virtual_stack_vars_rtx,
c41536f5 560 trunc_int_for_mode
56731d64 561 (slot_offset + bigend_correction,
c41536f5 562 Pmode));
6f086dfc 563
0358d788
RL
564 if (dynamic_align_addr)
565 {
566 addr = align_dynamic_address (addr, alignment_in_bits);
567 mark_reg_pointer (addr, alignment_in_bits);
568 }
569
38a448ca 570 x = gen_rtx_MEM (mode, addr);
76fe54f0 571 set_mem_align (x, alignment_in_bits);
be0c514c 572 MEM_NOTRAP_P (x) = 1;
6f086dfc 573
8c39f8ae 574 vec_safe_push (stack_slot_list, x);
e2ecd91c 575
bd60bab2
JH
576 if (frame_offset_overflow (frame_offset, current_function_decl))
577 frame_offset = 0;
9070115b 578
6f086dfc
RS
579 return x;
580}
2e3f842f
L
581
582/* Wrap up assign_stack_local_1 with last parameter as false. */
583
584rtx
f075bd95 585assign_stack_local (machine_mode mode, poly_int64 size, int align)
2e3f842f 586{
80a832cd 587 return assign_stack_local_1 (mode, size, align, ASLK_RECORD_PAD);
2e3f842f 588}
0aea6467 589\f
fb0703f7
SB
590/* In order to evaluate some expressions, such as function calls returning
591 structures in memory, we need to temporarily allocate stack locations.
592 We record each allocated temporary in the following structure.
593
594 Associated with each temporary slot is a nesting level. When we pop up
595 one level, all temporaries associated with the previous level are freed.
596 Normally, all temporaries are freed after the execution of the statement
597 in which they were created. However, if we are inside a ({...}) grouping,
598 the result may be in a temporary and hence must be preserved. If the
599 result could be in a temporary, we preserve it if we can determine which
600 one it is in. If we cannot determine which temporary may contain the
601 result, all temporaries are preserved. A temporary is preserved by
9474e8ab 602 pretending it was allocated at the previous nesting level. */
fb0703f7 603
d1b38208 604struct GTY(()) temp_slot {
fb0703f7
SB
605 /* Points to next temporary slot. */
606 struct temp_slot *next;
607 /* Points to previous temporary slot. */
608 struct temp_slot *prev;
609 /* The rtx to used to reference the slot. */
610 rtx slot;
fb0703f7 611 /* The size, in units, of the slot. */
f075bd95 612 poly_int64 size;
fb0703f7
SB
613 /* The type of the object in the slot, or zero if it doesn't correspond
614 to a type. We use this to determine whether a slot can be reused.
615 It can be reused if objects of the type of the new slot will always
616 conflict with objects of the type of the old slot. */
617 tree type;
8f5929e1
JJ
618 /* The alignment (in bits) of the slot. */
619 unsigned int align;
fb0703f7
SB
620 /* Nonzero if this temporary is currently in use. */
621 char in_use;
fb0703f7
SB
622 /* Nesting level at which this slot is being used. */
623 int level;
fb0703f7
SB
624 /* The offset of the slot from the frame_pointer, including extra space
625 for alignment. This info is for combine_temp_slots. */
f075bd95 626 poly_int64 base_offset;
fb0703f7
SB
627 /* The size of the slot, including extra space for alignment. This
628 info is for combine_temp_slots. */
f075bd95 629 poly_int64 full_size;
fb0703f7
SB
630};
631
2a22f99c
TS
632/* Entry for the below hash table. */
633struct GTY((for_user)) temp_slot_address_entry {
fb0703f7
SB
634 hashval_t hash;
635 rtx address;
636 struct temp_slot *temp_slot;
637};
638
ca752f39 639struct temp_address_hasher : ggc_ptr_hash<temp_slot_address_entry>
2a22f99c
TS
640{
641 static hashval_t hash (temp_slot_address_entry *);
642 static bool equal (temp_slot_address_entry *, temp_slot_address_entry *);
643};
644
645/* A table of addresses that represent a stack slot. The table is a mapping
646 from address RTXen to a temp slot. */
647static GTY(()) hash_table<temp_address_hasher> *temp_slot_address_table;
648static size_t n_temp_slots_in_use;
649
0aea6467
ZD
650/* Removes temporary slot TEMP from LIST. */
651
652static void
653cut_slot_from_list (struct temp_slot *temp, struct temp_slot **list)
654{
655 if (temp->next)
656 temp->next->prev = temp->prev;
657 if (temp->prev)
658 temp->prev->next = temp->next;
659 else
660 *list = temp->next;
661
662 temp->prev = temp->next = NULL;
663}
664
665/* Inserts temporary slot TEMP to LIST. */
666
667static void
668insert_slot_to_list (struct temp_slot *temp, struct temp_slot **list)
669{
670 temp->next = *list;
671 if (*list)
672 (*list)->prev = temp;
673 temp->prev = NULL;
674 *list = temp;
675}
676
677/* Returns the list of used temp slots at LEVEL. */
678
679static struct temp_slot **
680temp_slots_at_level (int level)
681{
9771b263
DN
682 if (level >= (int) vec_safe_length (used_temp_slots))
683 vec_safe_grow_cleared (used_temp_slots, level + 1);
0aea6467 684
9771b263 685 return &(*used_temp_slots)[level];
0aea6467
ZD
686}
687
688/* Returns the maximal temporary slot level. */
689
690static int
691max_slot_level (void)
692{
693 if (!used_temp_slots)
694 return -1;
695
9771b263 696 return used_temp_slots->length () - 1;
0aea6467
ZD
697}
698
699/* Moves temporary slot TEMP to LEVEL. */
700
701static void
702move_slot_to_level (struct temp_slot *temp, int level)
703{
704 cut_slot_from_list (temp, temp_slots_at_level (temp->level));
705 insert_slot_to_list (temp, temp_slots_at_level (level));
706 temp->level = level;
707}
708
709/* Make temporary slot TEMP available. */
710
711static void
712make_slot_available (struct temp_slot *temp)
713{
714 cut_slot_from_list (temp, temp_slots_at_level (temp->level));
715 insert_slot_to_list (temp, &avail_temp_slots);
716 temp->in_use = 0;
717 temp->level = -1;
f8395d62 718 n_temp_slots_in_use--;
0aea6467 719}
fb0703f7
SB
720
721/* Compute the hash value for an address -> temp slot mapping.
722 The value is cached on the mapping entry. */
723static hashval_t
724temp_slot_address_compute_hash (struct temp_slot_address_entry *t)
725{
726 int do_not_record = 0;
727 return hash_rtx (t->address, GET_MODE (t->address),
728 &do_not_record, NULL, false);
729}
730
731/* Return the hash value for an address -> temp slot mapping. */
2a22f99c
TS
732hashval_t
733temp_address_hasher::hash (temp_slot_address_entry *t)
fb0703f7 734{
fb0703f7
SB
735 return t->hash;
736}
737
738/* Compare two address -> temp slot mapping entries. */
2a22f99c
TS
739bool
740temp_address_hasher::equal (temp_slot_address_entry *t1,
741 temp_slot_address_entry *t2)
fb0703f7 742{
fb0703f7
SB
743 return exp_equiv_p (t1->address, t2->address, 0, true);
744}
745
746/* Add ADDRESS as an alias of TEMP_SLOT to the addess -> temp slot mapping. */
747static void
748insert_temp_slot_address (rtx address, struct temp_slot *temp_slot)
749{
766090c2 750 struct temp_slot_address_entry *t = ggc_alloc<temp_slot_address_entry> ();
fb0703f7
SB
751 t->address = address;
752 t->temp_slot = temp_slot;
753 t->hash = temp_slot_address_compute_hash (t);
2a22f99c 754 *temp_slot_address_table->find_slot_with_hash (t, t->hash, INSERT) = t;
fb0703f7
SB
755}
756
757/* Remove an address -> temp slot mapping entry if the temp slot is
758 not in use anymore. Callback for remove_unused_temp_slot_addresses. */
2a22f99c
TS
759int
760remove_unused_temp_slot_addresses_1 (temp_slot_address_entry **slot, void *)
fb0703f7 761{
2a22f99c 762 const struct temp_slot_address_entry *t = *slot;
fb0703f7 763 if (! t->temp_slot->in_use)
2a22f99c 764 temp_slot_address_table->clear_slot (slot);
fb0703f7
SB
765 return 1;
766}
767
768/* Remove all mappings of addresses to unused temp slots. */
769static void
770remove_unused_temp_slot_addresses (void)
771{
f8395d62
MM
772 /* Use quicker clearing if there aren't any active temp slots. */
773 if (n_temp_slots_in_use)
2a22f99c
TS
774 temp_slot_address_table->traverse
775 <void *, remove_unused_temp_slot_addresses_1> (NULL);
f8395d62 776 else
2a22f99c 777 temp_slot_address_table->empty ();
fb0703f7
SB
778}
779
780/* Find the temp slot corresponding to the object at address X. */
781
782static struct temp_slot *
783find_temp_slot_from_address (rtx x)
784{
785 struct temp_slot *p;
786 struct temp_slot_address_entry tmp, *t;
787
788 /* First try the easy way:
789 See if X exists in the address -> temp slot mapping. */
790 tmp.address = x;
791 tmp.temp_slot = NULL;
792 tmp.hash = temp_slot_address_compute_hash (&tmp);
2a22f99c 793 t = temp_slot_address_table->find_with_hash (&tmp, tmp.hash);
fb0703f7
SB
794 if (t)
795 return t->temp_slot;
796
797 /* If we have a sum involving a register, see if it points to a temp
798 slot. */
799 if (GET_CODE (x) == PLUS && REG_P (XEXP (x, 0))
800 && (p = find_temp_slot_from_address (XEXP (x, 0))) != 0)
801 return p;
802 else if (GET_CODE (x) == PLUS && REG_P (XEXP (x, 1))
803 && (p = find_temp_slot_from_address (XEXP (x, 1))) != 0)
804 return p;
805
806 /* Last resort: Address is a virtual stack var address. */
f075bd95
RS
807 poly_int64 offset;
808 if (strip_offset (x, &offset) == virtual_stack_vars_rtx)
fb0703f7
SB
809 {
810 int i;
811 for (i = max_slot_level (); i >= 0; i--)
812 for (p = *temp_slots_at_level (i); p; p = p->next)
f075bd95
RS
813 if (known_in_range_p (offset, p->base_offset, p->full_size))
814 return p;
fb0703f7
SB
815 }
816
817 return NULL;
818}
6f086dfc
RS
819\f
820/* Allocate a temporary stack slot and record it for possible later
821 reuse.
822
823 MODE is the machine mode to be given to the returned rtx.
824
825 SIZE is the size in units of the space required. We do no rounding here
826 since assign_stack_local will do any required rounding.
827
a4c6502a 828 TYPE is the type that will be used for the stack slot. */
6f086dfc 829
a06ef755 830rtx
f075bd95 831assign_stack_temp_for_type (machine_mode mode, poly_int64 size, tree type)
6f086dfc 832{
74e2819c 833 unsigned int align;
0aea6467 834 struct temp_slot *p, *best_p = 0, *selected = NULL, **pp;
faa964e5 835 rtx slot;
6f086dfc 836
f075bd95 837 gcc_assert (known_size_p (size));
303ec2aa 838
76fe54f0 839 align = get_stack_local_alignment (type, mode);
d16790f2
JW
840
841 /* Try to find an available, already-allocated temporary of the proper
842 mode which meets the size and alignment requirements. Choose the
3e8b0446 843 smallest one with the closest alignment.
b8698a0f 844
3e8b0446
ZD
845 If assign_stack_temp is called outside of the tree->rtl expansion,
846 we cannot reuse the stack slots (that may still refer to
847 VIRTUAL_STACK_VARS_REGNUM). */
848 if (!virtuals_instantiated)
0aea6467 849 {
3e8b0446 850 for (p = avail_temp_slots; p; p = p->next)
0aea6467 851 {
f075bd95
RS
852 if (p->align >= align
853 && known_ge (p->size, size)
3e8b0446
ZD
854 && GET_MODE (p->slot) == mode
855 && objects_must_conflict_p (p->type, type)
f075bd95
RS
856 && (best_p == 0
857 || (known_eq (best_p->size, p->size)
858 ? best_p->align > p->align
859 : known_ge (best_p->size, p->size))))
0aea6467 860 {
f075bd95 861 if (p->align == align && known_eq (p->size, size))
3e8b0446
ZD
862 {
863 selected = p;
864 cut_slot_from_list (selected, &avail_temp_slots);
865 best_p = 0;
866 break;
867 }
868 best_p = p;
0aea6467 869 }
0aea6467
ZD
870 }
871 }
6f086dfc
RS
872
873 /* Make our best, if any, the one to use. */
874 if (best_p)
a45035b6 875 {
0aea6467
ZD
876 selected = best_p;
877 cut_slot_from_list (selected, &avail_temp_slots);
878
a45035b6
JW
879 /* If there are enough aligned bytes left over, make them into a new
880 temp_slot so that the extra bytes don't get wasted. Do this only
881 for BLKmode slots, so that we can be sure of the alignment. */
3bdf5ad1 882 if (GET_MODE (best_p->slot) == BLKmode)
a45035b6 883 {
d16790f2 884 int alignment = best_p->align / BITS_PER_UNIT;
f075bd95 885 poly_int64 rounded_size = aligned_upper_bound (size, alignment);
a45035b6 886
f075bd95 887 if (known_ge (best_p->size - rounded_size, alignment))
a45035b6 888 {
766090c2 889 p = ggc_alloc<temp_slot> ();
9474e8ab 890 p->in_use = 0;
a45035b6 891 p->size = best_p->size - rounded_size;
307d8cd6
RK
892 p->base_offset = best_p->base_offset + rounded_size;
893 p->full_size = best_p->full_size - rounded_size;
be0c514c 894 p->slot = adjust_address_nv (best_p->slot, BLKmode, rounded_size);
d16790f2 895 p->align = best_p->align;
1da68f56 896 p->type = best_p->type;
0aea6467 897 insert_slot_to_list (p, &avail_temp_slots);
a45035b6 898
8c39f8ae 899 vec_safe_push (stack_slot_list, p->slot);
a45035b6
JW
900
901 best_p->size = rounded_size;
291dde90 902 best_p->full_size = rounded_size;
a45035b6
JW
903 }
904 }
a45035b6 905 }
718fe406 906
6f086dfc 907 /* If we still didn't find one, make a new temporary. */
0aea6467 908 if (selected == 0)
6f086dfc 909 {
f075bd95 910 poly_int64 frame_offset_old = frame_offset;
e5e809f4 911
766090c2 912 p = ggc_alloc<temp_slot> ();
e5e809f4 913
c87a0a39
JL
914 /* We are passing an explicit alignment request to assign_stack_local.
915 One side effect of that is assign_stack_local will not round SIZE
916 to ensure the frame offset remains suitably aligned.
917
918 So for requests which depended on the rounding of SIZE, we go ahead
919 and round it now. We also make sure ALIGNMENT is at least
920 BIGGEST_ALIGNMENT. */
0bccc606 921 gcc_assert (mode != BLKmode || align == BIGGEST_ALIGNMENT);
80a832cd
JJ
922 p->slot = assign_stack_local_1 (mode,
923 (mode == BLKmode
f075bd95
RS
924 ? aligned_upper_bound (size,
925 (int) align
926 / BITS_PER_UNIT)
80a832cd
JJ
927 : size),
928 align, 0);
d16790f2
JW
929
930 p->align = align;
e5e809f4 931
b2a80c0d
DE
932 /* The following slot size computation is necessary because we don't
933 know the actual size of the temporary slot until assign_stack_local
934 has performed all the frame alignment and size rounding for the
fc91b0d0
RK
935 requested temporary. Note that extra space added for alignment
936 can be either above or below this stack slot depending on which
937 way the frame grows. We include the extra space if and only if it
938 is above this slot. */
f62c8a5c
JJ
939 if (FRAME_GROWS_DOWNWARD)
940 p->size = frame_offset_old - frame_offset;
941 else
942 p->size = size;
e5e809f4 943
fc91b0d0 944 /* Now define the fields used by combine_temp_slots. */
f62c8a5c
JJ
945 if (FRAME_GROWS_DOWNWARD)
946 {
947 p->base_offset = frame_offset;
948 p->full_size = frame_offset_old - frame_offset;
949 }
950 else
951 {
952 p->base_offset = frame_offset_old;
953 p->full_size = frame_offset - frame_offset_old;
954 }
0aea6467
ZD
955
956 selected = p;
6f086dfc
RS
957 }
958
0aea6467 959 p = selected;
6f086dfc 960 p->in_use = 1;
1da68f56 961 p->type = type;
7efcb746 962 p->level = temp_slot_level;
f8395d62 963 n_temp_slots_in_use++;
1995f267 964
0aea6467
ZD
965 pp = temp_slots_at_level (p->level);
966 insert_slot_to_list (p, pp);
fb0703f7 967 insert_temp_slot_address (XEXP (p->slot, 0), p);
faa964e5
UW
968
969 /* Create a new MEM rtx to avoid clobbering MEM flags of old slots. */
970 slot = gen_rtx_MEM (mode, XEXP (p->slot, 0));
8c39f8ae 971 vec_safe_push (stack_slot_list, slot);
3bdf5ad1 972
1da68f56
RK
973 /* If we know the alias set for the memory that will be used, use
974 it. If there's no TYPE, then we don't know anything about the
975 alias set for the memory. */
faa964e5
UW
976 set_mem_alias_set (slot, type ? get_alias_set (type) : 0);
977 set_mem_align (slot, align);
1da68f56 978
30f7a378 979 /* If a type is specified, set the relevant flags. */
3bdf5ad1 980 if (type != 0)
55356334 981 MEM_VOLATILE_P (slot) = TYPE_VOLATILE (type);
be0c514c 982 MEM_NOTRAP_P (slot) = 1;
3bdf5ad1 983
faa964e5 984 return slot;
6f086dfc 985}
d16790f2
JW
986
987/* Allocate a temporary stack slot and record it for possible later
9474e8ab 988 reuse. First two arguments are same as in preceding function. */
d16790f2
JW
989
990rtx
f075bd95 991assign_stack_temp (machine_mode mode, poly_int64 size)
d16790f2 992{
9474e8ab 993 return assign_stack_temp_for_type (mode, size, NULL_TREE);
d16790f2 994}
638141a6 995\f
9432c136
EB
996/* Assign a temporary.
997 If TYPE_OR_DECL is a decl, then we are doing it on behalf of the decl
998 and so that should be used in error messages. In either case, we
999 allocate of the given type.
230f21b4 1000 MEMORY_REQUIRED is 1 if the result must be addressable stack memory;
b55d9ff8
RK
1001 it is 0 if a register is OK.
1002 DONT_PROMOTE is 1 if we should not promote values in register
1003 to wider modes. */
230f21b4
PB
1004
1005rtx
9474e8ab 1006assign_temp (tree type_or_decl, int memory_required,
fa8db1f7 1007 int dont_promote ATTRIBUTE_UNUSED)
230f21b4 1008{
9432c136 1009 tree type, decl;
ef4bddc2 1010 machine_mode mode;
9e1622ed 1011#ifdef PROMOTE_MODE
9432c136
EB
1012 int unsignedp;
1013#endif
1014
1015 if (DECL_P (type_or_decl))
1016 decl = type_or_decl, type = TREE_TYPE (decl);
1017 else
1018 decl = NULL, type = type_or_decl;
1019
1020 mode = TYPE_MODE (type);
9e1622ed 1021#ifdef PROMOTE_MODE
8df83eae 1022 unsignedp = TYPE_UNSIGNED (type);
0ce8a59c 1023#endif
638141a6 1024
45177337
JM
1025 /* Allocating temporaries of TREE_ADDRESSABLE type must be done in the front
1026 end. See also create_tmp_var for the gimplification-time check. */
1027 gcc_assert (!TREE_ADDRESSABLE (type) && COMPLETE_TYPE_P (type));
1028
230f21b4
PB
1029 if (mode == BLKmode || memory_required)
1030 {
6e246559 1031 poly_int64 size;
230f21b4
PB
1032 rtx tmp;
1033
1034 /* Unfortunately, we don't yet know how to allocate variable-sized
a441447f
OH
1035 temporaries. However, sometimes we can find a fixed upper limit on
1036 the size, so try that instead. */
6e246559 1037 if (!poly_int_tree_p (TYPE_SIZE_UNIT (type), &size))
a441447f 1038 size = max_int_size_in_bytes (type);
e30bb772 1039
6e246559
RS
1040 /* Zero sized arrays are a GNU C extension. Set size to 1 to avoid
1041 problems with allocating the stack space. */
1042 if (known_eq (size, 0))
1043 size = 1;
1044
9432c136
EB
1045 /* The size of the temporary may be too large to fit into an integer. */
1046 /* ??? Not sure this should happen except for user silliness, so limit
797a6ac1 1047 this to things that aren't compiler-generated temporaries. The
535a42b1 1048 rest of the time we'll die in assign_stack_temp_for_type. */
6e246559
RS
1049 if (decl
1050 && !known_size_p (size)
9432c136
EB
1051 && TREE_CODE (TYPE_SIZE_UNIT (type)) == INTEGER_CST)
1052 {
dee15844 1053 error ("size of variable %q+D is too large", decl);
9432c136
EB
1054 size = 1;
1055 }
1056
9474e8ab 1057 tmp = assign_stack_temp_for_type (mode, size, type);
230f21b4
PB
1058 return tmp;
1059 }
638141a6 1060
9e1622ed 1061#ifdef PROMOTE_MODE
b55d9ff8 1062 if (! dont_promote)
cde0f3fd 1063 mode = promote_mode (type, mode, &unsignedp);
230f21b4 1064#endif
638141a6 1065
230f21b4
PB
1066 return gen_reg_rtx (mode);
1067}
638141a6 1068\f
a45035b6
JW
1069/* Combine temporary stack slots which are adjacent on the stack.
1070
1071 This allows for better use of already allocated stack space. This is only
1072 done for BLKmode slots because we can be sure that we won't have alignment
1073 problems in this case. */
1074
6fe79279 1075static void
fa8db1f7 1076combine_temp_slots (void)
a45035b6 1077{
0aea6467 1078 struct temp_slot *p, *q, *next, *next_q;
e5e809f4
JL
1079 int num_slots;
1080
a4c6502a
MM
1081 /* We can't combine slots, because the information about which slot
1082 is in which alias set will be lost. */
1083 if (flag_strict_aliasing)
1084 return;
1085
718fe406 1086 /* If there are a lot of temp slots, don't do anything unless
d6a7951f 1087 high levels of optimization. */
e5e809f4 1088 if (! flag_expensive_optimizations)
0aea6467 1089 for (p = avail_temp_slots, num_slots = 0; p; p = p->next, num_slots++)
e5e809f4
JL
1090 if (num_slots > 100 || (num_slots > 10 && optimize == 0))
1091 return;
a45035b6 1092
0aea6467 1093 for (p = avail_temp_slots; p; p = next)
e9b7093a
RS
1094 {
1095 int delete_p = 0;
e5e809f4 1096
0aea6467
ZD
1097 next = p->next;
1098
1099 if (GET_MODE (p->slot) != BLKmode)
1100 continue;
1101
1102 for (q = p->next; q; q = next_q)
e9b7093a 1103 {
0aea6467
ZD
1104 int delete_q = 0;
1105
1106 next_q = q->next;
1107
1108 if (GET_MODE (q->slot) != BLKmode)
1109 continue;
1110
f075bd95 1111 if (known_eq (p->base_offset + p->full_size, q->base_offset))
0aea6467
ZD
1112 {
1113 /* Q comes after P; combine Q into P. */
1114 p->size += q->size;
1115 p->full_size += q->full_size;
1116 delete_q = 1;
1117 }
f075bd95 1118 else if (known_eq (q->base_offset + q->full_size, p->base_offset))
0aea6467
ZD
1119 {
1120 /* P comes after Q; combine P into Q. */
1121 q->size += p->size;
1122 q->full_size += p->full_size;
1123 delete_p = 1;
1124 break;
1125 }
1126 if (delete_q)
1127 cut_slot_from_list (q, &avail_temp_slots);
e9b7093a 1128 }
0aea6467
ZD
1129
1130 /* Either delete P or advance past it. */
1131 if (delete_p)
1132 cut_slot_from_list (p, &avail_temp_slots);
e9b7093a 1133 }
a45035b6 1134}
6f086dfc 1135\f
82d6e6fc
KG
1136/* Indicate that NEW_RTX is an alternate way of referring to the temp
1137 slot that previously was known by OLD_RTX. */
e5e76139
RK
1138
1139void
82d6e6fc 1140update_temp_slot_address (rtx old_rtx, rtx new_rtx)
e5e76139 1141{
14a774a9 1142 struct temp_slot *p;
e5e76139 1143
82d6e6fc 1144 if (rtx_equal_p (old_rtx, new_rtx))
e5e76139 1145 return;
14a774a9 1146
82d6e6fc 1147 p = find_temp_slot_from_address (old_rtx);
14a774a9 1148
82d6e6fc
KG
1149 /* If we didn't find one, see if both OLD_RTX is a PLUS. If so, and
1150 NEW_RTX is a register, see if one operand of the PLUS is a
1151 temporary location. If so, NEW_RTX points into it. Otherwise,
1152 if both OLD_RTX and NEW_RTX are a PLUS and if there is a register
1153 in common between them. If so, try a recursive call on those
1154 values. */
14a774a9
RK
1155 if (p == 0)
1156 {
82d6e6fc 1157 if (GET_CODE (old_rtx) != PLUS)
700f19f0
RK
1158 return;
1159
82d6e6fc 1160 if (REG_P (new_rtx))
700f19f0 1161 {
82d6e6fc
KG
1162 update_temp_slot_address (XEXP (old_rtx, 0), new_rtx);
1163 update_temp_slot_address (XEXP (old_rtx, 1), new_rtx);
700f19f0
RK
1164 return;
1165 }
82d6e6fc 1166 else if (GET_CODE (new_rtx) != PLUS)
14a774a9
RK
1167 return;
1168
82d6e6fc
KG
1169 if (rtx_equal_p (XEXP (old_rtx, 0), XEXP (new_rtx, 0)))
1170 update_temp_slot_address (XEXP (old_rtx, 1), XEXP (new_rtx, 1));
1171 else if (rtx_equal_p (XEXP (old_rtx, 1), XEXP (new_rtx, 0)))
1172 update_temp_slot_address (XEXP (old_rtx, 0), XEXP (new_rtx, 1));
1173 else if (rtx_equal_p (XEXP (old_rtx, 0), XEXP (new_rtx, 1)))
1174 update_temp_slot_address (XEXP (old_rtx, 1), XEXP (new_rtx, 0));
1175 else if (rtx_equal_p (XEXP (old_rtx, 1), XEXP (new_rtx, 1)))
1176 update_temp_slot_address (XEXP (old_rtx, 0), XEXP (new_rtx, 0));
14a774a9
RK
1177
1178 return;
1179 }
1180
718fe406 1181 /* Otherwise add an alias for the temp's address. */
fb0703f7 1182 insert_temp_slot_address (new_rtx, p);
e5e76139
RK
1183}
1184
9cca6a99
MS
1185/* If X could be a reference to a temporary slot, mark that slot as
1186 belonging to the to one level higher than the current level. If X
1187 matched one of our slots, just mark that one. Otherwise, we can't
9474e8ab 1188 easily predict which it is, so upgrade all of them.
6f086dfc
RS
1189
1190 This is called when an ({...}) construct occurs and a statement
1191 returns a value in memory. */
1192
1193void
fa8db1f7 1194preserve_temp_slots (rtx x)
6f086dfc 1195{
0aea6467 1196 struct temp_slot *p = 0, *next;
6f086dfc 1197
e3a77161 1198 if (x == 0)
9474e8ab 1199 return;
f7b6d104 1200
8fff4fc1 1201 /* If X is a register that is being used as a pointer, see if we have
9474e8ab 1202 a temporary slot we know it points to. */
8fff4fc1
RH
1203 if (REG_P (x) && REG_POINTER (x))
1204 p = find_temp_slot_from_address (x);
f7b6d104 1205
8fff4fc1 1206 /* If X is not in memory or is at a constant address, it cannot be in
9474e8ab 1207 a temporary slot. */
8fff4fc1 1208 if (p == 0 && (!MEM_P (x) || CONSTANT_P (XEXP (x, 0))))
9474e8ab 1209 return;
8fff4fc1
RH
1210
1211 /* First see if we can find a match. */
1212 if (p == 0)
1213 p = find_temp_slot_from_address (XEXP (x, 0));
1214
1215 if (p != 0)
1216 {
8fff4fc1 1217 if (p->level == temp_slot_level)
9474e8ab 1218 move_slot_to_level (p, temp_slot_level - 1);
8fff4fc1 1219 return;
f7b6d104 1220 }
e9a25f70 1221
8fff4fc1
RH
1222 /* Otherwise, preserve all non-kept slots at this level. */
1223 for (p = *temp_slots_at_level (temp_slot_level); p; p = next)
e9a25f70 1224 {
8fff4fc1 1225 next = p->next;
9474e8ab 1226 move_slot_to_level (p, temp_slot_level - 1);
8fff4fc1 1227 }
fe9b4957
MM
1228}
1229
8fff4fc1
RH
1230/* Free all temporaries used so far. This is normally called at the
1231 end of generating code for a statement. */
fe9b4957 1232
8fff4fc1
RH
1233void
1234free_temp_slots (void)
fe9b4957 1235{
8fff4fc1 1236 struct temp_slot *p, *next;
5d7cefe5 1237 bool some_available = false;
fe9b4957 1238
8fff4fc1
RH
1239 for (p = *temp_slots_at_level (temp_slot_level); p; p = next)
1240 {
1241 next = p->next;
9474e8ab
MM
1242 make_slot_available (p);
1243 some_available = true;
8fff4fc1 1244 }
fe9b4957 1245
5d7cefe5
MM
1246 if (some_available)
1247 {
1248 remove_unused_temp_slot_addresses ();
1249 combine_temp_slots ();
1250 }
8fff4fc1 1251}
fe9b4957 1252
8fff4fc1 1253/* Push deeper into the nesting level for stack temporaries. */
fe9b4957 1254
8fff4fc1
RH
1255void
1256push_temp_slots (void)
fe9b4957 1257{
8fff4fc1 1258 temp_slot_level++;
fe9b4957
MM
1259}
1260
8fff4fc1
RH
1261/* Pop a temporary nesting level. All slots in use in the current level
1262 are freed. */
fe9b4957 1263
8fff4fc1
RH
1264void
1265pop_temp_slots (void)
fe9b4957 1266{
9474e8ab 1267 free_temp_slots ();
8fff4fc1 1268 temp_slot_level--;
8c36698e
NC
1269}
1270
8fff4fc1 1271/* Initialize temporary slots. */
e9a25f70
JL
1272
1273void
8fff4fc1 1274init_temp_slots (void)
e9a25f70 1275{
8fff4fc1
RH
1276 /* We have not allocated any temporaries yet. */
1277 avail_temp_slots = 0;
9771b263 1278 vec_alloc (used_temp_slots, 0);
8fff4fc1 1279 temp_slot_level = 0;
f8395d62 1280 n_temp_slots_in_use = 0;
fb0703f7
SB
1281
1282 /* Set up the table to map addresses to temp slots. */
1283 if (! temp_slot_address_table)
2a22f99c 1284 temp_slot_address_table = hash_table<temp_address_hasher>::create_ggc (32);
fb0703f7 1285 else
2a22f99c 1286 temp_slot_address_table->empty ();
8fff4fc1
RH
1287}
1288\f
6399c0ab
SB
1289/* Functions and data structures to keep track of the values hard regs
1290 had at the start of the function. */
1291
1292/* Private type used by get_hard_reg_initial_reg, get_hard_reg_initial_val,
1293 and has_hard_reg_initial_val.. */
50686850 1294struct GTY(()) initial_value_pair {
6399c0ab
SB
1295 rtx hard_reg;
1296 rtx pseudo;
50686850 1297};
6399c0ab
SB
1298/* ??? This could be a VEC but there is currently no way to define an
1299 opaque VEC type. This could be worked around by defining struct
1300 initial_value_pair in function.h. */
50686850 1301struct GTY(()) initial_value_struct {
6399c0ab
SB
1302 int num_entries;
1303 int max_entries;
1304 initial_value_pair * GTY ((length ("%h.num_entries"))) entries;
50686850 1305};
6399c0ab
SB
1306
1307/* If a pseudo represents an initial hard reg (or expression), return
1308 it, else return NULL_RTX. */
1309
1310rtx
1311get_hard_reg_initial_reg (rtx reg)
1312{
1313 struct initial_value_struct *ivs = crtl->hard_reg_initial_vals;
1314 int i;
1315
1316 if (ivs == 0)
1317 return NULL_RTX;
1318
1319 for (i = 0; i < ivs->num_entries; i++)
1320 if (rtx_equal_p (ivs->entries[i].pseudo, reg))
1321 return ivs->entries[i].hard_reg;
1322
1323 return NULL_RTX;
1324}
1325
1326/* Make sure that there's a pseudo register of mode MODE that stores the
1327 initial value of hard register REGNO. Return an rtx for such a pseudo. */
1328
1329rtx
ef4bddc2 1330get_hard_reg_initial_val (machine_mode mode, unsigned int regno)
6399c0ab
SB
1331{
1332 struct initial_value_struct *ivs;
1333 rtx rv;
1334
1335 rv = has_hard_reg_initial_val (mode, regno);
1336 if (rv)
1337 return rv;
1338
1339 ivs = crtl->hard_reg_initial_vals;
1340 if (ivs == 0)
1341 {
766090c2 1342 ivs = ggc_alloc<initial_value_struct> ();
6399c0ab
SB
1343 ivs->num_entries = 0;
1344 ivs->max_entries = 5;
766090c2 1345 ivs->entries = ggc_vec_alloc<initial_value_pair> (5);
6399c0ab
SB
1346 crtl->hard_reg_initial_vals = ivs;
1347 }
1348
1349 if (ivs->num_entries >= ivs->max_entries)
1350 {
1351 ivs->max_entries += 5;
1352 ivs->entries = GGC_RESIZEVEC (initial_value_pair, ivs->entries,
1353 ivs->max_entries);
1354 }
1355
1356 ivs->entries[ivs->num_entries].hard_reg = gen_rtx_REG (mode, regno);
1357 ivs->entries[ivs->num_entries].pseudo = gen_reg_rtx (mode);
1358
1359 return ivs->entries[ivs->num_entries++].pseudo;
1360}
1361
1362/* See if get_hard_reg_initial_val has been used to create a pseudo
1363 for the initial value of hard register REGNO in mode MODE. Return
1364 the associated pseudo if so, otherwise return NULL. */
1365
1366rtx
ef4bddc2 1367has_hard_reg_initial_val (machine_mode mode, unsigned int regno)
6399c0ab
SB
1368{
1369 struct initial_value_struct *ivs;
1370 int i;
1371
1372 ivs = crtl->hard_reg_initial_vals;
1373 if (ivs != 0)
1374 for (i = 0; i < ivs->num_entries; i++)
1375 if (GET_MODE (ivs->entries[i].hard_reg) == mode
1376 && REGNO (ivs->entries[i].hard_reg) == regno)
1377 return ivs->entries[i].pseudo;
1378
1379 return NULL_RTX;
1380}
1381
1382unsigned int
1383emit_initial_value_sets (void)
1384{
1385 struct initial_value_struct *ivs = crtl->hard_reg_initial_vals;
1386 int i;
691fe203 1387 rtx_insn *seq;
6399c0ab
SB
1388
1389 if (ivs == 0)
1390 return 0;
1391
1392 start_sequence ();
1393 for (i = 0; i < ivs->num_entries; i++)
1394 emit_move_insn (ivs->entries[i].pseudo, ivs->entries[i].hard_reg);
1395 seq = get_insns ();
1396 end_sequence ();
1397
1398 emit_insn_at_entry (seq);
1399 return 0;
1400}
1401
1402/* Return the hardreg-pseudoreg initial values pair entry I and
1403 TRUE if I is a valid entry, or FALSE if I is not a valid entry. */
1404bool
1405initial_value_entry (int i, rtx *hreg, rtx *preg)
1406{
1407 struct initial_value_struct *ivs = crtl->hard_reg_initial_vals;
1408 if (!ivs || i >= ivs->num_entries)
1409 return false;
1410
1411 *hreg = ivs->entries[i].hard_reg;
1412 *preg = ivs->entries[i].pseudo;
1413 return true;
1414}
1415\f
8fff4fc1
RH
1416/* These routines are responsible for converting virtual register references
1417 to the actual hard register references once RTL generation is complete.
718fe406 1418
8fff4fc1
RH
1419 The following four variables are used for communication between the
1420 routines. They contain the offsets of the virtual registers from their
1421 respective hard registers. */
fe9b4957 1422
e6715081
RS
1423static poly_int64 in_arg_offset;
1424static poly_int64 var_offset;
1425static poly_int64 dynamic_offset;
1426static poly_int64 out_arg_offset;
1427static poly_int64 cfa_offset;
8a5275eb 1428
8fff4fc1
RH
1429/* In most machines, the stack pointer register is equivalent to the bottom
1430 of the stack. */
718fe406 1431
8fff4fc1
RH
1432#ifndef STACK_POINTER_OFFSET
1433#define STACK_POINTER_OFFSET 0
1434#endif
8c36698e 1435
ddbb449f
AM
1436#if defined (REG_PARM_STACK_SPACE) && !defined (INCOMING_REG_PARM_STACK_SPACE)
1437#define INCOMING_REG_PARM_STACK_SPACE REG_PARM_STACK_SPACE
1438#endif
1439
8fff4fc1
RH
1440/* If not defined, pick an appropriate default for the offset of dynamically
1441 allocated memory depending on the value of ACCUMULATE_OUTGOING_ARGS,
ddbb449f 1442 INCOMING_REG_PARM_STACK_SPACE, and OUTGOING_REG_PARM_STACK_SPACE. */
fe9b4957 1443
8fff4fc1 1444#ifndef STACK_DYNAMIC_OFFSET
8a5275eb 1445
8fff4fc1
RH
1446/* The bottom of the stack points to the actual arguments. If
1447 REG_PARM_STACK_SPACE is defined, this includes the space for the register
1448 parameters. However, if OUTGOING_REG_PARM_STACK space is not defined,
1449 stack space for register parameters is not pushed by the caller, but
1450 rather part of the fixed stack areas and hence not included in
38173d38 1451 `crtl->outgoing_args_size'. Nevertheless, we must allow
8fff4fc1 1452 for it when allocating stack dynamic objects. */
8a5275eb 1453
ddbb449f 1454#ifdef INCOMING_REG_PARM_STACK_SPACE
8fff4fc1
RH
1455#define STACK_DYNAMIC_OFFSET(FNDECL) \
1456((ACCUMULATE_OUTGOING_ARGS \
38173d38 1457 ? (crtl->outgoing_args_size \
81464b2c 1458 + (OUTGOING_REG_PARM_STACK_SPACE ((!(FNDECL) ? NULL_TREE : TREE_TYPE (FNDECL))) ? 0 \
ddbb449f 1459 : INCOMING_REG_PARM_STACK_SPACE (FNDECL))) \
ac294f0b 1460 : 0) + (STACK_POINTER_OFFSET))
8fff4fc1
RH
1461#else
1462#define STACK_DYNAMIC_OFFSET(FNDECL) \
a20c5714 1463 ((ACCUMULATE_OUTGOING_ARGS ? crtl->outgoing_args_size : poly_int64 (0)) \
8fff4fc1
RH
1464 + (STACK_POINTER_OFFSET))
1465#endif
1466#endif
4fa48eae 1467
659e47fb 1468\f
bbf9b913
RH
1469/* Given a piece of RTX and a pointer to a HOST_WIDE_INT, if the RTX
1470 is a virtual register, return the equivalent hard register and set the
1471 offset indirectly through the pointer. Otherwise, return 0. */
6f086dfc 1472
bbf9b913 1473static rtx
e6715081 1474instantiate_new_reg (rtx x, poly_int64_pod *poffset)
6f086dfc 1475{
82d6e6fc 1476 rtx new_rtx;
e6715081 1477 poly_int64 offset;
6f086dfc 1478
bbf9b913 1479 if (x == virtual_incoming_args_rtx)
2e3f842f 1480 {
d015f7cc 1481 if (stack_realign_drap)
2e3f842f 1482 {
d015f7cc
L
1483 /* Replace virtual_incoming_args_rtx with internal arg
1484 pointer if DRAP is used to realign stack. */
82d6e6fc 1485 new_rtx = crtl->args.internal_arg_pointer;
2e3f842f
L
1486 offset = 0;
1487 }
1488 else
82d6e6fc 1489 new_rtx = arg_pointer_rtx, offset = in_arg_offset;
2e3f842f 1490 }
bbf9b913 1491 else if (x == virtual_stack_vars_rtx)
82d6e6fc 1492 new_rtx = frame_pointer_rtx, offset = var_offset;
bbf9b913 1493 else if (x == virtual_stack_dynamic_rtx)
82d6e6fc 1494 new_rtx = stack_pointer_rtx, offset = dynamic_offset;
bbf9b913 1495 else if (x == virtual_outgoing_args_rtx)
82d6e6fc 1496 new_rtx = stack_pointer_rtx, offset = out_arg_offset;
bbf9b913 1497 else if (x == virtual_cfa_rtx)
f6672e8e
RH
1498 {
1499#ifdef FRAME_POINTER_CFA_OFFSET
82d6e6fc 1500 new_rtx = frame_pointer_rtx;
f6672e8e 1501#else
82d6e6fc 1502 new_rtx = arg_pointer_rtx;
f6672e8e
RH
1503#endif
1504 offset = cfa_offset;
1505 }
32990d5b
JJ
1506 else if (x == virtual_preferred_stack_boundary_rtx)
1507 {
1508 new_rtx = GEN_INT (crtl->preferred_stack_boundary / BITS_PER_UNIT);
1509 offset = 0;
1510 }
bbf9b913
RH
1511 else
1512 return NULL_RTX;
6f086dfc 1513
bbf9b913 1514 *poffset = offset;
82d6e6fc 1515 return new_rtx;
6f086dfc
RS
1516}
1517
b8704801
RS
1518/* A subroutine of instantiate_virtual_regs. Instantiate any virtual
1519 registers present inside of *LOC. The expression is simplified,
1520 as much as possible, but is not to be considered "valid" in any sense
1521 implied by the target. Return true if any change is made. */
6f086dfc 1522
b8704801
RS
1523static bool
1524instantiate_virtual_regs_in_rtx (rtx *loc)
6f086dfc 1525{
b8704801
RS
1526 if (!*loc)
1527 return false;
1528 bool changed = false;
1529 subrtx_ptr_iterator::array_type array;
1530 FOR_EACH_SUBRTX_PTR (iter, array, loc, NONCONST)
6f086dfc 1531 {
b8704801
RS
1532 rtx *loc = *iter;
1533 if (rtx x = *loc)
bbf9b913 1534 {
b8704801 1535 rtx new_rtx;
e6715081 1536 poly_int64 offset;
b8704801
RS
1537 switch (GET_CODE (x))
1538 {
1539 case REG:
1540 new_rtx = instantiate_new_reg (x, &offset);
1541 if (new_rtx)
1542 {
1543 *loc = plus_constant (GET_MODE (x), new_rtx, offset);
1544 changed = true;
1545 }
1546 iter.skip_subrtxes ();
1547 break;
bbf9b913 1548
b8704801
RS
1549 case PLUS:
1550 new_rtx = instantiate_new_reg (XEXP (x, 0), &offset);
1551 if (new_rtx)
1552 {
1553 XEXP (x, 0) = new_rtx;
1554 *loc = plus_constant (GET_MODE (x), x, offset, true);
1555 changed = true;
1556 iter.skip_subrtxes ();
1557 break;
1558 }
e5e809f4 1559
b8704801
RS
1560 /* FIXME -- from old code */
1561 /* If we have (plus (subreg (virtual-reg)) (const_int)), we know
1562 we can commute the PLUS and SUBREG because pointers into the
1563 frame are well-behaved. */
1564 break;
ce717ce4 1565
b8704801
RS
1566 default:
1567 break;
1568 }
1569 }
6f086dfc 1570 }
b8704801 1571 return changed;
6f086dfc
RS
1572}
1573
bbf9b913
RH
1574/* A subroutine of instantiate_virtual_regs_in_insn. Return true if X
1575 matches the predicate for insn CODE operand OPERAND. */
6f086dfc 1576
bbf9b913
RH
1577static int
1578safe_insn_predicate (int code, int operand, rtx x)
6f086dfc 1579{
2ef6ce06 1580 return code < 0 || insn_operand_matches ((enum insn_code) code, operand, x);
bbf9b913 1581}
5a73491b 1582
bbf9b913
RH
1583/* A subroutine of instantiate_virtual_regs. Instantiate any virtual
1584 registers present inside of insn. The result will be a valid insn. */
5a73491b
RK
1585
1586static void
691fe203 1587instantiate_virtual_regs_in_insn (rtx_insn *insn)
5a73491b 1588{
e6715081 1589 poly_int64 offset;
bbf9b913 1590 int insn_code, i;
9325973e 1591 bool any_change = false;
691fe203
DM
1592 rtx set, new_rtx, x;
1593 rtx_insn *seq;
32e66afd 1594
bbf9b913
RH
1595 /* There are some special cases to be handled first. */
1596 set = single_set (insn);
1597 if (set)
32e66afd 1598 {
bbf9b913
RH
1599 /* We're allowed to assign to a virtual register. This is interpreted
1600 to mean that the underlying register gets assigned the inverse
1601 transformation. This is used, for example, in the handling of
1602 non-local gotos. */
82d6e6fc
KG
1603 new_rtx = instantiate_new_reg (SET_DEST (set), &offset);
1604 if (new_rtx)
bbf9b913
RH
1605 {
1606 start_sequence ();
32e66afd 1607
b8704801 1608 instantiate_virtual_regs_in_rtx (&SET_SRC (set));
82d6e6fc 1609 x = simplify_gen_binary (PLUS, GET_MODE (new_rtx), SET_SRC (set),
69a59f0f 1610 gen_int_mode (-offset, GET_MODE (new_rtx)));
82d6e6fc
KG
1611 x = force_operand (x, new_rtx);
1612 if (x != new_rtx)
1613 emit_move_insn (new_rtx, x);
5a73491b 1614
bbf9b913
RH
1615 seq = get_insns ();
1616 end_sequence ();
5a73491b 1617
bbf9b913
RH
1618 emit_insn_before (seq, insn);
1619 delete_insn (insn);
1620 return;
1621 }
5a73491b 1622
bbf9b913
RH
1623 /* Handle a straight copy from a virtual register by generating a
1624 new add insn. The difference between this and falling through
1625 to the generic case is avoiding a new pseudo and eliminating a
1626 move insn in the initial rtl stream. */
82d6e6fc 1627 new_rtx = instantiate_new_reg (SET_SRC (set), &offset);
e6715081
RS
1628 if (new_rtx
1629 && maybe_ne (offset, 0)
bbf9b913
RH
1630 && REG_P (SET_DEST (set))
1631 && REGNO (SET_DEST (set)) > LAST_VIRTUAL_REGISTER)
1632 {
1633 start_sequence ();
5a73491b 1634
2f1cd2eb
RS
1635 x = expand_simple_binop (GET_MODE (SET_DEST (set)), PLUS, new_rtx,
1636 gen_int_mode (offset,
1637 GET_MODE (SET_DEST (set))),
1638 SET_DEST (set), 1, OPTAB_LIB_WIDEN);
bbf9b913
RH
1639 if (x != SET_DEST (set))
1640 emit_move_insn (SET_DEST (set), x);
770ae6cc 1641
bbf9b913
RH
1642 seq = get_insns ();
1643 end_sequence ();
87ce34d6 1644
bbf9b913
RH
1645 emit_insn_before (seq, insn);
1646 delete_insn (insn);
87ce34d6 1647 return;
bbf9b913 1648 }
5a73491b 1649
bbf9b913 1650 extract_insn (insn);
9325973e 1651 insn_code = INSN_CODE (insn);
5a73491b 1652
bbf9b913
RH
1653 /* Handle a plus involving a virtual register by determining if the
1654 operands remain valid if they're modified in place. */
e6715081 1655 poly_int64 delta;
bbf9b913
RH
1656 if (GET_CODE (SET_SRC (set)) == PLUS
1657 && recog_data.n_operands >= 3
1658 && recog_data.operand_loc[1] == &XEXP (SET_SRC (set), 0)
1659 && recog_data.operand_loc[2] == &XEXP (SET_SRC (set), 1)
e6715081 1660 && poly_int_rtx_p (recog_data.operand[2], &delta)
82d6e6fc 1661 && (new_rtx = instantiate_new_reg (recog_data.operand[1], &offset)))
bbf9b913 1662 {
e6715081 1663 offset += delta;
5a73491b 1664
bbf9b913 1665 /* If the sum is zero, then replace with a plain move. */
e6715081 1666 if (known_eq (offset, 0)
9325973e
RH
1667 && REG_P (SET_DEST (set))
1668 && REGNO (SET_DEST (set)) > LAST_VIRTUAL_REGISTER)
bbf9b913
RH
1669 {
1670 start_sequence ();
82d6e6fc 1671 emit_move_insn (SET_DEST (set), new_rtx);
bbf9b913
RH
1672 seq = get_insns ();
1673 end_sequence ();
d1405722 1674
bbf9b913
RH
1675 emit_insn_before (seq, insn);
1676 delete_insn (insn);
1677 return;
1678 }
d1405722 1679
bbf9b913 1680 x = gen_int_mode (offset, recog_data.operand_mode[2]);
bbf9b913
RH
1681
1682 /* Using validate_change and apply_change_group here leaves
1683 recog_data in an invalid state. Since we know exactly what
1684 we want to check, do those two by hand. */
82d6e6fc 1685 if (safe_insn_predicate (insn_code, 1, new_rtx)
bbf9b913
RH
1686 && safe_insn_predicate (insn_code, 2, x))
1687 {
82d6e6fc 1688 *recog_data.operand_loc[1] = recog_data.operand[1] = new_rtx;
bbf9b913
RH
1689 *recog_data.operand_loc[2] = recog_data.operand[2] = x;
1690 any_change = true;
9325973e
RH
1691
1692 /* Fall through into the regular operand fixup loop in
1693 order to take care of operands other than 1 and 2. */
bbf9b913
RH
1694 }
1695 }
1696 }
d1405722 1697 else
9325973e
RH
1698 {
1699 extract_insn (insn);
1700 insn_code = INSN_CODE (insn);
1701 }
5dc96d60 1702
bbf9b913
RH
1703 /* In the general case, we expect virtual registers to appear only in
1704 operands, and then only as either bare registers or inside memories. */
1705 for (i = 0; i < recog_data.n_operands; ++i)
1706 {
1707 x = recog_data.operand[i];
1708 switch (GET_CODE (x))
1709 {
1710 case MEM:
1711 {
1712 rtx addr = XEXP (x, 0);
bbf9b913 1713
b8704801 1714 if (!instantiate_virtual_regs_in_rtx (&addr))
bbf9b913
RH
1715 continue;
1716
1717 start_sequence ();
23b33725 1718 x = replace_equiv_address (x, addr, true);
a5bfb13a
MM
1719 /* It may happen that the address with the virtual reg
1720 was valid (e.g. based on the virtual stack reg, which might
1721 be acceptable to the predicates with all offsets), whereas
1722 the address now isn't anymore, for instance when the address
1723 is still offsetted, but the base reg isn't virtual-stack-reg
1724 anymore. Below we would do a force_reg on the whole operand,
1725 but this insn might actually only accept memory. Hence,
1726 before doing that last resort, try to reload the address into
1727 a register, so this operand stays a MEM. */
1728 if (!safe_insn_predicate (insn_code, i, x))
1729 {
1730 addr = force_reg (GET_MODE (addr), addr);
23b33725 1731 x = replace_equiv_address (x, addr, true);
a5bfb13a 1732 }
bbf9b913
RH
1733 seq = get_insns ();
1734 end_sequence ();
1735 if (seq)
1736 emit_insn_before (seq, insn);
1737 }
1738 break;
1739
1740 case REG:
82d6e6fc
KG
1741 new_rtx = instantiate_new_reg (x, &offset);
1742 if (new_rtx == NULL)
bbf9b913 1743 continue;
e6715081 1744 if (known_eq (offset, 0))
82d6e6fc 1745 x = new_rtx;
bbf9b913
RH
1746 else
1747 {
1748 start_sequence ();
6f086dfc 1749
bbf9b913
RH
1750 /* Careful, special mode predicates may have stuff in
1751 insn_data[insn_code].operand[i].mode that isn't useful
1752 to us for computing a new value. */
1753 /* ??? Recognize address_operand and/or "p" constraints
1754 to see if (plus new offset) is a valid before we put
1755 this through expand_simple_binop. */
82d6e6fc 1756 x = expand_simple_binop (GET_MODE (x), PLUS, new_rtx,
2f1cd2eb
RS
1757 gen_int_mode (offset, GET_MODE (x)),
1758 NULL_RTX, 1, OPTAB_LIB_WIDEN);
bbf9b913
RH
1759 seq = get_insns ();
1760 end_sequence ();
1761 emit_insn_before (seq, insn);
1762 }
1763 break;
6f086dfc 1764
bbf9b913 1765 case SUBREG:
82d6e6fc
KG
1766 new_rtx = instantiate_new_reg (SUBREG_REG (x), &offset);
1767 if (new_rtx == NULL)
bbf9b913 1768 continue;
e6715081 1769 if (maybe_ne (offset, 0))
bbf9b913
RH
1770 {
1771 start_sequence ();
2f1cd2eb
RS
1772 new_rtx = expand_simple_binop
1773 (GET_MODE (new_rtx), PLUS, new_rtx,
1774 gen_int_mode (offset, GET_MODE (new_rtx)),
1775 NULL_RTX, 1, OPTAB_LIB_WIDEN);
bbf9b913
RH
1776 seq = get_insns ();
1777 end_sequence ();
1778 emit_insn_before (seq, insn);
1779 }
82d6e6fc
KG
1780 x = simplify_gen_subreg (recog_data.operand_mode[i], new_rtx,
1781 GET_MODE (new_rtx), SUBREG_BYTE (x));
7314c7dd 1782 gcc_assert (x);
bbf9b913 1783 break;
6f086dfc 1784
bbf9b913
RH
1785 default:
1786 continue;
1787 }
6f086dfc 1788
bbf9b913
RH
1789 /* At this point, X contains the new value for the operand.
1790 Validate the new value vs the insn predicate. Note that
1791 asm insns will have insn_code -1 here. */
1792 if (!safe_insn_predicate (insn_code, i, x))
6ba1bd36
JM
1793 {
1794 start_sequence ();
f7ce0951
SE
1795 if (REG_P (x))
1796 {
1797 gcc_assert (REGNO (x) <= LAST_VIRTUAL_REGISTER);
1798 x = copy_to_reg (x);
1799 }
1800 else
1801 x = force_reg (insn_data[insn_code].operand[i].mode, x);
6ba1bd36
JM
1802 seq = get_insns ();
1803 end_sequence ();
1804 if (seq)
1805 emit_insn_before (seq, insn);
1806 }
6f086dfc 1807
bbf9b913
RH
1808 *recog_data.operand_loc[i] = recog_data.operand[i] = x;
1809 any_change = true;
1810 }
6f086dfc 1811
bbf9b913
RH
1812 if (any_change)
1813 {
1814 /* Propagate operand changes into the duplicates. */
1815 for (i = 0; i < recog_data.n_dups; ++i)
1816 *recog_data.dup_loc[i]
3e916873 1817 = copy_rtx (recog_data.operand[(unsigned)recog_data.dup_num[i]]);
5dc96d60 1818
bbf9b913
RH
1819 /* Force re-recognition of the instruction for validation. */
1820 INSN_CODE (insn) = -1;
1821 }
6f086dfc 1822
bbf9b913 1823 if (asm_noperands (PATTERN (insn)) >= 0)
6f086dfc 1824 {
bbf9b913 1825 if (!check_asm_operands (PATTERN (insn)))
6f086dfc 1826 {
bbf9b913 1827 error_for_asm (insn, "impossible constraint in %<asm%>");
5a860835
JJ
1828 /* For asm goto, instead of fixing up all the edges
1829 just clear the template and clear input operands
1830 (asm goto doesn't have any output operands). */
1831 if (JUMP_P (insn))
1832 {
1833 rtx asm_op = extract_asm_operands (PATTERN (insn));
1834 ASM_OPERANDS_TEMPLATE (asm_op) = ggc_strdup ("");
1835 ASM_OPERANDS_INPUT_VEC (asm_op) = rtvec_alloc (0);
1836 ASM_OPERANDS_INPUT_CONSTRAINT_VEC (asm_op) = rtvec_alloc (0);
1837 }
1838 else
1839 delete_insn (insn);
bbf9b913
RH
1840 }
1841 }
1842 else
1843 {
1844 if (recog_memoized (insn) < 0)
1845 fatal_insn_not_found (insn);
1846 }
1847}
14a774a9 1848
bbf9b913
RH
1849/* Subroutine of instantiate_decls. Given RTL representing a decl,
1850 do any instantiation required. */
14a774a9 1851
e41b2a33
PB
1852void
1853instantiate_decl_rtl (rtx x)
bbf9b913
RH
1854{
1855 rtx addr;
6f086dfc 1856
bbf9b913
RH
1857 if (x == 0)
1858 return;
6f086dfc 1859
bbf9b913
RH
1860 /* If this is a CONCAT, recurse for the pieces. */
1861 if (GET_CODE (x) == CONCAT)
1862 {
e41b2a33
PB
1863 instantiate_decl_rtl (XEXP (x, 0));
1864 instantiate_decl_rtl (XEXP (x, 1));
bbf9b913
RH
1865 return;
1866 }
6f086dfc 1867
bbf9b913
RH
1868 /* If this is not a MEM, no need to do anything. Similarly if the
1869 address is a constant or a register that is not a virtual register. */
1870 if (!MEM_P (x))
1871 return;
6f086dfc 1872
bbf9b913
RH
1873 addr = XEXP (x, 0);
1874 if (CONSTANT_P (addr)
1875 || (REG_P (addr)
1876 && (REGNO (addr) < FIRST_VIRTUAL_REGISTER
1877 || REGNO (addr) > LAST_VIRTUAL_REGISTER)))
1878 return;
6f086dfc 1879
b8704801 1880 instantiate_virtual_regs_in_rtx (&XEXP (x, 0));
bbf9b913 1881}
6f086dfc 1882
434eba35
JJ
1883/* Helper for instantiate_decls called via walk_tree: Process all decls
1884 in the given DECL_VALUE_EXPR. */
1885
1886static tree
1887instantiate_expr (tree *tp, int *walk_subtrees, void *data ATTRIBUTE_UNUSED)
1888{
1889 tree t = *tp;
726a989a 1890 if (! EXPR_P (t))
434eba35
JJ
1891 {
1892 *walk_subtrees = 0;
37d6a488
AO
1893 if (DECL_P (t))
1894 {
1895 if (DECL_RTL_SET_P (t))
1896 instantiate_decl_rtl (DECL_RTL (t));
1897 if (TREE_CODE (t) == PARM_DECL && DECL_NAMELESS (t)
1898 && DECL_INCOMING_RTL (t))
1899 instantiate_decl_rtl (DECL_INCOMING_RTL (t));
8813a647 1900 if ((VAR_P (t) || TREE_CODE (t) == RESULT_DECL)
37d6a488
AO
1901 && DECL_HAS_VALUE_EXPR_P (t))
1902 {
1903 tree v = DECL_VALUE_EXPR (t);
1904 walk_tree (&v, instantiate_expr, NULL, NULL);
1905 }
1906 }
434eba35
JJ
1907 }
1908 return NULL;
1909}
1910
bbf9b913
RH
1911/* Subroutine of instantiate_decls: Process all decls in the given
1912 BLOCK node and all its subblocks. */
6f086dfc 1913
bbf9b913
RH
1914static void
1915instantiate_decls_1 (tree let)
1916{
1917 tree t;
6f086dfc 1918
910ad8de 1919 for (t = BLOCK_VARS (let); t; t = DECL_CHAIN (t))
434eba35
JJ
1920 {
1921 if (DECL_RTL_SET_P (t))
e41b2a33 1922 instantiate_decl_rtl (DECL_RTL (t));
8813a647 1923 if (VAR_P (t) && DECL_HAS_VALUE_EXPR_P (t))
434eba35
JJ
1924 {
1925 tree v = DECL_VALUE_EXPR (t);
1926 walk_tree (&v, instantiate_expr, NULL, NULL);
1927 }
1928 }
6f086dfc 1929
bbf9b913 1930 /* Process all subblocks. */
87caf699 1931 for (t = BLOCK_SUBBLOCKS (let); t; t = BLOCK_CHAIN (t))
bbf9b913
RH
1932 instantiate_decls_1 (t);
1933}
6f086dfc 1934
bbf9b913
RH
1935/* Scan all decls in FNDECL (both variables and parameters) and instantiate
1936 all virtual registers in their DECL_RTL's. */
6f086dfc 1937
bbf9b913
RH
1938static void
1939instantiate_decls (tree fndecl)
1940{
c021f10b
NF
1941 tree decl;
1942 unsigned ix;
6f086dfc 1943
bbf9b913 1944 /* Process all parameters of the function. */
910ad8de 1945 for (decl = DECL_ARGUMENTS (fndecl); decl; decl = DECL_CHAIN (decl))
bbf9b913 1946 {
e41b2a33
PB
1947 instantiate_decl_rtl (DECL_RTL (decl));
1948 instantiate_decl_rtl (DECL_INCOMING_RTL (decl));
434eba35
JJ
1949 if (DECL_HAS_VALUE_EXPR_P (decl))
1950 {
1951 tree v = DECL_VALUE_EXPR (decl);
1952 walk_tree (&v, instantiate_expr, NULL, NULL);
1953 }
bbf9b913 1954 }
4fd796bb 1955
37d6a488
AO
1956 if ((decl = DECL_RESULT (fndecl))
1957 && TREE_CODE (decl) == RESULT_DECL)
1958 {
1959 if (DECL_RTL_SET_P (decl))
1960 instantiate_decl_rtl (DECL_RTL (decl));
1961 if (DECL_HAS_VALUE_EXPR_P (decl))
1962 {
1963 tree v = DECL_VALUE_EXPR (decl);
1964 walk_tree (&v, instantiate_expr, NULL, NULL);
1965 }
1966 }
1967
3fd48b12
EB
1968 /* Process the saved static chain if it exists. */
1969 decl = DECL_STRUCT_FUNCTION (fndecl)->static_chain_decl;
1970 if (decl && DECL_HAS_VALUE_EXPR_P (decl))
1971 instantiate_decl_rtl (DECL_RTL (DECL_VALUE_EXPR (decl)));
1972
bbf9b913 1973 /* Now process all variables defined in the function or its subblocks. */
51b86113
DM
1974 if (DECL_INITIAL (fndecl))
1975 instantiate_decls_1 (DECL_INITIAL (fndecl));
802e9f8e 1976
c021f10b
NF
1977 FOR_EACH_LOCAL_DECL (cfun, ix, decl)
1978 if (DECL_RTL_SET_P (decl))
1979 instantiate_decl_rtl (DECL_RTL (decl));
9771b263 1980 vec_free (cfun->local_decls);
bbf9b913 1981}
6f086dfc 1982
bbf9b913
RH
1983/* Pass through the INSNS of function FNDECL and convert virtual register
1984 references to hard register references. */
6f086dfc 1985
c2924966 1986static unsigned int
bbf9b913
RH
1987instantiate_virtual_regs (void)
1988{
691fe203 1989 rtx_insn *insn;
6f086dfc 1990
bbf9b913
RH
1991 /* Compute the offsets to use for this function. */
1992 in_arg_offset = FIRST_PARM_OFFSET (current_function_decl);
2a31c321 1993 var_offset = targetm.starting_frame_offset ();
bbf9b913
RH
1994 dynamic_offset = STACK_DYNAMIC_OFFSET (current_function_decl);
1995 out_arg_offset = STACK_POINTER_OFFSET;
f6672e8e
RH
1996#ifdef FRAME_POINTER_CFA_OFFSET
1997 cfa_offset = FRAME_POINTER_CFA_OFFSET (current_function_decl);
1998#else
bbf9b913 1999 cfa_offset = ARG_POINTER_CFA_OFFSET (current_function_decl);
f6672e8e 2000#endif
e9a25f70 2001
bbf9b913
RH
2002 /* Initialize recognition, indicating that volatile is OK. */
2003 init_recog ();
6f086dfc 2004
bbf9b913
RH
2005 /* Scan through all the insns, instantiating every virtual register still
2006 present. */
45dbce1b
NF
2007 for (insn = get_insns (); insn; insn = NEXT_INSN (insn))
2008 if (INSN_P (insn))
2009 {
2010 /* These patterns in the instruction stream can never be recognized.
2011 Fortunately, they shouldn't contain virtual registers either. */
39718607 2012 if (GET_CODE (PATTERN (insn)) == USE
45dbce1b 2013 || GET_CODE (PATTERN (insn)) == CLOBBER
36f52e8f
AO
2014 || GET_CODE (PATTERN (insn)) == ASM_INPUT
2015 || DEBUG_MARKER_INSN_P (insn))
45dbce1b 2016 continue;
36f52e8f
AO
2017 else if (DEBUG_BIND_INSN_P (insn))
2018 instantiate_virtual_regs_in_rtx (INSN_VAR_LOCATION_PTR (insn));
45dbce1b
NF
2019 else
2020 instantiate_virtual_regs_in_insn (insn);
ba4807a0 2021
4654c0cf 2022 if (insn->deleted ())
45dbce1b 2023 continue;
7114321e 2024
b8704801 2025 instantiate_virtual_regs_in_rtx (&REG_NOTES (insn));
ba4807a0 2026
45dbce1b
NF
2027 /* Instantiate any virtual registers in CALL_INSN_FUNCTION_USAGE. */
2028 if (CALL_P (insn))
b8704801 2029 instantiate_virtual_regs_in_rtx (&CALL_INSN_FUNCTION_USAGE (insn));
45dbce1b 2030 }
6f086dfc 2031
bbf9b913
RH
2032 /* Instantiate the virtual registers in the DECLs for debugging purposes. */
2033 instantiate_decls (current_function_decl);
2034
e41b2a33
PB
2035 targetm.instantiate_decls ();
2036
bbf9b913
RH
2037 /* Indicate that, from now on, assign_stack_local should use
2038 frame_pointer_rtx. */
2039 virtuals_instantiated = 1;
d3c12306 2040
c2924966 2041 return 0;
6f086dfc 2042}
ef330312 2043
27a4cd48
DM
2044namespace {
2045
2046const pass_data pass_data_instantiate_virtual_regs =
2047{
2048 RTL_PASS, /* type */
2049 "vregs", /* name */
2050 OPTGROUP_NONE, /* optinfo_flags */
27a4cd48
DM
2051 TV_NONE, /* tv_id */
2052 0, /* properties_required */
2053 0, /* properties_provided */
2054 0, /* properties_destroyed */
2055 0, /* todo_flags_start */
2056 0, /* todo_flags_finish */
ef330312
PB
2057};
2058
27a4cd48
DM
2059class pass_instantiate_virtual_regs : public rtl_opt_pass
2060{
2061public:
c3284718
RS
2062 pass_instantiate_virtual_regs (gcc::context *ctxt)
2063 : rtl_opt_pass (pass_data_instantiate_virtual_regs, ctxt)
27a4cd48
DM
2064 {}
2065
2066 /* opt_pass methods: */
be55bfe6
TS
2067 virtual unsigned int execute (function *)
2068 {
2069 return instantiate_virtual_regs ();
2070 }
27a4cd48
DM
2071
2072}; // class pass_instantiate_virtual_regs
2073
2074} // anon namespace
2075
2076rtl_opt_pass *
2077make_pass_instantiate_virtual_regs (gcc::context *ctxt)
2078{
2079 return new pass_instantiate_virtual_regs (ctxt);
2080}
2081
6f086dfc 2082\f
d181c154
RS
2083/* Return 1 if EXP is an aggregate type (or a value with aggregate type).
2084 This means a type for which function calls must pass an address to the
2085 function or get an address back from the function.
2086 EXP may be a type node or an expression (whose type is tested). */
6f086dfc
RS
2087
2088int
586de218 2089aggregate_value_p (const_tree exp, const_tree fntype)
6f086dfc 2090{
d47d0a8d 2091 const_tree type = (TYPE_P (exp)) ? exp : TREE_TYPE (exp);
9d790a4f
RS
2092 int i, regno, nregs;
2093 rtx reg;
2f939d94 2094
61f71b34
DD
2095 if (fntype)
2096 switch (TREE_CODE (fntype))
2097 {
2098 case CALL_EXPR:
d47d0a8d
EB
2099 {
2100 tree fndecl = get_callee_fndecl (fntype);
1304953e
JJ
2101 if (fndecl)
2102 fntype = TREE_TYPE (fndecl);
2103 else if (CALL_EXPR_FN (fntype))
2104 fntype = TREE_TYPE (TREE_TYPE (CALL_EXPR_FN (fntype)));
2105 else
2106 /* For internal functions, assume nothing needs to be
2107 returned in memory. */
2108 return 0;
d47d0a8d 2109 }
61f71b34
DD
2110 break;
2111 case FUNCTION_DECL:
d47d0a8d 2112 fntype = TREE_TYPE (fntype);
61f71b34
DD
2113 break;
2114 case FUNCTION_TYPE:
2115 case METHOD_TYPE:
2116 break;
2117 case IDENTIFIER_NODE:
d47d0a8d 2118 fntype = NULL_TREE;
61f71b34
DD
2119 break;
2120 default:
d47d0a8d 2121 /* We don't expect other tree types here. */
0bccc606 2122 gcc_unreachable ();
61f71b34
DD
2123 }
2124
d47d0a8d 2125 if (VOID_TYPE_P (type))
d7bf8ada 2126 return 0;
500c353d 2127
ebf0bf7f
JJ
2128 /* If a record should be passed the same as its first (and only) member
2129 don't pass it as an aggregate. */
2130 if (TREE_CODE (type) == RECORD_TYPE && TYPE_TRANSPARENT_AGGR (type))
2131 return aggregate_value_p (first_field (type), fntype);
2132
cc77ae10
JM
2133 /* If the front end has decided that this needs to be passed by
2134 reference, do so. */
2135 if ((TREE_CODE (exp) == PARM_DECL || TREE_CODE (exp) == RESULT_DECL)
2136 && DECL_BY_REFERENCE (exp))
2137 return 1;
500c353d 2138
d47d0a8d
EB
2139 /* Function types that are TREE_ADDRESSABLE force return in memory. */
2140 if (fntype && TREE_ADDRESSABLE (fntype))
500c353d 2141 return 1;
b8698a0f 2142
956d6950 2143 /* Types that are TREE_ADDRESSABLE must be constructed in memory,
49a2e5b2
DE
2144 and thus can't be returned in registers. */
2145 if (TREE_ADDRESSABLE (type))
2146 return 1;
d47d0a8d 2147
974aedcc
MP
2148 if (TYPE_EMPTY_P (type))
2149 return 0;
2150
05e3bdb9 2151 if (flag_pcc_struct_return && AGGREGATE_TYPE_P (type))
6f086dfc 2152 return 1;
d47d0a8d
EB
2153
2154 if (targetm.calls.return_in_memory (type, fntype))
2155 return 1;
2156
9d790a4f
RS
2157 /* Make sure we have suitable call-clobbered regs to return
2158 the value in; if not, we must return it in memory. */
1d636cc6 2159 reg = hard_function_value (type, 0, fntype, 0);
e71f7aa5
JW
2160
2161 /* If we have something other than a REG (e.g. a PARALLEL), then assume
2162 it is OK. */
f8cfc6aa 2163 if (!REG_P (reg))
e71f7aa5
JW
2164 return 0;
2165
9d790a4f 2166 regno = REGNO (reg);
ad474626 2167 nregs = hard_regno_nregs (regno, TYPE_MODE (type));
9d790a4f
RS
2168 for (i = 0; i < nregs; i++)
2169 if (! call_used_regs[regno + i])
2170 return 1;
d47d0a8d 2171
6f086dfc
RS
2172 return 0;
2173}
2174\f
8fff4fc1
RH
2175/* Return true if we should assign DECL a pseudo register; false if it
2176 should live on the local stack. */
2177
2178bool
fa233e34 2179use_register_for_decl (const_tree decl)
8fff4fc1 2180{
1f9ceff1
AO
2181 if (TREE_CODE (decl) == SSA_NAME)
2182 {
2183 /* We often try to use the SSA_NAME, instead of its underlying
2184 decl, to get type information and guide decisions, to avoid
2185 differences of behavior between anonymous and named
2186 variables, but in this one case we have to go for the actual
2187 variable if there is one. The main reason is that, at least
2188 at -O0, we want to place user variables on the stack, but we
2189 don't mind using pseudos for anonymous or ignored temps.
2190 Should we take the SSA_NAME, we'd conclude all SSA_NAMEs
2191 should go in pseudos, whereas their corresponding variables
2192 might have to go on the stack. So, disregarding the decl
2193 here would negatively impact debug info at -O0, enable
2194 coalescing between SSA_NAMEs that ought to get different
2195 stack/pseudo assignments, and get the incoming argument
2196 processing thoroughly confused by PARM_DECLs expected to live
2197 in stack slots but assigned to pseudos. */
2198 if (!SSA_NAME_VAR (decl))
2199 return TYPE_MODE (TREE_TYPE (decl)) != BLKmode
2200 && !(flag_float_store && FLOAT_TYPE_P (TREE_TYPE (decl)));
2201
2202 decl = SSA_NAME_VAR (decl);
2203 }
2204
8fff4fc1
RH
2205 /* Honor volatile. */
2206 if (TREE_SIDE_EFFECTS (decl))
2207 return false;
2208
2209 /* Honor addressability. */
2210 if (TREE_ADDRESSABLE (decl))
2211 return false;
2212
f11a7b6d
AO
2213 /* RESULT_DECLs are a bit special in that they're assigned without
2214 regard to use_register_for_decl, but we generally only store in
2215 them. If we coalesce their SSA NAMEs, we'd better return a
2216 result that matches the assignment in expand_function_start. */
2217 if (TREE_CODE (decl) == RESULT_DECL)
2218 {
2219 /* If it's not an aggregate, we're going to use a REG or a
2220 PARALLEL containing a REG. */
2221 if (!aggregate_value_p (decl, current_function_decl))
2222 return true;
2223
2224 /* If expand_function_start determines the return value, we'll
2225 use MEM if it's not by reference. */
2226 if (cfun->returns_pcc_struct
2227 || (targetm.calls.struct_value_rtx
2228 (TREE_TYPE (current_function_decl), 1)))
2229 return DECL_BY_REFERENCE (decl);
2230
2231 /* Otherwise, we're taking an extra all.function_result_decl
2232 argument. It's set up in assign_parms_augmented_arg_list,
2233 under the (negated) conditions above, and then it's used to
2234 set up the RESULT_DECL rtl in assign_params, after looping
2235 over all parameters. Now, if the RESULT_DECL is not by
2236 reference, we'll use a MEM either way. */
2237 if (!DECL_BY_REFERENCE (decl))
2238 return false;
2239
2240 /* Otherwise, if RESULT_DECL is DECL_BY_REFERENCE, it will take
2241 the function_result_decl's assignment. Since it's a pointer,
2242 we can short-circuit a number of the tests below, and we must
2243 duplicat e them because we don't have the
2244 function_result_decl to test. */
2245 if (!targetm.calls.allocate_stack_slots_for_args ())
2246 return true;
2247 /* We don't set DECL_IGNORED_P for the function_result_decl. */
2248 if (optimize)
2249 return true;
2250 /* We don't set DECL_REGISTER for the function_result_decl. */
2251 return false;
2252 }
2253
8fff4fc1
RH
2254 /* Only register-like things go in registers. */
2255 if (DECL_MODE (decl) == BLKmode)
2256 return false;
2257
2258 /* If -ffloat-store specified, don't put explicit float variables
2259 into registers. */
2260 /* ??? This should be checked after DECL_ARTIFICIAL, but tree-ssa
2261 propagates values across these stores, and it probably shouldn't. */
2262 if (flag_float_store && FLOAT_TYPE_P (TREE_TYPE (decl)))
2263 return false;
2264
de0fb905
AB
2265 if (!targetm.calls.allocate_stack_slots_for_args ())
2266 return true;
2267
78e0d62b
RH
2268 /* If we're not interested in tracking debugging information for
2269 this decl, then we can certainly put it in a register. */
2270 if (DECL_IGNORED_P (decl))
8fff4fc1
RH
2271 return true;
2272
d130d647
JJ
2273 if (optimize)
2274 return true;
2275
2276 if (!DECL_REGISTER (decl))
2277 return false;
2278
5aaa8fb4
NS
2279 /* When not optimizing, disregard register keyword for types that
2280 could have methods, otherwise the methods won't be callable from
2281 the debugger. */
2282 if (RECORD_OR_UNION_TYPE_P (TREE_TYPE (decl)))
2283 return false;
d130d647
JJ
2284
2285 return true;
8fff4fc1
RH
2286}
2287
6071dc7f
RH
2288/* Structures to communicate between the subroutines of assign_parms.
2289 The first holds data persistent across all parameters, the second
2290 is cleared out for each parameter. */
6f086dfc 2291
6071dc7f 2292struct assign_parm_data_all
6f086dfc 2293{
d5cc9181
JR
2294 /* When INIT_CUMULATIVE_ARGS gets revamped, allocating CUMULATIVE_ARGS
2295 should become a job of the target or otherwise encapsulated. */
2296 CUMULATIVE_ARGS args_so_far_v;
2297 cumulative_args_t args_so_far;
6f086dfc 2298 struct args_size stack_args_size;
6071dc7f
RH
2299 tree function_result_decl;
2300 tree orig_fnargs;
7a688d52
DM
2301 rtx_insn *first_conversion_insn;
2302 rtx_insn *last_conversion_insn;
6071dc7f
RH
2303 HOST_WIDE_INT pretend_args_size;
2304 HOST_WIDE_INT extra_pretend_bytes;
2305 int reg_parm_stack_space;
2306};
6f086dfc 2307
6071dc7f
RH
2308struct assign_parm_data_one
2309{
2310 tree nominal_type;
2311 tree passed_type;
2312 rtx entry_parm;
2313 rtx stack_parm;
ef4bddc2
RS
2314 machine_mode nominal_mode;
2315 machine_mode passed_mode;
2316 machine_mode promoted_mode;
6071dc7f
RH
2317 struct locate_and_pad_arg_data locate;
2318 int partial;
2319 BOOL_BITFIELD named_arg : 1;
6071dc7f
RH
2320 BOOL_BITFIELD passed_pointer : 1;
2321 BOOL_BITFIELD on_stack : 1;
2322 BOOL_BITFIELD loaded_in_reg : 1;
2323};
ebb904cb 2324
6071dc7f 2325/* A subroutine of assign_parms. Initialize ALL. */
6f086dfc 2326
6071dc7f
RH
2327static void
2328assign_parms_initialize_all (struct assign_parm_data_all *all)
2329{
fc2f1f53 2330 tree fntype ATTRIBUTE_UNUSED;
6f086dfc 2331
6071dc7f
RH
2332 memset (all, 0, sizeof (*all));
2333
2334 fntype = TREE_TYPE (current_function_decl);
2335
2336#ifdef INIT_CUMULATIVE_INCOMING_ARGS
d5cc9181 2337 INIT_CUMULATIVE_INCOMING_ARGS (all->args_so_far_v, fntype, NULL_RTX);
6071dc7f 2338#else
d5cc9181 2339 INIT_CUMULATIVE_ARGS (all->args_so_far_v, fntype, NULL_RTX,
6071dc7f
RH
2340 current_function_decl, -1);
2341#endif
d5cc9181 2342 all->args_so_far = pack_cumulative_args (&all->args_so_far_v);
6071dc7f 2343
ddbb449f
AM
2344#ifdef INCOMING_REG_PARM_STACK_SPACE
2345 all->reg_parm_stack_space
2346 = INCOMING_REG_PARM_STACK_SPACE (current_function_decl);
6071dc7f
RH
2347#endif
2348}
6f086dfc 2349
6071dc7f
RH
2350/* If ARGS contains entries with complex types, split the entry into two
2351 entries of the component type. Return a new list of substitutions are
2352 needed, else the old list. */
2353
3b3f318a 2354static void
f11a7b6d 2355split_complex_args (vec<tree> *args)
6071dc7f 2356{
3b3f318a 2357 unsigned i;
6071dc7f
RH
2358 tree p;
2359
9771b263 2360 FOR_EACH_VEC_ELT (*args, i, p)
6071dc7f
RH
2361 {
2362 tree type = TREE_TYPE (p);
2363 if (TREE_CODE (type) == COMPLEX_TYPE
2364 && targetm.calls.split_complex_arg (type))
2365 {
2366 tree decl;
2367 tree subtype = TREE_TYPE (type);
6ccd356e 2368 bool addressable = TREE_ADDRESSABLE (p);
6071dc7f
RH
2369
2370 /* Rewrite the PARM_DECL's type with its component. */
3b3f318a 2371 p = copy_node (p);
6071dc7f
RH
2372 TREE_TYPE (p) = subtype;
2373 DECL_ARG_TYPE (p) = TREE_TYPE (DECL_ARG_TYPE (p));
899ca90e 2374 SET_DECL_MODE (p, VOIDmode);
6071dc7f
RH
2375 DECL_SIZE (p) = NULL;
2376 DECL_SIZE_UNIT (p) = NULL;
6ccd356e
AM
2377 /* If this arg must go in memory, put it in a pseudo here.
2378 We can't allow it to go in memory as per normal parms,
2379 because the usual place might not have the imag part
2380 adjacent to the real part. */
2381 DECL_ARTIFICIAL (p) = addressable;
2382 DECL_IGNORED_P (p) = addressable;
2383 TREE_ADDRESSABLE (p) = 0;
6071dc7f 2384 layout_decl (p, 0);
9771b263 2385 (*args)[i] = p;
6071dc7f
RH
2386
2387 /* Build a second synthetic decl. */
c2255bc4
AH
2388 decl = build_decl (EXPR_LOCATION (p),
2389 PARM_DECL, NULL_TREE, subtype);
6071dc7f 2390 DECL_ARG_TYPE (decl) = DECL_ARG_TYPE (p);
6ccd356e
AM
2391 DECL_ARTIFICIAL (decl) = addressable;
2392 DECL_IGNORED_P (decl) = addressable;
6071dc7f 2393 layout_decl (decl, 0);
9771b263 2394 args->safe_insert (++i, decl);
6071dc7f
RH
2395 }
2396 }
6071dc7f
RH
2397}
2398
2399/* A subroutine of assign_parms. Adjust the parameter list to incorporate
2400 the hidden struct return argument, and (abi willing) complex args.
2401 Return the new parameter list. */
2402
9771b263 2403static vec<tree>
6071dc7f
RH
2404assign_parms_augmented_arg_list (struct assign_parm_data_all *all)
2405{
2406 tree fndecl = current_function_decl;
2407 tree fntype = TREE_TYPE (fndecl);
6e1aa848 2408 vec<tree> fnargs = vNULL;
3b3f318a
RG
2409 tree arg;
2410
910ad8de 2411 for (arg = DECL_ARGUMENTS (fndecl); arg; arg = DECL_CHAIN (arg))
9771b263 2412 fnargs.safe_push (arg);
3b3f318a
RG
2413
2414 all->orig_fnargs = DECL_ARGUMENTS (fndecl);
6f086dfc
RS
2415
2416 /* If struct value address is treated as the first argument, make it so. */
61f71b34 2417 if (aggregate_value_p (DECL_RESULT (fndecl), fndecl)
e3b5732b 2418 && ! cfun->returns_pcc_struct
61f71b34 2419 && targetm.calls.struct_value_rtx (TREE_TYPE (fndecl), 1) == 0)
6f086dfc 2420 {
f9f29478 2421 tree type = build_pointer_type (TREE_TYPE (fntype));
6071dc7f 2422 tree decl;
6f086dfc 2423
c2255bc4 2424 decl = build_decl (DECL_SOURCE_LOCATION (fndecl),
8dcfef8f 2425 PARM_DECL, get_identifier (".result_ptr"), type);
6071dc7f
RH
2426 DECL_ARG_TYPE (decl) = type;
2427 DECL_ARTIFICIAL (decl) = 1;
8dcfef8f
AO
2428 DECL_NAMELESS (decl) = 1;
2429 TREE_CONSTANT (decl) = 1;
f11a7b6d
AO
2430 /* We don't set DECL_IGNORED_P or DECL_REGISTER here. If this
2431 changes, the end of the RESULT_DECL handling block in
2432 use_register_for_decl must be adjusted to match. */
6f086dfc 2433
910ad8de 2434 DECL_CHAIN (decl) = all->orig_fnargs;
3b3f318a 2435 all->orig_fnargs = decl;
9771b263 2436 fnargs.safe_insert (0, decl);
3b3f318a 2437
6071dc7f 2438 all->function_result_decl = decl;
6f086dfc 2439 }
718fe406 2440
42ba5130
RH
2441 /* If the target wants to split complex arguments into scalars, do so. */
2442 if (targetm.calls.split_complex_arg)
f11a7b6d 2443 split_complex_args (&fnargs);
ded9bf77 2444
6071dc7f
RH
2445 return fnargs;
2446}
e7949876 2447
6071dc7f
RH
2448/* A subroutine of assign_parms. Examine PARM and pull out type and mode
2449 data for the parameter. Incorporate ABI specifics such as pass-by-
2450 reference and type promotion. */
6f086dfc 2451
6071dc7f
RH
2452static void
2453assign_parm_find_data_types (struct assign_parm_data_all *all, tree parm,
2454 struct assign_parm_data_one *data)
2455{
2456 tree nominal_type, passed_type;
ef4bddc2 2457 machine_mode nominal_mode, passed_mode, promoted_mode;
cde0f3fd 2458 int unsignedp;
6f086dfc 2459
6071dc7f
RH
2460 memset (data, 0, sizeof (*data));
2461
fa10beec 2462 /* NAMED_ARG is a misnomer. We really mean 'non-variadic'. */
e3b5732b 2463 if (!cfun->stdarg)
fa10beec 2464 data->named_arg = 1; /* No variadic parms. */
910ad8de 2465 else if (DECL_CHAIN (parm))
fa10beec 2466 data->named_arg = 1; /* Not the last non-variadic parm. */
d5cc9181 2467 else if (targetm.calls.strict_argument_naming (all->args_so_far))
fa10beec 2468 data->named_arg = 1; /* Only variadic ones are unnamed. */
6071dc7f 2469 else
fa10beec 2470 data->named_arg = 0; /* Treat as variadic. */
6071dc7f
RH
2471
2472 nominal_type = TREE_TYPE (parm);
2473 passed_type = DECL_ARG_TYPE (parm);
2474
2475 /* Look out for errors propagating this far. Also, if the parameter's
2476 type is void then its value doesn't matter. */
2477 if (TREE_TYPE (parm) == error_mark_node
2478 /* This can happen after weird syntax errors
2479 or if an enum type is defined among the parms. */
2480 || TREE_CODE (parm) != PARM_DECL
2481 || passed_type == NULL
2482 || VOID_TYPE_P (nominal_type))
2483 {
2484 nominal_type = passed_type = void_type_node;
2485 nominal_mode = passed_mode = promoted_mode = VOIDmode;
2486 goto egress;
2487 }
108b7d3d 2488
6071dc7f
RH
2489 /* Find mode of arg as it is passed, and mode of arg as it should be
2490 during execution of this function. */
2491 passed_mode = TYPE_MODE (passed_type);
2492 nominal_mode = TYPE_MODE (nominal_type);
2493
ebf0bf7f
JJ
2494 /* If the parm is to be passed as a transparent union or record, use the
2495 type of the first field for the tests below. We have already verified
2496 that the modes are the same. */
2497 if ((TREE_CODE (passed_type) == UNION_TYPE
2498 || TREE_CODE (passed_type) == RECORD_TYPE)
2499 && TYPE_TRANSPARENT_AGGR (passed_type))
2500 passed_type = TREE_TYPE (first_field (passed_type));
6071dc7f 2501
0976078c 2502 /* See if this arg was passed by invisible reference. */
d5cc9181 2503 if (pass_by_reference (&all->args_so_far_v, passed_mode,
0976078c 2504 passed_type, data->named_arg))
6071dc7f
RH
2505 {
2506 passed_type = nominal_type = build_pointer_type (passed_type);
2507 data->passed_pointer = true;
fd91cfe3 2508 passed_mode = nominal_mode = TYPE_MODE (nominal_type);
6071dc7f 2509 }
6f086dfc 2510
6071dc7f 2511 /* Find mode as it is passed by the ABI. */
cde0f3fd
PB
2512 unsignedp = TYPE_UNSIGNED (passed_type);
2513 promoted_mode = promote_function_mode (passed_type, passed_mode, &unsignedp,
2514 TREE_TYPE (current_function_decl), 0);
6f086dfc 2515
6071dc7f
RH
2516 egress:
2517 data->nominal_type = nominal_type;
2518 data->passed_type = passed_type;
2519 data->nominal_mode = nominal_mode;
2520 data->passed_mode = passed_mode;
2521 data->promoted_mode = promoted_mode;
2522}
16bae307 2523
6071dc7f 2524/* A subroutine of assign_parms. Invoke setup_incoming_varargs. */
6f086dfc 2525
6071dc7f
RH
2526static void
2527assign_parms_setup_varargs (struct assign_parm_data_all *all,
2528 struct assign_parm_data_one *data, bool no_rtl)
2529{
2530 int varargs_pretend_bytes = 0;
2531
d5cc9181 2532 targetm.calls.setup_incoming_varargs (all->args_so_far,
6071dc7f
RH
2533 data->promoted_mode,
2534 data->passed_type,
2535 &varargs_pretend_bytes, no_rtl);
2536
2537 /* If the back-end has requested extra stack space, record how much is
2538 needed. Do not change pretend_args_size otherwise since it may be
2539 nonzero from an earlier partial argument. */
2540 if (varargs_pretend_bytes > 0)
2541 all->pretend_args_size = varargs_pretend_bytes;
2542}
a53e14c0 2543
6071dc7f
RH
2544/* A subroutine of assign_parms. Set DATA->ENTRY_PARM corresponding to
2545 the incoming location of the current parameter. */
2546
2547static void
2548assign_parm_find_entry_rtl (struct assign_parm_data_all *all,
2549 struct assign_parm_data_one *data)
2550{
2551 HOST_WIDE_INT pretend_bytes = 0;
2552 rtx entry_parm;
2553 bool in_regs;
2554
2555 if (data->promoted_mode == VOIDmode)
2556 {
2557 data->entry_parm = data->stack_parm = const0_rtx;
2558 return;
2559 }
a53e14c0 2560
974aedcc
MP
2561 targetm.calls.warn_parameter_passing_abi (all->args_so_far,
2562 data->passed_type);
2563
d5cc9181 2564 entry_parm = targetm.calls.function_incoming_arg (all->args_so_far,
3c07301f
NF
2565 data->promoted_mode,
2566 data->passed_type,
2567 data->named_arg);
6f086dfc 2568
6071dc7f
RH
2569 if (entry_parm == 0)
2570 data->promoted_mode = data->passed_mode;
6f086dfc 2571
6071dc7f
RH
2572 /* Determine parm's home in the stack, in case it arrives in the stack
2573 or we should pretend it did. Compute the stack position and rtx where
2574 the argument arrives and its size.
6f086dfc 2575
6071dc7f
RH
2576 There is one complexity here: If this was a parameter that would
2577 have been passed in registers, but wasn't only because it is
2578 __builtin_va_alist, we want locate_and_pad_parm to treat it as if
2579 it came in a register so that REG_PARM_STACK_SPACE isn't skipped.
2580 In this case, we call FUNCTION_ARG with NAMED set to 1 instead of 0
2581 as it was the previous time. */
31db0fe0 2582 in_regs = (entry_parm != 0);
6f086dfc 2583#ifdef STACK_PARMS_IN_REG_PARM_AREA
6071dc7f 2584 in_regs = true;
e7949876 2585#endif
6071dc7f
RH
2586 if (!in_regs && !data->named_arg)
2587 {
d5cc9181 2588 if (targetm.calls.pretend_outgoing_varargs_named (all->args_so_far))
e7949876 2589 {
6071dc7f 2590 rtx tem;
d5cc9181 2591 tem = targetm.calls.function_incoming_arg (all->args_so_far,
3c07301f
NF
2592 data->promoted_mode,
2593 data->passed_type, true);
6071dc7f 2594 in_regs = tem != NULL;
e7949876 2595 }
6071dc7f 2596 }
e7949876 2597
6071dc7f
RH
2598 /* If this parameter was passed both in registers and in the stack, use
2599 the copy on the stack. */
fe984136
RH
2600 if (targetm.calls.must_pass_in_stack (data->promoted_mode,
2601 data->passed_type))
6071dc7f 2602 entry_parm = 0;
e7949876 2603
6071dc7f
RH
2604 if (entry_parm)
2605 {
2606 int partial;
2607
d5cc9181 2608 partial = targetm.calls.arg_partial_bytes (all->args_so_far,
78a52f11
RH
2609 data->promoted_mode,
2610 data->passed_type,
2611 data->named_arg);
6071dc7f
RH
2612 data->partial = partial;
2613
2614 /* The caller might already have allocated stack space for the
2615 register parameters. */
2616 if (partial != 0 && all->reg_parm_stack_space == 0)
975f3818 2617 {
6071dc7f
RH
2618 /* Part of this argument is passed in registers and part
2619 is passed on the stack. Ask the prologue code to extend
2620 the stack part so that we can recreate the full value.
2621
2622 PRETEND_BYTES is the size of the registers we need to store.
2623 CURRENT_FUNCTION_PRETEND_ARGS_SIZE is the amount of extra
2624 stack space that the prologue should allocate.
2625
2626 Internally, gcc assumes that the argument pointer is aligned
2627 to STACK_BOUNDARY bits. This is used both for alignment
2628 optimizations (see init_emit) and to locate arguments that are
2629 aligned to more than PARM_BOUNDARY bits. We must preserve this
2630 invariant by rounding CURRENT_FUNCTION_PRETEND_ARGS_SIZE up to
2631 a stack boundary. */
2632
2633 /* We assume at most one partial arg, and it must be the first
2634 argument on the stack. */
0bccc606 2635 gcc_assert (!all->extra_pretend_bytes && !all->pretend_args_size);
6071dc7f 2636
78a52f11 2637 pretend_bytes = partial;
6071dc7f
RH
2638 all->pretend_args_size = CEIL_ROUND (pretend_bytes, STACK_BYTES);
2639
2640 /* We want to align relative to the actual stack pointer, so
2641 don't include this in the stack size until later. */
2642 all->extra_pretend_bytes = all->pretend_args_size;
975f3818 2643 }
6071dc7f 2644 }
e7949876 2645
6071dc7f 2646 locate_and_pad_parm (data->promoted_mode, data->passed_type, in_regs,
2e4ceca5 2647 all->reg_parm_stack_space,
6071dc7f
RH
2648 entry_parm ? data->partial : 0, current_function_decl,
2649 &all->stack_args_size, &data->locate);
6f086dfc 2650
e94a448f
L
2651 /* Update parm_stack_boundary if this parameter is passed in the
2652 stack. */
2653 if (!in_regs && crtl->parm_stack_boundary < data->locate.boundary)
2654 crtl->parm_stack_boundary = data->locate.boundary;
2655
6071dc7f
RH
2656 /* Adjust offsets to include the pretend args. */
2657 pretend_bytes = all->extra_pretend_bytes - pretend_bytes;
2658 data->locate.slot_offset.constant += pretend_bytes;
2659 data->locate.offset.constant += pretend_bytes;
ebca59c3 2660
6071dc7f
RH
2661 data->entry_parm = entry_parm;
2662}
6f086dfc 2663
6071dc7f
RH
2664/* A subroutine of assign_parms. If there is actually space on the stack
2665 for this parm, count it in stack_args_size and return true. */
6f086dfc 2666
6071dc7f
RH
2667static bool
2668assign_parm_is_stack_parm (struct assign_parm_data_all *all,
2669 struct assign_parm_data_one *data)
2670{
2e6ae27f 2671 /* Trivially true if we've no incoming register. */
31db0fe0 2672 if (data->entry_parm == NULL)
6071dc7f
RH
2673 ;
2674 /* Also true if we're partially in registers and partially not,
2675 since we've arranged to drop the entire argument on the stack. */
2676 else if (data->partial != 0)
2677 ;
2678 /* Also true if the target says that it's passed in both registers
2679 and on the stack. */
2680 else if (GET_CODE (data->entry_parm) == PARALLEL
2681 && XEXP (XVECEXP (data->entry_parm, 0, 0), 0) == NULL_RTX)
2682 ;
2683 /* Also true if the target says that there's stack allocated for
2684 all register parameters. */
2685 else if (all->reg_parm_stack_space > 0)
2686 ;
2687 /* Otherwise, no, this parameter has no ABI defined stack slot. */
2688 else
2689 return false;
6f086dfc 2690
6071dc7f
RH
2691 all->stack_args_size.constant += data->locate.size.constant;
2692 if (data->locate.size.var)
2693 ADD_PARM_SIZE (all->stack_args_size, data->locate.size.var);
718fe406 2694
6071dc7f
RH
2695 return true;
2696}
0d1416c6 2697
6071dc7f
RH
2698/* A subroutine of assign_parms. Given that this parameter is allocated
2699 stack space by the ABI, find it. */
6f086dfc 2700
6071dc7f
RH
2701static void
2702assign_parm_find_stack_rtl (tree parm, struct assign_parm_data_one *data)
2703{
2704 rtx offset_rtx, stack_parm;
2705 unsigned int align, boundary;
6f086dfc 2706
6071dc7f
RH
2707 /* If we're passing this arg using a reg, make its stack home the
2708 aligned stack slot. */
2709 if (data->entry_parm)
2710 offset_rtx = ARGS_SIZE_RTX (data->locate.slot_offset);
2711 else
2712 offset_rtx = ARGS_SIZE_RTX (data->locate.offset);
2713
38173d38 2714 stack_parm = crtl->args.internal_arg_pointer;
6071dc7f
RH
2715 if (offset_rtx != const0_rtx)
2716 stack_parm = gen_rtx_PLUS (Pmode, stack_parm, offset_rtx);
2717 stack_parm = gen_rtx_MEM (data->promoted_mode, stack_parm);
2718
08ab0acf 2719 if (!data->passed_pointer)
997f78fb 2720 {
08ab0acf
JJ
2721 set_mem_attributes (stack_parm, parm, 1);
2722 /* set_mem_attributes could set MEM_SIZE to the passed mode's size,
2723 while promoted mode's size is needed. */
2724 if (data->promoted_mode != BLKmode
2725 && data->promoted_mode != DECL_MODE (parm))
997f78fb 2726 {
f5541398 2727 set_mem_size (stack_parm, GET_MODE_SIZE (data->promoted_mode));
527210c4 2728 if (MEM_EXPR (stack_parm) && MEM_OFFSET_KNOWN_P (stack_parm))
08ab0acf 2729 {
91914e56
RS
2730 poly_int64 offset = subreg_lowpart_offset (DECL_MODE (parm),
2731 data->promoted_mode);
2732 if (maybe_ne (offset, 0))
527210c4 2733 set_mem_offset (stack_parm, MEM_OFFSET (stack_parm) - offset);
08ab0acf 2734 }
997f78fb
JJ
2735 }
2736 }
6071dc7f 2737
bfc45551
AM
2738 boundary = data->locate.boundary;
2739 align = BITS_PER_UNIT;
6071dc7f
RH
2740
2741 /* If we're padding upward, we know that the alignment of the slot
c2ed6cf8 2742 is TARGET_FUNCTION_ARG_BOUNDARY. If we're using slot_offset, we're
6071dc7f
RH
2743 intentionally forcing upward padding. Otherwise we have to come
2744 up with a guess at the alignment based on OFFSET_RTX. */
a20c5714 2745 poly_int64 offset;
76b0cbf8 2746 if (data->locate.where_pad != PAD_DOWNWARD || data->entry_parm)
6071dc7f 2747 align = boundary;
a20c5714 2748 else if (poly_int_rtx_p (offset_rtx, &offset))
6071dc7f 2749 {
a20c5714
RS
2750 align = least_bit_hwi (boundary);
2751 unsigned int offset_align = known_alignment (offset) * BITS_PER_UNIT;
2752 if (offset_align != 0)
2753 align = MIN (align, offset_align);
6071dc7f 2754 }
bfc45551 2755 set_mem_align (stack_parm, align);
6071dc7f
RH
2756
2757 if (data->entry_parm)
2758 set_reg_attrs_for_parm (data->entry_parm, stack_parm);
2759
2760 data->stack_parm = stack_parm;
2761}
2762
2763/* A subroutine of assign_parms. Adjust DATA->ENTRY_RTL such that it's
2764 always valid and contiguous. */
2765
2766static void
2767assign_parm_adjust_entry_rtl (struct assign_parm_data_one *data)
2768{
2769 rtx entry_parm = data->entry_parm;
2770 rtx stack_parm = data->stack_parm;
2771
2772 /* If this parm was passed part in regs and part in memory, pretend it
2773 arrived entirely in memory by pushing the register-part onto the stack.
2774 In the special case of a DImode or DFmode that is split, we could put
2775 it together in a pseudoreg directly, but for now that's not worth
2776 bothering with. */
2777 if (data->partial != 0)
2778 {
2779 /* Handle calls that pass values in multiple non-contiguous
2780 locations. The Irix 6 ABI has examples of this. */
2781 if (GET_CODE (entry_parm) == PARALLEL)
1a8cb155 2782 emit_group_store (validize_mem (copy_rtx (stack_parm)), entry_parm,
b8698a0f 2783 data->passed_type,
6071dc7f 2784 int_size_in_bytes (data->passed_type));
6f086dfc 2785 else
78a52f11
RH
2786 {
2787 gcc_assert (data->partial % UNITS_PER_WORD == 0);
1a8cb155
RS
2788 move_block_from_reg (REGNO (entry_parm),
2789 validize_mem (copy_rtx (stack_parm)),
78a52f11
RH
2790 data->partial / UNITS_PER_WORD);
2791 }
6f086dfc 2792
6071dc7f
RH
2793 entry_parm = stack_parm;
2794 }
6f086dfc 2795
6071dc7f
RH
2796 /* If we didn't decide this parm came in a register, by default it came
2797 on the stack. */
2798 else if (entry_parm == NULL)
2799 entry_parm = stack_parm;
2800
2801 /* When an argument is passed in multiple locations, we can't make use
2802 of this information, but we can save some copying if the whole argument
2803 is passed in a single register. */
2804 else if (GET_CODE (entry_parm) == PARALLEL
2805 && data->nominal_mode != BLKmode
2806 && data->passed_mode != BLKmode)
2807 {
2808 size_t i, len = XVECLEN (entry_parm, 0);
2809
2810 for (i = 0; i < len; i++)
2811 if (XEXP (XVECEXP (entry_parm, 0, i), 0) != NULL_RTX
2812 && REG_P (XEXP (XVECEXP (entry_parm, 0, i), 0))
2813 && (GET_MODE (XEXP (XVECEXP (entry_parm, 0, i), 0))
2814 == data->passed_mode)
2815 && INTVAL (XEXP (XVECEXP (entry_parm, 0, i), 1)) == 0)
2816 {
2817 entry_parm = XEXP (XVECEXP (entry_parm, 0, i), 0);
2818 break;
2819 }
2820 }
e68a6ce1 2821
6071dc7f
RH
2822 data->entry_parm = entry_parm;
2823}
6f086dfc 2824
4d2a9850
DJ
2825/* A subroutine of assign_parms. Reconstitute any values which were
2826 passed in multiple registers and would fit in a single register. */
2827
2828static void
2829assign_parm_remove_parallels (struct assign_parm_data_one *data)
2830{
2831 rtx entry_parm = data->entry_parm;
2832
2833 /* Convert the PARALLEL to a REG of the same mode as the parallel.
2834 This can be done with register operations rather than on the
2835 stack, even if we will store the reconstituted parameter on the
2836 stack later. */
85776d60 2837 if (GET_CODE (entry_parm) == PARALLEL && GET_MODE (entry_parm) != BLKmode)
4d2a9850
DJ
2838 {
2839 rtx parmreg = gen_reg_rtx (GET_MODE (entry_parm));
bbd46fd5 2840 emit_group_store (parmreg, entry_parm, data->passed_type,
4d2a9850
DJ
2841 GET_MODE_SIZE (GET_MODE (entry_parm)));
2842 entry_parm = parmreg;
2843 }
2844
2845 data->entry_parm = entry_parm;
2846}
2847
6071dc7f
RH
2848/* A subroutine of assign_parms. Adjust DATA->STACK_RTL such that it's
2849 always valid and properly aligned. */
6f086dfc 2850
6071dc7f 2851static void
f11a7b6d 2852assign_parm_adjust_stack_rtl (struct assign_parm_data_one *data)
6071dc7f
RH
2853{
2854 rtx stack_parm = data->stack_parm;
2855
2856 /* If we can't trust the parm stack slot to be aligned enough for its
2857 ultimate type, don't use that slot after entry. We'll make another
2858 stack slot, if we need one. */
f11a7b6d
AO
2859 if (stack_parm
2860 && ((STRICT_ALIGNMENT
2861 && GET_MODE_ALIGNMENT (data->nominal_mode) > MEM_ALIGN (stack_parm))
2862 || (data->nominal_type
2863 && TYPE_ALIGN (data->nominal_type) > MEM_ALIGN (stack_parm)
2864 && MEM_ALIGN (stack_parm) < PREFERRED_STACK_BOUNDARY)))
6071dc7f
RH
2865 stack_parm = NULL;
2866
2867 /* If parm was passed in memory, and we need to convert it on entry,
2868 don't store it back in that same slot. */
2869 else if (data->entry_parm == stack_parm
2870 && data->nominal_mode != BLKmode
2871 && data->nominal_mode != data->passed_mode)
2872 stack_parm = NULL;
2873
7d69de61
RH
2874 /* If stack protection is in effect for this function, don't leave any
2875 pointers in their passed stack slots. */
cb91fab0 2876 else if (crtl->stack_protect_guard
7d69de61
RH
2877 && (flag_stack_protect == 2
2878 || data->passed_pointer
2879 || POINTER_TYPE_P (data->nominal_type)))
2880 stack_parm = NULL;
2881
6071dc7f
RH
2882 data->stack_parm = stack_parm;
2883}
a0506b54 2884
6071dc7f
RH
2885/* A subroutine of assign_parms. Return true if the current parameter
2886 should be stored as a BLKmode in the current frame. */
2887
2888static bool
2889assign_parm_setup_block_p (struct assign_parm_data_one *data)
2890{
2891 if (data->nominal_mode == BLKmode)
2892 return true;
85776d60
DJ
2893 if (GET_MODE (data->entry_parm) == BLKmode)
2894 return true;
531547e9 2895
6e985040 2896#ifdef BLOCK_REG_PADDING
ae8c9754
RS
2897 /* Only assign_parm_setup_block knows how to deal with register arguments
2898 that are padded at the least significant end. */
2899 if (REG_P (data->entry_parm)
cf098191 2900 && known_lt (GET_MODE_SIZE (data->promoted_mode), UNITS_PER_WORD)
ae8c9754 2901 && (BLOCK_REG_PADDING (data->passed_mode, data->passed_type, 1)
76b0cbf8 2902 == (BYTES_BIG_ENDIAN ? PAD_UPWARD : PAD_DOWNWARD)))
6071dc7f 2903 return true;
6e985040 2904#endif
6071dc7f
RH
2905
2906 return false;
2907}
2908
b8698a0f 2909/* A subroutine of assign_parms. Arrange for the parameter to be
6071dc7f
RH
2910 present and valid in DATA->STACK_RTL. */
2911
2912static void
27e29549
RH
2913assign_parm_setup_block (struct assign_parm_data_all *all,
2914 tree parm, struct assign_parm_data_one *data)
6071dc7f
RH
2915{
2916 rtx entry_parm = data->entry_parm;
2917 rtx stack_parm = data->stack_parm;
f11a7b6d 2918 rtx target_reg = NULL_RTX;
a029addd 2919 bool in_conversion_seq = false;
bfc45551
AM
2920 HOST_WIDE_INT size;
2921 HOST_WIDE_INT size_stored;
6071dc7f 2922
27e29549
RH
2923 if (GET_CODE (entry_parm) == PARALLEL)
2924 entry_parm = emit_group_move_into_temps (entry_parm);
2925
f11a7b6d
AO
2926 /* If we want the parameter in a pseudo, don't use a stack slot. */
2927 if (is_gimple_reg (parm) && use_register_for_decl (parm))
2928 {
2929 tree def = ssa_default_def (cfun, parm);
2930 gcc_assert (def);
2931 machine_mode mode = promote_ssa_mode (def, NULL);
2932 rtx reg = gen_reg_rtx (mode);
2933 if (GET_CODE (reg) != CONCAT)
2934 stack_parm = reg;
2935 else
a029addd
AO
2936 {
2937 target_reg = reg;
2938 /* Avoid allocating a stack slot, if there isn't one
2939 preallocated by the ABI. It might seem like we should
2940 always prefer a pseudo, but converting between
2941 floating-point and integer modes goes through the stack
2942 on various machines, so it's better to use the reserved
2943 stack slot than to risk wasting it and allocating more
2944 for the conversion. */
2945 if (stack_parm == NULL_RTX)
2946 {
2947 int save = generating_concat_p;
2948 generating_concat_p = 0;
2949 stack_parm = gen_reg_rtx (mode);
2950 generating_concat_p = save;
2951 }
2952 }
f11a7b6d
AO
2953 data->stack_parm = NULL;
2954 }
2955
bfc45551
AM
2956 size = int_size_in_bytes (data->passed_type);
2957 size_stored = CEIL_ROUND (size, UNITS_PER_WORD);
2958 if (stack_parm == 0)
2959 {
fe37c7af 2960 SET_DECL_ALIGN (parm, MAX (DECL_ALIGN (parm), BITS_PER_WORD));
f11a7b6d
AO
2961 stack_parm = assign_stack_local (BLKmode, size_stored,
2962 DECL_ALIGN (parm));
cf098191 2963 if (known_eq (GET_MODE_SIZE (GET_MODE (entry_parm)), size))
f11a7b6d
AO
2964 PUT_MODE (stack_parm, GET_MODE (entry_parm));
2965 set_mem_attributes (stack_parm, parm, 1);
bfc45551
AM
2966 }
2967
6071dc7f
RH
2968 /* If a BLKmode arrives in registers, copy it to a stack slot. Handle
2969 calls that pass values in multiple non-contiguous locations. */
2970 if (REG_P (entry_parm) || GET_CODE (entry_parm) == PARALLEL)
2971 {
6071dc7f
RH
2972 rtx mem;
2973
2974 /* Note that we will be storing an integral number of words.
2975 So we have to be careful to ensure that we allocate an
bfc45551 2976 integral number of words. We do this above when we call
6071dc7f
RH
2977 assign_stack_local if space was not allocated in the argument
2978 list. If it was, this will not work if PARM_BOUNDARY is not
2979 a multiple of BITS_PER_WORD. It isn't clear how to fix this
2980 if it becomes a problem. Exception is when BLKmode arrives
2981 with arguments not conforming to word_mode. */
2982
bfc45551
AM
2983 if (data->stack_parm == 0)
2984 ;
6071dc7f
RH
2985 else if (GET_CODE (entry_parm) == PARALLEL)
2986 ;
0bccc606
NS
2987 else
2988 gcc_assert (!size || !(PARM_BOUNDARY % BITS_PER_WORD));
6f086dfc 2989
1a8cb155 2990 mem = validize_mem (copy_rtx (stack_parm));
c6b97fac 2991
6071dc7f 2992 /* Handle values in multiple non-contiguous locations. */
a029addd
AO
2993 if (GET_CODE (entry_parm) == PARALLEL && !MEM_P (mem))
2994 emit_group_store (mem, entry_parm, data->passed_type, size);
2995 else if (GET_CODE (entry_parm) == PARALLEL)
27e29549 2996 {
bb27eeda
SE
2997 push_to_sequence2 (all->first_conversion_insn,
2998 all->last_conversion_insn);
27e29549 2999 emit_group_store (mem, entry_parm, data->passed_type, size);
bb27eeda
SE
3000 all->first_conversion_insn = get_insns ();
3001 all->last_conversion_insn = get_last_insn ();
27e29549 3002 end_sequence ();
a029addd 3003 in_conversion_seq = true;
27e29549 3004 }
c6b97fac 3005
6071dc7f
RH
3006 else if (size == 0)
3007 ;
5c07bd7a 3008
6071dc7f
RH
3009 /* If SIZE is that of a mode no bigger than a word, just use
3010 that mode's store operation. */
3011 else if (size <= UNITS_PER_WORD)
3012 {
f4b31647
RS
3013 unsigned int bits = size * BITS_PER_UNIT;
3014 machine_mode mode = int_mode_for_size (bits, 0).else_blk ();
c6b97fac 3015
6071dc7f 3016 if (mode != BLKmode
6e985040 3017#ifdef BLOCK_REG_PADDING
6071dc7f
RH
3018 && (size == UNITS_PER_WORD
3019 || (BLOCK_REG_PADDING (mode, data->passed_type, 1)
76b0cbf8 3020 != (BYTES_BIG_ENDIAN ? PAD_UPWARD : PAD_DOWNWARD)))
6e985040 3021#endif
6071dc7f
RH
3022 )
3023 {
208996c7
RS
3024 rtx reg;
3025
3026 /* We are really truncating a word_mode value containing
3027 SIZE bytes into a value of mode MODE. If such an
3028 operation requires no actual instructions, we can refer
3029 to the value directly in mode MODE, otherwise we must
3030 start with the register in word_mode and explicitly
3031 convert it. */
bb149ca2
RS
3032 if (targetm.truly_noop_truncation (size * BITS_PER_UNIT,
3033 BITS_PER_WORD))
208996c7
RS
3034 reg = gen_rtx_REG (mode, REGNO (entry_parm));
3035 else
3036 {
3037 reg = gen_rtx_REG (word_mode, REGNO (entry_parm));
3038 reg = convert_to_mode (mode, copy_to_reg (reg), 1);
3039 }
6071dc7f
RH
3040 emit_move_insn (change_address (mem, mode, 0), reg);
3041 }
c6b97fac 3042
1e5d7fd6
AO
3043#ifdef BLOCK_REG_PADDING
3044 /* Storing the register in memory as a full word, as
3045 move_block_from_reg below would do, and then using the
3046 MEM in a smaller mode, has the effect of shifting right
3047 if BYTES_BIG_ENDIAN. If we're bypassing memory, the
3048 shifting must be explicit. */
3049 else if (!MEM_P (mem))
3050 {
3051 rtx x;
3052
3053 /* If the assert below fails, we should have taken the
3054 mode != BLKmode path above, unless we have downward
3055 padding of smaller-than-word arguments on a machine
3056 with little-endian bytes, which would likely require
3057 additional changes to work correctly. */
3058 gcc_checking_assert (BYTES_BIG_ENDIAN
3059 && (BLOCK_REG_PADDING (mode,
3060 data->passed_type, 1)
76b0cbf8 3061 == PAD_UPWARD));
1e5d7fd6
AO
3062
3063 int by = (UNITS_PER_WORD - size) * BITS_PER_UNIT;
3064
3065 x = gen_rtx_REG (word_mode, REGNO (entry_parm));
3066 x = expand_shift (RSHIFT_EXPR, word_mode, x, by,
3067 NULL_RTX, 1);
3068 x = force_reg (word_mode, x);
3069 x = gen_lowpart_SUBREG (GET_MODE (mem), x);
3070
3071 emit_move_insn (mem, x);
3072 }
3073#endif
3074
6071dc7f
RH
3075 /* Blocks smaller than a word on a BYTES_BIG_ENDIAN
3076 machine must be aligned to the left before storing
3077 to memory. Note that the previous test doesn't
3078 handle all cases (e.g. SIZE == 3). */
3079 else if (size != UNITS_PER_WORD
6e985040 3080#ifdef BLOCK_REG_PADDING
6071dc7f 3081 && (BLOCK_REG_PADDING (mode, data->passed_type, 1)
76b0cbf8 3082 == PAD_DOWNWARD)
6e985040 3083#else
6071dc7f 3084 && BYTES_BIG_ENDIAN
6e985040 3085#endif
6071dc7f
RH
3086 )
3087 {
3088 rtx tem, x;
3089 int by = (UNITS_PER_WORD - size) * BITS_PER_UNIT;
65c844e2 3090 rtx reg = gen_rtx_REG (word_mode, REGNO (entry_parm));
6071dc7f 3091
eb6c3df1 3092 x = expand_shift (LSHIFT_EXPR, word_mode, reg, by, NULL_RTX, 1);
6071dc7f
RH
3093 tem = change_address (mem, word_mode, 0);
3094 emit_move_insn (tem, x);
6f086dfc 3095 }
6071dc7f 3096 else
27e29549 3097 move_block_from_reg (REGNO (entry_parm), mem,
6071dc7f 3098 size_stored / UNITS_PER_WORD);
6f086dfc 3099 }
f11a7b6d 3100 else if (!MEM_P (mem))
1e5d7fd6
AO
3101 {
3102 gcc_checking_assert (size > UNITS_PER_WORD);
3103#ifdef BLOCK_REG_PADDING
3104 gcc_checking_assert (BLOCK_REG_PADDING (GET_MODE (mem),
3105 data->passed_type, 0)
76b0cbf8 3106 == PAD_UPWARD);
1e5d7fd6
AO
3107#endif
3108 emit_move_insn (mem, entry_parm);
3109 }
6071dc7f 3110 else
27e29549 3111 move_block_from_reg (REGNO (entry_parm), mem,
6071dc7f
RH
3112 size_stored / UNITS_PER_WORD);
3113 }
bfc45551
AM
3114 else if (data->stack_parm == 0)
3115 {
bb27eeda 3116 push_to_sequence2 (all->first_conversion_insn, all->last_conversion_insn);
bfc45551
AM
3117 emit_block_move (stack_parm, data->entry_parm, GEN_INT (size),
3118 BLOCK_OP_NORMAL);
bb27eeda
SE
3119 all->first_conversion_insn = get_insns ();
3120 all->last_conversion_insn = get_last_insn ();
bfc45551 3121 end_sequence ();
a029addd 3122 in_conversion_seq = true;
bfc45551 3123 }
6071dc7f 3124
f11a7b6d
AO
3125 if (target_reg)
3126 {
a029addd
AO
3127 if (!in_conversion_seq)
3128 emit_move_insn (target_reg, stack_parm);
3129 else
3130 {
3131 push_to_sequence2 (all->first_conversion_insn,
3132 all->last_conversion_insn);
3133 emit_move_insn (target_reg, stack_parm);
3134 all->first_conversion_insn = get_insns ();
3135 all->last_conversion_insn = get_last_insn ();
3136 end_sequence ();
3137 }
f11a7b6d
AO
3138 stack_parm = target_reg;
3139 }
3140
bfc45551 3141 data->stack_parm = stack_parm;
f11a7b6d 3142 set_parm_rtl (parm, stack_parm);
6071dc7f
RH
3143}
3144
3145/* A subroutine of assign_parms. Allocate a pseudo to hold the current
3146 parameter. Get it there. Perform all ABI specified conversions. */
3147
3148static void
3149assign_parm_setup_reg (struct assign_parm_data_all *all, tree parm,
3150 struct assign_parm_data_one *data)
3151{
71008de4
BS
3152 rtx parmreg, validated_mem;
3153 rtx equiv_stack_parm;
ef4bddc2 3154 machine_mode promoted_nominal_mode;
6071dc7f
RH
3155 int unsignedp = TYPE_UNSIGNED (TREE_TYPE (parm));
3156 bool did_conversion = false;
71008de4 3157 bool need_conversion, moved;
f11a7b6d 3158 rtx rtl;
6071dc7f
RH
3159
3160 /* Store the parm in a pseudoregister during the function, but we may
666e3ceb
PB
3161 need to do it in a wider mode. Using 2 here makes the result
3162 consistent with promote_decl_mode and thus expand_expr_real_1. */
6071dc7f 3163 promoted_nominal_mode
cde0f3fd 3164 = promote_function_mode (data->nominal_type, data->nominal_mode, &unsignedp,
666e3ceb 3165 TREE_TYPE (current_function_decl), 2);
6071dc7f 3166
f11a7b6d
AO
3167 parmreg = gen_reg_rtx (promoted_nominal_mode);
3168 if (!DECL_ARTIFICIAL (parm))
3169 mark_user_reg (parmreg);
6071dc7f
RH
3170
3171 /* If this was an item that we received a pointer to,
f11a7b6d
AO
3172 set rtl appropriately. */
3173 if (data->passed_pointer)
6071dc7f 3174 {
f11a7b6d
AO
3175 rtl = gen_rtx_MEM (TYPE_MODE (TREE_TYPE (data->passed_type)), parmreg);
3176 set_mem_attributes (rtl, parm, 1);
6071dc7f
RH
3177 }
3178 else
f11a7b6d 3179 rtl = parmreg;
6071dc7f 3180
4d2a9850
DJ
3181 assign_parm_remove_parallels (data);
3182
666e3ceb
PB
3183 /* Copy the value into the register, thus bridging between
3184 assign_parm_find_data_types and expand_expr_real_1. */
6071dc7f 3185
71008de4 3186 equiv_stack_parm = data->stack_parm;
1a8cb155 3187 validated_mem = validize_mem (copy_rtx (data->entry_parm));
71008de4
BS
3188
3189 need_conversion = (data->nominal_mode != data->passed_mode
3190 || promoted_nominal_mode != data->promoted_mode);
3191 moved = false;
3192
dbb94435
BS
3193 if (need_conversion
3194 && GET_MODE_CLASS (data->nominal_mode) == MODE_INT
3195 && data->nominal_mode == data->passed_mode
3196 && data->nominal_mode == GET_MODE (data->entry_parm))
71008de4 3197 {
6071dc7f
RH
3198 /* ENTRY_PARM has been converted to PROMOTED_MODE, its
3199 mode, by the caller. We now have to convert it to
3200 NOMINAL_MODE, if different. However, PARMREG may be in
3201 a different mode than NOMINAL_MODE if it is being stored
3202 promoted.
3203
3204 If ENTRY_PARM is a hard register, it might be in a register
3205 not valid for operating in its mode (e.g., an odd-numbered
3206 register for a DFmode). In that case, moves are the only
3207 thing valid, so we can't do a convert from there. This
3208 occurs when the calling sequence allow such misaligned
3209 usages.
3210
3211 In addition, the conversion may involve a call, which could
3212 clobber parameters which haven't been copied to pseudo
71008de4
BS
3213 registers yet.
3214
3215 First, we try to emit an insn which performs the necessary
3216 conversion. We verify that this insn does not clobber any
3217 hard registers. */
3218
3219 enum insn_code icode;
3220 rtx op0, op1;
3221
3222 icode = can_extend_p (promoted_nominal_mode, data->passed_mode,
3223 unsignedp);
3224
3225 op0 = parmreg;
3226 op1 = validated_mem;
3227 if (icode != CODE_FOR_nothing
2ef6ce06
RS
3228 && insn_operand_matches (icode, 0, op0)
3229 && insn_operand_matches (icode, 1, op1))
71008de4
BS
3230 {
3231 enum rtx_code code = unsignedp ? ZERO_EXTEND : SIGN_EXTEND;
b32d5189
DM
3232 rtx_insn *insn, *insns;
3233 rtx t = op1;
71008de4
BS
3234 HARD_REG_SET hardregs;
3235
3236 start_sequence ();
f9fef349
JJ
3237 /* If op1 is a hard register that is likely spilled, first
3238 force it into a pseudo, otherwise combiner might extend
3239 its lifetime too much. */
3240 if (GET_CODE (t) == SUBREG)
3241 t = SUBREG_REG (t);
3242 if (REG_P (t)
3243 && HARD_REGISTER_P (t)
3244 && ! TEST_HARD_REG_BIT (fixed_reg_set, REGNO (t))
3245 && targetm.class_likely_spilled_p (REGNO_REG_CLASS (REGNO (t))))
3246 {
3247 t = gen_reg_rtx (GET_MODE (op1));
3248 emit_move_insn (t, op1);
3249 }
3250 else
3251 t = op1;
e67d1102
RS
3252 rtx_insn *pat = gen_extend_insn (op0, t, promoted_nominal_mode,
3253 data->passed_mode, unsignedp);
a11899b2 3254 emit_insn (pat);
71008de4
BS
3255 insns = get_insns ();
3256
3257 moved = true;
3258 CLEAR_HARD_REG_SET (hardregs);
3259 for (insn = insns; insn && moved; insn = NEXT_INSN (insn))
3260 {
3261 if (INSN_P (insn))
3262 note_stores (PATTERN (insn), record_hard_reg_sets,
3263 &hardregs);
3264 if (!hard_reg_set_empty_p (hardregs))
3265 moved = false;
3266 }
3267
3268 end_sequence ();
3269
3270 if (moved)
3271 {
3272 emit_insn (insns);
dbb94435
BS
3273 if (equiv_stack_parm != NULL_RTX)
3274 equiv_stack_parm = gen_rtx_fmt_e (code, GET_MODE (parmreg),
3275 equiv_stack_parm);
71008de4
BS
3276 }
3277 }
3278 }
3279
3280 if (moved)
3281 /* Nothing to do. */
3282 ;
3283 else if (need_conversion)
3284 {
3285 /* We did not have an insn to convert directly, or the sequence
3286 generated appeared unsafe. We must first copy the parm to a
3287 pseudo reg, and save the conversion until after all
6071dc7f
RH
3288 parameters have been moved. */
3289
71008de4 3290 int save_tree_used;
6071dc7f
RH
3291 rtx tempreg = gen_reg_rtx (GET_MODE (data->entry_parm));
3292
71008de4 3293 emit_move_insn (tempreg, validated_mem);
6071dc7f 3294
bb27eeda 3295 push_to_sequence2 (all->first_conversion_insn, all->last_conversion_insn);
6071dc7f
RH
3296 tempreg = convert_to_mode (data->nominal_mode, tempreg, unsignedp);
3297
bd4288c0 3298 if (partial_subreg_p (tempreg)
6071dc7f
RH
3299 && GET_MODE (tempreg) == data->nominal_mode
3300 && REG_P (SUBREG_REG (tempreg))
3301 && data->nominal_mode == data->passed_mode
bd4288c0 3302 && GET_MODE (SUBREG_REG (tempreg)) == GET_MODE (data->entry_parm))
6f086dfc 3303 {
6071dc7f
RH
3304 /* The argument is already sign/zero extended, so note it
3305 into the subreg. */
3306 SUBREG_PROMOTED_VAR_P (tempreg) = 1;
362d42dc 3307 SUBREG_PROMOTED_SET (tempreg, unsignedp);
6071dc7f 3308 }
00d8a4c1 3309
6071dc7f
RH
3310 /* TREE_USED gets set erroneously during expand_assignment. */
3311 save_tree_used = TREE_USED (parm);
f11a7b6d 3312 SET_DECL_RTL (parm, rtl);
79f5e442 3313 expand_assignment (parm, make_tree (data->nominal_type, tempreg), false);
f11a7b6d 3314 SET_DECL_RTL (parm, NULL_RTX);
6071dc7f 3315 TREE_USED (parm) = save_tree_used;
bb27eeda
SE
3316 all->first_conversion_insn = get_insns ();
3317 all->last_conversion_insn = get_last_insn ();
6071dc7f 3318 end_sequence ();
00d8a4c1 3319
6071dc7f
RH
3320 did_conversion = true;
3321 }
f11a7b6d 3322 else
71008de4 3323 emit_move_insn (parmreg, validated_mem);
6071dc7f
RH
3324
3325 /* If we were passed a pointer but the actual value can safely live
f7e088e7 3326 in a register, retrieve it and use it directly. */
f11a7b6d 3327 if (data->passed_pointer && TYPE_MODE (TREE_TYPE (parm)) != BLKmode)
6071dc7f
RH
3328 {
3329 /* We can't use nominal_mode, because it will have been set to
3330 Pmode above. We must use the actual mode of the parm. */
f11a7b6d 3331 if (use_register_for_decl (parm))
f7e088e7
EB
3332 {
3333 parmreg = gen_reg_rtx (TYPE_MODE (TREE_TYPE (parm)));
3334 mark_user_reg (parmreg);
3335 }
3336 else
3337 {
3338 int align = STACK_SLOT_ALIGNMENT (TREE_TYPE (parm),
3339 TYPE_MODE (TREE_TYPE (parm)),
3340 TYPE_ALIGN (TREE_TYPE (parm)));
3341 parmreg
3342 = assign_stack_local (TYPE_MODE (TREE_TYPE (parm)),
3343 GET_MODE_SIZE (TYPE_MODE (TREE_TYPE (parm))),
3344 align);
3345 set_mem_attributes (parmreg, parm, 1);
3346 }
cd5b3469 3347
951d8c8a
EB
3348 /* We need to preserve an address based on VIRTUAL_STACK_VARS_REGNUM for
3349 the debug info in case it is not legitimate. */
f11a7b6d 3350 if (GET_MODE (parmreg) != GET_MODE (rtl))
6071dc7f 3351 {
f11a7b6d 3352 rtx tempreg = gen_reg_rtx (GET_MODE (rtl));
6071dc7f
RH
3353 int unsigned_p = TYPE_UNSIGNED (TREE_TYPE (parm));
3354
bb27eeda
SE
3355 push_to_sequence2 (all->first_conversion_insn,
3356 all->last_conversion_insn);
f11a7b6d 3357 emit_move_insn (tempreg, rtl);
6071dc7f 3358 tempreg = convert_to_mode (GET_MODE (parmreg), tempreg, unsigned_p);
951d8c8a
EB
3359 emit_move_insn (MEM_P (parmreg) ? copy_rtx (parmreg) : parmreg,
3360 tempreg);
bb27eeda
SE
3361 all->first_conversion_insn = get_insns ();
3362 all->last_conversion_insn = get_last_insn ();
6071dc7f 3363 end_sequence ();
6f086dfc 3364
6071dc7f
RH
3365 did_conversion = true;
3366 }
3367 else
951d8c8a 3368 emit_move_insn (MEM_P (parmreg) ? copy_rtx (parmreg) : parmreg, rtl);
6f086dfc 3369
f11a7b6d 3370 rtl = parmreg;
797a6ac1 3371
6071dc7f
RH
3372 /* STACK_PARM is the pointer, not the parm, and PARMREG is
3373 now the parm. */
f11a7b6d 3374 data->stack_parm = NULL;
6071dc7f 3375 }
ddef6bc7 3376
f11a7b6d
AO
3377 set_parm_rtl (parm, rtl);
3378
6071dc7f
RH
3379 /* Mark the register as eliminable if we did no conversion and it was
3380 copied from memory at a fixed offset, and the arg pointer was not
3381 copied to a pseudo-reg. If the arg pointer is a pseudo reg or the
3382 offset formed an invalid address, such memory-equivalences as we
3383 make here would screw up life analysis for it. */
3384 if (data->nominal_mode == data->passed_mode
3385 && !did_conversion
f11a7b6d
AO
3386 && data->stack_parm != 0
3387 && MEM_P (data->stack_parm)
6071dc7f
RH
3388 && data->locate.offset.var == 0
3389 && reg_mentioned_p (virtual_incoming_args_rtx,
f11a7b6d 3390 XEXP (data->stack_parm, 0)))
6071dc7f 3391 {
691fe203
DM
3392 rtx_insn *linsn = get_last_insn ();
3393 rtx_insn *sinsn;
3394 rtx set;
a03caf76 3395
6071dc7f
RH
3396 /* Mark complex types separately. */
3397 if (GET_CODE (parmreg) == CONCAT)
3398 {
d21cefc2 3399 scalar_mode submode = GET_MODE_INNER (GET_MODE (parmreg));
1466e387
RH
3400 int regnor = REGNO (XEXP (parmreg, 0));
3401 int regnoi = REGNO (XEXP (parmreg, 1));
f11a7b6d
AO
3402 rtx stackr = adjust_address_nv (data->stack_parm, submode, 0);
3403 rtx stacki = adjust_address_nv (data->stack_parm, submode,
1466e387 3404 GET_MODE_SIZE (submode));
6071dc7f
RH
3405
3406 /* Scan backwards for the set of the real and
3407 imaginary parts. */
3408 for (sinsn = linsn; sinsn != 0;
3409 sinsn = prev_nonnote_insn (sinsn))
3410 {
3411 set = single_set (sinsn);
3412 if (set == 0)
3413 continue;
3414
3415 if (SET_DEST (set) == regno_reg_rtx [regnoi])
a31830a7 3416 set_unique_reg_note (sinsn, REG_EQUIV, stacki);
6071dc7f 3417 else if (SET_DEST (set) == regno_reg_rtx [regnor])
a31830a7 3418 set_unique_reg_note (sinsn, REG_EQUIV, stackr);
a03caf76 3419 }
6071dc7f 3420 }
f11a7b6d 3421 else
7543f918 3422 set_dst_reg_note (linsn, REG_EQUIV, equiv_stack_parm, parmreg);
6071dc7f
RH
3423 }
3424
3425 /* For pointer data type, suggest pointer register. */
3426 if (POINTER_TYPE_P (TREE_TYPE (parm)))
3427 mark_reg_pointer (parmreg,
3428 TYPE_ALIGN (TREE_TYPE (TREE_TYPE (parm))));
3429}
3430
3431/* A subroutine of assign_parms. Allocate stack space to hold the current
3432 parameter. Get it there. Perform all ABI specified conversions. */
3433
3434static void
3435assign_parm_setup_stack (struct assign_parm_data_all *all, tree parm,
3436 struct assign_parm_data_one *data)
3437{
3438 /* Value must be stored in the stack slot STACK_PARM during function
3439 execution. */
bfc45551 3440 bool to_conversion = false;
6071dc7f 3441
4d2a9850
DJ
3442 assign_parm_remove_parallels (data);
3443
6071dc7f
RH
3444 if (data->promoted_mode != data->nominal_mode)
3445 {
3446 /* Conversion is required. */
3447 rtx tempreg = gen_reg_rtx (GET_MODE (data->entry_parm));
6f086dfc 3448
1a8cb155 3449 emit_move_insn (tempreg, validize_mem (copy_rtx (data->entry_parm)));
6071dc7f 3450
bb27eeda 3451 push_to_sequence2 (all->first_conversion_insn, all->last_conversion_insn);
bfc45551
AM
3452 to_conversion = true;
3453
6071dc7f
RH
3454 data->entry_parm = convert_to_mode (data->nominal_mode, tempreg,
3455 TYPE_UNSIGNED (TREE_TYPE (parm)));
3456
3457 if (data->stack_parm)
dd67163f 3458 {
91914e56
RS
3459 poly_int64 offset
3460 = subreg_lowpart_offset (data->nominal_mode,
3461 GET_MODE (data->stack_parm));
dd67163f
JJ
3462 /* ??? This may need a big-endian conversion on sparc64. */
3463 data->stack_parm
3464 = adjust_address (data->stack_parm, data->nominal_mode, 0);
91914e56 3465 if (maybe_ne (offset, 0) && MEM_OFFSET_KNOWN_P (data->stack_parm))
dd67163f 3466 set_mem_offset (data->stack_parm,
527210c4 3467 MEM_OFFSET (data->stack_parm) + offset);
dd67163f 3468 }
6071dc7f
RH
3469 }
3470
3471 if (data->entry_parm != data->stack_parm)
3472 {
bfc45551 3473 rtx src, dest;
1f9ceff1 3474
6071dc7f
RH
3475 if (data->stack_parm == 0)
3476 {
3a695389
UW
3477 int align = STACK_SLOT_ALIGNMENT (data->passed_type,
3478 GET_MODE (data->entry_parm),
3479 TYPE_ALIGN (data->passed_type));
6071dc7f
RH
3480 data->stack_parm
3481 = assign_stack_local (GET_MODE (data->entry_parm),
3482 GET_MODE_SIZE (GET_MODE (data->entry_parm)),
3a695389 3483 align);
f11a7b6d 3484 set_mem_attributes (data->stack_parm, parm, 1);
6f086dfc 3485 }
6071dc7f 3486
1a8cb155
RS
3487 dest = validize_mem (copy_rtx (data->stack_parm));
3488 src = validize_mem (copy_rtx (data->entry_parm));
bfc45551
AM
3489
3490 if (MEM_P (src))
6f086dfc 3491 {
bfc45551
AM
3492 /* Use a block move to handle potentially misaligned entry_parm. */
3493 if (!to_conversion)
bb27eeda
SE
3494 push_to_sequence2 (all->first_conversion_insn,
3495 all->last_conversion_insn);
bfc45551
AM
3496 to_conversion = true;
3497
3498 emit_block_move (dest, src,
3499 GEN_INT (int_size_in_bytes (data->passed_type)),
3500 BLOCK_OP_NORMAL);
6071dc7f
RH
3501 }
3502 else
4a235312
L
3503 {
3504 if (!REG_P (src))
3505 src = force_reg (GET_MODE (src), src);
3506 emit_move_insn (dest, src);
3507 }
bfc45551
AM
3508 }
3509
3510 if (to_conversion)
3511 {
bb27eeda
SE
3512 all->first_conversion_insn = get_insns ();
3513 all->last_conversion_insn = get_last_insn ();
bfc45551 3514 end_sequence ();
6071dc7f 3515 }
6f086dfc 3516
f11a7b6d 3517 set_parm_rtl (parm, data->stack_parm);
6071dc7f 3518}
3412b298 3519
6071dc7f
RH
3520/* A subroutine of assign_parms. If the ABI splits complex arguments, then
3521 undo the frobbing that we did in assign_parms_augmented_arg_list. */
86f8eff3 3522
6071dc7f 3523static void
3b3f318a 3524assign_parms_unsplit_complex (struct assign_parm_data_all *all,
9771b263 3525 vec<tree> fnargs)
6071dc7f
RH
3526{
3527 tree parm;
6ccd356e 3528 tree orig_fnargs = all->orig_fnargs;
3b3f318a 3529 unsigned i = 0;
f4ef873c 3530
3b3f318a 3531 for (parm = orig_fnargs; parm; parm = TREE_CHAIN (parm), ++i)
6071dc7f
RH
3532 {
3533 if (TREE_CODE (TREE_TYPE (parm)) == COMPLEX_TYPE
3534 && targetm.calls.split_complex_arg (TREE_TYPE (parm)))
3535 {
3536 rtx tmp, real, imag;
d21cefc2 3537 scalar_mode inner = GET_MODE_INNER (DECL_MODE (parm));
6f086dfc 3538
9771b263
DN
3539 real = DECL_RTL (fnargs[i]);
3540 imag = DECL_RTL (fnargs[i + 1]);
6071dc7f 3541 if (inner != GET_MODE (real))
6f086dfc 3542 {
f11a7b6d
AO
3543 real = gen_lowpart_SUBREG (inner, real);
3544 imag = gen_lowpart_SUBREG (inner, imag);
6071dc7f 3545 }
6ccd356e 3546
f11a7b6d 3547 if (TREE_ADDRESSABLE (parm))
6ccd356e
AM
3548 {
3549 rtx rmem, imem;
3550 HOST_WIDE_INT size = int_size_in_bytes (TREE_TYPE (parm));
3a695389
UW
3551 int align = STACK_SLOT_ALIGNMENT (TREE_TYPE (parm),
3552 DECL_MODE (parm),
3553 TYPE_ALIGN (TREE_TYPE (parm)));
6ccd356e
AM
3554
3555 /* split_complex_arg put the real and imag parts in
3556 pseudos. Move them to memory. */
3a695389 3557 tmp = assign_stack_local (DECL_MODE (parm), size, align);
6ccd356e
AM
3558 set_mem_attributes (tmp, parm, 1);
3559 rmem = adjust_address_nv (tmp, inner, 0);
3560 imem = adjust_address_nv (tmp, inner, GET_MODE_SIZE (inner));
bb27eeda
SE
3561 push_to_sequence2 (all->first_conversion_insn,
3562 all->last_conversion_insn);
6ccd356e
AM
3563 emit_move_insn (rmem, real);
3564 emit_move_insn (imem, imag);
bb27eeda
SE
3565 all->first_conversion_insn = get_insns ();
3566 all->last_conversion_insn = get_last_insn ();
6ccd356e
AM
3567 end_sequence ();
3568 }
3569 else
3570 tmp = gen_rtx_CONCAT (DECL_MODE (parm), real, imag);
f11a7b6d 3571 set_parm_rtl (parm, tmp);
7e41ffa2 3572
9771b263
DN
3573 real = DECL_INCOMING_RTL (fnargs[i]);
3574 imag = DECL_INCOMING_RTL (fnargs[i + 1]);
6071dc7f
RH
3575 if (inner != GET_MODE (real))
3576 {
3577 real = gen_lowpart_SUBREG (inner, real);
3578 imag = gen_lowpart_SUBREG (inner, imag);
6f086dfc 3579 }
6071dc7f 3580 tmp = gen_rtx_CONCAT (DECL_MODE (parm), real, imag);
5141868d 3581 set_decl_incoming_rtl (parm, tmp, false);
3b3f318a 3582 i++;
6f086dfc 3583 }
6f086dfc 3584 }
6071dc7f
RH
3585}
3586
3587/* Assign RTL expressions to the function's parameters. This may involve
3588 copying them into registers and using those registers as the DECL_RTL. */
3589
6fe79279 3590static void
6071dc7f
RH
3591assign_parms (tree fndecl)
3592{
3593 struct assign_parm_data_all all;
3b3f318a 3594 tree parm;
9771b263 3595 vec<tree> fnargs;
31db0fe0 3596 unsigned i;
6f086dfc 3597
38173d38 3598 crtl->args.internal_arg_pointer
150cdc9e 3599 = targetm.calls.internal_arg_pointer ();
6071dc7f
RH
3600
3601 assign_parms_initialize_all (&all);
3602 fnargs = assign_parms_augmented_arg_list (&all);
3603
9771b263 3604 FOR_EACH_VEC_ELT (fnargs, i, parm)
ded9bf77 3605 {
6071dc7f
RH
3606 struct assign_parm_data_one data;
3607
3608 /* Extract the type of PARM; adjust it according to ABI. */
3609 assign_parm_find_data_types (&all, parm, &data);
3610
3611 /* Early out for errors and void parameters. */
3612 if (data.passed_mode == VOIDmode)
ded9bf77 3613 {
6071dc7f
RH
3614 SET_DECL_RTL (parm, const0_rtx);
3615 DECL_INCOMING_RTL (parm) = DECL_RTL (parm);
3616 continue;
3617 }
196c42cd 3618
2e3f842f
L
3619 /* Estimate stack alignment from parameter alignment. */
3620 if (SUPPORTS_STACK_ALIGNMENT)
3621 {
c2ed6cf8
NF
3622 unsigned int align
3623 = targetm.calls.function_arg_boundary (data.promoted_mode,
3624 data.passed_type);
ae58e548
JJ
3625 align = MINIMUM_ALIGNMENT (data.passed_type, data.promoted_mode,
3626 align);
2e3f842f 3627 if (TYPE_ALIGN (data.nominal_type) > align)
ae58e548
JJ
3628 align = MINIMUM_ALIGNMENT (data.nominal_type,
3629 TYPE_MODE (data.nominal_type),
3630 TYPE_ALIGN (data.nominal_type));
2e3f842f
L
3631 if (crtl->stack_alignment_estimated < align)
3632 {
3633 gcc_assert (!crtl->stack_realign_processed);
3634 crtl->stack_alignment_estimated = align;
3635 }
3636 }
b8698a0f 3637
6071dc7f
RH
3638 /* Find out where the parameter arrives in this function. */
3639 assign_parm_find_entry_rtl (&all, &data);
3640
3641 /* Find out where stack space for this parameter might be. */
3642 if (assign_parm_is_stack_parm (&all, &data))
3643 {
3644 assign_parm_find_stack_rtl (parm, &data);
3645 assign_parm_adjust_entry_rtl (&data);
ded9bf77 3646 }
6071dc7f 3647 /* Record permanently how this parm was passed. */
a82ff31f
JJ
3648 if (data.passed_pointer)
3649 {
3650 rtx incoming_rtl
3651 = gen_rtx_MEM (TYPE_MODE (TREE_TYPE (data.passed_type)),
3652 data.entry_parm);
3653 set_decl_incoming_rtl (parm, incoming_rtl, true);
3654 }
3655 else
3656 set_decl_incoming_rtl (parm, data.entry_parm, false);
6071dc7f 3657
f11a7b6d 3658 assign_parm_adjust_stack_rtl (&data);
1f9ceff1 3659
31db0fe0
ML
3660 if (assign_parm_setup_block_p (&data))
3661 assign_parm_setup_block (&all, parm, &data);
3662 else if (data.passed_pointer || use_register_for_decl (parm))
3663 assign_parm_setup_reg (&all, parm, &data);
d5e254e1 3664 else
31db0fe0 3665 assign_parm_setup_stack (&all, parm, &data);
d5e254e1
IE
3666
3667 if (cfun->stdarg && !DECL_CHAIN (parm))
31db0fe0 3668 assign_parms_setup_varargs (&all, &data, false);
d5e254e1 3669
6071dc7f 3670 /* Update info on where next arg arrives in registers. */
d5cc9181 3671 targetm.calls.function_arg_advance (all.args_so_far, data.promoted_mode,
3c07301f 3672 data.passed_type, data.named_arg);
ded9bf77
AH
3673 }
3674
3b3f318a 3675 if (targetm.calls.split_complex_arg)
6ccd356e 3676 assign_parms_unsplit_complex (&all, fnargs);
6071dc7f 3677
9771b263 3678 fnargs.release ();
3b3f318a 3679
3412b298
JW
3680 /* Output all parameter conversion instructions (possibly including calls)
3681 now that all parameters have been copied out of hard registers. */
bb27eeda 3682 emit_insn (all.first_conversion_insn);
3412b298 3683
2e3f842f
L
3684 /* Estimate reload stack alignment from scalar return mode. */
3685 if (SUPPORTS_STACK_ALIGNMENT)
3686 {
3687 if (DECL_RESULT (fndecl))
3688 {
3689 tree type = TREE_TYPE (DECL_RESULT (fndecl));
ef4bddc2 3690 machine_mode mode = TYPE_MODE (type);
2e3f842f
L
3691
3692 if (mode != BLKmode
3693 && mode != VOIDmode
3694 && !AGGREGATE_TYPE_P (type))
3695 {
3696 unsigned int align = GET_MODE_ALIGNMENT (mode);
3697 if (crtl->stack_alignment_estimated < align)
3698 {
3699 gcc_assert (!crtl->stack_realign_processed);
3700 crtl->stack_alignment_estimated = align;
3701 }
3702 }
b8698a0f 3703 }
2e3f842f
L
3704 }
3705
b36a8cc2
OH
3706 /* If we are receiving a struct value address as the first argument, set up
3707 the RTL for the function result. As this might require code to convert
3708 the transmitted address to Pmode, we do this here to ensure that possible
3709 preliminary conversions of the address have been emitted already. */
6071dc7f 3710 if (all.function_result_decl)
b36a8cc2 3711 {
6071dc7f
RH
3712 tree result = DECL_RESULT (current_function_decl);
3713 rtx addr = DECL_RTL (all.function_result_decl);
b36a8cc2 3714 rtx x;
fa8db1f7 3715
cc77ae10 3716 if (DECL_BY_REFERENCE (result))
8dcfef8f
AO
3717 {
3718 SET_DECL_VALUE_EXPR (result, all.function_result_decl);
3719 x = addr;
3720 }
cc77ae10
JM
3721 else
3722 {
8dcfef8f
AO
3723 SET_DECL_VALUE_EXPR (result,
3724 build1 (INDIRECT_REF, TREE_TYPE (result),
3725 all.function_result_decl));
cc77ae10
JM
3726 addr = convert_memory_address (Pmode, addr);
3727 x = gen_rtx_MEM (DECL_MODE (result), addr);
3728 set_mem_attributes (x, result, 1);
3729 }
8dcfef8f
AO
3730
3731 DECL_HAS_VALUE_EXPR_P (result) = 1;
3732
f11a7b6d 3733 set_parm_rtl (result, x);
b36a8cc2
OH
3734 }
3735
53c428d0 3736 /* We have aligned all the args, so add space for the pretend args. */
38173d38 3737 crtl->args.pretend_args_size = all.pretend_args_size;
6071dc7f 3738 all.stack_args_size.constant += all.extra_pretend_bytes;
38173d38 3739 crtl->args.size = all.stack_args_size.constant;
6f086dfc
RS
3740
3741 /* Adjust function incoming argument size for alignment and
3742 minimum length. */
3743
a20c5714
RS
3744 crtl->args.size = upper_bound (crtl->args.size, all.reg_parm_stack_space);
3745 crtl->args.size = aligned_upper_bound (crtl->args.size,
3746 PARM_BOUNDARY / BITS_PER_UNIT);
4433e339 3747
6dad9361
TS
3748 if (ARGS_GROW_DOWNWARD)
3749 {
3750 crtl->args.arg_offset_rtx
a20c5714
RS
3751 = (all.stack_args_size.var == 0
3752 ? gen_int_mode (-all.stack_args_size.constant, Pmode)
6dad9361
TS
3753 : expand_expr (size_diffop (all.stack_args_size.var,
3754 size_int (-all.stack_args_size.constant)),
3755 NULL_RTX, VOIDmode, EXPAND_NORMAL));
3756 }
3757 else
3758 crtl->args.arg_offset_rtx = ARGS_SIZE_RTX (all.stack_args_size);
6f086dfc
RS
3759
3760 /* See how many bytes, if any, of its args a function should try to pop
3761 on return. */
3762
079e7538
NF
3763 crtl->args.pops_args = targetm.calls.return_pops_args (fndecl,
3764 TREE_TYPE (fndecl),
3765 crtl->args.size);
6f086dfc 3766
3b69d50e
RK
3767 /* For stdarg.h function, save info about
3768 regs and stack space used by the named args. */
6f086dfc 3769
d5cc9181 3770 crtl->args.info = all.args_so_far_v;
6f086dfc
RS
3771
3772 /* Set the rtx used for the function return value. Put this in its
3773 own variable so any optimizers that need this information don't have
3774 to include tree.h. Do this here so it gets done when an inlined
3775 function gets output. */
3776
38173d38 3777 crtl->return_rtx
19e7881c
MM
3778 = (DECL_RTL_SET_P (DECL_RESULT (fndecl))
3779 ? DECL_RTL (DECL_RESULT (fndecl)) : NULL_RTX);
ce5e43d0
JJ
3780
3781 /* If scalar return value was computed in a pseudo-reg, or was a named
3782 return value that got dumped to the stack, copy that to the hard
3783 return register. */
3784 if (DECL_RTL_SET_P (DECL_RESULT (fndecl)))
3785 {
3786 tree decl_result = DECL_RESULT (fndecl);
3787 rtx decl_rtl = DECL_RTL (decl_result);
3788
3789 if (REG_P (decl_rtl)
3790 ? REGNO (decl_rtl) >= FIRST_PSEUDO_REGISTER
3791 : DECL_REGISTER (decl_result))
3792 {
3793 rtx real_decl_rtl;
3794
1d636cc6
RG
3795 real_decl_rtl = targetm.calls.function_value (TREE_TYPE (decl_result),
3796 fndecl, true);
ce5e43d0 3797 REG_FUNCTION_VALUE_P (real_decl_rtl) = 1;
38173d38 3798 /* The delay slot scheduler assumes that crtl->return_rtx
ce5e43d0
JJ
3799 holds the hard register containing the return value, not a
3800 temporary pseudo. */
38173d38 3801 crtl->return_rtx = real_decl_rtl;
ce5e43d0
JJ
3802 }
3803 }
6f086dfc 3804}
4744afba
RH
3805
3806/* A subroutine of gimplify_parameters, invoked via walk_tree.
3807 For all seen types, gimplify their sizes. */
3808
3809static tree
3810gimplify_parm_type (tree *tp, int *walk_subtrees, void *data)
3811{
3812 tree t = *tp;
3813
3814 *walk_subtrees = 0;
3815 if (TYPE_P (t))
3816 {
3817 if (POINTER_TYPE_P (t))
3818 *walk_subtrees = 1;
ad50bc8d
RH
3819 else if (TYPE_SIZE (t) && !TREE_CONSTANT (TYPE_SIZE (t))
3820 && !TYPE_SIZES_GIMPLIFIED (t))
4744afba 3821 {
726a989a 3822 gimplify_type_sizes (t, (gimple_seq *) data);
4744afba
RH
3823 *walk_subtrees = 1;
3824 }
3825 }
3826
3827 return NULL;
3828}
3829
3830/* Gimplify the parameter list for current_function_decl. This involves
3831 evaluating SAVE_EXPRs of variable sized parameters and generating code
726a989a
RB
3832 to implement callee-copies reference parameters. Returns a sequence of
3833 statements to add to the beginning of the function. */
4744afba 3834
726a989a 3835gimple_seq
6aee2fd0 3836gimplify_parameters (gimple_seq *cleanup)
4744afba
RH
3837{
3838 struct assign_parm_data_all all;
3b3f318a 3839 tree parm;
726a989a 3840 gimple_seq stmts = NULL;
9771b263 3841 vec<tree> fnargs;
3b3f318a 3842 unsigned i;
4744afba
RH
3843
3844 assign_parms_initialize_all (&all);
3845 fnargs = assign_parms_augmented_arg_list (&all);
3846
9771b263 3847 FOR_EACH_VEC_ELT (fnargs, i, parm)
4744afba
RH
3848 {
3849 struct assign_parm_data_one data;
3850
3851 /* Extract the type of PARM; adjust it according to ABI. */
3852 assign_parm_find_data_types (&all, parm, &data);
3853
3854 /* Early out for errors and void parameters. */
3855 if (data.passed_mode == VOIDmode || DECL_SIZE (parm) == NULL)
3856 continue;
3857
3858 /* Update info on where next arg arrives in registers. */
d5cc9181 3859 targetm.calls.function_arg_advance (all.args_so_far, data.promoted_mode,
3c07301f 3860 data.passed_type, data.named_arg);
4744afba
RH
3861
3862 /* ??? Once upon a time variable_size stuffed parameter list
3863 SAVE_EXPRs (amongst others) onto a pending sizes list. This
3864 turned out to be less than manageable in the gimple world.
3865 Now we have to hunt them down ourselves. */
3866 walk_tree_without_duplicates (&data.passed_type,
3867 gimplify_parm_type, &stmts);
3868
b38f3813 3869 if (TREE_CODE (DECL_SIZE_UNIT (parm)) != INTEGER_CST)
4744afba
RH
3870 {
3871 gimplify_one_sizepos (&DECL_SIZE (parm), &stmts);
3872 gimplify_one_sizepos (&DECL_SIZE_UNIT (parm), &stmts);
3873 }
3874
3875 if (data.passed_pointer)
3876 {
3877 tree type = TREE_TYPE (data.passed_type);
d5cc9181 3878 if (reference_callee_copied (&all.args_so_far_v, TYPE_MODE (type),
4744afba
RH
3879 type, data.named_arg))
3880 {
3881 tree local, t;
3882
b38f3813 3883 /* For constant-sized objects, this is trivial; for
4744afba 3884 variable-sized objects, we have to play games. */
b38f3813
EB
3885 if (TREE_CODE (DECL_SIZE_UNIT (parm)) == INTEGER_CST
3886 && !(flag_stack_check == GENERIC_STACK_CHECK
3887 && compare_tree_int (DECL_SIZE_UNIT (parm),
3888 STACK_CHECK_MAX_VAR_SIZE) > 0))
4744afba 3889 {
5dac1dae 3890 local = create_tmp_var (type, get_name (parm));
4744afba 3891 DECL_IGNORED_P (local) = 0;
04487a2f
JJ
3892 /* If PARM was addressable, move that flag over
3893 to the local copy, as its address will be taken,
37609bf0
RG
3894 not the PARMs. Keep the parms address taken
3895 as we'll query that flag during gimplification. */
04487a2f 3896 if (TREE_ADDRESSABLE (parm))
37609bf0 3897 TREE_ADDRESSABLE (local) = 1;
5dac1dae
JJ
3898 else if (TREE_CODE (type) == COMPLEX_TYPE
3899 || TREE_CODE (type) == VECTOR_TYPE)
3900 DECL_GIMPLE_REG_P (local) = 1;
6aee2fd0
JJ
3901
3902 if (!is_gimple_reg (local)
3903 && flag_stack_reuse != SR_NONE)
3904 {
3905 tree clobber = build_constructor (type, NULL);
3906 gimple *clobber_stmt;
3907 TREE_THIS_VOLATILE (clobber) = 1;
3908 clobber_stmt = gimple_build_assign (local, clobber);
3909 gimple_seq_add_stmt (cleanup, clobber_stmt);
3910 }
4744afba
RH
3911 }
3912 else
3913 {
5039610b 3914 tree ptr_type, addr;
4744afba
RH
3915
3916 ptr_type = build_pointer_type (type);
c98b08ff 3917 addr = create_tmp_reg (ptr_type, get_name (parm));
4744afba
RH
3918 DECL_IGNORED_P (addr) = 0;
3919 local = build_fold_indirect_ref (addr);
3920
9e878cf1
EB
3921 t = build_alloca_call_expr (DECL_SIZE_UNIT (parm),
3922 DECL_ALIGN (parm),
3923 max_int_size_in_bytes (type));
d3c12306 3924 /* The call has been built for a variable-sized object. */
63d2a353 3925 CALL_ALLOCA_FOR_VAR_P (t) = 1;
4744afba 3926 t = fold_convert (ptr_type, t);
726a989a 3927 t = build2 (MODIFY_EXPR, TREE_TYPE (addr), addr, t);
4744afba
RH
3928 gimplify_and_add (t, &stmts);
3929 }
3930
726a989a 3931 gimplify_assign (local, parm, &stmts);
4744afba 3932
833b3afe
DB
3933 SET_DECL_VALUE_EXPR (parm, local);
3934 DECL_HAS_VALUE_EXPR_P (parm) = 1;
4744afba
RH
3935 }
3936 }
3937 }
3938
9771b263 3939 fnargs.release ();
3b3f318a 3940
4744afba
RH
3941 return stmts;
3942}
75dc3319 3943\f
6f086dfc
RS
3944/* Compute the size and offset from the start of the stacked arguments for a
3945 parm passed in mode PASSED_MODE and with type TYPE.
3946
3947 INITIAL_OFFSET_PTR points to the current offset into the stacked
3948 arguments.
3949
e7949876
AM
3950 The starting offset and size for this parm are returned in
3951 LOCATE->OFFSET and LOCATE->SIZE, respectively. When IN_REGS is
3952 nonzero, the offset is that of stack slot, which is returned in
3953 LOCATE->SLOT_OFFSET. LOCATE->ALIGNMENT_PAD is the amount of
3954 padding required from the initial offset ptr to the stack slot.
6f086dfc 3955
cc2902df 3956 IN_REGS is nonzero if the argument will be passed in registers. It will
6f086dfc
RS
3957 never be set if REG_PARM_STACK_SPACE is not defined.
3958
2e4ceca5
UW
3959 REG_PARM_STACK_SPACE is the number of bytes of stack space reserved
3960 for arguments which are passed in registers.
3961
6f086dfc
RS
3962 FNDECL is the function in which the argument was defined.
3963
3964 There are two types of rounding that are done. The first, controlled by
c2ed6cf8
NF
3965 TARGET_FUNCTION_ARG_BOUNDARY, forces the offset from the start of the
3966 argument list to be aligned to the specific boundary (in bits). This
3967 rounding affects the initial and starting offsets, but not the argument
3968 size.
6f086dfc 3969
76b0cbf8 3970 The second, controlled by TARGET_FUNCTION_ARG_PADDING and PARM_BOUNDARY,
6f086dfc
RS
3971 optionally rounds the size of the parm to PARM_BOUNDARY. The
3972 initial offset is not affected by this rounding, while the size always
3973 is and the starting offset may be. */
3974
e7949876
AM
3975/* LOCATE->OFFSET will be negative for ARGS_GROW_DOWNWARD case;
3976 INITIAL_OFFSET_PTR is positive because locate_and_pad_parm's
6f086dfc 3977 callers pass in the total size of args so far as
e7949876 3978 INITIAL_OFFSET_PTR. LOCATE->SIZE is always positive. */
6f086dfc 3979
6f086dfc 3980void
ef4bddc2 3981locate_and_pad_parm (machine_mode passed_mode, tree type, int in_regs,
2e4ceca5
UW
3982 int reg_parm_stack_space, int partial,
3983 tree fndecl ATTRIBUTE_UNUSED,
fa8db1f7
AJ
3984 struct args_size *initial_offset_ptr,
3985 struct locate_and_pad_arg_data *locate)
6f086dfc 3986{
e7949876 3987 tree sizetree;
76b0cbf8 3988 pad_direction where_pad;
123148b5 3989 unsigned int boundary, round_boundary;
e7949876 3990 int part_size_in_regs;
6f086dfc 3991
6f086dfc
RS
3992 /* If we have found a stack parm before we reach the end of the
3993 area reserved for registers, skip that area. */
3994 if (! in_regs)
3995 {
6f086dfc
RS
3996 if (reg_parm_stack_space > 0)
3997 {
a20c5714
RS
3998 if (initial_offset_ptr->var
3999 || !ordered_p (initial_offset_ptr->constant,
4000 reg_parm_stack_space))
6f086dfc
RS
4001 {
4002 initial_offset_ptr->var
4003 = size_binop (MAX_EXPR, ARGS_SIZE_TREE (*initial_offset_ptr),
fed3cef0 4004 ssize_int (reg_parm_stack_space));
6f086dfc
RS
4005 initial_offset_ptr->constant = 0;
4006 }
a20c5714
RS
4007 else
4008 initial_offset_ptr->constant
4009 = ordered_max (initial_offset_ptr->constant,
4010 reg_parm_stack_space);
6f086dfc
RS
4011 }
4012 }
6f086dfc 4013
78a52f11 4014 part_size_in_regs = (reg_parm_stack_space == 0 ? partial : 0);
e7949876 4015
974aedcc
MP
4016 sizetree = (type
4017 ? arg_size_in_bytes (type)
4018 : size_int (GET_MODE_SIZE (passed_mode)));
76b0cbf8 4019 where_pad = targetm.calls.function_arg_padding (passed_mode, type);
c2ed6cf8 4020 boundary = targetm.calls.function_arg_boundary (passed_mode, type);
123148b5
BS
4021 round_boundary = targetm.calls.function_arg_round_boundary (passed_mode,
4022 type);
6e985040 4023 locate->where_pad = where_pad;
2e3f842f
L
4024
4025 /* Alignment can't exceed MAX_SUPPORTED_STACK_ALIGNMENT. */
4026 if (boundary > MAX_SUPPORTED_STACK_ALIGNMENT)
4027 boundary = MAX_SUPPORTED_STACK_ALIGNMENT;
4028
bfc45551 4029 locate->boundary = boundary;
6f086dfc 4030
2e3f842f
L
4031 if (SUPPORTS_STACK_ALIGNMENT)
4032 {
4033 /* stack_alignment_estimated can't change after stack has been
4034 realigned. */
4035 if (crtl->stack_alignment_estimated < boundary)
4036 {
4037 if (!crtl->stack_realign_processed)
4038 crtl->stack_alignment_estimated = boundary;
4039 else
4040 {
4041 /* If stack is realigned and stack alignment value
4042 hasn't been finalized, it is OK not to increase
4043 stack_alignment_estimated. The bigger alignment
4044 requirement is recorded in stack_alignment_needed
4045 below. */
4046 gcc_assert (!crtl->stack_realign_finalized
4047 && crtl->stack_realign_needed);
4048 }
4049 }
4050 }
4051
c7e777b5
RH
4052 /* Remember if the outgoing parameter requires extra alignment on the
4053 calling function side. */
cb91fab0
JH
4054 if (crtl->stack_alignment_needed < boundary)
4055 crtl->stack_alignment_needed = boundary;
2e3f842f
L
4056 if (crtl->preferred_stack_boundary < boundary)
4057 crtl->preferred_stack_boundary = boundary;
c7e777b5 4058
6dad9361
TS
4059 if (ARGS_GROW_DOWNWARD)
4060 {
4061 locate->slot_offset.constant = -initial_offset_ptr->constant;
4062 if (initial_offset_ptr->var)
4063 locate->slot_offset.var = size_binop (MINUS_EXPR, ssize_int (0),
4064 initial_offset_ptr->var);
4065
a589e68f
DM
4066 {
4067 tree s2 = sizetree;
76b0cbf8 4068 if (where_pad != PAD_NONE
a589e68f
DM
4069 && (!tree_fits_uhwi_p (sizetree)
4070 || (tree_to_uhwi (sizetree) * BITS_PER_UNIT) % round_boundary))
4071 s2 = round_up (s2, round_boundary / BITS_PER_UNIT);
4072 SUB_PARM_SIZE (locate->slot_offset, s2);
4073 }
6dad9361
TS
4074
4075 locate->slot_offset.constant += part_size_in_regs;
4076
4077 if (!in_regs || reg_parm_stack_space > 0)
4078 pad_to_arg_alignment (&locate->slot_offset, boundary,
4079 &locate->alignment_pad);
4080
4081 locate->size.constant = (-initial_offset_ptr->constant
4082 - locate->slot_offset.constant);
4083 if (initial_offset_ptr->var)
4084 locate->size.var = size_binop (MINUS_EXPR,
4085 size_binop (MINUS_EXPR,
4086 ssize_int (0),
4087 initial_offset_ptr->var),
4088 locate->slot_offset.var);
4089
4090 /* Pad_below needs the pre-rounded size to know how much to pad
4091 below. */
4092 locate->offset = locate->slot_offset;
76b0cbf8 4093 if (where_pad == PAD_DOWNWARD)
6dad9361
TS
4094 pad_below (&locate->offset, passed_mode, sizetree);
4095
4096 }
4097 else
4098 {
4099 if (!in_regs || reg_parm_stack_space > 0)
4100 pad_to_arg_alignment (initial_offset_ptr, boundary,
4101 &locate->alignment_pad);
4102 locate->slot_offset = *initial_offset_ptr;
6f086dfc
RS
4103
4104#ifdef PUSH_ROUNDING
6dad9361
TS
4105 if (passed_mode != BLKmode)
4106 sizetree = size_int (PUSH_ROUNDING (TREE_INT_CST_LOW (sizetree)));
6f086dfc
RS
4107#endif
4108
6dad9361
TS
4109 /* Pad_below needs the pre-rounded size to know how much to pad below
4110 so this must be done before rounding up. */
4111 locate->offset = locate->slot_offset;
76b0cbf8 4112 if (where_pad == PAD_DOWNWARD)
6dad9361 4113 pad_below (&locate->offset, passed_mode, sizetree);
d4b0a7a0 4114
76b0cbf8 4115 if (where_pad != PAD_NONE
6dad9361
TS
4116 && (!tree_fits_uhwi_p (sizetree)
4117 || (tree_to_uhwi (sizetree) * BITS_PER_UNIT) % round_boundary))
4118 sizetree = round_up (sizetree, round_boundary / BITS_PER_UNIT);
6f086dfc 4119
6dad9361 4120 ADD_PARM_SIZE (locate->size, sizetree);
e7949876 4121
6dad9361
TS
4122 locate->size.constant -= part_size_in_regs;
4123 }
099590dc 4124
870118b7
RS
4125 locate->offset.constant
4126 += targetm.calls.function_arg_offset (passed_mode, type);
6f086dfc
RS
4127}
4128
e16c591a
RS
4129/* Round the stack offset in *OFFSET_PTR up to a multiple of BOUNDARY.
4130 BOUNDARY is measured in bits, but must be a multiple of a storage unit. */
4131
6f086dfc 4132static void
fa8db1f7
AJ
4133pad_to_arg_alignment (struct args_size *offset_ptr, int boundary,
4134 struct args_size *alignment_pad)
6f086dfc 4135{
a544cfd2 4136 tree save_var = NULL_TREE;
a20c5714 4137 poly_int64 save_constant = 0;
a751cd5b 4138 int boundary_in_bytes = boundary / BITS_PER_UNIT;
a20c5714 4139 poly_int64 sp_offset = STACK_POINTER_OFFSET;
a594a19c
GK
4140
4141#ifdef SPARC_STACK_BOUNDARY_HACK
2358ff91
EB
4142 /* ??? The SPARC port may claim a STACK_BOUNDARY higher than
4143 the real alignment of %sp. However, when it does this, the
4144 alignment of %sp+STACK_POINTER_OFFSET is STACK_BOUNDARY. */
a594a19c
GK
4145 if (SPARC_STACK_BOUNDARY_HACK)
4146 sp_offset = 0;
4147#endif
4fc026cd 4148
6f6b8f81 4149 if (boundary > PARM_BOUNDARY)
4fc026cd
CM
4150 {
4151 save_var = offset_ptr->var;
4152 save_constant = offset_ptr->constant;
4153 }
4154
4155 alignment_pad->var = NULL_TREE;
4156 alignment_pad->constant = 0;
4fc026cd 4157
6f086dfc
RS
4158 if (boundary > BITS_PER_UNIT)
4159 {
a20c5714
RS
4160 int misalign;
4161 if (offset_ptr->var
4162 || !known_misalignment (offset_ptr->constant + sp_offset,
4163 boundary_in_bytes, &misalign))
6f086dfc 4164 {
a594a19c
GK
4165 tree sp_offset_tree = ssize_int (sp_offset);
4166 tree offset = size_binop (PLUS_EXPR,
4167 ARGS_SIZE_TREE (*offset_ptr),
4168 sp_offset_tree);
6dad9361
TS
4169 tree rounded;
4170 if (ARGS_GROW_DOWNWARD)
4171 rounded = round_down (offset, boundary / BITS_PER_UNIT);
4172 else
4173 rounded = round_up (offset, boundary / BITS_PER_UNIT);
a594a19c
GK
4174
4175 offset_ptr->var = size_binop (MINUS_EXPR, rounded, sp_offset_tree);
e7949876
AM
4176 /* ARGS_SIZE_TREE includes constant term. */
4177 offset_ptr->constant = 0;
6f6b8f81 4178 if (boundary > PARM_BOUNDARY)
dd3f0101 4179 alignment_pad->var = size_binop (MINUS_EXPR, offset_ptr->var,
fed3cef0 4180 save_var);
6f086dfc
RS
4181 }
4182 else
718fe406 4183 {
a20c5714
RS
4184 if (ARGS_GROW_DOWNWARD)
4185 offset_ptr->constant -= misalign;
4186 else
4187 offset_ptr->constant += -misalign & (boundary_in_bytes - 1);
6dad9361 4188
a20c5714
RS
4189 if (boundary > PARM_BOUNDARY)
4190 alignment_pad->constant = offset_ptr->constant - save_constant;
718fe406 4191 }
6f086dfc
RS
4192 }
4193}
4194
4195static void
ef4bddc2 4196pad_below (struct args_size *offset_ptr, machine_mode passed_mode, tree sizetree)
6f086dfc 4197{
b66fd4fc 4198 unsigned int align = PARM_BOUNDARY / BITS_PER_UNIT;
cf098191
RS
4199 int misalign;
4200 if (passed_mode != BLKmode
4201 && known_misalignment (GET_MODE_SIZE (passed_mode), align, &misalign))
4202 offset_ptr->constant += -misalign & (align - 1);
6f086dfc
RS
4203 else
4204 {
4205 if (TREE_CODE (sizetree) != INTEGER_CST
b66fd4fc 4206 || (TREE_INT_CST_LOW (sizetree) & (align - 1)) != 0)
6f086dfc
RS
4207 {
4208 /* Round the size up to multiple of PARM_BOUNDARY bits. */
b66fd4fc 4209 tree s2 = round_up (sizetree, align);
6f086dfc
RS
4210 /* Add it in. */
4211 ADD_PARM_SIZE (*offset_ptr, s2);
4212 SUB_PARM_SIZE (*offset_ptr, sizetree);
4213 }
4214 }
4215}
6f086dfc 4216\f
6f086dfc 4217
6fb5fa3c
DB
4218/* True if register REGNO was alive at a place where `setjmp' was
4219 called and was set more than once or is an argument. Such regs may
4220 be clobbered by `longjmp'. */
4221
4222static bool
4223regno_clobbered_at_setjmp (bitmap setjmp_crosses, int regno)
4224{
4225 /* There appear to be cases where some local vars never reach the
4226 backend but have bogus regnos. */
4227 if (regno >= max_reg_num ())
4228 return false;
4229
4230 return ((REG_N_SETS (regno) > 1
fefa31b5
DM
4231 || REGNO_REG_SET_P (df_get_live_out (ENTRY_BLOCK_PTR_FOR_FN (cfun)),
4232 regno))
6fb5fa3c
DB
4233 && REGNO_REG_SET_P (setjmp_crosses, regno));
4234}
4235
4236/* Walk the tree of blocks describing the binding levels within a
4237 function and warn about variables the might be killed by setjmp or
4238 vfork. This is done after calling flow_analysis before register
4239 allocation since that will clobber the pseudo-regs to hard
4240 regs. */
4241
4242static void
4243setjmp_vars_warning (bitmap setjmp_crosses, tree block)
6f086dfc 4244{
b3694847 4245 tree decl, sub;
6de9cd9a 4246
910ad8de 4247 for (decl = BLOCK_VARS (block); decl; decl = DECL_CHAIN (decl))
6f086dfc 4248 {
8813a647 4249 if (VAR_P (decl)
bc41842b 4250 && DECL_RTL_SET_P (decl)
f8cfc6aa 4251 && REG_P (DECL_RTL (decl))
6fb5fa3c 4252 && regno_clobbered_at_setjmp (setjmp_crosses, REGNO (DECL_RTL (decl))))
b8698a0f 4253 warning (OPT_Wclobbered, "variable %q+D might be clobbered by"
2b001724 4254 " %<longjmp%> or %<vfork%>", decl);
6f086dfc 4255 }
6de9cd9a 4256
87caf699 4257 for (sub = BLOCK_SUBBLOCKS (block); sub; sub = BLOCK_CHAIN (sub))
6fb5fa3c 4258 setjmp_vars_warning (setjmp_crosses, sub);
6f086dfc
RS
4259}
4260
6de9cd9a 4261/* Do the appropriate part of setjmp_vars_warning
6f086dfc
RS
4262 but for arguments instead of local variables. */
4263
6fb5fa3c
DB
4264static void
4265setjmp_args_warning (bitmap setjmp_crosses)
6f086dfc 4266{
b3694847 4267 tree decl;
6f086dfc 4268 for (decl = DECL_ARGUMENTS (current_function_decl);
910ad8de 4269 decl; decl = DECL_CHAIN (decl))
6f086dfc 4270 if (DECL_RTL (decl) != 0
f8cfc6aa 4271 && REG_P (DECL_RTL (decl))
6fb5fa3c 4272 && regno_clobbered_at_setjmp (setjmp_crosses, REGNO (DECL_RTL (decl))))
b8698a0f 4273 warning (OPT_Wclobbered,
2b001724 4274 "argument %q+D might be clobbered by %<longjmp%> or %<vfork%>",
dee15844 4275 decl);
6f086dfc
RS
4276}
4277
6fb5fa3c
DB
4278/* Generate warning messages for variables live across setjmp. */
4279
b8698a0f 4280void
6fb5fa3c
DB
4281generate_setjmp_warnings (void)
4282{
4283 bitmap setjmp_crosses = regstat_get_setjmp_crosses ();
4284
0cae8d31 4285 if (n_basic_blocks_for_fn (cfun) == NUM_FIXED_BLOCKS
6fb5fa3c
DB
4286 || bitmap_empty_p (setjmp_crosses))
4287 return;
4288
4289 setjmp_vars_warning (setjmp_crosses, DECL_INITIAL (current_function_decl));
4290 setjmp_args_warning (setjmp_crosses);
4291}
4292
6f086dfc 4293\f
3373692b 4294/* Reverse the order of elements in the fragment chain T of blocks,
1e3c1d95
JJ
4295 and return the new head of the chain (old last element).
4296 In addition to that clear BLOCK_SAME_RANGE flags when needed
4297 and adjust BLOCK_SUPERCONTEXT from the super fragment to
4298 its super fragment origin. */
3373692b
JJ
4299
4300static tree
4301block_fragments_nreverse (tree t)
4302{
1e3c1d95
JJ
4303 tree prev = 0, block, next, prev_super = 0;
4304 tree super = BLOCK_SUPERCONTEXT (t);
4305 if (BLOCK_FRAGMENT_ORIGIN (super))
4306 super = BLOCK_FRAGMENT_ORIGIN (super);
3373692b
JJ
4307 for (block = t; block; block = next)
4308 {
4309 next = BLOCK_FRAGMENT_CHAIN (block);
4310 BLOCK_FRAGMENT_CHAIN (block) = prev;
1e3c1d95
JJ
4311 if ((prev && !BLOCK_SAME_RANGE (prev))
4312 || (BLOCK_FRAGMENT_CHAIN (BLOCK_SUPERCONTEXT (block))
4313 != prev_super))
4314 BLOCK_SAME_RANGE (block) = 0;
4315 prev_super = BLOCK_SUPERCONTEXT (block);
4316 BLOCK_SUPERCONTEXT (block) = super;
3373692b
JJ
4317 prev = block;
4318 }
1e3c1d95
JJ
4319 t = BLOCK_FRAGMENT_ORIGIN (t);
4320 if (BLOCK_FRAGMENT_CHAIN (BLOCK_SUPERCONTEXT (t))
4321 != prev_super)
4322 BLOCK_SAME_RANGE (t) = 0;
4323 BLOCK_SUPERCONTEXT (t) = super;
3373692b
JJ
4324 return prev;
4325}
4326
4327/* Reverse the order of elements in the chain T of blocks,
4328 and return the new head of the chain (old last element).
4329 Also do the same on subblocks and reverse the order of elements
4330 in BLOCK_FRAGMENT_CHAIN as well. */
4331
4332static tree
4333blocks_nreverse_all (tree t)
4334{
4335 tree prev = 0, block, next;
4336 for (block = t; block; block = next)
4337 {
4338 next = BLOCK_CHAIN (block);
4339 BLOCK_CHAIN (block) = prev;
3373692b
JJ
4340 if (BLOCK_FRAGMENT_CHAIN (block)
4341 && BLOCK_FRAGMENT_ORIGIN (block) == NULL_TREE)
1e3c1d95
JJ
4342 {
4343 BLOCK_FRAGMENT_CHAIN (block)
4344 = block_fragments_nreverse (BLOCK_FRAGMENT_CHAIN (block));
4345 if (!BLOCK_SAME_RANGE (BLOCK_FRAGMENT_CHAIN (block)))
4346 BLOCK_SAME_RANGE (block) = 0;
4347 }
4348 BLOCK_SUBBLOCKS (block) = blocks_nreverse_all (BLOCK_SUBBLOCKS (block));
3373692b
JJ
4349 prev = block;
4350 }
4351 return prev;
4352}
4353
4354
a20612aa
RH
4355/* Identify BLOCKs referenced by more than one NOTE_INSN_BLOCK_{BEG,END},
4356 and create duplicate blocks. */
4357/* ??? Need an option to either create block fragments or to create
4358 abstract origin duplicates of a source block. It really depends
4359 on what optimization has been performed. */
467456d0 4360
116eebd6 4361void
fa8db1f7 4362reorder_blocks (void)
467456d0 4363{
116eebd6 4364 tree block = DECL_INITIAL (current_function_decl);
467456d0 4365
1a4450c7 4366 if (block == NULL_TREE)
116eebd6 4367 return;
fc289cd1 4368
00f96dc9 4369 auto_vec<tree, 10> block_stack;
18c038b9 4370
a20612aa 4371 /* Reset the TREE_ASM_WRITTEN bit for all blocks. */
6de9cd9a 4372 clear_block_marks (block);
a20612aa 4373
116eebd6
MM
4374 /* Prune the old trees away, so that they don't get in the way. */
4375 BLOCK_SUBBLOCKS (block) = NULL_TREE;
4376 BLOCK_CHAIN (block) = NULL_TREE;
fc289cd1 4377
a20612aa 4378 /* Recreate the block tree from the note nesting. */
116eebd6 4379 reorder_blocks_1 (get_insns (), block, &block_stack);
3373692b 4380 BLOCK_SUBBLOCKS (block) = blocks_nreverse_all (BLOCK_SUBBLOCKS (block));
467456d0
RS
4381}
4382
a20612aa 4383/* Helper function for reorder_blocks. Reset TREE_ASM_WRITTEN. */
0a1c58a2 4384
6de9cd9a
DN
4385void
4386clear_block_marks (tree block)
cc1fe44f 4387{
a20612aa 4388 while (block)
cc1fe44f 4389 {
a20612aa 4390 TREE_ASM_WRITTEN (block) = 0;
6de9cd9a 4391 clear_block_marks (BLOCK_SUBBLOCKS (block));
a20612aa 4392 block = BLOCK_CHAIN (block);
cc1fe44f
DD
4393 }
4394}
4395
0a1c58a2 4396static void
691fe203
DM
4397reorder_blocks_1 (rtx_insn *insns, tree current_block,
4398 vec<tree> *p_block_stack)
0a1c58a2 4399{
691fe203 4400 rtx_insn *insn;
1e3c1d95 4401 tree prev_beg = NULL_TREE, prev_end = NULL_TREE;
0a1c58a2
JL
4402
4403 for (insn = insns; insn; insn = NEXT_INSN (insn))
4404 {
4b4bf941 4405 if (NOTE_P (insn))
0a1c58a2 4406 {
a38e7aa5 4407 if (NOTE_KIND (insn) == NOTE_INSN_BLOCK_BEG)
0a1c58a2
JL
4408 {
4409 tree block = NOTE_BLOCK (insn);
51b7d006
DJ
4410 tree origin;
4411
3373692b
JJ
4412 gcc_assert (BLOCK_FRAGMENT_ORIGIN (block) == NULL_TREE);
4413 origin = block;
a20612aa 4414
1e3c1d95
JJ
4415 if (prev_end)
4416 BLOCK_SAME_RANGE (prev_end) = 0;
4417 prev_end = NULL_TREE;
4418
a20612aa
RH
4419 /* If we have seen this block before, that means it now
4420 spans multiple address regions. Create a new fragment. */
0a1c58a2
JL
4421 if (TREE_ASM_WRITTEN (block))
4422 {
a20612aa 4423 tree new_block = copy_node (block);
a20612aa 4424
1e3c1d95 4425 BLOCK_SAME_RANGE (new_block) = 0;
a20612aa
RH
4426 BLOCK_FRAGMENT_ORIGIN (new_block) = origin;
4427 BLOCK_FRAGMENT_CHAIN (new_block)
4428 = BLOCK_FRAGMENT_CHAIN (origin);
4429 BLOCK_FRAGMENT_CHAIN (origin) = new_block;
4430
4431 NOTE_BLOCK (insn) = new_block;
4432 block = new_block;
0a1c58a2 4433 }
a20612aa 4434
1e3c1d95
JJ
4435 if (prev_beg == current_block && prev_beg)
4436 BLOCK_SAME_RANGE (block) = 1;
4437
4438 prev_beg = origin;
4439
0a1c58a2
JL
4440 BLOCK_SUBBLOCKS (block) = 0;
4441 TREE_ASM_WRITTEN (block) = 1;
339a28b9
ZW
4442 /* When there's only one block for the entire function,
4443 current_block == block and we mustn't do this, it
4444 will cause infinite recursion. */
4445 if (block != current_block)
4446 {
1e3c1d95 4447 tree super;
51b7d006 4448 if (block != origin)
1e3c1d95
JJ
4449 gcc_assert (BLOCK_SUPERCONTEXT (origin) == current_block
4450 || BLOCK_FRAGMENT_ORIGIN (BLOCK_SUPERCONTEXT
4451 (origin))
4452 == current_block);
9771b263 4453 if (p_block_stack->is_empty ())
1e3c1d95
JJ
4454 super = current_block;
4455 else
4456 {
9771b263 4457 super = p_block_stack->last ();
1e3c1d95
JJ
4458 gcc_assert (super == current_block
4459 || BLOCK_FRAGMENT_ORIGIN (super)
4460 == current_block);
4461 }
4462 BLOCK_SUPERCONTEXT (block) = super;
339a28b9
ZW
4463 BLOCK_CHAIN (block) = BLOCK_SUBBLOCKS (current_block);
4464 BLOCK_SUBBLOCKS (current_block) = block;
51b7d006 4465 current_block = origin;
339a28b9 4466 }
9771b263 4467 p_block_stack->safe_push (block);
0a1c58a2 4468 }
a38e7aa5 4469 else if (NOTE_KIND (insn) == NOTE_INSN_BLOCK_END)
0a1c58a2 4470 {
9771b263 4471 NOTE_BLOCK (insn) = p_block_stack->pop ();
0a1c58a2 4472 current_block = BLOCK_SUPERCONTEXT (current_block);
1e3c1d95
JJ
4473 if (BLOCK_FRAGMENT_ORIGIN (current_block))
4474 current_block = BLOCK_FRAGMENT_ORIGIN (current_block);
4475 prev_beg = NULL_TREE;
4476 prev_end = BLOCK_SAME_RANGE (NOTE_BLOCK (insn))
4477 ? NOTE_BLOCK (insn) : NULL_TREE;
0a1c58a2
JL
4478 }
4479 }
1e3c1d95
JJ
4480 else
4481 {
4482 prev_beg = NULL_TREE;
4483 if (prev_end)
4484 BLOCK_SAME_RANGE (prev_end) = 0;
4485 prev_end = NULL_TREE;
4486 }
0a1c58a2
JL
4487 }
4488}
4489
467456d0
RS
4490/* Reverse the order of elements in the chain T of blocks,
4491 and return the new head of the chain (old last element). */
4492
6de9cd9a 4493tree
fa8db1f7 4494blocks_nreverse (tree t)
467456d0 4495{
3373692b
JJ
4496 tree prev = 0, block, next;
4497 for (block = t; block; block = next)
467456d0 4498 {
3373692b
JJ
4499 next = BLOCK_CHAIN (block);
4500 BLOCK_CHAIN (block) = prev;
4501 prev = block;
467456d0
RS
4502 }
4503 return prev;
4504}
4505
61e46a7d
NF
4506/* Concatenate two chains of blocks (chained through BLOCK_CHAIN)
4507 by modifying the last node in chain 1 to point to chain 2. */
4508
4509tree
4510block_chainon (tree op1, tree op2)
4511{
4512 tree t1;
4513
4514 if (!op1)
4515 return op2;
4516 if (!op2)
4517 return op1;
4518
4519 for (t1 = op1; BLOCK_CHAIN (t1); t1 = BLOCK_CHAIN (t1))
4520 continue;
4521 BLOCK_CHAIN (t1) = op2;
4522
4523#ifdef ENABLE_TREE_CHECKING
4524 {
4525 tree t2;
4526 for (t2 = op2; t2; t2 = BLOCK_CHAIN (t2))
4527 gcc_assert (t2 != t1);
4528 }
4529#endif
4530
4531 return op1;
4532}
4533
18c038b9
MM
4534/* Count the subblocks of the list starting with BLOCK. If VECTOR is
4535 non-NULL, list them all into VECTOR, in a depth-first preorder
4536 traversal of the block tree. Also clear TREE_ASM_WRITTEN in all
b2a59b15 4537 blocks. */
467456d0
RS
4538
4539static int
fa8db1f7 4540all_blocks (tree block, tree *vector)
467456d0 4541{
b2a59b15
MS
4542 int n_blocks = 0;
4543
a84efb51
JO
4544 while (block)
4545 {
4546 TREE_ASM_WRITTEN (block) = 0;
b2a59b15 4547
a84efb51
JO
4548 /* Record this block. */
4549 if (vector)
4550 vector[n_blocks] = block;
b2a59b15 4551
a84efb51 4552 ++n_blocks;
718fe406 4553
a84efb51
JO
4554 /* Record the subblocks, and their subblocks... */
4555 n_blocks += all_blocks (BLOCK_SUBBLOCKS (block),
4556 vector ? vector + n_blocks : 0);
4557 block = BLOCK_CHAIN (block);
4558 }
467456d0
RS
4559
4560 return n_blocks;
4561}
18c038b9
MM
4562
4563/* Return a vector containing all the blocks rooted at BLOCK. The
4564 number of elements in the vector is stored in N_BLOCKS_P. The
4565 vector is dynamically allocated; it is the caller's responsibility
4566 to call `free' on the pointer returned. */
718fe406 4567
18c038b9 4568static tree *
fa8db1f7 4569get_block_vector (tree block, int *n_blocks_p)
18c038b9
MM
4570{
4571 tree *block_vector;
4572
4573 *n_blocks_p = all_blocks (block, NULL);
5ed6ace5 4574 block_vector = XNEWVEC (tree, *n_blocks_p);
18c038b9
MM
4575 all_blocks (block, block_vector);
4576
4577 return block_vector;
4578}
4579
f83b236e 4580static GTY(()) int next_block_index = 2;
18c038b9
MM
4581
4582/* Set BLOCK_NUMBER for all the blocks in FN. */
4583
4584void
fa8db1f7 4585number_blocks (tree fn)
18c038b9
MM
4586{
4587 int i;
4588 int n_blocks;
4589 tree *block_vector;
4590
180295ed 4591 /* For XCOFF debugging output, we start numbering the blocks
18c038b9
MM
4592 from 1 within each function, rather than keeping a running
4593 count. */
180295ed
JW
4594#if defined (XCOFF_DEBUGGING_INFO)
4595 if (write_symbols == XCOFF_DEBUG)
b0e3a658 4596 next_block_index = 1;
18c038b9
MM
4597#endif
4598
4599 block_vector = get_block_vector (DECL_INITIAL (fn), &n_blocks);
4600
4601 /* The top-level BLOCK isn't numbered at all. */
4602 for (i = 1; i < n_blocks; ++i)
4603 /* We number the blocks from two. */
4604 BLOCK_NUMBER (block_vector[i]) = next_block_index++;
4605
4606 free (block_vector);
4607
4608 return;
4609}
df8992f8
RH
4610
4611/* If VAR is present in a subblock of BLOCK, return the subblock. */
4612
24e47c76 4613DEBUG_FUNCTION tree
fa8db1f7 4614debug_find_var_in_block_tree (tree var, tree block)
df8992f8
RH
4615{
4616 tree t;
4617
4618 for (t = BLOCK_VARS (block); t; t = TREE_CHAIN (t))
4619 if (t == var)
4620 return block;
4621
4622 for (t = BLOCK_SUBBLOCKS (block); t; t = TREE_CHAIN (t))
4623 {
4624 tree ret = debug_find_var_in_block_tree (var, t);
4625 if (ret)
4626 return ret;
4627 }
4628
4629 return NULL_TREE;
4630}
467456d0 4631\f
db2960f4
SL
4632/* Keep track of whether we're in a dummy function context. If we are,
4633 we don't want to invoke the set_current_function hook, because we'll
4634 get into trouble if the hook calls target_reinit () recursively or
4635 when the initial initialization is not yet complete. */
4636
4637static bool in_dummy_function;
4638
ab442df7
MM
4639/* Invoke the target hook when setting cfun. Update the optimization options
4640 if the function uses different options than the default. */
db2960f4
SL
4641
4642static void
4643invoke_set_current_function_hook (tree fndecl)
4644{
4645 if (!in_dummy_function)
ab442df7
MM
4646 {
4647 tree opts = ((fndecl)
4648 ? DECL_FUNCTION_SPECIFIC_OPTIMIZATION (fndecl)
4649 : optimization_default_node);
4650
4651 if (!opts)
4652 opts = optimization_default_node;
4653
4654 /* Change optimization options if needed. */
4655 if (optimization_current_node != opts)
4656 {
4657 optimization_current_node = opts;
46625112 4658 cl_optimization_restore (&global_options, TREE_OPTIMIZATION (opts));
ab442df7
MM
4659 }
4660
892c4745 4661 targetm.set_current_function (fndecl);
4b1baac8 4662 this_fn_optabs = this_target_optabs;
135204dd 4663
c518c102
ML
4664 /* Initialize global alignment variables after op. */
4665 parse_alignment_opts ();
4666
4b1baac8 4667 if (opts != optimization_default_node)
135204dd 4668 {
4b1baac8
RS
4669 init_tree_optimization_optabs (opts);
4670 if (TREE_OPTIMIZATION_OPTABS (opts))
4671 this_fn_optabs = (struct target_optabs *)
4672 TREE_OPTIMIZATION_OPTABS (opts);
135204dd 4673 }
ab442df7 4674 }
db2960f4
SL
4675}
4676
4677/* cfun should never be set directly; use this function. */
4678
4679void
77719b06 4680set_cfun (struct function *new_cfun, bool force)
db2960f4 4681{
77719b06 4682 if (cfun != new_cfun || force)
db2960f4
SL
4683 {
4684 cfun = new_cfun;
4685 invoke_set_current_function_hook (new_cfun ? new_cfun->decl : NULL_TREE);
b3e46655 4686 redirect_edge_var_map_empty ();
db2960f4
SL
4687 }
4688}
4689
db2960f4
SL
4690/* Initialized with NOGC, making this poisonous to the garbage collector. */
4691
526ceb68 4692static vec<function *> cfun_stack;
db2960f4 4693
af16bc76
MJ
4694/* Push the current cfun onto the stack, and set cfun to new_cfun. Also set
4695 current_function_decl accordingly. */
db2960f4
SL
4696
4697void
4698push_cfun (struct function *new_cfun)
4699{
af16bc76
MJ
4700 gcc_assert ((!cfun && !current_function_decl)
4701 || (cfun && current_function_decl == cfun->decl));
9771b263 4702 cfun_stack.safe_push (cfun);
af16bc76 4703 current_function_decl = new_cfun ? new_cfun->decl : NULL_TREE;
db2960f4
SL
4704 set_cfun (new_cfun);
4705}
4706
af16bc76 4707/* Pop cfun from the stack. Also set current_function_decl accordingly. */
db2960f4
SL
4708
4709void
4710pop_cfun (void)
4711{
9771b263 4712 struct function *new_cfun = cfun_stack.pop ();
af16bc76
MJ
4713 /* When in_dummy_function, we do have a cfun but current_function_decl is
4714 NULL. We also allow pushing NULL cfun and subsequently changing
4715 current_function_decl to something else and have both restored by
4716 pop_cfun. */
4717 gcc_checking_assert (in_dummy_function
4718 || !cfun
4719 || current_function_decl == cfun->decl);
38d34676 4720 set_cfun (new_cfun);
af16bc76 4721 current_function_decl = new_cfun ? new_cfun->decl : NULL_TREE;
db2960f4 4722}
3e87758a
RL
4723
4724/* Return value of funcdef and increase it. */
4725int
b8698a0f 4726get_next_funcdef_no (void)
3e87758a
RL
4727{
4728 return funcdef_no++;
4729}
4730
903d1e67
XDL
4731/* Return value of funcdef. */
4732int
4733get_last_funcdef_no (void)
4734{
4735 return funcdef_no;
4736}
4737
3a70d621 4738/* Allocate a function structure for FNDECL and set its contents
db2960f4
SL
4739 to the defaults. Set cfun to the newly-allocated object.
4740 Some of the helper functions invoked during initialization assume
4741 that cfun has already been set. Therefore, assign the new object
4742 directly into cfun and invoke the back end hook explicitly at the
4743 very end, rather than initializing a temporary and calling set_cfun
4744 on it.
182e0d71
AK
4745
4746 ABSTRACT_P is true if this is a function that will never be seen by
4747 the middle-end. Such functions are front-end concepts (like C++
4748 function templates) that do not correspond directly to functions
4749 placed in object files. */
7a80cf9a 4750
3a70d621 4751void
182e0d71 4752allocate_struct_function (tree fndecl, bool abstract_p)
6f086dfc 4753{
6de9cd9a 4754 tree fntype = fndecl ? TREE_TYPE (fndecl) : NULL_TREE;
6f086dfc 4755
766090c2 4756 cfun = ggc_cleared_alloc<function> ();
b384405b 4757
3a70d621 4758 init_eh_for_function ();
6f086dfc 4759
3a70d621
RH
4760 if (init_machine_status)
4761 cfun->machine = (*init_machine_status) ();
e2ecd91c 4762
7c800926
KT
4763#ifdef OVERRIDE_ABI_FORMAT
4764 OVERRIDE_ABI_FORMAT (fndecl);
4765#endif
4766
81464b2c 4767 if (fndecl != NULL_TREE)
3a70d621 4768 {
db2960f4
SL
4769 DECL_STRUCT_FUNCTION (fndecl) = cfun;
4770 cfun->decl = fndecl;
70cf5bc1 4771 current_function_funcdef_no = get_next_funcdef_no ();
5b9db1bc
MJ
4772 }
4773
4774 invoke_set_current_function_hook (fndecl);
db2960f4 4775
5b9db1bc
MJ
4776 if (fndecl != NULL_TREE)
4777 {
4778 tree result = DECL_RESULT (fndecl);
f11a7b6d
AO
4779
4780 if (!abstract_p)
4781 {
4782 /* Now that we have activated any function-specific attributes
4783 that might affect layout, particularly vector modes, relayout
4784 each of the parameters and the result. */
4785 relayout_decl (result);
4786 for (tree parm = DECL_ARGUMENTS (fndecl); parm;
4787 parm = DECL_CHAIN (parm))
4788 relayout_decl (parm);
63b0cb04
CB
4789
4790 /* Similarly relayout the function decl. */
4791 targetm.target_option.relayout_function (fndecl);
f11a7b6d
AO
4792 }
4793
182e0d71 4794 if (!abstract_p && aggregate_value_p (result, fndecl))
db2960f4 4795 {
3a70d621 4796#ifdef PCC_STATIC_STRUCT_RETURN
e3b5732b 4797 cfun->returns_pcc_struct = 1;
3a70d621 4798#endif
e3b5732b 4799 cfun->returns_struct = 1;
db2960f4
SL
4800 }
4801
f38958e8 4802 cfun->stdarg = stdarg_p (fntype);
b8698a0f 4803
db2960f4
SL
4804 /* Assume all registers in stdarg functions need to be saved. */
4805 cfun->va_list_gpr_size = VA_LIST_MAX_GPR_SIZE;
4806 cfun->va_list_fpr_size = VA_LIST_MAX_FPR_SIZE;
8f4f502f
EB
4807
4808 /* ??? This could be set on a per-function basis by the front-end
4809 but is this worth the hassle? */
4810 cfun->can_throw_non_call_exceptions = flag_non_call_exceptions;
d764963b 4811 cfun->can_delete_dead_exceptions = flag_delete_dead_exceptions;
0b37ba8a
AK
4812
4813 if (!profile_flag && !flag_instrument_function_entry_exit)
4814 DECL_NO_INSTRUMENT_FUNCTION_ENTRY_EXIT (fndecl) = 1;
3a70d621 4815 }
96a95ac1
AO
4816
4817 /* Don't enable begin stmt markers if var-tracking at assignments is
4818 disabled. The markers make little sense without the variable
4819 binding annotations among them. */
4820 cfun->debug_nonbind_markers = lang_hooks.emits_begin_stmt
4821 && MAY_HAVE_DEBUG_MARKER_STMTS;
db2960f4
SL
4822}
4823
4824/* This is like allocate_struct_function, but pushes a new cfun for FNDECL
4825 instead of just setting it. */
9d30f3c1 4826
db2960f4
SL
4827void
4828push_struct_function (tree fndecl)
4829{
af16bc76
MJ
4830 /* When in_dummy_function we might be in the middle of a pop_cfun and
4831 current_function_decl and cfun may not match. */
4832 gcc_assert (in_dummy_function
4833 || (!cfun && !current_function_decl)
4834 || (cfun && current_function_decl == cfun->decl));
9771b263 4835 cfun_stack.safe_push (cfun);
af16bc76 4836 current_function_decl = fndecl;
182e0d71 4837 allocate_struct_function (fndecl, false);
3a70d621 4838}
6f086dfc 4839
8f4f502f 4840/* Reset crtl and other non-struct-function variables to defaults as
2067c116 4841 appropriate for emitting rtl at the start of a function. */
6f086dfc 4842
3a70d621 4843static void
db2960f4 4844prepare_function_start (void)
3a70d621 4845{
614d5bd8 4846 gcc_assert (!get_last_insn ());
fb0703f7 4847 init_temp_slots ();
0de456a5 4848 init_emit ();
bd60bab2 4849 init_varasm_status ();
0de456a5 4850 init_expr ();
bf08ebeb 4851 default_rtl_profile ();
6f086dfc 4852
a11e0df4 4853 if (flag_stack_usage_info)
d3c12306 4854 {
766090c2 4855 cfun->su = ggc_cleared_alloc<stack_usage> ();
d3c12306
EB
4856 cfun->su->static_stack_size = -1;
4857 }
4858
3a70d621 4859 cse_not_expected = ! optimize;
6f086dfc 4860
3a70d621
RH
4861 /* Caller save not needed yet. */
4862 caller_save_needed = 0;
6f086dfc 4863
3a70d621
RH
4864 /* We haven't done register allocation yet. */
4865 reg_renumber = 0;
6f086dfc 4866
b384405b
BS
4867 /* Indicate that we have not instantiated virtual registers yet. */
4868 virtuals_instantiated = 0;
4869
1b3d8f8a
GK
4870 /* Indicate that we want CONCATs now. */
4871 generating_concat_p = 1;
4872
b384405b
BS
4873 /* Indicate we have no need of a frame pointer yet. */
4874 frame_pointer_needed = 0;
b384405b
BS
4875}
4876
5283d1ec
TV
4877void
4878push_dummy_function (bool with_decl)
4879{
4880 tree fn_decl, fn_type, fn_result_decl;
4881
4882 gcc_assert (!in_dummy_function);
4883 in_dummy_function = true;
4884
4885 if (with_decl)
4886 {
4887 fn_type = build_function_type_list (void_type_node, NULL_TREE);
4888 fn_decl = build_decl (UNKNOWN_LOCATION, FUNCTION_DECL, NULL_TREE,
4889 fn_type);
4890 fn_result_decl = build_decl (UNKNOWN_LOCATION, RESULT_DECL,
4891 NULL_TREE, void_type_node);
4892 DECL_RESULT (fn_decl) = fn_result_decl;
4893 }
4894 else
4895 fn_decl = NULL_TREE;
4896
4897 push_struct_function (fn_decl);
4898}
4899
b384405b
BS
4900/* Initialize the rtl expansion mechanism so that we can do simple things
4901 like generate sequences. This is used to provide a context during global
db2960f4
SL
4902 initialization of some passes. You must call expand_dummy_function_end
4903 to exit this context. */
4904
b384405b 4905void
fa8db1f7 4906init_dummy_function_start (void)
b384405b 4907{
5283d1ec 4908 push_dummy_function (false);
db2960f4 4909 prepare_function_start ();
b384405b
BS
4910}
4911
4912/* Generate RTL for the start of the function SUBR (a FUNCTION_DECL tree node)
4913 and initialize static variables for generating RTL for the statements
4914 of the function. */
4915
4916void
fa8db1f7 4917init_function_start (tree subr)
b384405b 4918{
b9b5f433
JH
4919 /* Initialize backend, if needed. */
4920 initialize_rtl ();
4921
db2960f4 4922 prepare_function_start ();
2c7eebae 4923 decide_function_section (subr);
b384405b 4924
6f086dfc
RS
4925 /* Warn if this value is an aggregate type,
4926 regardless of which calling convention we are using for it. */
ccf08a6e
DD
4927 if (AGGREGATE_TYPE_P (TREE_TYPE (DECL_RESULT (subr))))
4928 warning (OPT_Waggregate_return, "function returns an aggregate");
49ad7cfa 4929}
5c7675e9 4930
7d69de61
RH
4931/* Expand code to verify the stack_protect_guard. This is invoked at
4932 the end of a function to be protected. */
4933
b755446c 4934void
7d69de61
RH
4935stack_protect_epilogue (void)
4936{
4937 tree guard_decl = targetm.stack_protect_guard ();
19f8b229 4938 rtx_code_label *label = gen_label_rtx ();
9a24a3cc 4939 rtx x, y;
89d75572 4940 rtx_insn *seq = NULL;
7d69de61 4941
08d4cc33 4942 x = expand_normal (crtl->stack_protect_guard);
89d75572
TP
4943
4944 if (targetm.have_stack_protect_combined_test () && guard_decl)
4945 {
4946 gcc_assert (DECL_P (guard_decl));
4947 y = DECL_RTL (guard_decl);
4948 /* Allow the target to compute address of Y and compare it with X without
4949 leaking Y into a register. This combined address + compare pattern
4950 allows the target to prevent spilling of any intermediate results by
4951 splitting it after register allocator. */
4952 seq = targetm.gen_stack_protect_combined_test (x, y, label);
4953 }
a8b2130a 4954 else
89d75572
TP
4955 {
4956 if (guard_decl)
4957 y = expand_normal (guard_decl);
4958 else
4959 y = const0_rtx;
4960
4961 /* Allow the target to compare Y with X without leaking either into
4962 a register. */
4963 if (targetm.have_stack_protect_test ())
4964 seq = targetm.gen_stack_protect_test (x, y, label);
4965 }
a8b2130a 4966
89d75572 4967 if (seq)
a8b2130a
TP
4968 emit_insn (seq);
4969 else
4970 emit_cmp_and_jump_insns (x, y, EQ, NULL_RTX, ptr_mode, 1, label);
7d69de61
RH
4971
4972 /* The noreturn predictor has been moved to the tree level. The rtl-level
4973 predictors estimate this branch about 20%, which isn't enough to get
4974 things moved out of line. Since this is the only extant case of adding
4975 a noreturn function at the rtl level, it doesn't seem worth doing ought
4976 except adding the prediction by hand. */
9a24a3cc 4977 rtx_insn *tmp = get_last_insn ();
7d69de61 4978 if (JUMP_P (tmp))
9a24a3cc 4979 predict_insn_def (tmp, PRED_NORETURN, TAKEN);
7d69de61 4980
b3c144a3
SB
4981 expand_call (targetm.stack_protect_fail (), NULL_RTX, /*ignore=*/true);
4982 free_temp_slots ();
7d69de61
RH
4983 emit_label (label);
4984}
4985\f
6f086dfc
RS
4986/* Start the RTL for a new function, and set variables used for
4987 emitting RTL.
4988 SUBR is the FUNCTION_DECL node.
4989 PARMS_HAVE_CLEANUPS is nonzero if there are cleanups associated with
4990 the function's parameters, which must be run at any return statement. */
4991
4992void
b79c5284 4993expand_function_start (tree subr)
6f086dfc 4994{
6f086dfc
RS
4995 /* Make sure volatile mem refs aren't considered
4996 valid operands of arithmetic insns. */
4997 init_recog_no_volatile ();
4998
e3b5732b 4999 crtl->profile
70f4f91c
WC
5000 = (profile_flag
5001 && ! DECL_NO_INSTRUMENT_FUNCTION_ENTRY_EXIT (subr));
5002
e3b5732b 5003 crtl->limit_stack
a157febd
GK
5004 = (stack_limit_rtx != NULL_RTX && ! DECL_NO_LIMIT_STACK (subr));
5005
52a11cbf
RH
5006 /* Make the label for return statements to jump to. Do not special
5007 case machines with special return instructions -- they will be
5008 handled later during jump, ifcvt, or epilogue creation. */
6f086dfc 5009 return_label = gen_label_rtx ();
6f086dfc
RS
5010
5011 /* Initialize rtx used to return the value. */
5012 /* Do this before assign_parms so that we copy the struct value address
5013 before any library calls that assign parms might generate. */
5014
5015 /* Decide whether to return the value in memory or in a register. */
1f9ceff1 5016 tree res = DECL_RESULT (subr);
1f9ceff1 5017 if (aggregate_value_p (res, subr))
6f086dfc
RS
5018 {
5019 /* Returning something that won't go in a register. */
b3694847 5020 rtx value_address = 0;
6f086dfc
RS
5021
5022#ifdef PCC_STATIC_STRUCT_RETURN
e3b5732b 5023 if (cfun->returns_pcc_struct)
6f086dfc 5024 {
1f9ceff1 5025 int size = int_size_in_bytes (TREE_TYPE (res));
6f086dfc
RS
5026 value_address = assemble_static_space (size);
5027 }
5028 else
5029#endif
5030 {
2225b57c 5031 rtx sv = targetm.calls.struct_value_rtx (TREE_TYPE (subr), 2);
6f086dfc
RS
5032 /* Expect to be passed the address of a place to store the value.
5033 If it is passed as an argument, assign_parms will take care of
5034 it. */
61f71b34 5035 if (sv)
6f086dfc 5036 {
f11a7b6d 5037 value_address = gen_reg_rtx (Pmode);
61f71b34 5038 emit_move_insn (value_address, sv);
6f086dfc
RS
5039 }
5040 }
5041 if (value_address)
ccdecf58 5042 {
01c98570 5043 rtx x = value_address;
1f9ceff1 5044 if (!DECL_BY_REFERENCE (res))
01c98570 5045 {
f11a7b6d
AO
5046 x = gen_rtx_MEM (DECL_MODE (res), x);
5047 set_mem_attributes (x, res, 1);
01c98570 5048 }
f11a7b6d 5049 set_parm_rtl (res, x);
ccdecf58 5050 }
6f086dfc 5051 }
1f9ceff1 5052 else if (DECL_MODE (res) == VOIDmode)
6f086dfc 5053 /* If return mode is void, this decl rtl should not be used. */
f11a7b6d
AO
5054 set_parm_rtl (res, NULL_RTX);
5055 else
a53e14c0 5056 {
d5bf1143
RH
5057 /* Compute the return values into a pseudo reg, which we will copy
5058 into the true return register after the cleanups are done. */
1f9ceff1 5059 tree return_type = TREE_TYPE (res);
058c6384
EB
5060
5061 /* If we may coalesce this result, make sure it has the expected mode
5062 in case it was promoted. But we need not bother about BLKmode. */
5063 machine_mode promoted_mode
5064 = flag_tree_coalesce_vars && is_gimple_reg (res)
5065 ? promote_ssa_mode (ssa_default_def (cfun, res), NULL)
5066 : BLKmode;
5067
5068 if (promoted_mode != BLKmode)
5069 set_parm_rtl (res, gen_reg_rtx (promoted_mode));
1f9ceff1
AO
5070 else if (TYPE_MODE (return_type) != BLKmode
5071 && targetm.calls.return_in_msb (return_type))
bef5d8b6
RS
5072 /* expand_function_end will insert the appropriate padding in
5073 this case. Use the return value's natural (unpadded) mode
5074 within the function proper. */
f11a7b6d 5075 set_parm_rtl (res, gen_reg_rtx (TYPE_MODE (return_type)));
80a480ca 5076 else
0bccc606 5077 {
bef5d8b6
RS
5078 /* In order to figure out what mode to use for the pseudo, we
5079 figure out what the mode of the eventual return register will
5080 actually be, and use that. */
1d636cc6 5081 rtx hard_reg = hard_function_value (return_type, subr, 0, 1);
bef5d8b6
RS
5082
5083 /* Structures that are returned in registers are not
5084 aggregate_value_p, so we may see a PARALLEL or a REG. */
5085 if (REG_P (hard_reg))
f11a7b6d 5086 set_parm_rtl (res, gen_reg_rtx (GET_MODE (hard_reg)));
bef5d8b6
RS
5087 else
5088 {
5089 gcc_assert (GET_CODE (hard_reg) == PARALLEL);
f11a7b6d 5090 set_parm_rtl (res, gen_group_rtx (hard_reg));
bef5d8b6 5091 }
0bccc606 5092 }
a53e14c0 5093
084a1106
JDA
5094 /* Set DECL_REGISTER flag so that expand_function_end will copy the
5095 result to the real return register(s). */
1f9ceff1 5096 DECL_REGISTER (res) = 1;
a53e14c0 5097 }
6f086dfc
RS
5098
5099 /* Initialize rtx for parameters and local variables.
5100 In some cases this requires emitting insns. */
0d1416c6 5101 assign_parms (subr);
6f086dfc 5102
6de9cd9a
DN
5103 /* If function gets a static chain arg, store it. */
5104 if (cfun->static_chain_decl)
5105 {
7e140280 5106 tree parm = cfun->static_chain_decl;
21afc57d 5107 rtx local, chain;
f11a7b6d
AO
5108 rtx_insn *insn;
5109 int unsignedp;
7e140280 5110
f11a7b6d 5111 local = gen_reg_rtx (promote_decl_mode (parm, &unsignedp));
531ca746
RH
5112 chain = targetm.calls.static_chain (current_function_decl, true);
5113
5114 set_decl_incoming_rtl (parm, chain, false);
f11a7b6d 5115 set_parm_rtl (parm, local);
7e140280 5116 mark_reg_pointer (local, TYPE_ALIGN (TREE_TYPE (TREE_TYPE (parm))));
6de9cd9a 5117
f11a7b6d
AO
5118 if (GET_MODE (local) != GET_MODE (chain))
5119 {
5120 convert_move (local, chain, unsignedp);
5121 insn = get_last_insn ();
5122 }
5123 else
5124 insn = emit_move_insn (local, chain);
531ca746
RH
5125
5126 /* Mark the register as eliminable, similar to parameters. */
5127 if (MEM_P (chain)
5128 && reg_mentioned_p (arg_pointer_rtx, XEXP (chain, 0)))
7543f918 5129 set_dst_reg_note (insn, REG_EQUIV, chain, local);
3fd48b12
EB
5130
5131 /* If we aren't optimizing, save the static chain onto the stack. */
5132 if (!optimize)
5133 {
5134 tree saved_static_chain_decl
5135 = build_decl (DECL_SOURCE_LOCATION (parm), VAR_DECL,
5136 DECL_NAME (parm), TREE_TYPE (parm));
5137 rtx saved_static_chain_rtx
5138 = assign_stack_local (Pmode, GET_MODE_SIZE (Pmode), 0);
5139 SET_DECL_RTL (saved_static_chain_decl, saved_static_chain_rtx);
5140 emit_move_insn (saved_static_chain_rtx, chain);
5141 SET_DECL_VALUE_EXPR (parm, saved_static_chain_decl);
5142 DECL_HAS_VALUE_EXPR_P (parm) = 1;
5143 }
6de9cd9a
DN
5144 }
5145
108c3c88 5146 /* The following was moved from init_function_start.
180295ed 5147 The move was supposed to make sdb output more accurate. */
108c3c88
ML
5148 /* Indicate the beginning of the function body,
5149 as opposed to parm setup. */
5150 emit_note (NOTE_INSN_FUNCTION_BEG);
5151
5152 gcc_assert (NOTE_P (get_last_insn ()));
5153
5154 parm_birth_insn = get_last_insn ();
5155
6de9cd9a
DN
5156 /* If the function receives a non-local goto, then store the
5157 bits we need to restore the frame pointer. */
5158 if (cfun->nonlocal_goto_save_area)
5159 {
5160 tree t_save;
5161 rtx r_save;
5162
4846b435 5163 tree var = TREE_OPERAND (cfun->nonlocal_goto_save_area, 0);
ca5f4331 5164 gcc_assert (DECL_RTL_SET_P (var));
6de9cd9a 5165
6bbec3e1
L
5166 t_save = build4 (ARRAY_REF,
5167 TREE_TYPE (TREE_TYPE (cfun->nonlocal_goto_save_area)),
3244e67d
RS
5168 cfun->nonlocal_goto_save_area,
5169 integer_zero_node, NULL_TREE, NULL_TREE);
6de9cd9a 5170 r_save = expand_expr (t_save, NULL_RTX, VOIDmode, EXPAND_WRITE);
6bbec3e1 5171 gcc_assert (GET_MODE (r_save) == Pmode);
f0c51a1e 5172
88280cf9 5173 emit_move_insn (r_save, targetm.builtin_setjmp_frame_value ());
6de9cd9a
DN
5174 update_nonlocal_goto_save_area ();
5175 }
f0c51a1e 5176
e3b5732b 5177 if (crtl->profile)
f6f315fe 5178 {
f6f315fe 5179#ifdef PROFILE_HOOK
df696a75 5180 PROFILE_HOOK (current_function_funcdef_no);
411707f4 5181#endif
f6f315fe 5182 }
411707f4 5183
6d3cc8f0
EB
5184 /* If we are doing generic stack checking, the probe should go here. */
5185 if (flag_stack_check == GENERIC_STACK_CHECK)
ede497cf 5186 stack_check_probe_note = emit_note (NOTE_INSN_DELETED);
6f086dfc
RS
5187}
5188\f
5283d1ec
TV
5189void
5190pop_dummy_function (void)
5191{
5192 pop_cfun ();
5193 in_dummy_function = false;
5194}
5195
49ad7cfa
BS
5196/* Undo the effects of init_dummy_function_start. */
5197void
fa8db1f7 5198expand_dummy_function_end (void)
49ad7cfa 5199{
db2960f4
SL
5200 gcc_assert (in_dummy_function);
5201
49ad7cfa
BS
5202 /* End any sequences that failed to be closed due to syntax errors. */
5203 while (in_sequence_p ())
5204 end_sequence ();
5205
5206 /* Outside function body, can't compute type's actual size
5207 until next function's body starts. */
fa51b01b 5208
01d939e8
BS
5209 free_after_parsing (cfun);
5210 free_after_compilation (cfun);
5283d1ec 5211 pop_dummy_function ();
49ad7cfa
BS
5212}
5213
d5e254e1 5214/* Helper for diddle_return_value. */
bd695e1e
RH
5215
5216void
d5e254e1 5217diddle_return_value_1 (void (*doit) (rtx, void *), void *arg, rtx outgoing)
bd695e1e 5218{
c13fde05
RH
5219 if (! outgoing)
5220 return;
bd695e1e 5221
f8cfc6aa 5222 if (REG_P (outgoing))
c13fde05
RH
5223 (*doit) (outgoing, arg);
5224 else if (GET_CODE (outgoing) == PARALLEL)
5225 {
5226 int i;
bd695e1e 5227
c13fde05
RH
5228 for (i = 0; i < XVECLEN (outgoing, 0); i++)
5229 {
5230 rtx x = XEXP (XVECEXP (outgoing, 0, i), 0);
5231
f8cfc6aa 5232 if (REG_P (x) && REGNO (x) < FIRST_PSEUDO_REGISTER)
c13fde05 5233 (*doit) (x, arg);
bd695e1e
RH
5234 }
5235 }
5236}
5237
d5e254e1
IE
5238/* Call DOIT for each hard register used as a return value from
5239 the current function. */
5240
5241void
5242diddle_return_value (void (*doit) (rtx, void *), void *arg)
5243{
e9ae68af 5244 diddle_return_value_1 (doit, arg, crtl->return_rtx);
d5e254e1
IE
5245}
5246
c13fde05 5247static void
fa8db1f7 5248do_clobber_return_reg (rtx reg, void *arg ATTRIBUTE_UNUSED)
c13fde05 5249{
c41c1387 5250 emit_clobber (reg);
c13fde05
RH
5251}
5252
5253void
fa8db1f7 5254clobber_return_register (void)
c13fde05
RH
5255{
5256 diddle_return_value (do_clobber_return_reg, NULL);
9c65bbf4
JH
5257
5258 /* In case we do use pseudo to return value, clobber it too. */
5259 if (DECL_RTL_SET_P (DECL_RESULT (current_function_decl)))
5260 {
5261 tree decl_result = DECL_RESULT (current_function_decl);
5262 rtx decl_rtl = DECL_RTL (decl_result);
5263 if (REG_P (decl_rtl) && REGNO (decl_rtl) >= FIRST_PSEUDO_REGISTER)
5264 {
5265 do_clobber_return_reg (decl_rtl, NULL);
5266 }
5267 }
c13fde05
RH
5268}
5269
5270static void
fa8db1f7 5271do_use_return_reg (rtx reg, void *arg ATTRIBUTE_UNUSED)
c13fde05 5272{
c41c1387 5273 emit_use (reg);
c13fde05
RH
5274}
5275
0bf8477d 5276static void
fa8db1f7 5277use_return_register (void)
c13fde05
RH
5278{
5279 diddle_return_value (do_use_return_reg, NULL);
5280}
5281
862d0b35
DN
5282/* Set the location of the insn chain starting at INSN to LOC. */
5283
5284static void
dc01c3d1 5285set_insn_locations (rtx_insn *insn, int loc)
862d0b35 5286{
dc01c3d1 5287 while (insn != NULL)
862d0b35
DN
5288 {
5289 if (INSN_P (insn))
5290 INSN_LOCATION (insn) = loc;
5291 insn = NEXT_INSN (insn);
5292 }
5293}
5294
71c0e7fc 5295/* Generate RTL for the end of the current function. */
6f086dfc
RS
5296
5297void
fa8db1f7 5298expand_function_end (void)
6f086dfc 5299{
964be02f
RH
5300 /* If arg_pointer_save_area was referenced only from a nested
5301 function, we will not have initialized it yet. Do that now. */
e3b5732b 5302 if (arg_pointer_save_area && ! crtl->arg_pointer_save_area_init)
bd60bab2 5303 get_arg_pointer_save_area ();
964be02f 5304
b38f3813 5305 /* If we are doing generic stack checking and this function makes calls,
11044f66
RK
5306 do a stack probe at the start of the function to ensure we have enough
5307 space for another stack frame. */
b38f3813 5308 if (flag_stack_check == GENERIC_STACK_CHECK)
11044f66 5309 {
691fe203 5310 rtx_insn *insn, *seq;
11044f66
RK
5311
5312 for (insn = get_insns (); insn; insn = NEXT_INSN (insn))
4b4bf941 5313 if (CALL_P (insn))
11044f66 5314 {
c35af30f 5315 rtx max_frame_size = GEN_INT (STACK_CHECK_MAX_FRAME_SIZE);
11044f66 5316 start_sequence ();
c35af30f
EB
5317 if (STACK_CHECK_MOVING_SP)
5318 anti_adjust_stack_and_probe (max_frame_size, true);
5319 else
5320 probe_stack_range (STACK_OLD_CHECK_PROTECT, max_frame_size);
11044f66
RK
5321 seq = get_insns ();
5322 end_sequence ();
5368224f 5323 set_insn_locations (seq, prologue_location);
ede497cf 5324 emit_insn_before (seq, stack_check_probe_note);
11044f66
RK
5325 break;
5326 }
5327 }
5328
6f086dfc
RS
5329 /* End any sequences that failed to be closed due to syntax errors. */
5330 while (in_sequence_p ())
5f4f0e22 5331 end_sequence ();
6f086dfc 5332
6f086dfc
RS
5333 clear_pending_stack_adjust ();
5334 do_pending_stack_adjust ();
5335
6f086dfc 5336 /* Output a linenumber for the end of the function.
180295ed 5337 SDB depended on this. */
5368224f 5338 set_curr_insn_location (input_location);
6f086dfc 5339
fbffc70a 5340 /* Before the return label (if any), clobber the return
a1f300c0 5341 registers so that they are not propagated live to the rest of
fbffc70a
GK
5342 the function. This can only happen with functions that drop
5343 through; if there had been a return statement, there would
932f0847
JH
5344 have either been a return rtx, or a jump to the return label.
5345
5346 We delay actual code generation after the current_function_value_rtx
5347 is computed. */
e67d1102 5348 rtx_insn *clobber_after = get_last_insn ();
fbffc70a 5349
526c334b
KH
5350 /* Output the label for the actual return from the function. */
5351 emit_label (return_label);
6f086dfc 5352
677f3fa8 5353 if (targetm_common.except_unwind_info (&global_options) == UI_SJLJ)
815eb8f0
AM
5354 {
5355 /* Let except.c know where it should emit the call to unregister
5356 the function context for sjlj exceptions. */
5357 if (flag_exceptions)
5358 sjlj_emit_function_exit_after (get_last_insn ());
5359 }
0b59e81e 5360
652b0932
RH
5361 /* If this is an implementation of throw, do what's necessary to
5362 communicate between __builtin_eh_return and the epilogue. */
5363 expand_eh_return ();
5364
3e4eac3f
RH
5365 /* If scalar return value was computed in a pseudo-reg, or was a named
5366 return value that got dumped to the stack, copy that to the hard
5367 return register. */
19e7881c 5368 if (DECL_RTL_SET_P (DECL_RESULT (current_function_decl)))
6f086dfc 5369 {
3e4eac3f
RH
5370 tree decl_result = DECL_RESULT (current_function_decl);
5371 rtx decl_rtl = DECL_RTL (decl_result);
5372
5373 if (REG_P (decl_rtl)
5374 ? REGNO (decl_rtl) >= FIRST_PSEUDO_REGISTER
5375 : DECL_REGISTER (decl_result))
5376 {
38173d38 5377 rtx real_decl_rtl = crtl->return_rtx;
a97390bf 5378 complex_mode cmode;
6f086dfc 5379
ce5e43d0 5380 /* This should be set in assign_parms. */
0bccc606 5381 gcc_assert (REG_FUNCTION_VALUE_P (real_decl_rtl));
3e4eac3f
RH
5382
5383 /* If this is a BLKmode structure being returned in registers,
5384 then use the mode computed in expand_return. Note that if
797a6ac1 5385 decl_rtl is memory, then its mode may have been changed,
38173d38 5386 but that crtl->return_rtx has not. */
3e4eac3f 5387 if (GET_MODE (real_decl_rtl) == BLKmode)
ce5e43d0 5388 PUT_MODE (real_decl_rtl, GET_MODE (decl_rtl));
3e4eac3f 5389
bef5d8b6
RS
5390 /* If a non-BLKmode return value should be padded at the least
5391 significant end of the register, shift it left by the appropriate
5392 amount. BLKmode results are handled using the group load/store
5393 machinery. */
5394 if (TYPE_MODE (TREE_TYPE (decl_result)) != BLKmode
66de4d7c 5395 && REG_P (real_decl_rtl)
bef5d8b6
RS
5396 && targetm.calls.return_in_msb (TREE_TYPE (decl_result)))
5397 {
5398 emit_move_insn (gen_rtx_REG (GET_MODE (decl_rtl),
5399 REGNO (real_decl_rtl)),
5400 decl_rtl);
5401 shift_return_value (GET_MODE (decl_rtl), true, real_decl_rtl);
5402 }
aa570f54 5403 else if (GET_CODE (real_decl_rtl) == PARALLEL)
084a1106
JDA
5404 {
5405 /* If expand_function_start has created a PARALLEL for decl_rtl,
5406 move the result to the real return registers. Otherwise, do
5407 a group load from decl_rtl for a named return. */
5408 if (GET_CODE (decl_rtl) == PARALLEL)
5409 emit_group_move (real_decl_rtl, decl_rtl);
5410 else
5411 emit_group_load (real_decl_rtl, decl_rtl,
6e985040 5412 TREE_TYPE (decl_result),
084a1106
JDA
5413 int_size_in_bytes (TREE_TYPE (decl_result)));
5414 }
652b0932
RH
5415 /* In the case of complex integer modes smaller than a word, we'll
5416 need to generate some non-trivial bitfield insertions. Do that
5417 on a pseudo and not the hard register. */
5418 else if (GET_CODE (decl_rtl) == CONCAT
a97390bf
RS
5419 && is_complex_int_mode (GET_MODE (decl_rtl), &cmode)
5420 && GET_MODE_BITSIZE (cmode) <= BITS_PER_WORD)
652b0932
RH
5421 {
5422 int old_generating_concat_p;
5423 rtx tmp;
5424
5425 old_generating_concat_p = generating_concat_p;
5426 generating_concat_p = 0;
5427 tmp = gen_reg_rtx (GET_MODE (decl_rtl));
5428 generating_concat_p = old_generating_concat_p;
5429
5430 emit_move_insn (tmp, decl_rtl);
5431 emit_move_insn (real_decl_rtl, tmp);
5432 }
fc5851fe
AO
5433 /* If a named return value dumped decl_return to memory, then
5434 we may need to re-do the PROMOTE_MODE signed/unsigned
5435 extension. */
5436 else if (GET_MODE (real_decl_rtl) != GET_MODE (decl_rtl))
5437 {
5438 int unsignedp = TYPE_UNSIGNED (TREE_TYPE (decl_result));
5439 promote_function_mode (TREE_TYPE (decl_result),
5440 GET_MODE (decl_rtl), &unsignedp,
5441 TREE_TYPE (current_function_decl), 1);
5442
5443 convert_move (real_decl_rtl, decl_rtl, unsignedp);
5444 }
3e4eac3f
RH
5445 else
5446 emit_move_insn (real_decl_rtl, decl_rtl);
3e4eac3f 5447 }
6f086dfc
RS
5448 }
5449
5450 /* If returning a structure, arrange to return the address of the value
5451 in a place where debuggers expect to find it.
5452
5453 If returning a structure PCC style,
5454 the caller also depends on this value.
e3b5732b 5455 And cfun->returns_pcc_struct is not necessarily set. */
e0d14c39
BS
5456 if ((cfun->returns_struct || cfun->returns_pcc_struct)
5457 && !targetm.calls.omit_struct_return_reg)
6f086dfc 5458 {
cc77ae10 5459 rtx value_address = DECL_RTL (DECL_RESULT (current_function_decl));
6f086dfc 5460 tree type = TREE_TYPE (DECL_RESULT (current_function_decl));
cc77ae10
JM
5461 rtx outgoing;
5462
5463 if (DECL_BY_REFERENCE (DECL_RESULT (current_function_decl)))
5464 type = TREE_TYPE (type);
5465 else
5466 value_address = XEXP (value_address, 0);
5467
1d636cc6
RG
5468 outgoing = targetm.calls.function_value (build_pointer_type (type),
5469 current_function_decl, true);
6f086dfc
RS
5470
5471 /* Mark this as a function return value so integrate will delete the
5472 assignment and USE below when inlining this function. */
5473 REG_FUNCTION_VALUE_P (outgoing) = 1;
5474
d1608933 5475 /* The address may be ptr_mode and OUTGOING may be Pmode. */
c7ad039d
RS
5476 scalar_int_mode mode = as_a <scalar_int_mode> (GET_MODE (outgoing));
5477 value_address = convert_memory_address (mode, value_address);
d1608933 5478
6f086dfc 5479 emit_move_insn (outgoing, value_address);
d1608933
RK
5480
5481 /* Show return register used to hold result (in this case the address
5482 of the result. */
38173d38 5483 crtl->return_rtx = outgoing;
6f086dfc
RS
5484 }
5485
79c7fda6
JJ
5486 /* Emit the actual code to clobber return register. Don't emit
5487 it if clobber_after is a barrier, then the previous basic block
5488 certainly doesn't fall thru into the exit block. */
5489 if (!BARRIER_P (clobber_after))
5490 {
79c7fda6
JJ
5491 start_sequence ();
5492 clobber_return_register ();
e67d1102 5493 rtx_insn *seq = get_insns ();
79c7fda6 5494 end_sequence ();
932f0847 5495
79c7fda6
JJ
5496 emit_insn_after (seq, clobber_after);
5497 }
932f0847 5498
609c3937 5499 /* Output the label for the naked return from the function. */
4c33221c
UW
5500 if (naked_return_label)
5501 emit_label (naked_return_label);
6e3077c6 5502
25108646
AH
5503 /* @@@ This is a kludge. We want to ensure that instructions that
5504 may trap are not moved into the epilogue by scheduling, because
56d17681 5505 we don't always emit unwind information for the epilogue. */
f0a0390e 5506 if (cfun->can_throw_non_call_exceptions
677f3fa8 5507 && targetm_common.except_unwind_info (&global_options) != UI_SJLJ)
56d17681 5508 emit_insn (gen_blockage ());
25108646 5509
7d69de61 5510 /* If stack protection is enabled for this function, check the guard. */
87a5dc2d 5511 if (crtl->stack_protect_guard && targetm.stack_protect_runtime_enabled_p ())
7d69de61
RH
5512 stack_protect_epilogue ();
5513
40184445
BS
5514 /* If we had calls to alloca, and this machine needs
5515 an accurate stack pointer to exit the function,
5516 insert some code to save and restore the stack pointer. */
5517 if (! EXIT_IGNORE_STACK
e3b5732b 5518 && cfun->calls_alloca)
40184445 5519 {
e67d1102 5520 rtx tem = 0;
40184445 5521
9eac0f2a
RH
5522 start_sequence ();
5523 emit_stack_save (SAVE_FUNCTION, &tem);
e67d1102 5524 rtx_insn *seq = get_insns ();
9eac0f2a
RH
5525 end_sequence ();
5526 emit_insn_before (seq, parm_birth_insn);
5527
5528 emit_stack_restore (SAVE_FUNCTION, tem);
40184445
BS
5529 }
5530
c13fde05
RH
5531 /* ??? This should no longer be necessary since stupid is no longer with
5532 us, but there are some parts of the compiler (eg reload_combine, and
5533 sh mach_dep_reorg) that still try and compute their own lifetime info
5534 instead of using the general framework. */
5535 use_return_register ();
6f086dfc 5536}
278ed218
RH
5537
5538rtx
bd60bab2 5539get_arg_pointer_save_area (void)
278ed218 5540{
bd60bab2 5541 rtx ret = arg_pointer_save_area;
278ed218
RH
5542
5543 if (! ret)
5544 {
bd60bab2
JH
5545 ret = assign_stack_local (Pmode, GET_MODE_SIZE (Pmode), 0);
5546 arg_pointer_save_area = ret;
964be02f
RH
5547 }
5548
e3b5732b 5549 if (! crtl->arg_pointer_save_area_init)
964be02f 5550 {
797a6ac1 5551 /* Save the arg pointer at the beginning of the function. The
964be02f 5552 generated stack slot may not be a valid memory address, so we
278ed218
RH
5553 have to check it and fix it if necessary. */
5554 start_sequence ();
1a8cb155 5555 emit_move_insn (validize_mem (copy_rtx (ret)),
2e3f842f 5556 crtl->args.internal_arg_pointer);
e67d1102 5557 rtx_insn *seq = get_insns ();
278ed218
RH
5558 end_sequence ();
5559
964be02f 5560 push_topmost_sequence ();
1cb2fc7b 5561 emit_insn_after (seq, entry_of_function ());
964be02f 5562 pop_topmost_sequence ();
c1d9a70a
ILT
5563
5564 crtl->arg_pointer_save_area_init = true;
278ed218
RH
5565 }
5566
5567 return ret;
5568}
bdac5f58 5569\f
8a502a80
JL
5570
5571/* If debugging dumps are requested, dump information about how the
5572 target handled -fstack-check=clash for the prologue.
5573
5574 PROBES describes what if any probes were emitted.
5575
5576 RESIDUALS indicates if the prologue had any residual allocation
5577 (i.e. total allocation was not a multiple of PROBE_INTERVAL). */
5578
5579void
5580dump_stack_clash_frame_info (enum stack_clash_probes probes, bool residuals)
5581{
5582 if (!dump_file)
5583 return;
5584
5585 switch (probes)
5586 {
5587 case NO_PROBE_NO_FRAME:
5588 fprintf (dump_file,
5589 "Stack clash no probe no stack adjustment in prologue.\n");
5590 break;
5591 case NO_PROBE_SMALL_FRAME:
5592 fprintf (dump_file,
5593 "Stack clash no probe small stack adjustment in prologue.\n");
5594 break;
5595 case PROBE_INLINE:
5596 fprintf (dump_file, "Stack clash inline probes in prologue.\n");
5597 break;
5598 case PROBE_LOOP:
5599 fprintf (dump_file, "Stack clash probe loop in prologue.\n");
5600 break;
5601 }
5602
5603 if (residuals)
5604 fprintf (dump_file, "Stack clash residual allocation in prologue.\n");
5605 else
5606 fprintf (dump_file, "Stack clash no residual allocation in prologue.\n");
5607
5608 if (frame_pointer_needed)
5609 fprintf (dump_file, "Stack clash frame pointer needed.\n");
5610 else
5611 fprintf (dump_file, "Stack clash no frame pointer needed.\n");
5612
5613 if (TREE_THIS_VOLATILE (cfun->decl))
5614 fprintf (dump_file,
5615 "Stack clash noreturn prologue, assuming no implicit"
5616 " probes in caller.\n");
5617 else
5618 fprintf (dump_file,
5619 "Stack clash not noreturn prologue.\n");
5620}
5621
cd9c1ca8
RH
5622/* Add a list of INSNS to the hash HASHP, possibly allocating HASHP
5623 for the first time. */
bdac5f58 5624
0a1c58a2 5625static void
d242408f 5626record_insns (rtx_insn *insns, rtx end, hash_table<insn_cache_hasher> **hashp)
bdac5f58 5627{
dc01c3d1 5628 rtx_insn *tmp;
d242408f 5629 hash_table<insn_cache_hasher> *hash = *hashp;
0a1c58a2 5630
cd9c1ca8 5631 if (hash == NULL)
d242408f 5632 *hashp = hash = hash_table<insn_cache_hasher>::create_ggc (17);
cd9c1ca8
RH
5633
5634 for (tmp = insns; tmp != end; tmp = NEXT_INSN (tmp))
5635 {
d242408f 5636 rtx *slot = hash->find_slot (tmp, INSERT);
cd9c1ca8
RH
5637 gcc_assert (*slot == NULL);
5638 *slot = tmp;
5639 }
5640}
5641
cd400280
RH
5642/* INSN has been duplicated or replaced by as COPY, perhaps by duplicating a
5643 basic block, splitting or peepholes. If INSN is a prologue or epilogue
5644 insn, then record COPY as well. */
cd9c1ca8
RH
5645
5646void
cd400280 5647maybe_copy_prologue_epilogue_insn (rtx insn, rtx copy)
cd9c1ca8 5648{
d242408f
TS
5649 hash_table<insn_cache_hasher> *hash;
5650 rtx *slot;
cd9c1ca8 5651
cd400280 5652 hash = epilogue_insn_hash;
d242408f 5653 if (!hash || !hash->find (insn))
cd400280
RH
5654 {
5655 hash = prologue_insn_hash;
d242408f 5656 if (!hash || !hash->find (insn))
cd400280
RH
5657 return;
5658 }
cd9c1ca8 5659
d242408f 5660 slot = hash->find_slot (copy, INSERT);
cd9c1ca8
RH
5661 gcc_assert (*slot == NULL);
5662 *slot = copy;
bdac5f58
TW
5663}
5664
cd9c1ca8
RH
5665/* Determine if any INSNs in HASH are, or are part of, INSN. Because
5666 we can be running after reorg, SEQUENCE rtl is possible. */
bdac5f58 5667
cd9c1ca8 5668static bool
87ac59a0 5669contains (const rtx_insn *insn, hash_table<insn_cache_hasher> *hash)
bdac5f58 5670{
cd9c1ca8
RH
5671 if (hash == NULL)
5672 return false;
bdac5f58 5673
cd9c1ca8 5674 if (NONJUMP_INSN_P (insn) && GET_CODE (PATTERN (insn)) == SEQUENCE)
bdac5f58 5675 {
e0944870 5676 rtx_sequence *seq = as_a <rtx_sequence *> (PATTERN (insn));
cd9c1ca8 5677 int i;
e0944870 5678 for (i = seq->len () - 1; i >= 0; i--)
d242408f 5679 if (hash->find (seq->element (i)))
cd9c1ca8
RH
5680 return true;
5681 return false;
bdac5f58 5682 }
cd9c1ca8 5683
87ac59a0 5684 return hash->find (const_cast<rtx_insn *> (insn)) != NULL;
bdac5f58 5685}
5c7675e9 5686
64f6e1e1 5687int
87ac59a0 5688prologue_contains (const rtx_insn *insn)
64f6e1e1
SB
5689{
5690 return contains (insn, prologue_insn_hash);
5691}
5692
5693int
87ac59a0 5694epilogue_contains (const rtx_insn *insn)
64f6e1e1
SB
5695{
5696 return contains (insn, epilogue_insn_hash);
5697}
5698
5c7675e9 5699int
87ac59a0 5700prologue_epilogue_contains (const rtx_insn *insn)
5c7675e9 5701{
cd9c1ca8 5702 if (contains (insn, prologue_insn_hash))
5c7675e9 5703 return 1;
cd9c1ca8 5704 if (contains (insn, epilogue_insn_hash))
5c7675e9
RH
5705 return 1;
5706 return 0;
5707}
bdac5f58 5708
64f6e1e1
SB
5709void
5710record_prologue_seq (rtx_insn *seq)
5711{
5712 record_insns (seq, NULL, &prologue_insn_hash);
5713}
5714
5715void
5716record_epilogue_seq (rtx_insn *seq)
5717{
5718 record_insns (seq, NULL, &epilogue_insn_hash);
5719}
69732dcb 5720
387748de
AM
5721/* Set JUMP_LABEL for a return insn. */
5722
5723void
d38ff8dd 5724set_return_jump_label (rtx_insn *returnjump)
387748de
AM
5725{
5726 rtx pat = PATTERN (returnjump);
5727 if (GET_CODE (pat) == PARALLEL)
5728 pat = XVECEXP (pat, 0, 0);
5729 if (ANY_RETURN_P (pat))
5730 JUMP_LABEL (returnjump) = pat;
5731 else
5732 JUMP_LABEL (returnjump) = ret_rtx;
5733}
5734
fb42ed99
SB
5735/* Return a sequence to be used as the split prologue for the current
5736 function, or NULL. */
5737
5738static rtx_insn *
5739make_split_prologue_seq (void)
5740{
5741 if (!flag_split_stack
5742 || lookup_attribute ("no_split_stack", DECL_ATTRIBUTES (cfun->decl)))
5743 return NULL;
5744
5745 start_sequence ();
5746 emit_insn (targetm.gen_split_stack_prologue ());
5747 rtx_insn *seq = get_insns ();
5748 end_sequence ();
5749
5750 record_insns (seq, NULL, &prologue_insn_hash);
5751 set_insn_locations (seq, prologue_location);
5752
5753 return seq;
5754}
5755
5756/* Return a sequence to be used as the prologue for the current function,
5757 or NULL. */
5758
5759static rtx_insn *
5760make_prologue_seq (void)
5761{
5762 if (!targetm.have_prologue ())
5763 return NULL;
5764
5765 start_sequence ();
5766 rtx_insn *seq = targetm.gen_prologue ();
5767 emit_insn (seq);
5768
5769 /* Insert an explicit USE for the frame pointer
5770 if the profiling is on and the frame pointer is required. */
5771 if (crtl->profile && frame_pointer_needed)
5772 emit_use (hard_frame_pointer_rtx);
5773
5774 /* Retain a map of the prologue insns. */
5775 record_insns (seq, NULL, &prologue_insn_hash);
5776 emit_note (NOTE_INSN_PROLOGUE_END);
5777
5778 /* Ensure that instructions are not moved into the prologue when
5779 profiling is on. The call to the profiling routine can be
5780 emitted within the live range of a call-clobbered register. */
5781 if (!targetm.profile_before_prologue () && crtl->profile)
5782 emit_insn (gen_blockage ());
5783
5784 seq = get_insns ();
5785 end_sequence ();
5786 set_insn_locations (seq, prologue_location);
5787
5788 return seq;
5789}
5790
5791/* Return a sequence to be used as the epilogue for the current function,
5792 or NULL. */
5793
5794static rtx_insn *
33fec8d5 5795make_epilogue_seq (void)
fb42ed99
SB
5796{
5797 if (!targetm.have_epilogue ())
5798 return NULL;
5799
5800 start_sequence ();
33fec8d5 5801 emit_note (NOTE_INSN_EPILOGUE_BEG);
fb42ed99
SB
5802 rtx_insn *seq = targetm.gen_epilogue ();
5803 if (seq)
5804 emit_jump_insn (seq);
5805
5806 /* Retain a map of the epilogue insns. */
5807 record_insns (seq, NULL, &epilogue_insn_hash);
5808 set_insn_locations (seq, epilogue_location);
5809
5810 seq = get_insns ();
5811 rtx_insn *returnjump = get_last_insn ();
5812 end_sequence ();
5813
5814 if (JUMP_P (returnjump))
5815 set_return_jump_label (returnjump);
5816
5817 return seq;
5818}
5819
ffe14686 5820
9faa82d8 5821/* Generate the prologue and epilogue RTL if the machine supports it. Thread
bdac5f58 5822 this into place with notes indicating where the prologue ends and where
484db665
BS
5823 the epilogue begins. Update the basic block information when possible.
5824
5825 Notes on epilogue placement:
5826 There are several kinds of edges to the exit block:
5827 * a single fallthru edge from LAST_BB
5828 * possibly, edges from blocks containing sibcalls
5829 * possibly, fake edges from infinite loops
5830
5831 The epilogue is always emitted on the fallthru edge from the last basic
5832 block in the function, LAST_BB, into the exit block.
5833
5834 If LAST_BB is empty except for a label, it is the target of every
5835 other basic block in the function that ends in a return. If a
5836 target has a return or simple_return pattern (possibly with
5837 conditional variants), these basic blocks can be changed so that a
5838 return insn is emitted into them, and their target is adjusted to
5839 the real exit block.
5840
5841 Notes on shrink wrapping: We implement a fairly conservative
5842 version of shrink-wrapping rather than the textbook one. We only
5843 generate a single prologue and a single epilogue. This is
5844 sufficient to catch a number of interesting cases involving early
5845 exits.
5846
5847 First, we identify the blocks that require the prologue to occur before
5848 them. These are the ones that modify a call-saved register, or reference
5849 any of the stack or frame pointer registers. To simplify things, we then
5850 mark everything reachable from these blocks as also requiring a prologue.
5851 This takes care of loops automatically, and avoids the need to examine
5852 whether MEMs reference the frame, since it is sufficient to check for
5853 occurrences of the stack or frame pointer.
5854
5855 We then compute the set of blocks for which the need for a prologue
5856 is anticipatable (borrowing terminology from the shrink-wrapping
5857 description in Muchnick's book). These are the blocks which either
5858 require a prologue themselves, or those that have only successors
5859 where the prologue is anticipatable. The prologue needs to be
5860 inserted on all edges from BB1->BB2 where BB2 is in ANTIC and BB1
5861 is not. For the moment, we ensure that only one such edge exists.
5862
5863 The epilogue is placed as described above, but we make a
5864 distinction between inserting return and simple_return patterns
5865 when modifying other blocks that end in a return. Blocks that end
5866 in a sibcall omit the sibcall_epilogue if the block is not in
5867 ANTIC. */
bdac5f58 5868
c81b4a0e 5869void
6fb5fa3c 5870thread_prologue_and_epilogue_insns (void)
bdac5f58 5871{
484db665 5872 df_analyze ();
e881bb1b 5873
7458026b
ILT
5874 /* Can't deal with multiple successors of the entry block at the
5875 moment. Function should always have at least one entry
5876 point. */
fefa31b5 5877 gcc_assert (single_succ_p (ENTRY_BLOCK_PTR_FOR_FN (cfun)));
33fec8d5
SB
5878
5879 edge entry_edge = single_succ_edge (ENTRY_BLOCK_PTR_FOR_FN (cfun));
5880 edge orig_entry_edge = entry_edge;
484db665 5881
7dca85bf 5882 rtx_insn *split_prologue_seq = make_split_prologue_seq ();
fb42ed99 5883 rtx_insn *prologue_seq = make_prologue_seq ();
7dca85bf 5884 rtx_insn *epilogue_seq = make_epilogue_seq ();
484db665 5885
484db665
BS
5886 /* Try to perform a kind of shrink-wrapping, making sure the
5887 prologue/epilogue is emitted only around those parts of the
5888 function that require it. */
33fec8d5 5889 try_shrink_wrapping (&entry_edge, prologue_seq);
484db665 5890
c997869f
SB
5891 /* If the target can handle splitting the prologue/epilogue into separate
5892 components, try to shrink-wrap these components separately. */
5893 try_shrink_wrapping_separate (entry_edge->dest);
5894
5895 /* If that did anything for any component we now need the generate the
7dca85bf
SB
5896 "main" prologue again. Because some targets require some of these
5897 to be called in a specific order (i386 requires the split prologue
5898 to be first, for example), we create all three sequences again here.
5899 If this does not work for some target, that target should not enable
5900 separate shrink-wrapping. */
c997869f 5901 if (crtl->shrink_wrapped_separate)
7dca85bf
SB
5902 {
5903 split_prologue_seq = make_split_prologue_seq ();
5904 prologue_seq = make_prologue_seq ();
5905 epilogue_seq = make_epilogue_seq ();
5906 }
19d3c25c 5907
fefa31b5 5908 rtl_profile_for_bb (EXIT_BLOCK_PTR_FOR_FN (cfun));
484db665 5909
cd9c1ca8
RH
5910 /* A small fib -- epilogue is not yet completed, but we wish to re-use
5911 this marker for the splits of EH_RETURN patterns, and nothing else
5912 uses the flag in the meantime. */
5913 epilogue_completed = 1;
5914
cd9c1ca8
RH
5915 /* Find non-fallthru edges that end with EH_RETURN instructions. On
5916 some targets, these get split to a special version of the epilogue
5917 code. In order to be able to properly annotate these with unwind
5918 info, try to split them now. If we get a valid split, drop an
5919 EPILOGUE_BEG note and mark the insns as epilogue insns. */
33fec8d5
SB
5920 edge e;
5921 edge_iterator ei;
fefa31b5 5922 FOR_EACH_EDGE (e, ei, EXIT_BLOCK_PTR_FOR_FN (cfun)->preds)
cd9c1ca8 5923 {
691fe203 5924 rtx_insn *prev, *last, *trial;
cd9c1ca8
RH
5925
5926 if (e->flags & EDGE_FALLTHRU)
5927 continue;
5928 last = BB_END (e->src);
5929 if (!eh_returnjump_p (last))
5930 continue;
5931
5932 prev = PREV_INSN (last);
5933 trial = try_split (PATTERN (last), last, 1);
5934 if (trial == last)
5935 continue;
5936
5937 record_insns (NEXT_INSN (prev), NEXT_INSN (trial), &epilogue_insn_hash);
5938 emit_note_after (NOTE_INSN_EPILOGUE_BEG, prev);
5939 }
cd9c1ca8 5940
33fec8d5 5941 edge exit_fallthru_edge = find_fallthru_edge (EXIT_BLOCK_PTR_FOR_FN (cfun)->preds);
cc1f86f3 5942
33fec8d5 5943 if (exit_fallthru_edge)
623a66fa 5944 {
33fec8d5
SB
5945 if (epilogue_seq)
5946 {
5947 insert_insn_on_edge (epilogue_seq, exit_fallthru_edge);
e93044fc 5948 commit_edge_insertions ();
33fec8d5
SB
5949
5950 /* The epilogue insns we inserted may cause the exit edge to no longer
5951 be fallthru. */
5952 FOR_EACH_EDGE (e, ei, EXIT_BLOCK_PTR_FOR_FN (cfun)->preds)
5953 {
5954 if (((e->flags & EDGE_FALLTHRU) != 0)
5955 && returnjump_p (BB_END (e->src)))
5956 e->flags &= ~EDGE_FALLTHRU;
5957 }
5958 }
5959 else if (next_active_insn (BB_END (exit_fallthru_edge->src)))
5960 {
5961 /* We have a fall-through edge to the exit block, the source is not
5962 at the end of the function, and there will be an assembler epilogue
5963 at the end of the function.
5964 We can't use force_nonfallthru here, because that would try to
5965 use return. Inserting a jump 'by hand' is extremely messy, so
5966 we take advantage of cfg_layout_finalize using
5967 fixup_fallthru_exit_predecessor. */
5968 cfg_layout_initialize (0);
5969 basic_block cur_bb;
5970 FOR_EACH_BB_FN (cur_bb, cfun)
5971 if (cur_bb->index >= NUM_FIXED_BLOCKS
5972 && cur_bb->next_bb->index >= NUM_FIXED_BLOCKS)
5973 cur_bb->aux = cur_bb->next_bb;
5974 cfg_layout_finalize ();
5975 }
623a66fa 5976 }
cf103ca4 5977
33fec8d5 5978 /* Insert the prologue. */
484db665 5979
33fec8d5 5980 rtl_profile_for_bb (ENTRY_BLOCK_PTR_FOR_FN (cfun));
e881bb1b 5981
33fec8d5 5982 if (split_prologue_seq || prologue_seq)
30a873c3 5983 {
349721b7 5984 rtx_insn *split_prologue_insn = split_prologue_seq;
33fec8d5 5985 if (split_prologue_seq)
349721b7
JJ
5986 {
5987 while (split_prologue_insn && !NONDEBUG_INSN_P (split_prologue_insn))
5988 split_prologue_insn = NEXT_INSN (split_prologue_insn);
5989 insert_insn_on_edge (split_prologue_seq, orig_entry_edge);
5990 }
33fec8d5 5991
349721b7 5992 rtx_insn *prologue_insn = prologue_seq;
33fec8d5 5993 if (prologue_seq)
349721b7
JJ
5994 {
5995 while (prologue_insn && !NONDEBUG_INSN_P (prologue_insn))
5996 prologue_insn = NEXT_INSN (prologue_insn);
5997 insert_insn_on_edge (prologue_seq, entry_edge);
5998 }
cf103ca4 5999
30a873c3
ZD
6000 commit_edge_insertions ();
6001
cf103ca4 6002 /* Look for basic blocks within the prologue insns. */
349721b7
JJ
6003 if (split_prologue_insn
6004 && BLOCK_FOR_INSN (split_prologue_insn) == NULL)
6005 split_prologue_insn = NULL;
6006 if (prologue_insn
6007 && BLOCK_FOR_INSN (prologue_insn) == NULL)
6008 prologue_insn = NULL;
6009 if (split_prologue_insn || prologue_insn)
6010 {
6011 auto_sbitmap blocks (last_basic_block_for_fn (cfun));
6012 bitmap_clear (blocks);
6013 if (split_prologue_insn)
6014 bitmap_set_bit (blocks,
6015 BLOCK_FOR_INSN (split_prologue_insn)->index);
6016 if (prologue_insn)
6017 bitmap_set_bit (blocks, BLOCK_FOR_INSN (prologue_insn)->index);
6018 find_many_sub_basic_blocks (blocks);
6019 }
30a873c3 6020 }
0a1c58a2 6021
33fec8d5
SB
6022 default_rtl_profile ();
6023
0a1c58a2 6024 /* Emit sibling epilogues before any sibling call sites. */
33fec8d5
SB
6025 for (ei = ei_start (EXIT_BLOCK_PTR_FOR_FN (cfun)->preds);
6026 (e = ei_safe_edge (ei));
6027 ei_next (&ei))
0a1c58a2 6028 {
33fec8d5
SB
6029 /* Skip those already handled, the ones that run without prologue. */
6030 if (e->flags & EDGE_IGNORE)
628f6a4e 6031 {
33fec8d5 6032 e->flags &= ~EDGE_IGNORE;
628f6a4e
BE
6033 continue;
6034 }
0a1c58a2 6035
33fec8d5
SB
6036 rtx_insn *insn = BB_END (e->src);
6037
6038 if (!(CALL_P (insn) && SIBLING_CALL_P (insn)))
6039 continue;
6040
e86a9946 6041 if (rtx_insn *ep_seq = targetm.gen_sibcall_epilogue ())
484db665
BS
6042 {
6043 start_sequence ();
6044 emit_note (NOTE_INSN_EPILOGUE_BEG);
6045 emit_insn (ep_seq);
dc01c3d1 6046 rtx_insn *seq = get_insns ();
484db665 6047 end_sequence ();
0a1c58a2 6048
484db665
BS
6049 /* Retain a map of the epilogue insns. Used in life analysis to
6050 avoid getting rid of sibcall epilogue insns. Do this before we
6051 actually emit the sequence. */
6052 record_insns (seq, NULL, &epilogue_insn_hash);
5368224f 6053 set_insn_locations (seq, epilogue_location);
2f937369 6054
484db665
BS
6055 emit_insn_before (seq, insn);
6056 }
0a1c58a2 6057 }
ca1117cc 6058
33fec8d5 6059 if (epilogue_seq)
86c82654 6060 {
9c8348cf 6061 rtx_insn *insn, *next;
86c82654
RH
6062
6063 /* Similarly, move any line notes that appear after the epilogue.
ff7cc307 6064 There is no need, however, to be quite so anal about the existence
071a42f9 6065 of such a note. Also possibly move
84c1fa24
UW
6066 NOTE_INSN_FUNCTION_BEG notes, as those can be relevant for debug
6067 info generation. */
33fec8d5 6068 for (insn = epilogue_seq; insn; insn = next)
86c82654
RH
6069 {
6070 next = NEXT_INSN (insn);
b8698a0f 6071 if (NOTE_P (insn)
a38e7aa5 6072 && (NOTE_KIND (insn) == NOTE_INSN_FUNCTION_BEG))
33fec8d5 6073 reorder_insns (insn, insn, PREV_INSN (epilogue_seq));
86c82654
RH
6074 }
6075 }
6fb5fa3c
DB
6076
6077 /* Threading the prologue and epilogue changes the artificial refs
6078 in the entry and exit blocks. */
6079 epilogue_completed = 1;
6080 df_update_entry_exit_and_calls ();
bdac5f58
TW
6081}
6082
cd9c1ca8
RH
6083/* Reposition the prologue-end and epilogue-begin notes after
6084 instruction scheduling. */
bdac5f58
TW
6085
6086void
6fb5fa3c 6087reposition_prologue_and_epilogue_notes (void)
bdac5f58 6088{
e86a9946
RS
6089 if (!targetm.have_prologue ()
6090 && !targetm.have_epilogue ()
6091 && !targetm.have_sibcall_epilogue ())
5251b8b3 6092 return;
5251b8b3 6093
cd9c1ca8
RH
6094 /* Since the hash table is created on demand, the fact that it is
6095 non-null is a signal that it is non-empty. */
6096 if (prologue_insn_hash != NULL)
bdac5f58 6097 {
d242408f 6098 size_t len = prologue_insn_hash->elements ();
691fe203 6099 rtx_insn *insn, *last = NULL, *note = NULL;
bdac5f58 6100
cd9c1ca8
RH
6101 /* Scan from the beginning until we reach the last prologue insn. */
6102 /* ??? While we do have the CFG intact, there are two problems:
6103 (1) The prologue can contain loops (typically probing the stack),
6104 which means that the end of the prologue isn't in the first bb.
6105 (2) Sometimes the PROLOGUE_END note gets pushed into the next bb. */
6fb5fa3c 6106 for (insn = get_insns (); insn; insn = NEXT_INSN (insn))
bdac5f58 6107 {
4b4bf941 6108 if (NOTE_P (insn))
9392c110 6109 {
a38e7aa5 6110 if (NOTE_KIND (insn) == NOTE_INSN_PROLOGUE_END)
0a1c58a2
JL
6111 note = insn;
6112 }
cd9c1ca8 6113 else if (contains (insn, prologue_insn_hash))
0a1c58a2 6114 {
9f53e965
RH
6115 last = insn;
6116 if (--len == 0)
6117 break;
6118 }
6119 }
797a6ac1 6120
9f53e965
RH
6121 if (last)
6122 {
cd9c1ca8 6123 if (note == NULL)
9f53e965 6124 {
cd9c1ca8
RH
6125 /* Scan forward looking for the PROLOGUE_END note. It should
6126 be right at the beginning of the block, possibly with other
6127 insn notes that got moved there. */
6128 for (note = NEXT_INSN (last); ; note = NEXT_INSN (note))
6129 {
6130 if (NOTE_P (note)
6131 && NOTE_KIND (note) == NOTE_INSN_PROLOGUE_END)
6132 break;
6133 }
9f53e965 6134 }
c93b03c2 6135
9f53e965 6136 /* Avoid placing note between CODE_LABEL and BASIC_BLOCK note. */
4b4bf941 6137 if (LABEL_P (last))
9f53e965
RH
6138 last = NEXT_INSN (last);
6139 reorder_insns (note, note, last);
bdac5f58 6140 }
0a1c58a2
JL
6141 }
6142
cd9c1ca8 6143 if (epilogue_insn_hash != NULL)
0a1c58a2 6144 {
cd9c1ca8
RH
6145 edge_iterator ei;
6146 edge e;
bdac5f58 6147
fefa31b5 6148 FOR_EACH_EDGE (e, ei, EXIT_BLOCK_PTR_FOR_FN (cfun)->preds)
bdac5f58 6149 {
691fe203 6150 rtx_insn *insn, *first = NULL, *note = NULL;
997704f1 6151 basic_block bb = e->src;
c93b03c2 6152
997704f1 6153 /* Scan from the beginning until we reach the first epilogue insn. */
cd9c1ca8 6154 FOR_BB_INSNS (bb, insn)
9f53e965 6155 {
cd9c1ca8
RH
6156 if (NOTE_P (insn))
6157 {
6158 if (NOTE_KIND (insn) == NOTE_INSN_EPILOGUE_BEG)
6159 {
6160 note = insn;
997704f1 6161 if (first != NULL)
cd9c1ca8
RH
6162 break;
6163 }
6164 }
997704f1 6165 else if (first == NULL && contains (insn, epilogue_insn_hash))
cd9c1ca8 6166 {
997704f1 6167 first = insn;
cd9c1ca8
RH
6168 if (note != NULL)
6169 break;
6170 }
9392c110 6171 }
997704f1
RH
6172
6173 if (note)
6174 {
6175 /* If the function has a single basic block, and no real
b8698a0f 6176 epilogue insns (e.g. sibcall with no cleanup), the
997704f1
RH
6177 epilogue note can get scheduled before the prologue
6178 note. If we have frame related prologue insns, having
6179 them scanned during the epilogue will result in a crash.
6180 In this case re-order the epilogue note to just before
6181 the last insn in the block. */
6182 if (first == NULL)
6183 first = BB_END (bb);
6184
6185 if (PREV_INSN (first) != note)
6186 reorder_insns (note, note, PREV_INSN (first));
6187 }
bdac5f58
TW
6188 }
6189 }
bdac5f58 6190}
87ff9c8e 6191
df92c640
SB
6192/* Returns the name of function declared by FNDECL. */
6193const char *
6194fndecl_name (tree fndecl)
6195{
6196 if (fndecl == NULL)
6197 return "(nofn)";
60591d4e 6198 return lang_hooks.decl_printable_name (fndecl, 1);
df92c640
SB
6199}
6200
532aafad
SB
6201/* Returns the name of function FN. */
6202const char *
6203function_name (struct function *fn)
6204{
df92c640
SB
6205 tree fndecl = (fn == NULL) ? NULL : fn->decl;
6206 return fndecl_name (fndecl);
532aafad
SB
6207}
6208
faed5cc3
SB
6209/* Returns the name of the current function. */
6210const char *
6211current_function_name (void)
6212{
532aafad 6213 return function_name (cfun);
faed5cc3 6214}
ef330312
PB
6215\f
6216
c2924966 6217static unsigned int
ef330312
PB
6218rest_of_handle_check_leaf_regs (void)
6219{
6220#ifdef LEAF_REGISTERS
416ff32e 6221 crtl->uses_only_leaf_regs
ef330312
PB
6222 = optimize > 0 && only_leaf_regs_used () && leaf_function_p ();
6223#endif
c2924966 6224 return 0;
ef330312
PB
6225}
6226
8d8d1a28 6227/* Insert a TYPE into the used types hash table of CFUN. */
b646ba3f 6228
8d8d1a28
AH
6229static void
6230used_types_insert_helper (tree type, struct function *func)
33c9159e 6231{
8d8d1a28 6232 if (type != NULL && func != NULL)
33c9159e 6233 {
33c9159e 6234 if (func->used_types_hash == NULL)
b086d530
TS
6235 func->used_types_hash = hash_set<tree>::create_ggc (37);
6236
6237 func->used_types_hash->add (type);
33c9159e
AH
6238 }
6239}
6240
8d8d1a28
AH
6241/* Given a type, insert it into the used hash table in cfun. */
6242void
6243used_types_insert (tree t)
6244{
6245 while (POINTER_TYPE_P (t) || TREE_CODE (t) == ARRAY_TYPE)
095c7b3c
JJ
6246 if (TYPE_NAME (t))
6247 break;
6248 else
6249 t = TREE_TYPE (t);
29ce73cb
PB
6250 if (TREE_CODE (t) == ERROR_MARK)
6251 return;
095c7b3c
JJ
6252 if (TYPE_NAME (t) == NULL_TREE
6253 || TYPE_NAME (t) == TYPE_NAME (TYPE_MAIN_VARIANT (t)))
6254 t = TYPE_MAIN_VARIANT (t);
8d8d1a28 6255 if (debug_info_level > DINFO_LEVEL_NONE)
b646ba3f
DS
6256 {
6257 if (cfun)
6258 used_types_insert_helper (t, cfun);
6259 else
9771b263
DN
6260 {
6261 /* So this might be a type referenced by a global variable.
6262 Record that type so that we can later decide to emit its
6263 debug information. */
6264 vec_safe_push (types_used_by_cur_var_decl, t);
6265 }
b646ba3f
DS
6266 }
6267}
6268
6269/* Helper to Hash a struct types_used_by_vars_entry. */
6270
6271static hashval_t
6272hash_types_used_by_vars_entry (const struct types_used_by_vars_entry *entry)
6273{
6274 gcc_assert (entry && entry->var_decl && entry->type);
6275
6276 return iterative_hash_object (entry->type,
6277 iterative_hash_object (entry->var_decl, 0));
6278}
6279
6280/* Hash function of the types_used_by_vars_entry hash table. */
6281
6282hashval_t
2a22f99c 6283used_type_hasher::hash (types_used_by_vars_entry *entry)
b646ba3f 6284{
b646ba3f
DS
6285 return hash_types_used_by_vars_entry (entry);
6286}
6287
6288/*Equality function of the types_used_by_vars_entry hash table. */
6289
2a22f99c
TS
6290bool
6291used_type_hasher::equal (types_used_by_vars_entry *e1,
6292 types_used_by_vars_entry *e2)
b646ba3f 6293{
b646ba3f
DS
6294 return (e1->var_decl == e2->var_decl && e1->type == e2->type);
6295}
6296
6297/* Inserts an entry into the types_used_by_vars_hash hash table. */
6298
6299void
6300types_used_by_var_decl_insert (tree type, tree var_decl)
6301{
6302 if (type != NULL && var_decl != NULL)
6303 {
2a22f99c 6304 types_used_by_vars_entry **slot;
b646ba3f
DS
6305 struct types_used_by_vars_entry e;
6306 e.var_decl = var_decl;
6307 e.type = type;
6308 if (types_used_by_vars_hash == NULL)
2a22f99c
TS
6309 types_used_by_vars_hash
6310 = hash_table<used_type_hasher>::create_ggc (37);
6311
6312 slot = types_used_by_vars_hash->find_slot (&e, INSERT);
b646ba3f
DS
6313 if (*slot == NULL)
6314 {
6315 struct types_used_by_vars_entry *entry;
766090c2 6316 entry = ggc_alloc<types_used_by_vars_entry> ();
b646ba3f
DS
6317 entry->type = type;
6318 entry->var_decl = var_decl;
6319 *slot = entry;
6320 }
6321 }
8d8d1a28
AH
6322}
6323
27a4cd48
DM
6324namespace {
6325
6326const pass_data pass_data_leaf_regs =
6327{
6328 RTL_PASS, /* type */
6329 "*leaf_regs", /* name */
6330 OPTGROUP_NONE, /* optinfo_flags */
27a4cd48
DM
6331 TV_NONE, /* tv_id */
6332 0, /* properties_required */
6333 0, /* properties_provided */
6334 0, /* properties_destroyed */
6335 0, /* todo_flags_start */
6336 0, /* todo_flags_finish */
ef330312
PB
6337};
6338
27a4cd48
DM
6339class pass_leaf_regs : public rtl_opt_pass
6340{
6341public:
c3284718
RS
6342 pass_leaf_regs (gcc::context *ctxt)
6343 : rtl_opt_pass (pass_data_leaf_regs, ctxt)
27a4cd48
DM
6344 {}
6345
6346 /* opt_pass methods: */
be55bfe6
TS
6347 virtual unsigned int execute (function *)
6348 {
6349 return rest_of_handle_check_leaf_regs ();
6350 }
27a4cd48
DM
6351
6352}; // class pass_leaf_regs
6353
6354} // anon namespace
6355
6356rtl_opt_pass *
6357make_pass_leaf_regs (gcc::context *ctxt)
6358{
6359 return new pass_leaf_regs (ctxt);
6360}
6361
6fb5fa3c
DB
6362static unsigned int
6363rest_of_handle_thread_prologue_and_epilogue (void)
6364{
63d0f6ab
SB
6365 /* prepare_shrink_wrap is sensitive to the block structure of the control
6366 flow graph, so clean it up first. */
6fb5fa3c 6367 if (optimize)
63d0f6ab 6368 cleanup_cfg (0);
d3c12306 6369
6fb5fa3c
DB
6370 /* On some machines, the prologue and epilogue code, or parts thereof,
6371 can be represented as RTL. Doing so lets us schedule insns between
6372 it and the rest of the code and also allows delayed branch
6373 scheduling to operate in the epilogue. */
6fb5fa3c 6374 thread_prologue_and_epilogue_insns ();
d3c12306 6375
86b107ae
SB
6376 /* Some non-cold blocks may now be only reachable from cold blocks.
6377 Fix that up. */
6378 fixup_partitions ();
6379
bdc6e1ae
SB
6380 /* Shrink-wrapping can result in unreachable edges in the epilogue,
6381 see PR57320. */
63d0f6ab 6382 cleanup_cfg (optimize ? CLEANUP_EXPENSIVE : 0);
bdc6e1ae 6383
d3c12306 6384 /* The stack usage info is finalized during prologue expansion. */
a11e0df4 6385 if (flag_stack_usage_info)
d3c12306
EB
6386 output_stack_usage ();
6387
6fb5fa3c
DB
6388 return 0;
6389}
6390
27a4cd48
DM
6391namespace {
6392
6393const pass_data pass_data_thread_prologue_and_epilogue =
6394{
6395 RTL_PASS, /* type */
6396 "pro_and_epilogue", /* name */
6397 OPTGROUP_NONE, /* optinfo_flags */
27a4cd48
DM
6398 TV_THREAD_PROLOGUE_AND_EPILOGUE, /* tv_id */
6399 0, /* properties_required */
6400 0, /* properties_provided */
6401 0, /* properties_destroyed */
3bea341f
RB
6402 0, /* todo_flags_start */
6403 ( TODO_df_verify | TODO_df_finish ), /* todo_flags_finish */
6fb5fa3c 6404};
27a4cd48
DM
6405
6406class pass_thread_prologue_and_epilogue : public rtl_opt_pass
6407{
6408public:
c3284718
RS
6409 pass_thread_prologue_and_epilogue (gcc::context *ctxt)
6410 : rtl_opt_pass (pass_data_thread_prologue_and_epilogue, ctxt)
27a4cd48
DM
6411 {}
6412
6413 /* opt_pass methods: */
be55bfe6
TS
6414 virtual unsigned int execute (function *)
6415 {
6416 return rest_of_handle_thread_prologue_and_epilogue ();
6417 }
27a4cd48
DM
6418
6419}; // class pass_thread_prologue_and_epilogue
6420
6421} // anon namespace
6422
6423rtl_opt_pass *
6424make_pass_thread_prologue_and_epilogue (gcc::context *ctxt)
6425{
6426 return new pass_thread_prologue_and_epilogue (ctxt);
6427}
d8d72314
PB
6428\f
6429
6430/* This mini-pass fixes fall-out from SSA in asm statements that have
b8698a0f 6431 in-out constraints. Say you start with
d8d72314
PB
6432
6433 orig = inout;
6434 asm ("": "+mr" (inout));
6435 use (orig);
6436
6437 which is transformed very early to use explicit output and match operands:
6438
6439 orig = inout;
6440 asm ("": "=mr" (inout) : "0" (inout));
6441 use (orig);
6442
6443 Or, after SSA and copyprop,
6444
6445 asm ("": "=mr" (inout_2) : "0" (inout_1));
6446 use (inout_1);
6447
6448 Clearly inout_2 and inout_1 can't be coalesced easily anymore, as
6449 they represent two separate values, so they will get different pseudo
6450 registers during expansion. Then, since the two operands need to match
6451 per the constraints, but use different pseudo registers, reload can
6452 only register a reload for these operands. But reloads can only be
6453 satisfied by hardregs, not by memory, so we need a register for this
6454 reload, just because we are presented with non-matching operands.
6455 So, even though we allow memory for this operand, no memory can be
6456 used for it, just because the two operands don't match. This can
6457 cause reload failures on register-starved targets.
6458
6459 So it's a symptom of reload not being able to use memory for reloads
6460 or, alternatively it's also a symptom of both operands not coming into
6461 reload as matching (in which case the pseudo could go to memory just
6462 fine, as the alternative allows it, and no reload would be necessary).
6463 We fix the latter problem here, by transforming
6464
6465 asm ("": "=mr" (inout_2) : "0" (inout_1));
6466
6467 back to
6468
6469 inout_2 = inout_1;
6470 asm ("": "=mr" (inout_2) : "0" (inout_2)); */
6471
6472static void
691fe203 6473match_asm_constraints_1 (rtx_insn *insn, rtx *p_sets, int noutputs)
d8d72314
PB
6474{
6475 int i;
6476 bool changed = false;
6477 rtx op = SET_SRC (p_sets[0]);
6478 int ninputs = ASM_OPERANDS_INPUT_LENGTH (op);
6479 rtvec inputs = ASM_OPERANDS_INPUT_VEC (op);
1b4572a8 6480 bool *output_matched = XALLOCAVEC (bool, noutputs);
d8d72314 6481
d7b8033f 6482 memset (output_matched, 0, noutputs * sizeof (bool));
d8d72314
PB
6483 for (i = 0; i < ninputs; i++)
6484 {
691fe203
DM
6485 rtx input, output;
6486 rtx_insn *insns;
d8d72314
PB
6487 const char *constraint = ASM_OPERANDS_INPUT_CONSTRAINT (op, i);
6488 char *end;
53220215 6489 int match, j;
d8d72314 6490
70f16287
JJ
6491 if (*constraint == '%')
6492 constraint++;
6493
d8d72314
PB
6494 match = strtoul (constraint, &end, 10);
6495 if (end == constraint)
6496 continue;
6497
6498 gcc_assert (match < noutputs);
6499 output = SET_DEST (p_sets[match]);
6500 input = RTVEC_ELT (inputs, i);
53220215
MM
6501 /* Only do the transformation for pseudos. */
6502 if (! REG_P (output)
6503 || rtx_equal_p (output, input)
3c896da0 6504 || !(REG_P (input) || SUBREG_P (input)
b3d89380
JJ
6505 || MEM_P (input) || CONSTANT_P (input))
6506 || !general_operand (input, GET_MODE (output)))
d8d72314
PB
6507 continue;
6508
53220215
MM
6509 /* We can't do anything if the output is also used as input,
6510 as we're going to overwrite it. */
6511 for (j = 0; j < ninputs; j++)
6512 if (reg_overlap_mentioned_p (output, RTVEC_ELT (inputs, j)))
6513 break;
6514 if (j != ninputs)
6515 continue;
6516
d7b8033f
JJ
6517 /* Avoid changing the same input several times. For
6518 asm ("" : "=mr" (out1), "=mr" (out2) : "0" (in), "1" (in));
6519 only change in once (to out1), rather than changing it
6520 first to out1 and afterwards to out2. */
6521 if (i > 0)
6522 {
6523 for (j = 0; j < noutputs; j++)
6524 if (output_matched[j] && input == SET_DEST (p_sets[j]))
6525 break;
6526 if (j != noutputs)
6527 continue;
6528 }
6529 output_matched[match] = true;
6530
d8d72314 6531 start_sequence ();
6c4ff3c9 6532 emit_move_insn (output, copy_rtx (input));
d8d72314
PB
6533 insns = get_insns ();
6534 end_sequence ();
d8d72314 6535 emit_insn_before (insns, insn);
53220215
MM
6536
6537 /* Now replace all mentions of the input with output. We can't
fa10beec 6538 just replace the occurrence in inputs[i], as the register might
53220215
MM
6539 also be used in some other input (or even in an address of an
6540 output), which would mean possibly increasing the number of
6541 inputs by one (namely 'output' in addition), which might pose
6542 a too complicated problem for reload to solve. E.g. this situation:
6543
6544 asm ("" : "=r" (output), "=m" (input) : "0" (input))
6545
84fbffb2 6546 Here 'input' is used in two occurrences as input (once for the
53220215 6547 input operand, once for the address in the second output operand).
fa10beec 6548 If we would replace only the occurrence of the input operand (to
53220215
MM
6549 make the matching) we would be left with this:
6550
6551 output = input
6552 asm ("" : "=r" (output), "=m" (input) : "0" (output))
6553
6554 Now we suddenly have two different input values (containing the same
6555 value, but different pseudos) where we formerly had only one.
6556 With more complicated asms this might lead to reload failures
6557 which wouldn't have happen without this pass. So, iterate over
84fbffb2 6558 all operands and replace all occurrences of the register used. */
53220215 6559 for (j = 0; j < noutputs; j++)
1596d61e 6560 if (!rtx_equal_p (SET_DEST (p_sets[j]), input)
53220215
MM
6561 && reg_overlap_mentioned_p (input, SET_DEST (p_sets[j])))
6562 SET_DEST (p_sets[j]) = replace_rtx (SET_DEST (p_sets[j]),
6563 input, output);
6564 for (j = 0; j < ninputs; j++)
6565 if (reg_overlap_mentioned_p (input, RTVEC_ELT (inputs, j)))
6566 RTVEC_ELT (inputs, j) = replace_rtx (RTVEC_ELT (inputs, j),
6567 input, output);
6568
d8d72314
PB
6569 changed = true;
6570 }
6571
6572 if (changed)
6573 df_insn_rescan (insn);
6574}
6575
5cf18d25
ML
6576/* Add the decl D to the local_decls list of FUN. */
6577
6578void
6579add_local_decl (struct function *fun, tree d)
6580{
8813a647 6581 gcc_assert (VAR_P (d));
5cf18d25
ML
6582 vec_safe_push (fun->local_decls, d);
6583}
6584
be55bfe6
TS
6585namespace {
6586
6587const pass_data pass_data_match_asm_constraints =
6588{
6589 RTL_PASS, /* type */
6590 "asmcons", /* name */
6591 OPTGROUP_NONE, /* optinfo_flags */
be55bfe6
TS
6592 TV_NONE, /* tv_id */
6593 0, /* properties_required */
6594 0, /* properties_provided */
6595 0, /* properties_destroyed */
6596 0, /* todo_flags_start */
6597 0, /* todo_flags_finish */
6598};
6599
6600class pass_match_asm_constraints : public rtl_opt_pass
6601{
6602public:
6603 pass_match_asm_constraints (gcc::context *ctxt)
6604 : rtl_opt_pass (pass_data_match_asm_constraints, ctxt)
6605 {}
6606
6607 /* opt_pass methods: */
6608 virtual unsigned int execute (function *);
6609
6610}; // class pass_match_asm_constraints
6611
6612unsigned
6613pass_match_asm_constraints::execute (function *fun)
d8d72314
PB
6614{
6615 basic_block bb;
691fe203
DM
6616 rtx_insn *insn;
6617 rtx pat, *p_sets;
d8d72314
PB
6618 int noutputs;
6619
e3b5732b 6620 if (!crtl->has_asm_statement)
d8d72314
PB
6621 return 0;
6622
6623 df_set_flags (DF_DEFER_INSN_RESCAN);
be55bfe6 6624 FOR_EACH_BB_FN (bb, fun)
d8d72314
PB
6625 {
6626 FOR_BB_INSNS (bb, insn)
6627 {
6628 if (!INSN_P (insn))
6629 continue;
6630
6631 pat = PATTERN (insn);
6632 if (GET_CODE (pat) == PARALLEL)
6633 p_sets = &XVECEXP (pat, 0, 0), noutputs = XVECLEN (pat, 0);
6634 else if (GET_CODE (pat) == SET)
6635 p_sets = &PATTERN (insn), noutputs = 1;
6636 else
6637 continue;
6638
6639 if (GET_CODE (*p_sets) == SET
6640 && GET_CODE (SET_SRC (*p_sets)) == ASM_OPERANDS)
6641 match_asm_constraints_1 (insn, p_sets, noutputs);
6642 }
6643 }
6644
6645 return TODO_df_finish;
6646}
6647
27a4cd48
DM
6648} // anon namespace
6649
6650rtl_opt_pass *
6651make_pass_match_asm_constraints (gcc::context *ctxt)
6652{
6653 return new pass_match_asm_constraints (ctxt);
6654}
6655
faed5cc3 6656
e2500fed 6657#include "gt-function.h"