]> git.ipfire.org Git - thirdparty/gcc.git/blob - gcc/cp/lambda.c
Implement N4514, C++ Extensions for Transactional Memory.
[thirdparty/gcc.git] / gcc / cp / lambda.c
1 /* Perform the semantic phase of lambda parsing, i.e., the process of
2 building tree structure, checking semantic consistency, and
3 building RTL. These routines are used both during actual parsing
4 and during the instantiation of template functions.
5
6 Copyright (C) 1998-2015 Free Software Foundation, Inc.
7
8 This file is part of GCC.
9
10 GCC is free software; you can redistribute it and/or modify it
11 under the terms of the GNU General Public License as published by
12 the Free Software Foundation; either version 3, or (at your option)
13 any later version.
14
15 GCC is distributed in the hope that it will be useful, but
16 WITHOUT ANY WARRANTY; without even the implied warranty of
17 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
18 General Public License for more details.
19
20 You should have received a copy of the GNU General Public License
21 along with GCC; see the file COPYING3. If not see
22 <http://www.gnu.org/licenses/>. */
23
24 #include "config.h"
25 #include "system.h"
26 #include "coretypes.h"
27 #include "alias.h"
28 #include "tree.h"
29 #include "options.h"
30 #include "stringpool.h"
31 #include "tm.h"
32 #include "hard-reg-set.h"
33 #include "function.h"
34 #include "cgraph.h"
35 #include "tree-iterator.h"
36 #include "cp-tree.h"
37 #include "toplev.h"
38
39 /* Constructor for a lambda expression. */
40
41 tree
42 build_lambda_expr (void)
43 {
44 tree lambda = make_node (LAMBDA_EXPR);
45 LAMBDA_EXPR_DEFAULT_CAPTURE_MODE (lambda) = CPLD_NONE;
46 LAMBDA_EXPR_CAPTURE_LIST (lambda) = NULL_TREE;
47 LAMBDA_EXPR_THIS_CAPTURE (lambda) = NULL_TREE;
48 LAMBDA_EXPR_PENDING_PROXIES (lambda) = NULL;
49 LAMBDA_EXPR_RETURN_TYPE (lambda) = NULL_TREE;
50 LAMBDA_EXPR_MUTABLE_P (lambda) = false;
51 return lambda;
52 }
53
54 /* Create the closure object for a LAMBDA_EXPR. */
55
56 tree
57 build_lambda_object (tree lambda_expr)
58 {
59 /* Build aggregate constructor call.
60 - cp_parser_braced_list
61 - cp_parser_functional_cast */
62 vec<constructor_elt, va_gc> *elts = NULL;
63 tree node, expr, type;
64 location_t saved_loc;
65
66 if (processing_template_decl)
67 return lambda_expr;
68
69 /* Make sure any error messages refer to the lambda-introducer. */
70 saved_loc = input_location;
71 input_location = LAMBDA_EXPR_LOCATION (lambda_expr);
72
73 for (node = LAMBDA_EXPR_CAPTURE_LIST (lambda_expr);
74 node;
75 node = TREE_CHAIN (node))
76 {
77 tree field = TREE_PURPOSE (node);
78 tree val = TREE_VALUE (node);
79
80 if (field == error_mark_node)
81 {
82 expr = error_mark_node;
83 goto out;
84 }
85
86 if (DECL_P (val))
87 mark_used (val);
88
89 /* Mere mortals can't copy arrays with aggregate initialization, so
90 do some magic to make it work here. */
91 if (TREE_CODE (TREE_TYPE (field)) == ARRAY_TYPE)
92 val = build_array_copy (val);
93 else if (DECL_NORMAL_CAPTURE_P (field)
94 && !DECL_VLA_CAPTURE_P (field)
95 && TREE_CODE (TREE_TYPE (field)) != REFERENCE_TYPE)
96 {
97 /* "the entities that are captured by copy are used to
98 direct-initialize each corresponding non-static data
99 member of the resulting closure object."
100
101 There's normally no way to express direct-initialization
102 from an element of a CONSTRUCTOR, so we build up a special
103 TARGET_EXPR to bypass the usual copy-initialization. */
104 val = force_rvalue (val, tf_warning_or_error);
105 if (TREE_CODE (val) == TARGET_EXPR)
106 TARGET_EXPR_DIRECT_INIT_P (val) = true;
107 }
108
109 CONSTRUCTOR_APPEND_ELT (elts, DECL_NAME (field), val);
110 }
111
112 expr = build_constructor (init_list_type_node, elts);
113 CONSTRUCTOR_IS_DIRECT_INIT (expr) = 1;
114
115 /* N2927: "[The closure] class type is not an aggregate."
116 But we briefly treat it as an aggregate to make this simpler. */
117 type = LAMBDA_EXPR_CLOSURE (lambda_expr);
118 CLASSTYPE_NON_AGGREGATE (type) = 0;
119 expr = finish_compound_literal (type, expr, tf_warning_or_error);
120 CLASSTYPE_NON_AGGREGATE (type) = 1;
121
122 out:
123 input_location = saved_loc;
124 return expr;
125 }
126
127 /* Return an initialized RECORD_TYPE for LAMBDA.
128 LAMBDA must have its explicit captures already. */
129
130 tree
131 begin_lambda_type (tree lambda)
132 {
133 tree type;
134
135 {
136 /* Unique name. This is just like an unnamed class, but we cannot use
137 make_anon_name because of certain checks against TYPE_ANONYMOUS_P. */
138 tree name;
139 name = make_lambda_name ();
140
141 /* Create the new RECORD_TYPE for this lambda. */
142 type = xref_tag (/*tag_code=*/record_type,
143 name,
144 /*scope=*/ts_lambda,
145 /*template_header_p=*/false);
146 if (type == error_mark_node)
147 return error_mark_node;
148 }
149
150 /* Designate it as a struct so that we can use aggregate initialization. */
151 CLASSTYPE_DECLARED_CLASS (type) = false;
152
153 /* Cross-reference the expression and the type. */
154 LAMBDA_EXPR_CLOSURE (lambda) = type;
155 CLASSTYPE_LAMBDA_EXPR (type) = lambda;
156
157 /* Clear base types. */
158 xref_basetypes (type, /*bases=*/NULL_TREE);
159
160 /* Start the class. */
161 type = begin_class_definition (type);
162
163 return type;
164 }
165
166 /* Returns the type to use for the return type of the operator() of a
167 closure class. */
168
169 tree
170 lambda_return_type (tree expr)
171 {
172 if (expr == NULL_TREE)
173 return void_type_node;
174 if (type_unknown_p (expr)
175 || BRACE_ENCLOSED_INITIALIZER_P (expr))
176 {
177 cxx_incomplete_type_error (expr, TREE_TYPE (expr));
178 return void_type_node;
179 }
180 gcc_checking_assert (!type_dependent_expression_p (expr));
181 return cv_unqualified (type_decays_to (unlowered_expr_type (expr)));
182 }
183
184 /* Given a LAMBDA_EXPR or closure type LAMBDA, return the op() of the
185 closure type. */
186
187 tree
188 lambda_function (tree lambda)
189 {
190 tree type;
191 if (TREE_CODE (lambda) == LAMBDA_EXPR)
192 type = LAMBDA_EXPR_CLOSURE (lambda);
193 else
194 type = lambda;
195 gcc_assert (LAMBDA_TYPE_P (type));
196 /* Don't let debug_tree cause instantiation. */
197 if (CLASSTYPE_TEMPLATE_INSTANTIATION (type)
198 && !COMPLETE_OR_OPEN_TYPE_P (type))
199 return NULL_TREE;
200 lambda = lookup_member (type, ansi_opname (CALL_EXPR),
201 /*protect=*/0, /*want_type=*/false,
202 tf_warning_or_error);
203 if (lambda)
204 lambda = STRIP_TEMPLATE (get_first_fn (lambda));
205 return lambda;
206 }
207
208 /* Returns the type to use for the FIELD_DECL corresponding to the
209 capture of EXPR.
210 The caller should add REFERENCE_TYPE for capture by reference. */
211
212 tree
213 lambda_capture_field_type (tree expr, bool explicit_init_p)
214 {
215 tree type;
216 if (explicit_init_p)
217 {
218 type = make_auto ();
219 type = do_auto_deduction (type, expr, type);
220 }
221 else
222 type = non_reference (unlowered_expr_type (expr));
223 if (type_dependent_expression_p (expr)
224 && !is_this_parameter (tree_strip_nop_conversions (expr)))
225 {
226 type = cxx_make_type (DECLTYPE_TYPE);
227 DECLTYPE_TYPE_EXPR (type) = expr;
228 DECLTYPE_FOR_LAMBDA_CAPTURE (type) = true;
229 DECLTYPE_FOR_INIT_CAPTURE (type) = explicit_init_p;
230 SET_TYPE_STRUCTURAL_EQUALITY (type);
231 }
232 return type;
233 }
234
235 /* Returns true iff DECL is a lambda capture proxy variable created by
236 build_capture_proxy. */
237
238 bool
239 is_capture_proxy (tree decl)
240 {
241 return (VAR_P (decl)
242 && DECL_HAS_VALUE_EXPR_P (decl)
243 && !DECL_ANON_UNION_VAR_P (decl)
244 && LAMBDA_FUNCTION_P (DECL_CONTEXT (decl)));
245 }
246
247 /* Returns true iff DECL is a capture proxy for a normal capture
248 (i.e. without explicit initializer). */
249
250 bool
251 is_normal_capture_proxy (tree decl)
252 {
253 if (!is_capture_proxy (decl))
254 /* It's not a capture proxy. */
255 return false;
256
257 if (variably_modified_type_p (TREE_TYPE (decl), NULL_TREE))
258 /* VLA capture. */
259 return true;
260
261 /* It is a capture proxy, is it a normal capture? */
262 tree val = DECL_VALUE_EXPR (decl);
263 if (val == error_mark_node)
264 return true;
265
266 gcc_assert (TREE_CODE (val) == COMPONENT_REF);
267 val = TREE_OPERAND (val, 1);
268 return DECL_NORMAL_CAPTURE_P (val);
269 }
270
271 /* VAR is a capture proxy created by build_capture_proxy; add it to the
272 current function, which is the operator() for the appropriate lambda. */
273
274 void
275 insert_capture_proxy (tree var)
276 {
277 cp_binding_level *b;
278 tree stmt_list;
279
280 /* Put the capture proxy in the extra body block so that it won't clash
281 with a later local variable. */
282 b = current_binding_level;
283 for (;;)
284 {
285 cp_binding_level *n = b->level_chain;
286 if (n->kind == sk_function_parms)
287 break;
288 b = n;
289 }
290 pushdecl_with_scope (var, b, false);
291
292 /* And put a DECL_EXPR in the STATEMENT_LIST for the same block. */
293 var = build_stmt (DECL_SOURCE_LOCATION (var), DECL_EXPR, var);
294 stmt_list = (*stmt_list_stack)[1];
295 gcc_assert (stmt_list);
296 append_to_statement_list_force (var, &stmt_list);
297 }
298
299 /* We've just finished processing a lambda; if the containing scope is also
300 a lambda, insert any capture proxies that were created while processing
301 the nested lambda. */
302
303 void
304 insert_pending_capture_proxies (void)
305 {
306 tree lam;
307 vec<tree, va_gc> *proxies;
308 unsigned i;
309
310 if (!current_function_decl || !LAMBDA_FUNCTION_P (current_function_decl))
311 return;
312
313 lam = CLASSTYPE_LAMBDA_EXPR (DECL_CONTEXT (current_function_decl));
314 proxies = LAMBDA_EXPR_PENDING_PROXIES (lam);
315 for (i = 0; i < vec_safe_length (proxies); ++i)
316 {
317 tree var = (*proxies)[i];
318 insert_capture_proxy (var);
319 }
320 release_tree_vector (LAMBDA_EXPR_PENDING_PROXIES (lam));
321 LAMBDA_EXPR_PENDING_PROXIES (lam) = NULL;
322 }
323
324 /* Given REF, a COMPONENT_REF designating a field in the lambda closure,
325 return the type we want the proxy to have: the type of the field itself,
326 with added const-qualification if the lambda isn't mutable and the
327 capture is by value. */
328
329 tree
330 lambda_proxy_type (tree ref)
331 {
332 tree type;
333 if (ref == error_mark_node)
334 return error_mark_node;
335 if (REFERENCE_REF_P (ref))
336 ref = TREE_OPERAND (ref, 0);
337 gcc_assert (TREE_CODE (ref) == COMPONENT_REF);
338 type = TREE_TYPE (ref);
339 if (!type || WILDCARD_TYPE_P (non_reference (type)))
340 {
341 type = cxx_make_type (DECLTYPE_TYPE);
342 DECLTYPE_TYPE_EXPR (type) = ref;
343 DECLTYPE_FOR_LAMBDA_PROXY (type) = true;
344 SET_TYPE_STRUCTURAL_EQUALITY (type);
345 }
346 if (DECL_PACK_P (TREE_OPERAND (ref, 1)))
347 type = make_pack_expansion (type);
348 return type;
349 }
350
351 /* MEMBER is a capture field in a lambda closure class. Now that we're
352 inside the operator(), build a placeholder var for future lookups and
353 debugging. */
354
355 tree
356 build_capture_proxy (tree member)
357 {
358 tree var, object, fn, closure, name, lam, type;
359
360 if (PACK_EXPANSION_P (member))
361 member = PACK_EXPANSION_PATTERN (member);
362
363 closure = DECL_CONTEXT (member);
364 fn = lambda_function (closure);
365 lam = CLASSTYPE_LAMBDA_EXPR (closure);
366
367 /* The proxy variable forwards to the capture field. */
368 object = build_fold_indirect_ref (DECL_ARGUMENTS (fn));
369 object = finish_non_static_data_member (member, object, NULL_TREE);
370 if (REFERENCE_REF_P (object))
371 object = TREE_OPERAND (object, 0);
372
373 /* Remove the __ inserted by add_capture. */
374 name = get_identifier (IDENTIFIER_POINTER (DECL_NAME (member)) + 2);
375
376 type = lambda_proxy_type (object);
377
378 if (DECL_VLA_CAPTURE_P (member))
379 {
380 /* Rebuild the VLA type from the pointer and maxindex. */
381 tree field = next_initializable_field (TYPE_FIELDS (type));
382 tree ptr = build_simple_component_ref (object, field);
383 field = next_initializable_field (DECL_CHAIN (field));
384 tree max = build_simple_component_ref (object, field);
385 type = build_cplus_array_type (TREE_TYPE (TREE_TYPE (ptr)),
386 build_index_type (max));
387 type = build_reference_type (type);
388 REFERENCE_VLA_OK (type) = true;
389 object = convert (type, ptr);
390 }
391
392 var = build_decl (input_location, VAR_DECL, name, type);
393 SET_DECL_VALUE_EXPR (var, object);
394 DECL_HAS_VALUE_EXPR_P (var) = 1;
395 DECL_ARTIFICIAL (var) = 1;
396 TREE_USED (var) = 1;
397 DECL_CONTEXT (var) = fn;
398
399 if (name == this_identifier)
400 {
401 gcc_assert (LAMBDA_EXPR_THIS_CAPTURE (lam) == member);
402 LAMBDA_EXPR_THIS_CAPTURE (lam) = var;
403 }
404
405 if (fn == current_function_decl)
406 insert_capture_proxy (var);
407 else
408 vec_safe_push (LAMBDA_EXPR_PENDING_PROXIES (lam), var);
409
410 return var;
411 }
412
413 /* Return a struct containing a pointer and a length for lambda capture of
414 an array of runtime length. */
415
416 static tree
417 vla_capture_type (tree array_type)
418 {
419 static tree ptr_id, max_id;
420 tree type = xref_tag (record_type, make_anon_name (), ts_current, false);
421 xref_basetypes (type, NULL_TREE);
422 type = begin_class_definition (type);
423 if (!ptr_id)
424 {
425 ptr_id = get_identifier ("ptr");
426 max_id = get_identifier ("max");
427 }
428 tree ptrtype = build_pointer_type (TREE_TYPE (array_type));
429 tree field = build_decl (input_location, FIELD_DECL, ptr_id, ptrtype);
430 finish_member_declaration (field);
431 field = build_decl (input_location, FIELD_DECL, max_id, sizetype);
432 finish_member_declaration (field);
433 return finish_struct (type, NULL_TREE);
434 }
435
436 /* From an ID and INITIALIZER, create a capture (by reference if
437 BY_REFERENCE_P is true), add it to the capture-list for LAMBDA,
438 and return it. */
439
440 tree
441 add_capture (tree lambda, tree id, tree orig_init, bool by_reference_p,
442 bool explicit_init_p)
443 {
444 char *buf;
445 tree type, member, name;
446 bool vla = false;
447 bool variadic = false;
448 tree initializer = orig_init;
449
450 if (PACK_EXPANSION_P (initializer))
451 {
452 initializer = PACK_EXPANSION_PATTERN (initializer);
453 variadic = true;
454 }
455
456 if (TREE_CODE (initializer) == TREE_LIST)
457 initializer = build_x_compound_expr_from_list (initializer, ELK_INIT,
458 tf_warning_or_error);
459 type = TREE_TYPE (initializer);
460 if (type == error_mark_node)
461 return error_mark_node;
462
463 if (array_of_runtime_bound_p (type))
464 {
465 vla = true;
466 if (!by_reference_p)
467 error ("array of runtime bound cannot be captured by copy, "
468 "only by reference");
469
470 /* For a VLA, we capture the address of the first element and the
471 maximum index, and then reconstruct the VLA for the proxy. */
472 tree elt = cp_build_array_ref (input_location, initializer,
473 integer_zero_node, tf_warning_or_error);
474 initializer = build_constructor_va (init_list_type_node, 2,
475 NULL_TREE, build_address (elt),
476 NULL_TREE, array_type_nelts (type));
477 type = vla_capture_type (type);
478 }
479 else if (!dependent_type_p (type)
480 && variably_modified_type_p (type, NULL_TREE))
481 {
482 error ("capture of variable-size type %qT that is not an N3639 array "
483 "of runtime bound", type);
484 if (TREE_CODE (type) == ARRAY_TYPE
485 && variably_modified_type_p (TREE_TYPE (type), NULL_TREE))
486 inform (input_location, "because the array element type %qT has "
487 "variable size", TREE_TYPE (type));
488 type = error_mark_node;
489 }
490 else
491 {
492 type = lambda_capture_field_type (initializer, explicit_init_p);
493 if (by_reference_p)
494 {
495 type = build_reference_type (type);
496 if (!dependent_type_p (type) && !real_lvalue_p (initializer))
497 error ("cannot capture %qE by reference", initializer);
498 }
499 else
500 {
501 /* Capture by copy requires a complete type. */
502 type = complete_type (type);
503 if (!dependent_type_p (type) && !COMPLETE_TYPE_P (type))
504 {
505 error ("capture by copy of incomplete type %qT", type);
506 cxx_incomplete_type_inform (type);
507 return error_mark_node;
508 }
509 }
510 }
511
512 /* Add __ to the beginning of the field name so that user code
513 won't find the field with name lookup. We can't just leave the name
514 unset because template instantiation uses the name to find
515 instantiated fields. */
516 buf = (char *) alloca (IDENTIFIER_LENGTH (id) + 3);
517 buf[1] = buf[0] = '_';
518 memcpy (buf + 2, IDENTIFIER_POINTER (id),
519 IDENTIFIER_LENGTH (id) + 1);
520 name = get_identifier (buf);
521
522 /* If TREE_TYPE isn't set, we're still in the introducer, so check
523 for duplicates. */
524 if (!LAMBDA_EXPR_CLOSURE (lambda))
525 {
526 if (IDENTIFIER_MARKED (name))
527 {
528 pedwarn (input_location, 0,
529 "already captured %qD in lambda expression", id);
530 return NULL_TREE;
531 }
532 IDENTIFIER_MARKED (name) = true;
533 }
534
535 if (variadic)
536 type = make_pack_expansion (type);
537
538 /* Make member variable. */
539 member = build_decl (input_location, FIELD_DECL, name, type);
540 DECL_VLA_CAPTURE_P (member) = vla;
541
542 if (!explicit_init_p)
543 /* Normal captures are invisible to name lookup but uses are replaced
544 with references to the capture field; we implement this by only
545 really making them invisible in unevaluated context; see
546 qualify_lookup. For now, let's make explicitly initialized captures
547 always visible. */
548 DECL_NORMAL_CAPTURE_P (member) = true;
549
550 if (id == this_identifier)
551 LAMBDA_EXPR_THIS_CAPTURE (lambda) = member;
552
553 /* Add it to the appropriate closure class if we've started it. */
554 if (current_class_type
555 && current_class_type == LAMBDA_EXPR_CLOSURE (lambda))
556 finish_member_declaration (member);
557
558 tree listmem = member;
559 if (variadic)
560 {
561 listmem = make_pack_expansion (member);
562 initializer = orig_init;
563 }
564 LAMBDA_EXPR_CAPTURE_LIST (lambda)
565 = tree_cons (listmem, initializer, LAMBDA_EXPR_CAPTURE_LIST (lambda));
566
567 if (LAMBDA_EXPR_CLOSURE (lambda))
568 return build_capture_proxy (member);
569 /* For explicit captures we haven't started the function yet, so we wait
570 and build the proxy from cp_parser_lambda_body. */
571 return NULL_TREE;
572 }
573
574 /* Register all the capture members on the list CAPTURES, which is the
575 LAMBDA_EXPR_CAPTURE_LIST for the lambda after the introducer. */
576
577 void
578 register_capture_members (tree captures)
579 {
580 if (captures == NULL_TREE)
581 return;
582
583 register_capture_members (TREE_CHAIN (captures));
584
585 tree field = TREE_PURPOSE (captures);
586 if (PACK_EXPANSION_P (field))
587 field = PACK_EXPANSION_PATTERN (field);
588
589 /* We set this in add_capture to avoid duplicates. */
590 IDENTIFIER_MARKED (DECL_NAME (field)) = false;
591 finish_member_declaration (field);
592 }
593
594 /* Similar to add_capture, except this works on a stack of nested lambdas.
595 BY_REFERENCE_P in this case is derived from the default capture mode.
596 Returns the capture for the lambda at the bottom of the stack. */
597
598 tree
599 add_default_capture (tree lambda_stack, tree id, tree initializer)
600 {
601 bool this_capture_p = (id == this_identifier);
602
603 tree var = NULL_TREE;
604
605 tree saved_class_type = current_class_type;
606
607 tree node;
608
609 for (node = lambda_stack;
610 node;
611 node = TREE_CHAIN (node))
612 {
613 tree lambda = TREE_VALUE (node);
614
615 current_class_type = LAMBDA_EXPR_CLOSURE (lambda);
616 if (DECL_PACK_P (initializer))
617 initializer = make_pack_expansion (initializer);
618 var = add_capture (lambda,
619 id,
620 initializer,
621 /*by_reference_p=*/
622 (!this_capture_p
623 && (LAMBDA_EXPR_DEFAULT_CAPTURE_MODE (lambda)
624 == CPLD_REFERENCE)),
625 /*explicit_init_p=*/false);
626 initializer = convert_from_reference (var);
627 }
628
629 current_class_type = saved_class_type;
630
631 return var;
632 }
633
634 /* Return the capture pertaining to a use of 'this' in LAMBDA, in the
635 form of an INDIRECT_REF, possibly adding it through default
636 capturing, if ADD_CAPTURE_P is true. */
637
638 tree
639 lambda_expr_this_capture (tree lambda, bool add_capture_p)
640 {
641 tree result;
642
643 tree this_capture = LAMBDA_EXPR_THIS_CAPTURE (lambda);
644
645 /* In unevaluated context this isn't an odr-use, so don't capture. */
646 if (cp_unevaluated_operand)
647 add_capture_p = false;
648
649 /* Try to default capture 'this' if we can. */
650 if (!this_capture
651 && (!add_capture_p
652 || LAMBDA_EXPR_DEFAULT_CAPTURE_MODE (lambda) != CPLD_NONE))
653 {
654 tree lambda_stack = NULL_TREE;
655 tree init = NULL_TREE;
656
657 /* If we are in a lambda function, we can move out until we hit:
658 1. a non-lambda function or NSDMI,
659 2. a lambda function capturing 'this', or
660 3. a non-default capturing lambda function. */
661 for (tree tlambda = lambda; ;)
662 {
663 lambda_stack = tree_cons (NULL_TREE,
664 tlambda,
665 lambda_stack);
666
667 if (LAMBDA_EXPR_EXTRA_SCOPE (tlambda)
668 && TREE_CODE (LAMBDA_EXPR_EXTRA_SCOPE (tlambda)) == FIELD_DECL)
669 {
670 /* In an NSDMI, we don't have a function to look up the decl in,
671 but the fake 'this' pointer that we're using for parsing is
672 in scope_chain. */
673 init = scope_chain->x_current_class_ptr;
674 gcc_checking_assert
675 (init && (TREE_TYPE (TREE_TYPE (init))
676 == current_nonlambda_class_type ()));
677 break;
678 }
679
680 tree closure_decl = TYPE_NAME (LAMBDA_EXPR_CLOSURE (tlambda));
681 tree containing_function = decl_function_context (closure_decl);
682
683 if (containing_function == NULL_TREE)
684 /* We ran out of scopes; there's no 'this' to capture. */
685 break;
686
687 if (!LAMBDA_FUNCTION_P (containing_function))
688 {
689 /* We found a non-lambda function. */
690 if (DECL_NONSTATIC_MEMBER_FUNCTION_P (containing_function))
691 /* First parameter is 'this'. */
692 init = DECL_ARGUMENTS (containing_function);
693 break;
694 }
695
696 tlambda
697 = CLASSTYPE_LAMBDA_EXPR (DECL_CONTEXT (containing_function));
698
699 if (LAMBDA_EXPR_THIS_CAPTURE (tlambda))
700 {
701 /* An outer lambda has already captured 'this'. */
702 init = LAMBDA_EXPR_THIS_CAPTURE (tlambda);
703 break;
704 }
705
706 if (LAMBDA_EXPR_DEFAULT_CAPTURE_MODE (tlambda) == CPLD_NONE)
707 /* An outer lambda won't let us capture 'this'. */
708 break;
709 }
710
711 if (init)
712 {
713 if (add_capture_p)
714 this_capture = add_default_capture (lambda_stack,
715 /*id=*/this_identifier,
716 init);
717 else
718 this_capture = init;
719 }
720 }
721
722 if (cp_unevaluated_operand)
723 result = this_capture;
724 else if (!this_capture)
725 {
726 if (add_capture_p)
727 {
728 error ("%<this%> was not captured for this lambda function");
729 result = error_mark_node;
730 }
731 else
732 result = NULL_TREE;
733 }
734 else
735 {
736 /* To make sure that current_class_ref is for the lambda. */
737 gcc_assert (TYPE_MAIN_VARIANT (TREE_TYPE (current_class_ref))
738 == LAMBDA_EXPR_CLOSURE (lambda));
739
740 result = this_capture;
741
742 /* If 'this' is captured, each use of 'this' is transformed into an
743 access to the corresponding unnamed data member of the closure
744 type cast (_expr.cast_ 5.4) to the type of 'this'. [ The cast
745 ensures that the transformed expression is an rvalue. ] */
746 result = rvalue (result);
747 }
748
749 return result;
750 }
751
752 /* We don't want to capture 'this' until we know we need it, i.e. after
753 overload resolution has chosen a non-static member function. At that
754 point we call this function to turn a dummy object into a use of the
755 'this' capture. */
756
757 tree
758 maybe_resolve_dummy (tree object, bool add_capture_p)
759 {
760 if (!is_dummy_object (object))
761 return object;
762
763 tree type = TYPE_MAIN_VARIANT (TREE_TYPE (object));
764 gcc_assert (!TYPE_PTR_P (type));
765
766 if (type != current_class_type
767 && current_class_type
768 && LAMBDA_TYPE_P (current_class_type)
769 && lambda_function (current_class_type)
770 && DERIVED_FROM_P (type, current_nonlambda_class_type ()))
771 {
772 /* In a lambda, need to go through 'this' capture. */
773 tree lam = CLASSTYPE_LAMBDA_EXPR (current_class_type);
774 tree cap = lambda_expr_this_capture (lam, add_capture_p);
775 if (cap && cap != error_mark_node)
776 object = build_x_indirect_ref (EXPR_LOCATION (object), cap,
777 RO_NULL, tf_warning_or_error);
778 }
779
780 return object;
781 }
782
783 /* Returns the innermost non-lambda function. */
784
785 tree
786 current_nonlambda_function (void)
787 {
788 tree fn = current_function_decl;
789 while (fn && LAMBDA_FUNCTION_P (fn))
790 fn = decl_function_context (fn);
791 return fn;
792 }
793
794 /* Returns the method basetype of the innermost non-lambda function, or
795 NULL_TREE if none. */
796
797 tree
798 nonlambda_method_basetype (void)
799 {
800 tree fn, type;
801 if (!current_class_ref)
802 return NULL_TREE;
803
804 type = current_class_type;
805 if (!LAMBDA_TYPE_P (type))
806 return type;
807
808 /* Find the nearest enclosing non-lambda function. */
809 fn = TYPE_NAME (type);
810 do
811 fn = decl_function_context (fn);
812 while (fn && LAMBDA_FUNCTION_P (fn));
813
814 if (!fn || !DECL_NONSTATIC_MEMBER_FUNCTION_P (fn))
815 return NULL_TREE;
816
817 return TYPE_METHOD_BASETYPE (TREE_TYPE (fn));
818 }
819
820 /* Like current_scope, but looking through lambdas. */
821
822 tree
823 current_nonlambda_scope (void)
824 {
825 tree scope = current_scope ();
826 for (;;)
827 {
828 if (TREE_CODE (scope) == FUNCTION_DECL
829 && LAMBDA_FUNCTION_P (scope))
830 {
831 scope = CP_TYPE_CONTEXT (DECL_CONTEXT (scope));
832 continue;
833 }
834 else if (LAMBDA_TYPE_P (scope))
835 {
836 scope = CP_TYPE_CONTEXT (scope);
837 continue;
838 }
839 break;
840 }
841 return scope;
842 }
843
844 /* Helper function for maybe_add_lambda_conv_op; build a CALL_EXPR with
845 indicated FN and NARGS, but do not initialize the return type or any of the
846 argument slots. */
847
848 static tree
849 prepare_op_call (tree fn, int nargs)
850 {
851 tree t;
852
853 t = build_vl_exp (CALL_EXPR, nargs + 3);
854 CALL_EXPR_FN (t) = fn;
855 CALL_EXPR_STATIC_CHAIN (t) = NULL;
856
857 return t;
858 }
859
860 /* If the closure TYPE has a static op(), also add a conversion to function
861 pointer. */
862
863 void
864 maybe_add_lambda_conv_op (tree type)
865 {
866 bool nested = (cfun != NULL);
867 bool nested_def = decl_function_context (TYPE_MAIN_DECL (type));
868 tree callop = lambda_function (type);
869
870 if (LAMBDA_EXPR_CAPTURE_LIST (CLASSTYPE_LAMBDA_EXPR (type)) != NULL_TREE)
871 return;
872
873 if (processing_template_decl)
874 return;
875
876 bool const generic_lambda_p
877 = (DECL_TEMPLATE_INFO (callop)
878 && DECL_TEMPLATE_RESULT (DECL_TI_TEMPLATE (callop)) == callop);
879
880 if (!generic_lambda_p && DECL_INITIAL (callop) == NULL_TREE)
881 {
882 /* If the op() wasn't instantiated due to errors, give up. */
883 gcc_assert (errorcount || sorrycount);
884 return;
885 }
886
887 /* Non-template conversion operators are defined directly with build_call_a
888 and using DIRECT_ARGVEC for arguments (including 'this'). Templates are
889 deferred and the CALL is built in-place. In the case of a deduced return
890 call op, the decltype expression, DECLTYPE_CALL, used as a substitute for
891 the return type is also built in-place. The arguments of DECLTYPE_CALL in
892 the return expression may differ in flags from those in the body CALL. In
893 particular, parameter pack expansions are marked PACK_EXPANSION_LOCAL_P in
894 the body CALL, but not in DECLTYPE_CALL. */
895
896 vec<tree, va_gc> *direct_argvec = 0;
897 tree decltype_call = 0, call = 0;
898 tree optype = TREE_TYPE (callop);
899 tree fn_result = TREE_TYPE (optype);
900
901 if (generic_lambda_p)
902 {
903 /* Prepare the dependent member call for the static member function
904 '_FUN' and, potentially, prepare another call to be used in a decltype
905 return expression for a deduced return call op to allow for simple
906 implementation of the conversion operator. */
907
908 tree instance = build_nop (type, null_pointer_node);
909 tree objfn = build_min (COMPONENT_REF, NULL_TREE,
910 instance, DECL_NAME (callop), NULL_TREE);
911 int nargs = list_length (DECL_ARGUMENTS (callop)) - 1;
912
913 call = prepare_op_call (objfn, nargs);
914 if (type_uses_auto (fn_result))
915 decltype_call = prepare_op_call (objfn, nargs);
916 }
917 else
918 {
919 direct_argvec = make_tree_vector ();
920 direct_argvec->quick_push (build1 (NOP_EXPR,
921 TREE_TYPE (DECL_ARGUMENTS (callop)),
922 null_pointer_node));
923 }
924
925 /* Copy CALLOP's argument list (as per 'copy_list') as FN_ARGS in order to
926 declare the static member function "_FUN" below. For each arg append to
927 DIRECT_ARGVEC (for the non-template case) or populate the pre-allocated
928 call args (for the template case). If a parameter pack is found, expand
929 it, flagging it as PACK_EXPANSION_LOCAL_P for the body call. */
930
931 tree fn_args = NULL_TREE;
932 {
933 int ix = 0;
934 tree src = DECL_CHAIN (DECL_ARGUMENTS (callop));
935 tree tgt;
936
937 while (src)
938 {
939 tree new_node = copy_node (src);
940
941 if (!fn_args)
942 fn_args = tgt = new_node;
943 else
944 {
945 TREE_CHAIN (tgt) = new_node;
946 tgt = new_node;
947 }
948
949 mark_exp_read (tgt);
950
951 if (generic_lambda_p)
952 {
953 if (DECL_PACK_P (tgt))
954 {
955 tree a = make_pack_expansion (tgt);
956 if (decltype_call)
957 CALL_EXPR_ARG (decltype_call, ix) = copy_node (a);
958 PACK_EXPANSION_LOCAL_P (a) = true;
959 CALL_EXPR_ARG (call, ix) = a;
960 }
961 else
962 {
963 tree a = convert_from_reference (tgt);
964 CALL_EXPR_ARG (call, ix) = a;
965 if (decltype_call)
966 CALL_EXPR_ARG (decltype_call, ix) = copy_node (a);
967 }
968 ++ix;
969 }
970 else
971 vec_safe_push (direct_argvec, tgt);
972
973 src = TREE_CHAIN (src);
974 }
975 }
976
977
978 if (generic_lambda_p)
979 {
980 if (decltype_call)
981 {
982 ++processing_template_decl;
983 fn_result = finish_decltype_type
984 (decltype_call, /*id_expression_or_member_access_p=*/false,
985 tf_warning_or_error);
986 --processing_template_decl;
987 }
988 }
989 else
990 call = build_call_a (callop,
991 direct_argvec->length (),
992 direct_argvec->address ());
993
994 CALL_FROM_THUNK_P (call) = 1;
995
996 tree stattype = build_function_type (fn_result, FUNCTION_ARG_CHAIN (callop));
997 stattype = (cp_build_type_attribute_variant
998 (stattype, TYPE_ATTRIBUTES (optype)));
999
1000 /* First build up the conversion op. */
1001
1002 tree rettype = build_pointer_type (stattype);
1003 tree name = mangle_conv_op_name_for_type (rettype);
1004 tree thistype = cp_build_qualified_type (type, TYPE_QUAL_CONST);
1005 tree fntype = build_method_type_directly (thistype, rettype, void_list_node);
1006 tree convfn = build_lang_decl (FUNCTION_DECL, name, fntype);
1007 tree fn = convfn;
1008 DECL_SOURCE_LOCATION (fn) = DECL_SOURCE_LOCATION (callop);
1009
1010 if (TARGET_PTRMEMFUNC_VBIT_LOCATION == ptrmemfunc_vbit_in_pfn
1011 && DECL_ALIGN (fn) < 2 * BITS_PER_UNIT)
1012 DECL_ALIGN (fn) = 2 * BITS_PER_UNIT;
1013
1014 SET_OVERLOADED_OPERATOR_CODE (fn, TYPE_EXPR);
1015 grokclassfn (type, fn, NO_SPECIAL);
1016 set_linkage_according_to_type (type, fn);
1017 rest_of_decl_compilation (fn, toplevel_bindings_p (), at_eof);
1018 DECL_IN_AGGR_P (fn) = 1;
1019 DECL_ARTIFICIAL (fn) = 1;
1020 DECL_NOT_REALLY_EXTERN (fn) = 1;
1021 DECL_DECLARED_INLINE_P (fn) = 1;
1022 DECL_ARGUMENTS (fn) = build_this_parm (fntype, TYPE_QUAL_CONST);
1023 if (nested_def)
1024 DECL_INTERFACE_KNOWN (fn) = 1;
1025
1026 if (generic_lambda_p)
1027 fn = add_inherited_template_parms (fn, DECL_TI_TEMPLATE (callop));
1028
1029 add_method (type, fn, NULL_TREE);
1030
1031 /* Generic thunk code fails for varargs; we'll complain in mark_used if
1032 the conversion op is used. */
1033 if (varargs_function_p (callop))
1034 {
1035 DECL_DELETED_FN (fn) = 1;
1036 return;
1037 }
1038
1039 /* Now build up the thunk to be returned. */
1040
1041 name = get_identifier ("_FUN");
1042 tree statfn = build_lang_decl (FUNCTION_DECL, name, stattype);
1043 fn = statfn;
1044 DECL_SOURCE_LOCATION (fn) = DECL_SOURCE_LOCATION (callop);
1045 if (TARGET_PTRMEMFUNC_VBIT_LOCATION == ptrmemfunc_vbit_in_pfn
1046 && DECL_ALIGN (fn) < 2 * BITS_PER_UNIT)
1047 DECL_ALIGN (fn) = 2 * BITS_PER_UNIT;
1048 grokclassfn (type, fn, NO_SPECIAL);
1049 set_linkage_according_to_type (type, fn);
1050 rest_of_decl_compilation (fn, toplevel_bindings_p (), at_eof);
1051 DECL_IN_AGGR_P (fn) = 1;
1052 DECL_ARTIFICIAL (fn) = 1;
1053 DECL_NOT_REALLY_EXTERN (fn) = 1;
1054 DECL_DECLARED_INLINE_P (fn) = 1;
1055 DECL_STATIC_FUNCTION_P (fn) = 1;
1056 DECL_ARGUMENTS (fn) = fn_args;
1057 for (tree arg = fn_args; arg; arg = DECL_CHAIN (arg))
1058 {
1059 /* Avoid duplicate -Wshadow warnings. */
1060 DECL_NAME (arg) = NULL_TREE;
1061 DECL_CONTEXT (arg) = fn;
1062 }
1063 if (nested_def)
1064 DECL_INTERFACE_KNOWN (fn) = 1;
1065
1066 if (generic_lambda_p)
1067 fn = add_inherited_template_parms (fn, DECL_TI_TEMPLATE (callop));
1068
1069 add_method (type, fn, NULL_TREE);
1070
1071 if (nested)
1072 push_function_context ();
1073 else
1074 /* Still increment function_depth so that we don't GC in the
1075 middle of an expression. */
1076 ++function_depth;
1077
1078 /* Generate the body of the thunk. */
1079
1080 start_preparsed_function (statfn, NULL_TREE,
1081 SF_PRE_PARSED | SF_INCLASS_INLINE);
1082 if (DECL_ONE_ONLY (statfn))
1083 {
1084 /* Put the thunk in the same comdat group as the call op. */
1085 cgraph_node::get_create (statfn)->add_to_same_comdat_group
1086 (cgraph_node::get_create (callop));
1087 }
1088 tree body = begin_function_body ();
1089 tree compound_stmt = begin_compound_stmt (0);
1090 if (!generic_lambda_p)
1091 {
1092 set_flags_from_callee (call);
1093 if (MAYBE_CLASS_TYPE_P (TREE_TYPE (call)))
1094 call = build_cplus_new (TREE_TYPE (call), call, tf_warning_or_error);
1095 }
1096 call = convert_from_reference (call);
1097 finish_return_stmt (call);
1098
1099 finish_compound_stmt (compound_stmt);
1100 finish_function_body (body);
1101
1102 fn = finish_function (/*inline*/2);
1103 if (!generic_lambda_p)
1104 expand_or_defer_fn (fn);
1105
1106 /* Generate the body of the conversion op. */
1107
1108 start_preparsed_function (convfn, NULL_TREE,
1109 SF_PRE_PARSED | SF_INCLASS_INLINE);
1110 body = begin_function_body ();
1111 compound_stmt = begin_compound_stmt (0);
1112
1113 /* decl_needed_p needs to see that it's used. */
1114 TREE_USED (statfn) = 1;
1115 finish_return_stmt (decay_conversion (statfn, tf_warning_or_error));
1116
1117 finish_compound_stmt (compound_stmt);
1118 finish_function_body (body);
1119
1120 fn = finish_function (/*inline*/2);
1121 if (!generic_lambda_p)
1122 expand_or_defer_fn (fn);
1123
1124 if (nested)
1125 pop_function_context ();
1126 else
1127 --function_depth;
1128 }
1129
1130 /* Returns true iff VAL is a lambda-related declaration which should
1131 be ignored by unqualified lookup. */
1132
1133 bool
1134 is_lambda_ignored_entity (tree val)
1135 {
1136 /* In unevaluated context, look past normal capture proxies. */
1137 if (cp_unevaluated_operand && is_normal_capture_proxy (val))
1138 return true;
1139
1140 /* Always ignore lambda fields, their names are only for debugging. */
1141 if (TREE_CODE (val) == FIELD_DECL
1142 && CLASSTYPE_LAMBDA_EXPR (DECL_CONTEXT (val)))
1143 return true;
1144
1145 /* None of the lookups that use qualify_lookup want the op() from the
1146 lambda; they want the one from the enclosing class. */
1147 if (TREE_CODE (val) == FUNCTION_DECL && LAMBDA_FUNCTION_P (val))
1148 return true;
1149
1150 return false;
1151 }