]> git.ipfire.org Git - thirdparty/gcc.git/blob - gcc/convert.c
middle-end: Remove truly_noop_truncation check from convert.c
[thirdparty/gcc.git] / gcc / convert.c
1 /* Utility routines for data type conversion for GCC.
2 Copyright (C) 1987-2020 Free Software Foundation, Inc.
3
4 This file is part of GCC.
5
6 GCC is free software; you can redistribute it and/or modify it under
7 the terms of the GNU General Public License as published by the Free
8 Software Foundation; either version 3, or (at your option) any later
9 version.
10
11 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
12 WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
14 for more details.
15
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING3. If not see
18 <http://www.gnu.org/licenses/>. */
19
20
21 /* These routines are somewhat language-independent utility function
22 intended to be called by the language-specific convert () functions. */
23
24 #include "config.h"
25 #include "system.h"
26 #include "coretypes.h"
27 #include "target.h"
28 #include "tree.h"
29 #include "diagnostic-core.h"
30 #include "fold-const.h"
31 #include "stor-layout.h"
32 #include "convert.h"
33 #include "langhooks.h"
34 #include "builtins.h"
35 #include "ubsan.h"
36 #include "stringpool.h"
37 #include "attribs.h"
38 #include "asan.h"
39 #include "selftest.h"
40
41 #define maybe_fold_build1_loc(FOLD_P, LOC, CODE, TYPE, EXPR) \
42 ((FOLD_P) ? fold_build1_loc (LOC, CODE, TYPE, EXPR) \
43 : build1_loc (LOC, CODE, TYPE, EXPR))
44 #define maybe_fold_build2_loc(FOLD_P, LOC, CODE, TYPE, EXPR1, EXPR2) \
45 ((FOLD_P) ? fold_build2_loc (LOC, CODE, TYPE, EXPR1, EXPR2) \
46 : build2_loc (LOC, CODE, TYPE, EXPR1, EXPR2))
47
48 /* Convert EXPR to some pointer or reference type TYPE.
49 EXPR must be pointer, reference, integer, enumeral, or literal zero;
50 in other cases error is called. If FOLD_P is true, try to fold the
51 expression. */
52
53 static tree
54 convert_to_pointer_1 (tree type, tree expr, bool fold_p)
55 {
56 location_t loc = EXPR_LOCATION (expr);
57 if (TREE_TYPE (expr) == type)
58 return expr;
59
60 switch (TREE_CODE (TREE_TYPE (expr)))
61 {
62 case POINTER_TYPE:
63 case REFERENCE_TYPE:
64 {
65 /* If the pointers point to different address spaces, conversion needs
66 to be done via a ADDR_SPACE_CONVERT_EXPR instead of a NOP_EXPR. */
67 addr_space_t to_as = TYPE_ADDR_SPACE (TREE_TYPE (type));
68 addr_space_t from_as = TYPE_ADDR_SPACE (TREE_TYPE (TREE_TYPE (expr)));
69
70 if (to_as == from_as)
71 return maybe_fold_build1_loc (fold_p, loc, NOP_EXPR, type, expr);
72 else
73 return maybe_fold_build1_loc (fold_p, loc, ADDR_SPACE_CONVERT_EXPR,
74 type, expr);
75 }
76
77 case INTEGER_TYPE:
78 case ENUMERAL_TYPE:
79 case BOOLEAN_TYPE:
80 {
81 /* If the input precision differs from the target pointer type
82 precision, first convert the input expression to an integer type of
83 the target precision. Some targets, e.g. VMS, need several pointer
84 sizes to coexist so the latter isn't necessarily POINTER_SIZE. */
85 unsigned int pprec = TYPE_PRECISION (type);
86 unsigned int eprec = TYPE_PRECISION (TREE_TYPE (expr));
87
88 if (eprec != pprec)
89 expr
90 = maybe_fold_build1_loc (fold_p, loc, NOP_EXPR,
91 lang_hooks.types.type_for_size (pprec, 0),
92 expr);
93 }
94 return maybe_fold_build1_loc (fold_p, loc, CONVERT_EXPR, type, expr);
95
96 default:
97 error ("cannot convert to a pointer type");
98 return convert_to_pointer_1 (type, integer_zero_node, fold_p);
99 }
100 }
101
102 /* Subroutine of the various convert_to_*_maybe_fold routines.
103
104 If a location wrapper has been folded to a constant (presumably of
105 a different type), re-wrap the new constant with a location wrapper. */
106
107 tree
108 preserve_any_location_wrapper (tree result, tree orig_expr)
109 {
110 if (CONSTANT_CLASS_P (result) && location_wrapper_p (orig_expr))
111 {
112 if (result == TREE_OPERAND (orig_expr, 0))
113 return orig_expr;
114 else
115 return maybe_wrap_with_location (result, EXPR_LOCATION (orig_expr));
116 }
117
118 return result;
119 }
120
121 /* A wrapper around convert_to_pointer_1 that always folds the
122 expression. */
123
124 tree
125 convert_to_pointer (tree type, tree expr)
126 {
127 return convert_to_pointer_1 (type, expr, true);
128 }
129
130 /* A wrapper around convert_to_pointer_1 that only folds the
131 expression if DOFOLD, or if it is CONSTANT_CLASS_OR_WRAPPER_P. */
132
133 tree
134 convert_to_pointer_maybe_fold (tree type, tree expr, bool dofold)
135 {
136 tree result
137 = convert_to_pointer_1 (type, expr,
138 dofold || CONSTANT_CLASS_OR_WRAPPER_P (expr));
139 return preserve_any_location_wrapper (result, expr);
140 }
141
142 /* Convert EXPR to some floating-point type TYPE.
143
144 EXPR must be float, fixed-point, integer, or enumeral;
145 in other cases error is called. If FOLD_P is true, try to fold
146 the expression. */
147
148 static tree
149 convert_to_real_1 (tree type, tree expr, bool fold_p)
150 {
151 enum built_in_function fcode = builtin_mathfn_code (expr);
152 tree itype = TREE_TYPE (expr);
153 location_t loc = EXPR_LOCATION (expr);
154
155 if (TREE_CODE (expr) == COMPOUND_EXPR)
156 {
157 tree t = convert_to_real_1 (type, TREE_OPERAND (expr, 1), fold_p);
158 if (t == TREE_OPERAND (expr, 1))
159 return expr;
160 return build2_loc (EXPR_LOCATION (expr), COMPOUND_EXPR, TREE_TYPE (t),
161 TREE_OPERAND (expr, 0), t);
162 }
163
164 /* Disable until we figure out how to decide whether the functions are
165 present in runtime. */
166 /* Convert (float)sqrt((double)x) where x is float into sqrtf(x) */
167 if (optimize
168 && (TYPE_MODE (type) == TYPE_MODE (double_type_node)
169 || TYPE_MODE (type) == TYPE_MODE (float_type_node)))
170 {
171 switch (fcode)
172 {
173 #define CASE_MATHFN(FN) case BUILT_IN_##FN: case BUILT_IN_##FN##L:
174 CASE_MATHFN (COSH)
175 CASE_MATHFN (EXP)
176 CASE_MATHFN (EXP10)
177 CASE_MATHFN (EXP2)
178 CASE_MATHFN (EXPM1)
179 CASE_MATHFN (GAMMA)
180 CASE_MATHFN (J0)
181 CASE_MATHFN (J1)
182 CASE_MATHFN (LGAMMA)
183 CASE_MATHFN (POW10)
184 CASE_MATHFN (SINH)
185 CASE_MATHFN (TGAMMA)
186 CASE_MATHFN (Y0)
187 CASE_MATHFN (Y1)
188 /* The above functions may set errno differently with float
189 input or output so this transformation is not safe with
190 -fmath-errno. */
191 if (flag_errno_math)
192 break;
193 gcc_fallthrough ();
194 CASE_MATHFN (ACOS)
195 CASE_MATHFN (ACOSH)
196 CASE_MATHFN (ASIN)
197 CASE_MATHFN (ASINH)
198 CASE_MATHFN (ATAN)
199 CASE_MATHFN (ATANH)
200 CASE_MATHFN (CBRT)
201 CASE_MATHFN (COS)
202 CASE_MATHFN (ERF)
203 CASE_MATHFN (ERFC)
204 CASE_MATHFN (LOG)
205 CASE_MATHFN (LOG10)
206 CASE_MATHFN (LOG2)
207 CASE_MATHFN (LOG1P)
208 CASE_MATHFN (SIN)
209 CASE_MATHFN (TAN)
210 CASE_MATHFN (TANH)
211 /* The above functions are not safe to do this conversion. */
212 if (!flag_unsafe_math_optimizations)
213 break;
214 gcc_fallthrough ();
215 CASE_MATHFN (SQRT)
216 CASE_MATHFN (FABS)
217 CASE_MATHFN (LOGB)
218 #undef CASE_MATHFN
219 if (call_expr_nargs (expr) != 1
220 || !SCALAR_FLOAT_TYPE_P (TREE_TYPE (CALL_EXPR_ARG (expr, 0))))
221 break;
222 {
223 tree arg0 = strip_float_extensions (CALL_EXPR_ARG (expr, 0));
224 tree newtype = type;
225
226 /* We have (outertype)sqrt((innertype)x). Choose the wider mode
227 from the both as the safe type for operation. */
228 if (TYPE_PRECISION (TREE_TYPE (arg0)) > TYPE_PRECISION (type))
229 newtype = TREE_TYPE (arg0);
230
231 /* We consider to convert
232
233 (T1) sqrtT2 ((T2) exprT3)
234 to
235 (T1) sqrtT4 ((T4) exprT3)
236
237 , where T1 is TYPE, T2 is ITYPE, T3 is TREE_TYPE (ARG0),
238 and T4 is NEWTYPE. All those types are of floating-point types.
239 T4 (NEWTYPE) should be narrower than T2 (ITYPE). This conversion
240 is safe only if P1 >= P2*2+2, where P1 and P2 are precisions of
241 T2 and T4. See the following URL for a reference:
242 http://stackoverflow.com/questions/9235456/determining-
243 floating-point-square-root
244 */
245 if ((fcode == BUILT_IN_SQRT || fcode == BUILT_IN_SQRTL)
246 && !flag_unsafe_math_optimizations)
247 {
248 /* The following conversion is unsafe even the precision condition
249 below is satisfied:
250
251 (float) sqrtl ((long double) double_val) -> (float) sqrt (double_val)
252 */
253 if (TYPE_MODE (type) != TYPE_MODE (newtype))
254 break;
255
256 int p1 = REAL_MODE_FORMAT (TYPE_MODE (itype))->p;
257 int p2 = REAL_MODE_FORMAT (TYPE_MODE (newtype))->p;
258 if (p1 < p2 * 2 + 2)
259 break;
260 }
261
262 /* Be careful about integer to fp conversions.
263 These may overflow still. */
264 if (FLOAT_TYPE_P (TREE_TYPE (arg0))
265 && TYPE_PRECISION (newtype) < TYPE_PRECISION (itype)
266 && (TYPE_MODE (newtype) == TYPE_MODE (double_type_node)
267 || TYPE_MODE (newtype) == TYPE_MODE (float_type_node)))
268 {
269 tree fn = mathfn_built_in (newtype, fcode);
270 if (fn)
271 {
272 tree arg = convert_to_real_1 (newtype, arg0, fold_p);
273 expr = build_call_expr (fn, 1, arg);
274 if (newtype == type)
275 return expr;
276 }
277 }
278 }
279 default:
280 break;
281 }
282 }
283
284 /* Propagate the cast into the operation. */
285 if (itype != type && FLOAT_TYPE_P (type))
286 switch (TREE_CODE (expr))
287 {
288 /* Convert (float)-x into -(float)x. This is safe for
289 round-to-nearest rounding mode when the inner type is float. */
290 case ABS_EXPR:
291 case NEGATE_EXPR:
292 if (!flag_rounding_math
293 && FLOAT_TYPE_P (itype)
294 && TYPE_PRECISION (type) < TYPE_PRECISION (itype))
295 {
296 tree arg = convert_to_real_1 (type, TREE_OPERAND (expr, 0),
297 fold_p);
298 return build1 (TREE_CODE (expr), type, arg);
299 }
300 break;
301 default:
302 break;
303 }
304
305 switch (TREE_CODE (TREE_TYPE (expr)))
306 {
307 case REAL_TYPE:
308 /* Ignore the conversion if we don't need to store intermediate
309 results and neither type is a decimal float. */
310 return build1_loc (loc,
311 (flag_float_store
312 || DECIMAL_FLOAT_TYPE_P (type)
313 || DECIMAL_FLOAT_TYPE_P (itype))
314 ? CONVERT_EXPR : NOP_EXPR, type, expr);
315
316 case INTEGER_TYPE:
317 case ENUMERAL_TYPE:
318 case BOOLEAN_TYPE:
319 return build1 (FLOAT_EXPR, type, expr);
320
321 case FIXED_POINT_TYPE:
322 return build1 (FIXED_CONVERT_EXPR, type, expr);
323
324 case COMPLEX_TYPE:
325 return convert (type,
326 maybe_fold_build1_loc (fold_p, loc, REALPART_EXPR,
327 TREE_TYPE (TREE_TYPE (expr)),
328 expr));
329
330 case POINTER_TYPE:
331 case REFERENCE_TYPE:
332 error ("pointer value used where a floating-point was expected");
333 return convert_to_real_1 (type, integer_zero_node, fold_p);
334
335 default:
336 error ("aggregate value used where a floating-point was expected");
337 return convert_to_real_1 (type, integer_zero_node, fold_p);
338 }
339 }
340
341 /* A wrapper around convert_to_real_1 that always folds the
342 expression. */
343
344 tree
345 convert_to_real (tree type, tree expr)
346 {
347 return convert_to_real_1 (type, expr, true);
348 }
349
350 /* A wrapper around convert_to_real_1 that only folds the
351 expression if DOFOLD, or if it is CONSTANT_CLASS_OR_WRAPPER_P. */
352
353 tree
354 convert_to_real_maybe_fold (tree type, tree expr, bool dofold)
355 {
356 tree result
357 = convert_to_real_1 (type, expr,
358 dofold || CONSTANT_CLASS_OR_WRAPPER_P (expr));
359 return preserve_any_location_wrapper (result, expr);
360 }
361
362 /* Try to narrow EX_FORM ARG0 ARG1 in narrowed arg types producing a
363 result in TYPE. */
364
365 static tree
366 do_narrow (location_t loc,
367 enum tree_code ex_form, tree type, tree arg0, tree arg1,
368 tree expr, unsigned inprec, unsigned outprec, bool dofold)
369 {
370 /* Do the arithmetic in type TYPEX,
371 then convert result to TYPE. */
372 tree typex = type;
373
374 /* Can't do arithmetic in enumeral types
375 so use an integer type that will hold the values. */
376 if (TREE_CODE (typex) == ENUMERAL_TYPE)
377 typex = lang_hooks.types.type_for_size (TYPE_PRECISION (typex),
378 TYPE_UNSIGNED (typex));
379
380 /* The type demotion below might cause doing unsigned arithmetic
381 instead of signed, and thus hide overflow bugs. */
382 if ((ex_form == PLUS_EXPR || ex_form == MINUS_EXPR)
383 && !TYPE_UNSIGNED (typex)
384 && sanitize_flags_p (SANITIZE_SI_OVERFLOW))
385 return NULL_TREE;
386
387 /* But now perhaps TYPEX is as wide as INPREC.
388 In that case, do nothing special here.
389 (Otherwise would recurse infinitely in convert. */
390 if (TYPE_PRECISION (typex) != inprec)
391 {
392 /* Don't do unsigned arithmetic where signed was wanted,
393 or vice versa.
394 Exception: if both of the original operands were
395 unsigned then we can safely do the work as unsigned.
396 Exception: shift operations take their type solely
397 from the first argument.
398 Exception: the LSHIFT_EXPR case above requires that
399 we perform this operation unsigned lest we produce
400 signed-overflow undefinedness.
401 And we may need to do it as unsigned
402 if we truncate to the original size. */
403 if (TYPE_UNSIGNED (TREE_TYPE (expr))
404 || (TYPE_UNSIGNED (TREE_TYPE (arg0))
405 && (TYPE_UNSIGNED (TREE_TYPE (arg1))
406 || ex_form == LSHIFT_EXPR
407 || ex_form == RSHIFT_EXPR
408 || ex_form == LROTATE_EXPR
409 || ex_form == RROTATE_EXPR))
410 || ex_form == LSHIFT_EXPR
411 /* If we have !flag_wrapv, and either ARG0 or
412 ARG1 is of a signed type, we have to do
413 PLUS_EXPR, MINUS_EXPR or MULT_EXPR in an unsigned
414 type in case the operation in outprec precision
415 could overflow. Otherwise, we would introduce
416 signed-overflow undefinedness. */
417 || ((!(INTEGRAL_TYPE_P (TREE_TYPE (arg0))
418 && TYPE_OVERFLOW_WRAPS (TREE_TYPE (arg0)))
419 || !(INTEGRAL_TYPE_P (TREE_TYPE (arg1))
420 && TYPE_OVERFLOW_WRAPS (TREE_TYPE (arg1))))
421 && ((TYPE_PRECISION (TREE_TYPE (arg0)) * 2u
422 > outprec)
423 || (TYPE_PRECISION (TREE_TYPE (arg1)) * 2u
424 > outprec))
425 && (ex_form == PLUS_EXPR
426 || ex_form == MINUS_EXPR
427 || ex_form == MULT_EXPR)))
428 {
429 if (!TYPE_UNSIGNED (typex))
430 typex = unsigned_type_for (typex);
431 }
432 else
433 {
434 if (TYPE_UNSIGNED (typex))
435 typex = signed_type_for (typex);
436 }
437 /* We should do away with all this once we have a proper
438 type promotion/demotion pass, see PR45397. */
439 expr = maybe_fold_build2_loc (dofold, loc, ex_form, typex,
440 convert (typex, arg0),
441 convert (typex, arg1));
442 return convert (type, expr);
443 }
444
445 return NULL_TREE;
446 }
447
448 /* Convert EXPR to some integer (or enum) type TYPE.
449
450 EXPR must be pointer, integer, discrete (enum, char, or bool), float,
451 fixed-point or vector; in other cases error is called.
452
453 If DOFOLD is TRUE, we try to simplify newly-created patterns by folding.
454
455 The result of this is always supposed to be a newly created tree node
456 not in use in any existing structure. */
457
458 static tree
459 convert_to_integer_1 (tree type, tree expr, bool dofold)
460 {
461 enum tree_code ex_form = TREE_CODE (expr);
462 tree intype = TREE_TYPE (expr);
463 unsigned int inprec = element_precision (intype);
464 unsigned int outprec = element_precision (type);
465 location_t loc = EXPR_LOCATION (expr);
466
467 /* An INTEGER_TYPE cannot be incomplete, but an ENUMERAL_TYPE can
468 be. Consider `enum E = { a, b = (enum E) 3 };'. */
469 if (!COMPLETE_TYPE_P (type))
470 {
471 error ("conversion to incomplete type");
472 return error_mark_node;
473 }
474
475 if (ex_form == COMPOUND_EXPR)
476 {
477 tree t = convert_to_integer_1 (type, TREE_OPERAND (expr, 1), dofold);
478 if (t == TREE_OPERAND (expr, 1))
479 return expr;
480 return build2_loc (EXPR_LOCATION (expr), COMPOUND_EXPR, TREE_TYPE (t),
481 TREE_OPERAND (expr, 0), t);
482 }
483
484 /* Convert e.g. (long)round(d) -> lround(d). */
485 /* If we're converting to char, we may encounter differing behavior
486 between converting from double->char vs double->long->char.
487 We're in "undefined" territory but we prefer to be conservative,
488 so only proceed in "unsafe" math mode. */
489 if (optimize
490 && (flag_unsafe_math_optimizations
491 || (long_integer_type_node
492 && outprec >= TYPE_PRECISION (long_integer_type_node))))
493 {
494 tree s_expr = strip_float_extensions (expr);
495 tree s_intype = TREE_TYPE (s_expr);
496 const enum built_in_function fcode = builtin_mathfn_code (s_expr);
497 tree fn = 0;
498
499 switch (fcode)
500 {
501 CASE_FLT_FN (BUILT_IN_CEIL):
502 CASE_FLT_FN_FLOATN_NX (BUILT_IN_CEIL):
503 /* Only convert in ISO C99 mode. */
504 if (!targetm.libc_has_function (function_c99_misc))
505 break;
506 if (outprec < TYPE_PRECISION (integer_type_node)
507 || (outprec == TYPE_PRECISION (integer_type_node)
508 && !TYPE_UNSIGNED (type)))
509 fn = mathfn_built_in (s_intype, BUILT_IN_ICEIL);
510 else if (outprec == TYPE_PRECISION (long_integer_type_node)
511 && !TYPE_UNSIGNED (type))
512 fn = mathfn_built_in (s_intype, BUILT_IN_LCEIL);
513 else if (outprec == TYPE_PRECISION (long_long_integer_type_node)
514 && !TYPE_UNSIGNED (type))
515 fn = mathfn_built_in (s_intype, BUILT_IN_LLCEIL);
516 break;
517
518 CASE_FLT_FN (BUILT_IN_FLOOR):
519 CASE_FLT_FN_FLOATN_NX (BUILT_IN_FLOOR):
520 /* Only convert in ISO C99 mode. */
521 if (!targetm.libc_has_function (function_c99_misc))
522 break;
523 if (outprec < TYPE_PRECISION (integer_type_node)
524 || (outprec == TYPE_PRECISION (integer_type_node)
525 && !TYPE_UNSIGNED (type)))
526 fn = mathfn_built_in (s_intype, BUILT_IN_IFLOOR);
527 else if (outprec == TYPE_PRECISION (long_integer_type_node)
528 && !TYPE_UNSIGNED (type))
529 fn = mathfn_built_in (s_intype, BUILT_IN_LFLOOR);
530 else if (outprec == TYPE_PRECISION (long_long_integer_type_node)
531 && !TYPE_UNSIGNED (type))
532 fn = mathfn_built_in (s_intype, BUILT_IN_LLFLOOR);
533 break;
534
535 CASE_FLT_FN (BUILT_IN_ROUND):
536 CASE_FLT_FN_FLOATN_NX (BUILT_IN_ROUND):
537 /* Only convert in ISO C99 mode and with -fno-math-errno. */
538 if (!targetm.libc_has_function (function_c99_misc)
539 || flag_errno_math)
540 break;
541 if (outprec < TYPE_PRECISION (integer_type_node)
542 || (outprec == TYPE_PRECISION (integer_type_node)
543 && !TYPE_UNSIGNED (type)))
544 fn = mathfn_built_in (s_intype, BUILT_IN_IROUND);
545 else if (outprec == TYPE_PRECISION (long_integer_type_node)
546 && !TYPE_UNSIGNED (type))
547 fn = mathfn_built_in (s_intype, BUILT_IN_LROUND);
548 else if (outprec == TYPE_PRECISION (long_long_integer_type_node)
549 && !TYPE_UNSIGNED (type))
550 fn = mathfn_built_in (s_intype, BUILT_IN_LLROUND);
551 break;
552
553 CASE_FLT_FN (BUILT_IN_NEARBYINT):
554 CASE_FLT_FN_FLOATN_NX (BUILT_IN_NEARBYINT):
555 /* Only convert nearbyint* if we can ignore math exceptions. */
556 if (flag_trapping_math)
557 break;
558 gcc_fallthrough ();
559 CASE_FLT_FN (BUILT_IN_RINT):
560 CASE_FLT_FN_FLOATN_NX (BUILT_IN_RINT):
561 /* Only convert in ISO C99 mode and with -fno-math-errno. */
562 if (!targetm.libc_has_function (function_c99_misc)
563 || flag_errno_math)
564 break;
565 if (outprec < TYPE_PRECISION (integer_type_node)
566 || (outprec == TYPE_PRECISION (integer_type_node)
567 && !TYPE_UNSIGNED (type)))
568 fn = mathfn_built_in (s_intype, BUILT_IN_IRINT);
569 else if (outprec == TYPE_PRECISION (long_integer_type_node)
570 && !TYPE_UNSIGNED (type))
571 fn = mathfn_built_in (s_intype, BUILT_IN_LRINT);
572 else if (outprec == TYPE_PRECISION (long_long_integer_type_node)
573 && !TYPE_UNSIGNED (type))
574 fn = mathfn_built_in (s_intype, BUILT_IN_LLRINT);
575 break;
576
577 CASE_FLT_FN (BUILT_IN_TRUNC):
578 CASE_FLT_FN_FLOATN_NX (BUILT_IN_TRUNC):
579 if (call_expr_nargs (s_expr) != 1
580 || !SCALAR_FLOAT_TYPE_P (TREE_TYPE (CALL_EXPR_ARG (s_expr, 0))))
581 break;
582 return convert_to_integer_1 (type, CALL_EXPR_ARG (s_expr, 0),
583 dofold);
584
585 default:
586 break;
587 }
588
589 if (fn
590 && call_expr_nargs (s_expr) == 1
591 && SCALAR_FLOAT_TYPE_P (TREE_TYPE (CALL_EXPR_ARG (s_expr, 0))))
592 {
593 tree newexpr = build_call_expr (fn, 1, CALL_EXPR_ARG (s_expr, 0));
594 return convert_to_integer_1 (type, newexpr, dofold);
595 }
596 }
597
598 /* Convert (int)logb(d) -> ilogb(d). */
599 if (optimize
600 && flag_unsafe_math_optimizations
601 && !flag_trapping_math && !flag_errno_math && flag_finite_math_only
602 && integer_type_node
603 && (outprec > TYPE_PRECISION (integer_type_node)
604 || (outprec == TYPE_PRECISION (integer_type_node)
605 && !TYPE_UNSIGNED (type))))
606 {
607 tree s_expr = strip_float_extensions (expr);
608 tree s_intype = TREE_TYPE (s_expr);
609 const enum built_in_function fcode = builtin_mathfn_code (s_expr);
610 tree fn = 0;
611
612 switch (fcode)
613 {
614 CASE_FLT_FN (BUILT_IN_LOGB):
615 fn = mathfn_built_in (s_intype, BUILT_IN_ILOGB);
616 break;
617
618 default:
619 break;
620 }
621
622 if (fn
623 && call_expr_nargs (s_expr) == 1
624 && SCALAR_FLOAT_TYPE_P (TREE_TYPE (CALL_EXPR_ARG (s_expr, 0))))
625 {
626 tree newexpr = build_call_expr (fn, 1, CALL_EXPR_ARG (s_expr, 0));
627 return convert_to_integer_1 (type, newexpr, dofold);
628 }
629 }
630
631 switch (TREE_CODE (intype))
632 {
633 case POINTER_TYPE:
634 case REFERENCE_TYPE:
635 if (integer_zerop (expr)
636 && !TREE_OVERFLOW (tree_strip_any_location_wrapper (expr)))
637 return build_int_cst (type, 0);
638
639 /* Convert to an unsigned integer of the correct width first, and from
640 there widen/truncate to the required type. Some targets support the
641 coexistence of multiple valid pointer sizes, so fetch the one we need
642 from the type. */
643 if (!dofold)
644 return build1 (CONVERT_EXPR, type, expr);
645 expr = fold_build1 (CONVERT_EXPR,
646 lang_hooks.types.type_for_size
647 (TYPE_PRECISION (intype), 0),
648 expr);
649 return fold_convert (type, expr);
650
651 case INTEGER_TYPE:
652 case ENUMERAL_TYPE:
653 case BOOLEAN_TYPE:
654 case OFFSET_TYPE:
655 /* If this is a logical operation, which just returns 0 or 1, we can
656 change the type of the expression. */
657
658 if (TREE_CODE_CLASS (ex_form) == tcc_comparison)
659 {
660 expr = copy_node (expr);
661 TREE_TYPE (expr) = type;
662 return expr;
663 }
664
665 /* If we are widening the type, put in an explicit conversion.
666 Similarly if we are not changing the width. After this, we know
667 we are truncating EXPR. */
668
669 else if (outprec >= inprec)
670 {
671 enum tree_code code;
672
673 /* If the precision of the EXPR's type is K bits and the
674 destination mode has more bits, and the sign is changing,
675 it is not safe to use a NOP_EXPR. For example, suppose
676 that EXPR's type is a 3-bit unsigned integer type, the
677 TYPE is a 3-bit signed integer type, and the machine mode
678 for the types is 8-bit QImode. In that case, the
679 conversion necessitates an explicit sign-extension. In
680 the signed-to-unsigned case the high-order bits have to
681 be cleared. */
682 if (TYPE_UNSIGNED (type) != TYPE_UNSIGNED (TREE_TYPE (expr))
683 && !type_has_mode_precision_p (TREE_TYPE (expr)))
684 code = CONVERT_EXPR;
685 else
686 code = NOP_EXPR;
687
688 return maybe_fold_build1_loc (dofold, loc, code, type, expr);
689 }
690
691 /* If TYPE is an enumeral type or a type with a precision less
692 than the number of bits in its mode, do the conversion to the
693 type corresponding to its mode, then do a nop conversion
694 to TYPE. */
695 else if (TREE_CODE (type) == ENUMERAL_TYPE
696 || maybe_ne (outprec, GET_MODE_PRECISION (TYPE_MODE (type))))
697 {
698 expr
699 = convert_to_integer_1 (lang_hooks.types.type_for_mode
700 (TYPE_MODE (type), TYPE_UNSIGNED (type)),
701 expr, dofold);
702 return maybe_fold_build1_loc (dofold, loc, NOP_EXPR, type, expr);
703 }
704
705 /* Here detect when we can distribute the truncation down past some
706 arithmetic. For example, if adding two longs and converting to an
707 int, we can equally well convert both to ints and then add.
708 For the operations handled here, such truncation distribution
709 is always safe.
710 It is desirable in these cases:
711 1) when truncating down to full-word from a larger size
712 2) when truncating takes no work.
713 3) when at least one operand of the arithmetic has been extended
714 (as by C's default conversions). In this case we need two conversions
715 if we do the arithmetic as already requested, so we might as well
716 truncate both and then combine. Perhaps that way we need only one.
717
718 Note that in general we cannot do the arithmetic in a type
719 shorter than the desired result of conversion, even if the operands
720 are both extended from a shorter type, because they might overflow
721 if combined in that type. The exceptions to this--the times when
722 two narrow values can be combined in their narrow type even to
723 make a wider result--are handled by "shorten" in build_binary_op. */
724
725 if (dofold)
726 switch (ex_form)
727 {
728 case RSHIFT_EXPR:
729 /* We can pass truncation down through right shifting
730 when the shift count is a nonpositive constant. */
731 if (TREE_CODE (TREE_OPERAND (expr, 1)) == INTEGER_CST
732 && tree_int_cst_sgn (TREE_OPERAND (expr, 1)) <= 0)
733 goto trunc1;
734 break;
735
736 case LSHIFT_EXPR:
737 /* We can pass truncation down through left shifting
738 when the shift count is a nonnegative constant and
739 the target type is unsigned. */
740 if (TREE_CODE (TREE_OPERAND (expr, 1)) == INTEGER_CST
741 && tree_int_cst_sgn (TREE_OPERAND (expr, 1)) >= 0
742 && TYPE_UNSIGNED (type)
743 && TREE_CODE (TYPE_SIZE (type)) == INTEGER_CST)
744 {
745 /* If shift count is less than the width of the truncated type,
746 really shift. */
747 if (tree_int_cst_lt (TREE_OPERAND (expr, 1), TYPE_SIZE (type)))
748 /* In this case, shifting is like multiplication. */
749 goto trunc1;
750 else
751 {
752 /* If it is >= that width, result is zero.
753 Handling this with trunc1 would give the wrong result:
754 (int) ((long long) a << 32) is well defined (as 0)
755 but (int) a << 32 is undefined and would get a
756 warning. */
757
758 tree t = build_int_cst (type, 0);
759
760 /* If the original expression had side-effects, we must
761 preserve it. */
762 if (TREE_SIDE_EFFECTS (expr))
763 return build2 (COMPOUND_EXPR, type, expr, t);
764 else
765 return t;
766 }
767 }
768 break;
769
770 case TRUNC_DIV_EXPR:
771 {
772 tree arg0 = get_unwidened (TREE_OPERAND (expr, 0), NULL_TREE);
773 tree arg1 = get_unwidened (TREE_OPERAND (expr, 1), NULL_TREE);
774
775 /* Don't distribute unless the output precision is at least as
776 big as the actual inputs and it has the same signedness. */
777 if (outprec >= TYPE_PRECISION (TREE_TYPE (arg0))
778 && outprec >= TYPE_PRECISION (TREE_TYPE (arg1))
779 /* If signedness of arg0 and arg1 don't match,
780 we can't necessarily find a type to compare them in. */
781 && (TYPE_UNSIGNED (TREE_TYPE (arg0))
782 == TYPE_UNSIGNED (TREE_TYPE (arg1)))
783 /* Do not change the sign of the division. */
784 && (TYPE_UNSIGNED (TREE_TYPE (expr))
785 == TYPE_UNSIGNED (TREE_TYPE (arg0)))
786 /* Either require unsigned division or a division by
787 a constant that is not -1. */
788 && (TYPE_UNSIGNED (TREE_TYPE (arg0))
789 || (TREE_CODE (arg1) == INTEGER_CST
790 && !integer_all_onesp (arg1))))
791 {
792 tree tem = do_narrow (loc, ex_form, type, arg0, arg1,
793 expr, inprec, outprec, dofold);
794 if (tem)
795 return tem;
796 }
797 break;
798 }
799
800 case MAX_EXPR:
801 case MIN_EXPR:
802 case MULT_EXPR:
803 {
804 tree arg0 = get_unwidened (TREE_OPERAND (expr, 0), type);
805 tree arg1 = get_unwidened (TREE_OPERAND (expr, 1), type);
806
807 /* Don't distribute unless the output precision is at least as
808 big as the actual inputs. Otherwise, the comparison of the
809 truncated values will be wrong. */
810 if (outprec >= TYPE_PRECISION (TREE_TYPE (arg0))
811 && outprec >= TYPE_PRECISION (TREE_TYPE (arg1))
812 /* If signedness of arg0 and arg1 don't match,
813 we can't necessarily find a type to compare them in. */
814 && (TYPE_UNSIGNED (TREE_TYPE (arg0))
815 == TYPE_UNSIGNED (TREE_TYPE (arg1))))
816 goto trunc1;
817 break;
818 }
819
820 case PLUS_EXPR:
821 case MINUS_EXPR:
822 case BIT_AND_EXPR:
823 case BIT_IOR_EXPR:
824 case BIT_XOR_EXPR:
825 trunc1:
826 {
827 tree arg0 = get_unwidened (TREE_OPERAND (expr, 0), type);
828 tree arg1 = get_unwidened (TREE_OPERAND (expr, 1), type);
829
830 /* Do not try to narrow operands of pointer subtraction;
831 that will interfere with other folding. */
832 if (ex_form == MINUS_EXPR
833 && CONVERT_EXPR_P (arg0)
834 && CONVERT_EXPR_P (arg1)
835 && POINTER_TYPE_P (TREE_TYPE (TREE_OPERAND (arg0, 0)))
836 && POINTER_TYPE_P (TREE_TYPE (TREE_OPERAND (arg1, 0))))
837 break;
838
839 tree tem = do_narrow (loc, ex_form, type, arg0, arg1,
840 expr, inprec, outprec, dofold);
841 if (tem)
842 return tem;
843 }
844 break;
845
846 case NEGATE_EXPR:
847 /* Using unsigned arithmetic for signed types may hide overflow
848 bugs. */
849 if (!TYPE_UNSIGNED (TREE_TYPE (TREE_OPERAND (expr, 0)))
850 && sanitize_flags_p (SANITIZE_SI_OVERFLOW))
851 break;
852 /* Fall through. */
853 case BIT_NOT_EXPR:
854 /* This is not correct for ABS_EXPR,
855 since we must test the sign before truncation. */
856 {
857 /* Do the arithmetic in type TYPEX,
858 then convert result to TYPE. */
859 tree typex = type;
860
861 /* Can't do arithmetic in enumeral types
862 so use an integer type that will hold the values. */
863 if (TREE_CODE (typex) == ENUMERAL_TYPE)
864 typex
865 = lang_hooks.types.type_for_size (TYPE_PRECISION (typex),
866 TYPE_UNSIGNED (typex));
867
868 if (!TYPE_UNSIGNED (typex))
869 typex = unsigned_type_for (typex);
870 return convert (type,
871 fold_build1 (ex_form, typex,
872 convert (typex,
873 TREE_OPERAND (expr, 0))));
874 }
875
876 CASE_CONVERT:
877 {
878 tree argtype = TREE_TYPE (TREE_OPERAND (expr, 0));
879 /* Don't introduce a "can't convert between vector values
880 of different size" error. */
881 if (TREE_CODE (argtype) == VECTOR_TYPE
882 && maybe_ne (GET_MODE_SIZE (TYPE_MODE (argtype)),
883 GET_MODE_SIZE (TYPE_MODE (type))))
884 break;
885 }
886 /* If truncating after truncating, might as well do all at once.
887 If truncating after extending, we may get rid of wasted work. */
888 return convert (type, get_unwidened (TREE_OPERAND (expr, 0), type));
889
890 case COND_EXPR:
891 /* It is sometimes worthwhile to push the narrowing down through
892 the conditional and never loses. A COND_EXPR may have a throw
893 as one operand, which then has void type. Just leave void
894 operands as they are. */
895 return
896 fold_build3 (COND_EXPR, type, TREE_OPERAND (expr, 0),
897 VOID_TYPE_P (TREE_TYPE (TREE_OPERAND (expr, 1)))
898 ? TREE_OPERAND (expr, 1)
899 : convert (type, TREE_OPERAND (expr, 1)),
900 VOID_TYPE_P (TREE_TYPE (TREE_OPERAND (expr, 2)))
901 ? TREE_OPERAND (expr, 2)
902 : convert (type, TREE_OPERAND (expr, 2)));
903
904 default:
905 break;
906 }
907
908 /* When parsing long initializers, we might end up with a lot of casts.
909 Shortcut this. */
910 if (TREE_CODE (tree_strip_any_location_wrapper (expr)) == INTEGER_CST)
911 return fold_convert (type, expr);
912 return build1 (CONVERT_EXPR, type, expr);
913
914 case REAL_TYPE:
915 if (sanitize_flags_p (SANITIZE_FLOAT_CAST)
916 && current_function_decl != NULL_TREE)
917 {
918 expr = save_expr (expr);
919 tree check = ubsan_instrument_float_cast (loc, type, expr);
920 expr = build1 (FIX_TRUNC_EXPR, type, expr);
921 if (check == NULL_TREE)
922 return expr;
923 return maybe_fold_build2_loc (dofold, loc, COMPOUND_EXPR,
924 TREE_TYPE (expr), check, expr);
925 }
926 else
927 return build1 (FIX_TRUNC_EXPR, type, expr);
928
929 case FIXED_POINT_TYPE:
930 return build1 (FIXED_CONVERT_EXPR, type, expr);
931
932 case COMPLEX_TYPE:
933 expr = maybe_fold_build1_loc (dofold, loc, REALPART_EXPR,
934 TREE_TYPE (TREE_TYPE (expr)), expr);
935 return convert (type, expr);
936
937 case VECTOR_TYPE:
938 if (!tree_int_cst_equal (TYPE_SIZE (type), TYPE_SIZE (TREE_TYPE (expr))))
939 {
940 error ("cannot convert a vector of type %qT"
941 " to type %qT which has different size",
942 TREE_TYPE (expr), type);
943 return error_mark_node;
944 }
945 return build1 (VIEW_CONVERT_EXPR, type, expr);
946
947 default:
948 error ("aggregate value used where an integer was expected");
949 return convert (type, integer_zero_node);
950 }
951 }
952
953 /* Convert EXPR to some integer (or enum) type TYPE.
954
955 EXPR must be pointer, integer, discrete (enum, char, or bool), float,
956 fixed-point or vector; in other cases error is called.
957
958 The result of this is always supposed to be a newly created tree node
959 not in use in any existing structure. */
960
961 tree
962 convert_to_integer (tree type, tree expr)
963 {
964 return convert_to_integer_1 (type, expr, true);
965 }
966
967 /* A wrapper around convert_to_complex_1 that only folds the
968 expression if DOFOLD, or if it is CONSTANT_CLASS_OR_WRAPPER_P. */
969
970 tree
971 convert_to_integer_maybe_fold (tree type, tree expr, bool dofold)
972 {
973 tree result
974 = convert_to_integer_1 (type, expr,
975 dofold || CONSTANT_CLASS_OR_WRAPPER_P (expr));
976 return preserve_any_location_wrapper (result, expr);
977 }
978
979 /* Convert EXPR to the complex type TYPE in the usual ways. If FOLD_P is
980 true, try to fold the expression. */
981
982 static tree
983 convert_to_complex_1 (tree type, tree expr, bool fold_p)
984 {
985 location_t loc = EXPR_LOCATION (expr);
986 tree subtype = TREE_TYPE (type);
987
988 switch (TREE_CODE (TREE_TYPE (expr)))
989 {
990 case REAL_TYPE:
991 case FIXED_POINT_TYPE:
992 case INTEGER_TYPE:
993 case ENUMERAL_TYPE:
994 case BOOLEAN_TYPE:
995 return build2 (COMPLEX_EXPR, type, convert (subtype, expr),
996 convert (subtype, integer_zero_node));
997
998 case COMPLEX_TYPE:
999 {
1000 tree elt_type = TREE_TYPE (TREE_TYPE (expr));
1001
1002 if (TYPE_MAIN_VARIANT (elt_type) == TYPE_MAIN_VARIANT (subtype))
1003 return expr;
1004 else if (TREE_CODE (expr) == COMPOUND_EXPR)
1005 {
1006 tree t = convert_to_complex_1 (type, TREE_OPERAND (expr, 1),
1007 fold_p);
1008 if (t == TREE_OPERAND (expr, 1))
1009 return expr;
1010 return build2_loc (EXPR_LOCATION (expr), COMPOUND_EXPR,
1011 TREE_TYPE (t), TREE_OPERAND (expr, 0), t);
1012 }
1013 else if (TREE_CODE (expr) == COMPLEX_EXPR)
1014 return maybe_fold_build2_loc (fold_p, loc, COMPLEX_EXPR, type,
1015 convert (subtype,
1016 TREE_OPERAND (expr, 0)),
1017 convert (subtype,
1018 TREE_OPERAND (expr, 1)));
1019 else
1020 {
1021 expr = save_expr (expr);
1022 tree realp = maybe_fold_build1_loc (fold_p, loc, REALPART_EXPR,
1023 TREE_TYPE (TREE_TYPE (expr)),
1024 expr);
1025 tree imagp = maybe_fold_build1_loc (fold_p, loc, IMAGPART_EXPR,
1026 TREE_TYPE (TREE_TYPE (expr)),
1027 expr);
1028 return maybe_fold_build2_loc (fold_p, loc, COMPLEX_EXPR, type,
1029 convert (subtype, realp),
1030 convert (subtype, imagp));
1031 }
1032 }
1033
1034 case POINTER_TYPE:
1035 case REFERENCE_TYPE:
1036 error ("pointer value used where a complex was expected");
1037 return convert_to_complex_1 (type, integer_zero_node, fold_p);
1038
1039 default:
1040 error ("aggregate value used where a complex was expected");
1041 return convert_to_complex_1 (type, integer_zero_node, fold_p);
1042 }
1043 }
1044
1045 /* A wrapper around convert_to_complex_1 that always folds the
1046 expression. */
1047
1048 tree
1049 convert_to_complex (tree type, tree expr)
1050 {
1051 return convert_to_complex_1 (type, expr, true);
1052 }
1053
1054 /* A wrapper around convert_to_complex_1 that only folds the
1055 expression if DOFOLD, or if it is CONSTANT_CLASS_OR_WRAPPER_P. */
1056
1057 tree
1058 convert_to_complex_maybe_fold (tree type, tree expr, bool dofold)
1059 {
1060 tree result
1061 = convert_to_complex_1 (type, expr,
1062 dofold || CONSTANT_CLASS_OR_WRAPPER_P (expr));
1063 return preserve_any_location_wrapper (result, expr);
1064 }
1065
1066 /* Convert EXPR to the vector type TYPE in the usual ways. */
1067
1068 tree
1069 convert_to_vector (tree type, tree expr)
1070 {
1071 switch (TREE_CODE (TREE_TYPE (expr)))
1072 {
1073 case INTEGER_TYPE:
1074 case VECTOR_TYPE:
1075 if (!tree_int_cst_equal (TYPE_SIZE (type), TYPE_SIZE (TREE_TYPE (expr))))
1076 {
1077 error ("cannot convert a value of type %qT"
1078 " to vector type %qT which has different size",
1079 TREE_TYPE (expr), type);
1080 return error_mark_node;
1081 }
1082 return build1 (VIEW_CONVERT_EXPR, type, expr);
1083
1084 default:
1085 error ("cannot convert value to a vector");
1086 return error_mark_node;
1087 }
1088 }
1089
1090 /* Convert EXPR to some fixed-point type TYPE.
1091
1092 EXPR must be fixed-point, float, integer, or enumeral;
1093 in other cases error is called. */
1094
1095 tree
1096 convert_to_fixed (tree type, tree expr)
1097 {
1098 if (integer_zerop (expr))
1099 {
1100 tree fixed_zero_node = build_fixed (type, FCONST0 (TYPE_MODE (type)));
1101 return fixed_zero_node;
1102 }
1103 else if (integer_onep (expr) && ALL_SCALAR_ACCUM_MODE_P (TYPE_MODE (type)))
1104 {
1105 tree fixed_one_node = build_fixed (type, FCONST1 (TYPE_MODE (type)));
1106 return fixed_one_node;
1107 }
1108
1109 switch (TREE_CODE (TREE_TYPE (expr)))
1110 {
1111 case FIXED_POINT_TYPE:
1112 case INTEGER_TYPE:
1113 case ENUMERAL_TYPE:
1114 case BOOLEAN_TYPE:
1115 case REAL_TYPE:
1116 return build1 (FIXED_CONVERT_EXPR, type, expr);
1117
1118 case COMPLEX_TYPE:
1119 return convert (type,
1120 fold_build1 (REALPART_EXPR,
1121 TREE_TYPE (TREE_TYPE (expr)), expr));
1122
1123 default:
1124 error ("aggregate value used where a fixed-point was expected");
1125 return error_mark_node;
1126 }
1127 }
1128
1129 #if CHECKING_P
1130
1131 namespace selftest {
1132
1133 /* Selftests for conversions. */
1134
1135 static void
1136 test_convert_to_integer_maybe_fold (tree orig_type, tree new_type)
1137 {
1138 /* Calling convert_to_integer_maybe_fold on an INTEGER_CST. */
1139
1140 tree orig_cst = build_int_cst (orig_type, 42);
1141
1142 /* Verify that convert_to_integer_maybe_fold on a constant returns a new
1143 constant of the new type, unless the types are the same, in which
1144 case verify it's a no-op. */
1145 {
1146 tree result = convert_to_integer_maybe_fold (new_type,
1147 orig_cst, false);
1148 if (orig_type != new_type)
1149 {
1150 ASSERT_EQ (TREE_TYPE (result), new_type);
1151 ASSERT_EQ (TREE_CODE (result), INTEGER_CST);
1152 }
1153 else
1154 ASSERT_EQ (result, orig_cst);
1155 }
1156
1157 /* Calling convert_to_integer_maybe_fold on a location wrapper around
1158 an INTEGER_CST.
1159
1160 Verify that convert_to_integer_maybe_fold on a location wrapper
1161 around a constant returns a new location wrapper around an equivalent
1162 constant, both of the new type, unless the types are the same,
1163 in which case the original wrapper should be returned. */
1164 {
1165 const location_t loc = BUILTINS_LOCATION;
1166 tree wrapped_orig_cst = maybe_wrap_with_location (orig_cst, loc);
1167 tree result
1168 = convert_to_integer_maybe_fold (new_type, wrapped_orig_cst, false);
1169 ASSERT_EQ (TREE_TYPE (result), new_type);
1170 ASSERT_EQ (EXPR_LOCATION (result), loc);
1171 ASSERT_TRUE (location_wrapper_p (result));
1172 ASSERT_EQ (TREE_TYPE (TREE_OPERAND (result, 0)), new_type);
1173 ASSERT_EQ (TREE_CODE (TREE_OPERAND (result, 0)), INTEGER_CST);
1174
1175 if (orig_type == new_type)
1176 ASSERT_EQ (result, wrapped_orig_cst);
1177 }
1178 }
1179
1180 /* Verify that convert_to_integer_maybe_fold preserves locations. */
1181
1182 static void
1183 test_convert_to_integer_maybe_fold ()
1184 {
1185 /* char -> long. */
1186 test_convert_to_integer_maybe_fold (char_type_node, long_integer_type_node);
1187
1188 /* char -> char. */
1189 test_convert_to_integer_maybe_fold (char_type_node, char_type_node);
1190
1191 /* long -> char. */
1192 test_convert_to_integer_maybe_fold (char_type_node, long_integer_type_node);
1193
1194 /* long -> long. */
1195 test_convert_to_integer_maybe_fold (long_integer_type_node,
1196 long_integer_type_node);
1197 }
1198
1199 /* Run all of the selftests within this file. */
1200
1201 void
1202 convert_c_tests ()
1203 {
1204 test_convert_to_integer_maybe_fold ();
1205 }
1206
1207 } // namespace selftest
1208
1209 #endif /* CHECKING_P */