1 |
684 |
jeremybenn |
/* Expand builtin functions.
|
2 |
|
|
Copyright (C) 1988, 1992, 1993, 1994, 1995, 1996, 1997, 1998, 1999,
|
3 |
|
|
2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011,
|
4 |
|
|
2012 Free Software Foundation, Inc.
|
5 |
|
|
|
6 |
|
|
This file is part of GCC.
|
7 |
|
|
|
8 |
|
|
GCC is free software; you can redistribute it and/or modify it under
|
9 |
|
|
the terms of the GNU General Public License as published by the Free
|
10 |
|
|
Software Foundation; either version 3, or (at your option) any later
|
11 |
|
|
version.
|
12 |
|
|
|
13 |
|
|
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
|
14 |
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
15 |
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
16 |
|
|
for more details.
|
17 |
|
|
|
18 |
|
|
You should have received a copy of the GNU General Public License
|
19 |
|
|
along with GCC; see the file COPYING3. If not see
|
20 |
|
|
<http://www.gnu.org/licenses/>. */
|
21 |
|
|
|
22 |
|
|
#include "config.h"
|
23 |
|
|
#include "system.h"
|
24 |
|
|
#include "coretypes.h"
|
25 |
|
|
#include "tm.h"
|
26 |
|
|
#include "machmode.h"
|
27 |
|
|
#include "rtl.h"
|
28 |
|
|
#include "tree.h"
|
29 |
|
|
#include "realmpfr.h"
|
30 |
|
|
#include "gimple.h"
|
31 |
|
|
#include "flags.h"
|
32 |
|
|
#include "regs.h"
|
33 |
|
|
#include "hard-reg-set.h"
|
34 |
|
|
#include "except.h"
|
35 |
|
|
#include "function.h"
|
36 |
|
|
#include "insn-config.h"
|
37 |
|
|
#include "expr.h"
|
38 |
|
|
#include "optabs.h"
|
39 |
|
|
#include "libfuncs.h"
|
40 |
|
|
#include "recog.h"
|
41 |
|
|
#include "output.h"
|
42 |
|
|
#include "typeclass.h"
|
43 |
|
|
#include "predict.h"
|
44 |
|
|
#include "tm_p.h"
|
45 |
|
|
#include "target.h"
|
46 |
|
|
#include "langhooks.h"
|
47 |
|
|
#include "basic-block.h"
|
48 |
|
|
#include "tree-mudflap.h"
|
49 |
|
|
#include "tree-flow.h"
|
50 |
|
|
#include "value-prof.h"
|
51 |
|
|
#include "diagnostic-core.h"
|
52 |
|
|
#include "builtins.h"
|
53 |
|
|
|
54 |
|
|
|
55 |
|
|
#ifndef PAD_VARARGS_DOWN
|
56 |
|
|
#define PAD_VARARGS_DOWN BYTES_BIG_ENDIAN
|
57 |
|
|
#endif
|
58 |
|
|
static tree do_mpc_arg1 (tree, tree, int (*)(mpc_ptr, mpc_srcptr, mpc_rnd_t));
|
59 |
|
|
|
60 |
|
|
struct target_builtins default_target_builtins;
|
61 |
|
|
#if SWITCHABLE_TARGET
|
62 |
|
|
struct target_builtins *this_target_builtins = &default_target_builtins;
|
63 |
|
|
#endif
|
64 |
|
|
|
65 |
|
|
/* Define the names of the builtin function types and codes. */
|
66 |
|
|
const char *const built_in_class_names[4]
|
67 |
|
|
= {"NOT_BUILT_IN", "BUILT_IN_FRONTEND", "BUILT_IN_MD", "BUILT_IN_NORMAL"};
|
68 |
|
|
|
69 |
|
|
#define DEF_BUILTIN(X, N, C, T, LT, B, F, NA, AT, IM, COND) #X,
|
70 |
|
|
const char * built_in_names[(int) END_BUILTINS] =
|
71 |
|
|
{
|
72 |
|
|
#include "builtins.def"
|
73 |
|
|
};
|
74 |
|
|
#undef DEF_BUILTIN
|
75 |
|
|
|
76 |
|
|
/* Setup an array of _DECL trees, make sure each element is
|
77 |
|
|
initialized to NULL_TREE. */
|
78 |
|
|
builtin_info_type builtin_info;
|
79 |
|
|
|
80 |
|
|
static const char *c_getstr (tree);
|
81 |
|
|
static rtx c_readstr (const char *, enum machine_mode);
|
82 |
|
|
static int target_char_cast (tree, char *);
|
83 |
|
|
static rtx get_memory_rtx (tree, tree);
|
84 |
|
|
static int apply_args_size (void);
|
85 |
|
|
static int apply_result_size (void);
|
86 |
|
|
#if defined (HAVE_untyped_call) || defined (HAVE_untyped_return)
|
87 |
|
|
static rtx result_vector (int, rtx);
|
88 |
|
|
#endif
|
89 |
|
|
static void expand_builtin_update_setjmp_buf (rtx);
|
90 |
|
|
static void expand_builtin_prefetch (tree);
|
91 |
|
|
static rtx expand_builtin_apply_args (void);
|
92 |
|
|
static rtx expand_builtin_apply_args_1 (void);
|
93 |
|
|
static rtx expand_builtin_apply (rtx, rtx, rtx);
|
94 |
|
|
static void expand_builtin_return (rtx);
|
95 |
|
|
static enum type_class type_to_class (tree);
|
96 |
|
|
static rtx expand_builtin_classify_type (tree);
|
97 |
|
|
static void expand_errno_check (tree, rtx);
|
98 |
|
|
static rtx expand_builtin_mathfn (tree, rtx, rtx);
|
99 |
|
|
static rtx expand_builtin_mathfn_2 (tree, rtx, rtx);
|
100 |
|
|
static rtx expand_builtin_mathfn_3 (tree, rtx, rtx);
|
101 |
|
|
static rtx expand_builtin_mathfn_ternary (tree, rtx, rtx);
|
102 |
|
|
static rtx expand_builtin_interclass_mathfn (tree, rtx);
|
103 |
|
|
static rtx expand_builtin_sincos (tree);
|
104 |
|
|
static rtx expand_builtin_cexpi (tree, rtx);
|
105 |
|
|
static rtx expand_builtin_int_roundingfn (tree, rtx);
|
106 |
|
|
static rtx expand_builtin_int_roundingfn_2 (tree, rtx);
|
107 |
|
|
static rtx expand_builtin_next_arg (void);
|
108 |
|
|
static rtx expand_builtin_va_start (tree);
|
109 |
|
|
static rtx expand_builtin_va_end (tree);
|
110 |
|
|
static rtx expand_builtin_va_copy (tree);
|
111 |
|
|
static rtx expand_builtin_memcmp (tree, rtx, enum machine_mode);
|
112 |
|
|
static rtx expand_builtin_strcmp (tree, rtx);
|
113 |
|
|
static rtx expand_builtin_strncmp (tree, rtx, enum machine_mode);
|
114 |
|
|
static rtx builtin_memcpy_read_str (void *, HOST_WIDE_INT, enum machine_mode);
|
115 |
|
|
static rtx expand_builtin_memcpy (tree, rtx);
|
116 |
|
|
static rtx expand_builtin_mempcpy (tree, rtx, enum machine_mode);
|
117 |
|
|
static rtx expand_builtin_mempcpy_args (tree, tree, tree, rtx,
|
118 |
|
|
enum machine_mode, int);
|
119 |
|
|
static rtx expand_builtin_strcpy (tree, rtx);
|
120 |
|
|
static rtx expand_builtin_strcpy_args (tree, tree, rtx);
|
121 |
|
|
static rtx expand_builtin_stpcpy (tree, rtx, enum machine_mode);
|
122 |
|
|
static rtx expand_builtin_strncpy (tree, rtx);
|
123 |
|
|
static rtx builtin_memset_gen_str (void *, HOST_WIDE_INT, enum machine_mode);
|
124 |
|
|
static rtx expand_builtin_memset (tree, rtx, enum machine_mode);
|
125 |
|
|
static rtx expand_builtin_memset_args (tree, tree, tree, rtx, enum machine_mode, tree);
|
126 |
|
|
static rtx expand_builtin_bzero (tree);
|
127 |
|
|
static rtx expand_builtin_strlen (tree, rtx, enum machine_mode);
|
128 |
|
|
static rtx expand_builtin_alloca (tree, bool);
|
129 |
|
|
static rtx expand_builtin_unop (enum machine_mode, tree, rtx, rtx, optab);
|
130 |
|
|
static rtx expand_builtin_frame_address (tree, tree);
|
131 |
|
|
static tree stabilize_va_list_loc (location_t, tree, int);
|
132 |
|
|
static rtx expand_builtin_expect (tree, rtx);
|
133 |
|
|
static tree fold_builtin_constant_p (tree);
|
134 |
|
|
static tree fold_builtin_expect (location_t, tree, tree);
|
135 |
|
|
static tree fold_builtin_classify_type (tree);
|
136 |
|
|
static tree fold_builtin_strlen (location_t, tree, tree);
|
137 |
|
|
static tree fold_builtin_inf (location_t, tree, int);
|
138 |
|
|
static tree fold_builtin_nan (tree, tree, int);
|
139 |
|
|
static tree rewrite_call_expr (location_t, tree, int, tree, int, ...);
|
140 |
|
|
static bool validate_arg (const_tree, enum tree_code code);
|
141 |
|
|
static bool integer_valued_real_p (tree);
|
142 |
|
|
static tree fold_trunc_transparent_mathfn (location_t, tree, tree);
|
143 |
|
|
static bool readonly_data_expr (tree);
|
144 |
|
|
static rtx expand_builtin_fabs (tree, rtx, rtx);
|
145 |
|
|
static rtx expand_builtin_signbit (tree, rtx);
|
146 |
|
|
static tree fold_builtin_sqrt (location_t, tree, tree);
|
147 |
|
|
static tree fold_builtin_cbrt (location_t, tree, tree);
|
148 |
|
|
static tree fold_builtin_pow (location_t, tree, tree, tree, tree);
|
149 |
|
|
static tree fold_builtin_powi (location_t, tree, tree, tree, tree);
|
150 |
|
|
static tree fold_builtin_cos (location_t, tree, tree, tree);
|
151 |
|
|
static tree fold_builtin_cosh (location_t, tree, tree, tree);
|
152 |
|
|
static tree fold_builtin_tan (tree, tree);
|
153 |
|
|
static tree fold_builtin_trunc (location_t, tree, tree);
|
154 |
|
|
static tree fold_builtin_floor (location_t, tree, tree);
|
155 |
|
|
static tree fold_builtin_ceil (location_t, tree, tree);
|
156 |
|
|
static tree fold_builtin_round (location_t, tree, tree);
|
157 |
|
|
static tree fold_builtin_int_roundingfn (location_t, tree, tree);
|
158 |
|
|
static tree fold_builtin_bitop (tree, tree);
|
159 |
|
|
static tree fold_builtin_memory_op (location_t, tree, tree, tree, tree, bool, int);
|
160 |
|
|
static tree fold_builtin_strchr (location_t, tree, tree, tree);
|
161 |
|
|
static tree fold_builtin_memchr (location_t, tree, tree, tree, tree);
|
162 |
|
|
static tree fold_builtin_memcmp (location_t, tree, tree, tree);
|
163 |
|
|
static tree fold_builtin_strcmp (location_t, tree, tree);
|
164 |
|
|
static tree fold_builtin_strncmp (location_t, tree, tree, tree);
|
165 |
|
|
static tree fold_builtin_signbit (location_t, tree, tree);
|
166 |
|
|
static tree fold_builtin_copysign (location_t, tree, tree, tree, tree);
|
167 |
|
|
static tree fold_builtin_isascii (location_t, tree);
|
168 |
|
|
static tree fold_builtin_toascii (location_t, tree);
|
169 |
|
|
static tree fold_builtin_isdigit (location_t, tree);
|
170 |
|
|
static tree fold_builtin_fabs (location_t, tree, tree);
|
171 |
|
|
static tree fold_builtin_abs (location_t, tree, tree);
|
172 |
|
|
static tree fold_builtin_unordered_cmp (location_t, tree, tree, tree, enum tree_code,
|
173 |
|
|
enum tree_code);
|
174 |
|
|
static tree fold_builtin_n (location_t, tree, tree *, int, bool);
|
175 |
|
|
static tree fold_builtin_0 (location_t, tree, bool);
|
176 |
|
|
static tree fold_builtin_1 (location_t, tree, tree, bool);
|
177 |
|
|
static tree fold_builtin_2 (location_t, tree, tree, tree, bool);
|
178 |
|
|
static tree fold_builtin_3 (location_t, tree, tree, tree, tree, bool);
|
179 |
|
|
static tree fold_builtin_4 (location_t, tree, tree, tree, tree, tree, bool);
|
180 |
|
|
static tree fold_builtin_varargs (location_t, tree, tree, bool);
|
181 |
|
|
|
182 |
|
|
static tree fold_builtin_strpbrk (location_t, tree, tree, tree);
|
183 |
|
|
static tree fold_builtin_strstr (location_t, tree, tree, tree);
|
184 |
|
|
static tree fold_builtin_strrchr (location_t, tree, tree, tree);
|
185 |
|
|
static tree fold_builtin_strcat (location_t, tree, tree);
|
186 |
|
|
static tree fold_builtin_strncat (location_t, tree, tree, tree);
|
187 |
|
|
static tree fold_builtin_strspn (location_t, tree, tree);
|
188 |
|
|
static tree fold_builtin_strcspn (location_t, tree, tree);
|
189 |
|
|
static tree fold_builtin_sprintf (location_t, tree, tree, tree, int);
|
190 |
|
|
static tree fold_builtin_snprintf (location_t, tree, tree, tree, tree, int);
|
191 |
|
|
|
192 |
|
|
static rtx expand_builtin_object_size (tree);
|
193 |
|
|
static rtx expand_builtin_memory_chk (tree, rtx, enum machine_mode,
|
194 |
|
|
enum built_in_function);
|
195 |
|
|
static void maybe_emit_chk_warning (tree, enum built_in_function);
|
196 |
|
|
static void maybe_emit_sprintf_chk_warning (tree, enum built_in_function);
|
197 |
|
|
static void maybe_emit_free_warning (tree);
|
198 |
|
|
static tree fold_builtin_object_size (tree, tree);
|
199 |
|
|
static tree fold_builtin_strcat_chk (location_t, tree, tree, tree, tree);
|
200 |
|
|
static tree fold_builtin_strncat_chk (location_t, tree, tree, tree, tree, tree);
|
201 |
|
|
static tree fold_builtin_sprintf_chk (location_t, tree, enum built_in_function);
|
202 |
|
|
static tree fold_builtin_printf (location_t, tree, tree, tree, bool, enum built_in_function);
|
203 |
|
|
static tree fold_builtin_fprintf (location_t, tree, tree, tree, tree, bool,
|
204 |
|
|
enum built_in_function);
|
205 |
|
|
static bool init_target_chars (void);
|
206 |
|
|
|
207 |
|
|
static unsigned HOST_WIDE_INT target_newline;
|
208 |
|
|
static unsigned HOST_WIDE_INT target_percent;
|
209 |
|
|
static unsigned HOST_WIDE_INT target_c;
|
210 |
|
|
static unsigned HOST_WIDE_INT target_s;
|
211 |
|
|
static char target_percent_c[3];
|
212 |
|
|
static char target_percent_s[3];
|
213 |
|
|
static char target_percent_s_newline[4];
|
214 |
|
|
static tree do_mpfr_arg1 (tree, tree, int (*)(mpfr_ptr, mpfr_srcptr, mp_rnd_t),
|
215 |
|
|
const REAL_VALUE_TYPE *, const REAL_VALUE_TYPE *, bool);
|
216 |
|
|
static tree do_mpfr_arg2 (tree, tree, tree,
|
217 |
|
|
int (*)(mpfr_ptr, mpfr_srcptr, mpfr_srcptr, mp_rnd_t));
|
218 |
|
|
static tree do_mpfr_arg3 (tree, tree, tree, tree,
|
219 |
|
|
int (*)(mpfr_ptr, mpfr_srcptr, mpfr_srcptr, mpfr_srcptr, mp_rnd_t));
|
220 |
|
|
static tree do_mpfr_sincos (tree, tree, tree);
|
221 |
|
|
static tree do_mpfr_bessel_n (tree, tree, tree,
|
222 |
|
|
int (*)(mpfr_ptr, long, mpfr_srcptr, mp_rnd_t),
|
223 |
|
|
const REAL_VALUE_TYPE *, bool);
|
224 |
|
|
static tree do_mpfr_remquo (tree, tree, tree);
|
225 |
|
|
static tree do_mpfr_lgamma_r (tree, tree, tree);
|
226 |
|
|
static void expand_builtin_sync_synchronize (void);
|
227 |
|
|
|
228 |
|
|
/* Return true if NAME starts with __builtin_ or __sync_. */
|
229 |
|
|
|
230 |
|
|
static bool
|
231 |
|
|
is_builtin_name (const char *name)
|
232 |
|
|
{
|
233 |
|
|
if (strncmp (name, "__builtin_", 10) == 0)
|
234 |
|
|
return true;
|
235 |
|
|
if (strncmp (name, "__sync_", 7) == 0)
|
236 |
|
|
return true;
|
237 |
|
|
if (strncmp (name, "__atomic_", 9) == 0)
|
238 |
|
|
return true;
|
239 |
|
|
return false;
|
240 |
|
|
}
|
241 |
|
|
|
242 |
|
|
|
243 |
|
|
/* Return true if DECL is a function symbol representing a built-in. */
|
244 |
|
|
|
245 |
|
|
bool
|
246 |
|
|
is_builtin_fn (tree decl)
|
247 |
|
|
{
|
248 |
|
|
return TREE_CODE (decl) == FUNCTION_DECL && DECL_BUILT_IN (decl);
|
249 |
|
|
}
|
250 |
|
|
|
251 |
|
|
|
252 |
|
|
/* Return true if NODE should be considered for inline expansion regardless
|
253 |
|
|
of the optimization level. This means whenever a function is invoked with
|
254 |
|
|
its "internal" name, which normally contains the prefix "__builtin". */
|
255 |
|
|
|
256 |
|
|
static bool
|
257 |
|
|
called_as_built_in (tree node)
|
258 |
|
|
{
|
259 |
|
|
/* Note that we must use DECL_NAME, not DECL_ASSEMBLER_NAME_SET_P since
|
260 |
|
|
we want the name used to call the function, not the name it
|
261 |
|
|
will have. */
|
262 |
|
|
const char *name = IDENTIFIER_POINTER (DECL_NAME (node));
|
263 |
|
|
return is_builtin_name (name);
|
264 |
|
|
}
|
265 |
|
|
|
266 |
|
|
/* Compute values M and N such that M divides (address of EXP - N) and
|
267 |
|
|
such that N < M. Store N in *BITPOSP and return M.
|
268 |
|
|
|
269 |
|
|
Note that the address (and thus the alignment) computed here is based
|
270 |
|
|
on the address to which a symbol resolves, whereas DECL_ALIGN is based
|
271 |
|
|
on the address at which an object is actually located. These two
|
272 |
|
|
addresses are not always the same. For example, on ARM targets,
|
273 |
|
|
the address &foo of a Thumb function foo() has the lowest bit set,
|
274 |
|
|
whereas foo() itself starts on an even address. */
|
275 |
|
|
|
276 |
|
|
unsigned int
|
277 |
|
|
get_object_alignment_1 (tree exp, unsigned HOST_WIDE_INT *bitposp)
|
278 |
|
|
{
|
279 |
|
|
HOST_WIDE_INT bitsize, bitpos;
|
280 |
|
|
tree offset;
|
281 |
|
|
enum machine_mode mode;
|
282 |
|
|
int unsignedp, volatilep;
|
283 |
|
|
unsigned int align, inner;
|
284 |
|
|
|
285 |
|
|
/* Get the innermost object and the constant (bitpos) and possibly
|
286 |
|
|
variable (offset) offset of the access. */
|
287 |
|
|
exp = get_inner_reference (exp, &bitsize, &bitpos, &offset,
|
288 |
|
|
&mode, &unsignedp, &volatilep, true);
|
289 |
|
|
|
290 |
|
|
/* Extract alignment information from the innermost object and
|
291 |
|
|
possibly adjust bitpos and offset. */
|
292 |
|
|
if (TREE_CODE (exp) == CONST_DECL)
|
293 |
|
|
exp = DECL_INITIAL (exp);
|
294 |
|
|
if (DECL_P (exp)
|
295 |
|
|
&& TREE_CODE (exp) != LABEL_DECL)
|
296 |
|
|
{
|
297 |
|
|
if (TREE_CODE (exp) == FUNCTION_DECL)
|
298 |
|
|
{
|
299 |
|
|
/* Function addresses can encode extra information besides their
|
300 |
|
|
alignment. However, if TARGET_PTRMEMFUNC_VBIT_LOCATION
|
301 |
|
|
allows the low bit to be used as a virtual bit, we know
|
302 |
|
|
that the address itself must be 2-byte aligned. */
|
303 |
|
|
if (TARGET_PTRMEMFUNC_VBIT_LOCATION == ptrmemfunc_vbit_in_pfn)
|
304 |
|
|
align = 2 * BITS_PER_UNIT;
|
305 |
|
|
else
|
306 |
|
|
align = BITS_PER_UNIT;
|
307 |
|
|
}
|
308 |
|
|
else
|
309 |
|
|
align = DECL_ALIGN (exp);
|
310 |
|
|
}
|
311 |
|
|
else if (CONSTANT_CLASS_P (exp))
|
312 |
|
|
{
|
313 |
|
|
align = TYPE_ALIGN (TREE_TYPE (exp));
|
314 |
|
|
#ifdef CONSTANT_ALIGNMENT
|
315 |
|
|
align = (unsigned)CONSTANT_ALIGNMENT (exp, align);
|
316 |
|
|
#endif
|
317 |
|
|
}
|
318 |
|
|
else if (TREE_CODE (exp) == VIEW_CONVERT_EXPR)
|
319 |
|
|
align = TYPE_ALIGN (TREE_TYPE (exp));
|
320 |
|
|
else if (TREE_CODE (exp) == INDIRECT_REF)
|
321 |
|
|
align = TYPE_ALIGN (TREE_TYPE (exp));
|
322 |
|
|
else if (TREE_CODE (exp) == MEM_REF)
|
323 |
|
|
{
|
324 |
|
|
tree addr = TREE_OPERAND (exp, 0);
|
325 |
|
|
struct ptr_info_def *pi;
|
326 |
|
|
if (TREE_CODE (addr) == BIT_AND_EXPR
|
327 |
|
|
&& TREE_CODE (TREE_OPERAND (addr, 1)) == INTEGER_CST)
|
328 |
|
|
{
|
329 |
|
|
align = (TREE_INT_CST_LOW (TREE_OPERAND (addr, 1))
|
330 |
|
|
& -TREE_INT_CST_LOW (TREE_OPERAND (addr, 1)));
|
331 |
|
|
align *= BITS_PER_UNIT;
|
332 |
|
|
addr = TREE_OPERAND (addr, 0);
|
333 |
|
|
}
|
334 |
|
|
else
|
335 |
|
|
align = BITS_PER_UNIT;
|
336 |
|
|
if (TREE_CODE (addr) == SSA_NAME
|
337 |
|
|
&& (pi = SSA_NAME_PTR_INFO (addr)))
|
338 |
|
|
{
|
339 |
|
|
bitpos += (pi->misalign * BITS_PER_UNIT) & ~(align - 1);
|
340 |
|
|
align = MAX (pi->align * BITS_PER_UNIT, align);
|
341 |
|
|
}
|
342 |
|
|
else if (TREE_CODE (addr) == ADDR_EXPR)
|
343 |
|
|
align = MAX (align, get_object_alignment (TREE_OPERAND (addr, 0)));
|
344 |
|
|
bitpos += mem_ref_offset (exp).low * BITS_PER_UNIT;
|
345 |
|
|
}
|
346 |
|
|
else if (TREE_CODE (exp) == TARGET_MEM_REF)
|
347 |
|
|
{
|
348 |
|
|
struct ptr_info_def *pi;
|
349 |
|
|
tree addr = TMR_BASE (exp);
|
350 |
|
|
if (TREE_CODE (addr) == BIT_AND_EXPR
|
351 |
|
|
&& TREE_CODE (TREE_OPERAND (addr, 1)) == INTEGER_CST)
|
352 |
|
|
{
|
353 |
|
|
align = (TREE_INT_CST_LOW (TREE_OPERAND (addr, 1))
|
354 |
|
|
& -TREE_INT_CST_LOW (TREE_OPERAND (addr, 1)));
|
355 |
|
|
align *= BITS_PER_UNIT;
|
356 |
|
|
addr = TREE_OPERAND (addr, 0);
|
357 |
|
|
}
|
358 |
|
|
else
|
359 |
|
|
align = BITS_PER_UNIT;
|
360 |
|
|
if (TREE_CODE (addr) == SSA_NAME
|
361 |
|
|
&& (pi = SSA_NAME_PTR_INFO (addr)))
|
362 |
|
|
{
|
363 |
|
|
bitpos += (pi->misalign * BITS_PER_UNIT) & ~(align - 1);
|
364 |
|
|
align = MAX (pi->align * BITS_PER_UNIT, align);
|
365 |
|
|
}
|
366 |
|
|
else if (TREE_CODE (addr) == ADDR_EXPR)
|
367 |
|
|
align = MAX (align, get_object_alignment (TREE_OPERAND (addr, 0)));
|
368 |
|
|
if (TMR_OFFSET (exp))
|
369 |
|
|
bitpos += TREE_INT_CST_LOW (TMR_OFFSET (exp)) * BITS_PER_UNIT;
|
370 |
|
|
if (TMR_INDEX (exp) && TMR_STEP (exp))
|
371 |
|
|
{
|
372 |
|
|
unsigned HOST_WIDE_INT step = TREE_INT_CST_LOW (TMR_STEP (exp));
|
373 |
|
|
align = MIN (align, (step & -step) * BITS_PER_UNIT);
|
374 |
|
|
}
|
375 |
|
|
else if (TMR_INDEX (exp))
|
376 |
|
|
align = BITS_PER_UNIT;
|
377 |
|
|
if (TMR_INDEX2 (exp))
|
378 |
|
|
align = BITS_PER_UNIT;
|
379 |
|
|
}
|
380 |
|
|
else
|
381 |
|
|
align = BITS_PER_UNIT;
|
382 |
|
|
|
383 |
|
|
/* If there is a non-constant offset part extract the maximum
|
384 |
|
|
alignment that can prevail. */
|
385 |
|
|
inner = ~0U;
|
386 |
|
|
while (offset)
|
387 |
|
|
{
|
388 |
|
|
tree next_offset;
|
389 |
|
|
|
390 |
|
|
if (TREE_CODE (offset) == PLUS_EXPR)
|
391 |
|
|
{
|
392 |
|
|
next_offset = TREE_OPERAND (offset, 0);
|
393 |
|
|
offset = TREE_OPERAND (offset, 1);
|
394 |
|
|
}
|
395 |
|
|
else
|
396 |
|
|
next_offset = NULL;
|
397 |
|
|
if (host_integerp (offset, 1))
|
398 |
|
|
{
|
399 |
|
|
/* Any overflow in calculating offset_bits won't change
|
400 |
|
|
the alignment. */
|
401 |
|
|
unsigned offset_bits
|
402 |
|
|
= ((unsigned) tree_low_cst (offset, 1) * BITS_PER_UNIT);
|
403 |
|
|
|
404 |
|
|
if (offset_bits)
|
405 |
|
|
inner = MIN (inner, (offset_bits & -offset_bits));
|
406 |
|
|
}
|
407 |
|
|
else if (TREE_CODE (offset) == MULT_EXPR
|
408 |
|
|
&& host_integerp (TREE_OPERAND (offset, 1), 1))
|
409 |
|
|
{
|
410 |
|
|
/* Any overflow in calculating offset_factor won't change
|
411 |
|
|
the alignment. */
|
412 |
|
|
unsigned offset_factor
|
413 |
|
|
= ((unsigned) tree_low_cst (TREE_OPERAND (offset, 1), 1)
|
414 |
|
|
* BITS_PER_UNIT);
|
415 |
|
|
|
416 |
|
|
if (offset_factor)
|
417 |
|
|
inner = MIN (inner, (offset_factor & -offset_factor));
|
418 |
|
|
}
|
419 |
|
|
else
|
420 |
|
|
{
|
421 |
|
|
inner = MIN (inner, BITS_PER_UNIT);
|
422 |
|
|
break;
|
423 |
|
|
}
|
424 |
|
|
offset = next_offset;
|
425 |
|
|
}
|
426 |
|
|
|
427 |
|
|
/* Alignment is innermost object alignment adjusted by the constant
|
428 |
|
|
and non-constant offset parts. */
|
429 |
|
|
align = MIN (align, inner);
|
430 |
|
|
bitpos = bitpos & (align - 1);
|
431 |
|
|
|
432 |
|
|
*bitposp = bitpos;
|
433 |
|
|
return align;
|
434 |
|
|
}
|
435 |
|
|
|
436 |
|
|
/* Return the alignment in bits of EXP, an object. */
|
437 |
|
|
|
438 |
|
|
unsigned int
|
439 |
|
|
get_object_alignment (tree exp)
|
440 |
|
|
{
|
441 |
|
|
unsigned HOST_WIDE_INT bitpos = 0;
|
442 |
|
|
unsigned int align;
|
443 |
|
|
|
444 |
|
|
align = get_object_alignment_1 (exp, &bitpos);
|
445 |
|
|
|
446 |
|
|
/* align and bitpos now specify known low bits of the pointer.
|
447 |
|
|
ptr & (align - 1) == bitpos. */
|
448 |
|
|
|
449 |
|
|
if (bitpos != 0)
|
450 |
|
|
align = (bitpos & -bitpos);
|
451 |
|
|
|
452 |
|
|
return align;
|
453 |
|
|
}
|
454 |
|
|
|
455 |
|
|
/* Return the alignment of object EXP, also considering its type when we do
|
456 |
|
|
not know of explicit misalignment. Only handle MEM_REF and TARGET_MEM_REF.
|
457 |
|
|
|
458 |
|
|
??? Note that, in the general case, the type of an expression is not kept
|
459 |
|
|
consistent with misalignment information by the front-end, for example when
|
460 |
|
|
taking the address of a member of a packed structure. However, in most of
|
461 |
|
|
the cases, expressions have the alignment of their type so we optimistically
|
462 |
|
|
fall back to this alignment when we cannot compute a misalignment. */
|
463 |
|
|
|
464 |
|
|
unsigned int
|
465 |
|
|
get_object_or_type_alignment (tree exp)
|
466 |
|
|
{
|
467 |
|
|
unsigned HOST_WIDE_INT misalign;
|
468 |
|
|
unsigned int align = get_object_alignment_1 (exp, &misalign);
|
469 |
|
|
|
470 |
|
|
gcc_assert (TREE_CODE (exp) == MEM_REF || TREE_CODE (exp) == TARGET_MEM_REF);
|
471 |
|
|
|
472 |
|
|
if (misalign != 0)
|
473 |
|
|
align = (misalign & -misalign);
|
474 |
|
|
else
|
475 |
|
|
align = MAX (TYPE_ALIGN (TREE_TYPE (exp)), align);
|
476 |
|
|
|
477 |
|
|
return align;
|
478 |
|
|
}
|
479 |
|
|
|
480 |
|
|
/* For a pointer valued expression EXP compute values M and N such that
|
481 |
|
|
M divides (EXP - N) and such that N < M. Store N in *BITPOSP and return M.
|
482 |
|
|
|
483 |
|
|
If EXP is not a pointer, 0 is returned. */
|
484 |
|
|
|
485 |
|
|
unsigned int
|
486 |
|
|
get_pointer_alignment_1 (tree exp, unsigned HOST_WIDE_INT *bitposp)
|
487 |
|
|
{
|
488 |
|
|
STRIP_NOPS (exp);
|
489 |
|
|
|
490 |
|
|
if (TREE_CODE (exp) == ADDR_EXPR)
|
491 |
|
|
return get_object_alignment_1 (TREE_OPERAND (exp, 0), bitposp);
|
492 |
|
|
else if (TREE_CODE (exp) == SSA_NAME
|
493 |
|
|
&& POINTER_TYPE_P (TREE_TYPE (exp)))
|
494 |
|
|
{
|
495 |
|
|
struct ptr_info_def *pi = SSA_NAME_PTR_INFO (exp);
|
496 |
|
|
if (!pi)
|
497 |
|
|
{
|
498 |
|
|
*bitposp = 0;
|
499 |
|
|
return BITS_PER_UNIT;
|
500 |
|
|
}
|
501 |
|
|
*bitposp = pi->misalign * BITS_PER_UNIT;
|
502 |
|
|
return pi->align * BITS_PER_UNIT;
|
503 |
|
|
}
|
504 |
|
|
|
505 |
|
|
*bitposp = 0;
|
506 |
|
|
return POINTER_TYPE_P (TREE_TYPE (exp)) ? BITS_PER_UNIT : 0;
|
507 |
|
|
}
|
508 |
|
|
|
509 |
|
|
/* Return the alignment in bits of EXP, a pointer valued expression.
|
510 |
|
|
The alignment returned is, by default, the alignment of the thing that
|
511 |
|
|
EXP points to. If it is not a POINTER_TYPE, 0 is returned.
|
512 |
|
|
|
513 |
|
|
Otherwise, look at the expression to see if we can do better, i.e., if the
|
514 |
|
|
expression is actually pointing at an object whose alignment is tighter. */
|
515 |
|
|
|
516 |
|
|
unsigned int
|
517 |
|
|
get_pointer_alignment (tree exp)
|
518 |
|
|
{
|
519 |
|
|
unsigned HOST_WIDE_INT bitpos = 0;
|
520 |
|
|
unsigned int align;
|
521 |
|
|
|
522 |
|
|
align = get_pointer_alignment_1 (exp, &bitpos);
|
523 |
|
|
|
524 |
|
|
/* align and bitpos now specify known low bits of the pointer.
|
525 |
|
|
ptr & (align - 1) == bitpos. */
|
526 |
|
|
|
527 |
|
|
if (bitpos != 0)
|
528 |
|
|
align = (bitpos & -bitpos);
|
529 |
|
|
|
530 |
|
|
return align;
|
531 |
|
|
}
|
532 |
|
|
|
533 |
|
|
/* Compute the length of a C string. TREE_STRING_LENGTH is not the right
|
534 |
|
|
way, because it could contain a zero byte in the middle.
|
535 |
|
|
TREE_STRING_LENGTH is the size of the character array, not the string.
|
536 |
|
|
|
537 |
|
|
ONLY_VALUE should be nonzero if the result is not going to be emitted
|
538 |
|
|
into the instruction stream and zero if it is going to be expanded.
|
539 |
|
|
E.g. with i++ ? "foo" : "bar", if ONLY_VALUE is nonzero, constant 3
|
540 |
|
|
is returned, otherwise NULL, since
|
541 |
|
|
len = c_strlen (src, 1); if (len) expand_expr (len, ...); would not
|
542 |
|
|
evaluate the side-effects.
|
543 |
|
|
|
544 |
|
|
The value returned is of type `ssizetype'.
|
545 |
|
|
|
546 |
|
|
Unfortunately, string_constant can't access the values of const char
|
547 |
|
|
arrays with initializers, so neither can we do so here. */
|
548 |
|
|
|
549 |
|
|
tree
|
550 |
|
|
c_strlen (tree src, int only_value)
|
551 |
|
|
{
|
552 |
|
|
tree offset_node;
|
553 |
|
|
HOST_WIDE_INT offset;
|
554 |
|
|
int max;
|
555 |
|
|
const char *ptr;
|
556 |
|
|
location_t loc;
|
557 |
|
|
|
558 |
|
|
STRIP_NOPS (src);
|
559 |
|
|
if (TREE_CODE (src) == COND_EXPR
|
560 |
|
|
&& (only_value || !TREE_SIDE_EFFECTS (TREE_OPERAND (src, 0))))
|
561 |
|
|
{
|
562 |
|
|
tree len1, len2;
|
563 |
|
|
|
564 |
|
|
len1 = c_strlen (TREE_OPERAND (src, 1), only_value);
|
565 |
|
|
len2 = c_strlen (TREE_OPERAND (src, 2), only_value);
|
566 |
|
|
if (tree_int_cst_equal (len1, len2))
|
567 |
|
|
return len1;
|
568 |
|
|
}
|
569 |
|
|
|
570 |
|
|
if (TREE_CODE (src) == COMPOUND_EXPR
|
571 |
|
|
&& (only_value || !TREE_SIDE_EFFECTS (TREE_OPERAND (src, 0))))
|
572 |
|
|
return c_strlen (TREE_OPERAND (src, 1), only_value);
|
573 |
|
|
|
574 |
|
|
loc = EXPR_LOC_OR_HERE (src);
|
575 |
|
|
|
576 |
|
|
src = string_constant (src, &offset_node);
|
577 |
|
|
if (src == 0)
|
578 |
|
|
return NULL_TREE;
|
579 |
|
|
|
580 |
|
|
max = TREE_STRING_LENGTH (src) - 1;
|
581 |
|
|
ptr = TREE_STRING_POINTER (src);
|
582 |
|
|
|
583 |
|
|
if (offset_node && TREE_CODE (offset_node) != INTEGER_CST)
|
584 |
|
|
{
|
585 |
|
|
/* If the string has an internal zero byte (e.g., "foo\0bar"), we can't
|
586 |
|
|
compute the offset to the following null if we don't know where to
|
587 |
|
|
start searching for it. */
|
588 |
|
|
int i;
|
589 |
|
|
|
590 |
|
|
for (i = 0; i < max; i++)
|
591 |
|
|
if (ptr[i] == 0)
|
592 |
|
|
return NULL_TREE;
|
593 |
|
|
|
594 |
|
|
/* We don't know the starting offset, but we do know that the string
|
595 |
|
|
has no internal zero bytes. We can assume that the offset falls
|
596 |
|
|
within the bounds of the string; otherwise, the programmer deserves
|
597 |
|
|
what he gets. Subtract the offset from the length of the string,
|
598 |
|
|
and return that. This would perhaps not be valid if we were dealing
|
599 |
|
|
with named arrays in addition to literal string constants. */
|
600 |
|
|
|
601 |
|
|
return size_diffop_loc (loc, size_int (max), offset_node);
|
602 |
|
|
}
|
603 |
|
|
|
604 |
|
|
/* We have a known offset into the string. Start searching there for
|
605 |
|
|
a null character if we can represent it as a single HOST_WIDE_INT. */
|
606 |
|
|
if (offset_node == 0)
|
607 |
|
|
offset = 0;
|
608 |
|
|
else if (! host_integerp (offset_node, 0))
|
609 |
|
|
offset = -1;
|
610 |
|
|
else
|
611 |
|
|
offset = tree_low_cst (offset_node, 0);
|
612 |
|
|
|
613 |
|
|
/* If the offset is known to be out of bounds, warn, and call strlen at
|
614 |
|
|
runtime. */
|
615 |
|
|
if (offset < 0 || offset > max)
|
616 |
|
|
{
|
617 |
|
|
/* Suppress multiple warnings for propagated constant strings. */
|
618 |
|
|
if (! TREE_NO_WARNING (src))
|
619 |
|
|
{
|
620 |
|
|
warning_at (loc, 0, "offset outside bounds of constant string");
|
621 |
|
|
TREE_NO_WARNING (src) = 1;
|
622 |
|
|
}
|
623 |
|
|
return NULL_TREE;
|
624 |
|
|
}
|
625 |
|
|
|
626 |
|
|
/* Use strlen to search for the first zero byte. Since any strings
|
627 |
|
|
constructed with build_string will have nulls appended, we win even
|
628 |
|
|
if we get handed something like (char[4])"abcd".
|
629 |
|
|
|
630 |
|
|
Since OFFSET is our starting index into the string, no further
|
631 |
|
|
calculation is needed. */
|
632 |
|
|
return ssize_int (strlen (ptr + offset));
|
633 |
|
|
}
|
634 |
|
|
|
635 |
|
|
/* Return a char pointer for a C string if it is a string constant
|
636 |
|
|
or sum of string constant and integer constant. */
|
637 |
|
|
|
638 |
|
|
static const char *
|
639 |
|
|
c_getstr (tree src)
|
640 |
|
|
{
|
641 |
|
|
tree offset_node;
|
642 |
|
|
|
643 |
|
|
src = string_constant (src, &offset_node);
|
644 |
|
|
if (src == 0)
|
645 |
|
|
return 0;
|
646 |
|
|
|
647 |
|
|
if (offset_node == 0)
|
648 |
|
|
return TREE_STRING_POINTER (src);
|
649 |
|
|
else if (!host_integerp (offset_node, 1)
|
650 |
|
|
|| compare_tree_int (offset_node, TREE_STRING_LENGTH (src) - 1) > 0)
|
651 |
|
|
return 0;
|
652 |
|
|
|
653 |
|
|
return TREE_STRING_POINTER (src) + tree_low_cst (offset_node, 1);
|
654 |
|
|
}
|
655 |
|
|
|
656 |
|
|
/* Return a CONST_INT or CONST_DOUBLE corresponding to target reading
|
657 |
|
|
GET_MODE_BITSIZE (MODE) bits from string constant STR. */
|
658 |
|
|
|
659 |
|
|
static rtx
|
660 |
|
|
c_readstr (const char *str, enum machine_mode mode)
|
661 |
|
|
{
|
662 |
|
|
HOST_WIDE_INT c[2];
|
663 |
|
|
HOST_WIDE_INT ch;
|
664 |
|
|
unsigned int i, j;
|
665 |
|
|
|
666 |
|
|
gcc_assert (GET_MODE_CLASS (mode) == MODE_INT);
|
667 |
|
|
|
668 |
|
|
c[0] = 0;
|
669 |
|
|
c[1] = 0;
|
670 |
|
|
ch = 1;
|
671 |
|
|
for (i = 0; i < GET_MODE_SIZE (mode); i++)
|
672 |
|
|
{
|
673 |
|
|
j = i;
|
674 |
|
|
if (WORDS_BIG_ENDIAN)
|
675 |
|
|
j = GET_MODE_SIZE (mode) - i - 1;
|
676 |
|
|
if (BYTES_BIG_ENDIAN != WORDS_BIG_ENDIAN
|
677 |
|
|
&& GET_MODE_SIZE (mode) >= UNITS_PER_WORD)
|
678 |
|
|
j = j + UNITS_PER_WORD - 2 * (j % UNITS_PER_WORD) - 1;
|
679 |
|
|
j *= BITS_PER_UNIT;
|
680 |
|
|
gcc_assert (j < 2 * HOST_BITS_PER_WIDE_INT);
|
681 |
|
|
|
682 |
|
|
if (ch)
|
683 |
|
|
ch = (unsigned char) str[i];
|
684 |
|
|
c[j / HOST_BITS_PER_WIDE_INT] |= ch << (j % HOST_BITS_PER_WIDE_INT);
|
685 |
|
|
}
|
686 |
|
|
return immed_double_const (c[0], c[1], mode);
|
687 |
|
|
}
|
688 |
|
|
|
689 |
|
|
/* Cast a target constant CST to target CHAR and if that value fits into
|
690 |
|
|
host char type, return zero and put that value into variable pointed to by
|
691 |
|
|
P. */
|
692 |
|
|
|
693 |
|
|
static int
|
694 |
|
|
target_char_cast (tree cst, char *p)
|
695 |
|
|
{
|
696 |
|
|
unsigned HOST_WIDE_INT val, hostval;
|
697 |
|
|
|
698 |
|
|
if (TREE_CODE (cst) != INTEGER_CST
|
699 |
|
|
|| CHAR_TYPE_SIZE > HOST_BITS_PER_WIDE_INT)
|
700 |
|
|
return 1;
|
701 |
|
|
|
702 |
|
|
val = TREE_INT_CST_LOW (cst);
|
703 |
|
|
if (CHAR_TYPE_SIZE < HOST_BITS_PER_WIDE_INT)
|
704 |
|
|
val &= (((unsigned HOST_WIDE_INT) 1) << CHAR_TYPE_SIZE) - 1;
|
705 |
|
|
|
706 |
|
|
hostval = val;
|
707 |
|
|
if (HOST_BITS_PER_CHAR < HOST_BITS_PER_WIDE_INT)
|
708 |
|
|
hostval &= (((unsigned HOST_WIDE_INT) 1) << HOST_BITS_PER_CHAR) - 1;
|
709 |
|
|
|
710 |
|
|
if (val != hostval)
|
711 |
|
|
return 1;
|
712 |
|
|
|
713 |
|
|
*p = hostval;
|
714 |
|
|
return 0;
|
715 |
|
|
}
|
716 |
|
|
|
717 |
|
|
/* Similar to save_expr, but assumes that arbitrary code is not executed
|
718 |
|
|
in between the multiple evaluations. In particular, we assume that a
|
719 |
|
|
non-addressable local variable will not be modified. */
|
720 |
|
|
|
721 |
|
|
static tree
|
722 |
|
|
builtin_save_expr (tree exp)
|
723 |
|
|
{
|
724 |
|
|
if (TREE_CODE (exp) == SSA_NAME
|
725 |
|
|
|| (TREE_ADDRESSABLE (exp) == 0
|
726 |
|
|
&& (TREE_CODE (exp) == PARM_DECL
|
727 |
|
|
|| (TREE_CODE (exp) == VAR_DECL && !TREE_STATIC (exp)))))
|
728 |
|
|
return exp;
|
729 |
|
|
|
730 |
|
|
return save_expr (exp);
|
731 |
|
|
}
|
732 |
|
|
|
733 |
|
|
/* Given TEM, a pointer to a stack frame, follow the dynamic chain COUNT
|
734 |
|
|
times to get the address of either a higher stack frame, or a return
|
735 |
|
|
address located within it (depending on FNDECL_CODE). */
|
736 |
|
|
|
737 |
|
|
static rtx
|
738 |
|
|
expand_builtin_return_addr (enum built_in_function fndecl_code, int count)
|
739 |
|
|
{
|
740 |
|
|
int i;
|
741 |
|
|
|
742 |
|
|
#ifdef INITIAL_FRAME_ADDRESS_RTX
|
743 |
|
|
rtx tem = INITIAL_FRAME_ADDRESS_RTX;
|
744 |
|
|
#else
|
745 |
|
|
rtx tem;
|
746 |
|
|
|
747 |
|
|
/* For a zero count with __builtin_return_address, we don't care what
|
748 |
|
|
frame address we return, because target-specific definitions will
|
749 |
|
|
override us. Therefore frame pointer elimination is OK, and using
|
750 |
|
|
the soft frame pointer is OK.
|
751 |
|
|
|
752 |
|
|
For a nonzero count, or a zero count with __builtin_frame_address,
|
753 |
|
|
we require a stable offset from the current frame pointer to the
|
754 |
|
|
previous one, so we must use the hard frame pointer, and
|
755 |
|
|
we must disable frame pointer elimination. */
|
756 |
|
|
if (count == 0 && fndecl_code == BUILT_IN_RETURN_ADDRESS)
|
757 |
|
|
tem = frame_pointer_rtx;
|
758 |
|
|
else
|
759 |
|
|
{
|
760 |
|
|
tem = hard_frame_pointer_rtx;
|
761 |
|
|
|
762 |
|
|
/* Tell reload not to eliminate the frame pointer. */
|
763 |
|
|
crtl->accesses_prior_frames = 1;
|
764 |
|
|
}
|
765 |
|
|
#endif
|
766 |
|
|
|
767 |
|
|
/* Some machines need special handling before we can access
|
768 |
|
|
arbitrary frames. For example, on the SPARC, we must first flush
|
769 |
|
|
all register windows to the stack. */
|
770 |
|
|
#ifdef SETUP_FRAME_ADDRESSES
|
771 |
|
|
if (count > 0)
|
772 |
|
|
SETUP_FRAME_ADDRESSES ();
|
773 |
|
|
#endif
|
774 |
|
|
|
775 |
|
|
/* On the SPARC, the return address is not in the frame, it is in a
|
776 |
|
|
register. There is no way to access it off of the current frame
|
777 |
|
|
pointer, but it can be accessed off the previous frame pointer by
|
778 |
|
|
reading the value from the register window save area. */
|
779 |
|
|
#ifdef RETURN_ADDR_IN_PREVIOUS_FRAME
|
780 |
|
|
if (fndecl_code == BUILT_IN_RETURN_ADDRESS)
|
781 |
|
|
count--;
|
782 |
|
|
#endif
|
783 |
|
|
|
784 |
|
|
/* Scan back COUNT frames to the specified frame. */
|
785 |
|
|
for (i = 0; i < count; i++)
|
786 |
|
|
{
|
787 |
|
|
/* Assume the dynamic chain pointer is in the word that the
|
788 |
|
|
frame address points to, unless otherwise specified. */
|
789 |
|
|
#ifdef DYNAMIC_CHAIN_ADDRESS
|
790 |
|
|
tem = DYNAMIC_CHAIN_ADDRESS (tem);
|
791 |
|
|
#endif
|
792 |
|
|
tem = memory_address (Pmode, tem);
|
793 |
|
|
tem = gen_frame_mem (Pmode, tem);
|
794 |
|
|
tem = copy_to_reg (tem);
|
795 |
|
|
}
|
796 |
|
|
|
797 |
|
|
/* For __builtin_frame_address, return what we've got. But, on
|
798 |
|
|
the SPARC for example, we may have to add a bias. */
|
799 |
|
|
if (fndecl_code == BUILT_IN_FRAME_ADDRESS)
|
800 |
|
|
#ifdef FRAME_ADDR_RTX
|
801 |
|
|
return FRAME_ADDR_RTX (tem);
|
802 |
|
|
#else
|
803 |
|
|
return tem;
|
804 |
|
|
#endif
|
805 |
|
|
|
806 |
|
|
/* For __builtin_return_address, get the return address from that frame. */
|
807 |
|
|
#ifdef RETURN_ADDR_RTX
|
808 |
|
|
tem = RETURN_ADDR_RTX (count, tem);
|
809 |
|
|
#else
|
810 |
|
|
tem = memory_address (Pmode,
|
811 |
|
|
plus_constant (tem, GET_MODE_SIZE (Pmode)));
|
812 |
|
|
tem = gen_frame_mem (Pmode, tem);
|
813 |
|
|
#endif
|
814 |
|
|
return tem;
|
815 |
|
|
}
|
816 |
|
|
|
817 |
|
|
/* Alias set used for setjmp buffer. */
|
818 |
|
|
static alias_set_type setjmp_alias_set = -1;
|
819 |
|
|
|
820 |
|
|
/* Construct the leading half of a __builtin_setjmp call. Control will
|
821 |
|
|
return to RECEIVER_LABEL. This is also called directly by the SJLJ
|
822 |
|
|
exception handling code. */
|
823 |
|
|
|
824 |
|
|
void
|
825 |
|
|
expand_builtin_setjmp_setup (rtx buf_addr, rtx receiver_label)
|
826 |
|
|
{
|
827 |
|
|
enum machine_mode sa_mode = STACK_SAVEAREA_MODE (SAVE_NONLOCAL);
|
828 |
|
|
rtx stack_save;
|
829 |
|
|
rtx mem;
|
830 |
|
|
|
831 |
|
|
if (setjmp_alias_set == -1)
|
832 |
|
|
setjmp_alias_set = new_alias_set ();
|
833 |
|
|
|
834 |
|
|
buf_addr = convert_memory_address (Pmode, buf_addr);
|
835 |
|
|
|
836 |
|
|
buf_addr = force_reg (Pmode, force_operand (buf_addr, NULL_RTX));
|
837 |
|
|
|
838 |
|
|
/* We store the frame pointer and the address of receiver_label in
|
839 |
|
|
the buffer and use the rest of it for the stack save area, which
|
840 |
|
|
is machine-dependent. */
|
841 |
|
|
|
842 |
|
|
mem = gen_rtx_MEM (Pmode, buf_addr);
|
843 |
|
|
set_mem_alias_set (mem, setjmp_alias_set);
|
844 |
|
|
emit_move_insn (mem, targetm.builtin_setjmp_frame_value ());
|
845 |
|
|
|
846 |
|
|
mem = gen_rtx_MEM (Pmode, plus_constant (buf_addr, GET_MODE_SIZE (Pmode))),
|
847 |
|
|
set_mem_alias_set (mem, setjmp_alias_set);
|
848 |
|
|
|
849 |
|
|
emit_move_insn (validize_mem (mem),
|
850 |
|
|
force_reg (Pmode, gen_rtx_LABEL_REF (Pmode, receiver_label)));
|
851 |
|
|
|
852 |
|
|
stack_save = gen_rtx_MEM (sa_mode,
|
853 |
|
|
plus_constant (buf_addr,
|
854 |
|
|
2 * GET_MODE_SIZE (Pmode)));
|
855 |
|
|
set_mem_alias_set (stack_save, setjmp_alias_set);
|
856 |
|
|
emit_stack_save (SAVE_NONLOCAL, &stack_save);
|
857 |
|
|
|
858 |
|
|
/* If there is further processing to do, do it. */
|
859 |
|
|
#ifdef HAVE_builtin_setjmp_setup
|
860 |
|
|
if (HAVE_builtin_setjmp_setup)
|
861 |
|
|
emit_insn (gen_builtin_setjmp_setup (buf_addr));
|
862 |
|
|
#endif
|
863 |
|
|
|
864 |
|
|
/* We have a nonlocal label. */
|
865 |
|
|
cfun->has_nonlocal_label = 1;
|
866 |
|
|
}
|
867 |
|
|
|
868 |
|
|
/* Construct the trailing part of a __builtin_setjmp call. This is
|
869 |
|
|
also called directly by the SJLJ exception handling code. */
|
870 |
|
|
|
871 |
|
|
void
|
872 |
|
|
expand_builtin_setjmp_receiver (rtx receiver_label ATTRIBUTE_UNUSED)
|
873 |
|
|
{
|
874 |
|
|
rtx chain;
|
875 |
|
|
|
876 |
|
|
/* Clobber the FP when we get here, so we have to make sure it's
|
877 |
|
|
marked as used by this function. */
|
878 |
|
|
emit_use (hard_frame_pointer_rtx);
|
879 |
|
|
|
880 |
|
|
/* Mark the static chain as clobbered here so life information
|
881 |
|
|
doesn't get messed up for it. */
|
882 |
|
|
chain = targetm.calls.static_chain (current_function_decl, true);
|
883 |
|
|
if (chain && REG_P (chain))
|
884 |
|
|
emit_clobber (chain);
|
885 |
|
|
|
886 |
|
|
/* Now put in the code to restore the frame pointer, and argument
|
887 |
|
|
pointer, if needed. */
|
888 |
|
|
#ifdef HAVE_nonlocal_goto
|
889 |
|
|
if (! HAVE_nonlocal_goto)
|
890 |
|
|
#endif
|
891 |
|
|
{
|
892 |
|
|
emit_move_insn (virtual_stack_vars_rtx, hard_frame_pointer_rtx);
|
893 |
|
|
/* This might change the hard frame pointer in ways that aren't
|
894 |
|
|
apparent to early optimization passes, so force a clobber. */
|
895 |
|
|
emit_clobber (hard_frame_pointer_rtx);
|
896 |
|
|
}
|
897 |
|
|
|
898 |
|
|
#if !HARD_FRAME_POINTER_IS_ARG_POINTER
|
899 |
|
|
if (fixed_regs[ARG_POINTER_REGNUM])
|
900 |
|
|
{
|
901 |
|
|
#ifdef ELIMINABLE_REGS
|
902 |
|
|
size_t i;
|
903 |
|
|
static const struct elims {const int from, to;} elim_regs[] = ELIMINABLE_REGS;
|
904 |
|
|
|
905 |
|
|
for (i = 0; i < ARRAY_SIZE (elim_regs); i++)
|
906 |
|
|
if (elim_regs[i].from == ARG_POINTER_REGNUM
|
907 |
|
|
&& elim_regs[i].to == HARD_FRAME_POINTER_REGNUM)
|
908 |
|
|
break;
|
909 |
|
|
|
910 |
|
|
if (i == ARRAY_SIZE (elim_regs))
|
911 |
|
|
#endif
|
912 |
|
|
{
|
913 |
|
|
/* Now restore our arg pointer from the address at which it
|
914 |
|
|
was saved in our stack frame. */
|
915 |
|
|
emit_move_insn (crtl->args.internal_arg_pointer,
|
916 |
|
|
copy_to_reg (get_arg_pointer_save_area ()));
|
917 |
|
|
}
|
918 |
|
|
}
|
919 |
|
|
#endif
|
920 |
|
|
|
921 |
|
|
#ifdef HAVE_builtin_setjmp_receiver
|
922 |
|
|
if (HAVE_builtin_setjmp_receiver)
|
923 |
|
|
emit_insn (gen_builtin_setjmp_receiver (receiver_label));
|
924 |
|
|
else
|
925 |
|
|
#endif
|
926 |
|
|
#ifdef HAVE_nonlocal_goto_receiver
|
927 |
|
|
if (HAVE_nonlocal_goto_receiver)
|
928 |
|
|
emit_insn (gen_nonlocal_goto_receiver ());
|
929 |
|
|
else
|
930 |
|
|
#endif
|
931 |
|
|
{ /* Nothing */ }
|
932 |
|
|
|
933 |
|
|
/* We must not allow the code we just generated to be reordered by
|
934 |
|
|
scheduling. Specifically, the update of the frame pointer must
|
935 |
|
|
happen immediately, not later. */
|
936 |
|
|
emit_insn (gen_blockage ());
|
937 |
|
|
}
|
938 |
|
|
|
939 |
|
|
/* __builtin_longjmp is passed a pointer to an array of five words (not
|
940 |
|
|
all will be used on all machines). It operates similarly to the C
|
941 |
|
|
library function of the same name, but is more efficient. Much of
|
942 |
|
|
the code below is copied from the handling of non-local gotos. */
|
943 |
|
|
|
944 |
|
|
static void
|
945 |
|
|
expand_builtin_longjmp (rtx buf_addr, rtx value)
|
946 |
|
|
{
|
947 |
|
|
rtx fp, lab, stack, insn, last;
|
948 |
|
|
enum machine_mode sa_mode = STACK_SAVEAREA_MODE (SAVE_NONLOCAL);
|
949 |
|
|
|
950 |
|
|
/* DRAP is needed for stack realign if longjmp is expanded to current
|
951 |
|
|
function */
|
952 |
|
|
if (SUPPORTS_STACK_ALIGNMENT)
|
953 |
|
|
crtl->need_drap = true;
|
954 |
|
|
|
955 |
|
|
if (setjmp_alias_set == -1)
|
956 |
|
|
setjmp_alias_set = new_alias_set ();
|
957 |
|
|
|
958 |
|
|
buf_addr = convert_memory_address (Pmode, buf_addr);
|
959 |
|
|
|
960 |
|
|
buf_addr = force_reg (Pmode, buf_addr);
|
961 |
|
|
|
962 |
|
|
/* We require that the user must pass a second argument of 1, because
|
963 |
|
|
that is what builtin_setjmp will return. */
|
964 |
|
|
gcc_assert (value == const1_rtx);
|
965 |
|
|
|
966 |
|
|
last = get_last_insn ();
|
967 |
|
|
#ifdef HAVE_builtin_longjmp
|
968 |
|
|
if (HAVE_builtin_longjmp)
|
969 |
|
|
emit_insn (gen_builtin_longjmp (buf_addr));
|
970 |
|
|
else
|
971 |
|
|
#endif
|
972 |
|
|
{
|
973 |
|
|
fp = gen_rtx_MEM (Pmode, buf_addr);
|
974 |
|
|
lab = gen_rtx_MEM (Pmode, plus_constant (buf_addr,
|
975 |
|
|
GET_MODE_SIZE (Pmode)));
|
976 |
|
|
|
977 |
|
|
stack = gen_rtx_MEM (sa_mode, plus_constant (buf_addr,
|
978 |
|
|
2 * GET_MODE_SIZE (Pmode)));
|
979 |
|
|
set_mem_alias_set (fp, setjmp_alias_set);
|
980 |
|
|
set_mem_alias_set (lab, setjmp_alias_set);
|
981 |
|
|
set_mem_alias_set (stack, setjmp_alias_set);
|
982 |
|
|
|
983 |
|
|
/* Pick up FP, label, and SP from the block and jump. This code is
|
984 |
|
|
from expand_goto in stmt.c; see there for detailed comments. */
|
985 |
|
|
#ifdef HAVE_nonlocal_goto
|
986 |
|
|
if (HAVE_nonlocal_goto)
|
987 |
|
|
/* We have to pass a value to the nonlocal_goto pattern that will
|
988 |
|
|
get copied into the static_chain pointer, but it does not matter
|
989 |
|
|
what that value is, because builtin_setjmp does not use it. */
|
990 |
|
|
emit_insn (gen_nonlocal_goto (value, lab, stack, fp));
|
991 |
|
|
else
|
992 |
|
|
#endif
|
993 |
|
|
{
|
994 |
|
|
lab = copy_to_reg (lab);
|
995 |
|
|
|
996 |
|
|
emit_clobber (gen_rtx_MEM (BLKmode, gen_rtx_SCRATCH (VOIDmode)));
|
997 |
|
|
emit_clobber (gen_rtx_MEM (BLKmode, hard_frame_pointer_rtx));
|
998 |
|
|
|
999 |
|
|
emit_move_insn (hard_frame_pointer_rtx, fp);
|
1000 |
|
|
emit_stack_restore (SAVE_NONLOCAL, stack);
|
1001 |
|
|
|
1002 |
|
|
emit_use (hard_frame_pointer_rtx);
|
1003 |
|
|
emit_use (stack_pointer_rtx);
|
1004 |
|
|
emit_indirect_jump (lab);
|
1005 |
|
|
}
|
1006 |
|
|
}
|
1007 |
|
|
|
1008 |
|
|
/* Search backwards and mark the jump insn as a non-local goto.
|
1009 |
|
|
Note that this precludes the use of __builtin_longjmp to a
|
1010 |
|
|
__builtin_setjmp target in the same function. However, we've
|
1011 |
|
|
already cautioned the user that these functions are for
|
1012 |
|
|
internal exception handling use only. */
|
1013 |
|
|
for (insn = get_last_insn (); insn; insn = PREV_INSN (insn))
|
1014 |
|
|
{
|
1015 |
|
|
gcc_assert (insn != last);
|
1016 |
|
|
|
1017 |
|
|
if (JUMP_P (insn))
|
1018 |
|
|
{
|
1019 |
|
|
add_reg_note (insn, REG_NON_LOCAL_GOTO, const0_rtx);
|
1020 |
|
|
break;
|
1021 |
|
|
}
|
1022 |
|
|
else if (CALL_P (insn))
|
1023 |
|
|
break;
|
1024 |
|
|
}
|
1025 |
|
|
}
|
1026 |
|
|
|
1027 |
|
|
/* Expand a call to __builtin_nonlocal_goto. We're passed the target label
|
1028 |
|
|
and the address of the save area. */
|
1029 |
|
|
|
1030 |
|
|
static rtx
|
1031 |
|
|
expand_builtin_nonlocal_goto (tree exp)
|
1032 |
|
|
{
|
1033 |
|
|
tree t_label, t_save_area;
|
1034 |
|
|
rtx r_label, r_save_area, r_fp, r_sp, insn;
|
1035 |
|
|
|
1036 |
|
|
if (!validate_arglist (exp, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
1037 |
|
|
return NULL_RTX;
|
1038 |
|
|
|
1039 |
|
|
t_label = CALL_EXPR_ARG (exp, 0);
|
1040 |
|
|
t_save_area = CALL_EXPR_ARG (exp, 1);
|
1041 |
|
|
|
1042 |
|
|
r_label = expand_normal (t_label);
|
1043 |
|
|
r_label = convert_memory_address (Pmode, r_label);
|
1044 |
|
|
r_save_area = expand_normal (t_save_area);
|
1045 |
|
|
r_save_area = convert_memory_address (Pmode, r_save_area);
|
1046 |
|
|
/* Copy the address of the save location to a register just in case it was
|
1047 |
|
|
based on the frame pointer. */
|
1048 |
|
|
r_save_area = copy_to_reg (r_save_area);
|
1049 |
|
|
r_fp = gen_rtx_MEM (Pmode, r_save_area);
|
1050 |
|
|
r_sp = gen_rtx_MEM (STACK_SAVEAREA_MODE (SAVE_NONLOCAL),
|
1051 |
|
|
plus_constant (r_save_area, GET_MODE_SIZE (Pmode)));
|
1052 |
|
|
|
1053 |
|
|
crtl->has_nonlocal_goto = 1;
|
1054 |
|
|
|
1055 |
|
|
#ifdef HAVE_nonlocal_goto
|
1056 |
|
|
/* ??? We no longer need to pass the static chain value, afaik. */
|
1057 |
|
|
if (HAVE_nonlocal_goto)
|
1058 |
|
|
emit_insn (gen_nonlocal_goto (const0_rtx, r_label, r_sp, r_fp));
|
1059 |
|
|
else
|
1060 |
|
|
#endif
|
1061 |
|
|
{
|
1062 |
|
|
r_label = copy_to_reg (r_label);
|
1063 |
|
|
|
1064 |
|
|
emit_clobber (gen_rtx_MEM (BLKmode, gen_rtx_SCRATCH (VOIDmode)));
|
1065 |
|
|
emit_clobber (gen_rtx_MEM (BLKmode, hard_frame_pointer_rtx));
|
1066 |
|
|
|
1067 |
|
|
/* Restore frame pointer for containing function. */
|
1068 |
|
|
emit_move_insn (hard_frame_pointer_rtx, r_fp);
|
1069 |
|
|
emit_stack_restore (SAVE_NONLOCAL, r_sp);
|
1070 |
|
|
|
1071 |
|
|
/* USE of hard_frame_pointer_rtx added for consistency;
|
1072 |
|
|
not clear if really needed. */
|
1073 |
|
|
emit_use (hard_frame_pointer_rtx);
|
1074 |
|
|
emit_use (stack_pointer_rtx);
|
1075 |
|
|
|
1076 |
|
|
/* If the architecture is using a GP register, we must
|
1077 |
|
|
conservatively assume that the target function makes use of it.
|
1078 |
|
|
The prologue of functions with nonlocal gotos must therefore
|
1079 |
|
|
initialize the GP register to the appropriate value, and we
|
1080 |
|
|
must then make sure that this value is live at the point
|
1081 |
|
|
of the jump. (Note that this doesn't necessarily apply
|
1082 |
|
|
to targets with a nonlocal_goto pattern; they are free
|
1083 |
|
|
to implement it in their own way. Note also that this is
|
1084 |
|
|
a no-op if the GP register is a global invariant.) */
|
1085 |
|
|
if ((unsigned) PIC_OFFSET_TABLE_REGNUM != INVALID_REGNUM
|
1086 |
|
|
&& fixed_regs[PIC_OFFSET_TABLE_REGNUM])
|
1087 |
|
|
emit_use (pic_offset_table_rtx);
|
1088 |
|
|
|
1089 |
|
|
emit_indirect_jump (r_label);
|
1090 |
|
|
}
|
1091 |
|
|
|
1092 |
|
|
/* Search backwards to the jump insn and mark it as a
|
1093 |
|
|
non-local goto. */
|
1094 |
|
|
for (insn = get_last_insn (); insn; insn = PREV_INSN (insn))
|
1095 |
|
|
{
|
1096 |
|
|
if (JUMP_P (insn))
|
1097 |
|
|
{
|
1098 |
|
|
add_reg_note (insn, REG_NON_LOCAL_GOTO, const0_rtx);
|
1099 |
|
|
break;
|
1100 |
|
|
}
|
1101 |
|
|
else if (CALL_P (insn))
|
1102 |
|
|
break;
|
1103 |
|
|
}
|
1104 |
|
|
|
1105 |
|
|
return const0_rtx;
|
1106 |
|
|
}
|
1107 |
|
|
|
1108 |
|
|
/* __builtin_update_setjmp_buf is passed a pointer to an array of five words
|
1109 |
|
|
(not all will be used on all machines) that was passed to __builtin_setjmp.
|
1110 |
|
|
It updates the stack pointer in that block to correspond to the current
|
1111 |
|
|
stack pointer. */
|
1112 |
|
|
|
1113 |
|
|
static void
|
1114 |
|
|
expand_builtin_update_setjmp_buf (rtx buf_addr)
|
1115 |
|
|
{
|
1116 |
|
|
enum machine_mode sa_mode = STACK_SAVEAREA_MODE (SAVE_NONLOCAL);
|
1117 |
|
|
rtx stack_save
|
1118 |
|
|
= gen_rtx_MEM (sa_mode,
|
1119 |
|
|
memory_address
|
1120 |
|
|
(sa_mode,
|
1121 |
|
|
plus_constant (buf_addr, 2 * GET_MODE_SIZE (Pmode))));
|
1122 |
|
|
|
1123 |
|
|
emit_stack_save (SAVE_NONLOCAL, &stack_save);
|
1124 |
|
|
}
|
1125 |
|
|
|
1126 |
|
|
/* Expand a call to __builtin_prefetch. For a target that does not support
|
1127 |
|
|
data prefetch, evaluate the memory address argument in case it has side
|
1128 |
|
|
effects. */
|
1129 |
|
|
|
1130 |
|
|
static void
|
1131 |
|
|
expand_builtin_prefetch (tree exp)
|
1132 |
|
|
{
|
1133 |
|
|
tree arg0, arg1, arg2;
|
1134 |
|
|
int nargs;
|
1135 |
|
|
rtx op0, op1, op2;
|
1136 |
|
|
|
1137 |
|
|
if (!validate_arglist (exp, POINTER_TYPE, 0))
|
1138 |
|
|
return;
|
1139 |
|
|
|
1140 |
|
|
arg0 = CALL_EXPR_ARG (exp, 0);
|
1141 |
|
|
|
1142 |
|
|
/* Arguments 1 and 2 are optional; argument 1 (read/write) defaults to
|
1143 |
|
|
zero (read) and argument 2 (locality) defaults to 3 (high degree of
|
1144 |
|
|
locality). */
|
1145 |
|
|
nargs = call_expr_nargs (exp);
|
1146 |
|
|
if (nargs > 1)
|
1147 |
|
|
arg1 = CALL_EXPR_ARG (exp, 1);
|
1148 |
|
|
else
|
1149 |
|
|
arg1 = integer_zero_node;
|
1150 |
|
|
if (nargs > 2)
|
1151 |
|
|
arg2 = CALL_EXPR_ARG (exp, 2);
|
1152 |
|
|
else
|
1153 |
|
|
arg2 = integer_three_node;
|
1154 |
|
|
|
1155 |
|
|
/* Argument 0 is an address. */
|
1156 |
|
|
op0 = expand_expr (arg0, NULL_RTX, Pmode, EXPAND_NORMAL);
|
1157 |
|
|
|
1158 |
|
|
/* Argument 1 (read/write flag) must be a compile-time constant int. */
|
1159 |
|
|
if (TREE_CODE (arg1) != INTEGER_CST)
|
1160 |
|
|
{
|
1161 |
|
|
error ("second argument to %<__builtin_prefetch%> must be a constant");
|
1162 |
|
|
arg1 = integer_zero_node;
|
1163 |
|
|
}
|
1164 |
|
|
op1 = expand_normal (arg1);
|
1165 |
|
|
/* Argument 1 must be either zero or one. */
|
1166 |
|
|
if (INTVAL (op1) != 0 && INTVAL (op1) != 1)
|
1167 |
|
|
{
|
1168 |
|
|
warning (0, "invalid second argument to %<__builtin_prefetch%>;"
|
1169 |
|
|
" using zero");
|
1170 |
|
|
op1 = const0_rtx;
|
1171 |
|
|
}
|
1172 |
|
|
|
1173 |
|
|
/* Argument 2 (locality) must be a compile-time constant int. */
|
1174 |
|
|
if (TREE_CODE (arg2) != INTEGER_CST)
|
1175 |
|
|
{
|
1176 |
|
|
error ("third argument to %<__builtin_prefetch%> must be a constant");
|
1177 |
|
|
arg2 = integer_zero_node;
|
1178 |
|
|
}
|
1179 |
|
|
op2 = expand_normal (arg2);
|
1180 |
|
|
/* Argument 2 must be 0, 1, 2, or 3. */
|
1181 |
|
|
if (INTVAL (op2) < 0 || INTVAL (op2) > 3)
|
1182 |
|
|
{
|
1183 |
|
|
warning (0, "invalid third argument to %<__builtin_prefetch%>; using zero");
|
1184 |
|
|
op2 = const0_rtx;
|
1185 |
|
|
}
|
1186 |
|
|
|
1187 |
|
|
#ifdef HAVE_prefetch
|
1188 |
|
|
if (HAVE_prefetch)
|
1189 |
|
|
{
|
1190 |
|
|
struct expand_operand ops[3];
|
1191 |
|
|
|
1192 |
|
|
create_address_operand (&ops[0], op0);
|
1193 |
|
|
create_integer_operand (&ops[1], INTVAL (op1));
|
1194 |
|
|
create_integer_operand (&ops[2], INTVAL (op2));
|
1195 |
|
|
if (maybe_expand_insn (CODE_FOR_prefetch, 3, ops))
|
1196 |
|
|
return;
|
1197 |
|
|
}
|
1198 |
|
|
#endif
|
1199 |
|
|
|
1200 |
|
|
/* Don't do anything with direct references to volatile memory, but
|
1201 |
|
|
generate code to handle other side effects. */
|
1202 |
|
|
if (!MEM_P (op0) && side_effects_p (op0))
|
1203 |
|
|
emit_insn (op0);
|
1204 |
|
|
}
|
1205 |
|
|
|
1206 |
|
|
/* Get a MEM rtx for expression EXP which is the address of an operand
|
1207 |
|
|
to be used in a string instruction (cmpstrsi, movmemsi, ..). LEN is
|
1208 |
|
|
the maximum length of the block of memory that might be accessed or
|
1209 |
|
|
NULL if unknown. */
|
1210 |
|
|
|
1211 |
|
|
static rtx
|
1212 |
|
|
get_memory_rtx (tree exp, tree len)
|
1213 |
|
|
{
|
1214 |
|
|
tree orig_exp = exp;
|
1215 |
|
|
rtx addr, mem;
|
1216 |
|
|
HOST_WIDE_INT off;
|
1217 |
|
|
|
1218 |
|
|
/* When EXP is not resolved SAVE_EXPR, MEM_ATTRS can be still derived
|
1219 |
|
|
from its expression, for expr->a.b only <variable>.a.b is recorded. */
|
1220 |
|
|
if (TREE_CODE (exp) == SAVE_EXPR && !SAVE_EXPR_RESOLVED_P (exp))
|
1221 |
|
|
exp = TREE_OPERAND (exp, 0);
|
1222 |
|
|
|
1223 |
|
|
addr = expand_expr (orig_exp, NULL_RTX, ptr_mode, EXPAND_NORMAL);
|
1224 |
|
|
mem = gen_rtx_MEM (BLKmode, memory_address (BLKmode, addr));
|
1225 |
|
|
|
1226 |
|
|
/* Get an expression we can use to find the attributes to assign to MEM.
|
1227 |
|
|
If it is an ADDR_EXPR, use the operand. Otherwise, dereference it if
|
1228 |
|
|
we can. First remove any nops. */
|
1229 |
|
|
while (CONVERT_EXPR_P (exp)
|
1230 |
|
|
&& POINTER_TYPE_P (TREE_TYPE (TREE_OPERAND (exp, 0))))
|
1231 |
|
|
exp = TREE_OPERAND (exp, 0);
|
1232 |
|
|
|
1233 |
|
|
off = 0;
|
1234 |
|
|
if (TREE_CODE (exp) == POINTER_PLUS_EXPR
|
1235 |
|
|
&& TREE_CODE (TREE_OPERAND (exp, 0)) == ADDR_EXPR
|
1236 |
|
|
&& host_integerp (TREE_OPERAND (exp, 1), 0)
|
1237 |
|
|
&& (off = tree_low_cst (TREE_OPERAND (exp, 1), 0)) > 0)
|
1238 |
|
|
exp = TREE_OPERAND (TREE_OPERAND (exp, 0), 0);
|
1239 |
|
|
else if (TREE_CODE (exp) == ADDR_EXPR)
|
1240 |
|
|
exp = TREE_OPERAND (exp, 0);
|
1241 |
|
|
else if (POINTER_TYPE_P (TREE_TYPE (exp)))
|
1242 |
|
|
exp = build1 (INDIRECT_REF, TREE_TYPE (TREE_TYPE (exp)), exp);
|
1243 |
|
|
else
|
1244 |
|
|
exp = NULL;
|
1245 |
|
|
|
1246 |
|
|
/* Honor attributes derived from exp, except for the alias set
|
1247 |
|
|
(as builtin stringops may alias with anything) and the size
|
1248 |
|
|
(as stringops may access multiple array elements). */
|
1249 |
|
|
if (exp)
|
1250 |
|
|
{
|
1251 |
|
|
set_mem_attributes (mem, exp, 0);
|
1252 |
|
|
|
1253 |
|
|
if (off)
|
1254 |
|
|
mem = adjust_automodify_address_nv (mem, BLKmode, NULL, off);
|
1255 |
|
|
|
1256 |
|
|
/* Allow the string and memory builtins to overflow from one
|
1257 |
|
|
field into another, see http://gcc.gnu.org/PR23561.
|
1258 |
|
|
Thus avoid COMPONENT_REFs in MEM_EXPR unless we know the whole
|
1259 |
|
|
memory accessed by the string or memory builtin will fit
|
1260 |
|
|
within the field. */
|
1261 |
|
|
if (MEM_EXPR (mem) && TREE_CODE (MEM_EXPR (mem)) == COMPONENT_REF)
|
1262 |
|
|
{
|
1263 |
|
|
tree mem_expr = MEM_EXPR (mem);
|
1264 |
|
|
HOST_WIDE_INT offset = -1, length = -1;
|
1265 |
|
|
tree inner = exp;
|
1266 |
|
|
|
1267 |
|
|
while (TREE_CODE (inner) == ARRAY_REF
|
1268 |
|
|
|| CONVERT_EXPR_P (inner)
|
1269 |
|
|
|| TREE_CODE (inner) == VIEW_CONVERT_EXPR
|
1270 |
|
|
|| TREE_CODE (inner) == SAVE_EXPR)
|
1271 |
|
|
inner = TREE_OPERAND (inner, 0);
|
1272 |
|
|
|
1273 |
|
|
gcc_assert (TREE_CODE (inner) == COMPONENT_REF);
|
1274 |
|
|
|
1275 |
|
|
if (MEM_OFFSET_KNOWN_P (mem))
|
1276 |
|
|
offset = MEM_OFFSET (mem);
|
1277 |
|
|
|
1278 |
|
|
if (offset >= 0 && len && host_integerp (len, 0))
|
1279 |
|
|
length = tree_low_cst (len, 0);
|
1280 |
|
|
|
1281 |
|
|
while (TREE_CODE (inner) == COMPONENT_REF)
|
1282 |
|
|
{
|
1283 |
|
|
tree field = TREE_OPERAND (inner, 1);
|
1284 |
|
|
gcc_assert (TREE_CODE (mem_expr) == COMPONENT_REF);
|
1285 |
|
|
gcc_assert (field == TREE_OPERAND (mem_expr, 1));
|
1286 |
|
|
|
1287 |
|
|
/* Bitfields are generally not byte-addressable. */
|
1288 |
|
|
gcc_assert (!DECL_BIT_FIELD (field)
|
1289 |
|
|
|| ((tree_low_cst (DECL_FIELD_BIT_OFFSET (field), 1)
|
1290 |
|
|
% BITS_PER_UNIT) == 0
|
1291 |
|
|
&& host_integerp (DECL_SIZE (field), 0)
|
1292 |
|
|
&& (TREE_INT_CST_LOW (DECL_SIZE (field))
|
1293 |
|
|
% BITS_PER_UNIT) == 0));
|
1294 |
|
|
|
1295 |
|
|
/* If we can prove that the memory starting at XEXP (mem, 0) and
|
1296 |
|
|
ending at XEXP (mem, 0) + LENGTH will fit into this field, we
|
1297 |
|
|
can keep the COMPONENT_REF in MEM_EXPR. But be careful with
|
1298 |
|
|
fields without DECL_SIZE_UNIT like flexible array members. */
|
1299 |
|
|
if (length >= 0
|
1300 |
|
|
&& DECL_SIZE_UNIT (field)
|
1301 |
|
|
&& host_integerp (DECL_SIZE_UNIT (field), 0))
|
1302 |
|
|
{
|
1303 |
|
|
HOST_WIDE_INT size
|
1304 |
|
|
= TREE_INT_CST_LOW (DECL_SIZE_UNIT (field));
|
1305 |
|
|
if (offset <= size
|
1306 |
|
|
&& length <= size
|
1307 |
|
|
&& offset + length <= size)
|
1308 |
|
|
break;
|
1309 |
|
|
}
|
1310 |
|
|
|
1311 |
|
|
if (offset >= 0
|
1312 |
|
|
&& host_integerp (DECL_FIELD_OFFSET (field), 0))
|
1313 |
|
|
offset += TREE_INT_CST_LOW (DECL_FIELD_OFFSET (field))
|
1314 |
|
|
+ tree_low_cst (DECL_FIELD_BIT_OFFSET (field), 1)
|
1315 |
|
|
/ BITS_PER_UNIT;
|
1316 |
|
|
else
|
1317 |
|
|
{
|
1318 |
|
|
offset = -1;
|
1319 |
|
|
length = -1;
|
1320 |
|
|
}
|
1321 |
|
|
|
1322 |
|
|
mem_expr = TREE_OPERAND (mem_expr, 0);
|
1323 |
|
|
inner = TREE_OPERAND (inner, 0);
|
1324 |
|
|
}
|
1325 |
|
|
|
1326 |
|
|
if (mem_expr == NULL)
|
1327 |
|
|
offset = -1;
|
1328 |
|
|
if (mem_expr != MEM_EXPR (mem))
|
1329 |
|
|
{
|
1330 |
|
|
set_mem_expr (mem, mem_expr);
|
1331 |
|
|
if (offset >= 0)
|
1332 |
|
|
set_mem_offset (mem, offset);
|
1333 |
|
|
else
|
1334 |
|
|
clear_mem_offset (mem);
|
1335 |
|
|
}
|
1336 |
|
|
}
|
1337 |
|
|
set_mem_alias_set (mem, 0);
|
1338 |
|
|
clear_mem_size (mem);
|
1339 |
|
|
}
|
1340 |
|
|
|
1341 |
|
|
return mem;
|
1342 |
|
|
}
|
1343 |
|
|
|
1344 |
|
|
/* Built-in functions to perform an untyped call and return. */
|
1345 |
|
|
|
1346 |
|
|
#define apply_args_mode \
|
1347 |
|
|
(this_target_builtins->x_apply_args_mode)
|
1348 |
|
|
#define apply_result_mode \
|
1349 |
|
|
(this_target_builtins->x_apply_result_mode)
|
1350 |
|
|
|
1351 |
|
|
/* Return the size required for the block returned by __builtin_apply_args,
|
1352 |
|
|
and initialize apply_args_mode. */
|
1353 |
|
|
|
1354 |
|
|
static int
|
1355 |
|
|
apply_args_size (void)
|
1356 |
|
|
{
|
1357 |
|
|
static int size = -1;
|
1358 |
|
|
int align;
|
1359 |
|
|
unsigned int regno;
|
1360 |
|
|
enum machine_mode mode;
|
1361 |
|
|
|
1362 |
|
|
/* The values computed by this function never change. */
|
1363 |
|
|
if (size < 0)
|
1364 |
|
|
{
|
1365 |
|
|
/* The first value is the incoming arg-pointer. */
|
1366 |
|
|
size = GET_MODE_SIZE (Pmode);
|
1367 |
|
|
|
1368 |
|
|
/* The second value is the structure value address unless this is
|
1369 |
|
|
passed as an "invisible" first argument. */
|
1370 |
|
|
if (targetm.calls.struct_value_rtx (cfun ? TREE_TYPE (cfun->decl) : 0, 0))
|
1371 |
|
|
size += GET_MODE_SIZE (Pmode);
|
1372 |
|
|
|
1373 |
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
1374 |
|
|
if (FUNCTION_ARG_REGNO_P (regno))
|
1375 |
|
|
{
|
1376 |
|
|
mode = targetm.calls.get_raw_arg_mode (regno);
|
1377 |
|
|
|
1378 |
|
|
gcc_assert (mode != VOIDmode);
|
1379 |
|
|
|
1380 |
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
1381 |
|
|
if (size % align != 0)
|
1382 |
|
|
size = CEIL (size, align) * align;
|
1383 |
|
|
size += GET_MODE_SIZE (mode);
|
1384 |
|
|
apply_args_mode[regno] = mode;
|
1385 |
|
|
}
|
1386 |
|
|
else
|
1387 |
|
|
{
|
1388 |
|
|
apply_args_mode[regno] = VOIDmode;
|
1389 |
|
|
}
|
1390 |
|
|
}
|
1391 |
|
|
return size;
|
1392 |
|
|
}
|
1393 |
|
|
|
1394 |
|
|
/* Return the size required for the block returned by __builtin_apply,
|
1395 |
|
|
and initialize apply_result_mode. */
|
1396 |
|
|
|
1397 |
|
|
static int
|
1398 |
|
|
apply_result_size (void)
|
1399 |
|
|
{
|
1400 |
|
|
static int size = -1;
|
1401 |
|
|
int align, regno;
|
1402 |
|
|
enum machine_mode mode;
|
1403 |
|
|
|
1404 |
|
|
/* The values computed by this function never change. */
|
1405 |
|
|
if (size < 0)
|
1406 |
|
|
{
|
1407 |
|
|
size = 0;
|
1408 |
|
|
|
1409 |
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
1410 |
|
|
if (targetm.calls.function_value_regno_p (regno))
|
1411 |
|
|
{
|
1412 |
|
|
mode = targetm.calls.get_raw_result_mode (regno);
|
1413 |
|
|
|
1414 |
|
|
gcc_assert (mode != VOIDmode);
|
1415 |
|
|
|
1416 |
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
1417 |
|
|
if (size % align != 0)
|
1418 |
|
|
size = CEIL (size, align) * align;
|
1419 |
|
|
size += GET_MODE_SIZE (mode);
|
1420 |
|
|
apply_result_mode[regno] = mode;
|
1421 |
|
|
}
|
1422 |
|
|
else
|
1423 |
|
|
apply_result_mode[regno] = VOIDmode;
|
1424 |
|
|
|
1425 |
|
|
/* Allow targets that use untyped_call and untyped_return to override
|
1426 |
|
|
the size so that machine-specific information can be stored here. */
|
1427 |
|
|
#ifdef APPLY_RESULT_SIZE
|
1428 |
|
|
size = APPLY_RESULT_SIZE;
|
1429 |
|
|
#endif
|
1430 |
|
|
}
|
1431 |
|
|
return size;
|
1432 |
|
|
}
|
1433 |
|
|
|
1434 |
|
|
#if defined (HAVE_untyped_call) || defined (HAVE_untyped_return)
|
1435 |
|
|
/* Create a vector describing the result block RESULT. If SAVEP is true,
|
1436 |
|
|
the result block is used to save the values; otherwise it is used to
|
1437 |
|
|
restore the values. */
|
1438 |
|
|
|
1439 |
|
|
static rtx
|
1440 |
|
|
result_vector (int savep, rtx result)
|
1441 |
|
|
{
|
1442 |
|
|
int regno, size, align, nelts;
|
1443 |
|
|
enum machine_mode mode;
|
1444 |
|
|
rtx reg, mem;
|
1445 |
|
|
rtx *savevec = XALLOCAVEC (rtx, FIRST_PSEUDO_REGISTER);
|
1446 |
|
|
|
1447 |
|
|
size = nelts = 0;
|
1448 |
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
1449 |
|
|
if ((mode = apply_result_mode[regno]) != VOIDmode)
|
1450 |
|
|
{
|
1451 |
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
1452 |
|
|
if (size % align != 0)
|
1453 |
|
|
size = CEIL (size, align) * align;
|
1454 |
|
|
reg = gen_rtx_REG (mode, savep ? regno : INCOMING_REGNO (regno));
|
1455 |
|
|
mem = adjust_address (result, mode, size);
|
1456 |
|
|
savevec[nelts++] = (savep
|
1457 |
|
|
? gen_rtx_SET (VOIDmode, mem, reg)
|
1458 |
|
|
: gen_rtx_SET (VOIDmode, reg, mem));
|
1459 |
|
|
size += GET_MODE_SIZE (mode);
|
1460 |
|
|
}
|
1461 |
|
|
return gen_rtx_PARALLEL (VOIDmode, gen_rtvec_v (nelts, savevec));
|
1462 |
|
|
}
|
1463 |
|
|
#endif /* HAVE_untyped_call or HAVE_untyped_return */
|
1464 |
|
|
|
1465 |
|
|
/* Save the state required to perform an untyped call with the same
|
1466 |
|
|
arguments as were passed to the current function. */
|
1467 |
|
|
|
1468 |
|
|
static rtx
|
1469 |
|
|
expand_builtin_apply_args_1 (void)
|
1470 |
|
|
{
|
1471 |
|
|
rtx registers, tem;
|
1472 |
|
|
int size, align, regno;
|
1473 |
|
|
enum machine_mode mode;
|
1474 |
|
|
rtx struct_incoming_value = targetm.calls.struct_value_rtx (cfun ? TREE_TYPE (cfun->decl) : 0, 1);
|
1475 |
|
|
|
1476 |
|
|
/* Create a block where the arg-pointer, structure value address,
|
1477 |
|
|
and argument registers can be saved. */
|
1478 |
|
|
registers = assign_stack_local (BLKmode, apply_args_size (), -1);
|
1479 |
|
|
|
1480 |
|
|
/* Walk past the arg-pointer and structure value address. */
|
1481 |
|
|
size = GET_MODE_SIZE (Pmode);
|
1482 |
|
|
if (targetm.calls.struct_value_rtx (cfun ? TREE_TYPE (cfun->decl) : 0, 0))
|
1483 |
|
|
size += GET_MODE_SIZE (Pmode);
|
1484 |
|
|
|
1485 |
|
|
/* Save each register used in calling a function to the block. */
|
1486 |
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
1487 |
|
|
if ((mode = apply_args_mode[regno]) != VOIDmode)
|
1488 |
|
|
{
|
1489 |
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
1490 |
|
|
if (size % align != 0)
|
1491 |
|
|
size = CEIL (size, align) * align;
|
1492 |
|
|
|
1493 |
|
|
tem = gen_rtx_REG (mode, INCOMING_REGNO (regno));
|
1494 |
|
|
|
1495 |
|
|
emit_move_insn (adjust_address (registers, mode, size), tem);
|
1496 |
|
|
size += GET_MODE_SIZE (mode);
|
1497 |
|
|
}
|
1498 |
|
|
|
1499 |
|
|
/* Save the arg pointer to the block. */
|
1500 |
|
|
tem = copy_to_reg (crtl->args.internal_arg_pointer);
|
1501 |
|
|
#ifdef STACK_GROWS_DOWNWARD
|
1502 |
|
|
/* We need the pointer as the caller actually passed them to us, not
|
1503 |
|
|
as we might have pretended they were passed. Make sure it's a valid
|
1504 |
|
|
operand, as emit_move_insn isn't expected to handle a PLUS. */
|
1505 |
|
|
tem
|
1506 |
|
|
= force_operand (plus_constant (tem, crtl->args.pretend_args_size),
|
1507 |
|
|
NULL_RTX);
|
1508 |
|
|
#endif
|
1509 |
|
|
emit_move_insn (adjust_address (registers, Pmode, 0), tem);
|
1510 |
|
|
|
1511 |
|
|
size = GET_MODE_SIZE (Pmode);
|
1512 |
|
|
|
1513 |
|
|
/* Save the structure value address unless this is passed as an
|
1514 |
|
|
"invisible" first argument. */
|
1515 |
|
|
if (struct_incoming_value)
|
1516 |
|
|
{
|
1517 |
|
|
emit_move_insn (adjust_address (registers, Pmode, size),
|
1518 |
|
|
copy_to_reg (struct_incoming_value));
|
1519 |
|
|
size += GET_MODE_SIZE (Pmode);
|
1520 |
|
|
}
|
1521 |
|
|
|
1522 |
|
|
/* Return the address of the block. */
|
1523 |
|
|
return copy_addr_to_reg (XEXP (registers, 0));
|
1524 |
|
|
}
|
1525 |
|
|
|
1526 |
|
|
/* __builtin_apply_args returns block of memory allocated on
|
1527 |
|
|
the stack into which is stored the arg pointer, structure
|
1528 |
|
|
value address, static chain, and all the registers that might
|
1529 |
|
|
possibly be used in performing a function call. The code is
|
1530 |
|
|
moved to the start of the function so the incoming values are
|
1531 |
|
|
saved. */
|
1532 |
|
|
|
1533 |
|
|
static rtx
|
1534 |
|
|
expand_builtin_apply_args (void)
|
1535 |
|
|
{
|
1536 |
|
|
/* Don't do __builtin_apply_args more than once in a function.
|
1537 |
|
|
Save the result of the first call and reuse it. */
|
1538 |
|
|
if (apply_args_value != 0)
|
1539 |
|
|
return apply_args_value;
|
1540 |
|
|
{
|
1541 |
|
|
/* When this function is called, it means that registers must be
|
1542 |
|
|
saved on entry to this function. So we migrate the
|
1543 |
|
|
call to the first insn of this function. */
|
1544 |
|
|
rtx temp;
|
1545 |
|
|
rtx seq;
|
1546 |
|
|
|
1547 |
|
|
start_sequence ();
|
1548 |
|
|
temp = expand_builtin_apply_args_1 ();
|
1549 |
|
|
seq = get_insns ();
|
1550 |
|
|
end_sequence ();
|
1551 |
|
|
|
1552 |
|
|
apply_args_value = temp;
|
1553 |
|
|
|
1554 |
|
|
/* Put the insns after the NOTE that starts the function.
|
1555 |
|
|
If this is inside a start_sequence, make the outer-level insn
|
1556 |
|
|
chain current, so the code is placed at the start of the
|
1557 |
|
|
function. If internal_arg_pointer is a non-virtual pseudo,
|
1558 |
|
|
it needs to be placed after the function that initializes
|
1559 |
|
|
that pseudo. */
|
1560 |
|
|
push_topmost_sequence ();
|
1561 |
|
|
if (REG_P (crtl->args.internal_arg_pointer)
|
1562 |
|
|
&& REGNO (crtl->args.internal_arg_pointer) > LAST_VIRTUAL_REGISTER)
|
1563 |
|
|
emit_insn_before (seq, parm_birth_insn);
|
1564 |
|
|
else
|
1565 |
|
|
emit_insn_before (seq, NEXT_INSN (entry_of_function ()));
|
1566 |
|
|
pop_topmost_sequence ();
|
1567 |
|
|
return temp;
|
1568 |
|
|
}
|
1569 |
|
|
}
|
1570 |
|
|
|
1571 |
|
|
/* Perform an untyped call and save the state required to perform an
|
1572 |
|
|
untyped return of whatever value was returned by the given function. */
|
1573 |
|
|
|
1574 |
|
|
static rtx
|
1575 |
|
|
expand_builtin_apply (rtx function, rtx arguments, rtx argsize)
|
1576 |
|
|
{
|
1577 |
|
|
int size, align, regno;
|
1578 |
|
|
enum machine_mode mode;
|
1579 |
|
|
rtx incoming_args, result, reg, dest, src, call_insn;
|
1580 |
|
|
rtx old_stack_level = 0;
|
1581 |
|
|
rtx call_fusage = 0;
|
1582 |
|
|
rtx struct_value = targetm.calls.struct_value_rtx (cfun ? TREE_TYPE (cfun->decl) : 0, 0);
|
1583 |
|
|
|
1584 |
|
|
arguments = convert_memory_address (Pmode, arguments);
|
1585 |
|
|
|
1586 |
|
|
/* Create a block where the return registers can be saved. */
|
1587 |
|
|
result = assign_stack_local (BLKmode, apply_result_size (), -1);
|
1588 |
|
|
|
1589 |
|
|
/* Fetch the arg pointer from the ARGUMENTS block. */
|
1590 |
|
|
incoming_args = gen_reg_rtx (Pmode);
|
1591 |
|
|
emit_move_insn (incoming_args, gen_rtx_MEM (Pmode, arguments));
|
1592 |
|
|
#ifndef STACK_GROWS_DOWNWARD
|
1593 |
|
|
incoming_args = expand_simple_binop (Pmode, MINUS, incoming_args, argsize,
|
1594 |
|
|
incoming_args, 0, OPTAB_LIB_WIDEN);
|
1595 |
|
|
#endif
|
1596 |
|
|
|
1597 |
|
|
/* Push a new argument block and copy the arguments. Do not allow
|
1598 |
|
|
the (potential) memcpy call below to interfere with our stack
|
1599 |
|
|
manipulations. */
|
1600 |
|
|
do_pending_stack_adjust ();
|
1601 |
|
|
NO_DEFER_POP;
|
1602 |
|
|
|
1603 |
|
|
/* Save the stack with nonlocal if available. */
|
1604 |
|
|
#ifdef HAVE_save_stack_nonlocal
|
1605 |
|
|
if (HAVE_save_stack_nonlocal)
|
1606 |
|
|
emit_stack_save (SAVE_NONLOCAL, &old_stack_level);
|
1607 |
|
|
else
|
1608 |
|
|
#endif
|
1609 |
|
|
emit_stack_save (SAVE_BLOCK, &old_stack_level);
|
1610 |
|
|
|
1611 |
|
|
/* Allocate a block of memory onto the stack and copy the memory
|
1612 |
|
|
arguments to the outgoing arguments address. We can pass TRUE
|
1613 |
|
|
as the 4th argument because we just saved the stack pointer
|
1614 |
|
|
and will restore it right after the call. */
|
1615 |
|
|
allocate_dynamic_stack_space (argsize, 0, BIGGEST_ALIGNMENT, true);
|
1616 |
|
|
|
1617 |
|
|
/* Set DRAP flag to true, even though allocate_dynamic_stack_space
|
1618 |
|
|
may have already set current_function_calls_alloca to true.
|
1619 |
|
|
current_function_calls_alloca won't be set if argsize is zero,
|
1620 |
|
|
so we have to guarantee need_drap is true here. */
|
1621 |
|
|
if (SUPPORTS_STACK_ALIGNMENT)
|
1622 |
|
|
crtl->need_drap = true;
|
1623 |
|
|
|
1624 |
|
|
dest = virtual_outgoing_args_rtx;
|
1625 |
|
|
#ifndef STACK_GROWS_DOWNWARD
|
1626 |
|
|
if (CONST_INT_P (argsize))
|
1627 |
|
|
dest = plus_constant (dest, -INTVAL (argsize));
|
1628 |
|
|
else
|
1629 |
|
|
dest = gen_rtx_PLUS (Pmode, dest, negate_rtx (Pmode, argsize));
|
1630 |
|
|
#endif
|
1631 |
|
|
dest = gen_rtx_MEM (BLKmode, dest);
|
1632 |
|
|
set_mem_align (dest, PARM_BOUNDARY);
|
1633 |
|
|
src = gen_rtx_MEM (BLKmode, incoming_args);
|
1634 |
|
|
set_mem_align (src, PARM_BOUNDARY);
|
1635 |
|
|
emit_block_move (dest, src, argsize, BLOCK_OP_NORMAL);
|
1636 |
|
|
|
1637 |
|
|
/* Refer to the argument block. */
|
1638 |
|
|
apply_args_size ();
|
1639 |
|
|
arguments = gen_rtx_MEM (BLKmode, arguments);
|
1640 |
|
|
set_mem_align (arguments, PARM_BOUNDARY);
|
1641 |
|
|
|
1642 |
|
|
/* Walk past the arg-pointer and structure value address. */
|
1643 |
|
|
size = GET_MODE_SIZE (Pmode);
|
1644 |
|
|
if (struct_value)
|
1645 |
|
|
size += GET_MODE_SIZE (Pmode);
|
1646 |
|
|
|
1647 |
|
|
/* Restore each of the registers previously saved. Make USE insns
|
1648 |
|
|
for each of these registers for use in making the call. */
|
1649 |
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
1650 |
|
|
if ((mode = apply_args_mode[regno]) != VOIDmode)
|
1651 |
|
|
{
|
1652 |
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
1653 |
|
|
if (size % align != 0)
|
1654 |
|
|
size = CEIL (size, align) * align;
|
1655 |
|
|
reg = gen_rtx_REG (mode, regno);
|
1656 |
|
|
emit_move_insn (reg, adjust_address (arguments, mode, size));
|
1657 |
|
|
use_reg (&call_fusage, reg);
|
1658 |
|
|
size += GET_MODE_SIZE (mode);
|
1659 |
|
|
}
|
1660 |
|
|
|
1661 |
|
|
/* Restore the structure value address unless this is passed as an
|
1662 |
|
|
"invisible" first argument. */
|
1663 |
|
|
size = GET_MODE_SIZE (Pmode);
|
1664 |
|
|
if (struct_value)
|
1665 |
|
|
{
|
1666 |
|
|
rtx value = gen_reg_rtx (Pmode);
|
1667 |
|
|
emit_move_insn (value, adjust_address (arguments, Pmode, size));
|
1668 |
|
|
emit_move_insn (struct_value, value);
|
1669 |
|
|
if (REG_P (struct_value))
|
1670 |
|
|
use_reg (&call_fusage, struct_value);
|
1671 |
|
|
size += GET_MODE_SIZE (Pmode);
|
1672 |
|
|
}
|
1673 |
|
|
|
1674 |
|
|
/* All arguments and registers used for the call are set up by now! */
|
1675 |
|
|
function = prepare_call_address (NULL, function, NULL, &call_fusage, 0, 0);
|
1676 |
|
|
|
1677 |
|
|
/* Ensure address is valid. SYMBOL_REF is already valid, so no need,
|
1678 |
|
|
and we don't want to load it into a register as an optimization,
|
1679 |
|
|
because prepare_call_address already did it if it should be done. */
|
1680 |
|
|
if (GET_CODE (function) != SYMBOL_REF)
|
1681 |
|
|
function = memory_address (FUNCTION_MODE, function);
|
1682 |
|
|
|
1683 |
|
|
/* Generate the actual call instruction and save the return value. */
|
1684 |
|
|
#ifdef HAVE_untyped_call
|
1685 |
|
|
if (HAVE_untyped_call)
|
1686 |
|
|
emit_call_insn (gen_untyped_call (gen_rtx_MEM (FUNCTION_MODE, function),
|
1687 |
|
|
result, result_vector (1, result)));
|
1688 |
|
|
else
|
1689 |
|
|
#endif
|
1690 |
|
|
#ifdef HAVE_call_value
|
1691 |
|
|
if (HAVE_call_value)
|
1692 |
|
|
{
|
1693 |
|
|
rtx valreg = 0;
|
1694 |
|
|
|
1695 |
|
|
/* Locate the unique return register. It is not possible to
|
1696 |
|
|
express a call that sets more than one return register using
|
1697 |
|
|
call_value; use untyped_call for that. In fact, untyped_call
|
1698 |
|
|
only needs to save the return registers in the given block. */
|
1699 |
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
1700 |
|
|
if ((mode = apply_result_mode[regno]) != VOIDmode)
|
1701 |
|
|
{
|
1702 |
|
|
gcc_assert (!valreg); /* HAVE_untyped_call required. */
|
1703 |
|
|
|
1704 |
|
|
valreg = gen_rtx_REG (mode, regno);
|
1705 |
|
|
}
|
1706 |
|
|
|
1707 |
|
|
emit_call_insn (GEN_CALL_VALUE (valreg,
|
1708 |
|
|
gen_rtx_MEM (FUNCTION_MODE, function),
|
1709 |
|
|
const0_rtx, NULL_RTX, const0_rtx));
|
1710 |
|
|
|
1711 |
|
|
emit_move_insn (adjust_address (result, GET_MODE (valreg), 0), valreg);
|
1712 |
|
|
}
|
1713 |
|
|
else
|
1714 |
|
|
#endif
|
1715 |
|
|
gcc_unreachable ();
|
1716 |
|
|
|
1717 |
|
|
/* Find the CALL insn we just emitted, and attach the register usage
|
1718 |
|
|
information. */
|
1719 |
|
|
call_insn = last_call_insn ();
|
1720 |
|
|
add_function_usage_to (call_insn, call_fusage);
|
1721 |
|
|
|
1722 |
|
|
/* Restore the stack. */
|
1723 |
|
|
#ifdef HAVE_save_stack_nonlocal
|
1724 |
|
|
if (HAVE_save_stack_nonlocal)
|
1725 |
|
|
emit_stack_restore (SAVE_NONLOCAL, old_stack_level);
|
1726 |
|
|
else
|
1727 |
|
|
#endif
|
1728 |
|
|
emit_stack_restore (SAVE_BLOCK, old_stack_level);
|
1729 |
|
|
fixup_args_size_notes (call_insn, get_last_insn(), 0);
|
1730 |
|
|
|
1731 |
|
|
OK_DEFER_POP;
|
1732 |
|
|
|
1733 |
|
|
/* Return the address of the result block. */
|
1734 |
|
|
result = copy_addr_to_reg (XEXP (result, 0));
|
1735 |
|
|
return convert_memory_address (ptr_mode, result);
|
1736 |
|
|
}
|
1737 |
|
|
|
1738 |
|
|
/* Perform an untyped return. */
|
1739 |
|
|
|
1740 |
|
|
static void
|
1741 |
|
|
expand_builtin_return (rtx result)
|
1742 |
|
|
{
|
1743 |
|
|
int size, align, regno;
|
1744 |
|
|
enum machine_mode mode;
|
1745 |
|
|
rtx reg;
|
1746 |
|
|
rtx call_fusage = 0;
|
1747 |
|
|
|
1748 |
|
|
result = convert_memory_address (Pmode, result);
|
1749 |
|
|
|
1750 |
|
|
apply_result_size ();
|
1751 |
|
|
result = gen_rtx_MEM (BLKmode, result);
|
1752 |
|
|
|
1753 |
|
|
#ifdef HAVE_untyped_return
|
1754 |
|
|
if (HAVE_untyped_return)
|
1755 |
|
|
{
|
1756 |
|
|
emit_jump_insn (gen_untyped_return (result, result_vector (0, result)));
|
1757 |
|
|
emit_barrier ();
|
1758 |
|
|
return;
|
1759 |
|
|
}
|
1760 |
|
|
#endif
|
1761 |
|
|
|
1762 |
|
|
/* Restore the return value and note that each value is used. */
|
1763 |
|
|
size = 0;
|
1764 |
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
1765 |
|
|
if ((mode = apply_result_mode[regno]) != VOIDmode)
|
1766 |
|
|
{
|
1767 |
|
|
align = GET_MODE_ALIGNMENT (mode) / BITS_PER_UNIT;
|
1768 |
|
|
if (size % align != 0)
|
1769 |
|
|
size = CEIL (size, align) * align;
|
1770 |
|
|
reg = gen_rtx_REG (mode, INCOMING_REGNO (regno));
|
1771 |
|
|
emit_move_insn (reg, adjust_address (result, mode, size));
|
1772 |
|
|
|
1773 |
|
|
push_to_sequence (call_fusage);
|
1774 |
|
|
emit_use (reg);
|
1775 |
|
|
call_fusage = get_insns ();
|
1776 |
|
|
end_sequence ();
|
1777 |
|
|
size += GET_MODE_SIZE (mode);
|
1778 |
|
|
}
|
1779 |
|
|
|
1780 |
|
|
/* Put the USE insns before the return. */
|
1781 |
|
|
emit_insn (call_fusage);
|
1782 |
|
|
|
1783 |
|
|
/* Return whatever values was restored by jumping directly to the end
|
1784 |
|
|
of the function. */
|
1785 |
|
|
expand_naked_return ();
|
1786 |
|
|
}
|
1787 |
|
|
|
1788 |
|
|
/* Used by expand_builtin_classify_type and fold_builtin_classify_type. */
|
1789 |
|
|
|
1790 |
|
|
static enum type_class
|
1791 |
|
|
type_to_class (tree type)
|
1792 |
|
|
{
|
1793 |
|
|
switch (TREE_CODE (type))
|
1794 |
|
|
{
|
1795 |
|
|
case VOID_TYPE: return void_type_class;
|
1796 |
|
|
case INTEGER_TYPE: return integer_type_class;
|
1797 |
|
|
case ENUMERAL_TYPE: return enumeral_type_class;
|
1798 |
|
|
case BOOLEAN_TYPE: return boolean_type_class;
|
1799 |
|
|
case POINTER_TYPE: return pointer_type_class;
|
1800 |
|
|
case REFERENCE_TYPE: return reference_type_class;
|
1801 |
|
|
case OFFSET_TYPE: return offset_type_class;
|
1802 |
|
|
case REAL_TYPE: return real_type_class;
|
1803 |
|
|
case COMPLEX_TYPE: return complex_type_class;
|
1804 |
|
|
case FUNCTION_TYPE: return function_type_class;
|
1805 |
|
|
case METHOD_TYPE: return method_type_class;
|
1806 |
|
|
case RECORD_TYPE: return record_type_class;
|
1807 |
|
|
case UNION_TYPE:
|
1808 |
|
|
case QUAL_UNION_TYPE: return union_type_class;
|
1809 |
|
|
case ARRAY_TYPE: return (TYPE_STRING_FLAG (type)
|
1810 |
|
|
? string_type_class : array_type_class);
|
1811 |
|
|
case LANG_TYPE: return lang_type_class;
|
1812 |
|
|
default: return no_type_class;
|
1813 |
|
|
}
|
1814 |
|
|
}
|
1815 |
|
|
|
1816 |
|
|
/* Expand a call EXP to __builtin_classify_type. */
|
1817 |
|
|
|
1818 |
|
|
static rtx
|
1819 |
|
|
expand_builtin_classify_type (tree exp)
|
1820 |
|
|
{
|
1821 |
|
|
if (call_expr_nargs (exp))
|
1822 |
|
|
return GEN_INT (type_to_class (TREE_TYPE (CALL_EXPR_ARG (exp, 0))));
|
1823 |
|
|
return GEN_INT (no_type_class);
|
1824 |
|
|
}
|
1825 |
|
|
|
1826 |
|
|
/* This helper macro, meant to be used in mathfn_built_in below,
|
1827 |
|
|
determines which among a set of three builtin math functions is
|
1828 |
|
|
appropriate for a given type mode. The `F' and `L' cases are
|
1829 |
|
|
automatically generated from the `double' case. */
|
1830 |
|
|
#define CASE_MATHFN(BUILT_IN_MATHFN) \
|
1831 |
|
|
case BUILT_IN_MATHFN: case BUILT_IN_MATHFN##F: case BUILT_IN_MATHFN##L: \
|
1832 |
|
|
fcode = BUILT_IN_MATHFN; fcodef = BUILT_IN_MATHFN##F ; \
|
1833 |
|
|
fcodel = BUILT_IN_MATHFN##L ; break;
|
1834 |
|
|
/* Similar to above, but appends _R after any F/L suffix. */
|
1835 |
|
|
#define CASE_MATHFN_REENT(BUILT_IN_MATHFN) \
|
1836 |
|
|
case BUILT_IN_MATHFN##_R: case BUILT_IN_MATHFN##F_R: case BUILT_IN_MATHFN##L_R: \
|
1837 |
|
|
fcode = BUILT_IN_MATHFN##_R; fcodef = BUILT_IN_MATHFN##F_R ; \
|
1838 |
|
|
fcodel = BUILT_IN_MATHFN##L_R ; break;
|
1839 |
|
|
|
1840 |
|
|
/* Return mathematic function equivalent to FN but operating directly on TYPE,
|
1841 |
|
|
if available. If IMPLICIT is true use the implicit builtin declaration,
|
1842 |
|
|
otherwise use the explicit declaration. If we can't do the conversion,
|
1843 |
|
|
return zero. */
|
1844 |
|
|
|
1845 |
|
|
static tree
|
1846 |
|
|
mathfn_built_in_1 (tree type, enum built_in_function fn, bool implicit_p)
|
1847 |
|
|
{
|
1848 |
|
|
enum built_in_function fcode, fcodef, fcodel, fcode2;
|
1849 |
|
|
|
1850 |
|
|
switch (fn)
|
1851 |
|
|
{
|
1852 |
|
|
CASE_MATHFN (BUILT_IN_ACOS)
|
1853 |
|
|
CASE_MATHFN (BUILT_IN_ACOSH)
|
1854 |
|
|
CASE_MATHFN (BUILT_IN_ASIN)
|
1855 |
|
|
CASE_MATHFN (BUILT_IN_ASINH)
|
1856 |
|
|
CASE_MATHFN (BUILT_IN_ATAN)
|
1857 |
|
|
CASE_MATHFN (BUILT_IN_ATAN2)
|
1858 |
|
|
CASE_MATHFN (BUILT_IN_ATANH)
|
1859 |
|
|
CASE_MATHFN (BUILT_IN_CBRT)
|
1860 |
|
|
CASE_MATHFN (BUILT_IN_CEIL)
|
1861 |
|
|
CASE_MATHFN (BUILT_IN_CEXPI)
|
1862 |
|
|
CASE_MATHFN (BUILT_IN_COPYSIGN)
|
1863 |
|
|
CASE_MATHFN (BUILT_IN_COS)
|
1864 |
|
|
CASE_MATHFN (BUILT_IN_COSH)
|
1865 |
|
|
CASE_MATHFN (BUILT_IN_DREM)
|
1866 |
|
|
CASE_MATHFN (BUILT_IN_ERF)
|
1867 |
|
|
CASE_MATHFN (BUILT_IN_ERFC)
|
1868 |
|
|
CASE_MATHFN (BUILT_IN_EXP)
|
1869 |
|
|
CASE_MATHFN (BUILT_IN_EXP10)
|
1870 |
|
|
CASE_MATHFN (BUILT_IN_EXP2)
|
1871 |
|
|
CASE_MATHFN (BUILT_IN_EXPM1)
|
1872 |
|
|
CASE_MATHFN (BUILT_IN_FABS)
|
1873 |
|
|
CASE_MATHFN (BUILT_IN_FDIM)
|
1874 |
|
|
CASE_MATHFN (BUILT_IN_FLOOR)
|
1875 |
|
|
CASE_MATHFN (BUILT_IN_FMA)
|
1876 |
|
|
CASE_MATHFN (BUILT_IN_FMAX)
|
1877 |
|
|
CASE_MATHFN (BUILT_IN_FMIN)
|
1878 |
|
|
CASE_MATHFN (BUILT_IN_FMOD)
|
1879 |
|
|
CASE_MATHFN (BUILT_IN_FREXP)
|
1880 |
|
|
CASE_MATHFN (BUILT_IN_GAMMA)
|
1881 |
|
|
CASE_MATHFN_REENT (BUILT_IN_GAMMA) /* GAMMA_R */
|
1882 |
|
|
CASE_MATHFN (BUILT_IN_HUGE_VAL)
|
1883 |
|
|
CASE_MATHFN (BUILT_IN_HYPOT)
|
1884 |
|
|
CASE_MATHFN (BUILT_IN_ILOGB)
|
1885 |
|
|
CASE_MATHFN (BUILT_IN_ICEIL)
|
1886 |
|
|
CASE_MATHFN (BUILT_IN_IFLOOR)
|
1887 |
|
|
CASE_MATHFN (BUILT_IN_INF)
|
1888 |
|
|
CASE_MATHFN (BUILT_IN_IRINT)
|
1889 |
|
|
CASE_MATHFN (BUILT_IN_IROUND)
|
1890 |
|
|
CASE_MATHFN (BUILT_IN_ISINF)
|
1891 |
|
|
CASE_MATHFN (BUILT_IN_J0)
|
1892 |
|
|
CASE_MATHFN (BUILT_IN_J1)
|
1893 |
|
|
CASE_MATHFN (BUILT_IN_JN)
|
1894 |
|
|
CASE_MATHFN (BUILT_IN_LCEIL)
|
1895 |
|
|
CASE_MATHFN (BUILT_IN_LDEXP)
|
1896 |
|
|
CASE_MATHFN (BUILT_IN_LFLOOR)
|
1897 |
|
|
CASE_MATHFN (BUILT_IN_LGAMMA)
|
1898 |
|
|
CASE_MATHFN_REENT (BUILT_IN_LGAMMA) /* LGAMMA_R */
|
1899 |
|
|
CASE_MATHFN (BUILT_IN_LLCEIL)
|
1900 |
|
|
CASE_MATHFN (BUILT_IN_LLFLOOR)
|
1901 |
|
|
CASE_MATHFN (BUILT_IN_LLRINT)
|
1902 |
|
|
CASE_MATHFN (BUILT_IN_LLROUND)
|
1903 |
|
|
CASE_MATHFN (BUILT_IN_LOG)
|
1904 |
|
|
CASE_MATHFN (BUILT_IN_LOG10)
|
1905 |
|
|
CASE_MATHFN (BUILT_IN_LOG1P)
|
1906 |
|
|
CASE_MATHFN (BUILT_IN_LOG2)
|
1907 |
|
|
CASE_MATHFN (BUILT_IN_LOGB)
|
1908 |
|
|
CASE_MATHFN (BUILT_IN_LRINT)
|
1909 |
|
|
CASE_MATHFN (BUILT_IN_LROUND)
|
1910 |
|
|
CASE_MATHFN (BUILT_IN_MODF)
|
1911 |
|
|
CASE_MATHFN (BUILT_IN_NAN)
|
1912 |
|
|
CASE_MATHFN (BUILT_IN_NANS)
|
1913 |
|
|
CASE_MATHFN (BUILT_IN_NEARBYINT)
|
1914 |
|
|
CASE_MATHFN (BUILT_IN_NEXTAFTER)
|
1915 |
|
|
CASE_MATHFN (BUILT_IN_NEXTTOWARD)
|
1916 |
|
|
CASE_MATHFN (BUILT_IN_POW)
|
1917 |
|
|
CASE_MATHFN (BUILT_IN_POWI)
|
1918 |
|
|
CASE_MATHFN (BUILT_IN_POW10)
|
1919 |
|
|
CASE_MATHFN (BUILT_IN_REMAINDER)
|
1920 |
|
|
CASE_MATHFN (BUILT_IN_REMQUO)
|
1921 |
|
|
CASE_MATHFN (BUILT_IN_RINT)
|
1922 |
|
|
CASE_MATHFN (BUILT_IN_ROUND)
|
1923 |
|
|
CASE_MATHFN (BUILT_IN_SCALB)
|
1924 |
|
|
CASE_MATHFN (BUILT_IN_SCALBLN)
|
1925 |
|
|
CASE_MATHFN (BUILT_IN_SCALBN)
|
1926 |
|
|
CASE_MATHFN (BUILT_IN_SIGNBIT)
|
1927 |
|
|
CASE_MATHFN (BUILT_IN_SIGNIFICAND)
|
1928 |
|
|
CASE_MATHFN (BUILT_IN_SIN)
|
1929 |
|
|
CASE_MATHFN (BUILT_IN_SINCOS)
|
1930 |
|
|
CASE_MATHFN (BUILT_IN_SINH)
|
1931 |
|
|
CASE_MATHFN (BUILT_IN_SQRT)
|
1932 |
|
|
CASE_MATHFN (BUILT_IN_TAN)
|
1933 |
|
|
CASE_MATHFN (BUILT_IN_TANH)
|
1934 |
|
|
CASE_MATHFN (BUILT_IN_TGAMMA)
|
1935 |
|
|
CASE_MATHFN (BUILT_IN_TRUNC)
|
1936 |
|
|
CASE_MATHFN (BUILT_IN_Y0)
|
1937 |
|
|
CASE_MATHFN (BUILT_IN_Y1)
|
1938 |
|
|
CASE_MATHFN (BUILT_IN_YN)
|
1939 |
|
|
|
1940 |
|
|
default:
|
1941 |
|
|
return NULL_TREE;
|
1942 |
|
|
}
|
1943 |
|
|
|
1944 |
|
|
if (TYPE_MAIN_VARIANT (type) == double_type_node)
|
1945 |
|
|
fcode2 = fcode;
|
1946 |
|
|
else if (TYPE_MAIN_VARIANT (type) == float_type_node)
|
1947 |
|
|
fcode2 = fcodef;
|
1948 |
|
|
else if (TYPE_MAIN_VARIANT (type) == long_double_type_node)
|
1949 |
|
|
fcode2 = fcodel;
|
1950 |
|
|
else
|
1951 |
|
|
return NULL_TREE;
|
1952 |
|
|
|
1953 |
|
|
if (implicit_p && !builtin_decl_implicit_p (fcode2))
|
1954 |
|
|
return NULL_TREE;
|
1955 |
|
|
|
1956 |
|
|
return builtin_decl_explicit (fcode2);
|
1957 |
|
|
}
|
1958 |
|
|
|
1959 |
|
|
/* Like mathfn_built_in_1(), but always use the implicit array. */
|
1960 |
|
|
|
1961 |
|
|
tree
|
1962 |
|
|
mathfn_built_in (tree type, enum built_in_function fn)
|
1963 |
|
|
{
|
1964 |
|
|
return mathfn_built_in_1 (type, fn, /*implicit=*/ 1);
|
1965 |
|
|
}
|
1966 |
|
|
|
1967 |
|
|
/* If errno must be maintained, expand the RTL to check if the result,
|
1968 |
|
|
TARGET, of a built-in function call, EXP, is NaN, and if so set
|
1969 |
|
|
errno to EDOM. */
|
1970 |
|
|
|
1971 |
|
|
static void
|
1972 |
|
|
expand_errno_check (tree exp, rtx target)
|
1973 |
|
|
{
|
1974 |
|
|
rtx lab = gen_label_rtx ();
|
1975 |
|
|
|
1976 |
|
|
/* Test the result; if it is NaN, set errno=EDOM because
|
1977 |
|
|
the argument was not in the domain. */
|
1978 |
|
|
do_compare_rtx_and_jump (target, target, EQ, 0, GET_MODE (target),
|
1979 |
|
|
NULL_RTX, NULL_RTX, lab,
|
1980 |
|
|
/* The jump is very likely. */
|
1981 |
|
|
REG_BR_PROB_BASE - (REG_BR_PROB_BASE / 2000 - 1));
|
1982 |
|
|
|
1983 |
|
|
#ifdef TARGET_EDOM
|
1984 |
|
|
/* If this built-in doesn't throw an exception, set errno directly. */
|
1985 |
|
|
if (TREE_NOTHROW (TREE_OPERAND (CALL_EXPR_FN (exp), 0)))
|
1986 |
|
|
{
|
1987 |
|
|
#ifdef GEN_ERRNO_RTX
|
1988 |
|
|
rtx errno_rtx = GEN_ERRNO_RTX;
|
1989 |
|
|
#else
|
1990 |
|
|
rtx errno_rtx
|
1991 |
|
|
= gen_rtx_MEM (word_mode, gen_rtx_SYMBOL_REF (Pmode, "errno"));
|
1992 |
|
|
#endif
|
1993 |
|
|
emit_move_insn (errno_rtx, GEN_INT (TARGET_EDOM));
|
1994 |
|
|
emit_label (lab);
|
1995 |
|
|
return;
|
1996 |
|
|
}
|
1997 |
|
|
#endif
|
1998 |
|
|
|
1999 |
|
|
/* Make sure the library call isn't expanded as a tail call. */
|
2000 |
|
|
CALL_EXPR_TAILCALL (exp) = 0;
|
2001 |
|
|
|
2002 |
|
|
/* We can't set errno=EDOM directly; let the library call do it.
|
2003 |
|
|
Pop the arguments right away in case the call gets deleted. */
|
2004 |
|
|
NO_DEFER_POP;
|
2005 |
|
|
expand_call (exp, target, 0);
|
2006 |
|
|
OK_DEFER_POP;
|
2007 |
|
|
emit_label (lab);
|
2008 |
|
|
}
|
2009 |
|
|
|
2010 |
|
|
/* Expand a call to one of the builtin math functions (sqrt, exp, or log).
|
2011 |
|
|
Return NULL_RTX if a normal call should be emitted rather than expanding
|
2012 |
|
|
the function in-line. EXP is the expression that is a call to the builtin
|
2013 |
|
|
function; if convenient, the result should be placed in TARGET.
|
2014 |
|
|
SUBTARGET may be used as the target for computing one of EXP's operands. */
|
2015 |
|
|
|
2016 |
|
|
static rtx
|
2017 |
|
|
expand_builtin_mathfn (tree exp, rtx target, rtx subtarget)
|
2018 |
|
|
{
|
2019 |
|
|
optab builtin_optab;
|
2020 |
|
|
rtx op0, insns;
|
2021 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
2022 |
|
|
enum machine_mode mode;
|
2023 |
|
|
bool errno_set = false;
|
2024 |
|
|
tree arg;
|
2025 |
|
|
|
2026 |
|
|
if (!validate_arglist (exp, REAL_TYPE, VOID_TYPE))
|
2027 |
|
|
return NULL_RTX;
|
2028 |
|
|
|
2029 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
2030 |
|
|
|
2031 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
2032 |
|
|
{
|
2033 |
|
|
CASE_FLT_FN (BUILT_IN_SQRT):
|
2034 |
|
|
errno_set = ! tree_expr_nonnegative_p (arg);
|
2035 |
|
|
builtin_optab = sqrt_optab;
|
2036 |
|
|
break;
|
2037 |
|
|
CASE_FLT_FN (BUILT_IN_EXP):
|
2038 |
|
|
errno_set = true; builtin_optab = exp_optab; break;
|
2039 |
|
|
CASE_FLT_FN (BUILT_IN_EXP10):
|
2040 |
|
|
CASE_FLT_FN (BUILT_IN_POW10):
|
2041 |
|
|
errno_set = true; builtin_optab = exp10_optab; break;
|
2042 |
|
|
CASE_FLT_FN (BUILT_IN_EXP2):
|
2043 |
|
|
errno_set = true; builtin_optab = exp2_optab; break;
|
2044 |
|
|
CASE_FLT_FN (BUILT_IN_EXPM1):
|
2045 |
|
|
errno_set = true; builtin_optab = expm1_optab; break;
|
2046 |
|
|
CASE_FLT_FN (BUILT_IN_LOGB):
|
2047 |
|
|
errno_set = true; builtin_optab = logb_optab; break;
|
2048 |
|
|
CASE_FLT_FN (BUILT_IN_LOG):
|
2049 |
|
|
errno_set = true; builtin_optab = log_optab; break;
|
2050 |
|
|
CASE_FLT_FN (BUILT_IN_LOG10):
|
2051 |
|
|
errno_set = true; builtin_optab = log10_optab; break;
|
2052 |
|
|
CASE_FLT_FN (BUILT_IN_LOG2):
|
2053 |
|
|
errno_set = true; builtin_optab = log2_optab; break;
|
2054 |
|
|
CASE_FLT_FN (BUILT_IN_LOG1P):
|
2055 |
|
|
errno_set = true; builtin_optab = log1p_optab; break;
|
2056 |
|
|
CASE_FLT_FN (BUILT_IN_ASIN):
|
2057 |
|
|
builtin_optab = asin_optab; break;
|
2058 |
|
|
CASE_FLT_FN (BUILT_IN_ACOS):
|
2059 |
|
|
builtin_optab = acos_optab; break;
|
2060 |
|
|
CASE_FLT_FN (BUILT_IN_TAN):
|
2061 |
|
|
builtin_optab = tan_optab; break;
|
2062 |
|
|
CASE_FLT_FN (BUILT_IN_ATAN):
|
2063 |
|
|
builtin_optab = atan_optab; break;
|
2064 |
|
|
CASE_FLT_FN (BUILT_IN_FLOOR):
|
2065 |
|
|
builtin_optab = floor_optab; break;
|
2066 |
|
|
CASE_FLT_FN (BUILT_IN_CEIL):
|
2067 |
|
|
builtin_optab = ceil_optab; break;
|
2068 |
|
|
CASE_FLT_FN (BUILT_IN_TRUNC):
|
2069 |
|
|
builtin_optab = btrunc_optab; break;
|
2070 |
|
|
CASE_FLT_FN (BUILT_IN_ROUND):
|
2071 |
|
|
builtin_optab = round_optab; break;
|
2072 |
|
|
CASE_FLT_FN (BUILT_IN_NEARBYINT):
|
2073 |
|
|
builtin_optab = nearbyint_optab;
|
2074 |
|
|
if (flag_trapping_math)
|
2075 |
|
|
break;
|
2076 |
|
|
/* Else fallthrough and expand as rint. */
|
2077 |
|
|
CASE_FLT_FN (BUILT_IN_RINT):
|
2078 |
|
|
builtin_optab = rint_optab; break;
|
2079 |
|
|
CASE_FLT_FN (BUILT_IN_SIGNIFICAND):
|
2080 |
|
|
builtin_optab = significand_optab; break;
|
2081 |
|
|
default:
|
2082 |
|
|
gcc_unreachable ();
|
2083 |
|
|
}
|
2084 |
|
|
|
2085 |
|
|
/* Make a suitable register to place result in. */
|
2086 |
|
|
mode = TYPE_MODE (TREE_TYPE (exp));
|
2087 |
|
|
|
2088 |
|
|
if (! flag_errno_math || ! HONOR_NANS (mode))
|
2089 |
|
|
errno_set = false;
|
2090 |
|
|
|
2091 |
|
|
/* Before working hard, check whether the instruction is available. */
|
2092 |
|
|
if (optab_handler (builtin_optab, mode) != CODE_FOR_nothing
|
2093 |
|
|
&& (!errno_set || !optimize_insn_for_size_p ()))
|
2094 |
|
|
{
|
2095 |
|
|
target = gen_reg_rtx (mode);
|
2096 |
|
|
|
2097 |
|
|
/* Wrap the computation of the argument in a SAVE_EXPR, as we may
|
2098 |
|
|
need to expand the argument again. This way, we will not perform
|
2099 |
|
|
side-effects more the once. */
|
2100 |
|
|
CALL_EXPR_ARG (exp, 0) = arg = builtin_save_expr (arg);
|
2101 |
|
|
|
2102 |
|
|
op0 = expand_expr (arg, subtarget, VOIDmode, EXPAND_NORMAL);
|
2103 |
|
|
|
2104 |
|
|
start_sequence ();
|
2105 |
|
|
|
2106 |
|
|
/* Compute into TARGET.
|
2107 |
|
|
Set TARGET to wherever the result comes back. */
|
2108 |
|
|
target = expand_unop (mode, builtin_optab, op0, target, 0);
|
2109 |
|
|
|
2110 |
|
|
if (target != 0)
|
2111 |
|
|
{
|
2112 |
|
|
if (errno_set)
|
2113 |
|
|
expand_errno_check (exp, target);
|
2114 |
|
|
|
2115 |
|
|
/* Output the entire sequence. */
|
2116 |
|
|
insns = get_insns ();
|
2117 |
|
|
end_sequence ();
|
2118 |
|
|
emit_insn (insns);
|
2119 |
|
|
return target;
|
2120 |
|
|
}
|
2121 |
|
|
|
2122 |
|
|
/* If we were unable to expand via the builtin, stop the sequence
|
2123 |
|
|
(without outputting the insns) and call to the library function
|
2124 |
|
|
with the stabilized argument list. */
|
2125 |
|
|
end_sequence ();
|
2126 |
|
|
}
|
2127 |
|
|
|
2128 |
|
|
return expand_call (exp, target, target == const0_rtx);
|
2129 |
|
|
}
|
2130 |
|
|
|
2131 |
|
|
/* Expand a call to the builtin binary math functions (pow and atan2).
|
2132 |
|
|
Return NULL_RTX if a normal call should be emitted rather than expanding the
|
2133 |
|
|
function in-line. EXP is the expression that is a call to the builtin
|
2134 |
|
|
function; if convenient, the result should be placed in TARGET.
|
2135 |
|
|
SUBTARGET may be used as the target for computing one of EXP's
|
2136 |
|
|
operands. */
|
2137 |
|
|
|
2138 |
|
|
static rtx
|
2139 |
|
|
expand_builtin_mathfn_2 (tree exp, rtx target, rtx subtarget)
|
2140 |
|
|
{
|
2141 |
|
|
optab builtin_optab;
|
2142 |
|
|
rtx op0, op1, insns;
|
2143 |
|
|
int op1_type = REAL_TYPE;
|
2144 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
2145 |
|
|
tree arg0, arg1;
|
2146 |
|
|
enum machine_mode mode;
|
2147 |
|
|
bool errno_set = true;
|
2148 |
|
|
|
2149 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
2150 |
|
|
{
|
2151 |
|
|
CASE_FLT_FN (BUILT_IN_SCALBN):
|
2152 |
|
|
CASE_FLT_FN (BUILT_IN_SCALBLN):
|
2153 |
|
|
CASE_FLT_FN (BUILT_IN_LDEXP):
|
2154 |
|
|
op1_type = INTEGER_TYPE;
|
2155 |
|
|
default:
|
2156 |
|
|
break;
|
2157 |
|
|
}
|
2158 |
|
|
|
2159 |
|
|
if (!validate_arglist (exp, REAL_TYPE, op1_type, VOID_TYPE))
|
2160 |
|
|
return NULL_RTX;
|
2161 |
|
|
|
2162 |
|
|
arg0 = CALL_EXPR_ARG (exp, 0);
|
2163 |
|
|
arg1 = CALL_EXPR_ARG (exp, 1);
|
2164 |
|
|
|
2165 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
2166 |
|
|
{
|
2167 |
|
|
CASE_FLT_FN (BUILT_IN_POW):
|
2168 |
|
|
builtin_optab = pow_optab; break;
|
2169 |
|
|
CASE_FLT_FN (BUILT_IN_ATAN2):
|
2170 |
|
|
builtin_optab = atan2_optab; break;
|
2171 |
|
|
CASE_FLT_FN (BUILT_IN_SCALB):
|
2172 |
|
|
if (REAL_MODE_FORMAT (TYPE_MODE (TREE_TYPE (exp)))->b != 2)
|
2173 |
|
|
return 0;
|
2174 |
|
|
builtin_optab = scalb_optab; break;
|
2175 |
|
|
CASE_FLT_FN (BUILT_IN_SCALBN):
|
2176 |
|
|
CASE_FLT_FN (BUILT_IN_SCALBLN):
|
2177 |
|
|
if (REAL_MODE_FORMAT (TYPE_MODE (TREE_TYPE (exp)))->b != 2)
|
2178 |
|
|
return 0;
|
2179 |
|
|
/* Fall through... */
|
2180 |
|
|
CASE_FLT_FN (BUILT_IN_LDEXP):
|
2181 |
|
|
builtin_optab = ldexp_optab; break;
|
2182 |
|
|
CASE_FLT_FN (BUILT_IN_FMOD):
|
2183 |
|
|
builtin_optab = fmod_optab; break;
|
2184 |
|
|
CASE_FLT_FN (BUILT_IN_REMAINDER):
|
2185 |
|
|
CASE_FLT_FN (BUILT_IN_DREM):
|
2186 |
|
|
builtin_optab = remainder_optab; break;
|
2187 |
|
|
default:
|
2188 |
|
|
gcc_unreachable ();
|
2189 |
|
|
}
|
2190 |
|
|
|
2191 |
|
|
/* Make a suitable register to place result in. */
|
2192 |
|
|
mode = TYPE_MODE (TREE_TYPE (exp));
|
2193 |
|
|
|
2194 |
|
|
/* Before working hard, check whether the instruction is available. */
|
2195 |
|
|
if (optab_handler (builtin_optab, mode) == CODE_FOR_nothing)
|
2196 |
|
|
return NULL_RTX;
|
2197 |
|
|
|
2198 |
|
|
target = gen_reg_rtx (mode);
|
2199 |
|
|
|
2200 |
|
|
if (! flag_errno_math || ! HONOR_NANS (mode))
|
2201 |
|
|
errno_set = false;
|
2202 |
|
|
|
2203 |
|
|
if (errno_set && optimize_insn_for_size_p ())
|
2204 |
|
|
return 0;
|
2205 |
|
|
|
2206 |
|
|
/* Always stabilize the argument list. */
|
2207 |
|
|
CALL_EXPR_ARG (exp, 0) = arg0 = builtin_save_expr (arg0);
|
2208 |
|
|
CALL_EXPR_ARG (exp, 1) = arg1 = builtin_save_expr (arg1);
|
2209 |
|
|
|
2210 |
|
|
op0 = expand_expr (arg0, subtarget, VOIDmode, EXPAND_NORMAL);
|
2211 |
|
|
op1 = expand_normal (arg1);
|
2212 |
|
|
|
2213 |
|
|
start_sequence ();
|
2214 |
|
|
|
2215 |
|
|
/* Compute into TARGET.
|
2216 |
|
|
Set TARGET to wherever the result comes back. */
|
2217 |
|
|
target = expand_binop (mode, builtin_optab, op0, op1,
|
2218 |
|
|
target, 0, OPTAB_DIRECT);
|
2219 |
|
|
|
2220 |
|
|
/* If we were unable to expand via the builtin, stop the sequence
|
2221 |
|
|
(without outputting the insns) and call to the library function
|
2222 |
|
|
with the stabilized argument list. */
|
2223 |
|
|
if (target == 0)
|
2224 |
|
|
{
|
2225 |
|
|
end_sequence ();
|
2226 |
|
|
return expand_call (exp, target, target == const0_rtx);
|
2227 |
|
|
}
|
2228 |
|
|
|
2229 |
|
|
if (errno_set)
|
2230 |
|
|
expand_errno_check (exp, target);
|
2231 |
|
|
|
2232 |
|
|
/* Output the entire sequence. */
|
2233 |
|
|
insns = get_insns ();
|
2234 |
|
|
end_sequence ();
|
2235 |
|
|
emit_insn (insns);
|
2236 |
|
|
|
2237 |
|
|
return target;
|
2238 |
|
|
}
|
2239 |
|
|
|
2240 |
|
|
/* Expand a call to the builtin trinary math functions (fma).
|
2241 |
|
|
Return NULL_RTX if a normal call should be emitted rather than expanding the
|
2242 |
|
|
function in-line. EXP is the expression that is a call to the builtin
|
2243 |
|
|
function; if convenient, the result should be placed in TARGET.
|
2244 |
|
|
SUBTARGET may be used as the target for computing one of EXP's
|
2245 |
|
|
operands. */
|
2246 |
|
|
|
2247 |
|
|
static rtx
|
2248 |
|
|
expand_builtin_mathfn_ternary (tree exp, rtx target, rtx subtarget)
|
2249 |
|
|
{
|
2250 |
|
|
optab builtin_optab;
|
2251 |
|
|
rtx op0, op1, op2, insns;
|
2252 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
2253 |
|
|
tree arg0, arg1, arg2;
|
2254 |
|
|
enum machine_mode mode;
|
2255 |
|
|
|
2256 |
|
|
if (!validate_arglist (exp, REAL_TYPE, REAL_TYPE, REAL_TYPE, VOID_TYPE))
|
2257 |
|
|
return NULL_RTX;
|
2258 |
|
|
|
2259 |
|
|
arg0 = CALL_EXPR_ARG (exp, 0);
|
2260 |
|
|
arg1 = CALL_EXPR_ARG (exp, 1);
|
2261 |
|
|
arg2 = CALL_EXPR_ARG (exp, 2);
|
2262 |
|
|
|
2263 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
2264 |
|
|
{
|
2265 |
|
|
CASE_FLT_FN (BUILT_IN_FMA):
|
2266 |
|
|
builtin_optab = fma_optab; break;
|
2267 |
|
|
default:
|
2268 |
|
|
gcc_unreachable ();
|
2269 |
|
|
}
|
2270 |
|
|
|
2271 |
|
|
/* Make a suitable register to place result in. */
|
2272 |
|
|
mode = TYPE_MODE (TREE_TYPE (exp));
|
2273 |
|
|
|
2274 |
|
|
/* Before working hard, check whether the instruction is available. */
|
2275 |
|
|
if (optab_handler (builtin_optab, mode) == CODE_FOR_nothing)
|
2276 |
|
|
return NULL_RTX;
|
2277 |
|
|
|
2278 |
|
|
target = gen_reg_rtx (mode);
|
2279 |
|
|
|
2280 |
|
|
/* Always stabilize the argument list. */
|
2281 |
|
|
CALL_EXPR_ARG (exp, 0) = arg0 = builtin_save_expr (arg0);
|
2282 |
|
|
CALL_EXPR_ARG (exp, 1) = arg1 = builtin_save_expr (arg1);
|
2283 |
|
|
CALL_EXPR_ARG (exp, 2) = arg2 = builtin_save_expr (arg2);
|
2284 |
|
|
|
2285 |
|
|
op0 = expand_expr (arg0, subtarget, VOIDmode, EXPAND_NORMAL);
|
2286 |
|
|
op1 = expand_normal (arg1);
|
2287 |
|
|
op2 = expand_normal (arg2);
|
2288 |
|
|
|
2289 |
|
|
start_sequence ();
|
2290 |
|
|
|
2291 |
|
|
/* Compute into TARGET.
|
2292 |
|
|
Set TARGET to wherever the result comes back. */
|
2293 |
|
|
target = expand_ternary_op (mode, builtin_optab, op0, op1, op2,
|
2294 |
|
|
target, 0);
|
2295 |
|
|
|
2296 |
|
|
/* If we were unable to expand via the builtin, stop the sequence
|
2297 |
|
|
(without outputting the insns) and call to the library function
|
2298 |
|
|
with the stabilized argument list. */
|
2299 |
|
|
if (target == 0)
|
2300 |
|
|
{
|
2301 |
|
|
end_sequence ();
|
2302 |
|
|
return expand_call (exp, target, target == const0_rtx);
|
2303 |
|
|
}
|
2304 |
|
|
|
2305 |
|
|
/* Output the entire sequence. */
|
2306 |
|
|
insns = get_insns ();
|
2307 |
|
|
end_sequence ();
|
2308 |
|
|
emit_insn (insns);
|
2309 |
|
|
|
2310 |
|
|
return target;
|
2311 |
|
|
}
|
2312 |
|
|
|
2313 |
|
|
/* Expand a call to the builtin sin and cos math functions.
|
2314 |
|
|
Return NULL_RTX if a normal call should be emitted rather than expanding the
|
2315 |
|
|
function in-line. EXP is the expression that is a call to the builtin
|
2316 |
|
|
function; if convenient, the result should be placed in TARGET.
|
2317 |
|
|
SUBTARGET may be used as the target for computing one of EXP's
|
2318 |
|
|
operands. */
|
2319 |
|
|
|
2320 |
|
|
static rtx
|
2321 |
|
|
expand_builtin_mathfn_3 (tree exp, rtx target, rtx subtarget)
|
2322 |
|
|
{
|
2323 |
|
|
optab builtin_optab;
|
2324 |
|
|
rtx op0, insns;
|
2325 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
2326 |
|
|
enum machine_mode mode;
|
2327 |
|
|
tree arg;
|
2328 |
|
|
|
2329 |
|
|
if (!validate_arglist (exp, REAL_TYPE, VOID_TYPE))
|
2330 |
|
|
return NULL_RTX;
|
2331 |
|
|
|
2332 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
2333 |
|
|
|
2334 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
2335 |
|
|
{
|
2336 |
|
|
CASE_FLT_FN (BUILT_IN_SIN):
|
2337 |
|
|
CASE_FLT_FN (BUILT_IN_COS):
|
2338 |
|
|
builtin_optab = sincos_optab; break;
|
2339 |
|
|
default:
|
2340 |
|
|
gcc_unreachable ();
|
2341 |
|
|
}
|
2342 |
|
|
|
2343 |
|
|
/* Make a suitable register to place result in. */
|
2344 |
|
|
mode = TYPE_MODE (TREE_TYPE (exp));
|
2345 |
|
|
|
2346 |
|
|
/* Check if sincos insn is available, otherwise fallback
|
2347 |
|
|
to sin or cos insn. */
|
2348 |
|
|
if (optab_handler (builtin_optab, mode) == CODE_FOR_nothing)
|
2349 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
2350 |
|
|
{
|
2351 |
|
|
CASE_FLT_FN (BUILT_IN_SIN):
|
2352 |
|
|
builtin_optab = sin_optab; break;
|
2353 |
|
|
CASE_FLT_FN (BUILT_IN_COS):
|
2354 |
|
|
builtin_optab = cos_optab; break;
|
2355 |
|
|
default:
|
2356 |
|
|
gcc_unreachable ();
|
2357 |
|
|
}
|
2358 |
|
|
|
2359 |
|
|
/* Before working hard, check whether the instruction is available. */
|
2360 |
|
|
if (optab_handler (builtin_optab, mode) != CODE_FOR_nothing)
|
2361 |
|
|
{
|
2362 |
|
|
target = gen_reg_rtx (mode);
|
2363 |
|
|
|
2364 |
|
|
/* Wrap the computation of the argument in a SAVE_EXPR, as we may
|
2365 |
|
|
need to expand the argument again. This way, we will not perform
|
2366 |
|
|
side-effects more the once. */
|
2367 |
|
|
CALL_EXPR_ARG (exp, 0) = arg = builtin_save_expr (arg);
|
2368 |
|
|
|
2369 |
|
|
op0 = expand_expr (arg, subtarget, VOIDmode, EXPAND_NORMAL);
|
2370 |
|
|
|
2371 |
|
|
start_sequence ();
|
2372 |
|
|
|
2373 |
|
|
/* Compute into TARGET.
|
2374 |
|
|
Set TARGET to wherever the result comes back. */
|
2375 |
|
|
if (builtin_optab == sincos_optab)
|
2376 |
|
|
{
|
2377 |
|
|
int result;
|
2378 |
|
|
|
2379 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
2380 |
|
|
{
|
2381 |
|
|
CASE_FLT_FN (BUILT_IN_SIN):
|
2382 |
|
|
result = expand_twoval_unop (builtin_optab, op0, 0, target, 0);
|
2383 |
|
|
break;
|
2384 |
|
|
CASE_FLT_FN (BUILT_IN_COS):
|
2385 |
|
|
result = expand_twoval_unop (builtin_optab, op0, target, 0, 0);
|
2386 |
|
|
break;
|
2387 |
|
|
default:
|
2388 |
|
|
gcc_unreachable ();
|
2389 |
|
|
}
|
2390 |
|
|
gcc_assert (result);
|
2391 |
|
|
}
|
2392 |
|
|
else
|
2393 |
|
|
{
|
2394 |
|
|
target = expand_unop (mode, builtin_optab, op0, target, 0);
|
2395 |
|
|
}
|
2396 |
|
|
|
2397 |
|
|
if (target != 0)
|
2398 |
|
|
{
|
2399 |
|
|
/* Output the entire sequence. */
|
2400 |
|
|
insns = get_insns ();
|
2401 |
|
|
end_sequence ();
|
2402 |
|
|
emit_insn (insns);
|
2403 |
|
|
return target;
|
2404 |
|
|
}
|
2405 |
|
|
|
2406 |
|
|
/* If we were unable to expand via the builtin, stop the sequence
|
2407 |
|
|
(without outputting the insns) and call to the library function
|
2408 |
|
|
with the stabilized argument list. */
|
2409 |
|
|
end_sequence ();
|
2410 |
|
|
}
|
2411 |
|
|
|
2412 |
|
|
target = expand_call (exp, target, target == const0_rtx);
|
2413 |
|
|
|
2414 |
|
|
return target;
|
2415 |
|
|
}
|
2416 |
|
|
|
2417 |
|
|
/* Given an interclass math builtin decl FNDECL and it's argument ARG
|
2418 |
|
|
return an RTL instruction code that implements the functionality.
|
2419 |
|
|
If that isn't possible or available return CODE_FOR_nothing. */
|
2420 |
|
|
|
2421 |
|
|
static enum insn_code
|
2422 |
|
|
interclass_mathfn_icode (tree arg, tree fndecl)
|
2423 |
|
|
{
|
2424 |
|
|
bool errno_set = false;
|
2425 |
|
|
optab builtin_optab = 0;
|
2426 |
|
|
enum machine_mode mode;
|
2427 |
|
|
|
2428 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
2429 |
|
|
{
|
2430 |
|
|
CASE_FLT_FN (BUILT_IN_ILOGB):
|
2431 |
|
|
errno_set = true; builtin_optab = ilogb_optab; break;
|
2432 |
|
|
CASE_FLT_FN (BUILT_IN_ISINF):
|
2433 |
|
|
builtin_optab = isinf_optab; break;
|
2434 |
|
|
case BUILT_IN_ISNORMAL:
|
2435 |
|
|
case BUILT_IN_ISFINITE:
|
2436 |
|
|
CASE_FLT_FN (BUILT_IN_FINITE):
|
2437 |
|
|
case BUILT_IN_FINITED32:
|
2438 |
|
|
case BUILT_IN_FINITED64:
|
2439 |
|
|
case BUILT_IN_FINITED128:
|
2440 |
|
|
case BUILT_IN_ISINFD32:
|
2441 |
|
|
case BUILT_IN_ISINFD64:
|
2442 |
|
|
case BUILT_IN_ISINFD128:
|
2443 |
|
|
/* These builtins have no optabs (yet). */
|
2444 |
|
|
break;
|
2445 |
|
|
default:
|
2446 |
|
|
gcc_unreachable ();
|
2447 |
|
|
}
|
2448 |
|
|
|
2449 |
|
|
/* There's no easy way to detect the case we need to set EDOM. */
|
2450 |
|
|
if (flag_errno_math && errno_set)
|
2451 |
|
|
return CODE_FOR_nothing;
|
2452 |
|
|
|
2453 |
|
|
/* Optab mode depends on the mode of the input argument. */
|
2454 |
|
|
mode = TYPE_MODE (TREE_TYPE (arg));
|
2455 |
|
|
|
2456 |
|
|
if (builtin_optab)
|
2457 |
|
|
return optab_handler (builtin_optab, mode);
|
2458 |
|
|
return CODE_FOR_nothing;
|
2459 |
|
|
}
|
2460 |
|
|
|
2461 |
|
|
/* Expand a call to one of the builtin math functions that operate on
|
2462 |
|
|
floating point argument and output an integer result (ilogb, isinf,
|
2463 |
|
|
isnan, etc).
|
2464 |
|
|
Return 0 if a normal call should be emitted rather than expanding the
|
2465 |
|
|
function in-line. EXP is the expression that is a call to the builtin
|
2466 |
|
|
function; if convenient, the result should be placed in TARGET. */
|
2467 |
|
|
|
2468 |
|
|
static rtx
|
2469 |
|
|
expand_builtin_interclass_mathfn (tree exp, rtx target)
|
2470 |
|
|
{
|
2471 |
|
|
enum insn_code icode = CODE_FOR_nothing;
|
2472 |
|
|
rtx op0;
|
2473 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
2474 |
|
|
enum machine_mode mode;
|
2475 |
|
|
tree arg;
|
2476 |
|
|
|
2477 |
|
|
if (!validate_arglist (exp, REAL_TYPE, VOID_TYPE))
|
2478 |
|
|
return NULL_RTX;
|
2479 |
|
|
|
2480 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
2481 |
|
|
icode = interclass_mathfn_icode (arg, fndecl);
|
2482 |
|
|
mode = TYPE_MODE (TREE_TYPE (arg));
|
2483 |
|
|
|
2484 |
|
|
if (icode != CODE_FOR_nothing)
|
2485 |
|
|
{
|
2486 |
|
|
struct expand_operand ops[1];
|
2487 |
|
|
rtx last = get_last_insn ();
|
2488 |
|
|
tree orig_arg = arg;
|
2489 |
|
|
|
2490 |
|
|
/* Wrap the computation of the argument in a SAVE_EXPR, as we may
|
2491 |
|
|
need to expand the argument again. This way, we will not perform
|
2492 |
|
|
side-effects more the once. */
|
2493 |
|
|
CALL_EXPR_ARG (exp, 0) = arg = builtin_save_expr (arg);
|
2494 |
|
|
|
2495 |
|
|
op0 = expand_expr (arg, NULL_RTX, VOIDmode, EXPAND_NORMAL);
|
2496 |
|
|
|
2497 |
|
|
if (mode != GET_MODE (op0))
|
2498 |
|
|
op0 = convert_to_mode (mode, op0, 0);
|
2499 |
|
|
|
2500 |
|
|
create_output_operand (&ops[0], target, TYPE_MODE (TREE_TYPE (exp)));
|
2501 |
|
|
if (maybe_legitimize_operands (icode, 0, 1, ops)
|
2502 |
|
|
&& maybe_emit_unop_insn (icode, ops[0].value, op0, UNKNOWN))
|
2503 |
|
|
return ops[0].value;
|
2504 |
|
|
|
2505 |
|
|
delete_insns_since (last);
|
2506 |
|
|
CALL_EXPR_ARG (exp, 0) = orig_arg;
|
2507 |
|
|
}
|
2508 |
|
|
|
2509 |
|
|
return NULL_RTX;
|
2510 |
|
|
}
|
2511 |
|
|
|
2512 |
|
|
/* Expand a call to the builtin sincos math function.
|
2513 |
|
|
Return NULL_RTX if a normal call should be emitted rather than expanding the
|
2514 |
|
|
function in-line. EXP is the expression that is a call to the builtin
|
2515 |
|
|
function. */
|
2516 |
|
|
|
2517 |
|
|
static rtx
|
2518 |
|
|
expand_builtin_sincos (tree exp)
|
2519 |
|
|
{
|
2520 |
|
|
rtx op0, op1, op2, target1, target2;
|
2521 |
|
|
enum machine_mode mode;
|
2522 |
|
|
tree arg, sinp, cosp;
|
2523 |
|
|
int result;
|
2524 |
|
|
location_t loc = EXPR_LOCATION (exp);
|
2525 |
|
|
tree alias_type, alias_off;
|
2526 |
|
|
|
2527 |
|
|
if (!validate_arglist (exp, REAL_TYPE,
|
2528 |
|
|
POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
2529 |
|
|
return NULL_RTX;
|
2530 |
|
|
|
2531 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
2532 |
|
|
sinp = CALL_EXPR_ARG (exp, 1);
|
2533 |
|
|
cosp = CALL_EXPR_ARG (exp, 2);
|
2534 |
|
|
|
2535 |
|
|
/* Make a suitable register to place result in. */
|
2536 |
|
|
mode = TYPE_MODE (TREE_TYPE (arg));
|
2537 |
|
|
|
2538 |
|
|
/* Check if sincos insn is available, otherwise emit the call. */
|
2539 |
|
|
if (optab_handler (sincos_optab, mode) == CODE_FOR_nothing)
|
2540 |
|
|
return NULL_RTX;
|
2541 |
|
|
|
2542 |
|
|
target1 = gen_reg_rtx (mode);
|
2543 |
|
|
target2 = gen_reg_rtx (mode);
|
2544 |
|
|
|
2545 |
|
|
op0 = expand_normal (arg);
|
2546 |
|
|
alias_type = build_pointer_type_for_mode (TREE_TYPE (arg), ptr_mode, true);
|
2547 |
|
|
alias_off = build_int_cst (alias_type, 0);
|
2548 |
|
|
op1 = expand_normal (fold_build2_loc (loc, MEM_REF, TREE_TYPE (arg),
|
2549 |
|
|
sinp, alias_off));
|
2550 |
|
|
op2 = expand_normal (fold_build2_loc (loc, MEM_REF, TREE_TYPE (arg),
|
2551 |
|
|
cosp, alias_off));
|
2552 |
|
|
|
2553 |
|
|
/* Compute into target1 and target2.
|
2554 |
|
|
Set TARGET to wherever the result comes back. */
|
2555 |
|
|
result = expand_twoval_unop (sincos_optab, op0, target2, target1, 0);
|
2556 |
|
|
gcc_assert (result);
|
2557 |
|
|
|
2558 |
|
|
/* Move target1 and target2 to the memory locations indicated
|
2559 |
|
|
by op1 and op2. */
|
2560 |
|
|
emit_move_insn (op1, target1);
|
2561 |
|
|
emit_move_insn (op2, target2);
|
2562 |
|
|
|
2563 |
|
|
return const0_rtx;
|
2564 |
|
|
}
|
2565 |
|
|
|
2566 |
|
|
/* Expand a call to the internal cexpi builtin to the sincos math function.
|
2567 |
|
|
EXP is the expression that is a call to the builtin function; if convenient,
|
2568 |
|
|
the result should be placed in TARGET. */
|
2569 |
|
|
|
2570 |
|
|
static rtx
|
2571 |
|
|
expand_builtin_cexpi (tree exp, rtx target)
|
2572 |
|
|
{
|
2573 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
2574 |
|
|
tree arg, type;
|
2575 |
|
|
enum machine_mode mode;
|
2576 |
|
|
rtx op0, op1, op2;
|
2577 |
|
|
location_t loc = EXPR_LOCATION (exp);
|
2578 |
|
|
|
2579 |
|
|
if (!validate_arglist (exp, REAL_TYPE, VOID_TYPE))
|
2580 |
|
|
return NULL_RTX;
|
2581 |
|
|
|
2582 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
2583 |
|
|
type = TREE_TYPE (arg);
|
2584 |
|
|
mode = TYPE_MODE (TREE_TYPE (arg));
|
2585 |
|
|
|
2586 |
|
|
/* Try expanding via a sincos optab, fall back to emitting a libcall
|
2587 |
|
|
to sincos or cexp. We are sure we have sincos or cexp because cexpi
|
2588 |
|
|
is only generated from sincos, cexp or if we have either of them. */
|
2589 |
|
|
if (optab_handler (sincos_optab, mode) != CODE_FOR_nothing)
|
2590 |
|
|
{
|
2591 |
|
|
op1 = gen_reg_rtx (mode);
|
2592 |
|
|
op2 = gen_reg_rtx (mode);
|
2593 |
|
|
|
2594 |
|
|
op0 = expand_expr (arg, NULL_RTX, VOIDmode, EXPAND_NORMAL);
|
2595 |
|
|
|
2596 |
|
|
/* Compute into op1 and op2. */
|
2597 |
|
|
expand_twoval_unop (sincos_optab, op0, op2, op1, 0);
|
2598 |
|
|
}
|
2599 |
|
|
else if (TARGET_HAS_SINCOS)
|
2600 |
|
|
{
|
2601 |
|
|
tree call, fn = NULL_TREE;
|
2602 |
|
|
tree top1, top2;
|
2603 |
|
|
rtx op1a, op2a;
|
2604 |
|
|
|
2605 |
|
|
if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_CEXPIF)
|
2606 |
|
|
fn = builtin_decl_explicit (BUILT_IN_SINCOSF);
|
2607 |
|
|
else if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_CEXPI)
|
2608 |
|
|
fn = builtin_decl_explicit (BUILT_IN_SINCOS);
|
2609 |
|
|
else if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_CEXPIL)
|
2610 |
|
|
fn = builtin_decl_explicit (BUILT_IN_SINCOSL);
|
2611 |
|
|
else
|
2612 |
|
|
gcc_unreachable ();
|
2613 |
|
|
|
2614 |
|
|
op1 = assign_temp (TREE_TYPE (arg), 0, 1, 1);
|
2615 |
|
|
op2 = assign_temp (TREE_TYPE (arg), 0, 1, 1);
|
2616 |
|
|
op1a = copy_to_mode_reg (Pmode, XEXP (op1, 0));
|
2617 |
|
|
op2a = copy_to_mode_reg (Pmode, XEXP (op2, 0));
|
2618 |
|
|
top1 = make_tree (build_pointer_type (TREE_TYPE (arg)), op1a);
|
2619 |
|
|
top2 = make_tree (build_pointer_type (TREE_TYPE (arg)), op2a);
|
2620 |
|
|
|
2621 |
|
|
/* Make sure not to fold the sincos call again. */
|
2622 |
|
|
call = build1 (ADDR_EXPR, build_pointer_type (TREE_TYPE (fn)), fn);
|
2623 |
|
|
expand_normal (build_call_nary (TREE_TYPE (TREE_TYPE (fn)),
|
2624 |
|
|
call, 3, arg, top1, top2));
|
2625 |
|
|
}
|
2626 |
|
|
else
|
2627 |
|
|
{
|
2628 |
|
|
tree call, fn = NULL_TREE, narg;
|
2629 |
|
|
tree ctype = build_complex_type (type);
|
2630 |
|
|
|
2631 |
|
|
if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_CEXPIF)
|
2632 |
|
|
fn = builtin_decl_explicit (BUILT_IN_CEXPF);
|
2633 |
|
|
else if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_CEXPI)
|
2634 |
|
|
fn = builtin_decl_explicit (BUILT_IN_CEXP);
|
2635 |
|
|
else if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_CEXPIL)
|
2636 |
|
|
fn = builtin_decl_explicit (BUILT_IN_CEXPL);
|
2637 |
|
|
else
|
2638 |
|
|
gcc_unreachable ();
|
2639 |
|
|
|
2640 |
|
|
/* If we don't have a decl for cexp create one. This is the
|
2641 |
|
|
friendliest fallback if the user calls __builtin_cexpi
|
2642 |
|
|
without full target C99 function support. */
|
2643 |
|
|
if (fn == NULL_TREE)
|
2644 |
|
|
{
|
2645 |
|
|
tree fntype;
|
2646 |
|
|
const char *name = NULL;
|
2647 |
|
|
|
2648 |
|
|
if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_CEXPIF)
|
2649 |
|
|
name = "cexpf";
|
2650 |
|
|
else if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_CEXPI)
|
2651 |
|
|
name = "cexp";
|
2652 |
|
|
else if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_CEXPIL)
|
2653 |
|
|
name = "cexpl";
|
2654 |
|
|
|
2655 |
|
|
fntype = build_function_type_list (ctype, ctype, NULL_TREE);
|
2656 |
|
|
fn = build_fn_decl (name, fntype);
|
2657 |
|
|
}
|
2658 |
|
|
|
2659 |
|
|
narg = fold_build2_loc (loc, COMPLEX_EXPR, ctype,
|
2660 |
|
|
build_real (type, dconst0), arg);
|
2661 |
|
|
|
2662 |
|
|
/* Make sure not to fold the cexp call again. */
|
2663 |
|
|
call = build1 (ADDR_EXPR, build_pointer_type (TREE_TYPE (fn)), fn);
|
2664 |
|
|
return expand_expr (build_call_nary (ctype, call, 1, narg),
|
2665 |
|
|
target, VOIDmode, EXPAND_NORMAL);
|
2666 |
|
|
}
|
2667 |
|
|
|
2668 |
|
|
/* Now build the proper return type. */
|
2669 |
|
|
return expand_expr (build2 (COMPLEX_EXPR, build_complex_type (type),
|
2670 |
|
|
make_tree (TREE_TYPE (arg), op2),
|
2671 |
|
|
make_tree (TREE_TYPE (arg), op1)),
|
2672 |
|
|
target, VOIDmode, EXPAND_NORMAL);
|
2673 |
|
|
}
|
2674 |
|
|
|
2675 |
|
|
/* Conveniently construct a function call expression. FNDECL names the
|
2676 |
|
|
function to be called, N is the number of arguments, and the "..."
|
2677 |
|
|
parameters are the argument expressions. Unlike build_call_exr
|
2678 |
|
|
this doesn't fold the call, hence it will always return a CALL_EXPR. */
|
2679 |
|
|
|
2680 |
|
|
static tree
|
2681 |
|
|
build_call_nofold_loc (location_t loc, tree fndecl, int n, ...)
|
2682 |
|
|
{
|
2683 |
|
|
va_list ap;
|
2684 |
|
|
tree fntype = TREE_TYPE (fndecl);
|
2685 |
|
|
tree fn = build1 (ADDR_EXPR, build_pointer_type (fntype), fndecl);
|
2686 |
|
|
|
2687 |
|
|
va_start (ap, n);
|
2688 |
|
|
fn = build_call_valist (TREE_TYPE (fntype), fn, n, ap);
|
2689 |
|
|
va_end (ap);
|
2690 |
|
|
SET_EXPR_LOCATION (fn, loc);
|
2691 |
|
|
return fn;
|
2692 |
|
|
}
|
2693 |
|
|
|
2694 |
|
|
/* Expand a call to one of the builtin rounding functions gcc defines
|
2695 |
|
|
as an extension (lfloor and lceil). As these are gcc extensions we
|
2696 |
|
|
do not need to worry about setting errno to EDOM.
|
2697 |
|
|
If expanding via optab fails, lower expression to (int)(floor(x)).
|
2698 |
|
|
EXP is the expression that is a call to the builtin function;
|
2699 |
|
|
if convenient, the result should be placed in TARGET. */
|
2700 |
|
|
|
2701 |
|
|
static rtx
|
2702 |
|
|
expand_builtin_int_roundingfn (tree exp, rtx target)
|
2703 |
|
|
{
|
2704 |
|
|
convert_optab builtin_optab;
|
2705 |
|
|
rtx op0, insns, tmp;
|
2706 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
2707 |
|
|
enum built_in_function fallback_fn;
|
2708 |
|
|
tree fallback_fndecl;
|
2709 |
|
|
enum machine_mode mode;
|
2710 |
|
|
tree arg;
|
2711 |
|
|
|
2712 |
|
|
if (!validate_arglist (exp, REAL_TYPE, VOID_TYPE))
|
2713 |
|
|
gcc_unreachable ();
|
2714 |
|
|
|
2715 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
2716 |
|
|
|
2717 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
2718 |
|
|
{
|
2719 |
|
|
CASE_FLT_FN (BUILT_IN_ICEIL):
|
2720 |
|
|
CASE_FLT_FN (BUILT_IN_LCEIL):
|
2721 |
|
|
CASE_FLT_FN (BUILT_IN_LLCEIL):
|
2722 |
|
|
builtin_optab = lceil_optab;
|
2723 |
|
|
fallback_fn = BUILT_IN_CEIL;
|
2724 |
|
|
break;
|
2725 |
|
|
|
2726 |
|
|
CASE_FLT_FN (BUILT_IN_IFLOOR):
|
2727 |
|
|
CASE_FLT_FN (BUILT_IN_LFLOOR):
|
2728 |
|
|
CASE_FLT_FN (BUILT_IN_LLFLOOR):
|
2729 |
|
|
builtin_optab = lfloor_optab;
|
2730 |
|
|
fallback_fn = BUILT_IN_FLOOR;
|
2731 |
|
|
break;
|
2732 |
|
|
|
2733 |
|
|
default:
|
2734 |
|
|
gcc_unreachable ();
|
2735 |
|
|
}
|
2736 |
|
|
|
2737 |
|
|
/* Make a suitable register to place result in. */
|
2738 |
|
|
mode = TYPE_MODE (TREE_TYPE (exp));
|
2739 |
|
|
|
2740 |
|
|
target = gen_reg_rtx (mode);
|
2741 |
|
|
|
2742 |
|
|
/* Wrap the computation of the argument in a SAVE_EXPR, as we may
|
2743 |
|
|
need to expand the argument again. This way, we will not perform
|
2744 |
|
|
side-effects more the once. */
|
2745 |
|
|
CALL_EXPR_ARG (exp, 0) = arg = builtin_save_expr (arg);
|
2746 |
|
|
|
2747 |
|
|
op0 = expand_expr (arg, NULL, VOIDmode, EXPAND_NORMAL);
|
2748 |
|
|
|
2749 |
|
|
start_sequence ();
|
2750 |
|
|
|
2751 |
|
|
/* Compute into TARGET. */
|
2752 |
|
|
if (expand_sfix_optab (target, op0, builtin_optab))
|
2753 |
|
|
{
|
2754 |
|
|
/* Output the entire sequence. */
|
2755 |
|
|
insns = get_insns ();
|
2756 |
|
|
end_sequence ();
|
2757 |
|
|
emit_insn (insns);
|
2758 |
|
|
return target;
|
2759 |
|
|
}
|
2760 |
|
|
|
2761 |
|
|
/* If we were unable to expand via the builtin, stop the sequence
|
2762 |
|
|
(without outputting the insns). */
|
2763 |
|
|
end_sequence ();
|
2764 |
|
|
|
2765 |
|
|
/* Fall back to floating point rounding optab. */
|
2766 |
|
|
fallback_fndecl = mathfn_built_in (TREE_TYPE (arg), fallback_fn);
|
2767 |
|
|
|
2768 |
|
|
/* For non-C99 targets we may end up without a fallback fndecl here
|
2769 |
|
|
if the user called __builtin_lfloor directly. In this case emit
|
2770 |
|
|
a call to the floor/ceil variants nevertheless. This should result
|
2771 |
|
|
in the best user experience for not full C99 targets. */
|
2772 |
|
|
if (fallback_fndecl == NULL_TREE)
|
2773 |
|
|
{
|
2774 |
|
|
tree fntype;
|
2775 |
|
|
const char *name = NULL;
|
2776 |
|
|
|
2777 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
2778 |
|
|
{
|
2779 |
|
|
case BUILT_IN_ICEIL:
|
2780 |
|
|
case BUILT_IN_LCEIL:
|
2781 |
|
|
case BUILT_IN_LLCEIL:
|
2782 |
|
|
name = "ceil";
|
2783 |
|
|
break;
|
2784 |
|
|
case BUILT_IN_ICEILF:
|
2785 |
|
|
case BUILT_IN_LCEILF:
|
2786 |
|
|
case BUILT_IN_LLCEILF:
|
2787 |
|
|
name = "ceilf";
|
2788 |
|
|
break;
|
2789 |
|
|
case BUILT_IN_ICEILL:
|
2790 |
|
|
case BUILT_IN_LCEILL:
|
2791 |
|
|
case BUILT_IN_LLCEILL:
|
2792 |
|
|
name = "ceill";
|
2793 |
|
|
break;
|
2794 |
|
|
case BUILT_IN_IFLOOR:
|
2795 |
|
|
case BUILT_IN_LFLOOR:
|
2796 |
|
|
case BUILT_IN_LLFLOOR:
|
2797 |
|
|
name = "floor";
|
2798 |
|
|
break;
|
2799 |
|
|
case BUILT_IN_IFLOORF:
|
2800 |
|
|
case BUILT_IN_LFLOORF:
|
2801 |
|
|
case BUILT_IN_LLFLOORF:
|
2802 |
|
|
name = "floorf";
|
2803 |
|
|
break;
|
2804 |
|
|
case BUILT_IN_IFLOORL:
|
2805 |
|
|
case BUILT_IN_LFLOORL:
|
2806 |
|
|
case BUILT_IN_LLFLOORL:
|
2807 |
|
|
name = "floorl";
|
2808 |
|
|
break;
|
2809 |
|
|
default:
|
2810 |
|
|
gcc_unreachable ();
|
2811 |
|
|
}
|
2812 |
|
|
|
2813 |
|
|
fntype = build_function_type_list (TREE_TYPE (arg),
|
2814 |
|
|
TREE_TYPE (arg), NULL_TREE);
|
2815 |
|
|
fallback_fndecl = build_fn_decl (name, fntype);
|
2816 |
|
|
}
|
2817 |
|
|
|
2818 |
|
|
exp = build_call_nofold_loc (EXPR_LOCATION (exp), fallback_fndecl, 1, arg);
|
2819 |
|
|
|
2820 |
|
|
tmp = expand_normal (exp);
|
2821 |
|
|
|
2822 |
|
|
/* Truncate the result of floating point optab to integer
|
2823 |
|
|
via expand_fix (). */
|
2824 |
|
|
target = gen_reg_rtx (mode);
|
2825 |
|
|
expand_fix (target, tmp, 0);
|
2826 |
|
|
|
2827 |
|
|
return target;
|
2828 |
|
|
}
|
2829 |
|
|
|
2830 |
|
|
/* Expand a call to one of the builtin math functions doing integer
|
2831 |
|
|
conversion (lrint).
|
2832 |
|
|
Return 0 if a normal call should be emitted rather than expanding the
|
2833 |
|
|
function in-line. EXP is the expression that is a call to the builtin
|
2834 |
|
|
function; if convenient, the result should be placed in TARGET. */
|
2835 |
|
|
|
2836 |
|
|
static rtx
|
2837 |
|
|
expand_builtin_int_roundingfn_2 (tree exp, rtx target)
|
2838 |
|
|
{
|
2839 |
|
|
convert_optab builtin_optab;
|
2840 |
|
|
rtx op0, insns;
|
2841 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
2842 |
|
|
tree arg;
|
2843 |
|
|
enum machine_mode mode;
|
2844 |
|
|
|
2845 |
|
|
/* There's no easy way to detect the case we need to set EDOM. */
|
2846 |
|
|
if (flag_errno_math)
|
2847 |
|
|
return NULL_RTX;
|
2848 |
|
|
|
2849 |
|
|
if (!validate_arglist (exp, REAL_TYPE, VOID_TYPE))
|
2850 |
|
|
gcc_unreachable ();
|
2851 |
|
|
|
2852 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
2853 |
|
|
|
2854 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
2855 |
|
|
{
|
2856 |
|
|
CASE_FLT_FN (BUILT_IN_IRINT):
|
2857 |
|
|
CASE_FLT_FN (BUILT_IN_LRINT):
|
2858 |
|
|
CASE_FLT_FN (BUILT_IN_LLRINT):
|
2859 |
|
|
builtin_optab = lrint_optab; break;
|
2860 |
|
|
|
2861 |
|
|
CASE_FLT_FN (BUILT_IN_IROUND):
|
2862 |
|
|
CASE_FLT_FN (BUILT_IN_LROUND):
|
2863 |
|
|
CASE_FLT_FN (BUILT_IN_LLROUND):
|
2864 |
|
|
builtin_optab = lround_optab; break;
|
2865 |
|
|
|
2866 |
|
|
default:
|
2867 |
|
|
gcc_unreachable ();
|
2868 |
|
|
}
|
2869 |
|
|
|
2870 |
|
|
/* Make a suitable register to place result in. */
|
2871 |
|
|
mode = TYPE_MODE (TREE_TYPE (exp));
|
2872 |
|
|
|
2873 |
|
|
target = gen_reg_rtx (mode);
|
2874 |
|
|
|
2875 |
|
|
/* Wrap the computation of the argument in a SAVE_EXPR, as we may
|
2876 |
|
|
need to expand the argument again. This way, we will not perform
|
2877 |
|
|
side-effects more the once. */
|
2878 |
|
|
CALL_EXPR_ARG (exp, 0) = arg = builtin_save_expr (arg);
|
2879 |
|
|
|
2880 |
|
|
op0 = expand_expr (arg, NULL, VOIDmode, EXPAND_NORMAL);
|
2881 |
|
|
|
2882 |
|
|
start_sequence ();
|
2883 |
|
|
|
2884 |
|
|
if (expand_sfix_optab (target, op0, builtin_optab))
|
2885 |
|
|
{
|
2886 |
|
|
/* Output the entire sequence. */
|
2887 |
|
|
insns = get_insns ();
|
2888 |
|
|
end_sequence ();
|
2889 |
|
|
emit_insn (insns);
|
2890 |
|
|
return target;
|
2891 |
|
|
}
|
2892 |
|
|
|
2893 |
|
|
/* If we were unable to expand via the builtin, stop the sequence
|
2894 |
|
|
(without outputting the insns) and call to the library function
|
2895 |
|
|
with the stabilized argument list. */
|
2896 |
|
|
end_sequence ();
|
2897 |
|
|
|
2898 |
|
|
target = expand_call (exp, target, target == const0_rtx);
|
2899 |
|
|
|
2900 |
|
|
return target;
|
2901 |
|
|
}
|
2902 |
|
|
|
2903 |
|
|
/* Expand a call to the powi built-in mathematical function. Return NULL_RTX if
|
2904 |
|
|
a normal call should be emitted rather than expanding the function
|
2905 |
|
|
in-line. EXP is the expression that is a call to the builtin
|
2906 |
|
|
function; if convenient, the result should be placed in TARGET. */
|
2907 |
|
|
|
2908 |
|
|
static rtx
|
2909 |
|
|
expand_builtin_powi (tree exp, rtx target)
|
2910 |
|
|
{
|
2911 |
|
|
tree arg0, arg1;
|
2912 |
|
|
rtx op0, op1;
|
2913 |
|
|
enum machine_mode mode;
|
2914 |
|
|
enum machine_mode mode2;
|
2915 |
|
|
|
2916 |
|
|
if (! validate_arglist (exp, REAL_TYPE, INTEGER_TYPE, VOID_TYPE))
|
2917 |
|
|
return NULL_RTX;
|
2918 |
|
|
|
2919 |
|
|
arg0 = CALL_EXPR_ARG (exp, 0);
|
2920 |
|
|
arg1 = CALL_EXPR_ARG (exp, 1);
|
2921 |
|
|
mode = TYPE_MODE (TREE_TYPE (exp));
|
2922 |
|
|
|
2923 |
|
|
/* Emit a libcall to libgcc. */
|
2924 |
|
|
|
2925 |
|
|
/* Mode of the 2nd argument must match that of an int. */
|
2926 |
|
|
mode2 = mode_for_size (INT_TYPE_SIZE, MODE_INT, 0);
|
2927 |
|
|
|
2928 |
|
|
if (target == NULL_RTX)
|
2929 |
|
|
target = gen_reg_rtx (mode);
|
2930 |
|
|
|
2931 |
|
|
op0 = expand_expr (arg0, NULL_RTX, mode, EXPAND_NORMAL);
|
2932 |
|
|
if (GET_MODE (op0) != mode)
|
2933 |
|
|
op0 = convert_to_mode (mode, op0, 0);
|
2934 |
|
|
op1 = expand_expr (arg1, NULL_RTX, mode2, EXPAND_NORMAL);
|
2935 |
|
|
if (GET_MODE (op1) != mode2)
|
2936 |
|
|
op1 = convert_to_mode (mode2, op1, 0);
|
2937 |
|
|
|
2938 |
|
|
target = emit_library_call_value (optab_libfunc (powi_optab, mode),
|
2939 |
|
|
target, LCT_CONST, mode, 2,
|
2940 |
|
|
op0, mode, op1, mode2);
|
2941 |
|
|
|
2942 |
|
|
return target;
|
2943 |
|
|
}
|
2944 |
|
|
|
2945 |
|
|
/* Expand expression EXP which is a call to the strlen builtin. Return
|
2946 |
|
|
NULL_RTX if we failed the caller should emit a normal call, otherwise
|
2947 |
|
|
try to get the result in TARGET, if convenient. */
|
2948 |
|
|
|
2949 |
|
|
static rtx
|
2950 |
|
|
expand_builtin_strlen (tree exp, rtx target,
|
2951 |
|
|
enum machine_mode target_mode)
|
2952 |
|
|
{
|
2953 |
|
|
if (!validate_arglist (exp, POINTER_TYPE, VOID_TYPE))
|
2954 |
|
|
return NULL_RTX;
|
2955 |
|
|
else
|
2956 |
|
|
{
|
2957 |
|
|
struct expand_operand ops[4];
|
2958 |
|
|
rtx pat;
|
2959 |
|
|
tree len;
|
2960 |
|
|
tree src = CALL_EXPR_ARG (exp, 0);
|
2961 |
|
|
rtx src_reg, before_strlen;
|
2962 |
|
|
enum machine_mode insn_mode = target_mode;
|
2963 |
|
|
enum insn_code icode = CODE_FOR_nothing;
|
2964 |
|
|
unsigned int align;
|
2965 |
|
|
|
2966 |
|
|
/* If the length can be computed at compile-time, return it. */
|
2967 |
|
|
len = c_strlen (src, 0);
|
2968 |
|
|
if (len)
|
2969 |
|
|
return expand_expr (len, target, target_mode, EXPAND_NORMAL);
|
2970 |
|
|
|
2971 |
|
|
/* If the length can be computed at compile-time and is constant
|
2972 |
|
|
integer, but there are side-effects in src, evaluate
|
2973 |
|
|
src for side-effects, then return len.
|
2974 |
|
|
E.g. x = strlen (i++ ? "xfoo" + 1 : "bar");
|
2975 |
|
|
can be optimized into: i++; x = 3; */
|
2976 |
|
|
len = c_strlen (src, 1);
|
2977 |
|
|
if (len && TREE_CODE (len) == INTEGER_CST)
|
2978 |
|
|
{
|
2979 |
|
|
expand_expr (src, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
2980 |
|
|
return expand_expr (len, target, target_mode, EXPAND_NORMAL);
|
2981 |
|
|
}
|
2982 |
|
|
|
2983 |
|
|
align = get_pointer_alignment (src) / BITS_PER_UNIT;
|
2984 |
|
|
|
2985 |
|
|
/* If SRC is not a pointer type, don't do this operation inline. */
|
2986 |
|
|
if (align == 0)
|
2987 |
|
|
return NULL_RTX;
|
2988 |
|
|
|
2989 |
|
|
/* Bail out if we can't compute strlen in the right mode. */
|
2990 |
|
|
while (insn_mode != VOIDmode)
|
2991 |
|
|
{
|
2992 |
|
|
icode = optab_handler (strlen_optab, insn_mode);
|
2993 |
|
|
if (icode != CODE_FOR_nothing)
|
2994 |
|
|
break;
|
2995 |
|
|
|
2996 |
|
|
insn_mode = GET_MODE_WIDER_MODE (insn_mode);
|
2997 |
|
|
}
|
2998 |
|
|
if (insn_mode == VOIDmode)
|
2999 |
|
|
return NULL_RTX;
|
3000 |
|
|
|
3001 |
|
|
/* Make a place to hold the source address. We will not expand
|
3002 |
|
|
the actual source until we are sure that the expansion will
|
3003 |
|
|
not fail -- there are trees that cannot be expanded twice. */
|
3004 |
|
|
src_reg = gen_reg_rtx (Pmode);
|
3005 |
|
|
|
3006 |
|
|
/* Mark the beginning of the strlen sequence so we can emit the
|
3007 |
|
|
source operand later. */
|
3008 |
|
|
before_strlen = get_last_insn ();
|
3009 |
|
|
|
3010 |
|
|
create_output_operand (&ops[0], target, insn_mode);
|
3011 |
|
|
create_fixed_operand (&ops[1], gen_rtx_MEM (BLKmode, src_reg));
|
3012 |
|
|
create_integer_operand (&ops[2], 0);
|
3013 |
|
|
create_integer_operand (&ops[3], align);
|
3014 |
|
|
if (!maybe_expand_insn (icode, 4, ops))
|
3015 |
|
|
return NULL_RTX;
|
3016 |
|
|
|
3017 |
|
|
/* Now that we are assured of success, expand the source. */
|
3018 |
|
|
start_sequence ();
|
3019 |
|
|
pat = expand_expr (src, src_reg, Pmode, EXPAND_NORMAL);
|
3020 |
|
|
if (pat != src_reg)
|
3021 |
|
|
{
|
3022 |
|
|
#ifdef POINTERS_EXTEND_UNSIGNED
|
3023 |
|
|
if (GET_MODE (pat) != Pmode)
|
3024 |
|
|
pat = convert_to_mode (Pmode, pat,
|
3025 |
|
|
POINTERS_EXTEND_UNSIGNED);
|
3026 |
|
|
#endif
|
3027 |
|
|
emit_move_insn (src_reg, pat);
|
3028 |
|
|
}
|
3029 |
|
|
pat = get_insns ();
|
3030 |
|
|
end_sequence ();
|
3031 |
|
|
|
3032 |
|
|
if (before_strlen)
|
3033 |
|
|
emit_insn_after (pat, before_strlen);
|
3034 |
|
|
else
|
3035 |
|
|
emit_insn_before (pat, get_insns ());
|
3036 |
|
|
|
3037 |
|
|
/* Return the value in the proper mode for this function. */
|
3038 |
|
|
if (GET_MODE (ops[0].value) == target_mode)
|
3039 |
|
|
target = ops[0].value;
|
3040 |
|
|
else if (target != 0)
|
3041 |
|
|
convert_move (target, ops[0].value, 0);
|
3042 |
|
|
else
|
3043 |
|
|
target = convert_to_mode (target_mode, ops[0].value, 0);
|
3044 |
|
|
|
3045 |
|
|
return target;
|
3046 |
|
|
}
|
3047 |
|
|
}
|
3048 |
|
|
|
3049 |
|
|
/* Callback routine for store_by_pieces. Read GET_MODE_BITSIZE (MODE)
|
3050 |
|
|
bytes from constant string DATA + OFFSET and return it as target
|
3051 |
|
|
constant. */
|
3052 |
|
|
|
3053 |
|
|
static rtx
|
3054 |
|
|
builtin_memcpy_read_str (void *data, HOST_WIDE_INT offset,
|
3055 |
|
|
enum machine_mode mode)
|
3056 |
|
|
{
|
3057 |
|
|
const char *str = (const char *) data;
|
3058 |
|
|
|
3059 |
|
|
gcc_assert (offset >= 0
|
3060 |
|
|
&& ((unsigned HOST_WIDE_INT) offset + GET_MODE_SIZE (mode)
|
3061 |
|
|
<= strlen (str) + 1));
|
3062 |
|
|
|
3063 |
|
|
return c_readstr (str + offset, mode);
|
3064 |
|
|
}
|
3065 |
|
|
|
3066 |
|
|
/* Expand a call EXP to the memcpy builtin.
|
3067 |
|
|
Return NULL_RTX if we failed, the caller should emit a normal call,
|
3068 |
|
|
otherwise try to get the result in TARGET, if convenient (and in
|
3069 |
|
|
mode MODE if that's convenient). */
|
3070 |
|
|
|
3071 |
|
|
static rtx
|
3072 |
|
|
expand_builtin_memcpy (tree exp, rtx target)
|
3073 |
|
|
{
|
3074 |
|
|
if (!validate_arglist (exp,
|
3075 |
|
|
POINTER_TYPE, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
3076 |
|
|
return NULL_RTX;
|
3077 |
|
|
else
|
3078 |
|
|
{
|
3079 |
|
|
tree dest = CALL_EXPR_ARG (exp, 0);
|
3080 |
|
|
tree src = CALL_EXPR_ARG (exp, 1);
|
3081 |
|
|
tree len = CALL_EXPR_ARG (exp, 2);
|
3082 |
|
|
const char *src_str;
|
3083 |
|
|
unsigned int src_align = get_pointer_alignment (src);
|
3084 |
|
|
unsigned int dest_align = get_pointer_alignment (dest);
|
3085 |
|
|
rtx dest_mem, src_mem, dest_addr, len_rtx;
|
3086 |
|
|
HOST_WIDE_INT expected_size = -1;
|
3087 |
|
|
unsigned int expected_align = 0;
|
3088 |
|
|
|
3089 |
|
|
/* If DEST is not a pointer type, call the normal function. */
|
3090 |
|
|
if (dest_align == 0)
|
3091 |
|
|
return NULL_RTX;
|
3092 |
|
|
|
3093 |
|
|
/* If either SRC is not a pointer type, don't do this
|
3094 |
|
|
operation in-line. */
|
3095 |
|
|
if (src_align == 0)
|
3096 |
|
|
return NULL_RTX;
|
3097 |
|
|
|
3098 |
|
|
if (currently_expanding_gimple_stmt)
|
3099 |
|
|
stringop_block_profile (currently_expanding_gimple_stmt,
|
3100 |
|
|
&expected_align, &expected_size);
|
3101 |
|
|
|
3102 |
|
|
if (expected_align < dest_align)
|
3103 |
|
|
expected_align = dest_align;
|
3104 |
|
|
dest_mem = get_memory_rtx (dest, len);
|
3105 |
|
|
set_mem_align (dest_mem, dest_align);
|
3106 |
|
|
len_rtx = expand_normal (len);
|
3107 |
|
|
src_str = c_getstr (src);
|
3108 |
|
|
|
3109 |
|
|
/* If SRC is a string constant and block move would be done
|
3110 |
|
|
by pieces, we can avoid loading the string from memory
|
3111 |
|
|
and only stored the computed constants. */
|
3112 |
|
|
if (src_str
|
3113 |
|
|
&& CONST_INT_P (len_rtx)
|
3114 |
|
|
&& (unsigned HOST_WIDE_INT) INTVAL (len_rtx) <= strlen (src_str) + 1
|
3115 |
|
|
&& can_store_by_pieces (INTVAL (len_rtx), builtin_memcpy_read_str,
|
3116 |
|
|
CONST_CAST (char *, src_str),
|
3117 |
|
|
dest_align, false))
|
3118 |
|
|
{
|
3119 |
|
|
dest_mem = store_by_pieces (dest_mem, INTVAL (len_rtx),
|
3120 |
|
|
builtin_memcpy_read_str,
|
3121 |
|
|
CONST_CAST (char *, src_str),
|
3122 |
|
|
dest_align, false, 0);
|
3123 |
|
|
dest_mem = force_operand (XEXP (dest_mem, 0), target);
|
3124 |
|
|
dest_mem = convert_memory_address (ptr_mode, dest_mem);
|
3125 |
|
|
return dest_mem;
|
3126 |
|
|
}
|
3127 |
|
|
|
3128 |
|
|
src_mem = get_memory_rtx (src, len);
|
3129 |
|
|
set_mem_align (src_mem, src_align);
|
3130 |
|
|
|
3131 |
|
|
/* Copy word part most expediently. */
|
3132 |
|
|
dest_addr = emit_block_move_hints (dest_mem, src_mem, len_rtx,
|
3133 |
|
|
CALL_EXPR_TAILCALL (exp)
|
3134 |
|
|
? BLOCK_OP_TAILCALL : BLOCK_OP_NORMAL,
|
3135 |
|
|
expected_align, expected_size);
|
3136 |
|
|
|
3137 |
|
|
if (dest_addr == 0)
|
3138 |
|
|
{
|
3139 |
|
|
dest_addr = force_operand (XEXP (dest_mem, 0), target);
|
3140 |
|
|
dest_addr = convert_memory_address (ptr_mode, dest_addr);
|
3141 |
|
|
}
|
3142 |
|
|
return dest_addr;
|
3143 |
|
|
}
|
3144 |
|
|
}
|
3145 |
|
|
|
3146 |
|
|
/* Expand a call EXP to the mempcpy builtin.
|
3147 |
|
|
Return NULL_RTX if we failed; the caller should emit a normal call,
|
3148 |
|
|
otherwise try to get the result in TARGET, if convenient (and in
|
3149 |
|
|
mode MODE if that's convenient). If ENDP is 0 return the
|
3150 |
|
|
destination pointer, if ENDP is 1 return the end pointer ala
|
3151 |
|
|
mempcpy, and if ENDP is 2 return the end pointer minus one ala
|
3152 |
|
|
stpcpy. */
|
3153 |
|
|
|
3154 |
|
|
static rtx
|
3155 |
|
|
expand_builtin_mempcpy (tree exp, rtx target, enum machine_mode mode)
|
3156 |
|
|
{
|
3157 |
|
|
if (!validate_arglist (exp,
|
3158 |
|
|
POINTER_TYPE, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
3159 |
|
|
return NULL_RTX;
|
3160 |
|
|
else
|
3161 |
|
|
{
|
3162 |
|
|
tree dest = CALL_EXPR_ARG (exp, 0);
|
3163 |
|
|
tree src = CALL_EXPR_ARG (exp, 1);
|
3164 |
|
|
tree len = CALL_EXPR_ARG (exp, 2);
|
3165 |
|
|
return expand_builtin_mempcpy_args (dest, src, len,
|
3166 |
|
|
target, mode, /*endp=*/ 1);
|
3167 |
|
|
}
|
3168 |
|
|
}
|
3169 |
|
|
|
3170 |
|
|
/* Helper function to do the actual work for expand_builtin_mempcpy. The
|
3171 |
|
|
arguments to the builtin_mempcpy call DEST, SRC, and LEN are broken out
|
3172 |
|
|
so that this can also be called without constructing an actual CALL_EXPR.
|
3173 |
|
|
The other arguments and return value are the same as for
|
3174 |
|
|
expand_builtin_mempcpy. */
|
3175 |
|
|
|
3176 |
|
|
static rtx
|
3177 |
|
|
expand_builtin_mempcpy_args (tree dest, tree src, tree len,
|
3178 |
|
|
rtx target, enum machine_mode mode, int endp)
|
3179 |
|
|
{
|
3180 |
|
|
/* If return value is ignored, transform mempcpy into memcpy. */
|
3181 |
|
|
if (target == const0_rtx && builtin_decl_implicit_p (BUILT_IN_MEMCPY))
|
3182 |
|
|
{
|
3183 |
|
|
tree fn = builtin_decl_implicit (BUILT_IN_MEMCPY);
|
3184 |
|
|
tree result = build_call_nofold_loc (UNKNOWN_LOCATION, fn, 3,
|
3185 |
|
|
dest, src, len);
|
3186 |
|
|
return expand_expr (result, target, mode, EXPAND_NORMAL);
|
3187 |
|
|
}
|
3188 |
|
|
else
|
3189 |
|
|
{
|
3190 |
|
|
const char *src_str;
|
3191 |
|
|
unsigned int src_align = get_pointer_alignment (src);
|
3192 |
|
|
unsigned int dest_align = get_pointer_alignment (dest);
|
3193 |
|
|
rtx dest_mem, src_mem, len_rtx;
|
3194 |
|
|
|
3195 |
|
|
/* If either SRC or DEST is not a pointer type, don't do this
|
3196 |
|
|
operation in-line. */
|
3197 |
|
|
if (dest_align == 0 || src_align == 0)
|
3198 |
|
|
return NULL_RTX;
|
3199 |
|
|
|
3200 |
|
|
/* If LEN is not constant, call the normal function. */
|
3201 |
|
|
if (! host_integerp (len, 1))
|
3202 |
|
|
return NULL_RTX;
|
3203 |
|
|
|
3204 |
|
|
len_rtx = expand_normal (len);
|
3205 |
|
|
src_str = c_getstr (src);
|
3206 |
|
|
|
3207 |
|
|
/* If SRC is a string constant and block move would be done
|
3208 |
|
|
by pieces, we can avoid loading the string from memory
|
3209 |
|
|
and only stored the computed constants. */
|
3210 |
|
|
if (src_str
|
3211 |
|
|
&& CONST_INT_P (len_rtx)
|
3212 |
|
|
&& (unsigned HOST_WIDE_INT) INTVAL (len_rtx) <= strlen (src_str) + 1
|
3213 |
|
|
&& can_store_by_pieces (INTVAL (len_rtx), builtin_memcpy_read_str,
|
3214 |
|
|
CONST_CAST (char *, src_str),
|
3215 |
|
|
dest_align, false))
|
3216 |
|
|
{
|
3217 |
|
|
dest_mem = get_memory_rtx (dest, len);
|
3218 |
|
|
set_mem_align (dest_mem, dest_align);
|
3219 |
|
|
dest_mem = store_by_pieces (dest_mem, INTVAL (len_rtx),
|
3220 |
|
|
builtin_memcpy_read_str,
|
3221 |
|
|
CONST_CAST (char *, src_str),
|
3222 |
|
|
dest_align, false, endp);
|
3223 |
|
|
dest_mem = force_operand (XEXP (dest_mem, 0), NULL_RTX);
|
3224 |
|
|
dest_mem = convert_memory_address (ptr_mode, dest_mem);
|
3225 |
|
|
return dest_mem;
|
3226 |
|
|
}
|
3227 |
|
|
|
3228 |
|
|
if (CONST_INT_P (len_rtx)
|
3229 |
|
|
&& can_move_by_pieces (INTVAL (len_rtx),
|
3230 |
|
|
MIN (dest_align, src_align)))
|
3231 |
|
|
{
|
3232 |
|
|
dest_mem = get_memory_rtx (dest, len);
|
3233 |
|
|
set_mem_align (dest_mem, dest_align);
|
3234 |
|
|
src_mem = get_memory_rtx (src, len);
|
3235 |
|
|
set_mem_align (src_mem, src_align);
|
3236 |
|
|
dest_mem = move_by_pieces (dest_mem, src_mem, INTVAL (len_rtx),
|
3237 |
|
|
MIN (dest_align, src_align), endp);
|
3238 |
|
|
dest_mem = force_operand (XEXP (dest_mem, 0), NULL_RTX);
|
3239 |
|
|
dest_mem = convert_memory_address (ptr_mode, dest_mem);
|
3240 |
|
|
return dest_mem;
|
3241 |
|
|
}
|
3242 |
|
|
|
3243 |
|
|
return NULL_RTX;
|
3244 |
|
|
}
|
3245 |
|
|
}
|
3246 |
|
|
|
3247 |
|
|
#ifndef HAVE_movstr
|
3248 |
|
|
# define HAVE_movstr 0
|
3249 |
|
|
# define CODE_FOR_movstr CODE_FOR_nothing
|
3250 |
|
|
#endif
|
3251 |
|
|
|
3252 |
|
|
/* Expand into a movstr instruction, if one is available. Return NULL_RTX if
|
3253 |
|
|
we failed, the caller should emit a normal call, otherwise try to
|
3254 |
|
|
get the result in TARGET, if convenient. If ENDP is 0 return the
|
3255 |
|
|
destination pointer, if ENDP is 1 return the end pointer ala
|
3256 |
|
|
mempcpy, and if ENDP is 2 return the end pointer minus one ala
|
3257 |
|
|
stpcpy. */
|
3258 |
|
|
|
3259 |
|
|
static rtx
|
3260 |
|
|
expand_movstr (tree dest, tree src, rtx target, int endp)
|
3261 |
|
|
{
|
3262 |
|
|
struct expand_operand ops[3];
|
3263 |
|
|
rtx dest_mem;
|
3264 |
|
|
rtx src_mem;
|
3265 |
|
|
|
3266 |
|
|
if (!HAVE_movstr)
|
3267 |
|
|
return NULL_RTX;
|
3268 |
|
|
|
3269 |
|
|
dest_mem = get_memory_rtx (dest, NULL);
|
3270 |
|
|
src_mem = get_memory_rtx (src, NULL);
|
3271 |
|
|
if (!endp)
|
3272 |
|
|
{
|
3273 |
|
|
target = force_reg (Pmode, XEXP (dest_mem, 0));
|
3274 |
|
|
dest_mem = replace_equiv_address (dest_mem, target);
|
3275 |
|
|
}
|
3276 |
|
|
|
3277 |
|
|
create_output_operand (&ops[0], endp ? target : NULL_RTX, Pmode);
|
3278 |
|
|
create_fixed_operand (&ops[1], dest_mem);
|
3279 |
|
|
create_fixed_operand (&ops[2], src_mem);
|
3280 |
|
|
expand_insn (CODE_FOR_movstr, 3, ops);
|
3281 |
|
|
|
3282 |
|
|
if (endp && target != const0_rtx)
|
3283 |
|
|
{
|
3284 |
|
|
target = ops[0].value;
|
3285 |
|
|
/* movstr is supposed to set end to the address of the NUL
|
3286 |
|
|
terminator. If the caller requested a mempcpy-like return value,
|
3287 |
|
|
adjust it. */
|
3288 |
|
|
if (endp == 1)
|
3289 |
|
|
{
|
3290 |
|
|
rtx tem = plus_constant (gen_lowpart (GET_MODE (target), target), 1);
|
3291 |
|
|
emit_move_insn (target, force_operand (tem, NULL_RTX));
|
3292 |
|
|
}
|
3293 |
|
|
}
|
3294 |
|
|
return target;
|
3295 |
|
|
}
|
3296 |
|
|
|
3297 |
|
|
/* Expand expression EXP, which is a call to the strcpy builtin. Return
|
3298 |
|
|
NULL_RTX if we failed the caller should emit a normal call, otherwise
|
3299 |
|
|
try to get the result in TARGET, if convenient (and in mode MODE if that's
|
3300 |
|
|
convenient). */
|
3301 |
|
|
|
3302 |
|
|
static rtx
|
3303 |
|
|
expand_builtin_strcpy (tree exp, rtx target)
|
3304 |
|
|
{
|
3305 |
|
|
if (validate_arglist (exp, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
3306 |
|
|
{
|
3307 |
|
|
tree dest = CALL_EXPR_ARG (exp, 0);
|
3308 |
|
|
tree src = CALL_EXPR_ARG (exp, 1);
|
3309 |
|
|
return expand_builtin_strcpy_args (dest, src, target);
|
3310 |
|
|
}
|
3311 |
|
|
return NULL_RTX;
|
3312 |
|
|
}
|
3313 |
|
|
|
3314 |
|
|
/* Helper function to do the actual work for expand_builtin_strcpy. The
|
3315 |
|
|
arguments to the builtin_strcpy call DEST and SRC are broken out
|
3316 |
|
|
so that this can also be called without constructing an actual CALL_EXPR.
|
3317 |
|
|
The other arguments and return value are the same as for
|
3318 |
|
|
expand_builtin_strcpy. */
|
3319 |
|
|
|
3320 |
|
|
static rtx
|
3321 |
|
|
expand_builtin_strcpy_args (tree dest, tree src, rtx target)
|
3322 |
|
|
{
|
3323 |
|
|
return expand_movstr (dest, src, target, /*endp=*/0);
|
3324 |
|
|
}
|
3325 |
|
|
|
3326 |
|
|
/* Expand a call EXP to the stpcpy builtin.
|
3327 |
|
|
Return NULL_RTX if we failed the caller should emit a normal call,
|
3328 |
|
|
otherwise try to get the result in TARGET, if convenient (and in
|
3329 |
|
|
mode MODE if that's convenient). */
|
3330 |
|
|
|
3331 |
|
|
static rtx
|
3332 |
|
|
expand_builtin_stpcpy (tree exp, rtx target, enum machine_mode mode)
|
3333 |
|
|
{
|
3334 |
|
|
tree dst, src;
|
3335 |
|
|
location_t loc = EXPR_LOCATION (exp);
|
3336 |
|
|
|
3337 |
|
|
if (!validate_arglist (exp, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
3338 |
|
|
return NULL_RTX;
|
3339 |
|
|
|
3340 |
|
|
dst = CALL_EXPR_ARG (exp, 0);
|
3341 |
|
|
src = CALL_EXPR_ARG (exp, 1);
|
3342 |
|
|
|
3343 |
|
|
/* If return value is ignored, transform stpcpy into strcpy. */
|
3344 |
|
|
if (target == const0_rtx && builtin_decl_implicit (BUILT_IN_STRCPY))
|
3345 |
|
|
{
|
3346 |
|
|
tree fn = builtin_decl_implicit (BUILT_IN_STRCPY);
|
3347 |
|
|
tree result = build_call_nofold_loc (loc, fn, 2, dst, src);
|
3348 |
|
|
return expand_expr (result, target, mode, EXPAND_NORMAL);
|
3349 |
|
|
}
|
3350 |
|
|
else
|
3351 |
|
|
{
|
3352 |
|
|
tree len, lenp1;
|
3353 |
|
|
rtx ret;
|
3354 |
|
|
|
3355 |
|
|
/* Ensure we get an actual string whose length can be evaluated at
|
3356 |
|
|
compile-time, not an expression containing a string. This is
|
3357 |
|
|
because the latter will potentially produce pessimized code
|
3358 |
|
|
when used to produce the return value. */
|
3359 |
|
|
if (! c_getstr (src) || ! (len = c_strlen (src, 0)))
|
3360 |
|
|
return expand_movstr (dst, src, target, /*endp=*/2);
|
3361 |
|
|
|
3362 |
|
|
lenp1 = size_binop_loc (loc, PLUS_EXPR, len, ssize_int (1));
|
3363 |
|
|
ret = expand_builtin_mempcpy_args (dst, src, lenp1,
|
3364 |
|
|
target, mode, /*endp=*/2);
|
3365 |
|
|
|
3366 |
|
|
if (ret)
|
3367 |
|
|
return ret;
|
3368 |
|
|
|
3369 |
|
|
if (TREE_CODE (len) == INTEGER_CST)
|
3370 |
|
|
{
|
3371 |
|
|
rtx len_rtx = expand_normal (len);
|
3372 |
|
|
|
3373 |
|
|
if (CONST_INT_P (len_rtx))
|
3374 |
|
|
{
|
3375 |
|
|
ret = expand_builtin_strcpy_args (dst, src, target);
|
3376 |
|
|
|
3377 |
|
|
if (ret)
|
3378 |
|
|
{
|
3379 |
|
|
if (! target)
|
3380 |
|
|
{
|
3381 |
|
|
if (mode != VOIDmode)
|
3382 |
|
|
target = gen_reg_rtx (mode);
|
3383 |
|
|
else
|
3384 |
|
|
target = gen_reg_rtx (GET_MODE (ret));
|
3385 |
|
|
}
|
3386 |
|
|
if (GET_MODE (target) != GET_MODE (ret))
|
3387 |
|
|
ret = gen_lowpart (GET_MODE (target), ret);
|
3388 |
|
|
|
3389 |
|
|
ret = plus_constant (ret, INTVAL (len_rtx));
|
3390 |
|
|
ret = emit_move_insn (target, force_operand (ret, NULL_RTX));
|
3391 |
|
|
gcc_assert (ret);
|
3392 |
|
|
|
3393 |
|
|
return target;
|
3394 |
|
|
}
|
3395 |
|
|
}
|
3396 |
|
|
}
|
3397 |
|
|
|
3398 |
|
|
return expand_movstr (dst, src, target, /*endp=*/2);
|
3399 |
|
|
}
|
3400 |
|
|
}
|
3401 |
|
|
|
3402 |
|
|
/* Callback routine for store_by_pieces. Read GET_MODE_BITSIZE (MODE)
|
3403 |
|
|
bytes from constant string DATA + OFFSET and return it as target
|
3404 |
|
|
constant. */
|
3405 |
|
|
|
3406 |
|
|
rtx
|
3407 |
|
|
builtin_strncpy_read_str (void *data, HOST_WIDE_INT offset,
|
3408 |
|
|
enum machine_mode mode)
|
3409 |
|
|
{
|
3410 |
|
|
const char *str = (const char *) data;
|
3411 |
|
|
|
3412 |
|
|
if ((unsigned HOST_WIDE_INT) offset > strlen (str))
|
3413 |
|
|
return const0_rtx;
|
3414 |
|
|
|
3415 |
|
|
return c_readstr (str + offset, mode);
|
3416 |
|
|
}
|
3417 |
|
|
|
3418 |
|
|
/* Expand expression EXP, which is a call to the strncpy builtin. Return
|
3419 |
|
|
NULL_RTX if we failed the caller should emit a normal call. */
|
3420 |
|
|
|
3421 |
|
|
static rtx
|
3422 |
|
|
expand_builtin_strncpy (tree exp, rtx target)
|
3423 |
|
|
{
|
3424 |
|
|
location_t loc = EXPR_LOCATION (exp);
|
3425 |
|
|
|
3426 |
|
|
if (validate_arglist (exp,
|
3427 |
|
|
POINTER_TYPE, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
3428 |
|
|
{
|
3429 |
|
|
tree dest = CALL_EXPR_ARG (exp, 0);
|
3430 |
|
|
tree src = CALL_EXPR_ARG (exp, 1);
|
3431 |
|
|
tree len = CALL_EXPR_ARG (exp, 2);
|
3432 |
|
|
tree slen = c_strlen (src, 1);
|
3433 |
|
|
|
3434 |
|
|
/* We must be passed a constant len and src parameter. */
|
3435 |
|
|
if (!host_integerp (len, 1) || !slen || !host_integerp (slen, 1))
|
3436 |
|
|
return NULL_RTX;
|
3437 |
|
|
|
3438 |
|
|
slen = size_binop_loc (loc, PLUS_EXPR, slen, ssize_int (1));
|
3439 |
|
|
|
3440 |
|
|
/* We're required to pad with trailing zeros if the requested
|
3441 |
|
|
len is greater than strlen(s2)+1. In that case try to
|
3442 |
|
|
use store_by_pieces, if it fails, punt. */
|
3443 |
|
|
if (tree_int_cst_lt (slen, len))
|
3444 |
|
|
{
|
3445 |
|
|
unsigned int dest_align = get_pointer_alignment (dest);
|
3446 |
|
|
const char *p = c_getstr (src);
|
3447 |
|
|
rtx dest_mem;
|
3448 |
|
|
|
3449 |
|
|
if (!p || dest_align == 0 || !host_integerp (len, 1)
|
3450 |
|
|
|| !can_store_by_pieces (tree_low_cst (len, 1),
|
3451 |
|
|
builtin_strncpy_read_str,
|
3452 |
|
|
CONST_CAST (char *, p),
|
3453 |
|
|
dest_align, false))
|
3454 |
|
|
return NULL_RTX;
|
3455 |
|
|
|
3456 |
|
|
dest_mem = get_memory_rtx (dest, len);
|
3457 |
|
|
store_by_pieces (dest_mem, tree_low_cst (len, 1),
|
3458 |
|
|
builtin_strncpy_read_str,
|
3459 |
|
|
CONST_CAST (char *, p), dest_align, false, 0);
|
3460 |
|
|
dest_mem = force_operand (XEXP (dest_mem, 0), target);
|
3461 |
|
|
dest_mem = convert_memory_address (ptr_mode, dest_mem);
|
3462 |
|
|
return dest_mem;
|
3463 |
|
|
}
|
3464 |
|
|
}
|
3465 |
|
|
return NULL_RTX;
|
3466 |
|
|
}
|
3467 |
|
|
|
3468 |
|
|
/* Callback routine for store_by_pieces. Read GET_MODE_BITSIZE (MODE)
|
3469 |
|
|
bytes from constant string DATA + OFFSET and return it as target
|
3470 |
|
|
constant. */
|
3471 |
|
|
|
3472 |
|
|
rtx
|
3473 |
|
|
builtin_memset_read_str (void *data, HOST_WIDE_INT offset ATTRIBUTE_UNUSED,
|
3474 |
|
|
enum machine_mode mode)
|
3475 |
|
|
{
|
3476 |
|
|
const char *c = (const char *) data;
|
3477 |
|
|
char *p = XALLOCAVEC (char, GET_MODE_SIZE (mode));
|
3478 |
|
|
|
3479 |
|
|
memset (p, *c, GET_MODE_SIZE (mode));
|
3480 |
|
|
|
3481 |
|
|
return c_readstr (p, mode);
|
3482 |
|
|
}
|
3483 |
|
|
|
3484 |
|
|
/* Callback routine for store_by_pieces. Return the RTL of a register
|
3485 |
|
|
containing GET_MODE_SIZE (MODE) consecutive copies of the unsigned
|
3486 |
|
|
char value given in the RTL register data. For example, if mode is
|
3487 |
|
|
4 bytes wide, return the RTL for 0x01010101*data. */
|
3488 |
|
|
|
3489 |
|
|
static rtx
|
3490 |
|
|
builtin_memset_gen_str (void *data, HOST_WIDE_INT offset ATTRIBUTE_UNUSED,
|
3491 |
|
|
enum machine_mode mode)
|
3492 |
|
|
{
|
3493 |
|
|
rtx target, coeff;
|
3494 |
|
|
size_t size;
|
3495 |
|
|
char *p;
|
3496 |
|
|
|
3497 |
|
|
size = GET_MODE_SIZE (mode);
|
3498 |
|
|
if (size == 1)
|
3499 |
|
|
return (rtx) data;
|
3500 |
|
|
|
3501 |
|
|
p = XALLOCAVEC (char, size);
|
3502 |
|
|
memset (p, 1, size);
|
3503 |
|
|
coeff = c_readstr (p, mode);
|
3504 |
|
|
|
3505 |
|
|
target = convert_to_mode (mode, (rtx) data, 1);
|
3506 |
|
|
target = expand_mult (mode, target, coeff, NULL_RTX, 1);
|
3507 |
|
|
return force_reg (mode, target);
|
3508 |
|
|
}
|
3509 |
|
|
|
3510 |
|
|
/* Expand expression EXP, which is a call to the memset builtin. Return
|
3511 |
|
|
NULL_RTX if we failed the caller should emit a normal call, otherwise
|
3512 |
|
|
try to get the result in TARGET, if convenient (and in mode MODE if that's
|
3513 |
|
|
convenient). */
|
3514 |
|
|
|
3515 |
|
|
static rtx
|
3516 |
|
|
expand_builtin_memset (tree exp, rtx target, enum machine_mode mode)
|
3517 |
|
|
{
|
3518 |
|
|
if (!validate_arglist (exp,
|
3519 |
|
|
POINTER_TYPE, INTEGER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
3520 |
|
|
return NULL_RTX;
|
3521 |
|
|
else
|
3522 |
|
|
{
|
3523 |
|
|
tree dest = CALL_EXPR_ARG (exp, 0);
|
3524 |
|
|
tree val = CALL_EXPR_ARG (exp, 1);
|
3525 |
|
|
tree len = CALL_EXPR_ARG (exp, 2);
|
3526 |
|
|
return expand_builtin_memset_args (dest, val, len, target, mode, exp);
|
3527 |
|
|
}
|
3528 |
|
|
}
|
3529 |
|
|
|
3530 |
|
|
/* Helper function to do the actual work for expand_builtin_memset. The
|
3531 |
|
|
arguments to the builtin_memset call DEST, VAL, and LEN are broken out
|
3532 |
|
|
so that this can also be called without constructing an actual CALL_EXPR.
|
3533 |
|
|
The other arguments and return value are the same as for
|
3534 |
|
|
expand_builtin_memset. */
|
3535 |
|
|
|
3536 |
|
|
static rtx
|
3537 |
|
|
expand_builtin_memset_args (tree dest, tree val, tree len,
|
3538 |
|
|
rtx target, enum machine_mode mode, tree orig_exp)
|
3539 |
|
|
{
|
3540 |
|
|
tree fndecl, fn;
|
3541 |
|
|
enum built_in_function fcode;
|
3542 |
|
|
enum machine_mode val_mode;
|
3543 |
|
|
char c;
|
3544 |
|
|
unsigned int dest_align;
|
3545 |
|
|
rtx dest_mem, dest_addr, len_rtx;
|
3546 |
|
|
HOST_WIDE_INT expected_size = -1;
|
3547 |
|
|
unsigned int expected_align = 0;
|
3548 |
|
|
|
3549 |
|
|
dest_align = get_pointer_alignment (dest);
|
3550 |
|
|
|
3551 |
|
|
/* If DEST is not a pointer type, don't do this operation in-line. */
|
3552 |
|
|
if (dest_align == 0)
|
3553 |
|
|
return NULL_RTX;
|
3554 |
|
|
|
3555 |
|
|
if (currently_expanding_gimple_stmt)
|
3556 |
|
|
stringop_block_profile (currently_expanding_gimple_stmt,
|
3557 |
|
|
&expected_align, &expected_size);
|
3558 |
|
|
|
3559 |
|
|
if (expected_align < dest_align)
|
3560 |
|
|
expected_align = dest_align;
|
3561 |
|
|
|
3562 |
|
|
/* If the LEN parameter is zero, return DEST. */
|
3563 |
|
|
if (integer_zerop (len))
|
3564 |
|
|
{
|
3565 |
|
|
/* Evaluate and ignore VAL in case it has side-effects. */
|
3566 |
|
|
expand_expr (val, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
3567 |
|
|
return expand_expr (dest, target, mode, EXPAND_NORMAL);
|
3568 |
|
|
}
|
3569 |
|
|
|
3570 |
|
|
/* Stabilize the arguments in case we fail. */
|
3571 |
|
|
dest = builtin_save_expr (dest);
|
3572 |
|
|
val = builtin_save_expr (val);
|
3573 |
|
|
len = builtin_save_expr (len);
|
3574 |
|
|
|
3575 |
|
|
len_rtx = expand_normal (len);
|
3576 |
|
|
dest_mem = get_memory_rtx (dest, len);
|
3577 |
|
|
val_mode = TYPE_MODE (unsigned_char_type_node);
|
3578 |
|
|
|
3579 |
|
|
if (TREE_CODE (val) != INTEGER_CST)
|
3580 |
|
|
{
|
3581 |
|
|
rtx val_rtx;
|
3582 |
|
|
|
3583 |
|
|
val_rtx = expand_normal (val);
|
3584 |
|
|
val_rtx = convert_to_mode (val_mode, val_rtx, 0);
|
3585 |
|
|
|
3586 |
|
|
/* Assume that we can memset by pieces if we can store
|
3587 |
|
|
* the coefficients by pieces (in the required modes).
|
3588 |
|
|
* We can't pass builtin_memset_gen_str as that emits RTL. */
|
3589 |
|
|
c = 1;
|
3590 |
|
|
if (host_integerp (len, 1)
|
3591 |
|
|
&& can_store_by_pieces (tree_low_cst (len, 1),
|
3592 |
|
|
builtin_memset_read_str, &c, dest_align,
|
3593 |
|
|
true))
|
3594 |
|
|
{
|
3595 |
|
|
val_rtx = force_reg (val_mode, val_rtx);
|
3596 |
|
|
store_by_pieces (dest_mem, tree_low_cst (len, 1),
|
3597 |
|
|
builtin_memset_gen_str, val_rtx, dest_align,
|
3598 |
|
|
true, 0);
|
3599 |
|
|
}
|
3600 |
|
|
else if (!set_storage_via_setmem (dest_mem, len_rtx, val_rtx,
|
3601 |
|
|
dest_align, expected_align,
|
3602 |
|
|
expected_size))
|
3603 |
|
|
goto do_libcall;
|
3604 |
|
|
|
3605 |
|
|
dest_mem = force_operand (XEXP (dest_mem, 0), NULL_RTX);
|
3606 |
|
|
dest_mem = convert_memory_address (ptr_mode, dest_mem);
|
3607 |
|
|
return dest_mem;
|
3608 |
|
|
}
|
3609 |
|
|
|
3610 |
|
|
if (target_char_cast (val, &c))
|
3611 |
|
|
goto do_libcall;
|
3612 |
|
|
|
3613 |
|
|
if (c)
|
3614 |
|
|
{
|
3615 |
|
|
if (host_integerp (len, 1)
|
3616 |
|
|
&& can_store_by_pieces (tree_low_cst (len, 1),
|
3617 |
|
|
builtin_memset_read_str, &c, dest_align,
|
3618 |
|
|
true))
|
3619 |
|
|
store_by_pieces (dest_mem, tree_low_cst (len, 1),
|
3620 |
|
|
builtin_memset_read_str, &c, dest_align, true, 0);
|
3621 |
|
|
else if (!set_storage_via_setmem (dest_mem, len_rtx,
|
3622 |
|
|
gen_int_mode (c, val_mode),
|
3623 |
|
|
dest_align, expected_align,
|
3624 |
|
|
expected_size))
|
3625 |
|
|
goto do_libcall;
|
3626 |
|
|
|
3627 |
|
|
dest_mem = force_operand (XEXP (dest_mem, 0), NULL_RTX);
|
3628 |
|
|
dest_mem = convert_memory_address (ptr_mode, dest_mem);
|
3629 |
|
|
return dest_mem;
|
3630 |
|
|
}
|
3631 |
|
|
|
3632 |
|
|
set_mem_align (dest_mem, dest_align);
|
3633 |
|
|
dest_addr = clear_storage_hints (dest_mem, len_rtx,
|
3634 |
|
|
CALL_EXPR_TAILCALL (orig_exp)
|
3635 |
|
|
? BLOCK_OP_TAILCALL : BLOCK_OP_NORMAL,
|
3636 |
|
|
expected_align, expected_size);
|
3637 |
|
|
|
3638 |
|
|
if (dest_addr == 0)
|
3639 |
|
|
{
|
3640 |
|
|
dest_addr = force_operand (XEXP (dest_mem, 0), NULL_RTX);
|
3641 |
|
|
dest_addr = convert_memory_address (ptr_mode, dest_addr);
|
3642 |
|
|
}
|
3643 |
|
|
|
3644 |
|
|
return dest_addr;
|
3645 |
|
|
|
3646 |
|
|
do_libcall:
|
3647 |
|
|
fndecl = get_callee_fndecl (orig_exp);
|
3648 |
|
|
fcode = DECL_FUNCTION_CODE (fndecl);
|
3649 |
|
|
if (fcode == BUILT_IN_MEMSET)
|
3650 |
|
|
fn = build_call_nofold_loc (EXPR_LOCATION (orig_exp), fndecl, 3,
|
3651 |
|
|
dest, val, len);
|
3652 |
|
|
else if (fcode == BUILT_IN_BZERO)
|
3653 |
|
|
fn = build_call_nofold_loc (EXPR_LOCATION (orig_exp), fndecl, 2,
|
3654 |
|
|
dest, len);
|
3655 |
|
|
else
|
3656 |
|
|
gcc_unreachable ();
|
3657 |
|
|
gcc_assert (TREE_CODE (fn) == CALL_EXPR);
|
3658 |
|
|
CALL_EXPR_TAILCALL (fn) = CALL_EXPR_TAILCALL (orig_exp);
|
3659 |
|
|
return expand_call (fn, target, target == const0_rtx);
|
3660 |
|
|
}
|
3661 |
|
|
|
3662 |
|
|
/* Expand expression EXP, which is a call to the bzero builtin. Return
|
3663 |
|
|
NULL_RTX if we failed the caller should emit a normal call. */
|
3664 |
|
|
|
3665 |
|
|
static rtx
|
3666 |
|
|
expand_builtin_bzero (tree exp)
|
3667 |
|
|
{
|
3668 |
|
|
tree dest, size;
|
3669 |
|
|
location_t loc = EXPR_LOCATION (exp);
|
3670 |
|
|
|
3671 |
|
|
if (!validate_arglist (exp, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
3672 |
|
|
return NULL_RTX;
|
3673 |
|
|
|
3674 |
|
|
dest = CALL_EXPR_ARG (exp, 0);
|
3675 |
|
|
size = CALL_EXPR_ARG (exp, 1);
|
3676 |
|
|
|
3677 |
|
|
/* New argument list transforming bzero(ptr x, int y) to
|
3678 |
|
|
memset(ptr x, int 0, size_t y). This is done this way
|
3679 |
|
|
so that if it isn't expanded inline, we fallback to
|
3680 |
|
|
calling bzero instead of memset. */
|
3681 |
|
|
|
3682 |
|
|
return expand_builtin_memset_args (dest, integer_zero_node,
|
3683 |
|
|
fold_convert_loc (loc,
|
3684 |
|
|
size_type_node, size),
|
3685 |
|
|
const0_rtx, VOIDmode, exp);
|
3686 |
|
|
}
|
3687 |
|
|
|
3688 |
|
|
/* Expand expression EXP, which is a call to the memcmp built-in function.
|
3689 |
|
|
Return NULL_RTX if we failed and the caller should emit a normal call,
|
3690 |
|
|
otherwise try to get the result in TARGET, if convenient (and in mode
|
3691 |
|
|
MODE, if that's convenient). */
|
3692 |
|
|
|
3693 |
|
|
static rtx
|
3694 |
|
|
expand_builtin_memcmp (tree exp, ATTRIBUTE_UNUSED rtx target,
|
3695 |
|
|
ATTRIBUTE_UNUSED enum machine_mode mode)
|
3696 |
|
|
{
|
3697 |
|
|
location_t loc ATTRIBUTE_UNUSED = EXPR_LOCATION (exp);
|
3698 |
|
|
|
3699 |
|
|
if (!validate_arglist (exp,
|
3700 |
|
|
POINTER_TYPE, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
3701 |
|
|
return NULL_RTX;
|
3702 |
|
|
|
3703 |
|
|
/* Note: The cmpstrnsi pattern, if it exists, is not suitable for
|
3704 |
|
|
implementing memcmp because it will stop if it encounters two
|
3705 |
|
|
zero bytes. */
|
3706 |
|
|
#if defined HAVE_cmpmemsi
|
3707 |
|
|
{
|
3708 |
|
|
rtx arg1_rtx, arg2_rtx, arg3_rtx;
|
3709 |
|
|
rtx result;
|
3710 |
|
|
rtx insn;
|
3711 |
|
|
tree arg1 = CALL_EXPR_ARG (exp, 0);
|
3712 |
|
|
tree arg2 = CALL_EXPR_ARG (exp, 1);
|
3713 |
|
|
tree len = CALL_EXPR_ARG (exp, 2);
|
3714 |
|
|
|
3715 |
|
|
unsigned int arg1_align = get_pointer_alignment (arg1) / BITS_PER_UNIT;
|
3716 |
|
|
unsigned int arg2_align = get_pointer_alignment (arg2) / BITS_PER_UNIT;
|
3717 |
|
|
enum machine_mode insn_mode;
|
3718 |
|
|
|
3719 |
|
|
if (HAVE_cmpmemsi)
|
3720 |
|
|
insn_mode = insn_data[(int) CODE_FOR_cmpmemsi].operand[0].mode;
|
3721 |
|
|
else
|
3722 |
|
|
return NULL_RTX;
|
3723 |
|
|
|
3724 |
|
|
/* If we don't have POINTER_TYPE, call the function. */
|
3725 |
|
|
if (arg1_align == 0 || arg2_align == 0)
|
3726 |
|
|
return NULL_RTX;
|
3727 |
|
|
|
3728 |
|
|
/* Make a place to write the result of the instruction. */
|
3729 |
|
|
result = target;
|
3730 |
|
|
if (! (result != 0
|
3731 |
|
|
&& REG_P (result) && GET_MODE (result) == insn_mode
|
3732 |
|
|
&& REGNO (result) >= FIRST_PSEUDO_REGISTER))
|
3733 |
|
|
result = gen_reg_rtx (insn_mode);
|
3734 |
|
|
|
3735 |
|
|
arg1_rtx = get_memory_rtx (arg1, len);
|
3736 |
|
|
arg2_rtx = get_memory_rtx (arg2, len);
|
3737 |
|
|
arg3_rtx = expand_normal (fold_convert_loc (loc, sizetype, len));
|
3738 |
|
|
|
3739 |
|
|
/* Set MEM_SIZE as appropriate. */
|
3740 |
|
|
if (CONST_INT_P (arg3_rtx))
|
3741 |
|
|
{
|
3742 |
|
|
set_mem_size (arg1_rtx, INTVAL (arg3_rtx));
|
3743 |
|
|
set_mem_size (arg2_rtx, INTVAL (arg3_rtx));
|
3744 |
|
|
}
|
3745 |
|
|
|
3746 |
|
|
if (HAVE_cmpmemsi)
|
3747 |
|
|
insn = gen_cmpmemsi (result, arg1_rtx, arg2_rtx, arg3_rtx,
|
3748 |
|
|
GEN_INT (MIN (arg1_align, arg2_align)));
|
3749 |
|
|
else
|
3750 |
|
|
gcc_unreachable ();
|
3751 |
|
|
|
3752 |
|
|
if (insn)
|
3753 |
|
|
emit_insn (insn);
|
3754 |
|
|
else
|
3755 |
|
|
emit_library_call_value (memcmp_libfunc, result, LCT_PURE,
|
3756 |
|
|
TYPE_MODE (integer_type_node), 3,
|
3757 |
|
|
XEXP (arg1_rtx, 0), Pmode,
|
3758 |
|
|
XEXP (arg2_rtx, 0), Pmode,
|
3759 |
|
|
convert_to_mode (TYPE_MODE (sizetype), arg3_rtx,
|
3760 |
|
|
TYPE_UNSIGNED (sizetype)),
|
3761 |
|
|
TYPE_MODE (sizetype));
|
3762 |
|
|
|
3763 |
|
|
/* Return the value in the proper mode for this function. */
|
3764 |
|
|
mode = TYPE_MODE (TREE_TYPE (exp));
|
3765 |
|
|
if (GET_MODE (result) == mode)
|
3766 |
|
|
return result;
|
3767 |
|
|
else if (target != 0)
|
3768 |
|
|
{
|
3769 |
|
|
convert_move (target, result, 0);
|
3770 |
|
|
return target;
|
3771 |
|
|
}
|
3772 |
|
|
else
|
3773 |
|
|
return convert_to_mode (mode, result, 0);
|
3774 |
|
|
}
|
3775 |
|
|
#endif /* HAVE_cmpmemsi. */
|
3776 |
|
|
|
3777 |
|
|
return NULL_RTX;
|
3778 |
|
|
}
|
3779 |
|
|
|
3780 |
|
|
/* Expand expression EXP, which is a call to the strcmp builtin. Return NULL_RTX
|
3781 |
|
|
if we failed the caller should emit a normal call, otherwise try to get
|
3782 |
|
|
the result in TARGET, if convenient. */
|
3783 |
|
|
|
3784 |
|
|
static rtx
|
3785 |
|
|
expand_builtin_strcmp (tree exp, ATTRIBUTE_UNUSED rtx target)
|
3786 |
|
|
{
|
3787 |
|
|
if (!validate_arglist (exp, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
3788 |
|
|
return NULL_RTX;
|
3789 |
|
|
|
3790 |
|
|
#if defined HAVE_cmpstrsi || defined HAVE_cmpstrnsi
|
3791 |
|
|
if (direct_optab_handler (cmpstr_optab, SImode) != CODE_FOR_nothing
|
3792 |
|
|
|| direct_optab_handler (cmpstrn_optab, SImode) != CODE_FOR_nothing)
|
3793 |
|
|
{
|
3794 |
|
|
rtx arg1_rtx, arg2_rtx;
|
3795 |
|
|
rtx result, insn = NULL_RTX;
|
3796 |
|
|
tree fndecl, fn;
|
3797 |
|
|
tree arg1 = CALL_EXPR_ARG (exp, 0);
|
3798 |
|
|
tree arg2 = CALL_EXPR_ARG (exp, 1);
|
3799 |
|
|
|
3800 |
|
|
unsigned int arg1_align = get_pointer_alignment (arg1) / BITS_PER_UNIT;
|
3801 |
|
|
unsigned int arg2_align = get_pointer_alignment (arg2) / BITS_PER_UNIT;
|
3802 |
|
|
|
3803 |
|
|
/* If we don't have POINTER_TYPE, call the function. */
|
3804 |
|
|
if (arg1_align == 0 || arg2_align == 0)
|
3805 |
|
|
return NULL_RTX;
|
3806 |
|
|
|
3807 |
|
|
/* Stabilize the arguments in case gen_cmpstr(n)si fail. */
|
3808 |
|
|
arg1 = builtin_save_expr (arg1);
|
3809 |
|
|
arg2 = builtin_save_expr (arg2);
|
3810 |
|
|
|
3811 |
|
|
arg1_rtx = get_memory_rtx (arg1, NULL);
|
3812 |
|
|
arg2_rtx = get_memory_rtx (arg2, NULL);
|
3813 |
|
|
|
3814 |
|
|
#ifdef HAVE_cmpstrsi
|
3815 |
|
|
/* Try to call cmpstrsi. */
|
3816 |
|
|
if (HAVE_cmpstrsi)
|
3817 |
|
|
{
|
3818 |
|
|
enum machine_mode insn_mode
|
3819 |
|
|
= insn_data[(int) CODE_FOR_cmpstrsi].operand[0].mode;
|
3820 |
|
|
|
3821 |
|
|
/* Make a place to write the result of the instruction. */
|
3822 |
|
|
result = target;
|
3823 |
|
|
if (! (result != 0
|
3824 |
|
|
&& REG_P (result) && GET_MODE (result) == insn_mode
|
3825 |
|
|
&& REGNO (result) >= FIRST_PSEUDO_REGISTER))
|
3826 |
|
|
result = gen_reg_rtx (insn_mode);
|
3827 |
|
|
|
3828 |
|
|
insn = gen_cmpstrsi (result, arg1_rtx, arg2_rtx,
|
3829 |
|
|
GEN_INT (MIN (arg1_align, arg2_align)));
|
3830 |
|
|
}
|
3831 |
|
|
#endif
|
3832 |
|
|
#ifdef HAVE_cmpstrnsi
|
3833 |
|
|
/* Try to determine at least one length and call cmpstrnsi. */
|
3834 |
|
|
if (!insn && HAVE_cmpstrnsi)
|
3835 |
|
|
{
|
3836 |
|
|
tree len;
|
3837 |
|
|
rtx arg3_rtx;
|
3838 |
|
|
|
3839 |
|
|
enum machine_mode insn_mode
|
3840 |
|
|
= insn_data[(int) CODE_FOR_cmpstrnsi].operand[0].mode;
|
3841 |
|
|
tree len1 = c_strlen (arg1, 1);
|
3842 |
|
|
tree len2 = c_strlen (arg2, 1);
|
3843 |
|
|
|
3844 |
|
|
if (len1)
|
3845 |
|
|
len1 = size_binop (PLUS_EXPR, ssize_int (1), len1);
|
3846 |
|
|
if (len2)
|
3847 |
|
|
len2 = size_binop (PLUS_EXPR, ssize_int (1), len2);
|
3848 |
|
|
|
3849 |
|
|
/* If we don't have a constant length for the first, use the length
|
3850 |
|
|
of the second, if we know it. We don't require a constant for
|
3851 |
|
|
this case; some cost analysis could be done if both are available
|
3852 |
|
|
but neither is constant. For now, assume they're equally cheap,
|
3853 |
|
|
unless one has side effects. If both strings have constant lengths,
|
3854 |
|
|
use the smaller. */
|
3855 |
|
|
|
3856 |
|
|
if (!len1)
|
3857 |
|
|
len = len2;
|
3858 |
|
|
else if (!len2)
|
3859 |
|
|
len = len1;
|
3860 |
|
|
else if (TREE_SIDE_EFFECTS (len1))
|
3861 |
|
|
len = len2;
|
3862 |
|
|
else if (TREE_SIDE_EFFECTS (len2))
|
3863 |
|
|
len = len1;
|
3864 |
|
|
else if (TREE_CODE (len1) != INTEGER_CST)
|
3865 |
|
|
len = len2;
|
3866 |
|
|
else if (TREE_CODE (len2) != INTEGER_CST)
|
3867 |
|
|
len = len1;
|
3868 |
|
|
else if (tree_int_cst_lt (len1, len2))
|
3869 |
|
|
len = len1;
|
3870 |
|
|
else
|
3871 |
|
|
len = len2;
|
3872 |
|
|
|
3873 |
|
|
/* If both arguments have side effects, we cannot optimize. */
|
3874 |
|
|
if (!len || TREE_SIDE_EFFECTS (len))
|
3875 |
|
|
goto do_libcall;
|
3876 |
|
|
|
3877 |
|
|
arg3_rtx = expand_normal (len);
|
3878 |
|
|
|
3879 |
|
|
/* Make a place to write the result of the instruction. */
|
3880 |
|
|
result = target;
|
3881 |
|
|
if (! (result != 0
|
3882 |
|
|
&& REG_P (result) && GET_MODE (result) == insn_mode
|
3883 |
|
|
&& REGNO (result) >= FIRST_PSEUDO_REGISTER))
|
3884 |
|
|
result = gen_reg_rtx (insn_mode);
|
3885 |
|
|
|
3886 |
|
|
insn = gen_cmpstrnsi (result, arg1_rtx, arg2_rtx, arg3_rtx,
|
3887 |
|
|
GEN_INT (MIN (arg1_align, arg2_align)));
|
3888 |
|
|
}
|
3889 |
|
|
#endif
|
3890 |
|
|
|
3891 |
|
|
if (insn)
|
3892 |
|
|
{
|
3893 |
|
|
enum machine_mode mode;
|
3894 |
|
|
emit_insn (insn);
|
3895 |
|
|
|
3896 |
|
|
/* Return the value in the proper mode for this function. */
|
3897 |
|
|
mode = TYPE_MODE (TREE_TYPE (exp));
|
3898 |
|
|
if (GET_MODE (result) == mode)
|
3899 |
|
|
return result;
|
3900 |
|
|
if (target == 0)
|
3901 |
|
|
return convert_to_mode (mode, result, 0);
|
3902 |
|
|
convert_move (target, result, 0);
|
3903 |
|
|
return target;
|
3904 |
|
|
}
|
3905 |
|
|
|
3906 |
|
|
/* Expand the library call ourselves using a stabilized argument
|
3907 |
|
|
list to avoid re-evaluating the function's arguments twice. */
|
3908 |
|
|
#ifdef HAVE_cmpstrnsi
|
3909 |
|
|
do_libcall:
|
3910 |
|
|
#endif
|
3911 |
|
|
fndecl = get_callee_fndecl (exp);
|
3912 |
|
|
fn = build_call_nofold_loc (EXPR_LOCATION (exp), fndecl, 2, arg1, arg2);
|
3913 |
|
|
gcc_assert (TREE_CODE (fn) == CALL_EXPR);
|
3914 |
|
|
CALL_EXPR_TAILCALL (fn) = CALL_EXPR_TAILCALL (exp);
|
3915 |
|
|
return expand_call (fn, target, target == const0_rtx);
|
3916 |
|
|
}
|
3917 |
|
|
#endif
|
3918 |
|
|
return NULL_RTX;
|
3919 |
|
|
}
|
3920 |
|
|
|
3921 |
|
|
/* Expand expression EXP, which is a call to the strncmp builtin. Return
|
3922 |
|
|
NULL_RTX if we failed the caller should emit a normal call, otherwise try to get
|
3923 |
|
|
the result in TARGET, if convenient. */
|
3924 |
|
|
|
3925 |
|
|
static rtx
|
3926 |
|
|
expand_builtin_strncmp (tree exp, ATTRIBUTE_UNUSED rtx target,
|
3927 |
|
|
ATTRIBUTE_UNUSED enum machine_mode mode)
|
3928 |
|
|
{
|
3929 |
|
|
location_t loc ATTRIBUTE_UNUSED = EXPR_LOCATION (exp);
|
3930 |
|
|
|
3931 |
|
|
if (!validate_arglist (exp,
|
3932 |
|
|
POINTER_TYPE, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
3933 |
|
|
return NULL_RTX;
|
3934 |
|
|
|
3935 |
|
|
/* If c_strlen can determine an expression for one of the string
|
3936 |
|
|
lengths, and it doesn't have side effects, then emit cmpstrnsi
|
3937 |
|
|
using length MIN(strlen(string)+1, arg3). */
|
3938 |
|
|
#ifdef HAVE_cmpstrnsi
|
3939 |
|
|
if (HAVE_cmpstrnsi)
|
3940 |
|
|
{
|
3941 |
|
|
tree len, len1, len2;
|
3942 |
|
|
rtx arg1_rtx, arg2_rtx, arg3_rtx;
|
3943 |
|
|
rtx result, insn;
|
3944 |
|
|
tree fndecl, fn;
|
3945 |
|
|
tree arg1 = CALL_EXPR_ARG (exp, 0);
|
3946 |
|
|
tree arg2 = CALL_EXPR_ARG (exp, 1);
|
3947 |
|
|
tree arg3 = CALL_EXPR_ARG (exp, 2);
|
3948 |
|
|
|
3949 |
|
|
unsigned int arg1_align = get_pointer_alignment (arg1) / BITS_PER_UNIT;
|
3950 |
|
|
unsigned int arg2_align = get_pointer_alignment (arg2) / BITS_PER_UNIT;
|
3951 |
|
|
enum machine_mode insn_mode
|
3952 |
|
|
= insn_data[(int) CODE_FOR_cmpstrnsi].operand[0].mode;
|
3953 |
|
|
|
3954 |
|
|
len1 = c_strlen (arg1, 1);
|
3955 |
|
|
len2 = c_strlen (arg2, 1);
|
3956 |
|
|
|
3957 |
|
|
if (len1)
|
3958 |
|
|
len1 = size_binop_loc (loc, PLUS_EXPR, ssize_int (1), len1);
|
3959 |
|
|
if (len2)
|
3960 |
|
|
len2 = size_binop_loc (loc, PLUS_EXPR, ssize_int (1), len2);
|
3961 |
|
|
|
3962 |
|
|
/* If we don't have a constant length for the first, use the length
|
3963 |
|
|
of the second, if we know it. We don't require a constant for
|
3964 |
|
|
this case; some cost analysis could be done if both are available
|
3965 |
|
|
but neither is constant. For now, assume they're equally cheap,
|
3966 |
|
|
unless one has side effects. If both strings have constant lengths,
|
3967 |
|
|
use the smaller. */
|
3968 |
|
|
|
3969 |
|
|
if (!len1)
|
3970 |
|
|
len = len2;
|
3971 |
|
|
else if (!len2)
|
3972 |
|
|
len = len1;
|
3973 |
|
|
else if (TREE_SIDE_EFFECTS (len1))
|
3974 |
|
|
len = len2;
|
3975 |
|
|
else if (TREE_SIDE_EFFECTS (len2))
|
3976 |
|
|
len = len1;
|
3977 |
|
|
else if (TREE_CODE (len1) != INTEGER_CST)
|
3978 |
|
|
len = len2;
|
3979 |
|
|
else if (TREE_CODE (len2) != INTEGER_CST)
|
3980 |
|
|
len = len1;
|
3981 |
|
|
else if (tree_int_cst_lt (len1, len2))
|
3982 |
|
|
len = len1;
|
3983 |
|
|
else
|
3984 |
|
|
len = len2;
|
3985 |
|
|
|
3986 |
|
|
/* If both arguments have side effects, we cannot optimize. */
|
3987 |
|
|
if (!len || TREE_SIDE_EFFECTS (len))
|
3988 |
|
|
return NULL_RTX;
|
3989 |
|
|
|
3990 |
|
|
/* The actual new length parameter is MIN(len,arg3). */
|
3991 |
|
|
len = fold_build2_loc (loc, MIN_EXPR, TREE_TYPE (len), len,
|
3992 |
|
|
fold_convert_loc (loc, TREE_TYPE (len), arg3));
|
3993 |
|
|
|
3994 |
|
|
/* If we don't have POINTER_TYPE, call the function. */
|
3995 |
|
|
if (arg1_align == 0 || arg2_align == 0)
|
3996 |
|
|
return NULL_RTX;
|
3997 |
|
|
|
3998 |
|
|
/* Make a place to write the result of the instruction. */
|
3999 |
|
|
result = target;
|
4000 |
|
|
if (! (result != 0
|
4001 |
|
|
&& REG_P (result) && GET_MODE (result) == insn_mode
|
4002 |
|
|
&& REGNO (result) >= FIRST_PSEUDO_REGISTER))
|
4003 |
|
|
result = gen_reg_rtx (insn_mode);
|
4004 |
|
|
|
4005 |
|
|
/* Stabilize the arguments in case gen_cmpstrnsi fails. */
|
4006 |
|
|
arg1 = builtin_save_expr (arg1);
|
4007 |
|
|
arg2 = builtin_save_expr (arg2);
|
4008 |
|
|
len = builtin_save_expr (len);
|
4009 |
|
|
|
4010 |
|
|
arg1_rtx = get_memory_rtx (arg1, len);
|
4011 |
|
|
arg2_rtx = get_memory_rtx (arg2, len);
|
4012 |
|
|
arg3_rtx = expand_normal (len);
|
4013 |
|
|
insn = gen_cmpstrnsi (result, arg1_rtx, arg2_rtx, arg3_rtx,
|
4014 |
|
|
GEN_INT (MIN (arg1_align, arg2_align)));
|
4015 |
|
|
if (insn)
|
4016 |
|
|
{
|
4017 |
|
|
emit_insn (insn);
|
4018 |
|
|
|
4019 |
|
|
/* Return the value in the proper mode for this function. */
|
4020 |
|
|
mode = TYPE_MODE (TREE_TYPE (exp));
|
4021 |
|
|
if (GET_MODE (result) == mode)
|
4022 |
|
|
return result;
|
4023 |
|
|
if (target == 0)
|
4024 |
|
|
return convert_to_mode (mode, result, 0);
|
4025 |
|
|
convert_move (target, result, 0);
|
4026 |
|
|
return target;
|
4027 |
|
|
}
|
4028 |
|
|
|
4029 |
|
|
/* Expand the library call ourselves using a stabilized argument
|
4030 |
|
|
list to avoid re-evaluating the function's arguments twice. */
|
4031 |
|
|
fndecl = get_callee_fndecl (exp);
|
4032 |
|
|
fn = build_call_nofold_loc (EXPR_LOCATION (exp), fndecl, 3,
|
4033 |
|
|
arg1, arg2, len);
|
4034 |
|
|
gcc_assert (TREE_CODE (fn) == CALL_EXPR);
|
4035 |
|
|
CALL_EXPR_TAILCALL (fn) = CALL_EXPR_TAILCALL (exp);
|
4036 |
|
|
return expand_call (fn, target, target == const0_rtx);
|
4037 |
|
|
}
|
4038 |
|
|
#endif
|
4039 |
|
|
return NULL_RTX;
|
4040 |
|
|
}
|
4041 |
|
|
|
4042 |
|
|
/* Expand a call to __builtin_saveregs, generating the result in TARGET,
|
4043 |
|
|
if that's convenient. */
|
4044 |
|
|
|
4045 |
|
|
rtx
|
4046 |
|
|
expand_builtin_saveregs (void)
|
4047 |
|
|
{
|
4048 |
|
|
rtx val, seq;
|
4049 |
|
|
|
4050 |
|
|
/* Don't do __builtin_saveregs more than once in a function.
|
4051 |
|
|
Save the result of the first call and reuse it. */
|
4052 |
|
|
if (saveregs_value != 0)
|
4053 |
|
|
return saveregs_value;
|
4054 |
|
|
|
4055 |
|
|
/* When this function is called, it means that registers must be
|
4056 |
|
|
saved on entry to this function. So we migrate the call to the
|
4057 |
|
|
first insn of this function. */
|
4058 |
|
|
|
4059 |
|
|
start_sequence ();
|
4060 |
|
|
|
4061 |
|
|
/* Do whatever the machine needs done in this case. */
|
4062 |
|
|
val = targetm.calls.expand_builtin_saveregs ();
|
4063 |
|
|
|
4064 |
|
|
seq = get_insns ();
|
4065 |
|
|
end_sequence ();
|
4066 |
|
|
|
4067 |
|
|
saveregs_value = val;
|
4068 |
|
|
|
4069 |
|
|
/* Put the insns after the NOTE that starts the function. If this
|
4070 |
|
|
is inside a start_sequence, make the outer-level insn chain current, so
|
4071 |
|
|
the code is placed at the start of the function. */
|
4072 |
|
|
push_topmost_sequence ();
|
4073 |
|
|
emit_insn_after (seq, entry_of_function ());
|
4074 |
|
|
pop_topmost_sequence ();
|
4075 |
|
|
|
4076 |
|
|
return val;
|
4077 |
|
|
}
|
4078 |
|
|
|
4079 |
|
|
/* Expand a call to __builtin_next_arg. */
|
4080 |
|
|
|
4081 |
|
|
static rtx
|
4082 |
|
|
expand_builtin_next_arg (void)
|
4083 |
|
|
{
|
4084 |
|
|
/* Checking arguments is already done in fold_builtin_next_arg
|
4085 |
|
|
that must be called before this function. */
|
4086 |
|
|
return expand_binop (ptr_mode, add_optab,
|
4087 |
|
|
crtl->args.internal_arg_pointer,
|
4088 |
|
|
crtl->args.arg_offset_rtx,
|
4089 |
|
|
NULL_RTX, 0, OPTAB_LIB_WIDEN);
|
4090 |
|
|
}
|
4091 |
|
|
|
4092 |
|
|
/* Make it easier for the backends by protecting the valist argument
|
4093 |
|
|
from multiple evaluations. */
|
4094 |
|
|
|
4095 |
|
|
static tree
|
4096 |
|
|
stabilize_va_list_loc (location_t loc, tree valist, int needs_lvalue)
|
4097 |
|
|
{
|
4098 |
|
|
tree vatype = targetm.canonical_va_list_type (TREE_TYPE (valist));
|
4099 |
|
|
|
4100 |
|
|
/* The current way of determining the type of valist is completely
|
4101 |
|
|
bogus. We should have the information on the va builtin instead. */
|
4102 |
|
|
if (!vatype)
|
4103 |
|
|
vatype = targetm.fn_abi_va_list (cfun->decl);
|
4104 |
|
|
|
4105 |
|
|
if (TREE_CODE (vatype) == ARRAY_TYPE)
|
4106 |
|
|
{
|
4107 |
|
|
if (TREE_SIDE_EFFECTS (valist))
|
4108 |
|
|
valist = save_expr (valist);
|
4109 |
|
|
|
4110 |
|
|
/* For this case, the backends will be expecting a pointer to
|
4111 |
|
|
vatype, but it's possible we've actually been given an array
|
4112 |
|
|
(an actual TARGET_CANONICAL_VA_LIST_TYPE (valist)).
|
4113 |
|
|
So fix it. */
|
4114 |
|
|
if (TREE_CODE (TREE_TYPE (valist)) == ARRAY_TYPE)
|
4115 |
|
|
{
|
4116 |
|
|
tree p1 = build_pointer_type (TREE_TYPE (vatype));
|
4117 |
|
|
valist = build_fold_addr_expr_with_type_loc (loc, valist, p1);
|
4118 |
|
|
}
|
4119 |
|
|
}
|
4120 |
|
|
else
|
4121 |
|
|
{
|
4122 |
|
|
tree pt = build_pointer_type (vatype);
|
4123 |
|
|
|
4124 |
|
|
if (! needs_lvalue)
|
4125 |
|
|
{
|
4126 |
|
|
if (! TREE_SIDE_EFFECTS (valist))
|
4127 |
|
|
return valist;
|
4128 |
|
|
|
4129 |
|
|
valist = fold_build1_loc (loc, ADDR_EXPR, pt, valist);
|
4130 |
|
|
TREE_SIDE_EFFECTS (valist) = 1;
|
4131 |
|
|
}
|
4132 |
|
|
|
4133 |
|
|
if (TREE_SIDE_EFFECTS (valist))
|
4134 |
|
|
valist = save_expr (valist);
|
4135 |
|
|
valist = fold_build2_loc (loc, MEM_REF,
|
4136 |
|
|
vatype, valist, build_int_cst (pt, 0));
|
4137 |
|
|
}
|
4138 |
|
|
|
4139 |
|
|
return valist;
|
4140 |
|
|
}
|
4141 |
|
|
|
4142 |
|
|
/* The "standard" definition of va_list is void*. */
|
4143 |
|
|
|
4144 |
|
|
tree
|
4145 |
|
|
std_build_builtin_va_list (void)
|
4146 |
|
|
{
|
4147 |
|
|
return ptr_type_node;
|
4148 |
|
|
}
|
4149 |
|
|
|
4150 |
|
|
/* The "standard" abi va_list is va_list_type_node. */
|
4151 |
|
|
|
4152 |
|
|
tree
|
4153 |
|
|
std_fn_abi_va_list (tree fndecl ATTRIBUTE_UNUSED)
|
4154 |
|
|
{
|
4155 |
|
|
return va_list_type_node;
|
4156 |
|
|
}
|
4157 |
|
|
|
4158 |
|
|
/* The "standard" type of va_list is va_list_type_node. */
|
4159 |
|
|
|
4160 |
|
|
tree
|
4161 |
|
|
std_canonical_va_list_type (tree type)
|
4162 |
|
|
{
|
4163 |
|
|
tree wtype, htype;
|
4164 |
|
|
|
4165 |
|
|
if (INDIRECT_REF_P (type))
|
4166 |
|
|
type = TREE_TYPE (type);
|
4167 |
|
|
else if (POINTER_TYPE_P (type) && POINTER_TYPE_P (TREE_TYPE(type)))
|
4168 |
|
|
type = TREE_TYPE (type);
|
4169 |
|
|
wtype = va_list_type_node;
|
4170 |
|
|
htype = type;
|
4171 |
|
|
/* Treat structure va_list types. */
|
4172 |
|
|
if (TREE_CODE (wtype) == RECORD_TYPE && POINTER_TYPE_P (htype))
|
4173 |
|
|
htype = TREE_TYPE (htype);
|
4174 |
|
|
else if (TREE_CODE (wtype) == ARRAY_TYPE)
|
4175 |
|
|
{
|
4176 |
|
|
/* If va_list is an array type, the argument may have decayed
|
4177 |
|
|
to a pointer type, e.g. by being passed to another function.
|
4178 |
|
|
In that case, unwrap both types so that we can compare the
|
4179 |
|
|
underlying records. */
|
4180 |
|
|
if (TREE_CODE (htype) == ARRAY_TYPE
|
4181 |
|
|
|| POINTER_TYPE_P (htype))
|
4182 |
|
|
{
|
4183 |
|
|
wtype = TREE_TYPE (wtype);
|
4184 |
|
|
htype = TREE_TYPE (htype);
|
4185 |
|
|
}
|
4186 |
|
|
}
|
4187 |
|
|
if (TYPE_MAIN_VARIANT (wtype) == TYPE_MAIN_VARIANT (htype))
|
4188 |
|
|
return va_list_type_node;
|
4189 |
|
|
|
4190 |
|
|
return NULL_TREE;
|
4191 |
|
|
}
|
4192 |
|
|
|
4193 |
|
|
/* The "standard" implementation of va_start: just assign `nextarg' to
|
4194 |
|
|
the variable. */
|
4195 |
|
|
|
4196 |
|
|
void
|
4197 |
|
|
std_expand_builtin_va_start (tree valist, rtx nextarg)
|
4198 |
|
|
{
|
4199 |
|
|
rtx va_r = expand_expr (valist, NULL_RTX, VOIDmode, EXPAND_WRITE);
|
4200 |
|
|
convert_move (va_r, nextarg, 0);
|
4201 |
|
|
}
|
4202 |
|
|
|
4203 |
|
|
/* Expand EXP, a call to __builtin_va_start. */
|
4204 |
|
|
|
4205 |
|
|
static rtx
|
4206 |
|
|
expand_builtin_va_start (tree exp)
|
4207 |
|
|
{
|
4208 |
|
|
rtx nextarg;
|
4209 |
|
|
tree valist;
|
4210 |
|
|
location_t loc = EXPR_LOCATION (exp);
|
4211 |
|
|
|
4212 |
|
|
if (call_expr_nargs (exp) < 2)
|
4213 |
|
|
{
|
4214 |
|
|
error_at (loc, "too few arguments to function %<va_start%>");
|
4215 |
|
|
return const0_rtx;
|
4216 |
|
|
}
|
4217 |
|
|
|
4218 |
|
|
if (fold_builtin_next_arg (exp, true))
|
4219 |
|
|
return const0_rtx;
|
4220 |
|
|
|
4221 |
|
|
nextarg = expand_builtin_next_arg ();
|
4222 |
|
|
valist = stabilize_va_list_loc (loc, CALL_EXPR_ARG (exp, 0), 1);
|
4223 |
|
|
|
4224 |
|
|
if (targetm.expand_builtin_va_start)
|
4225 |
|
|
targetm.expand_builtin_va_start (valist, nextarg);
|
4226 |
|
|
else
|
4227 |
|
|
std_expand_builtin_va_start (valist, nextarg);
|
4228 |
|
|
|
4229 |
|
|
return const0_rtx;
|
4230 |
|
|
}
|
4231 |
|
|
|
4232 |
|
|
/* The "standard" implementation of va_arg: read the value from the
|
4233 |
|
|
current (padded) address and increment by the (padded) size. */
|
4234 |
|
|
|
4235 |
|
|
tree
|
4236 |
|
|
std_gimplify_va_arg_expr (tree valist, tree type, gimple_seq *pre_p,
|
4237 |
|
|
gimple_seq *post_p)
|
4238 |
|
|
{
|
4239 |
|
|
tree addr, t, type_size, rounded_size, valist_tmp;
|
4240 |
|
|
unsigned HOST_WIDE_INT align, boundary;
|
4241 |
|
|
bool indirect;
|
4242 |
|
|
|
4243 |
|
|
#ifdef ARGS_GROW_DOWNWARD
|
4244 |
|
|
/* All of the alignment and movement below is for args-grow-up machines.
|
4245 |
|
|
As of 2004, there are only 3 ARGS_GROW_DOWNWARD targets, and they all
|
4246 |
|
|
implement their own specialized gimplify_va_arg_expr routines. */
|
4247 |
|
|
gcc_unreachable ();
|
4248 |
|
|
#endif
|
4249 |
|
|
|
4250 |
|
|
indirect = pass_by_reference (NULL, TYPE_MODE (type), type, false);
|
4251 |
|
|
if (indirect)
|
4252 |
|
|
type = build_pointer_type (type);
|
4253 |
|
|
|
4254 |
|
|
align = PARM_BOUNDARY / BITS_PER_UNIT;
|
4255 |
|
|
boundary = targetm.calls.function_arg_boundary (TYPE_MODE (type), type);
|
4256 |
|
|
|
4257 |
|
|
/* When we align parameter on stack for caller, if the parameter
|
4258 |
|
|
alignment is beyond MAX_SUPPORTED_STACK_ALIGNMENT, it will be
|
4259 |
|
|
aligned at MAX_SUPPORTED_STACK_ALIGNMENT. We will match callee
|
4260 |
|
|
here with caller. */
|
4261 |
|
|
if (boundary > MAX_SUPPORTED_STACK_ALIGNMENT)
|
4262 |
|
|
boundary = MAX_SUPPORTED_STACK_ALIGNMENT;
|
4263 |
|
|
|
4264 |
|
|
boundary /= BITS_PER_UNIT;
|
4265 |
|
|
|
4266 |
|
|
/* Hoist the valist value into a temporary for the moment. */
|
4267 |
|
|
valist_tmp = get_initialized_tmp_var (valist, pre_p, NULL);
|
4268 |
|
|
|
4269 |
|
|
/* va_list pointer is aligned to PARM_BOUNDARY. If argument actually
|
4270 |
|
|
requires greater alignment, we must perform dynamic alignment. */
|
4271 |
|
|
if (boundary > align
|
4272 |
|
|
&& !integer_zerop (TYPE_SIZE (type)))
|
4273 |
|
|
{
|
4274 |
|
|
t = build2 (MODIFY_EXPR, TREE_TYPE (valist), valist_tmp,
|
4275 |
|
|
fold_build_pointer_plus_hwi (valist_tmp, boundary - 1));
|
4276 |
|
|
gimplify_and_add (t, pre_p);
|
4277 |
|
|
|
4278 |
|
|
t = build2 (MODIFY_EXPR, TREE_TYPE (valist), valist_tmp,
|
4279 |
|
|
fold_build2 (BIT_AND_EXPR, TREE_TYPE (valist),
|
4280 |
|
|
valist_tmp,
|
4281 |
|
|
build_int_cst (TREE_TYPE (valist), -boundary)));
|
4282 |
|
|
gimplify_and_add (t, pre_p);
|
4283 |
|
|
}
|
4284 |
|
|
else
|
4285 |
|
|
boundary = align;
|
4286 |
|
|
|
4287 |
|
|
/* If the actual alignment is less than the alignment of the type,
|
4288 |
|
|
adjust the type accordingly so that we don't assume strict alignment
|
4289 |
|
|
when dereferencing the pointer. */
|
4290 |
|
|
boundary *= BITS_PER_UNIT;
|
4291 |
|
|
if (boundary < TYPE_ALIGN (type))
|
4292 |
|
|
{
|
4293 |
|
|
type = build_variant_type_copy (type);
|
4294 |
|
|
TYPE_ALIGN (type) = boundary;
|
4295 |
|
|
}
|
4296 |
|
|
|
4297 |
|
|
/* Compute the rounded size of the type. */
|
4298 |
|
|
type_size = size_in_bytes (type);
|
4299 |
|
|
rounded_size = round_up (type_size, align);
|
4300 |
|
|
|
4301 |
|
|
/* Reduce rounded_size so it's sharable with the postqueue. */
|
4302 |
|
|
gimplify_expr (&rounded_size, pre_p, post_p, is_gimple_val, fb_rvalue);
|
4303 |
|
|
|
4304 |
|
|
/* Get AP. */
|
4305 |
|
|
addr = valist_tmp;
|
4306 |
|
|
if (PAD_VARARGS_DOWN && !integer_zerop (rounded_size))
|
4307 |
|
|
{
|
4308 |
|
|
/* Small args are padded downward. */
|
4309 |
|
|
t = fold_build2_loc (input_location, GT_EXPR, sizetype,
|
4310 |
|
|
rounded_size, size_int (align));
|
4311 |
|
|
t = fold_build3 (COND_EXPR, sizetype, t, size_zero_node,
|
4312 |
|
|
size_binop (MINUS_EXPR, rounded_size, type_size));
|
4313 |
|
|
addr = fold_build_pointer_plus (addr, t);
|
4314 |
|
|
}
|
4315 |
|
|
|
4316 |
|
|
/* Compute new value for AP. */
|
4317 |
|
|
t = fold_build_pointer_plus (valist_tmp, rounded_size);
|
4318 |
|
|
t = build2 (MODIFY_EXPR, TREE_TYPE (valist), valist, t);
|
4319 |
|
|
gimplify_and_add (t, pre_p);
|
4320 |
|
|
|
4321 |
|
|
addr = fold_convert (build_pointer_type (type), addr);
|
4322 |
|
|
|
4323 |
|
|
if (indirect)
|
4324 |
|
|
addr = build_va_arg_indirect_ref (addr);
|
4325 |
|
|
|
4326 |
|
|
return build_va_arg_indirect_ref (addr);
|
4327 |
|
|
}
|
4328 |
|
|
|
4329 |
|
|
/* Build an indirect-ref expression over the given TREE, which represents a
|
4330 |
|
|
piece of a va_arg() expansion. */
|
4331 |
|
|
tree
|
4332 |
|
|
build_va_arg_indirect_ref (tree addr)
|
4333 |
|
|
{
|
4334 |
|
|
addr = build_simple_mem_ref_loc (EXPR_LOCATION (addr), addr);
|
4335 |
|
|
|
4336 |
|
|
if (flag_mudflap) /* Don't instrument va_arg INDIRECT_REF. */
|
4337 |
|
|
mf_mark (addr);
|
4338 |
|
|
|
4339 |
|
|
return addr;
|
4340 |
|
|
}
|
4341 |
|
|
|
4342 |
|
|
/* Return a dummy expression of type TYPE in order to keep going after an
|
4343 |
|
|
error. */
|
4344 |
|
|
|
4345 |
|
|
static tree
|
4346 |
|
|
dummy_object (tree type)
|
4347 |
|
|
{
|
4348 |
|
|
tree t = build_int_cst (build_pointer_type (type), 0);
|
4349 |
|
|
return build2 (MEM_REF, type, t, t);
|
4350 |
|
|
}
|
4351 |
|
|
|
4352 |
|
|
/* Gimplify __builtin_va_arg, aka VA_ARG_EXPR, which is not really a
|
4353 |
|
|
builtin function, but a very special sort of operator. */
|
4354 |
|
|
|
4355 |
|
|
enum gimplify_status
|
4356 |
|
|
gimplify_va_arg_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p)
|
4357 |
|
|
{
|
4358 |
|
|
tree promoted_type, have_va_type;
|
4359 |
|
|
tree valist = TREE_OPERAND (*expr_p, 0);
|
4360 |
|
|
tree type = TREE_TYPE (*expr_p);
|
4361 |
|
|
tree t;
|
4362 |
|
|
location_t loc = EXPR_LOCATION (*expr_p);
|
4363 |
|
|
|
4364 |
|
|
/* Verify that valist is of the proper type. */
|
4365 |
|
|
have_va_type = TREE_TYPE (valist);
|
4366 |
|
|
if (have_va_type == error_mark_node)
|
4367 |
|
|
return GS_ERROR;
|
4368 |
|
|
have_va_type = targetm.canonical_va_list_type (have_va_type);
|
4369 |
|
|
|
4370 |
|
|
if (have_va_type == NULL_TREE)
|
4371 |
|
|
{
|
4372 |
|
|
error_at (loc, "first argument to %<va_arg%> not of type %<va_list%>");
|
4373 |
|
|
return GS_ERROR;
|
4374 |
|
|
}
|
4375 |
|
|
|
4376 |
|
|
/* Generate a diagnostic for requesting data of a type that cannot
|
4377 |
|
|
be passed through `...' due to type promotion at the call site. */
|
4378 |
|
|
if ((promoted_type = lang_hooks.types.type_promotes_to (type))
|
4379 |
|
|
!= type)
|
4380 |
|
|
{
|
4381 |
|
|
static bool gave_help;
|
4382 |
|
|
bool warned;
|
4383 |
|
|
|
4384 |
|
|
/* Unfortunately, this is merely undefined, rather than a constraint
|
4385 |
|
|
violation, so we cannot make this an error. If this call is never
|
4386 |
|
|
executed, the program is still strictly conforming. */
|
4387 |
|
|
warned = warning_at (loc, 0,
|
4388 |
|
|
"%qT is promoted to %qT when passed through %<...%>",
|
4389 |
|
|
type, promoted_type);
|
4390 |
|
|
if (!gave_help && warned)
|
4391 |
|
|
{
|
4392 |
|
|
gave_help = true;
|
4393 |
|
|
inform (loc, "(so you should pass %qT not %qT to %<va_arg%>)",
|
4394 |
|
|
promoted_type, type);
|
4395 |
|
|
}
|
4396 |
|
|
|
4397 |
|
|
/* We can, however, treat "undefined" any way we please.
|
4398 |
|
|
Call abort to encourage the user to fix the program. */
|
4399 |
|
|
if (warned)
|
4400 |
|
|
inform (loc, "if this code is reached, the program will abort");
|
4401 |
|
|
/* Before the abort, allow the evaluation of the va_list
|
4402 |
|
|
expression to exit or longjmp. */
|
4403 |
|
|
gimplify_and_add (valist, pre_p);
|
4404 |
|
|
t = build_call_expr_loc (loc,
|
4405 |
|
|
builtin_decl_implicit (BUILT_IN_TRAP), 0);
|
4406 |
|
|
gimplify_and_add (t, pre_p);
|
4407 |
|
|
|
4408 |
|
|
/* This is dead code, but go ahead and finish so that the
|
4409 |
|
|
mode of the result comes out right. */
|
4410 |
|
|
*expr_p = dummy_object (type);
|
4411 |
|
|
return GS_ALL_DONE;
|
4412 |
|
|
}
|
4413 |
|
|
else
|
4414 |
|
|
{
|
4415 |
|
|
/* Make it easier for the backends by protecting the valist argument
|
4416 |
|
|
from multiple evaluations. */
|
4417 |
|
|
if (TREE_CODE (have_va_type) == ARRAY_TYPE)
|
4418 |
|
|
{
|
4419 |
|
|
/* For this case, the backends will be expecting a pointer to
|
4420 |
|
|
TREE_TYPE (abi), but it's possible we've
|
4421 |
|
|
actually been given an array (an actual TARGET_FN_ABI_VA_LIST).
|
4422 |
|
|
So fix it. */
|
4423 |
|
|
if (TREE_CODE (TREE_TYPE (valist)) == ARRAY_TYPE)
|
4424 |
|
|
{
|
4425 |
|
|
tree p1 = build_pointer_type (TREE_TYPE (have_va_type));
|
4426 |
|
|
valist = fold_convert_loc (loc, p1,
|
4427 |
|
|
build_fold_addr_expr_loc (loc, valist));
|
4428 |
|
|
}
|
4429 |
|
|
|
4430 |
|
|
gimplify_expr (&valist, pre_p, post_p, is_gimple_val, fb_rvalue);
|
4431 |
|
|
}
|
4432 |
|
|
else
|
4433 |
|
|
gimplify_expr (&valist, pre_p, post_p, is_gimple_min_lval, fb_lvalue);
|
4434 |
|
|
|
4435 |
|
|
if (!targetm.gimplify_va_arg_expr)
|
4436 |
|
|
/* FIXME: Once most targets are converted we should merely
|
4437 |
|
|
assert this is non-null. */
|
4438 |
|
|
return GS_ALL_DONE;
|
4439 |
|
|
|
4440 |
|
|
*expr_p = targetm.gimplify_va_arg_expr (valist, type, pre_p, post_p);
|
4441 |
|
|
return GS_OK;
|
4442 |
|
|
}
|
4443 |
|
|
}
|
4444 |
|
|
|
4445 |
|
|
/* Expand EXP, a call to __builtin_va_end. */
|
4446 |
|
|
|
4447 |
|
|
static rtx
|
4448 |
|
|
expand_builtin_va_end (tree exp)
|
4449 |
|
|
{
|
4450 |
|
|
tree valist = CALL_EXPR_ARG (exp, 0);
|
4451 |
|
|
|
4452 |
|
|
/* Evaluate for side effects, if needed. I hate macros that don't
|
4453 |
|
|
do that. */
|
4454 |
|
|
if (TREE_SIDE_EFFECTS (valist))
|
4455 |
|
|
expand_expr (valist, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
4456 |
|
|
|
4457 |
|
|
return const0_rtx;
|
4458 |
|
|
}
|
4459 |
|
|
|
4460 |
|
|
/* Expand EXP, a call to __builtin_va_copy. We do this as a
|
4461 |
|
|
builtin rather than just as an assignment in stdarg.h because of the
|
4462 |
|
|
nastiness of array-type va_list types. */
|
4463 |
|
|
|
4464 |
|
|
static rtx
|
4465 |
|
|
expand_builtin_va_copy (tree exp)
|
4466 |
|
|
{
|
4467 |
|
|
tree dst, src, t;
|
4468 |
|
|
location_t loc = EXPR_LOCATION (exp);
|
4469 |
|
|
|
4470 |
|
|
dst = CALL_EXPR_ARG (exp, 0);
|
4471 |
|
|
src = CALL_EXPR_ARG (exp, 1);
|
4472 |
|
|
|
4473 |
|
|
dst = stabilize_va_list_loc (loc, dst, 1);
|
4474 |
|
|
src = stabilize_va_list_loc (loc, src, 0);
|
4475 |
|
|
|
4476 |
|
|
gcc_assert (cfun != NULL && cfun->decl != NULL_TREE);
|
4477 |
|
|
|
4478 |
|
|
if (TREE_CODE (targetm.fn_abi_va_list (cfun->decl)) != ARRAY_TYPE)
|
4479 |
|
|
{
|
4480 |
|
|
t = build2 (MODIFY_EXPR, targetm.fn_abi_va_list (cfun->decl), dst, src);
|
4481 |
|
|
TREE_SIDE_EFFECTS (t) = 1;
|
4482 |
|
|
expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
4483 |
|
|
}
|
4484 |
|
|
else
|
4485 |
|
|
{
|
4486 |
|
|
rtx dstb, srcb, size;
|
4487 |
|
|
|
4488 |
|
|
/* Evaluate to pointers. */
|
4489 |
|
|
dstb = expand_expr (dst, NULL_RTX, Pmode, EXPAND_NORMAL);
|
4490 |
|
|
srcb = expand_expr (src, NULL_RTX, Pmode, EXPAND_NORMAL);
|
4491 |
|
|
size = expand_expr (TYPE_SIZE_UNIT (targetm.fn_abi_va_list (cfun->decl)),
|
4492 |
|
|
NULL_RTX, VOIDmode, EXPAND_NORMAL);
|
4493 |
|
|
|
4494 |
|
|
dstb = convert_memory_address (Pmode, dstb);
|
4495 |
|
|
srcb = convert_memory_address (Pmode, srcb);
|
4496 |
|
|
|
4497 |
|
|
/* "Dereference" to BLKmode memories. */
|
4498 |
|
|
dstb = gen_rtx_MEM (BLKmode, dstb);
|
4499 |
|
|
set_mem_alias_set (dstb, get_alias_set (TREE_TYPE (TREE_TYPE (dst))));
|
4500 |
|
|
set_mem_align (dstb, TYPE_ALIGN (targetm.fn_abi_va_list (cfun->decl)));
|
4501 |
|
|
srcb = gen_rtx_MEM (BLKmode, srcb);
|
4502 |
|
|
set_mem_alias_set (srcb, get_alias_set (TREE_TYPE (TREE_TYPE (src))));
|
4503 |
|
|
set_mem_align (srcb, TYPE_ALIGN (targetm.fn_abi_va_list (cfun->decl)));
|
4504 |
|
|
|
4505 |
|
|
/* Copy. */
|
4506 |
|
|
emit_block_move (dstb, srcb, size, BLOCK_OP_NORMAL);
|
4507 |
|
|
}
|
4508 |
|
|
|
4509 |
|
|
return const0_rtx;
|
4510 |
|
|
}
|
4511 |
|
|
|
4512 |
|
|
/* Expand a call to one of the builtin functions __builtin_frame_address or
|
4513 |
|
|
__builtin_return_address. */
|
4514 |
|
|
|
4515 |
|
|
static rtx
|
4516 |
|
|
expand_builtin_frame_address (tree fndecl, tree exp)
|
4517 |
|
|
{
|
4518 |
|
|
/* The argument must be a nonnegative integer constant.
|
4519 |
|
|
It counts the number of frames to scan up the stack.
|
4520 |
|
|
The value is the return address saved in that frame. */
|
4521 |
|
|
if (call_expr_nargs (exp) == 0)
|
4522 |
|
|
/* Warning about missing arg was already issued. */
|
4523 |
|
|
return const0_rtx;
|
4524 |
|
|
else if (! host_integerp (CALL_EXPR_ARG (exp, 0), 1))
|
4525 |
|
|
{
|
4526 |
|
|
if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_FRAME_ADDRESS)
|
4527 |
|
|
error ("invalid argument to %<__builtin_frame_address%>");
|
4528 |
|
|
else
|
4529 |
|
|
error ("invalid argument to %<__builtin_return_address%>");
|
4530 |
|
|
return const0_rtx;
|
4531 |
|
|
}
|
4532 |
|
|
else
|
4533 |
|
|
{
|
4534 |
|
|
rtx tem
|
4535 |
|
|
= expand_builtin_return_addr (DECL_FUNCTION_CODE (fndecl),
|
4536 |
|
|
tree_low_cst (CALL_EXPR_ARG (exp, 0), 1));
|
4537 |
|
|
|
4538 |
|
|
/* Some ports cannot access arbitrary stack frames. */
|
4539 |
|
|
if (tem == NULL)
|
4540 |
|
|
{
|
4541 |
|
|
if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_FRAME_ADDRESS)
|
4542 |
|
|
warning (0, "unsupported argument to %<__builtin_frame_address%>");
|
4543 |
|
|
else
|
4544 |
|
|
warning (0, "unsupported argument to %<__builtin_return_address%>");
|
4545 |
|
|
return const0_rtx;
|
4546 |
|
|
}
|
4547 |
|
|
|
4548 |
|
|
/* For __builtin_frame_address, return what we've got. */
|
4549 |
|
|
if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_FRAME_ADDRESS)
|
4550 |
|
|
return tem;
|
4551 |
|
|
|
4552 |
|
|
if (!REG_P (tem)
|
4553 |
|
|
&& ! CONSTANT_P (tem))
|
4554 |
|
|
tem = copy_to_mode_reg (Pmode, tem);
|
4555 |
|
|
return tem;
|
4556 |
|
|
}
|
4557 |
|
|
}
|
4558 |
|
|
|
4559 |
|
|
/* Expand EXP, a call to the alloca builtin. Return NULL_RTX if we
|
4560 |
|
|
failed and the caller should emit a normal call. CANNOT_ACCUMULATE
|
4561 |
|
|
is the same as for allocate_dynamic_stack_space. */
|
4562 |
|
|
|
4563 |
|
|
static rtx
|
4564 |
|
|
expand_builtin_alloca (tree exp, bool cannot_accumulate)
|
4565 |
|
|
{
|
4566 |
|
|
rtx op0;
|
4567 |
|
|
rtx result;
|
4568 |
|
|
bool valid_arglist;
|
4569 |
|
|
unsigned int align;
|
4570 |
|
|
bool alloca_with_align = (DECL_FUNCTION_CODE (get_callee_fndecl (exp))
|
4571 |
|
|
== BUILT_IN_ALLOCA_WITH_ALIGN);
|
4572 |
|
|
|
4573 |
|
|
/* Emit normal call if we use mudflap. */
|
4574 |
|
|
if (flag_mudflap)
|
4575 |
|
|
return NULL_RTX;
|
4576 |
|
|
|
4577 |
|
|
valid_arglist
|
4578 |
|
|
= (alloca_with_align
|
4579 |
|
|
? validate_arglist (exp, INTEGER_TYPE, INTEGER_TYPE, VOID_TYPE)
|
4580 |
|
|
: validate_arglist (exp, INTEGER_TYPE, VOID_TYPE));
|
4581 |
|
|
|
4582 |
|
|
if (!valid_arglist)
|
4583 |
|
|
return NULL_RTX;
|
4584 |
|
|
|
4585 |
|
|
/* Compute the argument. */
|
4586 |
|
|
op0 = expand_normal (CALL_EXPR_ARG (exp, 0));
|
4587 |
|
|
|
4588 |
|
|
/* Compute the alignment. */
|
4589 |
|
|
align = (alloca_with_align
|
4590 |
|
|
? TREE_INT_CST_LOW (CALL_EXPR_ARG (exp, 1))
|
4591 |
|
|
: BIGGEST_ALIGNMENT);
|
4592 |
|
|
|
4593 |
|
|
/* Allocate the desired space. */
|
4594 |
|
|
result = allocate_dynamic_stack_space (op0, 0, align, cannot_accumulate);
|
4595 |
|
|
result = convert_memory_address (ptr_mode, result);
|
4596 |
|
|
|
4597 |
|
|
return result;
|
4598 |
|
|
}
|
4599 |
|
|
|
4600 |
|
|
/* Expand a call to a bswap builtin with argument ARG0. MODE
|
4601 |
|
|
is the mode to expand with. */
|
4602 |
|
|
|
4603 |
|
|
static rtx
|
4604 |
|
|
expand_builtin_bswap (tree exp, rtx target, rtx subtarget)
|
4605 |
|
|
{
|
4606 |
|
|
enum machine_mode mode;
|
4607 |
|
|
tree arg;
|
4608 |
|
|
rtx op0;
|
4609 |
|
|
|
4610 |
|
|
if (!validate_arglist (exp, INTEGER_TYPE, VOID_TYPE))
|
4611 |
|
|
return NULL_RTX;
|
4612 |
|
|
|
4613 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
4614 |
|
|
mode = TYPE_MODE (TREE_TYPE (arg));
|
4615 |
|
|
op0 = expand_expr (arg, subtarget, VOIDmode, EXPAND_NORMAL);
|
4616 |
|
|
|
4617 |
|
|
target = expand_unop (mode, bswap_optab, op0, target, 1);
|
4618 |
|
|
|
4619 |
|
|
gcc_assert (target);
|
4620 |
|
|
|
4621 |
|
|
return convert_to_mode (mode, target, 0);
|
4622 |
|
|
}
|
4623 |
|
|
|
4624 |
|
|
/* Expand a call to a unary builtin in EXP.
|
4625 |
|
|
Return NULL_RTX if a normal call should be emitted rather than expanding the
|
4626 |
|
|
function in-line. If convenient, the result should be placed in TARGET.
|
4627 |
|
|
SUBTARGET may be used as the target for computing one of EXP's operands. */
|
4628 |
|
|
|
4629 |
|
|
static rtx
|
4630 |
|
|
expand_builtin_unop (enum machine_mode target_mode, tree exp, rtx target,
|
4631 |
|
|
rtx subtarget, optab op_optab)
|
4632 |
|
|
{
|
4633 |
|
|
rtx op0;
|
4634 |
|
|
|
4635 |
|
|
if (!validate_arglist (exp, INTEGER_TYPE, VOID_TYPE))
|
4636 |
|
|
return NULL_RTX;
|
4637 |
|
|
|
4638 |
|
|
/* Compute the argument. */
|
4639 |
|
|
op0 = expand_expr (CALL_EXPR_ARG (exp, 0),
|
4640 |
|
|
(subtarget
|
4641 |
|
|
&& (TYPE_MODE (TREE_TYPE (CALL_EXPR_ARG (exp, 0)))
|
4642 |
|
|
== GET_MODE (subtarget))) ? subtarget : NULL_RTX,
|
4643 |
|
|
VOIDmode, EXPAND_NORMAL);
|
4644 |
|
|
/* Compute op, into TARGET if possible.
|
4645 |
|
|
Set TARGET to wherever the result comes back. */
|
4646 |
|
|
target = expand_unop (TYPE_MODE (TREE_TYPE (CALL_EXPR_ARG (exp, 0))),
|
4647 |
|
|
op_optab, op0, target, op_optab != clrsb_optab);
|
4648 |
|
|
gcc_assert (target);
|
4649 |
|
|
|
4650 |
|
|
return convert_to_mode (target_mode, target, 0);
|
4651 |
|
|
}
|
4652 |
|
|
|
4653 |
|
|
/* Expand a call to __builtin_expect. We just return our argument
|
4654 |
|
|
as the builtin_expect semantic should've been already executed by
|
4655 |
|
|
tree branch prediction pass. */
|
4656 |
|
|
|
4657 |
|
|
static rtx
|
4658 |
|
|
expand_builtin_expect (tree exp, rtx target)
|
4659 |
|
|
{
|
4660 |
|
|
tree arg;
|
4661 |
|
|
|
4662 |
|
|
if (call_expr_nargs (exp) < 2)
|
4663 |
|
|
return const0_rtx;
|
4664 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
4665 |
|
|
|
4666 |
|
|
target = expand_expr (arg, target, VOIDmode, EXPAND_NORMAL);
|
4667 |
|
|
/* When guessing was done, the hints should be already stripped away. */
|
4668 |
|
|
gcc_assert (!flag_guess_branch_prob
|
4669 |
|
|
|| optimize == 0 || seen_error ());
|
4670 |
|
|
return target;
|
4671 |
|
|
}
|
4672 |
|
|
|
4673 |
|
|
/* Expand a call to __builtin_assume_aligned. We just return our first
|
4674 |
|
|
argument as the builtin_assume_aligned semantic should've been already
|
4675 |
|
|
executed by CCP. */
|
4676 |
|
|
|
4677 |
|
|
static rtx
|
4678 |
|
|
expand_builtin_assume_aligned (tree exp, rtx target)
|
4679 |
|
|
{
|
4680 |
|
|
if (call_expr_nargs (exp) < 2)
|
4681 |
|
|
return const0_rtx;
|
4682 |
|
|
target = expand_expr (CALL_EXPR_ARG (exp, 0), target, VOIDmode,
|
4683 |
|
|
EXPAND_NORMAL);
|
4684 |
|
|
gcc_assert (!TREE_SIDE_EFFECTS (CALL_EXPR_ARG (exp, 1))
|
4685 |
|
|
&& (call_expr_nargs (exp) < 3
|
4686 |
|
|
|| !TREE_SIDE_EFFECTS (CALL_EXPR_ARG (exp, 2))));
|
4687 |
|
|
return target;
|
4688 |
|
|
}
|
4689 |
|
|
|
4690 |
|
|
void
|
4691 |
|
|
expand_builtin_trap (void)
|
4692 |
|
|
{
|
4693 |
|
|
#ifdef HAVE_trap
|
4694 |
|
|
if (HAVE_trap)
|
4695 |
|
|
emit_insn (gen_trap ());
|
4696 |
|
|
else
|
4697 |
|
|
#endif
|
4698 |
|
|
emit_library_call (abort_libfunc, LCT_NORETURN, VOIDmode, 0);
|
4699 |
|
|
emit_barrier ();
|
4700 |
|
|
}
|
4701 |
|
|
|
4702 |
|
|
/* Expand a call to __builtin_unreachable. We do nothing except emit
|
4703 |
|
|
a barrier saying that control flow will not pass here.
|
4704 |
|
|
|
4705 |
|
|
It is the responsibility of the program being compiled to ensure
|
4706 |
|
|
that control flow does never reach __builtin_unreachable. */
|
4707 |
|
|
static void
|
4708 |
|
|
expand_builtin_unreachable (void)
|
4709 |
|
|
{
|
4710 |
|
|
emit_barrier ();
|
4711 |
|
|
}
|
4712 |
|
|
|
4713 |
|
|
/* Expand EXP, a call to fabs, fabsf or fabsl.
|
4714 |
|
|
Return NULL_RTX if a normal call should be emitted rather than expanding
|
4715 |
|
|
the function inline. If convenient, the result should be placed
|
4716 |
|
|
in TARGET. SUBTARGET may be used as the target for computing
|
4717 |
|
|
the operand. */
|
4718 |
|
|
|
4719 |
|
|
static rtx
|
4720 |
|
|
expand_builtin_fabs (tree exp, rtx target, rtx subtarget)
|
4721 |
|
|
{
|
4722 |
|
|
enum machine_mode mode;
|
4723 |
|
|
tree arg;
|
4724 |
|
|
rtx op0;
|
4725 |
|
|
|
4726 |
|
|
if (!validate_arglist (exp, REAL_TYPE, VOID_TYPE))
|
4727 |
|
|
return NULL_RTX;
|
4728 |
|
|
|
4729 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
4730 |
|
|
CALL_EXPR_ARG (exp, 0) = arg = builtin_save_expr (arg);
|
4731 |
|
|
mode = TYPE_MODE (TREE_TYPE (arg));
|
4732 |
|
|
op0 = expand_expr (arg, subtarget, VOIDmode, EXPAND_NORMAL);
|
4733 |
|
|
return expand_abs (mode, op0, target, 0, safe_from_p (target, arg, 1));
|
4734 |
|
|
}
|
4735 |
|
|
|
4736 |
|
|
/* Expand EXP, a call to copysign, copysignf, or copysignl.
|
4737 |
|
|
Return NULL is a normal call should be emitted rather than expanding the
|
4738 |
|
|
function inline. If convenient, the result should be placed in TARGET.
|
4739 |
|
|
SUBTARGET may be used as the target for computing the operand. */
|
4740 |
|
|
|
4741 |
|
|
static rtx
|
4742 |
|
|
expand_builtin_copysign (tree exp, rtx target, rtx subtarget)
|
4743 |
|
|
{
|
4744 |
|
|
rtx op0, op1;
|
4745 |
|
|
tree arg;
|
4746 |
|
|
|
4747 |
|
|
if (!validate_arglist (exp, REAL_TYPE, REAL_TYPE, VOID_TYPE))
|
4748 |
|
|
return NULL_RTX;
|
4749 |
|
|
|
4750 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
4751 |
|
|
op0 = expand_expr (arg, subtarget, VOIDmode, EXPAND_NORMAL);
|
4752 |
|
|
|
4753 |
|
|
arg = CALL_EXPR_ARG (exp, 1);
|
4754 |
|
|
op1 = expand_normal (arg);
|
4755 |
|
|
|
4756 |
|
|
return expand_copysign (op0, op1, target);
|
4757 |
|
|
}
|
4758 |
|
|
|
4759 |
|
|
/* Create a new constant string literal and return a char* pointer to it.
|
4760 |
|
|
The STRING_CST value is the LEN characters at STR. */
|
4761 |
|
|
tree
|
4762 |
|
|
build_string_literal (int len, const char *str)
|
4763 |
|
|
{
|
4764 |
|
|
tree t, elem, index, type;
|
4765 |
|
|
|
4766 |
|
|
t = build_string (len, str);
|
4767 |
|
|
elem = build_type_variant (char_type_node, 1, 0);
|
4768 |
|
|
index = build_index_type (size_int (len - 1));
|
4769 |
|
|
type = build_array_type (elem, index);
|
4770 |
|
|
TREE_TYPE (t) = type;
|
4771 |
|
|
TREE_CONSTANT (t) = 1;
|
4772 |
|
|
TREE_READONLY (t) = 1;
|
4773 |
|
|
TREE_STATIC (t) = 1;
|
4774 |
|
|
|
4775 |
|
|
type = build_pointer_type (elem);
|
4776 |
|
|
t = build1 (ADDR_EXPR, type,
|
4777 |
|
|
build4 (ARRAY_REF, elem,
|
4778 |
|
|
t, integer_zero_node, NULL_TREE, NULL_TREE));
|
4779 |
|
|
return t;
|
4780 |
|
|
}
|
4781 |
|
|
|
4782 |
|
|
/* Expand a call to __builtin___clear_cache. */
|
4783 |
|
|
|
4784 |
|
|
static rtx
|
4785 |
|
|
expand_builtin___clear_cache (tree exp ATTRIBUTE_UNUSED)
|
4786 |
|
|
{
|
4787 |
|
|
#ifndef HAVE_clear_cache
|
4788 |
|
|
#ifdef CLEAR_INSN_CACHE
|
4789 |
|
|
/* There is no "clear_cache" insn, and __clear_cache() in libgcc
|
4790 |
|
|
does something. Just do the default expansion to a call to
|
4791 |
|
|
__clear_cache(). */
|
4792 |
|
|
return NULL_RTX;
|
4793 |
|
|
#else
|
4794 |
|
|
/* There is no "clear_cache" insn, and __clear_cache() in libgcc
|
4795 |
|
|
does nothing. There is no need to call it. Do nothing. */
|
4796 |
|
|
return const0_rtx;
|
4797 |
|
|
#endif /* CLEAR_INSN_CACHE */
|
4798 |
|
|
#else
|
4799 |
|
|
/* We have a "clear_cache" insn, and it will handle everything. */
|
4800 |
|
|
tree begin, end;
|
4801 |
|
|
rtx begin_rtx, end_rtx;
|
4802 |
|
|
|
4803 |
|
|
/* We must not expand to a library call. If we did, any
|
4804 |
|
|
fallback library function in libgcc that might contain a call to
|
4805 |
|
|
__builtin___clear_cache() would recurse infinitely. */
|
4806 |
|
|
if (!validate_arglist (exp, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
4807 |
|
|
{
|
4808 |
|
|
error ("both arguments to %<__builtin___clear_cache%> must be pointers");
|
4809 |
|
|
return const0_rtx;
|
4810 |
|
|
}
|
4811 |
|
|
|
4812 |
|
|
if (HAVE_clear_cache)
|
4813 |
|
|
{
|
4814 |
|
|
struct expand_operand ops[2];
|
4815 |
|
|
|
4816 |
|
|
begin = CALL_EXPR_ARG (exp, 0);
|
4817 |
|
|
begin_rtx = expand_expr (begin, NULL_RTX, Pmode, EXPAND_NORMAL);
|
4818 |
|
|
|
4819 |
|
|
end = CALL_EXPR_ARG (exp, 1);
|
4820 |
|
|
end_rtx = expand_expr (end, NULL_RTX, Pmode, EXPAND_NORMAL);
|
4821 |
|
|
|
4822 |
|
|
create_address_operand (&ops[0], begin_rtx);
|
4823 |
|
|
create_address_operand (&ops[1], end_rtx);
|
4824 |
|
|
if (maybe_expand_insn (CODE_FOR_clear_cache, 2, ops))
|
4825 |
|
|
return const0_rtx;
|
4826 |
|
|
}
|
4827 |
|
|
return const0_rtx;
|
4828 |
|
|
#endif /* HAVE_clear_cache */
|
4829 |
|
|
}
|
4830 |
|
|
|
4831 |
|
|
/* Given a trampoline address, make sure it satisfies TRAMPOLINE_ALIGNMENT. */
|
4832 |
|
|
|
4833 |
|
|
static rtx
|
4834 |
|
|
round_trampoline_addr (rtx tramp)
|
4835 |
|
|
{
|
4836 |
|
|
rtx temp, addend, mask;
|
4837 |
|
|
|
4838 |
|
|
/* If we don't need too much alignment, we'll have been guaranteed
|
4839 |
|
|
proper alignment by get_trampoline_type. */
|
4840 |
|
|
if (TRAMPOLINE_ALIGNMENT <= STACK_BOUNDARY)
|
4841 |
|
|
return tramp;
|
4842 |
|
|
|
4843 |
|
|
/* Round address up to desired boundary. */
|
4844 |
|
|
temp = gen_reg_rtx (Pmode);
|
4845 |
|
|
addend = GEN_INT (TRAMPOLINE_ALIGNMENT / BITS_PER_UNIT - 1);
|
4846 |
|
|
mask = GEN_INT (-TRAMPOLINE_ALIGNMENT / BITS_PER_UNIT);
|
4847 |
|
|
|
4848 |
|
|
temp = expand_simple_binop (Pmode, PLUS, tramp, addend,
|
4849 |
|
|
temp, 0, OPTAB_LIB_WIDEN);
|
4850 |
|
|
tramp = expand_simple_binop (Pmode, AND, temp, mask,
|
4851 |
|
|
temp, 0, OPTAB_LIB_WIDEN);
|
4852 |
|
|
|
4853 |
|
|
return tramp;
|
4854 |
|
|
}
|
4855 |
|
|
|
4856 |
|
|
static rtx
|
4857 |
|
|
expand_builtin_init_trampoline (tree exp, bool onstack)
|
4858 |
|
|
{
|
4859 |
|
|
tree t_tramp, t_func, t_chain;
|
4860 |
|
|
rtx m_tramp, r_tramp, r_chain, tmp;
|
4861 |
|
|
|
4862 |
|
|
if (!validate_arglist (exp, POINTER_TYPE, POINTER_TYPE,
|
4863 |
|
|
POINTER_TYPE, VOID_TYPE))
|
4864 |
|
|
return NULL_RTX;
|
4865 |
|
|
|
4866 |
|
|
t_tramp = CALL_EXPR_ARG (exp, 0);
|
4867 |
|
|
t_func = CALL_EXPR_ARG (exp, 1);
|
4868 |
|
|
t_chain = CALL_EXPR_ARG (exp, 2);
|
4869 |
|
|
|
4870 |
|
|
r_tramp = expand_normal (t_tramp);
|
4871 |
|
|
m_tramp = gen_rtx_MEM (BLKmode, r_tramp);
|
4872 |
|
|
MEM_NOTRAP_P (m_tramp) = 1;
|
4873 |
|
|
|
4874 |
|
|
/* If ONSTACK, the TRAMP argument should be the address of a field
|
4875 |
|
|
within the local function's FRAME decl. Either way, let's see if
|
4876 |
|
|
we can fill in the MEM_ATTRs for this memory. */
|
4877 |
|
|
if (TREE_CODE (t_tramp) == ADDR_EXPR)
|
4878 |
|
|
set_mem_attributes_minus_bitpos (m_tramp, TREE_OPERAND (t_tramp, 0),
|
4879 |
|
|
true, 0);
|
4880 |
|
|
|
4881 |
|
|
/* Creator of a heap trampoline is responsible for making sure the
|
4882 |
|
|
address is aligned to at least STACK_BOUNDARY. Normally malloc
|
4883 |
|
|
will ensure this anyhow. */
|
4884 |
|
|
tmp = round_trampoline_addr (r_tramp);
|
4885 |
|
|
if (tmp != r_tramp)
|
4886 |
|
|
{
|
4887 |
|
|
m_tramp = change_address (m_tramp, BLKmode, tmp);
|
4888 |
|
|
set_mem_align (m_tramp, TRAMPOLINE_ALIGNMENT);
|
4889 |
|
|
set_mem_size (m_tramp, TRAMPOLINE_SIZE);
|
4890 |
|
|
}
|
4891 |
|
|
|
4892 |
|
|
/* The FUNC argument should be the address of the nested function.
|
4893 |
|
|
Extract the actual function decl to pass to the hook. */
|
4894 |
|
|
gcc_assert (TREE_CODE (t_func) == ADDR_EXPR);
|
4895 |
|
|
t_func = TREE_OPERAND (t_func, 0);
|
4896 |
|
|
gcc_assert (TREE_CODE (t_func) == FUNCTION_DECL);
|
4897 |
|
|
|
4898 |
|
|
r_chain = expand_normal (t_chain);
|
4899 |
|
|
|
4900 |
|
|
/* Generate insns to initialize the trampoline. */
|
4901 |
|
|
targetm.calls.trampoline_init (m_tramp, t_func, r_chain);
|
4902 |
|
|
|
4903 |
|
|
if (onstack)
|
4904 |
|
|
{
|
4905 |
|
|
trampolines_created = 1;
|
4906 |
|
|
|
4907 |
|
|
warning_at (DECL_SOURCE_LOCATION (t_func), OPT_Wtrampolines,
|
4908 |
|
|
"trampoline generated for nested function %qD", t_func);
|
4909 |
|
|
}
|
4910 |
|
|
|
4911 |
|
|
return const0_rtx;
|
4912 |
|
|
}
|
4913 |
|
|
|
4914 |
|
|
static rtx
|
4915 |
|
|
expand_builtin_adjust_trampoline (tree exp)
|
4916 |
|
|
{
|
4917 |
|
|
rtx tramp;
|
4918 |
|
|
|
4919 |
|
|
if (!validate_arglist (exp, POINTER_TYPE, VOID_TYPE))
|
4920 |
|
|
return NULL_RTX;
|
4921 |
|
|
|
4922 |
|
|
tramp = expand_normal (CALL_EXPR_ARG (exp, 0));
|
4923 |
|
|
tramp = round_trampoline_addr (tramp);
|
4924 |
|
|
if (targetm.calls.trampoline_adjust_address)
|
4925 |
|
|
tramp = targetm.calls.trampoline_adjust_address (tramp);
|
4926 |
|
|
|
4927 |
|
|
return tramp;
|
4928 |
|
|
}
|
4929 |
|
|
|
4930 |
|
|
/* Expand the call EXP to the built-in signbit, signbitf or signbitl
|
4931 |
|
|
function. The function first checks whether the back end provides
|
4932 |
|
|
an insn to implement signbit for the respective mode. If not, it
|
4933 |
|
|
checks whether the floating point format of the value is such that
|
4934 |
|
|
the sign bit can be extracted. If that is not the case, the
|
4935 |
|
|
function returns NULL_RTX to indicate that a normal call should be
|
4936 |
|
|
emitted rather than expanding the function in-line. EXP is the
|
4937 |
|
|
expression that is a call to the builtin function; if convenient,
|
4938 |
|
|
the result should be placed in TARGET. */
|
4939 |
|
|
static rtx
|
4940 |
|
|
expand_builtin_signbit (tree exp, rtx target)
|
4941 |
|
|
{
|
4942 |
|
|
const struct real_format *fmt;
|
4943 |
|
|
enum machine_mode fmode, imode, rmode;
|
4944 |
|
|
tree arg;
|
4945 |
|
|
int word, bitpos;
|
4946 |
|
|
enum insn_code icode;
|
4947 |
|
|
rtx temp;
|
4948 |
|
|
location_t loc = EXPR_LOCATION (exp);
|
4949 |
|
|
|
4950 |
|
|
if (!validate_arglist (exp, REAL_TYPE, VOID_TYPE))
|
4951 |
|
|
return NULL_RTX;
|
4952 |
|
|
|
4953 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
4954 |
|
|
fmode = TYPE_MODE (TREE_TYPE (arg));
|
4955 |
|
|
rmode = TYPE_MODE (TREE_TYPE (exp));
|
4956 |
|
|
fmt = REAL_MODE_FORMAT (fmode);
|
4957 |
|
|
|
4958 |
|
|
arg = builtin_save_expr (arg);
|
4959 |
|
|
|
4960 |
|
|
/* Expand the argument yielding a RTX expression. */
|
4961 |
|
|
temp = expand_normal (arg);
|
4962 |
|
|
|
4963 |
|
|
/* Check if the back end provides an insn that handles signbit for the
|
4964 |
|
|
argument's mode. */
|
4965 |
|
|
icode = optab_handler (signbit_optab, fmode);
|
4966 |
|
|
if (icode != CODE_FOR_nothing)
|
4967 |
|
|
{
|
4968 |
|
|
rtx last = get_last_insn ();
|
4969 |
|
|
target = gen_reg_rtx (TYPE_MODE (TREE_TYPE (exp)));
|
4970 |
|
|
if (maybe_emit_unop_insn (icode, target, temp, UNKNOWN))
|
4971 |
|
|
return target;
|
4972 |
|
|
delete_insns_since (last);
|
4973 |
|
|
}
|
4974 |
|
|
|
4975 |
|
|
/* For floating point formats without a sign bit, implement signbit
|
4976 |
|
|
as "ARG < 0.0". */
|
4977 |
|
|
bitpos = fmt->signbit_ro;
|
4978 |
|
|
if (bitpos < 0)
|
4979 |
|
|
{
|
4980 |
|
|
/* But we can't do this if the format supports signed zero. */
|
4981 |
|
|
if (fmt->has_signed_zero && HONOR_SIGNED_ZEROS (fmode))
|
4982 |
|
|
return NULL_RTX;
|
4983 |
|
|
|
4984 |
|
|
arg = fold_build2_loc (loc, LT_EXPR, TREE_TYPE (exp), arg,
|
4985 |
|
|
build_real (TREE_TYPE (arg), dconst0));
|
4986 |
|
|
return expand_expr (arg, target, VOIDmode, EXPAND_NORMAL);
|
4987 |
|
|
}
|
4988 |
|
|
|
4989 |
|
|
if (GET_MODE_SIZE (fmode) <= UNITS_PER_WORD)
|
4990 |
|
|
{
|
4991 |
|
|
imode = int_mode_for_mode (fmode);
|
4992 |
|
|
if (imode == BLKmode)
|
4993 |
|
|
return NULL_RTX;
|
4994 |
|
|
temp = gen_lowpart (imode, temp);
|
4995 |
|
|
}
|
4996 |
|
|
else
|
4997 |
|
|
{
|
4998 |
|
|
imode = word_mode;
|
4999 |
|
|
/* Handle targets with different FP word orders. */
|
5000 |
|
|
if (FLOAT_WORDS_BIG_ENDIAN)
|
5001 |
|
|
word = (GET_MODE_BITSIZE (fmode) - bitpos) / BITS_PER_WORD;
|
5002 |
|
|
else
|
5003 |
|
|
word = bitpos / BITS_PER_WORD;
|
5004 |
|
|
temp = operand_subword_force (temp, word, fmode);
|
5005 |
|
|
bitpos = bitpos % BITS_PER_WORD;
|
5006 |
|
|
}
|
5007 |
|
|
|
5008 |
|
|
/* Force the intermediate word_mode (or narrower) result into a
|
5009 |
|
|
register. This avoids attempting to create paradoxical SUBREGs
|
5010 |
|
|
of floating point modes below. */
|
5011 |
|
|
temp = force_reg (imode, temp);
|
5012 |
|
|
|
5013 |
|
|
/* If the bitpos is within the "result mode" lowpart, the operation
|
5014 |
|
|
can be implement with a single bitwise AND. Otherwise, we need
|
5015 |
|
|
a right shift and an AND. */
|
5016 |
|
|
|
5017 |
|
|
if (bitpos < GET_MODE_BITSIZE (rmode))
|
5018 |
|
|
{
|
5019 |
|
|
double_int mask = double_int_setbit (double_int_zero, bitpos);
|
5020 |
|
|
|
5021 |
|
|
if (GET_MODE_SIZE (imode) > GET_MODE_SIZE (rmode))
|
5022 |
|
|
temp = gen_lowpart (rmode, temp);
|
5023 |
|
|
temp = expand_binop (rmode, and_optab, temp,
|
5024 |
|
|
immed_double_int_const (mask, rmode),
|
5025 |
|
|
NULL_RTX, 1, OPTAB_LIB_WIDEN);
|
5026 |
|
|
}
|
5027 |
|
|
else
|
5028 |
|
|
{
|
5029 |
|
|
/* Perform a logical right shift to place the signbit in the least
|
5030 |
|
|
significant bit, then truncate the result to the desired mode
|
5031 |
|
|
and mask just this bit. */
|
5032 |
|
|
temp = expand_shift (RSHIFT_EXPR, imode, temp, bitpos, NULL_RTX, 1);
|
5033 |
|
|
temp = gen_lowpart (rmode, temp);
|
5034 |
|
|
temp = expand_binop (rmode, and_optab, temp, const1_rtx,
|
5035 |
|
|
NULL_RTX, 1, OPTAB_LIB_WIDEN);
|
5036 |
|
|
}
|
5037 |
|
|
|
5038 |
|
|
return temp;
|
5039 |
|
|
}
|
5040 |
|
|
|
5041 |
|
|
/* Expand fork or exec calls. TARGET is the desired target of the
|
5042 |
|
|
call. EXP is the call. FN is the
|
5043 |
|
|
identificator of the actual function. IGNORE is nonzero if the
|
5044 |
|
|
value is to be ignored. */
|
5045 |
|
|
|
5046 |
|
|
static rtx
|
5047 |
|
|
expand_builtin_fork_or_exec (tree fn, tree exp, rtx target, int ignore)
|
5048 |
|
|
{
|
5049 |
|
|
tree id, decl;
|
5050 |
|
|
tree call;
|
5051 |
|
|
|
5052 |
|
|
/* If we are not profiling, just call the function. */
|
5053 |
|
|
if (!profile_arc_flag)
|
5054 |
|
|
return NULL_RTX;
|
5055 |
|
|
|
5056 |
|
|
/* Otherwise call the wrapper. This should be equivalent for the rest of
|
5057 |
|
|
compiler, so the code does not diverge, and the wrapper may run the
|
5058 |
|
|
code necessary for keeping the profiling sane. */
|
5059 |
|
|
|
5060 |
|
|
switch (DECL_FUNCTION_CODE (fn))
|
5061 |
|
|
{
|
5062 |
|
|
case BUILT_IN_FORK:
|
5063 |
|
|
id = get_identifier ("__gcov_fork");
|
5064 |
|
|
break;
|
5065 |
|
|
|
5066 |
|
|
case BUILT_IN_EXECL:
|
5067 |
|
|
id = get_identifier ("__gcov_execl");
|
5068 |
|
|
break;
|
5069 |
|
|
|
5070 |
|
|
case BUILT_IN_EXECV:
|
5071 |
|
|
id = get_identifier ("__gcov_execv");
|
5072 |
|
|
break;
|
5073 |
|
|
|
5074 |
|
|
case BUILT_IN_EXECLP:
|
5075 |
|
|
id = get_identifier ("__gcov_execlp");
|
5076 |
|
|
break;
|
5077 |
|
|
|
5078 |
|
|
case BUILT_IN_EXECLE:
|
5079 |
|
|
id = get_identifier ("__gcov_execle");
|
5080 |
|
|
break;
|
5081 |
|
|
|
5082 |
|
|
case BUILT_IN_EXECVP:
|
5083 |
|
|
id = get_identifier ("__gcov_execvp");
|
5084 |
|
|
break;
|
5085 |
|
|
|
5086 |
|
|
case BUILT_IN_EXECVE:
|
5087 |
|
|
id = get_identifier ("__gcov_execve");
|
5088 |
|
|
break;
|
5089 |
|
|
|
5090 |
|
|
default:
|
5091 |
|
|
gcc_unreachable ();
|
5092 |
|
|
}
|
5093 |
|
|
|
5094 |
|
|
decl = build_decl (DECL_SOURCE_LOCATION (fn),
|
5095 |
|
|
FUNCTION_DECL, id, TREE_TYPE (fn));
|
5096 |
|
|
DECL_EXTERNAL (decl) = 1;
|
5097 |
|
|
TREE_PUBLIC (decl) = 1;
|
5098 |
|
|
DECL_ARTIFICIAL (decl) = 1;
|
5099 |
|
|
TREE_NOTHROW (decl) = 1;
|
5100 |
|
|
DECL_VISIBILITY (decl) = VISIBILITY_DEFAULT;
|
5101 |
|
|
DECL_VISIBILITY_SPECIFIED (decl) = 1;
|
5102 |
|
|
call = rewrite_call_expr (EXPR_LOCATION (exp), exp, 0, decl, 0);
|
5103 |
|
|
return expand_call (call, target, ignore);
|
5104 |
|
|
}
|
5105 |
|
|
|
5106 |
|
|
|
5107 |
|
|
|
5108 |
|
|
/* Reconstitute a mode for a __sync intrinsic operation. Since the type of
|
5109 |
|
|
the pointer in these functions is void*, the tree optimizers may remove
|
5110 |
|
|
casts. The mode computed in expand_builtin isn't reliable either, due
|
5111 |
|
|
to __sync_bool_compare_and_swap.
|
5112 |
|
|
|
5113 |
|
|
FCODE_DIFF should be fcode - base, where base is the FOO_1 code for the
|
5114 |
|
|
group of builtins. This gives us log2 of the mode size. */
|
5115 |
|
|
|
5116 |
|
|
static inline enum machine_mode
|
5117 |
|
|
get_builtin_sync_mode (int fcode_diff)
|
5118 |
|
|
{
|
5119 |
|
|
/* The size is not negotiable, so ask not to get BLKmode in return
|
5120 |
|
|
if the target indicates that a smaller size would be better. */
|
5121 |
|
|
return mode_for_size (BITS_PER_UNIT << fcode_diff, MODE_INT, 0);
|
5122 |
|
|
}
|
5123 |
|
|
|
5124 |
|
|
/* Expand the memory expression LOC and return the appropriate memory operand
|
5125 |
|
|
for the builtin_sync operations. */
|
5126 |
|
|
|
5127 |
|
|
static rtx
|
5128 |
|
|
get_builtin_sync_mem (tree loc, enum machine_mode mode)
|
5129 |
|
|
{
|
5130 |
|
|
rtx addr, mem;
|
5131 |
|
|
|
5132 |
|
|
addr = expand_expr (loc, NULL_RTX, ptr_mode, EXPAND_SUM);
|
5133 |
|
|
addr = convert_memory_address (Pmode, addr);
|
5134 |
|
|
|
5135 |
|
|
/* Note that we explicitly do not want any alias information for this
|
5136 |
|
|
memory, so that we kill all other live memories. Otherwise we don't
|
5137 |
|
|
satisfy the full barrier semantics of the intrinsic. */
|
5138 |
|
|
mem = validize_mem (gen_rtx_MEM (mode, addr));
|
5139 |
|
|
|
5140 |
|
|
/* The alignment needs to be at least according to that of the mode. */
|
5141 |
|
|
set_mem_align (mem, MAX (GET_MODE_ALIGNMENT (mode),
|
5142 |
|
|
get_pointer_alignment (loc)));
|
5143 |
|
|
set_mem_alias_set (mem, ALIAS_SET_MEMORY_BARRIER);
|
5144 |
|
|
MEM_VOLATILE_P (mem) = 1;
|
5145 |
|
|
|
5146 |
|
|
return mem;
|
5147 |
|
|
}
|
5148 |
|
|
|
5149 |
|
|
/* Make sure an argument is in the right mode.
|
5150 |
|
|
EXP is the tree argument.
|
5151 |
|
|
MODE is the mode it should be in. */
|
5152 |
|
|
|
5153 |
|
|
static rtx
|
5154 |
|
|
expand_expr_force_mode (tree exp, enum machine_mode mode)
|
5155 |
|
|
{
|
5156 |
|
|
rtx val;
|
5157 |
|
|
enum machine_mode old_mode;
|
5158 |
|
|
|
5159 |
|
|
val = expand_expr (exp, NULL_RTX, mode, EXPAND_NORMAL);
|
5160 |
|
|
/* If VAL is promoted to a wider mode, convert it back to MODE. Take care
|
5161 |
|
|
of CONST_INTs, where we know the old_mode only from the call argument. */
|
5162 |
|
|
|
5163 |
|
|
old_mode = GET_MODE (val);
|
5164 |
|
|
if (old_mode == VOIDmode)
|
5165 |
|
|
old_mode = TYPE_MODE (TREE_TYPE (exp));
|
5166 |
|
|
val = convert_modes (mode, old_mode, val, 1);
|
5167 |
|
|
return val;
|
5168 |
|
|
}
|
5169 |
|
|
|
5170 |
|
|
|
5171 |
|
|
/* Expand the __sync_xxx_and_fetch and __sync_fetch_and_xxx intrinsics.
|
5172 |
|
|
EXP is the CALL_EXPR. CODE is the rtx code
|
5173 |
|
|
that corresponds to the arithmetic or logical operation from the name;
|
5174 |
|
|
an exception here is that NOT actually means NAND. TARGET is an optional
|
5175 |
|
|
place for us to store the results; AFTER is true if this is the
|
5176 |
|
|
fetch_and_xxx form. */
|
5177 |
|
|
|
5178 |
|
|
static rtx
|
5179 |
|
|
expand_builtin_sync_operation (enum machine_mode mode, tree exp,
|
5180 |
|
|
enum rtx_code code, bool after,
|
5181 |
|
|
rtx target)
|
5182 |
|
|
{
|
5183 |
|
|
rtx val, mem;
|
5184 |
|
|
location_t loc = EXPR_LOCATION (exp);
|
5185 |
|
|
|
5186 |
|
|
if (code == NOT && warn_sync_nand)
|
5187 |
|
|
{
|
5188 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
5189 |
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
5190 |
|
|
|
5191 |
|
|
static bool warned_f_a_n, warned_n_a_f;
|
5192 |
|
|
|
5193 |
|
|
switch (fcode)
|
5194 |
|
|
{
|
5195 |
|
|
case BUILT_IN_SYNC_FETCH_AND_NAND_1:
|
5196 |
|
|
case BUILT_IN_SYNC_FETCH_AND_NAND_2:
|
5197 |
|
|
case BUILT_IN_SYNC_FETCH_AND_NAND_4:
|
5198 |
|
|
case BUILT_IN_SYNC_FETCH_AND_NAND_8:
|
5199 |
|
|
case BUILT_IN_SYNC_FETCH_AND_NAND_16:
|
5200 |
|
|
if (warned_f_a_n)
|
5201 |
|
|
break;
|
5202 |
|
|
|
5203 |
|
|
fndecl = builtin_decl_implicit (BUILT_IN_SYNC_FETCH_AND_NAND_N);
|
5204 |
|
|
inform (loc, "%qD changed semantics in GCC 4.4", fndecl);
|
5205 |
|
|
warned_f_a_n = true;
|
5206 |
|
|
break;
|
5207 |
|
|
|
5208 |
|
|
case BUILT_IN_SYNC_NAND_AND_FETCH_1:
|
5209 |
|
|
case BUILT_IN_SYNC_NAND_AND_FETCH_2:
|
5210 |
|
|
case BUILT_IN_SYNC_NAND_AND_FETCH_4:
|
5211 |
|
|
case BUILT_IN_SYNC_NAND_AND_FETCH_8:
|
5212 |
|
|
case BUILT_IN_SYNC_NAND_AND_FETCH_16:
|
5213 |
|
|
if (warned_n_a_f)
|
5214 |
|
|
break;
|
5215 |
|
|
|
5216 |
|
|
fndecl = builtin_decl_implicit (BUILT_IN_SYNC_NAND_AND_FETCH_N);
|
5217 |
|
|
inform (loc, "%qD changed semantics in GCC 4.4", fndecl);
|
5218 |
|
|
warned_n_a_f = true;
|
5219 |
|
|
break;
|
5220 |
|
|
|
5221 |
|
|
default:
|
5222 |
|
|
gcc_unreachable ();
|
5223 |
|
|
}
|
5224 |
|
|
}
|
5225 |
|
|
|
5226 |
|
|
/* Expand the operands. */
|
5227 |
|
|
mem = get_builtin_sync_mem (CALL_EXPR_ARG (exp, 0), mode);
|
5228 |
|
|
val = expand_expr_force_mode (CALL_EXPR_ARG (exp, 1), mode);
|
5229 |
|
|
|
5230 |
|
|
return expand_atomic_fetch_op (target, mem, val, code, MEMMODEL_SEQ_CST,
|
5231 |
|
|
after);
|
5232 |
|
|
}
|
5233 |
|
|
|
5234 |
|
|
/* Expand the __sync_val_compare_and_swap and __sync_bool_compare_and_swap
|
5235 |
|
|
intrinsics. EXP is the CALL_EXPR. IS_BOOL is
|
5236 |
|
|
true if this is the boolean form. TARGET is a place for us to store the
|
5237 |
|
|
results; this is NOT optional if IS_BOOL is true. */
|
5238 |
|
|
|
5239 |
|
|
static rtx
|
5240 |
|
|
expand_builtin_compare_and_swap (enum machine_mode mode, tree exp,
|
5241 |
|
|
bool is_bool, rtx target)
|
5242 |
|
|
{
|
5243 |
|
|
rtx old_val, new_val, mem;
|
5244 |
|
|
rtx *pbool, *poval;
|
5245 |
|
|
|
5246 |
|
|
/* Expand the operands. */
|
5247 |
|
|
mem = get_builtin_sync_mem (CALL_EXPR_ARG (exp, 0), mode);
|
5248 |
|
|
old_val = expand_expr_force_mode (CALL_EXPR_ARG (exp, 1), mode);
|
5249 |
|
|
new_val = expand_expr_force_mode (CALL_EXPR_ARG (exp, 2), mode);
|
5250 |
|
|
|
5251 |
|
|
pbool = poval = NULL;
|
5252 |
|
|
if (target != const0_rtx)
|
5253 |
|
|
{
|
5254 |
|
|
if (is_bool)
|
5255 |
|
|
pbool = ⌖
|
5256 |
|
|
else
|
5257 |
|
|
poval = ⌖
|
5258 |
|
|
}
|
5259 |
|
|
if (!expand_atomic_compare_and_swap (pbool, poval, mem, old_val, new_val,
|
5260 |
|
|
false, MEMMODEL_SEQ_CST,
|
5261 |
|
|
MEMMODEL_SEQ_CST))
|
5262 |
|
|
return NULL_RTX;
|
5263 |
|
|
|
5264 |
|
|
return target;
|
5265 |
|
|
}
|
5266 |
|
|
|
5267 |
|
|
/* Expand the __sync_lock_test_and_set intrinsic. Note that the most
|
5268 |
|
|
general form is actually an atomic exchange, and some targets only
|
5269 |
|
|
support a reduced form with the second argument being a constant 1.
|
5270 |
|
|
EXP is the CALL_EXPR; TARGET is an optional place for us to store
|
5271 |
|
|
the results. */
|
5272 |
|
|
|
5273 |
|
|
static rtx
|
5274 |
|
|
expand_builtin_sync_lock_test_and_set (enum machine_mode mode, tree exp,
|
5275 |
|
|
rtx target)
|
5276 |
|
|
{
|
5277 |
|
|
rtx val, mem;
|
5278 |
|
|
|
5279 |
|
|
/* Expand the operands. */
|
5280 |
|
|
mem = get_builtin_sync_mem (CALL_EXPR_ARG (exp, 0), mode);
|
5281 |
|
|
val = expand_expr_force_mode (CALL_EXPR_ARG (exp, 1), mode);
|
5282 |
|
|
|
5283 |
|
|
return expand_sync_lock_test_and_set (target, mem, val);
|
5284 |
|
|
}
|
5285 |
|
|
|
5286 |
|
|
/* Expand the __sync_lock_release intrinsic. EXP is the CALL_EXPR. */
|
5287 |
|
|
|
5288 |
|
|
static void
|
5289 |
|
|
expand_builtin_sync_lock_release (enum machine_mode mode, tree exp)
|
5290 |
|
|
{
|
5291 |
|
|
rtx mem;
|
5292 |
|
|
|
5293 |
|
|
/* Expand the operands. */
|
5294 |
|
|
mem = get_builtin_sync_mem (CALL_EXPR_ARG (exp, 0), mode);
|
5295 |
|
|
|
5296 |
|
|
expand_atomic_store (mem, const0_rtx, MEMMODEL_RELEASE, true);
|
5297 |
|
|
}
|
5298 |
|
|
|
5299 |
|
|
/* Given an integer representing an ``enum memmodel'', verify its
|
5300 |
|
|
correctness and return the memory model enum. */
|
5301 |
|
|
|
5302 |
|
|
static enum memmodel
|
5303 |
|
|
get_memmodel (tree exp)
|
5304 |
|
|
{
|
5305 |
|
|
rtx op;
|
5306 |
|
|
|
5307 |
|
|
/* If the parameter is not a constant, it's a run time value so we'll just
|
5308 |
|
|
convert it to MEMMODEL_SEQ_CST to avoid annoying runtime checking. */
|
5309 |
|
|
if (TREE_CODE (exp) != INTEGER_CST)
|
5310 |
|
|
return MEMMODEL_SEQ_CST;
|
5311 |
|
|
|
5312 |
|
|
op = expand_normal (exp);
|
5313 |
|
|
if (INTVAL (op) < 0 || INTVAL (op) >= MEMMODEL_LAST)
|
5314 |
|
|
{
|
5315 |
|
|
warning (OPT_Winvalid_memory_model,
|
5316 |
|
|
"invalid memory model argument to builtin");
|
5317 |
|
|
return MEMMODEL_SEQ_CST;
|
5318 |
|
|
}
|
5319 |
|
|
return (enum memmodel) INTVAL (op);
|
5320 |
|
|
}
|
5321 |
|
|
|
5322 |
|
|
/* Expand the __atomic_exchange intrinsic:
|
5323 |
|
|
TYPE __atomic_exchange (TYPE *object, TYPE desired, enum memmodel)
|
5324 |
|
|
EXP is the CALL_EXPR.
|
5325 |
|
|
TARGET is an optional place for us to store the results. */
|
5326 |
|
|
|
5327 |
|
|
static rtx
|
5328 |
|
|
expand_builtin_atomic_exchange (enum machine_mode mode, tree exp, rtx target)
|
5329 |
|
|
{
|
5330 |
|
|
rtx val, mem;
|
5331 |
|
|
enum memmodel model;
|
5332 |
|
|
|
5333 |
|
|
model = get_memmodel (CALL_EXPR_ARG (exp, 2));
|
5334 |
|
|
if (model == MEMMODEL_CONSUME)
|
5335 |
|
|
{
|
5336 |
|
|
error ("invalid memory model for %<__atomic_exchange%>");
|
5337 |
|
|
return NULL_RTX;
|
5338 |
|
|
}
|
5339 |
|
|
|
5340 |
|
|
if (!flag_inline_atomics)
|
5341 |
|
|
return NULL_RTX;
|
5342 |
|
|
|
5343 |
|
|
/* Expand the operands. */
|
5344 |
|
|
mem = get_builtin_sync_mem (CALL_EXPR_ARG (exp, 0), mode);
|
5345 |
|
|
val = expand_expr_force_mode (CALL_EXPR_ARG (exp, 1), mode);
|
5346 |
|
|
|
5347 |
|
|
return expand_atomic_exchange (target, mem, val, model);
|
5348 |
|
|
}
|
5349 |
|
|
|
5350 |
|
|
/* Expand the __atomic_compare_exchange intrinsic:
|
5351 |
|
|
bool __atomic_compare_exchange (TYPE *object, TYPE *expect,
|
5352 |
|
|
TYPE desired, BOOL weak,
|
5353 |
|
|
enum memmodel success,
|
5354 |
|
|
enum memmodel failure)
|
5355 |
|
|
EXP is the CALL_EXPR.
|
5356 |
|
|
TARGET is an optional place for us to store the results. */
|
5357 |
|
|
|
5358 |
|
|
static rtx
|
5359 |
|
|
expand_builtin_atomic_compare_exchange (enum machine_mode mode, tree exp,
|
5360 |
|
|
rtx target)
|
5361 |
|
|
{
|
5362 |
|
|
rtx expect, desired, mem, oldval;
|
5363 |
|
|
enum memmodel success, failure;
|
5364 |
|
|
tree weak;
|
5365 |
|
|
bool is_weak;
|
5366 |
|
|
|
5367 |
|
|
success = get_memmodel (CALL_EXPR_ARG (exp, 4));
|
5368 |
|
|
failure = get_memmodel (CALL_EXPR_ARG (exp, 5));
|
5369 |
|
|
|
5370 |
|
|
if (failure == MEMMODEL_RELEASE || failure == MEMMODEL_ACQ_REL)
|
5371 |
|
|
{
|
5372 |
|
|
error ("invalid failure memory model for %<__atomic_compare_exchange%>");
|
5373 |
|
|
return NULL_RTX;
|
5374 |
|
|
}
|
5375 |
|
|
|
5376 |
|
|
if (failure > success)
|
5377 |
|
|
{
|
5378 |
|
|
error ("failure memory model cannot be stronger than success "
|
5379 |
|
|
"memory model for %<__atomic_compare_exchange%>");
|
5380 |
|
|
return NULL_RTX;
|
5381 |
|
|
}
|
5382 |
|
|
|
5383 |
|
|
if (!flag_inline_atomics)
|
5384 |
|
|
return NULL_RTX;
|
5385 |
|
|
|
5386 |
|
|
/* Expand the operands. */
|
5387 |
|
|
mem = get_builtin_sync_mem (CALL_EXPR_ARG (exp, 0), mode);
|
5388 |
|
|
|
5389 |
|
|
expect = expand_normal (CALL_EXPR_ARG (exp, 1));
|
5390 |
|
|
expect = convert_memory_address (Pmode, expect);
|
5391 |
|
|
desired = expand_expr_force_mode (CALL_EXPR_ARG (exp, 2), mode);
|
5392 |
|
|
|
5393 |
|
|
weak = CALL_EXPR_ARG (exp, 3);
|
5394 |
|
|
is_weak = false;
|
5395 |
|
|
if (host_integerp (weak, 0) && tree_low_cst (weak, 0) != 0)
|
5396 |
|
|
is_weak = true;
|
5397 |
|
|
|
5398 |
|
|
oldval = copy_to_reg (gen_rtx_MEM (mode, expect));
|
5399 |
|
|
|
5400 |
|
|
if (!expand_atomic_compare_and_swap ((target == const0_rtx ? NULL : &target),
|
5401 |
|
|
&oldval, mem, oldval, desired,
|
5402 |
|
|
is_weak, success, failure))
|
5403 |
|
|
return NULL_RTX;
|
5404 |
|
|
|
5405 |
|
|
emit_move_insn (gen_rtx_MEM (mode, expect), oldval);
|
5406 |
|
|
return target;
|
5407 |
|
|
}
|
5408 |
|
|
|
5409 |
|
|
/* Expand the __atomic_load intrinsic:
|
5410 |
|
|
TYPE __atomic_load (TYPE *object, enum memmodel)
|
5411 |
|
|
EXP is the CALL_EXPR.
|
5412 |
|
|
TARGET is an optional place for us to store the results. */
|
5413 |
|
|
|
5414 |
|
|
static rtx
|
5415 |
|
|
expand_builtin_atomic_load (enum machine_mode mode, tree exp, rtx target)
|
5416 |
|
|
{
|
5417 |
|
|
rtx mem;
|
5418 |
|
|
enum memmodel model;
|
5419 |
|
|
|
5420 |
|
|
model = get_memmodel (CALL_EXPR_ARG (exp, 1));
|
5421 |
|
|
if (model == MEMMODEL_RELEASE
|
5422 |
|
|
|| model == MEMMODEL_ACQ_REL)
|
5423 |
|
|
{
|
5424 |
|
|
error ("invalid memory model for %<__atomic_load%>");
|
5425 |
|
|
return NULL_RTX;
|
5426 |
|
|
}
|
5427 |
|
|
|
5428 |
|
|
if (!flag_inline_atomics)
|
5429 |
|
|
return NULL_RTX;
|
5430 |
|
|
|
5431 |
|
|
/* Expand the operand. */
|
5432 |
|
|
mem = get_builtin_sync_mem (CALL_EXPR_ARG (exp, 0), mode);
|
5433 |
|
|
|
5434 |
|
|
return expand_atomic_load (target, mem, model);
|
5435 |
|
|
}
|
5436 |
|
|
|
5437 |
|
|
|
5438 |
|
|
/* Expand the __atomic_store intrinsic:
|
5439 |
|
|
void __atomic_store (TYPE *object, TYPE desired, enum memmodel)
|
5440 |
|
|
EXP is the CALL_EXPR.
|
5441 |
|
|
TARGET is an optional place for us to store the results. */
|
5442 |
|
|
|
5443 |
|
|
static rtx
|
5444 |
|
|
expand_builtin_atomic_store (enum machine_mode mode, tree exp)
|
5445 |
|
|
{
|
5446 |
|
|
rtx mem, val;
|
5447 |
|
|
enum memmodel model;
|
5448 |
|
|
|
5449 |
|
|
model = get_memmodel (CALL_EXPR_ARG (exp, 2));
|
5450 |
|
|
if (model != MEMMODEL_RELAXED
|
5451 |
|
|
&& model != MEMMODEL_SEQ_CST
|
5452 |
|
|
&& model != MEMMODEL_RELEASE)
|
5453 |
|
|
{
|
5454 |
|
|
error ("invalid memory model for %<__atomic_store%>");
|
5455 |
|
|
return NULL_RTX;
|
5456 |
|
|
}
|
5457 |
|
|
|
5458 |
|
|
if (!flag_inline_atomics)
|
5459 |
|
|
return NULL_RTX;
|
5460 |
|
|
|
5461 |
|
|
/* Expand the operands. */
|
5462 |
|
|
mem = get_builtin_sync_mem (CALL_EXPR_ARG (exp, 0), mode);
|
5463 |
|
|
val = expand_expr_force_mode (CALL_EXPR_ARG (exp, 1), mode);
|
5464 |
|
|
|
5465 |
|
|
return expand_atomic_store (mem, val, model, false);
|
5466 |
|
|
}
|
5467 |
|
|
|
5468 |
|
|
/* Expand the __atomic_fetch_XXX intrinsic:
|
5469 |
|
|
TYPE __atomic_fetch_XXX (TYPE *object, TYPE val, enum memmodel)
|
5470 |
|
|
EXP is the CALL_EXPR.
|
5471 |
|
|
TARGET is an optional place for us to store the results.
|
5472 |
|
|
CODE is the operation, PLUS, MINUS, ADD, XOR, or IOR.
|
5473 |
|
|
FETCH_AFTER is true if returning the result of the operation.
|
5474 |
|
|
FETCH_AFTER is false if returning the value before the operation.
|
5475 |
|
|
IGNORE is true if the result is not used.
|
5476 |
|
|
EXT_CALL is the correct builtin for an external call if this cannot be
|
5477 |
|
|
resolved to an instruction sequence. */
|
5478 |
|
|
|
5479 |
|
|
static rtx
|
5480 |
|
|
expand_builtin_atomic_fetch_op (enum machine_mode mode, tree exp, rtx target,
|
5481 |
|
|
enum rtx_code code, bool fetch_after,
|
5482 |
|
|
bool ignore, enum built_in_function ext_call)
|
5483 |
|
|
{
|
5484 |
|
|
rtx val, mem, ret;
|
5485 |
|
|
enum memmodel model;
|
5486 |
|
|
tree fndecl;
|
5487 |
|
|
tree addr;
|
5488 |
|
|
|
5489 |
|
|
model = get_memmodel (CALL_EXPR_ARG (exp, 2));
|
5490 |
|
|
|
5491 |
|
|
/* Expand the operands. */
|
5492 |
|
|
mem = get_builtin_sync_mem (CALL_EXPR_ARG (exp, 0), mode);
|
5493 |
|
|
val = expand_expr_force_mode (CALL_EXPR_ARG (exp, 1), mode);
|
5494 |
|
|
|
5495 |
|
|
/* Only try generating instructions if inlining is turned on. */
|
5496 |
|
|
if (flag_inline_atomics)
|
5497 |
|
|
{
|
5498 |
|
|
ret = expand_atomic_fetch_op (target, mem, val, code, model, fetch_after);
|
5499 |
|
|
if (ret)
|
5500 |
|
|
return ret;
|
5501 |
|
|
}
|
5502 |
|
|
|
5503 |
|
|
/* Return if a different routine isn't needed for the library call. */
|
5504 |
|
|
if (ext_call == BUILT_IN_NONE)
|
5505 |
|
|
return NULL_RTX;
|
5506 |
|
|
|
5507 |
|
|
/* Change the call to the specified function. */
|
5508 |
|
|
fndecl = get_callee_fndecl (exp);
|
5509 |
|
|
addr = CALL_EXPR_FN (exp);
|
5510 |
|
|
STRIP_NOPS (addr);
|
5511 |
|
|
|
5512 |
|
|
gcc_assert (TREE_OPERAND (addr, 0) == fndecl);
|
5513 |
|
|
TREE_OPERAND (addr, 0) = builtin_decl_explicit(ext_call);
|
5514 |
|
|
|
5515 |
|
|
/* Expand the call here so we can emit trailing code. */
|
5516 |
|
|
ret = expand_call (exp, target, ignore);
|
5517 |
|
|
|
5518 |
|
|
/* Replace the original function just in case it matters. */
|
5519 |
|
|
TREE_OPERAND (addr, 0) = fndecl;
|
5520 |
|
|
|
5521 |
|
|
/* Then issue the arithmetic correction to return the right result. */
|
5522 |
|
|
if (!ignore)
|
5523 |
|
|
{
|
5524 |
|
|
if (code == NOT)
|
5525 |
|
|
{
|
5526 |
|
|
ret = expand_simple_binop (mode, AND, ret, val, NULL_RTX, true,
|
5527 |
|
|
OPTAB_LIB_WIDEN);
|
5528 |
|
|
ret = expand_simple_unop (mode, NOT, ret, target, true);
|
5529 |
|
|
}
|
5530 |
|
|
else
|
5531 |
|
|
ret = expand_simple_binop (mode, code, ret, val, target, true,
|
5532 |
|
|
OPTAB_LIB_WIDEN);
|
5533 |
|
|
}
|
5534 |
|
|
return ret;
|
5535 |
|
|
}
|
5536 |
|
|
|
5537 |
|
|
|
5538 |
|
|
#ifndef HAVE_atomic_clear
|
5539 |
|
|
# define HAVE_atomic_clear 0
|
5540 |
|
|
# define gen_atomic_clear(x,y) (gcc_unreachable (), NULL_RTX)
|
5541 |
|
|
#endif
|
5542 |
|
|
|
5543 |
|
|
/* Expand an atomic clear operation.
|
5544 |
|
|
void _atomic_clear (BOOL *obj, enum memmodel)
|
5545 |
|
|
EXP is the call expression. */
|
5546 |
|
|
|
5547 |
|
|
static rtx
|
5548 |
|
|
expand_builtin_atomic_clear (tree exp)
|
5549 |
|
|
{
|
5550 |
|
|
enum machine_mode mode;
|
5551 |
|
|
rtx mem, ret;
|
5552 |
|
|
enum memmodel model;
|
5553 |
|
|
|
5554 |
|
|
mode = mode_for_size (BOOL_TYPE_SIZE, MODE_INT, 0);
|
5555 |
|
|
mem = get_builtin_sync_mem (CALL_EXPR_ARG (exp, 0), mode);
|
5556 |
|
|
model = get_memmodel (CALL_EXPR_ARG (exp, 1));
|
5557 |
|
|
|
5558 |
|
|
if (model == MEMMODEL_ACQUIRE || model == MEMMODEL_ACQ_REL)
|
5559 |
|
|
{
|
5560 |
|
|
error ("invalid memory model for %<__atomic_store%>");
|
5561 |
|
|
return const0_rtx;
|
5562 |
|
|
}
|
5563 |
|
|
|
5564 |
|
|
if (HAVE_atomic_clear)
|
5565 |
|
|
{
|
5566 |
|
|
emit_insn (gen_atomic_clear (mem, model));
|
5567 |
|
|
return const0_rtx;
|
5568 |
|
|
}
|
5569 |
|
|
|
5570 |
|
|
/* Try issuing an __atomic_store, and allow fallback to __sync_lock_release.
|
5571 |
|
|
Failing that, a store is issued by __atomic_store. The only way this can
|
5572 |
|
|
fail is if the bool type is larger than a word size. Unlikely, but
|
5573 |
|
|
handle it anyway for completeness. Assume a single threaded model since
|
5574 |
|
|
there is no atomic support in this case, and no barriers are required. */
|
5575 |
|
|
ret = expand_atomic_store (mem, const0_rtx, model, true);
|
5576 |
|
|
if (!ret)
|
5577 |
|
|
emit_move_insn (mem, const0_rtx);
|
5578 |
|
|
return const0_rtx;
|
5579 |
|
|
}
|
5580 |
|
|
|
5581 |
|
|
/* Expand an atomic test_and_set operation.
|
5582 |
|
|
bool _atomic_test_and_set (BOOL *obj, enum memmodel)
|
5583 |
|
|
EXP is the call expression. */
|
5584 |
|
|
|
5585 |
|
|
static rtx
|
5586 |
|
|
expand_builtin_atomic_test_and_set (tree exp, rtx target)
|
5587 |
|
|
{
|
5588 |
|
|
rtx mem;
|
5589 |
|
|
enum memmodel model;
|
5590 |
|
|
enum machine_mode mode;
|
5591 |
|
|
|
5592 |
|
|
mode = mode_for_size (BOOL_TYPE_SIZE, MODE_INT, 0);
|
5593 |
|
|
mem = get_builtin_sync_mem (CALL_EXPR_ARG (exp, 0), mode);
|
5594 |
|
|
model = get_memmodel (CALL_EXPR_ARG (exp, 1));
|
5595 |
|
|
|
5596 |
|
|
return expand_atomic_test_and_set (target, mem, model);
|
5597 |
|
|
}
|
5598 |
|
|
|
5599 |
|
|
|
5600 |
|
|
/* Return true if (optional) argument ARG1 of size ARG0 is always lock free on
|
5601 |
|
|
this architecture. If ARG1 is NULL, use typical alignment for size ARG0. */
|
5602 |
|
|
|
5603 |
|
|
static tree
|
5604 |
|
|
fold_builtin_atomic_always_lock_free (tree arg0, tree arg1)
|
5605 |
|
|
{
|
5606 |
|
|
int size;
|
5607 |
|
|
enum machine_mode mode;
|
5608 |
|
|
unsigned int mode_align, type_align;
|
5609 |
|
|
|
5610 |
|
|
if (TREE_CODE (arg0) != INTEGER_CST)
|
5611 |
|
|
return NULL_TREE;
|
5612 |
|
|
|
5613 |
|
|
size = INTVAL (expand_normal (arg0)) * BITS_PER_UNIT;
|
5614 |
|
|
mode = mode_for_size (size, MODE_INT, 0);
|
5615 |
|
|
mode_align = GET_MODE_ALIGNMENT (mode);
|
5616 |
|
|
|
5617 |
|
|
if (TREE_CODE (arg1) == INTEGER_CST && INTVAL (expand_normal (arg1)) == 0)
|
5618 |
|
|
type_align = mode_align;
|
5619 |
|
|
else
|
5620 |
|
|
{
|
5621 |
|
|
tree ttype = TREE_TYPE (arg1);
|
5622 |
|
|
|
5623 |
|
|
/* This function is usually invoked and folded immediately by the front
|
5624 |
|
|
end before anything else has a chance to look at it. The pointer
|
5625 |
|
|
parameter at this point is usually cast to a void *, so check for that
|
5626 |
|
|
and look past the cast. */
|
5627 |
|
|
if (TREE_CODE (arg1) == NOP_EXPR && POINTER_TYPE_P (ttype)
|
5628 |
|
|
&& VOID_TYPE_P (TREE_TYPE (ttype)))
|
5629 |
|
|
arg1 = TREE_OPERAND (arg1, 0);
|
5630 |
|
|
|
5631 |
|
|
ttype = TREE_TYPE (arg1);
|
5632 |
|
|
gcc_assert (POINTER_TYPE_P (ttype));
|
5633 |
|
|
|
5634 |
|
|
/* Get the underlying type of the object. */
|
5635 |
|
|
ttype = TREE_TYPE (ttype);
|
5636 |
|
|
type_align = TYPE_ALIGN (ttype);
|
5637 |
|
|
}
|
5638 |
|
|
|
5639 |
|
|
/* If the object has smaller alignment, the the lock free routines cannot
|
5640 |
|
|
be used. */
|
5641 |
|
|
if (type_align < mode_align)
|
5642 |
|
|
return boolean_false_node;
|
5643 |
|
|
|
5644 |
|
|
/* Check if a compare_and_swap pattern exists for the mode which represents
|
5645 |
|
|
the required size. The pattern is not allowed to fail, so the existence
|
5646 |
|
|
of the pattern indicates support is present. */
|
5647 |
|
|
if (can_compare_and_swap_p (mode, true))
|
5648 |
|
|
return boolean_true_node;
|
5649 |
|
|
else
|
5650 |
|
|
return boolean_false_node;
|
5651 |
|
|
}
|
5652 |
|
|
|
5653 |
|
|
/* Return true if the parameters to call EXP represent an object which will
|
5654 |
|
|
always generate lock free instructions. The first argument represents the
|
5655 |
|
|
size of the object, and the second parameter is a pointer to the object
|
5656 |
|
|
itself. If NULL is passed for the object, then the result is based on
|
5657 |
|
|
typical alignment for an object of the specified size. Otherwise return
|
5658 |
|
|
false. */
|
5659 |
|
|
|
5660 |
|
|
static rtx
|
5661 |
|
|
expand_builtin_atomic_always_lock_free (tree exp)
|
5662 |
|
|
{
|
5663 |
|
|
tree size;
|
5664 |
|
|
tree arg0 = CALL_EXPR_ARG (exp, 0);
|
5665 |
|
|
tree arg1 = CALL_EXPR_ARG (exp, 1);
|
5666 |
|
|
|
5667 |
|
|
if (TREE_CODE (arg0) != INTEGER_CST)
|
5668 |
|
|
{
|
5669 |
|
|
error ("non-constant argument 1 to __atomic_always_lock_free");
|
5670 |
|
|
return const0_rtx;
|
5671 |
|
|
}
|
5672 |
|
|
|
5673 |
|
|
size = fold_builtin_atomic_always_lock_free (arg0, arg1);
|
5674 |
|
|
if (size == boolean_true_node)
|
5675 |
|
|
return const1_rtx;
|
5676 |
|
|
return const0_rtx;
|
5677 |
|
|
}
|
5678 |
|
|
|
5679 |
|
|
/* Return a one or zero if it can be determined that object ARG1 of size ARG
|
5680 |
|
|
is lock free on this architecture. */
|
5681 |
|
|
|
5682 |
|
|
static tree
|
5683 |
|
|
fold_builtin_atomic_is_lock_free (tree arg0, tree arg1)
|
5684 |
|
|
{
|
5685 |
|
|
if (!flag_inline_atomics)
|
5686 |
|
|
return NULL_TREE;
|
5687 |
|
|
|
5688 |
|
|
/* If it isn't always lock free, don't generate a result. */
|
5689 |
|
|
if (fold_builtin_atomic_always_lock_free (arg0, arg1) == boolean_true_node)
|
5690 |
|
|
return boolean_true_node;
|
5691 |
|
|
|
5692 |
|
|
return NULL_TREE;
|
5693 |
|
|
}
|
5694 |
|
|
|
5695 |
|
|
/* Return true if the parameters to call EXP represent an object which will
|
5696 |
|
|
always generate lock free instructions. The first argument represents the
|
5697 |
|
|
size of the object, and the second parameter is a pointer to the object
|
5698 |
|
|
itself. If NULL is passed for the object, then the result is based on
|
5699 |
|
|
typical alignment for an object of the specified size. Otherwise return
|
5700 |
|
|
NULL*/
|
5701 |
|
|
|
5702 |
|
|
static rtx
|
5703 |
|
|
expand_builtin_atomic_is_lock_free (tree exp)
|
5704 |
|
|
{
|
5705 |
|
|
tree size;
|
5706 |
|
|
tree arg0 = CALL_EXPR_ARG (exp, 0);
|
5707 |
|
|
tree arg1 = CALL_EXPR_ARG (exp, 1);
|
5708 |
|
|
|
5709 |
|
|
if (!INTEGRAL_TYPE_P (TREE_TYPE (arg0)))
|
5710 |
|
|
{
|
5711 |
|
|
error ("non-integer argument 1 to __atomic_is_lock_free");
|
5712 |
|
|
return NULL_RTX;
|
5713 |
|
|
}
|
5714 |
|
|
|
5715 |
|
|
if (!flag_inline_atomics)
|
5716 |
|
|
return NULL_RTX;
|
5717 |
|
|
|
5718 |
|
|
/* If the value is known at compile time, return the RTX for it. */
|
5719 |
|
|
size = fold_builtin_atomic_is_lock_free (arg0, arg1);
|
5720 |
|
|
if (size == boolean_true_node)
|
5721 |
|
|
return const1_rtx;
|
5722 |
|
|
|
5723 |
|
|
return NULL_RTX;
|
5724 |
|
|
}
|
5725 |
|
|
|
5726 |
|
|
/* Expand the __atomic_thread_fence intrinsic:
|
5727 |
|
|
void __atomic_thread_fence (enum memmodel)
|
5728 |
|
|
EXP is the CALL_EXPR. */
|
5729 |
|
|
|
5730 |
|
|
static void
|
5731 |
|
|
expand_builtin_atomic_thread_fence (tree exp)
|
5732 |
|
|
{
|
5733 |
|
|
enum memmodel model = get_memmodel (CALL_EXPR_ARG (exp, 0));
|
5734 |
|
|
expand_mem_thread_fence (model);
|
5735 |
|
|
}
|
5736 |
|
|
|
5737 |
|
|
/* Expand the __atomic_signal_fence intrinsic:
|
5738 |
|
|
void __atomic_signal_fence (enum memmodel)
|
5739 |
|
|
EXP is the CALL_EXPR. */
|
5740 |
|
|
|
5741 |
|
|
static void
|
5742 |
|
|
expand_builtin_atomic_signal_fence (tree exp)
|
5743 |
|
|
{
|
5744 |
|
|
enum memmodel model = get_memmodel (CALL_EXPR_ARG (exp, 0));
|
5745 |
|
|
expand_mem_signal_fence (model);
|
5746 |
|
|
}
|
5747 |
|
|
|
5748 |
|
|
/* Expand the __sync_synchronize intrinsic. */
|
5749 |
|
|
|
5750 |
|
|
static void
|
5751 |
|
|
expand_builtin_sync_synchronize (void)
|
5752 |
|
|
{
|
5753 |
|
|
expand_mem_thread_fence (MEMMODEL_SEQ_CST);
|
5754 |
|
|
}
|
5755 |
|
|
|
5756 |
|
|
|
5757 |
|
|
/* Expand an expression EXP that calls a built-in function,
|
5758 |
|
|
with result going to TARGET if that's convenient
|
5759 |
|
|
(and in mode MODE if that's convenient).
|
5760 |
|
|
SUBTARGET may be used as the target for computing one of EXP's operands.
|
5761 |
|
|
IGNORE is nonzero if the value is to be ignored. */
|
5762 |
|
|
|
5763 |
|
|
rtx
|
5764 |
|
|
expand_builtin (tree exp, rtx target, rtx subtarget, enum machine_mode mode,
|
5765 |
|
|
int ignore)
|
5766 |
|
|
{
|
5767 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
5768 |
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
5769 |
|
|
enum machine_mode target_mode = TYPE_MODE (TREE_TYPE (exp));
|
5770 |
|
|
int flags;
|
5771 |
|
|
|
5772 |
|
|
if (DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_MD)
|
5773 |
|
|
return targetm.expand_builtin (exp, target, subtarget, mode, ignore);
|
5774 |
|
|
|
5775 |
|
|
/* When not optimizing, generate calls to library functions for a certain
|
5776 |
|
|
set of builtins. */
|
5777 |
|
|
if (!optimize
|
5778 |
|
|
&& !called_as_built_in (fndecl)
|
5779 |
|
|
&& fcode != BUILT_IN_ALLOCA
|
5780 |
|
|
&& fcode != BUILT_IN_ALLOCA_WITH_ALIGN
|
5781 |
|
|
&& fcode != BUILT_IN_FREE)
|
5782 |
|
|
return expand_call (exp, target, ignore);
|
5783 |
|
|
|
5784 |
|
|
/* The built-in function expanders test for target == const0_rtx
|
5785 |
|
|
to determine whether the function's result will be ignored. */
|
5786 |
|
|
if (ignore)
|
5787 |
|
|
target = const0_rtx;
|
5788 |
|
|
|
5789 |
|
|
/* If the result of a pure or const built-in function is ignored, and
|
5790 |
|
|
none of its arguments are volatile, we can avoid expanding the
|
5791 |
|
|
built-in call and just evaluate the arguments for side-effects. */
|
5792 |
|
|
if (target == const0_rtx
|
5793 |
|
|
&& ((flags = flags_from_decl_or_type (fndecl)) & (ECF_CONST | ECF_PURE))
|
5794 |
|
|
&& !(flags & ECF_LOOPING_CONST_OR_PURE))
|
5795 |
|
|
{
|
5796 |
|
|
bool volatilep = false;
|
5797 |
|
|
tree arg;
|
5798 |
|
|
call_expr_arg_iterator iter;
|
5799 |
|
|
|
5800 |
|
|
FOR_EACH_CALL_EXPR_ARG (arg, iter, exp)
|
5801 |
|
|
if (TREE_THIS_VOLATILE (arg))
|
5802 |
|
|
{
|
5803 |
|
|
volatilep = true;
|
5804 |
|
|
break;
|
5805 |
|
|
}
|
5806 |
|
|
|
5807 |
|
|
if (! volatilep)
|
5808 |
|
|
{
|
5809 |
|
|
FOR_EACH_CALL_EXPR_ARG (arg, iter, exp)
|
5810 |
|
|
expand_expr (arg, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
5811 |
|
|
return const0_rtx;
|
5812 |
|
|
}
|
5813 |
|
|
}
|
5814 |
|
|
|
5815 |
|
|
switch (fcode)
|
5816 |
|
|
{
|
5817 |
|
|
CASE_FLT_FN (BUILT_IN_FABS):
|
5818 |
|
|
target = expand_builtin_fabs (exp, target, subtarget);
|
5819 |
|
|
if (target)
|
5820 |
|
|
return target;
|
5821 |
|
|
break;
|
5822 |
|
|
|
5823 |
|
|
CASE_FLT_FN (BUILT_IN_COPYSIGN):
|
5824 |
|
|
target = expand_builtin_copysign (exp, target, subtarget);
|
5825 |
|
|
if (target)
|
5826 |
|
|
return target;
|
5827 |
|
|
break;
|
5828 |
|
|
|
5829 |
|
|
/* Just do a normal library call if we were unable to fold
|
5830 |
|
|
the values. */
|
5831 |
|
|
CASE_FLT_FN (BUILT_IN_CABS):
|
5832 |
|
|
break;
|
5833 |
|
|
|
5834 |
|
|
CASE_FLT_FN (BUILT_IN_EXP):
|
5835 |
|
|
CASE_FLT_FN (BUILT_IN_EXP10):
|
5836 |
|
|
CASE_FLT_FN (BUILT_IN_POW10):
|
5837 |
|
|
CASE_FLT_FN (BUILT_IN_EXP2):
|
5838 |
|
|
CASE_FLT_FN (BUILT_IN_EXPM1):
|
5839 |
|
|
CASE_FLT_FN (BUILT_IN_LOGB):
|
5840 |
|
|
CASE_FLT_FN (BUILT_IN_LOG):
|
5841 |
|
|
CASE_FLT_FN (BUILT_IN_LOG10):
|
5842 |
|
|
CASE_FLT_FN (BUILT_IN_LOG2):
|
5843 |
|
|
CASE_FLT_FN (BUILT_IN_LOG1P):
|
5844 |
|
|
CASE_FLT_FN (BUILT_IN_TAN):
|
5845 |
|
|
CASE_FLT_FN (BUILT_IN_ASIN):
|
5846 |
|
|
CASE_FLT_FN (BUILT_IN_ACOS):
|
5847 |
|
|
CASE_FLT_FN (BUILT_IN_ATAN):
|
5848 |
|
|
CASE_FLT_FN (BUILT_IN_SIGNIFICAND):
|
5849 |
|
|
/* Treat these like sqrt only if unsafe math optimizations are allowed,
|
5850 |
|
|
because of possible accuracy problems. */
|
5851 |
|
|
if (! flag_unsafe_math_optimizations)
|
5852 |
|
|
break;
|
5853 |
|
|
CASE_FLT_FN (BUILT_IN_SQRT):
|
5854 |
|
|
CASE_FLT_FN (BUILT_IN_FLOOR):
|
5855 |
|
|
CASE_FLT_FN (BUILT_IN_CEIL):
|
5856 |
|
|
CASE_FLT_FN (BUILT_IN_TRUNC):
|
5857 |
|
|
CASE_FLT_FN (BUILT_IN_ROUND):
|
5858 |
|
|
CASE_FLT_FN (BUILT_IN_NEARBYINT):
|
5859 |
|
|
CASE_FLT_FN (BUILT_IN_RINT):
|
5860 |
|
|
target = expand_builtin_mathfn (exp, target, subtarget);
|
5861 |
|
|
if (target)
|
5862 |
|
|
return target;
|
5863 |
|
|
break;
|
5864 |
|
|
|
5865 |
|
|
CASE_FLT_FN (BUILT_IN_FMA):
|
5866 |
|
|
target = expand_builtin_mathfn_ternary (exp, target, subtarget);
|
5867 |
|
|
if (target)
|
5868 |
|
|
return target;
|
5869 |
|
|
break;
|
5870 |
|
|
|
5871 |
|
|
CASE_FLT_FN (BUILT_IN_ILOGB):
|
5872 |
|
|
if (! flag_unsafe_math_optimizations)
|
5873 |
|
|
break;
|
5874 |
|
|
CASE_FLT_FN (BUILT_IN_ISINF):
|
5875 |
|
|
CASE_FLT_FN (BUILT_IN_FINITE):
|
5876 |
|
|
case BUILT_IN_ISFINITE:
|
5877 |
|
|
case BUILT_IN_ISNORMAL:
|
5878 |
|
|
target = expand_builtin_interclass_mathfn (exp, target);
|
5879 |
|
|
if (target)
|
5880 |
|
|
return target;
|
5881 |
|
|
break;
|
5882 |
|
|
|
5883 |
|
|
CASE_FLT_FN (BUILT_IN_ICEIL):
|
5884 |
|
|
CASE_FLT_FN (BUILT_IN_LCEIL):
|
5885 |
|
|
CASE_FLT_FN (BUILT_IN_LLCEIL):
|
5886 |
|
|
CASE_FLT_FN (BUILT_IN_LFLOOR):
|
5887 |
|
|
CASE_FLT_FN (BUILT_IN_IFLOOR):
|
5888 |
|
|
CASE_FLT_FN (BUILT_IN_LLFLOOR):
|
5889 |
|
|
target = expand_builtin_int_roundingfn (exp, target);
|
5890 |
|
|
if (target)
|
5891 |
|
|
return target;
|
5892 |
|
|
break;
|
5893 |
|
|
|
5894 |
|
|
CASE_FLT_FN (BUILT_IN_IRINT):
|
5895 |
|
|
CASE_FLT_FN (BUILT_IN_LRINT):
|
5896 |
|
|
CASE_FLT_FN (BUILT_IN_LLRINT):
|
5897 |
|
|
CASE_FLT_FN (BUILT_IN_IROUND):
|
5898 |
|
|
CASE_FLT_FN (BUILT_IN_LROUND):
|
5899 |
|
|
CASE_FLT_FN (BUILT_IN_LLROUND):
|
5900 |
|
|
target = expand_builtin_int_roundingfn_2 (exp, target);
|
5901 |
|
|
if (target)
|
5902 |
|
|
return target;
|
5903 |
|
|
break;
|
5904 |
|
|
|
5905 |
|
|
CASE_FLT_FN (BUILT_IN_POWI):
|
5906 |
|
|
target = expand_builtin_powi (exp, target);
|
5907 |
|
|
if (target)
|
5908 |
|
|
return target;
|
5909 |
|
|
break;
|
5910 |
|
|
|
5911 |
|
|
CASE_FLT_FN (BUILT_IN_ATAN2):
|
5912 |
|
|
CASE_FLT_FN (BUILT_IN_LDEXP):
|
5913 |
|
|
CASE_FLT_FN (BUILT_IN_SCALB):
|
5914 |
|
|
CASE_FLT_FN (BUILT_IN_SCALBN):
|
5915 |
|
|
CASE_FLT_FN (BUILT_IN_SCALBLN):
|
5916 |
|
|
if (! flag_unsafe_math_optimizations)
|
5917 |
|
|
break;
|
5918 |
|
|
|
5919 |
|
|
CASE_FLT_FN (BUILT_IN_FMOD):
|
5920 |
|
|
CASE_FLT_FN (BUILT_IN_REMAINDER):
|
5921 |
|
|
CASE_FLT_FN (BUILT_IN_DREM):
|
5922 |
|
|
CASE_FLT_FN (BUILT_IN_POW):
|
5923 |
|
|
target = expand_builtin_mathfn_2 (exp, target, subtarget);
|
5924 |
|
|
if (target)
|
5925 |
|
|
return target;
|
5926 |
|
|
break;
|
5927 |
|
|
|
5928 |
|
|
CASE_FLT_FN (BUILT_IN_CEXPI):
|
5929 |
|
|
target = expand_builtin_cexpi (exp, target);
|
5930 |
|
|
gcc_assert (target);
|
5931 |
|
|
return target;
|
5932 |
|
|
|
5933 |
|
|
CASE_FLT_FN (BUILT_IN_SIN):
|
5934 |
|
|
CASE_FLT_FN (BUILT_IN_COS):
|
5935 |
|
|
if (! flag_unsafe_math_optimizations)
|
5936 |
|
|
break;
|
5937 |
|
|
target = expand_builtin_mathfn_3 (exp, target, subtarget);
|
5938 |
|
|
if (target)
|
5939 |
|
|
return target;
|
5940 |
|
|
break;
|
5941 |
|
|
|
5942 |
|
|
CASE_FLT_FN (BUILT_IN_SINCOS):
|
5943 |
|
|
if (! flag_unsafe_math_optimizations)
|
5944 |
|
|
break;
|
5945 |
|
|
target = expand_builtin_sincos (exp);
|
5946 |
|
|
if (target)
|
5947 |
|
|
return target;
|
5948 |
|
|
break;
|
5949 |
|
|
|
5950 |
|
|
case BUILT_IN_APPLY_ARGS:
|
5951 |
|
|
return expand_builtin_apply_args ();
|
5952 |
|
|
|
5953 |
|
|
/* __builtin_apply (FUNCTION, ARGUMENTS, ARGSIZE) invokes
|
5954 |
|
|
FUNCTION with a copy of the parameters described by
|
5955 |
|
|
ARGUMENTS, and ARGSIZE. It returns a block of memory
|
5956 |
|
|
allocated on the stack into which is stored all the registers
|
5957 |
|
|
that might possibly be used for returning the result of a
|
5958 |
|
|
function. ARGUMENTS is the value returned by
|
5959 |
|
|
__builtin_apply_args. ARGSIZE is the number of bytes of
|
5960 |
|
|
arguments that must be copied. ??? How should this value be
|
5961 |
|
|
computed? We'll also need a safe worst case value for varargs
|
5962 |
|
|
functions. */
|
5963 |
|
|
case BUILT_IN_APPLY:
|
5964 |
|
|
if (!validate_arglist (exp, POINTER_TYPE,
|
5965 |
|
|
POINTER_TYPE, INTEGER_TYPE, VOID_TYPE)
|
5966 |
|
|
&& !validate_arglist (exp, REFERENCE_TYPE,
|
5967 |
|
|
POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
5968 |
|
|
return const0_rtx;
|
5969 |
|
|
else
|
5970 |
|
|
{
|
5971 |
|
|
rtx ops[3];
|
5972 |
|
|
|
5973 |
|
|
ops[0] = expand_normal (CALL_EXPR_ARG (exp, 0));
|
5974 |
|
|
ops[1] = expand_normal (CALL_EXPR_ARG (exp, 1));
|
5975 |
|
|
ops[2] = expand_normal (CALL_EXPR_ARG (exp, 2));
|
5976 |
|
|
|
5977 |
|
|
return expand_builtin_apply (ops[0], ops[1], ops[2]);
|
5978 |
|
|
}
|
5979 |
|
|
|
5980 |
|
|
/* __builtin_return (RESULT) causes the function to return the
|
5981 |
|
|
value described by RESULT. RESULT is address of the block of
|
5982 |
|
|
memory returned by __builtin_apply. */
|
5983 |
|
|
case BUILT_IN_RETURN:
|
5984 |
|
|
if (validate_arglist (exp, POINTER_TYPE, VOID_TYPE))
|
5985 |
|
|
expand_builtin_return (expand_normal (CALL_EXPR_ARG (exp, 0)));
|
5986 |
|
|
return const0_rtx;
|
5987 |
|
|
|
5988 |
|
|
case BUILT_IN_SAVEREGS:
|
5989 |
|
|
return expand_builtin_saveregs ();
|
5990 |
|
|
|
5991 |
|
|
case BUILT_IN_VA_ARG_PACK:
|
5992 |
|
|
/* All valid uses of __builtin_va_arg_pack () are removed during
|
5993 |
|
|
inlining. */
|
5994 |
|
|
error ("%Kinvalid use of %<__builtin_va_arg_pack ()%>", exp);
|
5995 |
|
|
return const0_rtx;
|
5996 |
|
|
|
5997 |
|
|
case BUILT_IN_VA_ARG_PACK_LEN:
|
5998 |
|
|
/* All valid uses of __builtin_va_arg_pack_len () are removed during
|
5999 |
|
|
inlining. */
|
6000 |
|
|
error ("%Kinvalid use of %<__builtin_va_arg_pack_len ()%>", exp);
|
6001 |
|
|
return const0_rtx;
|
6002 |
|
|
|
6003 |
|
|
/* Return the address of the first anonymous stack arg. */
|
6004 |
|
|
case BUILT_IN_NEXT_ARG:
|
6005 |
|
|
if (fold_builtin_next_arg (exp, false))
|
6006 |
|
|
return const0_rtx;
|
6007 |
|
|
return expand_builtin_next_arg ();
|
6008 |
|
|
|
6009 |
|
|
case BUILT_IN_CLEAR_CACHE:
|
6010 |
|
|
target = expand_builtin___clear_cache (exp);
|
6011 |
|
|
if (target)
|
6012 |
|
|
return target;
|
6013 |
|
|
break;
|
6014 |
|
|
|
6015 |
|
|
case BUILT_IN_CLASSIFY_TYPE:
|
6016 |
|
|
return expand_builtin_classify_type (exp);
|
6017 |
|
|
|
6018 |
|
|
case BUILT_IN_CONSTANT_P:
|
6019 |
|
|
return const0_rtx;
|
6020 |
|
|
|
6021 |
|
|
case BUILT_IN_FRAME_ADDRESS:
|
6022 |
|
|
case BUILT_IN_RETURN_ADDRESS:
|
6023 |
|
|
return expand_builtin_frame_address (fndecl, exp);
|
6024 |
|
|
|
6025 |
|
|
/* Returns the address of the area where the structure is returned.
|
6026 |
|
|
|
6027 |
|
|
case BUILT_IN_AGGREGATE_INCOMING_ADDRESS:
|
6028 |
|
|
if (call_expr_nargs (exp) != 0
|
6029 |
|
|
|| ! AGGREGATE_TYPE_P (TREE_TYPE (TREE_TYPE (current_function_decl)))
|
6030 |
|
|
|| !MEM_P (DECL_RTL (DECL_RESULT (current_function_decl))))
|
6031 |
|
|
return const0_rtx;
|
6032 |
|
|
else
|
6033 |
|
|
return XEXP (DECL_RTL (DECL_RESULT (current_function_decl)), 0);
|
6034 |
|
|
|
6035 |
|
|
case BUILT_IN_ALLOCA:
|
6036 |
|
|
case BUILT_IN_ALLOCA_WITH_ALIGN:
|
6037 |
|
|
/* If the allocation stems from the declaration of a variable-sized
|
6038 |
|
|
object, it cannot accumulate. */
|
6039 |
|
|
target = expand_builtin_alloca (exp, CALL_ALLOCA_FOR_VAR_P (exp));
|
6040 |
|
|
if (target)
|
6041 |
|
|
return target;
|
6042 |
|
|
break;
|
6043 |
|
|
|
6044 |
|
|
case BUILT_IN_STACK_SAVE:
|
6045 |
|
|
return expand_stack_save ();
|
6046 |
|
|
|
6047 |
|
|
case BUILT_IN_STACK_RESTORE:
|
6048 |
|
|
expand_stack_restore (CALL_EXPR_ARG (exp, 0));
|
6049 |
|
|
return const0_rtx;
|
6050 |
|
|
|
6051 |
|
|
case BUILT_IN_BSWAP32:
|
6052 |
|
|
case BUILT_IN_BSWAP64:
|
6053 |
|
|
target = expand_builtin_bswap (exp, target, subtarget);
|
6054 |
|
|
|
6055 |
|
|
if (target)
|
6056 |
|
|
return target;
|
6057 |
|
|
break;
|
6058 |
|
|
|
6059 |
|
|
CASE_INT_FN (BUILT_IN_FFS):
|
6060 |
|
|
case BUILT_IN_FFSIMAX:
|
6061 |
|
|
target = expand_builtin_unop (target_mode, exp, target,
|
6062 |
|
|
subtarget, ffs_optab);
|
6063 |
|
|
if (target)
|
6064 |
|
|
return target;
|
6065 |
|
|
break;
|
6066 |
|
|
|
6067 |
|
|
CASE_INT_FN (BUILT_IN_CLZ):
|
6068 |
|
|
case BUILT_IN_CLZIMAX:
|
6069 |
|
|
target = expand_builtin_unop (target_mode, exp, target,
|
6070 |
|
|
subtarget, clz_optab);
|
6071 |
|
|
if (target)
|
6072 |
|
|
return target;
|
6073 |
|
|
break;
|
6074 |
|
|
|
6075 |
|
|
CASE_INT_FN (BUILT_IN_CTZ):
|
6076 |
|
|
case BUILT_IN_CTZIMAX:
|
6077 |
|
|
target = expand_builtin_unop (target_mode, exp, target,
|
6078 |
|
|
subtarget, ctz_optab);
|
6079 |
|
|
if (target)
|
6080 |
|
|
return target;
|
6081 |
|
|
break;
|
6082 |
|
|
|
6083 |
|
|
CASE_INT_FN (BUILT_IN_CLRSB):
|
6084 |
|
|
case BUILT_IN_CLRSBIMAX:
|
6085 |
|
|
target = expand_builtin_unop (target_mode, exp, target,
|
6086 |
|
|
subtarget, clrsb_optab);
|
6087 |
|
|
if (target)
|
6088 |
|
|
return target;
|
6089 |
|
|
break;
|
6090 |
|
|
|
6091 |
|
|
CASE_INT_FN (BUILT_IN_POPCOUNT):
|
6092 |
|
|
case BUILT_IN_POPCOUNTIMAX:
|
6093 |
|
|
target = expand_builtin_unop (target_mode, exp, target,
|
6094 |
|
|
subtarget, popcount_optab);
|
6095 |
|
|
if (target)
|
6096 |
|
|
return target;
|
6097 |
|
|
break;
|
6098 |
|
|
|
6099 |
|
|
CASE_INT_FN (BUILT_IN_PARITY):
|
6100 |
|
|
case BUILT_IN_PARITYIMAX:
|
6101 |
|
|
target = expand_builtin_unop (target_mode, exp, target,
|
6102 |
|
|
subtarget, parity_optab);
|
6103 |
|
|
if (target)
|
6104 |
|
|
return target;
|
6105 |
|
|
break;
|
6106 |
|
|
|
6107 |
|
|
case BUILT_IN_STRLEN:
|
6108 |
|
|
target = expand_builtin_strlen (exp, target, target_mode);
|
6109 |
|
|
if (target)
|
6110 |
|
|
return target;
|
6111 |
|
|
break;
|
6112 |
|
|
|
6113 |
|
|
case BUILT_IN_STRCPY:
|
6114 |
|
|
target = expand_builtin_strcpy (exp, target);
|
6115 |
|
|
if (target)
|
6116 |
|
|
return target;
|
6117 |
|
|
break;
|
6118 |
|
|
|
6119 |
|
|
case BUILT_IN_STRNCPY:
|
6120 |
|
|
target = expand_builtin_strncpy (exp, target);
|
6121 |
|
|
if (target)
|
6122 |
|
|
return target;
|
6123 |
|
|
break;
|
6124 |
|
|
|
6125 |
|
|
case BUILT_IN_STPCPY:
|
6126 |
|
|
target = expand_builtin_stpcpy (exp, target, mode);
|
6127 |
|
|
if (target)
|
6128 |
|
|
return target;
|
6129 |
|
|
break;
|
6130 |
|
|
|
6131 |
|
|
case BUILT_IN_MEMCPY:
|
6132 |
|
|
target = expand_builtin_memcpy (exp, target);
|
6133 |
|
|
if (target)
|
6134 |
|
|
return target;
|
6135 |
|
|
break;
|
6136 |
|
|
|
6137 |
|
|
case BUILT_IN_MEMPCPY:
|
6138 |
|
|
target = expand_builtin_mempcpy (exp, target, mode);
|
6139 |
|
|
if (target)
|
6140 |
|
|
return target;
|
6141 |
|
|
break;
|
6142 |
|
|
|
6143 |
|
|
case BUILT_IN_MEMSET:
|
6144 |
|
|
target = expand_builtin_memset (exp, target, mode);
|
6145 |
|
|
if (target)
|
6146 |
|
|
return target;
|
6147 |
|
|
break;
|
6148 |
|
|
|
6149 |
|
|
case BUILT_IN_BZERO:
|
6150 |
|
|
target = expand_builtin_bzero (exp);
|
6151 |
|
|
if (target)
|
6152 |
|
|
return target;
|
6153 |
|
|
break;
|
6154 |
|
|
|
6155 |
|
|
case BUILT_IN_STRCMP:
|
6156 |
|
|
target = expand_builtin_strcmp (exp, target);
|
6157 |
|
|
if (target)
|
6158 |
|
|
return target;
|
6159 |
|
|
break;
|
6160 |
|
|
|
6161 |
|
|
case BUILT_IN_STRNCMP:
|
6162 |
|
|
target = expand_builtin_strncmp (exp, target, mode);
|
6163 |
|
|
if (target)
|
6164 |
|
|
return target;
|
6165 |
|
|
break;
|
6166 |
|
|
|
6167 |
|
|
case BUILT_IN_BCMP:
|
6168 |
|
|
case BUILT_IN_MEMCMP:
|
6169 |
|
|
target = expand_builtin_memcmp (exp, target, mode);
|
6170 |
|
|
if (target)
|
6171 |
|
|
return target;
|
6172 |
|
|
break;
|
6173 |
|
|
|
6174 |
|
|
case BUILT_IN_SETJMP:
|
6175 |
|
|
/* This should have been lowered to the builtins below. */
|
6176 |
|
|
gcc_unreachable ();
|
6177 |
|
|
|
6178 |
|
|
case BUILT_IN_SETJMP_SETUP:
|
6179 |
|
|
/* __builtin_setjmp_setup is passed a pointer to an array of five words
|
6180 |
|
|
and the receiver label. */
|
6181 |
|
|
if (validate_arglist (exp, POINTER_TYPE, POINTER_TYPE, VOID_TYPE))
|
6182 |
|
|
{
|
6183 |
|
|
rtx buf_addr = expand_expr (CALL_EXPR_ARG (exp, 0), subtarget,
|
6184 |
|
|
VOIDmode, EXPAND_NORMAL);
|
6185 |
|
|
tree label = TREE_OPERAND (CALL_EXPR_ARG (exp, 1), 0);
|
6186 |
|
|
rtx label_r = label_rtx (label);
|
6187 |
|
|
|
6188 |
|
|
/* This is copied from the handling of non-local gotos. */
|
6189 |
|
|
expand_builtin_setjmp_setup (buf_addr, label_r);
|
6190 |
|
|
nonlocal_goto_handler_labels
|
6191 |
|
|
= gen_rtx_EXPR_LIST (VOIDmode, label_r,
|
6192 |
|
|
nonlocal_goto_handler_labels);
|
6193 |
|
|
/* ??? Do not let expand_label treat us as such since we would
|
6194 |
|
|
not want to be both on the list of non-local labels and on
|
6195 |
|
|
the list of forced labels. */
|
6196 |
|
|
FORCED_LABEL (label) = 0;
|
6197 |
|
|
return const0_rtx;
|
6198 |
|
|
}
|
6199 |
|
|
break;
|
6200 |
|
|
|
6201 |
|
|
case BUILT_IN_SETJMP_DISPATCHER:
|
6202 |
|
|
/* __builtin_setjmp_dispatcher is passed the dispatcher label. */
|
6203 |
|
|
if (validate_arglist (exp, POINTER_TYPE, VOID_TYPE))
|
6204 |
|
|
{
|
6205 |
|
|
tree label = TREE_OPERAND (CALL_EXPR_ARG (exp, 0), 0);
|
6206 |
|
|
rtx label_r = label_rtx (label);
|
6207 |
|
|
|
6208 |
|
|
/* Remove the dispatcher label from the list of non-local labels
|
6209 |
|
|
since the receiver labels have been added to it above. */
|
6210 |
|
|
remove_node_from_expr_list (label_r, &nonlocal_goto_handler_labels);
|
6211 |
|
|
return const0_rtx;
|
6212 |
|
|
}
|
6213 |
|
|
break;
|
6214 |
|
|
|
6215 |
|
|
case BUILT_IN_SETJMP_RECEIVER:
|
6216 |
|
|
/* __builtin_setjmp_receiver is passed the receiver label. */
|
6217 |
|
|
if (validate_arglist (exp, POINTER_TYPE, VOID_TYPE))
|
6218 |
|
|
{
|
6219 |
|
|
tree label = TREE_OPERAND (CALL_EXPR_ARG (exp, 0), 0);
|
6220 |
|
|
rtx label_r = label_rtx (label);
|
6221 |
|
|
|
6222 |
|
|
expand_builtin_setjmp_receiver (label_r);
|
6223 |
|
|
return const0_rtx;
|
6224 |
|
|
}
|
6225 |
|
|
break;
|
6226 |
|
|
|
6227 |
|
|
/* __builtin_longjmp is passed a pointer to an array of five words.
|
6228 |
|
|
It's similar to the C library longjmp function but works with
|
6229 |
|
|
__builtin_setjmp above. */
|
6230 |
|
|
case BUILT_IN_LONGJMP:
|
6231 |
|
|
if (validate_arglist (exp, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
6232 |
|
|
{
|
6233 |
|
|
rtx buf_addr = expand_expr (CALL_EXPR_ARG (exp, 0), subtarget,
|
6234 |
|
|
VOIDmode, EXPAND_NORMAL);
|
6235 |
|
|
rtx value = expand_normal (CALL_EXPR_ARG (exp, 1));
|
6236 |
|
|
|
6237 |
|
|
if (value != const1_rtx)
|
6238 |
|
|
{
|
6239 |
|
|
error ("%<__builtin_longjmp%> second argument must be 1");
|
6240 |
|
|
return const0_rtx;
|
6241 |
|
|
}
|
6242 |
|
|
|
6243 |
|
|
expand_builtin_longjmp (buf_addr, value);
|
6244 |
|
|
return const0_rtx;
|
6245 |
|
|
}
|
6246 |
|
|
break;
|
6247 |
|
|
|
6248 |
|
|
case BUILT_IN_NONLOCAL_GOTO:
|
6249 |
|
|
target = expand_builtin_nonlocal_goto (exp);
|
6250 |
|
|
if (target)
|
6251 |
|
|
return target;
|
6252 |
|
|
break;
|
6253 |
|
|
|
6254 |
|
|
/* This updates the setjmp buffer that is its argument with the value
|
6255 |
|
|
of the current stack pointer. */
|
6256 |
|
|
case BUILT_IN_UPDATE_SETJMP_BUF:
|
6257 |
|
|
if (validate_arglist (exp, POINTER_TYPE, VOID_TYPE))
|
6258 |
|
|
{
|
6259 |
|
|
rtx buf_addr
|
6260 |
|
|
= expand_normal (CALL_EXPR_ARG (exp, 0));
|
6261 |
|
|
|
6262 |
|
|
expand_builtin_update_setjmp_buf (buf_addr);
|
6263 |
|
|
return const0_rtx;
|
6264 |
|
|
}
|
6265 |
|
|
break;
|
6266 |
|
|
|
6267 |
|
|
case BUILT_IN_TRAP:
|
6268 |
|
|
expand_builtin_trap ();
|
6269 |
|
|
return const0_rtx;
|
6270 |
|
|
|
6271 |
|
|
case BUILT_IN_UNREACHABLE:
|
6272 |
|
|
expand_builtin_unreachable ();
|
6273 |
|
|
return const0_rtx;
|
6274 |
|
|
|
6275 |
|
|
CASE_FLT_FN (BUILT_IN_SIGNBIT):
|
6276 |
|
|
case BUILT_IN_SIGNBITD32:
|
6277 |
|
|
case BUILT_IN_SIGNBITD64:
|
6278 |
|
|
case BUILT_IN_SIGNBITD128:
|
6279 |
|
|
target = expand_builtin_signbit (exp, target);
|
6280 |
|
|
if (target)
|
6281 |
|
|
return target;
|
6282 |
|
|
break;
|
6283 |
|
|
|
6284 |
|
|
/* Various hooks for the DWARF 2 __throw routine. */
|
6285 |
|
|
case BUILT_IN_UNWIND_INIT:
|
6286 |
|
|
expand_builtin_unwind_init ();
|
6287 |
|
|
return const0_rtx;
|
6288 |
|
|
case BUILT_IN_DWARF_CFA:
|
6289 |
|
|
return virtual_cfa_rtx;
|
6290 |
|
|
#ifdef DWARF2_UNWIND_INFO
|
6291 |
|
|
case BUILT_IN_DWARF_SP_COLUMN:
|
6292 |
|
|
return expand_builtin_dwarf_sp_column ();
|
6293 |
|
|
case BUILT_IN_INIT_DWARF_REG_SIZES:
|
6294 |
|
|
expand_builtin_init_dwarf_reg_sizes (CALL_EXPR_ARG (exp, 0));
|
6295 |
|
|
return const0_rtx;
|
6296 |
|
|
#endif
|
6297 |
|
|
case BUILT_IN_FROB_RETURN_ADDR:
|
6298 |
|
|
return expand_builtin_frob_return_addr (CALL_EXPR_ARG (exp, 0));
|
6299 |
|
|
case BUILT_IN_EXTRACT_RETURN_ADDR:
|
6300 |
|
|
return expand_builtin_extract_return_addr (CALL_EXPR_ARG (exp, 0));
|
6301 |
|
|
case BUILT_IN_EH_RETURN:
|
6302 |
|
|
expand_builtin_eh_return (CALL_EXPR_ARG (exp, 0),
|
6303 |
|
|
CALL_EXPR_ARG (exp, 1));
|
6304 |
|
|
return const0_rtx;
|
6305 |
|
|
#ifdef EH_RETURN_DATA_REGNO
|
6306 |
|
|
case BUILT_IN_EH_RETURN_DATA_REGNO:
|
6307 |
|
|
return expand_builtin_eh_return_data_regno (exp);
|
6308 |
|
|
#endif
|
6309 |
|
|
case BUILT_IN_EXTEND_POINTER:
|
6310 |
|
|
return expand_builtin_extend_pointer (CALL_EXPR_ARG (exp, 0));
|
6311 |
|
|
case BUILT_IN_EH_POINTER:
|
6312 |
|
|
return expand_builtin_eh_pointer (exp);
|
6313 |
|
|
case BUILT_IN_EH_FILTER:
|
6314 |
|
|
return expand_builtin_eh_filter (exp);
|
6315 |
|
|
case BUILT_IN_EH_COPY_VALUES:
|
6316 |
|
|
return expand_builtin_eh_copy_values (exp);
|
6317 |
|
|
|
6318 |
|
|
case BUILT_IN_VA_START:
|
6319 |
|
|
return expand_builtin_va_start (exp);
|
6320 |
|
|
case BUILT_IN_VA_END:
|
6321 |
|
|
return expand_builtin_va_end (exp);
|
6322 |
|
|
case BUILT_IN_VA_COPY:
|
6323 |
|
|
return expand_builtin_va_copy (exp);
|
6324 |
|
|
case BUILT_IN_EXPECT:
|
6325 |
|
|
return expand_builtin_expect (exp, target);
|
6326 |
|
|
case BUILT_IN_ASSUME_ALIGNED:
|
6327 |
|
|
return expand_builtin_assume_aligned (exp, target);
|
6328 |
|
|
case BUILT_IN_PREFETCH:
|
6329 |
|
|
expand_builtin_prefetch (exp);
|
6330 |
|
|
return const0_rtx;
|
6331 |
|
|
|
6332 |
|
|
case BUILT_IN_INIT_TRAMPOLINE:
|
6333 |
|
|
return expand_builtin_init_trampoline (exp, true);
|
6334 |
|
|
case BUILT_IN_INIT_HEAP_TRAMPOLINE:
|
6335 |
|
|
return expand_builtin_init_trampoline (exp, false);
|
6336 |
|
|
case BUILT_IN_ADJUST_TRAMPOLINE:
|
6337 |
|
|
return expand_builtin_adjust_trampoline (exp);
|
6338 |
|
|
|
6339 |
|
|
case BUILT_IN_FORK:
|
6340 |
|
|
case BUILT_IN_EXECL:
|
6341 |
|
|
case BUILT_IN_EXECV:
|
6342 |
|
|
case BUILT_IN_EXECLP:
|
6343 |
|
|
case BUILT_IN_EXECLE:
|
6344 |
|
|
case BUILT_IN_EXECVP:
|
6345 |
|
|
case BUILT_IN_EXECVE:
|
6346 |
|
|
target = expand_builtin_fork_or_exec (fndecl, exp, target, ignore);
|
6347 |
|
|
if (target)
|
6348 |
|
|
return target;
|
6349 |
|
|
break;
|
6350 |
|
|
|
6351 |
|
|
case BUILT_IN_SYNC_FETCH_AND_ADD_1:
|
6352 |
|
|
case BUILT_IN_SYNC_FETCH_AND_ADD_2:
|
6353 |
|
|
case BUILT_IN_SYNC_FETCH_AND_ADD_4:
|
6354 |
|
|
case BUILT_IN_SYNC_FETCH_AND_ADD_8:
|
6355 |
|
|
case BUILT_IN_SYNC_FETCH_AND_ADD_16:
|
6356 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_FETCH_AND_ADD_1);
|
6357 |
|
|
target = expand_builtin_sync_operation (mode, exp, PLUS, false, target);
|
6358 |
|
|
if (target)
|
6359 |
|
|
return target;
|
6360 |
|
|
break;
|
6361 |
|
|
|
6362 |
|
|
case BUILT_IN_SYNC_FETCH_AND_SUB_1:
|
6363 |
|
|
case BUILT_IN_SYNC_FETCH_AND_SUB_2:
|
6364 |
|
|
case BUILT_IN_SYNC_FETCH_AND_SUB_4:
|
6365 |
|
|
case BUILT_IN_SYNC_FETCH_AND_SUB_8:
|
6366 |
|
|
case BUILT_IN_SYNC_FETCH_AND_SUB_16:
|
6367 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_FETCH_AND_SUB_1);
|
6368 |
|
|
target = expand_builtin_sync_operation (mode, exp, MINUS, false, target);
|
6369 |
|
|
if (target)
|
6370 |
|
|
return target;
|
6371 |
|
|
break;
|
6372 |
|
|
|
6373 |
|
|
case BUILT_IN_SYNC_FETCH_AND_OR_1:
|
6374 |
|
|
case BUILT_IN_SYNC_FETCH_AND_OR_2:
|
6375 |
|
|
case BUILT_IN_SYNC_FETCH_AND_OR_4:
|
6376 |
|
|
case BUILT_IN_SYNC_FETCH_AND_OR_8:
|
6377 |
|
|
case BUILT_IN_SYNC_FETCH_AND_OR_16:
|
6378 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_FETCH_AND_OR_1);
|
6379 |
|
|
target = expand_builtin_sync_operation (mode, exp, IOR, false, target);
|
6380 |
|
|
if (target)
|
6381 |
|
|
return target;
|
6382 |
|
|
break;
|
6383 |
|
|
|
6384 |
|
|
case BUILT_IN_SYNC_FETCH_AND_AND_1:
|
6385 |
|
|
case BUILT_IN_SYNC_FETCH_AND_AND_2:
|
6386 |
|
|
case BUILT_IN_SYNC_FETCH_AND_AND_4:
|
6387 |
|
|
case BUILT_IN_SYNC_FETCH_AND_AND_8:
|
6388 |
|
|
case BUILT_IN_SYNC_FETCH_AND_AND_16:
|
6389 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_FETCH_AND_AND_1);
|
6390 |
|
|
target = expand_builtin_sync_operation (mode, exp, AND, false, target);
|
6391 |
|
|
if (target)
|
6392 |
|
|
return target;
|
6393 |
|
|
break;
|
6394 |
|
|
|
6395 |
|
|
case BUILT_IN_SYNC_FETCH_AND_XOR_1:
|
6396 |
|
|
case BUILT_IN_SYNC_FETCH_AND_XOR_2:
|
6397 |
|
|
case BUILT_IN_SYNC_FETCH_AND_XOR_4:
|
6398 |
|
|
case BUILT_IN_SYNC_FETCH_AND_XOR_8:
|
6399 |
|
|
case BUILT_IN_SYNC_FETCH_AND_XOR_16:
|
6400 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_FETCH_AND_XOR_1);
|
6401 |
|
|
target = expand_builtin_sync_operation (mode, exp, XOR, false, target);
|
6402 |
|
|
if (target)
|
6403 |
|
|
return target;
|
6404 |
|
|
break;
|
6405 |
|
|
|
6406 |
|
|
case BUILT_IN_SYNC_FETCH_AND_NAND_1:
|
6407 |
|
|
case BUILT_IN_SYNC_FETCH_AND_NAND_2:
|
6408 |
|
|
case BUILT_IN_SYNC_FETCH_AND_NAND_4:
|
6409 |
|
|
case BUILT_IN_SYNC_FETCH_AND_NAND_8:
|
6410 |
|
|
case BUILT_IN_SYNC_FETCH_AND_NAND_16:
|
6411 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_FETCH_AND_NAND_1);
|
6412 |
|
|
target = expand_builtin_sync_operation (mode, exp, NOT, false, target);
|
6413 |
|
|
if (target)
|
6414 |
|
|
return target;
|
6415 |
|
|
break;
|
6416 |
|
|
|
6417 |
|
|
case BUILT_IN_SYNC_ADD_AND_FETCH_1:
|
6418 |
|
|
case BUILT_IN_SYNC_ADD_AND_FETCH_2:
|
6419 |
|
|
case BUILT_IN_SYNC_ADD_AND_FETCH_4:
|
6420 |
|
|
case BUILT_IN_SYNC_ADD_AND_FETCH_8:
|
6421 |
|
|
case BUILT_IN_SYNC_ADD_AND_FETCH_16:
|
6422 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_ADD_AND_FETCH_1);
|
6423 |
|
|
target = expand_builtin_sync_operation (mode, exp, PLUS, true, target);
|
6424 |
|
|
if (target)
|
6425 |
|
|
return target;
|
6426 |
|
|
break;
|
6427 |
|
|
|
6428 |
|
|
case BUILT_IN_SYNC_SUB_AND_FETCH_1:
|
6429 |
|
|
case BUILT_IN_SYNC_SUB_AND_FETCH_2:
|
6430 |
|
|
case BUILT_IN_SYNC_SUB_AND_FETCH_4:
|
6431 |
|
|
case BUILT_IN_SYNC_SUB_AND_FETCH_8:
|
6432 |
|
|
case BUILT_IN_SYNC_SUB_AND_FETCH_16:
|
6433 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_SUB_AND_FETCH_1);
|
6434 |
|
|
target = expand_builtin_sync_operation (mode, exp, MINUS, true, target);
|
6435 |
|
|
if (target)
|
6436 |
|
|
return target;
|
6437 |
|
|
break;
|
6438 |
|
|
|
6439 |
|
|
case BUILT_IN_SYNC_OR_AND_FETCH_1:
|
6440 |
|
|
case BUILT_IN_SYNC_OR_AND_FETCH_2:
|
6441 |
|
|
case BUILT_IN_SYNC_OR_AND_FETCH_4:
|
6442 |
|
|
case BUILT_IN_SYNC_OR_AND_FETCH_8:
|
6443 |
|
|
case BUILT_IN_SYNC_OR_AND_FETCH_16:
|
6444 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_OR_AND_FETCH_1);
|
6445 |
|
|
target = expand_builtin_sync_operation (mode, exp, IOR, true, target);
|
6446 |
|
|
if (target)
|
6447 |
|
|
return target;
|
6448 |
|
|
break;
|
6449 |
|
|
|
6450 |
|
|
case BUILT_IN_SYNC_AND_AND_FETCH_1:
|
6451 |
|
|
case BUILT_IN_SYNC_AND_AND_FETCH_2:
|
6452 |
|
|
case BUILT_IN_SYNC_AND_AND_FETCH_4:
|
6453 |
|
|
case BUILT_IN_SYNC_AND_AND_FETCH_8:
|
6454 |
|
|
case BUILT_IN_SYNC_AND_AND_FETCH_16:
|
6455 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_AND_AND_FETCH_1);
|
6456 |
|
|
target = expand_builtin_sync_operation (mode, exp, AND, true, target);
|
6457 |
|
|
if (target)
|
6458 |
|
|
return target;
|
6459 |
|
|
break;
|
6460 |
|
|
|
6461 |
|
|
case BUILT_IN_SYNC_XOR_AND_FETCH_1:
|
6462 |
|
|
case BUILT_IN_SYNC_XOR_AND_FETCH_2:
|
6463 |
|
|
case BUILT_IN_SYNC_XOR_AND_FETCH_4:
|
6464 |
|
|
case BUILT_IN_SYNC_XOR_AND_FETCH_8:
|
6465 |
|
|
case BUILT_IN_SYNC_XOR_AND_FETCH_16:
|
6466 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_XOR_AND_FETCH_1);
|
6467 |
|
|
target = expand_builtin_sync_operation (mode, exp, XOR, true, target);
|
6468 |
|
|
if (target)
|
6469 |
|
|
return target;
|
6470 |
|
|
break;
|
6471 |
|
|
|
6472 |
|
|
case BUILT_IN_SYNC_NAND_AND_FETCH_1:
|
6473 |
|
|
case BUILT_IN_SYNC_NAND_AND_FETCH_2:
|
6474 |
|
|
case BUILT_IN_SYNC_NAND_AND_FETCH_4:
|
6475 |
|
|
case BUILT_IN_SYNC_NAND_AND_FETCH_8:
|
6476 |
|
|
case BUILT_IN_SYNC_NAND_AND_FETCH_16:
|
6477 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_NAND_AND_FETCH_1);
|
6478 |
|
|
target = expand_builtin_sync_operation (mode, exp, NOT, true, target);
|
6479 |
|
|
if (target)
|
6480 |
|
|
return target;
|
6481 |
|
|
break;
|
6482 |
|
|
|
6483 |
|
|
case BUILT_IN_SYNC_BOOL_COMPARE_AND_SWAP_1:
|
6484 |
|
|
case BUILT_IN_SYNC_BOOL_COMPARE_AND_SWAP_2:
|
6485 |
|
|
case BUILT_IN_SYNC_BOOL_COMPARE_AND_SWAP_4:
|
6486 |
|
|
case BUILT_IN_SYNC_BOOL_COMPARE_AND_SWAP_8:
|
6487 |
|
|
case BUILT_IN_SYNC_BOOL_COMPARE_AND_SWAP_16:
|
6488 |
|
|
if (mode == VOIDmode)
|
6489 |
|
|
mode = TYPE_MODE (boolean_type_node);
|
6490 |
|
|
if (!target || !register_operand (target, mode))
|
6491 |
|
|
target = gen_reg_rtx (mode);
|
6492 |
|
|
|
6493 |
|
|
mode = get_builtin_sync_mode
|
6494 |
|
|
(fcode - BUILT_IN_SYNC_BOOL_COMPARE_AND_SWAP_1);
|
6495 |
|
|
target = expand_builtin_compare_and_swap (mode, exp, true, target);
|
6496 |
|
|
if (target)
|
6497 |
|
|
return target;
|
6498 |
|
|
break;
|
6499 |
|
|
|
6500 |
|
|
case BUILT_IN_SYNC_VAL_COMPARE_AND_SWAP_1:
|
6501 |
|
|
case BUILT_IN_SYNC_VAL_COMPARE_AND_SWAP_2:
|
6502 |
|
|
case BUILT_IN_SYNC_VAL_COMPARE_AND_SWAP_4:
|
6503 |
|
|
case BUILT_IN_SYNC_VAL_COMPARE_AND_SWAP_8:
|
6504 |
|
|
case BUILT_IN_SYNC_VAL_COMPARE_AND_SWAP_16:
|
6505 |
|
|
mode = get_builtin_sync_mode
|
6506 |
|
|
(fcode - BUILT_IN_SYNC_VAL_COMPARE_AND_SWAP_1);
|
6507 |
|
|
target = expand_builtin_compare_and_swap (mode, exp, false, target);
|
6508 |
|
|
if (target)
|
6509 |
|
|
return target;
|
6510 |
|
|
break;
|
6511 |
|
|
|
6512 |
|
|
case BUILT_IN_SYNC_LOCK_TEST_AND_SET_1:
|
6513 |
|
|
case BUILT_IN_SYNC_LOCK_TEST_AND_SET_2:
|
6514 |
|
|
case BUILT_IN_SYNC_LOCK_TEST_AND_SET_4:
|
6515 |
|
|
case BUILT_IN_SYNC_LOCK_TEST_AND_SET_8:
|
6516 |
|
|
case BUILT_IN_SYNC_LOCK_TEST_AND_SET_16:
|
6517 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_LOCK_TEST_AND_SET_1);
|
6518 |
|
|
target = expand_builtin_sync_lock_test_and_set (mode, exp, target);
|
6519 |
|
|
if (target)
|
6520 |
|
|
return target;
|
6521 |
|
|
break;
|
6522 |
|
|
|
6523 |
|
|
case BUILT_IN_SYNC_LOCK_RELEASE_1:
|
6524 |
|
|
case BUILT_IN_SYNC_LOCK_RELEASE_2:
|
6525 |
|
|
case BUILT_IN_SYNC_LOCK_RELEASE_4:
|
6526 |
|
|
case BUILT_IN_SYNC_LOCK_RELEASE_8:
|
6527 |
|
|
case BUILT_IN_SYNC_LOCK_RELEASE_16:
|
6528 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_SYNC_LOCK_RELEASE_1);
|
6529 |
|
|
expand_builtin_sync_lock_release (mode, exp);
|
6530 |
|
|
return const0_rtx;
|
6531 |
|
|
|
6532 |
|
|
case BUILT_IN_SYNC_SYNCHRONIZE:
|
6533 |
|
|
expand_builtin_sync_synchronize ();
|
6534 |
|
|
return const0_rtx;
|
6535 |
|
|
|
6536 |
|
|
case BUILT_IN_ATOMIC_EXCHANGE_1:
|
6537 |
|
|
case BUILT_IN_ATOMIC_EXCHANGE_2:
|
6538 |
|
|
case BUILT_IN_ATOMIC_EXCHANGE_4:
|
6539 |
|
|
case BUILT_IN_ATOMIC_EXCHANGE_8:
|
6540 |
|
|
case BUILT_IN_ATOMIC_EXCHANGE_16:
|
6541 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_EXCHANGE_1);
|
6542 |
|
|
target = expand_builtin_atomic_exchange (mode, exp, target);
|
6543 |
|
|
if (target)
|
6544 |
|
|
return target;
|
6545 |
|
|
break;
|
6546 |
|
|
|
6547 |
|
|
case BUILT_IN_ATOMIC_COMPARE_EXCHANGE_1:
|
6548 |
|
|
case BUILT_IN_ATOMIC_COMPARE_EXCHANGE_2:
|
6549 |
|
|
case BUILT_IN_ATOMIC_COMPARE_EXCHANGE_4:
|
6550 |
|
|
case BUILT_IN_ATOMIC_COMPARE_EXCHANGE_8:
|
6551 |
|
|
case BUILT_IN_ATOMIC_COMPARE_EXCHANGE_16:
|
6552 |
|
|
{
|
6553 |
|
|
unsigned int nargs, z;
|
6554 |
|
|
VEC(tree,gc) *vec;
|
6555 |
|
|
|
6556 |
|
|
mode =
|
6557 |
|
|
get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_COMPARE_EXCHANGE_1);
|
6558 |
|
|
target = expand_builtin_atomic_compare_exchange (mode, exp, target);
|
6559 |
|
|
if (target)
|
6560 |
|
|
return target;
|
6561 |
|
|
|
6562 |
|
|
/* If this is turned into an external library call, the weak parameter
|
6563 |
|
|
must be dropped to match the expected parameter list. */
|
6564 |
|
|
nargs = call_expr_nargs (exp);
|
6565 |
|
|
vec = VEC_alloc (tree, gc, nargs - 1);
|
6566 |
|
|
for (z = 0; z < 3; z++)
|
6567 |
|
|
VEC_quick_push (tree, vec, CALL_EXPR_ARG (exp, z));
|
6568 |
|
|
/* Skip the boolean weak parameter. */
|
6569 |
|
|
for (z = 4; z < 6; z++)
|
6570 |
|
|
VEC_quick_push (tree, vec, CALL_EXPR_ARG (exp, z));
|
6571 |
|
|
exp = build_call_vec (TREE_TYPE (exp), CALL_EXPR_FN (exp), vec);
|
6572 |
|
|
break;
|
6573 |
|
|
}
|
6574 |
|
|
|
6575 |
|
|
case BUILT_IN_ATOMIC_LOAD_1:
|
6576 |
|
|
case BUILT_IN_ATOMIC_LOAD_2:
|
6577 |
|
|
case BUILT_IN_ATOMIC_LOAD_4:
|
6578 |
|
|
case BUILT_IN_ATOMIC_LOAD_8:
|
6579 |
|
|
case BUILT_IN_ATOMIC_LOAD_16:
|
6580 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_LOAD_1);
|
6581 |
|
|
target = expand_builtin_atomic_load (mode, exp, target);
|
6582 |
|
|
if (target)
|
6583 |
|
|
return target;
|
6584 |
|
|
break;
|
6585 |
|
|
|
6586 |
|
|
case BUILT_IN_ATOMIC_STORE_1:
|
6587 |
|
|
case BUILT_IN_ATOMIC_STORE_2:
|
6588 |
|
|
case BUILT_IN_ATOMIC_STORE_4:
|
6589 |
|
|
case BUILT_IN_ATOMIC_STORE_8:
|
6590 |
|
|
case BUILT_IN_ATOMIC_STORE_16:
|
6591 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_STORE_1);
|
6592 |
|
|
target = expand_builtin_atomic_store (mode, exp);
|
6593 |
|
|
if (target)
|
6594 |
|
|
return const0_rtx;
|
6595 |
|
|
break;
|
6596 |
|
|
|
6597 |
|
|
case BUILT_IN_ATOMIC_ADD_FETCH_1:
|
6598 |
|
|
case BUILT_IN_ATOMIC_ADD_FETCH_2:
|
6599 |
|
|
case BUILT_IN_ATOMIC_ADD_FETCH_4:
|
6600 |
|
|
case BUILT_IN_ATOMIC_ADD_FETCH_8:
|
6601 |
|
|
case BUILT_IN_ATOMIC_ADD_FETCH_16:
|
6602 |
|
|
{
|
6603 |
|
|
enum built_in_function lib;
|
6604 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_ADD_FETCH_1);
|
6605 |
|
|
lib = (enum built_in_function)((int)BUILT_IN_ATOMIC_FETCH_ADD_1 +
|
6606 |
|
|
(fcode - BUILT_IN_ATOMIC_ADD_FETCH_1));
|
6607 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, PLUS, true,
|
6608 |
|
|
ignore, lib);
|
6609 |
|
|
if (target)
|
6610 |
|
|
return target;
|
6611 |
|
|
break;
|
6612 |
|
|
}
|
6613 |
|
|
case BUILT_IN_ATOMIC_SUB_FETCH_1:
|
6614 |
|
|
case BUILT_IN_ATOMIC_SUB_FETCH_2:
|
6615 |
|
|
case BUILT_IN_ATOMIC_SUB_FETCH_4:
|
6616 |
|
|
case BUILT_IN_ATOMIC_SUB_FETCH_8:
|
6617 |
|
|
case BUILT_IN_ATOMIC_SUB_FETCH_16:
|
6618 |
|
|
{
|
6619 |
|
|
enum built_in_function lib;
|
6620 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_SUB_FETCH_1);
|
6621 |
|
|
lib = (enum built_in_function)((int)BUILT_IN_ATOMIC_FETCH_SUB_1 +
|
6622 |
|
|
(fcode - BUILT_IN_ATOMIC_SUB_FETCH_1));
|
6623 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, MINUS, true,
|
6624 |
|
|
ignore, lib);
|
6625 |
|
|
if (target)
|
6626 |
|
|
return target;
|
6627 |
|
|
break;
|
6628 |
|
|
}
|
6629 |
|
|
case BUILT_IN_ATOMIC_AND_FETCH_1:
|
6630 |
|
|
case BUILT_IN_ATOMIC_AND_FETCH_2:
|
6631 |
|
|
case BUILT_IN_ATOMIC_AND_FETCH_4:
|
6632 |
|
|
case BUILT_IN_ATOMIC_AND_FETCH_8:
|
6633 |
|
|
case BUILT_IN_ATOMIC_AND_FETCH_16:
|
6634 |
|
|
{
|
6635 |
|
|
enum built_in_function lib;
|
6636 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_AND_FETCH_1);
|
6637 |
|
|
lib = (enum built_in_function)((int)BUILT_IN_ATOMIC_FETCH_AND_1 +
|
6638 |
|
|
(fcode - BUILT_IN_ATOMIC_AND_FETCH_1));
|
6639 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, AND, true,
|
6640 |
|
|
ignore, lib);
|
6641 |
|
|
if (target)
|
6642 |
|
|
return target;
|
6643 |
|
|
break;
|
6644 |
|
|
}
|
6645 |
|
|
case BUILT_IN_ATOMIC_NAND_FETCH_1:
|
6646 |
|
|
case BUILT_IN_ATOMIC_NAND_FETCH_2:
|
6647 |
|
|
case BUILT_IN_ATOMIC_NAND_FETCH_4:
|
6648 |
|
|
case BUILT_IN_ATOMIC_NAND_FETCH_8:
|
6649 |
|
|
case BUILT_IN_ATOMIC_NAND_FETCH_16:
|
6650 |
|
|
{
|
6651 |
|
|
enum built_in_function lib;
|
6652 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_NAND_FETCH_1);
|
6653 |
|
|
lib = (enum built_in_function)((int)BUILT_IN_ATOMIC_FETCH_NAND_1 +
|
6654 |
|
|
(fcode - BUILT_IN_ATOMIC_NAND_FETCH_1));
|
6655 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, NOT, true,
|
6656 |
|
|
ignore, lib);
|
6657 |
|
|
if (target)
|
6658 |
|
|
return target;
|
6659 |
|
|
break;
|
6660 |
|
|
}
|
6661 |
|
|
case BUILT_IN_ATOMIC_XOR_FETCH_1:
|
6662 |
|
|
case BUILT_IN_ATOMIC_XOR_FETCH_2:
|
6663 |
|
|
case BUILT_IN_ATOMIC_XOR_FETCH_4:
|
6664 |
|
|
case BUILT_IN_ATOMIC_XOR_FETCH_8:
|
6665 |
|
|
case BUILT_IN_ATOMIC_XOR_FETCH_16:
|
6666 |
|
|
{
|
6667 |
|
|
enum built_in_function lib;
|
6668 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_XOR_FETCH_1);
|
6669 |
|
|
lib = (enum built_in_function)((int)BUILT_IN_ATOMIC_FETCH_XOR_1 +
|
6670 |
|
|
(fcode - BUILT_IN_ATOMIC_XOR_FETCH_1));
|
6671 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, XOR, true,
|
6672 |
|
|
ignore, lib);
|
6673 |
|
|
if (target)
|
6674 |
|
|
return target;
|
6675 |
|
|
break;
|
6676 |
|
|
}
|
6677 |
|
|
case BUILT_IN_ATOMIC_OR_FETCH_1:
|
6678 |
|
|
case BUILT_IN_ATOMIC_OR_FETCH_2:
|
6679 |
|
|
case BUILT_IN_ATOMIC_OR_FETCH_4:
|
6680 |
|
|
case BUILT_IN_ATOMIC_OR_FETCH_8:
|
6681 |
|
|
case BUILT_IN_ATOMIC_OR_FETCH_16:
|
6682 |
|
|
{
|
6683 |
|
|
enum built_in_function lib;
|
6684 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_OR_FETCH_1);
|
6685 |
|
|
lib = (enum built_in_function)((int)BUILT_IN_ATOMIC_FETCH_OR_1 +
|
6686 |
|
|
(fcode - BUILT_IN_ATOMIC_OR_FETCH_1));
|
6687 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, IOR, true,
|
6688 |
|
|
ignore, lib);
|
6689 |
|
|
if (target)
|
6690 |
|
|
return target;
|
6691 |
|
|
break;
|
6692 |
|
|
}
|
6693 |
|
|
case BUILT_IN_ATOMIC_FETCH_ADD_1:
|
6694 |
|
|
case BUILT_IN_ATOMIC_FETCH_ADD_2:
|
6695 |
|
|
case BUILT_IN_ATOMIC_FETCH_ADD_4:
|
6696 |
|
|
case BUILT_IN_ATOMIC_FETCH_ADD_8:
|
6697 |
|
|
case BUILT_IN_ATOMIC_FETCH_ADD_16:
|
6698 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_FETCH_ADD_1);
|
6699 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, PLUS, false,
|
6700 |
|
|
ignore, BUILT_IN_NONE);
|
6701 |
|
|
if (target)
|
6702 |
|
|
return target;
|
6703 |
|
|
break;
|
6704 |
|
|
|
6705 |
|
|
case BUILT_IN_ATOMIC_FETCH_SUB_1:
|
6706 |
|
|
case BUILT_IN_ATOMIC_FETCH_SUB_2:
|
6707 |
|
|
case BUILT_IN_ATOMIC_FETCH_SUB_4:
|
6708 |
|
|
case BUILT_IN_ATOMIC_FETCH_SUB_8:
|
6709 |
|
|
case BUILT_IN_ATOMIC_FETCH_SUB_16:
|
6710 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_FETCH_SUB_1);
|
6711 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, MINUS, false,
|
6712 |
|
|
ignore, BUILT_IN_NONE);
|
6713 |
|
|
if (target)
|
6714 |
|
|
return target;
|
6715 |
|
|
break;
|
6716 |
|
|
|
6717 |
|
|
case BUILT_IN_ATOMIC_FETCH_AND_1:
|
6718 |
|
|
case BUILT_IN_ATOMIC_FETCH_AND_2:
|
6719 |
|
|
case BUILT_IN_ATOMIC_FETCH_AND_4:
|
6720 |
|
|
case BUILT_IN_ATOMIC_FETCH_AND_8:
|
6721 |
|
|
case BUILT_IN_ATOMIC_FETCH_AND_16:
|
6722 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_FETCH_AND_1);
|
6723 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, AND, false,
|
6724 |
|
|
ignore, BUILT_IN_NONE);
|
6725 |
|
|
if (target)
|
6726 |
|
|
return target;
|
6727 |
|
|
break;
|
6728 |
|
|
|
6729 |
|
|
case BUILT_IN_ATOMIC_FETCH_NAND_1:
|
6730 |
|
|
case BUILT_IN_ATOMIC_FETCH_NAND_2:
|
6731 |
|
|
case BUILT_IN_ATOMIC_FETCH_NAND_4:
|
6732 |
|
|
case BUILT_IN_ATOMIC_FETCH_NAND_8:
|
6733 |
|
|
case BUILT_IN_ATOMIC_FETCH_NAND_16:
|
6734 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_FETCH_NAND_1);
|
6735 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, NOT, false,
|
6736 |
|
|
ignore, BUILT_IN_NONE);
|
6737 |
|
|
if (target)
|
6738 |
|
|
return target;
|
6739 |
|
|
break;
|
6740 |
|
|
|
6741 |
|
|
case BUILT_IN_ATOMIC_FETCH_XOR_1:
|
6742 |
|
|
case BUILT_IN_ATOMIC_FETCH_XOR_2:
|
6743 |
|
|
case BUILT_IN_ATOMIC_FETCH_XOR_4:
|
6744 |
|
|
case BUILT_IN_ATOMIC_FETCH_XOR_8:
|
6745 |
|
|
case BUILT_IN_ATOMIC_FETCH_XOR_16:
|
6746 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_FETCH_XOR_1);
|
6747 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, XOR, false,
|
6748 |
|
|
ignore, BUILT_IN_NONE);
|
6749 |
|
|
if (target)
|
6750 |
|
|
return target;
|
6751 |
|
|
break;
|
6752 |
|
|
|
6753 |
|
|
case BUILT_IN_ATOMIC_FETCH_OR_1:
|
6754 |
|
|
case BUILT_IN_ATOMIC_FETCH_OR_2:
|
6755 |
|
|
case BUILT_IN_ATOMIC_FETCH_OR_4:
|
6756 |
|
|
case BUILT_IN_ATOMIC_FETCH_OR_8:
|
6757 |
|
|
case BUILT_IN_ATOMIC_FETCH_OR_16:
|
6758 |
|
|
mode = get_builtin_sync_mode (fcode - BUILT_IN_ATOMIC_FETCH_OR_1);
|
6759 |
|
|
target = expand_builtin_atomic_fetch_op (mode, exp, target, IOR, false,
|
6760 |
|
|
ignore, BUILT_IN_NONE);
|
6761 |
|
|
if (target)
|
6762 |
|
|
return target;
|
6763 |
|
|
break;
|
6764 |
|
|
|
6765 |
|
|
case BUILT_IN_ATOMIC_TEST_AND_SET:
|
6766 |
|
|
return expand_builtin_atomic_test_and_set (exp, target);
|
6767 |
|
|
|
6768 |
|
|
case BUILT_IN_ATOMIC_CLEAR:
|
6769 |
|
|
return expand_builtin_atomic_clear (exp);
|
6770 |
|
|
|
6771 |
|
|
case BUILT_IN_ATOMIC_ALWAYS_LOCK_FREE:
|
6772 |
|
|
return expand_builtin_atomic_always_lock_free (exp);
|
6773 |
|
|
|
6774 |
|
|
case BUILT_IN_ATOMIC_IS_LOCK_FREE:
|
6775 |
|
|
target = expand_builtin_atomic_is_lock_free (exp);
|
6776 |
|
|
if (target)
|
6777 |
|
|
return target;
|
6778 |
|
|
break;
|
6779 |
|
|
|
6780 |
|
|
case BUILT_IN_ATOMIC_THREAD_FENCE:
|
6781 |
|
|
expand_builtin_atomic_thread_fence (exp);
|
6782 |
|
|
return const0_rtx;
|
6783 |
|
|
|
6784 |
|
|
case BUILT_IN_ATOMIC_SIGNAL_FENCE:
|
6785 |
|
|
expand_builtin_atomic_signal_fence (exp);
|
6786 |
|
|
return const0_rtx;
|
6787 |
|
|
|
6788 |
|
|
case BUILT_IN_OBJECT_SIZE:
|
6789 |
|
|
return expand_builtin_object_size (exp);
|
6790 |
|
|
|
6791 |
|
|
case BUILT_IN_MEMCPY_CHK:
|
6792 |
|
|
case BUILT_IN_MEMPCPY_CHK:
|
6793 |
|
|
case BUILT_IN_MEMMOVE_CHK:
|
6794 |
|
|
case BUILT_IN_MEMSET_CHK:
|
6795 |
|
|
target = expand_builtin_memory_chk (exp, target, mode, fcode);
|
6796 |
|
|
if (target)
|
6797 |
|
|
return target;
|
6798 |
|
|
break;
|
6799 |
|
|
|
6800 |
|
|
case BUILT_IN_STRCPY_CHK:
|
6801 |
|
|
case BUILT_IN_STPCPY_CHK:
|
6802 |
|
|
case BUILT_IN_STRNCPY_CHK:
|
6803 |
|
|
case BUILT_IN_STPNCPY_CHK:
|
6804 |
|
|
case BUILT_IN_STRCAT_CHK:
|
6805 |
|
|
case BUILT_IN_STRNCAT_CHK:
|
6806 |
|
|
case BUILT_IN_SNPRINTF_CHK:
|
6807 |
|
|
case BUILT_IN_VSNPRINTF_CHK:
|
6808 |
|
|
maybe_emit_chk_warning (exp, fcode);
|
6809 |
|
|
break;
|
6810 |
|
|
|
6811 |
|
|
case BUILT_IN_SPRINTF_CHK:
|
6812 |
|
|
case BUILT_IN_VSPRINTF_CHK:
|
6813 |
|
|
maybe_emit_sprintf_chk_warning (exp, fcode);
|
6814 |
|
|
break;
|
6815 |
|
|
|
6816 |
|
|
case BUILT_IN_FREE:
|
6817 |
|
|
if (warn_free_nonheap_object)
|
6818 |
|
|
maybe_emit_free_warning (exp);
|
6819 |
|
|
break;
|
6820 |
|
|
|
6821 |
|
|
default: /* just do library call, if unknown builtin */
|
6822 |
|
|
break;
|
6823 |
|
|
}
|
6824 |
|
|
|
6825 |
|
|
/* The switch statement above can drop through to cause the function
|
6826 |
|
|
to be called normally. */
|
6827 |
|
|
return expand_call (exp, target, ignore);
|
6828 |
|
|
}
|
6829 |
|
|
|
6830 |
|
|
/* Determine whether a tree node represents a call to a built-in
|
6831 |
|
|
function. If the tree T is a call to a built-in function with
|
6832 |
|
|
the right number of arguments of the appropriate types, return
|
6833 |
|
|
the DECL_FUNCTION_CODE of the call, e.g. BUILT_IN_SQRT.
|
6834 |
|
|
Otherwise the return value is END_BUILTINS. */
|
6835 |
|
|
|
6836 |
|
|
enum built_in_function
|
6837 |
|
|
builtin_mathfn_code (const_tree t)
|
6838 |
|
|
{
|
6839 |
|
|
const_tree fndecl, arg, parmlist;
|
6840 |
|
|
const_tree argtype, parmtype;
|
6841 |
|
|
const_call_expr_arg_iterator iter;
|
6842 |
|
|
|
6843 |
|
|
if (TREE_CODE (t) != CALL_EXPR
|
6844 |
|
|
|| TREE_CODE (CALL_EXPR_FN (t)) != ADDR_EXPR)
|
6845 |
|
|
return END_BUILTINS;
|
6846 |
|
|
|
6847 |
|
|
fndecl = get_callee_fndecl (t);
|
6848 |
|
|
if (fndecl == NULL_TREE
|
6849 |
|
|
|| TREE_CODE (fndecl) != FUNCTION_DECL
|
6850 |
|
|
|| ! DECL_BUILT_IN (fndecl)
|
6851 |
|
|
|| DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_MD)
|
6852 |
|
|
return END_BUILTINS;
|
6853 |
|
|
|
6854 |
|
|
parmlist = TYPE_ARG_TYPES (TREE_TYPE (fndecl));
|
6855 |
|
|
init_const_call_expr_arg_iterator (t, &iter);
|
6856 |
|
|
for (; parmlist; parmlist = TREE_CHAIN (parmlist))
|
6857 |
|
|
{
|
6858 |
|
|
/* If a function doesn't take a variable number of arguments,
|
6859 |
|
|
the last element in the list will have type `void'. */
|
6860 |
|
|
parmtype = TREE_VALUE (parmlist);
|
6861 |
|
|
if (VOID_TYPE_P (parmtype))
|
6862 |
|
|
{
|
6863 |
|
|
if (more_const_call_expr_args_p (&iter))
|
6864 |
|
|
return END_BUILTINS;
|
6865 |
|
|
return DECL_FUNCTION_CODE (fndecl);
|
6866 |
|
|
}
|
6867 |
|
|
|
6868 |
|
|
if (! more_const_call_expr_args_p (&iter))
|
6869 |
|
|
return END_BUILTINS;
|
6870 |
|
|
|
6871 |
|
|
arg = next_const_call_expr_arg (&iter);
|
6872 |
|
|
argtype = TREE_TYPE (arg);
|
6873 |
|
|
|
6874 |
|
|
if (SCALAR_FLOAT_TYPE_P (parmtype))
|
6875 |
|
|
{
|
6876 |
|
|
if (! SCALAR_FLOAT_TYPE_P (argtype))
|
6877 |
|
|
return END_BUILTINS;
|
6878 |
|
|
}
|
6879 |
|
|
else if (COMPLEX_FLOAT_TYPE_P (parmtype))
|
6880 |
|
|
{
|
6881 |
|
|
if (! COMPLEX_FLOAT_TYPE_P (argtype))
|
6882 |
|
|
return END_BUILTINS;
|
6883 |
|
|
}
|
6884 |
|
|
else if (POINTER_TYPE_P (parmtype))
|
6885 |
|
|
{
|
6886 |
|
|
if (! POINTER_TYPE_P (argtype))
|
6887 |
|
|
return END_BUILTINS;
|
6888 |
|
|
}
|
6889 |
|
|
else if (INTEGRAL_TYPE_P (parmtype))
|
6890 |
|
|
{
|
6891 |
|
|
if (! INTEGRAL_TYPE_P (argtype))
|
6892 |
|
|
return END_BUILTINS;
|
6893 |
|
|
}
|
6894 |
|
|
else
|
6895 |
|
|
return END_BUILTINS;
|
6896 |
|
|
}
|
6897 |
|
|
|
6898 |
|
|
/* Variable-length argument list. */
|
6899 |
|
|
return DECL_FUNCTION_CODE (fndecl);
|
6900 |
|
|
}
|
6901 |
|
|
|
6902 |
|
|
/* Fold a call to __builtin_constant_p, if we know its argument ARG will
|
6903 |
|
|
evaluate to a constant. */
|
6904 |
|
|
|
6905 |
|
|
static tree
|
6906 |
|
|
fold_builtin_constant_p (tree arg)
|
6907 |
|
|
{
|
6908 |
|
|
/* We return 1 for a numeric type that's known to be a constant
|
6909 |
|
|
value at compile-time or for an aggregate type that's a
|
6910 |
|
|
literal constant. */
|
6911 |
|
|
STRIP_NOPS (arg);
|
6912 |
|
|
|
6913 |
|
|
/* If we know this is a constant, emit the constant of one. */
|
6914 |
|
|
if (CONSTANT_CLASS_P (arg)
|
6915 |
|
|
|| (TREE_CODE (arg) == CONSTRUCTOR
|
6916 |
|
|
&& TREE_CONSTANT (arg)))
|
6917 |
|
|
return integer_one_node;
|
6918 |
|
|
if (TREE_CODE (arg) == ADDR_EXPR)
|
6919 |
|
|
{
|
6920 |
|
|
tree op = TREE_OPERAND (arg, 0);
|
6921 |
|
|
if (TREE_CODE (op) == STRING_CST
|
6922 |
|
|
|| (TREE_CODE (op) == ARRAY_REF
|
6923 |
|
|
&& integer_zerop (TREE_OPERAND (op, 1))
|
6924 |
|
|
&& TREE_CODE (TREE_OPERAND (op, 0)) == STRING_CST))
|
6925 |
|
|
return integer_one_node;
|
6926 |
|
|
}
|
6927 |
|
|
|
6928 |
|
|
/* If this expression has side effects, show we don't know it to be a
|
6929 |
|
|
constant. Likewise if it's a pointer or aggregate type since in
|
6930 |
|
|
those case we only want literals, since those are only optimized
|
6931 |
|
|
when generating RTL, not later.
|
6932 |
|
|
And finally, if we are compiling an initializer, not code, we
|
6933 |
|
|
need to return a definite result now; there's not going to be any
|
6934 |
|
|
more optimization done. */
|
6935 |
|
|
if (TREE_SIDE_EFFECTS (arg)
|
6936 |
|
|
|| AGGREGATE_TYPE_P (TREE_TYPE (arg))
|
6937 |
|
|
|| POINTER_TYPE_P (TREE_TYPE (arg))
|
6938 |
|
|
|| cfun == 0
|
6939 |
|
|
|| folding_initializer)
|
6940 |
|
|
return integer_zero_node;
|
6941 |
|
|
|
6942 |
|
|
return NULL_TREE;
|
6943 |
|
|
}
|
6944 |
|
|
|
6945 |
|
|
/* Create builtin_expect with PRED and EXPECTED as its arguments and
|
6946 |
|
|
return it as a truthvalue. */
|
6947 |
|
|
|
6948 |
|
|
static tree
|
6949 |
|
|
build_builtin_expect_predicate (location_t loc, tree pred, tree expected)
|
6950 |
|
|
{
|
6951 |
|
|
tree fn, arg_types, pred_type, expected_type, call_expr, ret_type;
|
6952 |
|
|
|
6953 |
|
|
fn = builtin_decl_explicit (BUILT_IN_EXPECT);
|
6954 |
|
|
arg_types = TYPE_ARG_TYPES (TREE_TYPE (fn));
|
6955 |
|
|
ret_type = TREE_TYPE (TREE_TYPE (fn));
|
6956 |
|
|
pred_type = TREE_VALUE (arg_types);
|
6957 |
|
|
expected_type = TREE_VALUE (TREE_CHAIN (arg_types));
|
6958 |
|
|
|
6959 |
|
|
pred = fold_convert_loc (loc, pred_type, pred);
|
6960 |
|
|
expected = fold_convert_loc (loc, expected_type, expected);
|
6961 |
|
|
call_expr = build_call_expr_loc (loc, fn, 2, pred, expected);
|
6962 |
|
|
|
6963 |
|
|
return build2 (NE_EXPR, TREE_TYPE (pred), call_expr,
|
6964 |
|
|
build_int_cst (ret_type, 0));
|
6965 |
|
|
}
|
6966 |
|
|
|
6967 |
|
|
/* Fold a call to builtin_expect with arguments ARG0 and ARG1. Return
|
6968 |
|
|
NULL_TREE if no simplification is possible. */
|
6969 |
|
|
|
6970 |
|
|
static tree
|
6971 |
|
|
fold_builtin_expect (location_t loc, tree arg0, tree arg1)
|
6972 |
|
|
{
|
6973 |
|
|
tree inner, fndecl, inner_arg0;
|
6974 |
|
|
enum tree_code code;
|
6975 |
|
|
|
6976 |
|
|
/* Distribute the expected value over short-circuiting operators.
|
6977 |
|
|
See through the cast from truthvalue_type_node to long. */
|
6978 |
|
|
inner_arg0 = arg0;
|
6979 |
|
|
while (TREE_CODE (inner_arg0) == NOP_EXPR
|
6980 |
|
|
&& INTEGRAL_TYPE_P (TREE_TYPE (inner_arg0))
|
6981 |
|
|
&& INTEGRAL_TYPE_P (TREE_TYPE (TREE_OPERAND (inner_arg0, 0))))
|
6982 |
|
|
inner_arg0 = TREE_OPERAND (inner_arg0, 0);
|
6983 |
|
|
|
6984 |
|
|
/* If this is a builtin_expect within a builtin_expect keep the
|
6985 |
|
|
inner one. See through a comparison against a constant. It
|
6986 |
|
|
might have been added to create a thruthvalue. */
|
6987 |
|
|
inner = inner_arg0;
|
6988 |
|
|
|
6989 |
|
|
if (COMPARISON_CLASS_P (inner)
|
6990 |
|
|
&& TREE_CODE (TREE_OPERAND (inner, 1)) == INTEGER_CST)
|
6991 |
|
|
inner = TREE_OPERAND (inner, 0);
|
6992 |
|
|
|
6993 |
|
|
if (TREE_CODE (inner) == CALL_EXPR
|
6994 |
|
|
&& (fndecl = get_callee_fndecl (inner))
|
6995 |
|
|
&& DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_NORMAL
|
6996 |
|
|
&& DECL_FUNCTION_CODE (fndecl) == BUILT_IN_EXPECT)
|
6997 |
|
|
return arg0;
|
6998 |
|
|
|
6999 |
|
|
inner = inner_arg0;
|
7000 |
|
|
code = TREE_CODE (inner);
|
7001 |
|
|
if (code == TRUTH_ANDIF_EXPR || code == TRUTH_ORIF_EXPR)
|
7002 |
|
|
{
|
7003 |
|
|
tree op0 = TREE_OPERAND (inner, 0);
|
7004 |
|
|
tree op1 = TREE_OPERAND (inner, 1);
|
7005 |
|
|
|
7006 |
|
|
op0 = build_builtin_expect_predicate (loc, op0, arg1);
|
7007 |
|
|
op1 = build_builtin_expect_predicate (loc, op1, arg1);
|
7008 |
|
|
inner = build2 (code, TREE_TYPE (inner), op0, op1);
|
7009 |
|
|
|
7010 |
|
|
return fold_convert_loc (loc, TREE_TYPE (arg0), inner);
|
7011 |
|
|
}
|
7012 |
|
|
|
7013 |
|
|
/* If the argument isn't invariant then there's nothing else we can do. */
|
7014 |
|
|
if (!TREE_CONSTANT (inner_arg0))
|
7015 |
|
|
return NULL_TREE;
|
7016 |
|
|
|
7017 |
|
|
/* If we expect that a comparison against the argument will fold to
|
7018 |
|
|
a constant return the constant. In practice, this means a true
|
7019 |
|
|
constant or the address of a non-weak symbol. */
|
7020 |
|
|
inner = inner_arg0;
|
7021 |
|
|
STRIP_NOPS (inner);
|
7022 |
|
|
if (TREE_CODE (inner) == ADDR_EXPR)
|
7023 |
|
|
{
|
7024 |
|
|
do
|
7025 |
|
|
{
|
7026 |
|
|
inner = TREE_OPERAND (inner, 0);
|
7027 |
|
|
}
|
7028 |
|
|
while (TREE_CODE (inner) == COMPONENT_REF
|
7029 |
|
|
|| TREE_CODE (inner) == ARRAY_REF);
|
7030 |
|
|
if ((TREE_CODE (inner) == VAR_DECL
|
7031 |
|
|
|| TREE_CODE (inner) == FUNCTION_DECL)
|
7032 |
|
|
&& DECL_WEAK (inner))
|
7033 |
|
|
return NULL_TREE;
|
7034 |
|
|
}
|
7035 |
|
|
|
7036 |
|
|
/* Otherwise, ARG0 already has the proper type for the return value. */
|
7037 |
|
|
return arg0;
|
7038 |
|
|
}
|
7039 |
|
|
|
7040 |
|
|
/* Fold a call to __builtin_classify_type with argument ARG. */
|
7041 |
|
|
|
7042 |
|
|
static tree
|
7043 |
|
|
fold_builtin_classify_type (tree arg)
|
7044 |
|
|
{
|
7045 |
|
|
if (arg == 0)
|
7046 |
|
|
return build_int_cst (integer_type_node, no_type_class);
|
7047 |
|
|
|
7048 |
|
|
return build_int_cst (integer_type_node, type_to_class (TREE_TYPE (arg)));
|
7049 |
|
|
}
|
7050 |
|
|
|
7051 |
|
|
/* Fold a call to __builtin_strlen with argument ARG. */
|
7052 |
|
|
|
7053 |
|
|
static tree
|
7054 |
|
|
fold_builtin_strlen (location_t loc, tree type, tree arg)
|
7055 |
|
|
{
|
7056 |
|
|
if (!validate_arg (arg, POINTER_TYPE))
|
7057 |
|
|
return NULL_TREE;
|
7058 |
|
|
else
|
7059 |
|
|
{
|
7060 |
|
|
tree len = c_strlen (arg, 0);
|
7061 |
|
|
|
7062 |
|
|
if (len)
|
7063 |
|
|
return fold_convert_loc (loc, type, len);
|
7064 |
|
|
|
7065 |
|
|
return NULL_TREE;
|
7066 |
|
|
}
|
7067 |
|
|
}
|
7068 |
|
|
|
7069 |
|
|
/* Fold a call to __builtin_inf or __builtin_huge_val. */
|
7070 |
|
|
|
7071 |
|
|
static tree
|
7072 |
|
|
fold_builtin_inf (location_t loc, tree type, int warn)
|
7073 |
|
|
{
|
7074 |
|
|
REAL_VALUE_TYPE real;
|
7075 |
|
|
|
7076 |
|
|
/* __builtin_inff is intended to be usable to define INFINITY on all
|
7077 |
|
|
targets. If an infinity is not available, INFINITY expands "to a
|
7078 |
|
|
positive constant of type float that overflows at translation
|
7079 |
|
|
time", footnote "In this case, using INFINITY will violate the
|
7080 |
|
|
constraint in 6.4.4 and thus require a diagnostic." (C99 7.12#4).
|
7081 |
|
|
Thus we pedwarn to ensure this constraint violation is
|
7082 |
|
|
diagnosed. */
|
7083 |
|
|
if (!MODE_HAS_INFINITIES (TYPE_MODE (type)) && warn)
|
7084 |
|
|
pedwarn (loc, 0, "target format does not support infinity");
|
7085 |
|
|
|
7086 |
|
|
real_inf (&real);
|
7087 |
|
|
return build_real (type, real);
|
7088 |
|
|
}
|
7089 |
|
|
|
7090 |
|
|
/* Fold a call to __builtin_nan or __builtin_nans with argument ARG. */
|
7091 |
|
|
|
7092 |
|
|
static tree
|
7093 |
|
|
fold_builtin_nan (tree arg, tree type, int quiet)
|
7094 |
|
|
{
|
7095 |
|
|
REAL_VALUE_TYPE real;
|
7096 |
|
|
const char *str;
|
7097 |
|
|
|
7098 |
|
|
if (!validate_arg (arg, POINTER_TYPE))
|
7099 |
|
|
return NULL_TREE;
|
7100 |
|
|
str = c_getstr (arg);
|
7101 |
|
|
if (!str)
|
7102 |
|
|
return NULL_TREE;
|
7103 |
|
|
|
7104 |
|
|
if (!real_nan (&real, str, quiet, TYPE_MODE (type)))
|
7105 |
|
|
return NULL_TREE;
|
7106 |
|
|
|
7107 |
|
|
return build_real (type, real);
|
7108 |
|
|
}
|
7109 |
|
|
|
7110 |
|
|
/* Return true if the floating point expression T has an integer value.
|
7111 |
|
|
We also allow +Inf, -Inf and NaN to be considered integer values. */
|
7112 |
|
|
|
7113 |
|
|
static bool
|
7114 |
|
|
integer_valued_real_p (tree t)
|
7115 |
|
|
{
|
7116 |
|
|
switch (TREE_CODE (t))
|
7117 |
|
|
{
|
7118 |
|
|
case FLOAT_EXPR:
|
7119 |
|
|
return true;
|
7120 |
|
|
|
7121 |
|
|
case ABS_EXPR:
|
7122 |
|
|
case SAVE_EXPR:
|
7123 |
|
|
return integer_valued_real_p (TREE_OPERAND (t, 0));
|
7124 |
|
|
|
7125 |
|
|
case COMPOUND_EXPR:
|
7126 |
|
|
case MODIFY_EXPR:
|
7127 |
|
|
case BIND_EXPR:
|
7128 |
|
|
return integer_valued_real_p (TREE_OPERAND (t, 1));
|
7129 |
|
|
|
7130 |
|
|
case PLUS_EXPR:
|
7131 |
|
|
case MINUS_EXPR:
|
7132 |
|
|
case MULT_EXPR:
|
7133 |
|
|
case MIN_EXPR:
|
7134 |
|
|
case MAX_EXPR:
|
7135 |
|
|
return integer_valued_real_p (TREE_OPERAND (t, 0))
|
7136 |
|
|
&& integer_valued_real_p (TREE_OPERAND (t, 1));
|
7137 |
|
|
|
7138 |
|
|
case COND_EXPR:
|
7139 |
|
|
return integer_valued_real_p (TREE_OPERAND (t, 1))
|
7140 |
|
|
&& integer_valued_real_p (TREE_OPERAND (t, 2));
|
7141 |
|
|
|
7142 |
|
|
case REAL_CST:
|
7143 |
|
|
return real_isinteger (TREE_REAL_CST_PTR (t), TYPE_MODE (TREE_TYPE (t)));
|
7144 |
|
|
|
7145 |
|
|
case NOP_EXPR:
|
7146 |
|
|
{
|
7147 |
|
|
tree type = TREE_TYPE (TREE_OPERAND (t, 0));
|
7148 |
|
|
if (TREE_CODE (type) == INTEGER_TYPE)
|
7149 |
|
|
return true;
|
7150 |
|
|
if (TREE_CODE (type) == REAL_TYPE)
|
7151 |
|
|
return integer_valued_real_p (TREE_OPERAND (t, 0));
|
7152 |
|
|
break;
|
7153 |
|
|
}
|
7154 |
|
|
|
7155 |
|
|
case CALL_EXPR:
|
7156 |
|
|
switch (builtin_mathfn_code (t))
|
7157 |
|
|
{
|
7158 |
|
|
CASE_FLT_FN (BUILT_IN_CEIL):
|
7159 |
|
|
CASE_FLT_FN (BUILT_IN_FLOOR):
|
7160 |
|
|
CASE_FLT_FN (BUILT_IN_NEARBYINT):
|
7161 |
|
|
CASE_FLT_FN (BUILT_IN_RINT):
|
7162 |
|
|
CASE_FLT_FN (BUILT_IN_ROUND):
|
7163 |
|
|
CASE_FLT_FN (BUILT_IN_TRUNC):
|
7164 |
|
|
return true;
|
7165 |
|
|
|
7166 |
|
|
CASE_FLT_FN (BUILT_IN_FMIN):
|
7167 |
|
|
CASE_FLT_FN (BUILT_IN_FMAX):
|
7168 |
|
|
return integer_valued_real_p (CALL_EXPR_ARG (t, 0))
|
7169 |
|
|
&& integer_valued_real_p (CALL_EXPR_ARG (t, 1));
|
7170 |
|
|
|
7171 |
|
|
default:
|
7172 |
|
|
break;
|
7173 |
|
|
}
|
7174 |
|
|
break;
|
7175 |
|
|
|
7176 |
|
|
default:
|
7177 |
|
|
break;
|
7178 |
|
|
}
|
7179 |
|
|
return false;
|
7180 |
|
|
}
|
7181 |
|
|
|
7182 |
|
|
/* FNDECL is assumed to be a builtin where truncation can be propagated
|
7183 |
|
|
across (for instance floor((double)f) == (double)floorf (f).
|
7184 |
|
|
Do the transformation for a call with argument ARG. */
|
7185 |
|
|
|
7186 |
|
|
static tree
|
7187 |
|
|
fold_trunc_transparent_mathfn (location_t loc, tree fndecl, tree arg)
|
7188 |
|
|
{
|
7189 |
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
7190 |
|
|
|
7191 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
7192 |
|
|
return NULL_TREE;
|
7193 |
|
|
|
7194 |
|
|
/* Integer rounding functions are idempotent. */
|
7195 |
|
|
if (fcode == builtin_mathfn_code (arg))
|
7196 |
|
|
return arg;
|
7197 |
|
|
|
7198 |
|
|
/* If argument is already integer valued, and we don't need to worry
|
7199 |
|
|
about setting errno, there's no need to perform rounding. */
|
7200 |
|
|
if (! flag_errno_math && integer_valued_real_p (arg))
|
7201 |
|
|
return arg;
|
7202 |
|
|
|
7203 |
|
|
if (optimize)
|
7204 |
|
|
{
|
7205 |
|
|
tree arg0 = strip_float_extensions (arg);
|
7206 |
|
|
tree ftype = TREE_TYPE (TREE_TYPE (fndecl));
|
7207 |
|
|
tree newtype = TREE_TYPE (arg0);
|
7208 |
|
|
tree decl;
|
7209 |
|
|
|
7210 |
|
|
if (TYPE_PRECISION (newtype) < TYPE_PRECISION (ftype)
|
7211 |
|
|
&& (decl = mathfn_built_in (newtype, fcode)))
|
7212 |
|
|
return fold_convert_loc (loc, ftype,
|
7213 |
|
|
build_call_expr_loc (loc, decl, 1,
|
7214 |
|
|
fold_convert_loc (loc,
|
7215 |
|
|
newtype,
|
7216 |
|
|
arg0)));
|
7217 |
|
|
}
|
7218 |
|
|
return NULL_TREE;
|
7219 |
|
|
}
|
7220 |
|
|
|
7221 |
|
|
/* FNDECL is assumed to be builtin which can narrow the FP type of
|
7222 |
|
|
the argument, for instance lround((double)f) -> lroundf (f).
|
7223 |
|
|
Do the transformation for a call with argument ARG. */
|
7224 |
|
|
|
7225 |
|
|
static tree
|
7226 |
|
|
fold_fixed_mathfn (location_t loc, tree fndecl, tree arg)
|
7227 |
|
|
{
|
7228 |
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
7229 |
|
|
|
7230 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
7231 |
|
|
return NULL_TREE;
|
7232 |
|
|
|
7233 |
|
|
/* If argument is already integer valued, and we don't need to worry
|
7234 |
|
|
about setting errno, there's no need to perform rounding. */
|
7235 |
|
|
if (! flag_errno_math && integer_valued_real_p (arg))
|
7236 |
|
|
return fold_build1_loc (loc, FIX_TRUNC_EXPR,
|
7237 |
|
|
TREE_TYPE (TREE_TYPE (fndecl)), arg);
|
7238 |
|
|
|
7239 |
|
|
if (optimize)
|
7240 |
|
|
{
|
7241 |
|
|
tree ftype = TREE_TYPE (arg);
|
7242 |
|
|
tree arg0 = strip_float_extensions (arg);
|
7243 |
|
|
tree newtype = TREE_TYPE (arg0);
|
7244 |
|
|
tree decl;
|
7245 |
|
|
|
7246 |
|
|
if (TYPE_PRECISION (newtype) < TYPE_PRECISION (ftype)
|
7247 |
|
|
&& (decl = mathfn_built_in (newtype, fcode)))
|
7248 |
|
|
return build_call_expr_loc (loc, decl, 1,
|
7249 |
|
|
fold_convert_loc (loc, newtype, arg0));
|
7250 |
|
|
}
|
7251 |
|
|
|
7252 |
|
|
/* Canonicalize iround (x) to lround (x) on ILP32 targets where
|
7253 |
|
|
sizeof (int) == sizeof (long). */
|
7254 |
|
|
if (TYPE_PRECISION (integer_type_node)
|
7255 |
|
|
== TYPE_PRECISION (long_integer_type_node))
|
7256 |
|
|
{
|
7257 |
|
|
tree newfn = NULL_TREE;
|
7258 |
|
|
switch (fcode)
|
7259 |
|
|
{
|
7260 |
|
|
CASE_FLT_FN (BUILT_IN_ICEIL):
|
7261 |
|
|
newfn = mathfn_built_in (TREE_TYPE (arg), BUILT_IN_LCEIL);
|
7262 |
|
|
break;
|
7263 |
|
|
|
7264 |
|
|
CASE_FLT_FN (BUILT_IN_IFLOOR):
|
7265 |
|
|
newfn = mathfn_built_in (TREE_TYPE (arg), BUILT_IN_LFLOOR);
|
7266 |
|
|
break;
|
7267 |
|
|
|
7268 |
|
|
CASE_FLT_FN (BUILT_IN_IROUND):
|
7269 |
|
|
newfn = mathfn_built_in (TREE_TYPE (arg), BUILT_IN_LROUND);
|
7270 |
|
|
break;
|
7271 |
|
|
|
7272 |
|
|
CASE_FLT_FN (BUILT_IN_IRINT):
|
7273 |
|
|
newfn = mathfn_built_in (TREE_TYPE (arg), BUILT_IN_LRINT);
|
7274 |
|
|
break;
|
7275 |
|
|
|
7276 |
|
|
default:
|
7277 |
|
|
break;
|
7278 |
|
|
}
|
7279 |
|
|
|
7280 |
|
|
if (newfn)
|
7281 |
|
|
{
|
7282 |
|
|
tree newcall = build_call_expr_loc (loc, newfn, 1, arg);
|
7283 |
|
|
return fold_convert_loc (loc,
|
7284 |
|
|
TREE_TYPE (TREE_TYPE (fndecl)), newcall);
|
7285 |
|
|
}
|
7286 |
|
|
}
|
7287 |
|
|
|
7288 |
|
|
/* Canonicalize llround (x) to lround (x) on LP64 targets where
|
7289 |
|
|
sizeof (long long) == sizeof (long). */
|
7290 |
|
|
if (TYPE_PRECISION (long_long_integer_type_node)
|
7291 |
|
|
== TYPE_PRECISION (long_integer_type_node))
|
7292 |
|
|
{
|
7293 |
|
|
tree newfn = NULL_TREE;
|
7294 |
|
|
switch (fcode)
|
7295 |
|
|
{
|
7296 |
|
|
CASE_FLT_FN (BUILT_IN_LLCEIL):
|
7297 |
|
|
newfn = mathfn_built_in (TREE_TYPE (arg), BUILT_IN_LCEIL);
|
7298 |
|
|
break;
|
7299 |
|
|
|
7300 |
|
|
CASE_FLT_FN (BUILT_IN_LLFLOOR):
|
7301 |
|
|
newfn = mathfn_built_in (TREE_TYPE (arg), BUILT_IN_LFLOOR);
|
7302 |
|
|
break;
|
7303 |
|
|
|
7304 |
|
|
CASE_FLT_FN (BUILT_IN_LLROUND):
|
7305 |
|
|
newfn = mathfn_built_in (TREE_TYPE (arg), BUILT_IN_LROUND);
|
7306 |
|
|
break;
|
7307 |
|
|
|
7308 |
|
|
CASE_FLT_FN (BUILT_IN_LLRINT):
|
7309 |
|
|
newfn = mathfn_built_in (TREE_TYPE (arg), BUILT_IN_LRINT);
|
7310 |
|
|
break;
|
7311 |
|
|
|
7312 |
|
|
default:
|
7313 |
|
|
break;
|
7314 |
|
|
}
|
7315 |
|
|
|
7316 |
|
|
if (newfn)
|
7317 |
|
|
{
|
7318 |
|
|
tree newcall = build_call_expr_loc (loc, newfn, 1, arg);
|
7319 |
|
|
return fold_convert_loc (loc,
|
7320 |
|
|
TREE_TYPE (TREE_TYPE (fndecl)), newcall);
|
7321 |
|
|
}
|
7322 |
|
|
}
|
7323 |
|
|
|
7324 |
|
|
return NULL_TREE;
|
7325 |
|
|
}
|
7326 |
|
|
|
7327 |
|
|
/* Fold call to builtin cabs, cabsf or cabsl with argument ARG. TYPE is the
|
7328 |
|
|
return type. Return NULL_TREE if no simplification can be made. */
|
7329 |
|
|
|
7330 |
|
|
static tree
|
7331 |
|
|
fold_builtin_cabs (location_t loc, tree arg, tree type, tree fndecl)
|
7332 |
|
|
{
|
7333 |
|
|
tree res;
|
7334 |
|
|
|
7335 |
|
|
if (!validate_arg (arg, COMPLEX_TYPE)
|
7336 |
|
|
|| TREE_CODE (TREE_TYPE (TREE_TYPE (arg))) != REAL_TYPE)
|
7337 |
|
|
return NULL_TREE;
|
7338 |
|
|
|
7339 |
|
|
/* Calculate the result when the argument is a constant. */
|
7340 |
|
|
if (TREE_CODE (arg) == COMPLEX_CST
|
7341 |
|
|
&& (res = do_mpfr_arg2 (TREE_REALPART (arg), TREE_IMAGPART (arg),
|
7342 |
|
|
type, mpfr_hypot)))
|
7343 |
|
|
return res;
|
7344 |
|
|
|
7345 |
|
|
if (TREE_CODE (arg) == COMPLEX_EXPR)
|
7346 |
|
|
{
|
7347 |
|
|
tree real = TREE_OPERAND (arg, 0);
|
7348 |
|
|
tree imag = TREE_OPERAND (arg, 1);
|
7349 |
|
|
|
7350 |
|
|
/* If either part is zero, cabs is fabs of the other. */
|
7351 |
|
|
if (real_zerop (real))
|
7352 |
|
|
return fold_build1_loc (loc, ABS_EXPR, type, imag);
|
7353 |
|
|
if (real_zerop (imag))
|
7354 |
|
|
return fold_build1_loc (loc, ABS_EXPR, type, real);
|
7355 |
|
|
|
7356 |
|
|
/* cabs(x+xi) -> fabs(x)*sqrt(2). */
|
7357 |
|
|
if (flag_unsafe_math_optimizations
|
7358 |
|
|
&& operand_equal_p (real, imag, OEP_PURE_SAME))
|
7359 |
|
|
{
|
7360 |
|
|
const REAL_VALUE_TYPE sqrt2_trunc
|
7361 |
|
|
= real_value_truncate (TYPE_MODE (type), dconst_sqrt2 ());
|
7362 |
|
|
STRIP_NOPS (real);
|
7363 |
|
|
return fold_build2_loc (loc, MULT_EXPR, type,
|
7364 |
|
|
fold_build1_loc (loc, ABS_EXPR, type, real),
|
7365 |
|
|
build_real (type, sqrt2_trunc));
|
7366 |
|
|
}
|
7367 |
|
|
}
|
7368 |
|
|
|
7369 |
|
|
/* Optimize cabs(-z) and cabs(conj(z)) as cabs(z). */
|
7370 |
|
|
if (TREE_CODE (arg) == NEGATE_EXPR
|
7371 |
|
|
|| TREE_CODE (arg) == CONJ_EXPR)
|
7372 |
|
|
return build_call_expr_loc (loc, fndecl, 1, TREE_OPERAND (arg, 0));
|
7373 |
|
|
|
7374 |
|
|
/* Don't do this when optimizing for size. */
|
7375 |
|
|
if (flag_unsafe_math_optimizations
|
7376 |
|
|
&& optimize && optimize_function_for_speed_p (cfun))
|
7377 |
|
|
{
|
7378 |
|
|
tree sqrtfn = mathfn_built_in (type, BUILT_IN_SQRT);
|
7379 |
|
|
|
7380 |
|
|
if (sqrtfn != NULL_TREE)
|
7381 |
|
|
{
|
7382 |
|
|
tree rpart, ipart, result;
|
7383 |
|
|
|
7384 |
|
|
arg = builtin_save_expr (arg);
|
7385 |
|
|
|
7386 |
|
|
rpart = fold_build1_loc (loc, REALPART_EXPR, type, arg);
|
7387 |
|
|
ipart = fold_build1_loc (loc, IMAGPART_EXPR, type, arg);
|
7388 |
|
|
|
7389 |
|
|
rpart = builtin_save_expr (rpart);
|
7390 |
|
|
ipart = builtin_save_expr (ipart);
|
7391 |
|
|
|
7392 |
|
|
result = fold_build2_loc (loc, PLUS_EXPR, type,
|
7393 |
|
|
fold_build2_loc (loc, MULT_EXPR, type,
|
7394 |
|
|
rpart, rpart),
|
7395 |
|
|
fold_build2_loc (loc, MULT_EXPR, type,
|
7396 |
|
|
ipart, ipart));
|
7397 |
|
|
|
7398 |
|
|
return build_call_expr_loc (loc, sqrtfn, 1, result);
|
7399 |
|
|
}
|
7400 |
|
|
}
|
7401 |
|
|
|
7402 |
|
|
return NULL_TREE;
|
7403 |
|
|
}
|
7404 |
|
|
|
7405 |
|
|
/* Build a complex (inf +- 0i) for the result of cproj. TYPE is the
|
7406 |
|
|
complex tree type of the result. If NEG is true, the imaginary
|
7407 |
|
|
zero is negative. */
|
7408 |
|
|
|
7409 |
|
|
static tree
|
7410 |
|
|
build_complex_cproj (tree type, bool neg)
|
7411 |
|
|
{
|
7412 |
|
|
REAL_VALUE_TYPE rinf, rzero = dconst0;
|
7413 |
|
|
|
7414 |
|
|
real_inf (&rinf);
|
7415 |
|
|
rzero.sign = neg;
|
7416 |
|
|
return build_complex (type, build_real (TREE_TYPE (type), rinf),
|
7417 |
|
|
build_real (TREE_TYPE (type), rzero));
|
7418 |
|
|
}
|
7419 |
|
|
|
7420 |
|
|
/* Fold call to builtin cproj, cprojf or cprojl with argument ARG. TYPE is the
|
7421 |
|
|
return type. Return NULL_TREE if no simplification can be made. */
|
7422 |
|
|
|
7423 |
|
|
static tree
|
7424 |
|
|
fold_builtin_cproj (location_t loc, tree arg, tree type)
|
7425 |
|
|
{
|
7426 |
|
|
if (!validate_arg (arg, COMPLEX_TYPE)
|
7427 |
|
|
|| TREE_CODE (TREE_TYPE (TREE_TYPE (arg))) != REAL_TYPE)
|
7428 |
|
|
return NULL_TREE;
|
7429 |
|
|
|
7430 |
|
|
/* If there are no infinities, return arg. */
|
7431 |
|
|
if (! HONOR_INFINITIES (TYPE_MODE (TREE_TYPE (type))))
|
7432 |
|
|
return non_lvalue_loc (loc, arg);
|
7433 |
|
|
|
7434 |
|
|
/* Calculate the result when the argument is a constant. */
|
7435 |
|
|
if (TREE_CODE (arg) == COMPLEX_CST)
|
7436 |
|
|
{
|
7437 |
|
|
const REAL_VALUE_TYPE *real = TREE_REAL_CST_PTR (TREE_REALPART (arg));
|
7438 |
|
|
const REAL_VALUE_TYPE *imag = TREE_REAL_CST_PTR (TREE_IMAGPART (arg));
|
7439 |
|
|
|
7440 |
|
|
if (real_isinf (real) || real_isinf (imag))
|
7441 |
|
|
return build_complex_cproj (type, imag->sign);
|
7442 |
|
|
else
|
7443 |
|
|
return arg;
|
7444 |
|
|
}
|
7445 |
|
|
else if (TREE_CODE (arg) == COMPLEX_EXPR)
|
7446 |
|
|
{
|
7447 |
|
|
tree real = TREE_OPERAND (arg, 0);
|
7448 |
|
|
tree imag = TREE_OPERAND (arg, 1);
|
7449 |
|
|
|
7450 |
|
|
STRIP_NOPS (real);
|
7451 |
|
|
STRIP_NOPS (imag);
|
7452 |
|
|
|
7453 |
|
|
/* If the real part is inf and the imag part is known to be
|
7454 |
|
|
nonnegative, return (inf + 0i). Remember side-effects are
|
7455 |
|
|
possible in the imag part. */
|
7456 |
|
|
if (TREE_CODE (real) == REAL_CST
|
7457 |
|
|
&& real_isinf (TREE_REAL_CST_PTR (real))
|
7458 |
|
|
&& tree_expr_nonnegative_p (imag))
|
7459 |
|
|
return omit_one_operand_loc (loc, type,
|
7460 |
|
|
build_complex_cproj (type, false),
|
7461 |
|
|
arg);
|
7462 |
|
|
|
7463 |
|
|
/* If the imag part is inf, return (inf+I*copysign(0,imag)).
|
7464 |
|
|
Remember side-effects are possible in the real part. */
|
7465 |
|
|
if (TREE_CODE (imag) == REAL_CST
|
7466 |
|
|
&& real_isinf (TREE_REAL_CST_PTR (imag)))
|
7467 |
|
|
return
|
7468 |
|
|
omit_one_operand_loc (loc, type,
|
7469 |
|
|
build_complex_cproj (type, TREE_REAL_CST_PTR
|
7470 |
|
|
(imag)->sign), arg);
|
7471 |
|
|
}
|
7472 |
|
|
|
7473 |
|
|
return NULL_TREE;
|
7474 |
|
|
}
|
7475 |
|
|
|
7476 |
|
|
/* Fold a builtin function call to sqrt, sqrtf, or sqrtl with argument ARG.
|
7477 |
|
|
Return NULL_TREE if no simplification can be made. */
|
7478 |
|
|
|
7479 |
|
|
static tree
|
7480 |
|
|
fold_builtin_sqrt (location_t loc, tree arg, tree type)
|
7481 |
|
|
{
|
7482 |
|
|
|
7483 |
|
|
enum built_in_function fcode;
|
7484 |
|
|
tree res;
|
7485 |
|
|
|
7486 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
7487 |
|
|
return NULL_TREE;
|
7488 |
|
|
|
7489 |
|
|
/* Calculate the result when the argument is a constant. */
|
7490 |
|
|
if ((res = do_mpfr_arg1 (arg, type, mpfr_sqrt, &dconst0, NULL, true)))
|
7491 |
|
|
return res;
|
7492 |
|
|
|
7493 |
|
|
/* Optimize sqrt(expN(x)) = expN(x*0.5). */
|
7494 |
|
|
fcode = builtin_mathfn_code (arg);
|
7495 |
|
|
if (flag_unsafe_math_optimizations && BUILTIN_EXPONENT_P (fcode))
|
7496 |
|
|
{
|
7497 |
|
|
tree expfn = TREE_OPERAND (CALL_EXPR_FN (arg), 0);
|
7498 |
|
|
arg = fold_build2_loc (loc, MULT_EXPR, type,
|
7499 |
|
|
CALL_EXPR_ARG (arg, 0),
|
7500 |
|
|
build_real (type, dconsthalf));
|
7501 |
|
|
return build_call_expr_loc (loc, expfn, 1, arg);
|
7502 |
|
|
}
|
7503 |
|
|
|
7504 |
|
|
/* Optimize sqrt(Nroot(x)) -> pow(x,1/(2*N)). */
|
7505 |
|
|
if (flag_unsafe_math_optimizations && BUILTIN_ROOT_P (fcode))
|
7506 |
|
|
{
|
7507 |
|
|
tree powfn = mathfn_built_in (type, BUILT_IN_POW);
|
7508 |
|
|
|
7509 |
|
|
if (powfn)
|
7510 |
|
|
{
|
7511 |
|
|
tree arg0 = CALL_EXPR_ARG (arg, 0);
|
7512 |
|
|
tree tree_root;
|
7513 |
|
|
/* The inner root was either sqrt or cbrt. */
|
7514 |
|
|
/* This was a conditional expression but it triggered a bug
|
7515 |
|
|
in Sun C 5.5. */
|
7516 |
|
|
REAL_VALUE_TYPE dconstroot;
|
7517 |
|
|
if (BUILTIN_SQRT_P (fcode))
|
7518 |
|
|
dconstroot = dconsthalf;
|
7519 |
|
|
else
|
7520 |
|
|
dconstroot = dconst_third ();
|
7521 |
|
|
|
7522 |
|
|
/* Adjust for the outer root. */
|
7523 |
|
|
SET_REAL_EXP (&dconstroot, REAL_EXP (&dconstroot) - 1);
|
7524 |
|
|
dconstroot = real_value_truncate (TYPE_MODE (type), dconstroot);
|
7525 |
|
|
tree_root = build_real (type, dconstroot);
|
7526 |
|
|
return build_call_expr_loc (loc, powfn, 2, arg0, tree_root);
|
7527 |
|
|
}
|
7528 |
|
|
}
|
7529 |
|
|
|
7530 |
|
|
/* Optimize sqrt(pow(x,y)) = pow(|x|,y*0.5). */
|
7531 |
|
|
if (flag_unsafe_math_optimizations
|
7532 |
|
|
&& (fcode == BUILT_IN_POW
|
7533 |
|
|
|| fcode == BUILT_IN_POWF
|
7534 |
|
|
|| fcode == BUILT_IN_POWL))
|
7535 |
|
|
{
|
7536 |
|
|
tree powfn = TREE_OPERAND (CALL_EXPR_FN (arg), 0);
|
7537 |
|
|
tree arg0 = CALL_EXPR_ARG (arg, 0);
|
7538 |
|
|
tree arg1 = CALL_EXPR_ARG (arg, 1);
|
7539 |
|
|
tree narg1;
|
7540 |
|
|
if (!tree_expr_nonnegative_p (arg0))
|
7541 |
|
|
arg0 = build1 (ABS_EXPR, type, arg0);
|
7542 |
|
|
narg1 = fold_build2_loc (loc, MULT_EXPR, type, arg1,
|
7543 |
|
|
build_real (type, dconsthalf));
|
7544 |
|
|
return build_call_expr_loc (loc, powfn, 2, arg0, narg1);
|
7545 |
|
|
}
|
7546 |
|
|
|
7547 |
|
|
return NULL_TREE;
|
7548 |
|
|
}
|
7549 |
|
|
|
7550 |
|
|
/* Fold a builtin function call to cbrt, cbrtf, or cbrtl with argument ARG.
|
7551 |
|
|
Return NULL_TREE if no simplification can be made. */
|
7552 |
|
|
|
7553 |
|
|
static tree
|
7554 |
|
|
fold_builtin_cbrt (location_t loc, tree arg, tree type)
|
7555 |
|
|
{
|
7556 |
|
|
const enum built_in_function fcode = builtin_mathfn_code (arg);
|
7557 |
|
|
tree res;
|
7558 |
|
|
|
7559 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
7560 |
|
|
return NULL_TREE;
|
7561 |
|
|
|
7562 |
|
|
/* Calculate the result when the argument is a constant. */
|
7563 |
|
|
if ((res = do_mpfr_arg1 (arg, type, mpfr_cbrt, NULL, NULL, 0)))
|
7564 |
|
|
return res;
|
7565 |
|
|
|
7566 |
|
|
if (flag_unsafe_math_optimizations)
|
7567 |
|
|
{
|
7568 |
|
|
/* Optimize cbrt(expN(x)) -> expN(x/3). */
|
7569 |
|
|
if (BUILTIN_EXPONENT_P (fcode))
|
7570 |
|
|
{
|
7571 |
|
|
tree expfn = TREE_OPERAND (CALL_EXPR_FN (arg), 0);
|
7572 |
|
|
const REAL_VALUE_TYPE third_trunc =
|
7573 |
|
|
real_value_truncate (TYPE_MODE (type), dconst_third ());
|
7574 |
|
|
arg = fold_build2_loc (loc, MULT_EXPR, type,
|
7575 |
|
|
CALL_EXPR_ARG (arg, 0),
|
7576 |
|
|
build_real (type, third_trunc));
|
7577 |
|
|
return build_call_expr_loc (loc, expfn, 1, arg);
|
7578 |
|
|
}
|
7579 |
|
|
|
7580 |
|
|
/* Optimize cbrt(sqrt(x)) -> pow(x,1/6). */
|
7581 |
|
|
if (BUILTIN_SQRT_P (fcode))
|
7582 |
|
|
{
|
7583 |
|
|
tree powfn = mathfn_built_in (type, BUILT_IN_POW);
|
7584 |
|
|
|
7585 |
|
|
if (powfn)
|
7586 |
|
|
{
|
7587 |
|
|
tree arg0 = CALL_EXPR_ARG (arg, 0);
|
7588 |
|
|
tree tree_root;
|
7589 |
|
|
REAL_VALUE_TYPE dconstroot = dconst_third ();
|
7590 |
|
|
|
7591 |
|
|
SET_REAL_EXP (&dconstroot, REAL_EXP (&dconstroot) - 1);
|
7592 |
|
|
dconstroot = real_value_truncate (TYPE_MODE (type), dconstroot);
|
7593 |
|
|
tree_root = build_real (type, dconstroot);
|
7594 |
|
|
return build_call_expr_loc (loc, powfn, 2, arg0, tree_root);
|
7595 |
|
|
}
|
7596 |
|
|
}
|
7597 |
|
|
|
7598 |
|
|
/* Optimize cbrt(cbrt(x)) -> pow(x,1/9) iff x is nonnegative. */
|
7599 |
|
|
if (BUILTIN_CBRT_P (fcode))
|
7600 |
|
|
{
|
7601 |
|
|
tree arg0 = CALL_EXPR_ARG (arg, 0);
|
7602 |
|
|
if (tree_expr_nonnegative_p (arg0))
|
7603 |
|
|
{
|
7604 |
|
|
tree powfn = mathfn_built_in (type, BUILT_IN_POW);
|
7605 |
|
|
|
7606 |
|
|
if (powfn)
|
7607 |
|
|
{
|
7608 |
|
|
tree tree_root;
|
7609 |
|
|
REAL_VALUE_TYPE dconstroot;
|
7610 |
|
|
|
7611 |
|
|
real_arithmetic (&dconstroot, MULT_EXPR,
|
7612 |
|
|
dconst_third_ptr (), dconst_third_ptr ());
|
7613 |
|
|
dconstroot = real_value_truncate (TYPE_MODE (type), dconstroot);
|
7614 |
|
|
tree_root = build_real (type, dconstroot);
|
7615 |
|
|
return build_call_expr_loc (loc, powfn, 2, arg0, tree_root);
|
7616 |
|
|
}
|
7617 |
|
|
}
|
7618 |
|
|
}
|
7619 |
|
|
|
7620 |
|
|
/* Optimize cbrt(pow(x,y)) -> pow(x,y/3) iff x is nonnegative. */
|
7621 |
|
|
if (fcode == BUILT_IN_POW
|
7622 |
|
|
|| fcode == BUILT_IN_POWF
|
7623 |
|
|
|| fcode == BUILT_IN_POWL)
|
7624 |
|
|
{
|
7625 |
|
|
tree arg00 = CALL_EXPR_ARG (arg, 0);
|
7626 |
|
|
tree arg01 = CALL_EXPR_ARG (arg, 1);
|
7627 |
|
|
if (tree_expr_nonnegative_p (arg00))
|
7628 |
|
|
{
|
7629 |
|
|
tree powfn = TREE_OPERAND (CALL_EXPR_FN (arg), 0);
|
7630 |
|
|
const REAL_VALUE_TYPE dconstroot
|
7631 |
|
|
= real_value_truncate (TYPE_MODE (type), dconst_third ());
|
7632 |
|
|
tree narg01 = fold_build2_loc (loc, MULT_EXPR, type, arg01,
|
7633 |
|
|
build_real (type, dconstroot));
|
7634 |
|
|
return build_call_expr_loc (loc, powfn, 2, arg00, narg01);
|
7635 |
|
|
}
|
7636 |
|
|
}
|
7637 |
|
|
}
|
7638 |
|
|
return NULL_TREE;
|
7639 |
|
|
}
|
7640 |
|
|
|
7641 |
|
|
/* Fold function call to builtin cos, cosf, or cosl with argument ARG.
|
7642 |
|
|
TYPE is the type of the return value. Return NULL_TREE if no
|
7643 |
|
|
simplification can be made. */
|
7644 |
|
|
|
7645 |
|
|
static tree
|
7646 |
|
|
fold_builtin_cos (location_t loc,
|
7647 |
|
|
tree arg, tree type, tree fndecl)
|
7648 |
|
|
{
|
7649 |
|
|
tree res, narg;
|
7650 |
|
|
|
7651 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
7652 |
|
|
return NULL_TREE;
|
7653 |
|
|
|
7654 |
|
|
/* Calculate the result when the argument is a constant. */
|
7655 |
|
|
if ((res = do_mpfr_arg1 (arg, type, mpfr_cos, NULL, NULL, 0)))
|
7656 |
|
|
return res;
|
7657 |
|
|
|
7658 |
|
|
/* Optimize cos(-x) into cos (x). */
|
7659 |
|
|
if ((narg = fold_strip_sign_ops (arg)))
|
7660 |
|
|
return build_call_expr_loc (loc, fndecl, 1, narg);
|
7661 |
|
|
|
7662 |
|
|
return NULL_TREE;
|
7663 |
|
|
}
|
7664 |
|
|
|
7665 |
|
|
/* Fold function call to builtin cosh, coshf, or coshl with argument ARG.
|
7666 |
|
|
Return NULL_TREE if no simplification can be made. */
|
7667 |
|
|
|
7668 |
|
|
static tree
|
7669 |
|
|
fold_builtin_cosh (location_t loc, tree arg, tree type, tree fndecl)
|
7670 |
|
|
{
|
7671 |
|
|
if (validate_arg (arg, REAL_TYPE))
|
7672 |
|
|
{
|
7673 |
|
|
tree res, narg;
|
7674 |
|
|
|
7675 |
|
|
/* Calculate the result when the argument is a constant. */
|
7676 |
|
|
if ((res = do_mpfr_arg1 (arg, type, mpfr_cosh, NULL, NULL, 0)))
|
7677 |
|
|
return res;
|
7678 |
|
|
|
7679 |
|
|
/* Optimize cosh(-x) into cosh (x). */
|
7680 |
|
|
if ((narg = fold_strip_sign_ops (arg)))
|
7681 |
|
|
return build_call_expr_loc (loc, fndecl, 1, narg);
|
7682 |
|
|
}
|
7683 |
|
|
|
7684 |
|
|
return NULL_TREE;
|
7685 |
|
|
}
|
7686 |
|
|
|
7687 |
|
|
/* Fold function call to builtin ccos (or ccosh if HYPER is TRUE) with
|
7688 |
|
|
argument ARG. TYPE is the type of the return value. Return
|
7689 |
|
|
NULL_TREE if no simplification can be made. */
|
7690 |
|
|
|
7691 |
|
|
static tree
|
7692 |
|
|
fold_builtin_ccos (location_t loc, tree arg, tree type, tree fndecl,
|
7693 |
|
|
bool hyper)
|
7694 |
|
|
{
|
7695 |
|
|
if (validate_arg (arg, COMPLEX_TYPE)
|
7696 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg))) == REAL_TYPE)
|
7697 |
|
|
{
|
7698 |
|
|
tree tmp;
|
7699 |
|
|
|
7700 |
|
|
/* Calculate the result when the argument is a constant. */
|
7701 |
|
|
if ((tmp = do_mpc_arg1 (arg, type, (hyper ? mpc_cosh : mpc_cos))))
|
7702 |
|
|
return tmp;
|
7703 |
|
|
|
7704 |
|
|
/* Optimize fn(-x) into fn(x). */
|
7705 |
|
|
if ((tmp = fold_strip_sign_ops (arg)))
|
7706 |
|
|
return build_call_expr_loc (loc, fndecl, 1, tmp);
|
7707 |
|
|
}
|
7708 |
|
|
|
7709 |
|
|
return NULL_TREE;
|
7710 |
|
|
}
|
7711 |
|
|
|
7712 |
|
|
/* Fold function call to builtin tan, tanf, or tanl with argument ARG.
|
7713 |
|
|
Return NULL_TREE if no simplification can be made. */
|
7714 |
|
|
|
7715 |
|
|
static tree
|
7716 |
|
|
fold_builtin_tan (tree arg, tree type)
|
7717 |
|
|
{
|
7718 |
|
|
enum built_in_function fcode;
|
7719 |
|
|
tree res;
|
7720 |
|
|
|
7721 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
7722 |
|
|
return NULL_TREE;
|
7723 |
|
|
|
7724 |
|
|
/* Calculate the result when the argument is a constant. */
|
7725 |
|
|
if ((res = do_mpfr_arg1 (arg, type, mpfr_tan, NULL, NULL, 0)))
|
7726 |
|
|
return res;
|
7727 |
|
|
|
7728 |
|
|
/* Optimize tan(atan(x)) = x. */
|
7729 |
|
|
fcode = builtin_mathfn_code (arg);
|
7730 |
|
|
if (flag_unsafe_math_optimizations
|
7731 |
|
|
&& (fcode == BUILT_IN_ATAN
|
7732 |
|
|
|| fcode == BUILT_IN_ATANF
|
7733 |
|
|
|| fcode == BUILT_IN_ATANL))
|
7734 |
|
|
return CALL_EXPR_ARG (arg, 0);
|
7735 |
|
|
|
7736 |
|
|
return NULL_TREE;
|
7737 |
|
|
}
|
7738 |
|
|
|
7739 |
|
|
/* Fold function call to builtin sincos, sincosf, or sincosl. Return
|
7740 |
|
|
NULL_TREE if no simplification can be made. */
|
7741 |
|
|
|
7742 |
|
|
static tree
|
7743 |
|
|
fold_builtin_sincos (location_t loc,
|
7744 |
|
|
tree arg0, tree arg1, tree arg2)
|
7745 |
|
|
{
|
7746 |
|
|
tree type;
|
7747 |
|
|
tree res, fn, call;
|
7748 |
|
|
|
7749 |
|
|
if (!validate_arg (arg0, REAL_TYPE)
|
7750 |
|
|
|| !validate_arg (arg1, POINTER_TYPE)
|
7751 |
|
|
|| !validate_arg (arg2, POINTER_TYPE))
|
7752 |
|
|
return NULL_TREE;
|
7753 |
|
|
|
7754 |
|
|
type = TREE_TYPE (arg0);
|
7755 |
|
|
|
7756 |
|
|
/* Calculate the result when the argument is a constant. */
|
7757 |
|
|
if ((res = do_mpfr_sincos (arg0, arg1, arg2)))
|
7758 |
|
|
return res;
|
7759 |
|
|
|
7760 |
|
|
/* Canonicalize sincos to cexpi. */
|
7761 |
|
|
if (!TARGET_C99_FUNCTIONS)
|
7762 |
|
|
return NULL_TREE;
|
7763 |
|
|
fn = mathfn_built_in (type, BUILT_IN_CEXPI);
|
7764 |
|
|
if (!fn)
|
7765 |
|
|
return NULL_TREE;
|
7766 |
|
|
|
7767 |
|
|
call = build_call_expr_loc (loc, fn, 1, arg0);
|
7768 |
|
|
call = builtin_save_expr (call);
|
7769 |
|
|
|
7770 |
|
|
return build2 (COMPOUND_EXPR, void_type_node,
|
7771 |
|
|
build2 (MODIFY_EXPR, void_type_node,
|
7772 |
|
|
build_fold_indirect_ref_loc (loc, arg1),
|
7773 |
|
|
build1 (IMAGPART_EXPR, type, call)),
|
7774 |
|
|
build2 (MODIFY_EXPR, void_type_node,
|
7775 |
|
|
build_fold_indirect_ref_loc (loc, arg2),
|
7776 |
|
|
build1 (REALPART_EXPR, type, call)));
|
7777 |
|
|
}
|
7778 |
|
|
|
7779 |
|
|
/* Fold function call to builtin cexp, cexpf, or cexpl. Return
|
7780 |
|
|
NULL_TREE if no simplification can be made. */
|
7781 |
|
|
|
7782 |
|
|
static tree
|
7783 |
|
|
fold_builtin_cexp (location_t loc, tree arg0, tree type)
|
7784 |
|
|
{
|
7785 |
|
|
tree rtype;
|
7786 |
|
|
tree realp, imagp, ifn;
|
7787 |
|
|
tree res;
|
7788 |
|
|
|
7789 |
|
|
if (!validate_arg (arg0, COMPLEX_TYPE)
|
7790 |
|
|
|| TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) != REAL_TYPE)
|
7791 |
|
|
return NULL_TREE;
|
7792 |
|
|
|
7793 |
|
|
/* Calculate the result when the argument is a constant. */
|
7794 |
|
|
if ((res = do_mpc_arg1 (arg0, type, mpc_exp)))
|
7795 |
|
|
return res;
|
7796 |
|
|
|
7797 |
|
|
rtype = TREE_TYPE (TREE_TYPE (arg0));
|
7798 |
|
|
|
7799 |
|
|
/* In case we can figure out the real part of arg0 and it is constant zero
|
7800 |
|
|
fold to cexpi. */
|
7801 |
|
|
if (!TARGET_C99_FUNCTIONS)
|
7802 |
|
|
return NULL_TREE;
|
7803 |
|
|
ifn = mathfn_built_in (rtype, BUILT_IN_CEXPI);
|
7804 |
|
|
if (!ifn)
|
7805 |
|
|
return NULL_TREE;
|
7806 |
|
|
|
7807 |
|
|
if ((realp = fold_unary_loc (loc, REALPART_EXPR, rtype, arg0))
|
7808 |
|
|
&& real_zerop (realp))
|
7809 |
|
|
{
|
7810 |
|
|
tree narg = fold_build1_loc (loc, IMAGPART_EXPR, rtype, arg0);
|
7811 |
|
|
return build_call_expr_loc (loc, ifn, 1, narg);
|
7812 |
|
|
}
|
7813 |
|
|
|
7814 |
|
|
/* In case we can easily decompose real and imaginary parts split cexp
|
7815 |
|
|
to exp (r) * cexpi (i). */
|
7816 |
|
|
if (flag_unsafe_math_optimizations
|
7817 |
|
|
&& realp)
|
7818 |
|
|
{
|
7819 |
|
|
tree rfn, rcall, icall;
|
7820 |
|
|
|
7821 |
|
|
rfn = mathfn_built_in (rtype, BUILT_IN_EXP);
|
7822 |
|
|
if (!rfn)
|
7823 |
|
|
return NULL_TREE;
|
7824 |
|
|
|
7825 |
|
|
imagp = fold_unary_loc (loc, IMAGPART_EXPR, rtype, arg0);
|
7826 |
|
|
if (!imagp)
|
7827 |
|
|
return NULL_TREE;
|
7828 |
|
|
|
7829 |
|
|
icall = build_call_expr_loc (loc, ifn, 1, imagp);
|
7830 |
|
|
icall = builtin_save_expr (icall);
|
7831 |
|
|
rcall = build_call_expr_loc (loc, rfn, 1, realp);
|
7832 |
|
|
rcall = builtin_save_expr (rcall);
|
7833 |
|
|
return fold_build2_loc (loc, COMPLEX_EXPR, type,
|
7834 |
|
|
fold_build2_loc (loc, MULT_EXPR, rtype,
|
7835 |
|
|
rcall,
|
7836 |
|
|
fold_build1_loc (loc, REALPART_EXPR,
|
7837 |
|
|
rtype, icall)),
|
7838 |
|
|
fold_build2_loc (loc, MULT_EXPR, rtype,
|
7839 |
|
|
rcall,
|
7840 |
|
|
fold_build1_loc (loc, IMAGPART_EXPR,
|
7841 |
|
|
rtype, icall)));
|
7842 |
|
|
}
|
7843 |
|
|
|
7844 |
|
|
return NULL_TREE;
|
7845 |
|
|
}
|
7846 |
|
|
|
7847 |
|
|
/* Fold function call to builtin trunc, truncf or truncl with argument ARG.
|
7848 |
|
|
Return NULL_TREE if no simplification can be made. */
|
7849 |
|
|
|
7850 |
|
|
static tree
|
7851 |
|
|
fold_builtin_trunc (location_t loc, tree fndecl, tree arg)
|
7852 |
|
|
{
|
7853 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
7854 |
|
|
return NULL_TREE;
|
7855 |
|
|
|
7856 |
|
|
/* Optimize trunc of constant value. */
|
7857 |
|
|
if (TREE_CODE (arg) == REAL_CST && !TREE_OVERFLOW (arg))
|
7858 |
|
|
{
|
7859 |
|
|
REAL_VALUE_TYPE r, x;
|
7860 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
7861 |
|
|
|
7862 |
|
|
x = TREE_REAL_CST (arg);
|
7863 |
|
|
real_trunc (&r, TYPE_MODE (type), &x);
|
7864 |
|
|
return build_real (type, r);
|
7865 |
|
|
}
|
7866 |
|
|
|
7867 |
|
|
return fold_trunc_transparent_mathfn (loc, fndecl, arg);
|
7868 |
|
|
}
|
7869 |
|
|
|
7870 |
|
|
/* Fold function call to builtin floor, floorf or floorl with argument ARG.
|
7871 |
|
|
Return NULL_TREE if no simplification can be made. */
|
7872 |
|
|
|
7873 |
|
|
static tree
|
7874 |
|
|
fold_builtin_floor (location_t loc, tree fndecl, tree arg)
|
7875 |
|
|
{
|
7876 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
7877 |
|
|
return NULL_TREE;
|
7878 |
|
|
|
7879 |
|
|
/* Optimize floor of constant value. */
|
7880 |
|
|
if (TREE_CODE (arg) == REAL_CST && !TREE_OVERFLOW (arg))
|
7881 |
|
|
{
|
7882 |
|
|
REAL_VALUE_TYPE x;
|
7883 |
|
|
|
7884 |
|
|
x = TREE_REAL_CST (arg);
|
7885 |
|
|
if (! REAL_VALUE_ISNAN (x) || ! flag_errno_math)
|
7886 |
|
|
{
|
7887 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
7888 |
|
|
REAL_VALUE_TYPE r;
|
7889 |
|
|
|
7890 |
|
|
real_floor (&r, TYPE_MODE (type), &x);
|
7891 |
|
|
return build_real (type, r);
|
7892 |
|
|
}
|
7893 |
|
|
}
|
7894 |
|
|
|
7895 |
|
|
/* Fold floor (x) where x is nonnegative to trunc (x). */
|
7896 |
|
|
if (tree_expr_nonnegative_p (arg))
|
7897 |
|
|
{
|
7898 |
|
|
tree truncfn = mathfn_built_in (TREE_TYPE (arg), BUILT_IN_TRUNC);
|
7899 |
|
|
if (truncfn)
|
7900 |
|
|
return build_call_expr_loc (loc, truncfn, 1, arg);
|
7901 |
|
|
}
|
7902 |
|
|
|
7903 |
|
|
return fold_trunc_transparent_mathfn (loc, fndecl, arg);
|
7904 |
|
|
}
|
7905 |
|
|
|
7906 |
|
|
/* Fold function call to builtin ceil, ceilf or ceill with argument ARG.
|
7907 |
|
|
Return NULL_TREE if no simplification can be made. */
|
7908 |
|
|
|
7909 |
|
|
static tree
|
7910 |
|
|
fold_builtin_ceil (location_t loc, tree fndecl, tree arg)
|
7911 |
|
|
{
|
7912 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
7913 |
|
|
return NULL_TREE;
|
7914 |
|
|
|
7915 |
|
|
/* Optimize ceil of constant value. */
|
7916 |
|
|
if (TREE_CODE (arg) == REAL_CST && !TREE_OVERFLOW (arg))
|
7917 |
|
|
{
|
7918 |
|
|
REAL_VALUE_TYPE x;
|
7919 |
|
|
|
7920 |
|
|
x = TREE_REAL_CST (arg);
|
7921 |
|
|
if (! REAL_VALUE_ISNAN (x) || ! flag_errno_math)
|
7922 |
|
|
{
|
7923 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
7924 |
|
|
REAL_VALUE_TYPE r;
|
7925 |
|
|
|
7926 |
|
|
real_ceil (&r, TYPE_MODE (type), &x);
|
7927 |
|
|
return build_real (type, r);
|
7928 |
|
|
}
|
7929 |
|
|
}
|
7930 |
|
|
|
7931 |
|
|
return fold_trunc_transparent_mathfn (loc, fndecl, arg);
|
7932 |
|
|
}
|
7933 |
|
|
|
7934 |
|
|
/* Fold function call to builtin round, roundf or roundl with argument ARG.
|
7935 |
|
|
Return NULL_TREE if no simplification can be made. */
|
7936 |
|
|
|
7937 |
|
|
static tree
|
7938 |
|
|
fold_builtin_round (location_t loc, tree fndecl, tree arg)
|
7939 |
|
|
{
|
7940 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
7941 |
|
|
return NULL_TREE;
|
7942 |
|
|
|
7943 |
|
|
/* Optimize round of constant value. */
|
7944 |
|
|
if (TREE_CODE (arg) == REAL_CST && !TREE_OVERFLOW (arg))
|
7945 |
|
|
{
|
7946 |
|
|
REAL_VALUE_TYPE x;
|
7947 |
|
|
|
7948 |
|
|
x = TREE_REAL_CST (arg);
|
7949 |
|
|
if (! REAL_VALUE_ISNAN (x) || ! flag_errno_math)
|
7950 |
|
|
{
|
7951 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
7952 |
|
|
REAL_VALUE_TYPE r;
|
7953 |
|
|
|
7954 |
|
|
real_round (&r, TYPE_MODE (type), &x);
|
7955 |
|
|
return build_real (type, r);
|
7956 |
|
|
}
|
7957 |
|
|
}
|
7958 |
|
|
|
7959 |
|
|
return fold_trunc_transparent_mathfn (loc, fndecl, arg);
|
7960 |
|
|
}
|
7961 |
|
|
|
7962 |
|
|
/* Fold function call to builtin lround, lroundf or lroundl (or the
|
7963 |
|
|
corresponding long long versions) and other rounding functions. ARG
|
7964 |
|
|
is the argument to the call. Return NULL_TREE if no simplification
|
7965 |
|
|
can be made. */
|
7966 |
|
|
|
7967 |
|
|
static tree
|
7968 |
|
|
fold_builtin_int_roundingfn (location_t loc, tree fndecl, tree arg)
|
7969 |
|
|
{
|
7970 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
7971 |
|
|
return NULL_TREE;
|
7972 |
|
|
|
7973 |
|
|
/* Optimize lround of constant value. */
|
7974 |
|
|
if (TREE_CODE (arg) == REAL_CST && !TREE_OVERFLOW (arg))
|
7975 |
|
|
{
|
7976 |
|
|
const REAL_VALUE_TYPE x = TREE_REAL_CST (arg);
|
7977 |
|
|
|
7978 |
|
|
if (real_isfinite (&x))
|
7979 |
|
|
{
|
7980 |
|
|
tree itype = TREE_TYPE (TREE_TYPE (fndecl));
|
7981 |
|
|
tree ftype = TREE_TYPE (arg);
|
7982 |
|
|
double_int val;
|
7983 |
|
|
REAL_VALUE_TYPE r;
|
7984 |
|
|
|
7985 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
7986 |
|
|
{
|
7987 |
|
|
CASE_FLT_FN (BUILT_IN_IFLOOR):
|
7988 |
|
|
CASE_FLT_FN (BUILT_IN_LFLOOR):
|
7989 |
|
|
CASE_FLT_FN (BUILT_IN_LLFLOOR):
|
7990 |
|
|
real_floor (&r, TYPE_MODE (ftype), &x);
|
7991 |
|
|
break;
|
7992 |
|
|
|
7993 |
|
|
CASE_FLT_FN (BUILT_IN_ICEIL):
|
7994 |
|
|
CASE_FLT_FN (BUILT_IN_LCEIL):
|
7995 |
|
|
CASE_FLT_FN (BUILT_IN_LLCEIL):
|
7996 |
|
|
real_ceil (&r, TYPE_MODE (ftype), &x);
|
7997 |
|
|
break;
|
7998 |
|
|
|
7999 |
|
|
CASE_FLT_FN (BUILT_IN_IROUND):
|
8000 |
|
|
CASE_FLT_FN (BUILT_IN_LROUND):
|
8001 |
|
|
CASE_FLT_FN (BUILT_IN_LLROUND):
|
8002 |
|
|
real_round (&r, TYPE_MODE (ftype), &x);
|
8003 |
|
|
break;
|
8004 |
|
|
|
8005 |
|
|
default:
|
8006 |
|
|
gcc_unreachable ();
|
8007 |
|
|
}
|
8008 |
|
|
|
8009 |
|
|
real_to_integer2 ((HOST_WIDE_INT *)&val.low, &val.high, &r);
|
8010 |
|
|
if (double_int_fits_to_tree_p (itype, val))
|
8011 |
|
|
return double_int_to_tree (itype, val);
|
8012 |
|
|
}
|
8013 |
|
|
}
|
8014 |
|
|
|
8015 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
8016 |
|
|
{
|
8017 |
|
|
CASE_FLT_FN (BUILT_IN_LFLOOR):
|
8018 |
|
|
CASE_FLT_FN (BUILT_IN_LLFLOOR):
|
8019 |
|
|
/* Fold lfloor (x) where x is nonnegative to FIX_TRUNC (x). */
|
8020 |
|
|
if (tree_expr_nonnegative_p (arg))
|
8021 |
|
|
return fold_build1_loc (loc, FIX_TRUNC_EXPR,
|
8022 |
|
|
TREE_TYPE (TREE_TYPE (fndecl)), arg);
|
8023 |
|
|
break;
|
8024 |
|
|
default:;
|
8025 |
|
|
}
|
8026 |
|
|
|
8027 |
|
|
return fold_fixed_mathfn (loc, fndecl, arg);
|
8028 |
|
|
}
|
8029 |
|
|
|
8030 |
|
|
/* Fold function call to builtin ffs, clz, ctz, popcount and parity
|
8031 |
|
|
and their long and long long variants (i.e. ffsl and ffsll). ARG is
|
8032 |
|
|
the argument to the call. Return NULL_TREE if no simplification can
|
8033 |
|
|
be made. */
|
8034 |
|
|
|
8035 |
|
|
static tree
|
8036 |
|
|
fold_builtin_bitop (tree fndecl, tree arg)
|
8037 |
|
|
{
|
8038 |
|
|
if (!validate_arg (arg, INTEGER_TYPE))
|
8039 |
|
|
return NULL_TREE;
|
8040 |
|
|
|
8041 |
|
|
/* Optimize for constant argument. */
|
8042 |
|
|
if (TREE_CODE (arg) == INTEGER_CST && !TREE_OVERFLOW (arg))
|
8043 |
|
|
{
|
8044 |
|
|
HOST_WIDE_INT hi, width, result;
|
8045 |
|
|
unsigned HOST_WIDE_INT lo;
|
8046 |
|
|
tree type;
|
8047 |
|
|
|
8048 |
|
|
type = TREE_TYPE (arg);
|
8049 |
|
|
width = TYPE_PRECISION (type);
|
8050 |
|
|
lo = TREE_INT_CST_LOW (arg);
|
8051 |
|
|
|
8052 |
|
|
/* Clear all the bits that are beyond the type's precision. */
|
8053 |
|
|
if (width > HOST_BITS_PER_WIDE_INT)
|
8054 |
|
|
{
|
8055 |
|
|
hi = TREE_INT_CST_HIGH (arg);
|
8056 |
|
|
if (width < 2 * HOST_BITS_PER_WIDE_INT)
|
8057 |
|
|
hi &= ~((unsigned HOST_WIDE_INT) (-1)
|
8058 |
|
|
<< (width - HOST_BITS_PER_WIDE_INT));
|
8059 |
|
|
}
|
8060 |
|
|
else
|
8061 |
|
|
{
|
8062 |
|
|
hi = 0;
|
8063 |
|
|
if (width < HOST_BITS_PER_WIDE_INT)
|
8064 |
|
|
lo &= ~((unsigned HOST_WIDE_INT) (-1) << width);
|
8065 |
|
|
}
|
8066 |
|
|
|
8067 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
8068 |
|
|
{
|
8069 |
|
|
CASE_INT_FN (BUILT_IN_FFS):
|
8070 |
|
|
if (lo != 0)
|
8071 |
|
|
result = ffs_hwi (lo);
|
8072 |
|
|
else if (hi != 0)
|
8073 |
|
|
result = HOST_BITS_PER_WIDE_INT + ffs_hwi (hi);
|
8074 |
|
|
else
|
8075 |
|
|
result = 0;
|
8076 |
|
|
break;
|
8077 |
|
|
|
8078 |
|
|
CASE_INT_FN (BUILT_IN_CLZ):
|
8079 |
|
|
if (hi != 0)
|
8080 |
|
|
result = width - floor_log2 (hi) - 1 - HOST_BITS_PER_WIDE_INT;
|
8081 |
|
|
else if (lo != 0)
|
8082 |
|
|
result = width - floor_log2 (lo) - 1;
|
8083 |
|
|
else if (! CLZ_DEFINED_VALUE_AT_ZERO (TYPE_MODE (type), result))
|
8084 |
|
|
result = width;
|
8085 |
|
|
break;
|
8086 |
|
|
|
8087 |
|
|
CASE_INT_FN (BUILT_IN_CTZ):
|
8088 |
|
|
if (lo != 0)
|
8089 |
|
|
result = ctz_hwi (lo);
|
8090 |
|
|
else if (hi != 0)
|
8091 |
|
|
result = HOST_BITS_PER_WIDE_INT + ctz_hwi (hi);
|
8092 |
|
|
else if (! CTZ_DEFINED_VALUE_AT_ZERO (TYPE_MODE (type), result))
|
8093 |
|
|
result = width;
|
8094 |
|
|
break;
|
8095 |
|
|
|
8096 |
|
|
CASE_INT_FN (BUILT_IN_CLRSB):
|
8097 |
|
|
if (width > HOST_BITS_PER_WIDE_INT
|
8098 |
|
|
&& (hi & ((unsigned HOST_WIDE_INT) 1
|
8099 |
|
|
<< (width - HOST_BITS_PER_WIDE_INT - 1))) != 0)
|
8100 |
|
|
{
|
8101 |
|
|
hi = ~hi & ~((unsigned HOST_WIDE_INT) (-1)
|
8102 |
|
|
<< (width - HOST_BITS_PER_WIDE_INT - 1));
|
8103 |
|
|
lo = ~lo;
|
8104 |
|
|
}
|
8105 |
|
|
else if (width <= HOST_BITS_PER_WIDE_INT
|
8106 |
|
|
&& (lo & ((unsigned HOST_WIDE_INT) 1 << (width - 1))) != 0)
|
8107 |
|
|
lo = ~lo & ~((unsigned HOST_WIDE_INT) (-1) << (width - 1));
|
8108 |
|
|
if (hi != 0)
|
8109 |
|
|
result = width - floor_log2 (hi) - 2 - HOST_BITS_PER_WIDE_INT;
|
8110 |
|
|
else if (lo != 0)
|
8111 |
|
|
result = width - floor_log2 (lo) - 2;
|
8112 |
|
|
else
|
8113 |
|
|
result = width - 1;
|
8114 |
|
|
break;
|
8115 |
|
|
|
8116 |
|
|
CASE_INT_FN (BUILT_IN_POPCOUNT):
|
8117 |
|
|
result = 0;
|
8118 |
|
|
while (lo)
|
8119 |
|
|
result++, lo &= lo - 1;
|
8120 |
|
|
while (hi)
|
8121 |
|
|
result++, hi &= (unsigned HOST_WIDE_INT) hi - 1;
|
8122 |
|
|
break;
|
8123 |
|
|
|
8124 |
|
|
CASE_INT_FN (BUILT_IN_PARITY):
|
8125 |
|
|
result = 0;
|
8126 |
|
|
while (lo)
|
8127 |
|
|
result++, lo &= lo - 1;
|
8128 |
|
|
while (hi)
|
8129 |
|
|
result++, hi &= (unsigned HOST_WIDE_INT) hi - 1;
|
8130 |
|
|
result &= 1;
|
8131 |
|
|
break;
|
8132 |
|
|
|
8133 |
|
|
default:
|
8134 |
|
|
gcc_unreachable ();
|
8135 |
|
|
}
|
8136 |
|
|
|
8137 |
|
|
return build_int_cst (TREE_TYPE (TREE_TYPE (fndecl)), result);
|
8138 |
|
|
}
|
8139 |
|
|
|
8140 |
|
|
return NULL_TREE;
|
8141 |
|
|
}
|
8142 |
|
|
|
8143 |
|
|
/* Fold function call to builtin_bswap and the long and long long
|
8144 |
|
|
variants. Return NULL_TREE if no simplification can be made. */
|
8145 |
|
|
static tree
|
8146 |
|
|
fold_builtin_bswap (tree fndecl, tree arg)
|
8147 |
|
|
{
|
8148 |
|
|
if (! validate_arg (arg, INTEGER_TYPE))
|
8149 |
|
|
return NULL_TREE;
|
8150 |
|
|
|
8151 |
|
|
/* Optimize constant value. */
|
8152 |
|
|
if (TREE_CODE (arg) == INTEGER_CST && !TREE_OVERFLOW (arg))
|
8153 |
|
|
{
|
8154 |
|
|
HOST_WIDE_INT hi, width, r_hi = 0;
|
8155 |
|
|
unsigned HOST_WIDE_INT lo, r_lo = 0;
|
8156 |
|
|
tree type;
|
8157 |
|
|
|
8158 |
|
|
type = TREE_TYPE (arg);
|
8159 |
|
|
width = TYPE_PRECISION (type);
|
8160 |
|
|
lo = TREE_INT_CST_LOW (arg);
|
8161 |
|
|
hi = TREE_INT_CST_HIGH (arg);
|
8162 |
|
|
|
8163 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
8164 |
|
|
{
|
8165 |
|
|
case BUILT_IN_BSWAP32:
|
8166 |
|
|
case BUILT_IN_BSWAP64:
|
8167 |
|
|
{
|
8168 |
|
|
int s;
|
8169 |
|
|
|
8170 |
|
|
for (s = 0; s < width; s += 8)
|
8171 |
|
|
{
|
8172 |
|
|
int d = width - s - 8;
|
8173 |
|
|
unsigned HOST_WIDE_INT byte;
|
8174 |
|
|
|
8175 |
|
|
if (s < HOST_BITS_PER_WIDE_INT)
|
8176 |
|
|
byte = (lo >> s) & 0xff;
|
8177 |
|
|
else
|
8178 |
|
|
byte = (hi >> (s - HOST_BITS_PER_WIDE_INT)) & 0xff;
|
8179 |
|
|
|
8180 |
|
|
if (d < HOST_BITS_PER_WIDE_INT)
|
8181 |
|
|
r_lo |= byte << d;
|
8182 |
|
|
else
|
8183 |
|
|
r_hi |= byte << (d - HOST_BITS_PER_WIDE_INT);
|
8184 |
|
|
}
|
8185 |
|
|
}
|
8186 |
|
|
|
8187 |
|
|
break;
|
8188 |
|
|
|
8189 |
|
|
default:
|
8190 |
|
|
gcc_unreachable ();
|
8191 |
|
|
}
|
8192 |
|
|
|
8193 |
|
|
if (width < HOST_BITS_PER_WIDE_INT)
|
8194 |
|
|
return build_int_cst (TREE_TYPE (TREE_TYPE (fndecl)), r_lo);
|
8195 |
|
|
else
|
8196 |
|
|
return build_int_cst_wide (TREE_TYPE (TREE_TYPE (fndecl)), r_lo, r_hi);
|
8197 |
|
|
}
|
8198 |
|
|
|
8199 |
|
|
return NULL_TREE;
|
8200 |
|
|
}
|
8201 |
|
|
|
8202 |
|
|
/* A subroutine of fold_builtin to fold the various logarithmic
|
8203 |
|
|
functions. Return NULL_TREE if no simplification can me made.
|
8204 |
|
|
FUNC is the corresponding MPFR logarithm function. */
|
8205 |
|
|
|
8206 |
|
|
static tree
|
8207 |
|
|
fold_builtin_logarithm (location_t loc, tree fndecl, tree arg,
|
8208 |
|
|
int (*func)(mpfr_ptr, mpfr_srcptr, mp_rnd_t))
|
8209 |
|
|
{
|
8210 |
|
|
if (validate_arg (arg, REAL_TYPE))
|
8211 |
|
|
{
|
8212 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
8213 |
|
|
tree res;
|
8214 |
|
|
const enum built_in_function fcode = builtin_mathfn_code (arg);
|
8215 |
|
|
|
8216 |
|
|
/* Calculate the result when the argument is a constant. */
|
8217 |
|
|
if ((res = do_mpfr_arg1 (arg, type, func, &dconst0, NULL, false)))
|
8218 |
|
|
return res;
|
8219 |
|
|
|
8220 |
|
|
/* Special case, optimize logN(expN(x)) = x. */
|
8221 |
|
|
if (flag_unsafe_math_optimizations
|
8222 |
|
|
&& ((func == mpfr_log
|
8223 |
|
|
&& (fcode == BUILT_IN_EXP
|
8224 |
|
|
|| fcode == BUILT_IN_EXPF
|
8225 |
|
|
|| fcode == BUILT_IN_EXPL))
|
8226 |
|
|
|| (func == mpfr_log2
|
8227 |
|
|
&& (fcode == BUILT_IN_EXP2
|
8228 |
|
|
|| fcode == BUILT_IN_EXP2F
|
8229 |
|
|
|| fcode == BUILT_IN_EXP2L))
|
8230 |
|
|
|| (func == mpfr_log10 && (BUILTIN_EXP10_P (fcode)))))
|
8231 |
|
|
return fold_convert_loc (loc, type, CALL_EXPR_ARG (arg, 0));
|
8232 |
|
|
|
8233 |
|
|
/* Optimize logN(func()) for various exponential functions. We
|
8234 |
|
|
want to determine the value "x" and the power "exponent" in
|
8235 |
|
|
order to transform logN(x**exponent) into exponent*logN(x). */
|
8236 |
|
|
if (flag_unsafe_math_optimizations)
|
8237 |
|
|
{
|
8238 |
|
|
tree exponent = 0, x = 0;
|
8239 |
|
|
|
8240 |
|
|
switch (fcode)
|
8241 |
|
|
{
|
8242 |
|
|
CASE_FLT_FN (BUILT_IN_EXP):
|
8243 |
|
|
/* Prepare to do logN(exp(exponent) -> exponent*logN(e). */
|
8244 |
|
|
x = build_real (type, real_value_truncate (TYPE_MODE (type),
|
8245 |
|
|
dconst_e ()));
|
8246 |
|
|
exponent = CALL_EXPR_ARG (arg, 0);
|
8247 |
|
|
break;
|
8248 |
|
|
CASE_FLT_FN (BUILT_IN_EXP2):
|
8249 |
|
|
/* Prepare to do logN(exp2(exponent) -> exponent*logN(2). */
|
8250 |
|
|
x = build_real (type, dconst2);
|
8251 |
|
|
exponent = CALL_EXPR_ARG (arg, 0);
|
8252 |
|
|
break;
|
8253 |
|
|
CASE_FLT_FN (BUILT_IN_EXP10):
|
8254 |
|
|
CASE_FLT_FN (BUILT_IN_POW10):
|
8255 |
|
|
/* Prepare to do logN(exp10(exponent) -> exponent*logN(10). */
|
8256 |
|
|
{
|
8257 |
|
|
REAL_VALUE_TYPE dconst10;
|
8258 |
|
|
real_from_integer (&dconst10, VOIDmode, 10, 0, 0);
|
8259 |
|
|
x = build_real (type, dconst10);
|
8260 |
|
|
}
|
8261 |
|
|
exponent = CALL_EXPR_ARG (arg, 0);
|
8262 |
|
|
break;
|
8263 |
|
|
CASE_FLT_FN (BUILT_IN_SQRT):
|
8264 |
|
|
/* Prepare to do logN(sqrt(x) -> 0.5*logN(x). */
|
8265 |
|
|
x = CALL_EXPR_ARG (arg, 0);
|
8266 |
|
|
exponent = build_real (type, dconsthalf);
|
8267 |
|
|
break;
|
8268 |
|
|
CASE_FLT_FN (BUILT_IN_CBRT):
|
8269 |
|
|
/* Prepare to do logN(cbrt(x) -> (1/3)*logN(x). */
|
8270 |
|
|
x = CALL_EXPR_ARG (arg, 0);
|
8271 |
|
|
exponent = build_real (type, real_value_truncate (TYPE_MODE (type),
|
8272 |
|
|
dconst_third ()));
|
8273 |
|
|
break;
|
8274 |
|
|
CASE_FLT_FN (BUILT_IN_POW):
|
8275 |
|
|
/* Prepare to do logN(pow(x,exponent) -> exponent*logN(x). */
|
8276 |
|
|
x = CALL_EXPR_ARG (arg, 0);
|
8277 |
|
|
exponent = CALL_EXPR_ARG (arg, 1);
|
8278 |
|
|
break;
|
8279 |
|
|
default:
|
8280 |
|
|
break;
|
8281 |
|
|
}
|
8282 |
|
|
|
8283 |
|
|
/* Now perform the optimization. */
|
8284 |
|
|
if (x && exponent)
|
8285 |
|
|
{
|
8286 |
|
|
tree logfn = build_call_expr_loc (loc, fndecl, 1, x);
|
8287 |
|
|
return fold_build2_loc (loc, MULT_EXPR, type, exponent, logfn);
|
8288 |
|
|
}
|
8289 |
|
|
}
|
8290 |
|
|
}
|
8291 |
|
|
|
8292 |
|
|
return NULL_TREE;
|
8293 |
|
|
}
|
8294 |
|
|
|
8295 |
|
|
/* Fold a builtin function call to hypot, hypotf, or hypotl. Return
|
8296 |
|
|
NULL_TREE if no simplification can be made. */
|
8297 |
|
|
|
8298 |
|
|
static tree
|
8299 |
|
|
fold_builtin_hypot (location_t loc, tree fndecl,
|
8300 |
|
|
tree arg0, tree arg1, tree type)
|
8301 |
|
|
{
|
8302 |
|
|
tree res, narg0, narg1;
|
8303 |
|
|
|
8304 |
|
|
if (!validate_arg (arg0, REAL_TYPE)
|
8305 |
|
|
|| !validate_arg (arg1, REAL_TYPE))
|
8306 |
|
|
return NULL_TREE;
|
8307 |
|
|
|
8308 |
|
|
/* Calculate the result when the argument is a constant. */
|
8309 |
|
|
if ((res = do_mpfr_arg2 (arg0, arg1, type, mpfr_hypot)))
|
8310 |
|
|
return res;
|
8311 |
|
|
|
8312 |
|
|
/* If either argument to hypot has a negate or abs, strip that off.
|
8313 |
|
|
E.g. hypot(-x,fabs(y)) -> hypot(x,y). */
|
8314 |
|
|
narg0 = fold_strip_sign_ops (arg0);
|
8315 |
|
|
narg1 = fold_strip_sign_ops (arg1);
|
8316 |
|
|
if (narg0 || narg1)
|
8317 |
|
|
{
|
8318 |
|
|
return build_call_expr_loc (loc, fndecl, 2, narg0 ? narg0 : arg0,
|
8319 |
|
|
narg1 ? narg1 : arg1);
|
8320 |
|
|
}
|
8321 |
|
|
|
8322 |
|
|
/* If either argument is zero, hypot is fabs of the other. */
|
8323 |
|
|
if (real_zerop (arg0))
|
8324 |
|
|
return fold_build1_loc (loc, ABS_EXPR, type, arg1);
|
8325 |
|
|
else if (real_zerop (arg1))
|
8326 |
|
|
return fold_build1_loc (loc, ABS_EXPR, type, arg0);
|
8327 |
|
|
|
8328 |
|
|
/* hypot(x,x) -> fabs(x)*sqrt(2). */
|
8329 |
|
|
if (flag_unsafe_math_optimizations
|
8330 |
|
|
&& operand_equal_p (arg0, arg1, OEP_PURE_SAME))
|
8331 |
|
|
{
|
8332 |
|
|
const REAL_VALUE_TYPE sqrt2_trunc
|
8333 |
|
|
= real_value_truncate (TYPE_MODE (type), dconst_sqrt2 ());
|
8334 |
|
|
return fold_build2_loc (loc, MULT_EXPR, type,
|
8335 |
|
|
fold_build1_loc (loc, ABS_EXPR, type, arg0),
|
8336 |
|
|
build_real (type, sqrt2_trunc));
|
8337 |
|
|
}
|
8338 |
|
|
|
8339 |
|
|
return NULL_TREE;
|
8340 |
|
|
}
|
8341 |
|
|
|
8342 |
|
|
|
8343 |
|
|
/* Fold a builtin function call to pow, powf, or powl. Return
|
8344 |
|
|
NULL_TREE if no simplification can be made. */
|
8345 |
|
|
static tree
|
8346 |
|
|
fold_builtin_pow (location_t loc, tree fndecl, tree arg0, tree arg1, tree type)
|
8347 |
|
|
{
|
8348 |
|
|
tree res;
|
8349 |
|
|
|
8350 |
|
|
if (!validate_arg (arg0, REAL_TYPE)
|
8351 |
|
|
|| !validate_arg (arg1, REAL_TYPE))
|
8352 |
|
|
return NULL_TREE;
|
8353 |
|
|
|
8354 |
|
|
/* Calculate the result when the argument is a constant. */
|
8355 |
|
|
if ((res = do_mpfr_arg2 (arg0, arg1, type, mpfr_pow)))
|
8356 |
|
|
return res;
|
8357 |
|
|
|
8358 |
|
|
/* Optimize pow(1.0,y) = 1.0. */
|
8359 |
|
|
if (real_onep (arg0))
|
8360 |
|
|
return omit_one_operand_loc (loc, type, build_real (type, dconst1), arg1);
|
8361 |
|
|
|
8362 |
|
|
if (TREE_CODE (arg1) == REAL_CST
|
8363 |
|
|
&& !TREE_OVERFLOW (arg1))
|
8364 |
|
|
{
|
8365 |
|
|
REAL_VALUE_TYPE cint;
|
8366 |
|
|
REAL_VALUE_TYPE c;
|
8367 |
|
|
HOST_WIDE_INT n;
|
8368 |
|
|
|
8369 |
|
|
c = TREE_REAL_CST (arg1);
|
8370 |
|
|
|
8371 |
|
|
/* Optimize pow(x,0.0) = 1.0. */
|
8372 |
|
|
if (REAL_VALUES_EQUAL (c, dconst0))
|
8373 |
|
|
return omit_one_operand_loc (loc, type, build_real (type, dconst1),
|
8374 |
|
|
arg0);
|
8375 |
|
|
|
8376 |
|
|
/* Optimize pow(x,1.0) = x. */
|
8377 |
|
|
if (REAL_VALUES_EQUAL (c, dconst1))
|
8378 |
|
|
return arg0;
|
8379 |
|
|
|
8380 |
|
|
/* Optimize pow(x,-1.0) = 1.0/x. */
|
8381 |
|
|
if (REAL_VALUES_EQUAL (c, dconstm1))
|
8382 |
|
|
return fold_build2_loc (loc, RDIV_EXPR, type,
|
8383 |
|
|
build_real (type, dconst1), arg0);
|
8384 |
|
|
|
8385 |
|
|
/* Optimize pow(x,0.5) = sqrt(x). */
|
8386 |
|
|
if (flag_unsafe_math_optimizations
|
8387 |
|
|
&& REAL_VALUES_EQUAL (c, dconsthalf))
|
8388 |
|
|
{
|
8389 |
|
|
tree sqrtfn = mathfn_built_in (type, BUILT_IN_SQRT);
|
8390 |
|
|
|
8391 |
|
|
if (sqrtfn != NULL_TREE)
|
8392 |
|
|
return build_call_expr_loc (loc, sqrtfn, 1, arg0);
|
8393 |
|
|
}
|
8394 |
|
|
|
8395 |
|
|
/* Optimize pow(x,1.0/3.0) = cbrt(x). */
|
8396 |
|
|
if (flag_unsafe_math_optimizations)
|
8397 |
|
|
{
|
8398 |
|
|
const REAL_VALUE_TYPE dconstroot
|
8399 |
|
|
= real_value_truncate (TYPE_MODE (type), dconst_third ());
|
8400 |
|
|
|
8401 |
|
|
if (REAL_VALUES_EQUAL (c, dconstroot))
|
8402 |
|
|
{
|
8403 |
|
|
tree cbrtfn = mathfn_built_in (type, BUILT_IN_CBRT);
|
8404 |
|
|
if (cbrtfn != NULL_TREE)
|
8405 |
|
|
return build_call_expr_loc (loc, cbrtfn, 1, arg0);
|
8406 |
|
|
}
|
8407 |
|
|
}
|
8408 |
|
|
|
8409 |
|
|
/* Check for an integer exponent. */
|
8410 |
|
|
n = real_to_integer (&c);
|
8411 |
|
|
real_from_integer (&cint, VOIDmode, n, n < 0 ? -1 : 0, 0);
|
8412 |
|
|
if (real_identical (&c, &cint))
|
8413 |
|
|
{
|
8414 |
|
|
/* Attempt to evaluate pow at compile-time, unless this should
|
8415 |
|
|
raise an exception. */
|
8416 |
|
|
if (TREE_CODE (arg0) == REAL_CST
|
8417 |
|
|
&& !TREE_OVERFLOW (arg0)
|
8418 |
|
|
&& (n > 0
|
8419 |
|
|
|| (!flag_trapping_math && !flag_errno_math)
|
8420 |
|
|
|| !REAL_VALUES_EQUAL (TREE_REAL_CST (arg0), dconst0)))
|
8421 |
|
|
{
|
8422 |
|
|
REAL_VALUE_TYPE x;
|
8423 |
|
|
bool inexact;
|
8424 |
|
|
|
8425 |
|
|
x = TREE_REAL_CST (arg0);
|
8426 |
|
|
inexact = real_powi (&x, TYPE_MODE (type), &x, n);
|
8427 |
|
|
if (flag_unsafe_math_optimizations || !inexact)
|
8428 |
|
|
return build_real (type, x);
|
8429 |
|
|
}
|
8430 |
|
|
|
8431 |
|
|
/* Strip sign ops from even integer powers. */
|
8432 |
|
|
if ((n & 1) == 0 && flag_unsafe_math_optimizations)
|
8433 |
|
|
{
|
8434 |
|
|
tree narg0 = fold_strip_sign_ops (arg0);
|
8435 |
|
|
if (narg0)
|
8436 |
|
|
return build_call_expr_loc (loc, fndecl, 2, narg0, arg1);
|
8437 |
|
|
}
|
8438 |
|
|
}
|
8439 |
|
|
}
|
8440 |
|
|
|
8441 |
|
|
if (flag_unsafe_math_optimizations)
|
8442 |
|
|
{
|
8443 |
|
|
const enum built_in_function fcode = builtin_mathfn_code (arg0);
|
8444 |
|
|
|
8445 |
|
|
/* Optimize pow(expN(x),y) = expN(x*y). */
|
8446 |
|
|
if (BUILTIN_EXPONENT_P (fcode))
|
8447 |
|
|
{
|
8448 |
|
|
tree expfn = TREE_OPERAND (CALL_EXPR_FN (arg0), 0);
|
8449 |
|
|
tree arg = CALL_EXPR_ARG (arg0, 0);
|
8450 |
|
|
arg = fold_build2_loc (loc, MULT_EXPR, type, arg, arg1);
|
8451 |
|
|
return build_call_expr_loc (loc, expfn, 1, arg);
|
8452 |
|
|
}
|
8453 |
|
|
|
8454 |
|
|
/* Optimize pow(sqrt(x),y) = pow(x,y*0.5). */
|
8455 |
|
|
if (BUILTIN_SQRT_P (fcode))
|
8456 |
|
|
{
|
8457 |
|
|
tree narg0 = CALL_EXPR_ARG (arg0, 0);
|
8458 |
|
|
tree narg1 = fold_build2_loc (loc, MULT_EXPR, type, arg1,
|
8459 |
|
|
build_real (type, dconsthalf));
|
8460 |
|
|
return build_call_expr_loc (loc, fndecl, 2, narg0, narg1);
|
8461 |
|
|
}
|
8462 |
|
|
|
8463 |
|
|
/* Optimize pow(cbrt(x),y) = pow(x,y/3) iff x is nonnegative. */
|
8464 |
|
|
if (BUILTIN_CBRT_P (fcode))
|
8465 |
|
|
{
|
8466 |
|
|
tree arg = CALL_EXPR_ARG (arg0, 0);
|
8467 |
|
|
if (tree_expr_nonnegative_p (arg))
|
8468 |
|
|
{
|
8469 |
|
|
const REAL_VALUE_TYPE dconstroot
|
8470 |
|
|
= real_value_truncate (TYPE_MODE (type), dconst_third ());
|
8471 |
|
|
tree narg1 = fold_build2_loc (loc, MULT_EXPR, type, arg1,
|
8472 |
|
|
build_real (type, dconstroot));
|
8473 |
|
|
return build_call_expr_loc (loc, fndecl, 2, arg, narg1);
|
8474 |
|
|
}
|
8475 |
|
|
}
|
8476 |
|
|
|
8477 |
|
|
/* Optimize pow(pow(x,y),z) = pow(x,y*z) iff x is nonnegative. */
|
8478 |
|
|
if (fcode == BUILT_IN_POW
|
8479 |
|
|
|| fcode == BUILT_IN_POWF
|
8480 |
|
|
|| fcode == BUILT_IN_POWL)
|
8481 |
|
|
{
|
8482 |
|
|
tree arg00 = CALL_EXPR_ARG (arg0, 0);
|
8483 |
|
|
if (tree_expr_nonnegative_p (arg00))
|
8484 |
|
|
{
|
8485 |
|
|
tree arg01 = CALL_EXPR_ARG (arg0, 1);
|
8486 |
|
|
tree narg1 = fold_build2_loc (loc, MULT_EXPR, type, arg01, arg1);
|
8487 |
|
|
return build_call_expr_loc (loc, fndecl, 2, arg00, narg1);
|
8488 |
|
|
}
|
8489 |
|
|
}
|
8490 |
|
|
}
|
8491 |
|
|
|
8492 |
|
|
return NULL_TREE;
|
8493 |
|
|
}
|
8494 |
|
|
|
8495 |
|
|
/* Fold a builtin function call to powi, powif, or powil with argument ARG.
|
8496 |
|
|
Return NULL_TREE if no simplification can be made. */
|
8497 |
|
|
static tree
|
8498 |
|
|
fold_builtin_powi (location_t loc, tree fndecl ATTRIBUTE_UNUSED,
|
8499 |
|
|
tree arg0, tree arg1, tree type)
|
8500 |
|
|
{
|
8501 |
|
|
if (!validate_arg (arg0, REAL_TYPE)
|
8502 |
|
|
|| !validate_arg (arg1, INTEGER_TYPE))
|
8503 |
|
|
return NULL_TREE;
|
8504 |
|
|
|
8505 |
|
|
/* Optimize pow(1.0,y) = 1.0. */
|
8506 |
|
|
if (real_onep (arg0))
|
8507 |
|
|
return omit_one_operand_loc (loc, type, build_real (type, dconst1), arg1);
|
8508 |
|
|
|
8509 |
|
|
if (host_integerp (arg1, 0))
|
8510 |
|
|
{
|
8511 |
|
|
HOST_WIDE_INT c = TREE_INT_CST_LOW (arg1);
|
8512 |
|
|
|
8513 |
|
|
/* Evaluate powi at compile-time. */
|
8514 |
|
|
if (TREE_CODE (arg0) == REAL_CST
|
8515 |
|
|
&& !TREE_OVERFLOW (arg0))
|
8516 |
|
|
{
|
8517 |
|
|
REAL_VALUE_TYPE x;
|
8518 |
|
|
x = TREE_REAL_CST (arg0);
|
8519 |
|
|
real_powi (&x, TYPE_MODE (type), &x, c);
|
8520 |
|
|
return build_real (type, x);
|
8521 |
|
|
}
|
8522 |
|
|
|
8523 |
|
|
/* Optimize pow(x,0) = 1.0. */
|
8524 |
|
|
if (c == 0)
|
8525 |
|
|
return omit_one_operand_loc (loc, type, build_real (type, dconst1),
|
8526 |
|
|
arg0);
|
8527 |
|
|
|
8528 |
|
|
/* Optimize pow(x,1) = x. */
|
8529 |
|
|
if (c == 1)
|
8530 |
|
|
return arg0;
|
8531 |
|
|
|
8532 |
|
|
/* Optimize pow(x,-1) = 1.0/x. */
|
8533 |
|
|
if (c == -1)
|
8534 |
|
|
return fold_build2_loc (loc, RDIV_EXPR, type,
|
8535 |
|
|
build_real (type, dconst1), arg0);
|
8536 |
|
|
}
|
8537 |
|
|
|
8538 |
|
|
return NULL_TREE;
|
8539 |
|
|
}
|
8540 |
|
|
|
8541 |
|
|
/* A subroutine of fold_builtin to fold the various exponent
|
8542 |
|
|
functions. Return NULL_TREE if no simplification can be made.
|
8543 |
|
|
FUNC is the corresponding MPFR exponent function. */
|
8544 |
|
|
|
8545 |
|
|
static tree
|
8546 |
|
|
fold_builtin_exponent (location_t loc, tree fndecl, tree arg,
|
8547 |
|
|
int (*func)(mpfr_ptr, mpfr_srcptr, mp_rnd_t))
|
8548 |
|
|
{
|
8549 |
|
|
if (validate_arg (arg, REAL_TYPE))
|
8550 |
|
|
{
|
8551 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
8552 |
|
|
tree res;
|
8553 |
|
|
|
8554 |
|
|
/* Calculate the result when the argument is a constant. */
|
8555 |
|
|
if ((res = do_mpfr_arg1 (arg, type, func, NULL, NULL, 0)))
|
8556 |
|
|
return res;
|
8557 |
|
|
|
8558 |
|
|
/* Optimize expN(logN(x)) = x. */
|
8559 |
|
|
if (flag_unsafe_math_optimizations)
|
8560 |
|
|
{
|
8561 |
|
|
const enum built_in_function fcode = builtin_mathfn_code (arg);
|
8562 |
|
|
|
8563 |
|
|
if ((func == mpfr_exp
|
8564 |
|
|
&& (fcode == BUILT_IN_LOG
|
8565 |
|
|
|| fcode == BUILT_IN_LOGF
|
8566 |
|
|
|| fcode == BUILT_IN_LOGL))
|
8567 |
|
|
|| (func == mpfr_exp2
|
8568 |
|
|
&& (fcode == BUILT_IN_LOG2
|
8569 |
|
|
|| fcode == BUILT_IN_LOG2F
|
8570 |
|
|
|| fcode == BUILT_IN_LOG2L))
|
8571 |
|
|
|| (func == mpfr_exp10
|
8572 |
|
|
&& (fcode == BUILT_IN_LOG10
|
8573 |
|
|
|| fcode == BUILT_IN_LOG10F
|
8574 |
|
|
|| fcode == BUILT_IN_LOG10L)))
|
8575 |
|
|
return fold_convert_loc (loc, type, CALL_EXPR_ARG (arg, 0));
|
8576 |
|
|
}
|
8577 |
|
|
}
|
8578 |
|
|
|
8579 |
|
|
return NULL_TREE;
|
8580 |
|
|
}
|
8581 |
|
|
|
8582 |
|
|
/* Return true if VAR is a VAR_DECL or a component thereof. */
|
8583 |
|
|
|
8584 |
|
|
static bool
|
8585 |
|
|
var_decl_component_p (tree var)
|
8586 |
|
|
{
|
8587 |
|
|
tree inner = var;
|
8588 |
|
|
while (handled_component_p (inner))
|
8589 |
|
|
inner = TREE_OPERAND (inner, 0);
|
8590 |
|
|
return SSA_VAR_P (inner);
|
8591 |
|
|
}
|
8592 |
|
|
|
8593 |
|
|
/* Fold function call to builtin memset. Return
|
8594 |
|
|
NULL_TREE if no simplification can be made. */
|
8595 |
|
|
|
8596 |
|
|
static tree
|
8597 |
|
|
fold_builtin_memset (location_t loc, tree dest, tree c, tree len,
|
8598 |
|
|
tree type, bool ignore)
|
8599 |
|
|
{
|
8600 |
|
|
tree var, ret, etype;
|
8601 |
|
|
unsigned HOST_WIDE_INT length, cval;
|
8602 |
|
|
|
8603 |
|
|
if (! validate_arg (dest, POINTER_TYPE)
|
8604 |
|
|
|| ! validate_arg (c, INTEGER_TYPE)
|
8605 |
|
|
|| ! validate_arg (len, INTEGER_TYPE))
|
8606 |
|
|
return NULL_TREE;
|
8607 |
|
|
|
8608 |
|
|
if (! host_integerp (len, 1))
|
8609 |
|
|
return NULL_TREE;
|
8610 |
|
|
|
8611 |
|
|
/* If the LEN parameter is zero, return DEST. */
|
8612 |
|
|
if (integer_zerop (len))
|
8613 |
|
|
return omit_one_operand_loc (loc, type, dest, c);
|
8614 |
|
|
|
8615 |
|
|
if (TREE_CODE (c) != INTEGER_CST || TREE_SIDE_EFFECTS (dest))
|
8616 |
|
|
return NULL_TREE;
|
8617 |
|
|
|
8618 |
|
|
var = dest;
|
8619 |
|
|
STRIP_NOPS (var);
|
8620 |
|
|
if (TREE_CODE (var) != ADDR_EXPR)
|
8621 |
|
|
return NULL_TREE;
|
8622 |
|
|
|
8623 |
|
|
var = TREE_OPERAND (var, 0);
|
8624 |
|
|
if (TREE_THIS_VOLATILE (var))
|
8625 |
|
|
return NULL_TREE;
|
8626 |
|
|
|
8627 |
|
|
etype = TREE_TYPE (var);
|
8628 |
|
|
if (TREE_CODE (etype) == ARRAY_TYPE)
|
8629 |
|
|
etype = TREE_TYPE (etype);
|
8630 |
|
|
|
8631 |
|
|
if (!INTEGRAL_TYPE_P (etype)
|
8632 |
|
|
&& !POINTER_TYPE_P (etype))
|
8633 |
|
|
return NULL_TREE;
|
8634 |
|
|
|
8635 |
|
|
if (! var_decl_component_p (var))
|
8636 |
|
|
return NULL_TREE;
|
8637 |
|
|
|
8638 |
|
|
length = tree_low_cst (len, 1);
|
8639 |
|
|
if (GET_MODE_SIZE (TYPE_MODE (etype)) != length
|
8640 |
|
|
|| get_pointer_alignment (dest) / BITS_PER_UNIT < length)
|
8641 |
|
|
return NULL_TREE;
|
8642 |
|
|
|
8643 |
|
|
if (length > HOST_BITS_PER_WIDE_INT / BITS_PER_UNIT)
|
8644 |
|
|
return NULL_TREE;
|
8645 |
|
|
|
8646 |
|
|
if (integer_zerop (c))
|
8647 |
|
|
cval = 0;
|
8648 |
|
|
else
|
8649 |
|
|
{
|
8650 |
|
|
if (CHAR_BIT != 8 || BITS_PER_UNIT != 8 || HOST_BITS_PER_WIDE_INT > 64)
|
8651 |
|
|
return NULL_TREE;
|
8652 |
|
|
|
8653 |
|
|
cval = TREE_INT_CST_LOW (c);
|
8654 |
|
|
cval &= 0xff;
|
8655 |
|
|
cval |= cval << 8;
|
8656 |
|
|
cval |= cval << 16;
|
8657 |
|
|
cval |= (cval << 31) << 1;
|
8658 |
|
|
}
|
8659 |
|
|
|
8660 |
|
|
ret = build_int_cst_type (etype, cval);
|
8661 |
|
|
var = build_fold_indirect_ref_loc (loc,
|
8662 |
|
|
fold_convert_loc (loc,
|
8663 |
|
|
build_pointer_type (etype),
|
8664 |
|
|
dest));
|
8665 |
|
|
ret = build2 (MODIFY_EXPR, etype, var, ret);
|
8666 |
|
|
if (ignore)
|
8667 |
|
|
return ret;
|
8668 |
|
|
|
8669 |
|
|
return omit_one_operand_loc (loc, type, dest, ret);
|
8670 |
|
|
}
|
8671 |
|
|
|
8672 |
|
|
/* Fold function call to builtin memset. Return
|
8673 |
|
|
NULL_TREE if no simplification can be made. */
|
8674 |
|
|
|
8675 |
|
|
static tree
|
8676 |
|
|
fold_builtin_bzero (location_t loc, tree dest, tree size, bool ignore)
|
8677 |
|
|
{
|
8678 |
|
|
if (! validate_arg (dest, POINTER_TYPE)
|
8679 |
|
|
|| ! validate_arg (size, INTEGER_TYPE))
|
8680 |
|
|
return NULL_TREE;
|
8681 |
|
|
|
8682 |
|
|
if (!ignore)
|
8683 |
|
|
return NULL_TREE;
|
8684 |
|
|
|
8685 |
|
|
/* New argument list transforming bzero(ptr x, int y) to
|
8686 |
|
|
memset(ptr x, int 0, size_t y). This is done this way
|
8687 |
|
|
so that if it isn't expanded inline, we fallback to
|
8688 |
|
|
calling bzero instead of memset. */
|
8689 |
|
|
|
8690 |
|
|
return fold_builtin_memset (loc, dest, integer_zero_node,
|
8691 |
|
|
fold_convert_loc (loc, size_type_node, size),
|
8692 |
|
|
void_type_node, ignore);
|
8693 |
|
|
}
|
8694 |
|
|
|
8695 |
|
|
/* Fold function call to builtin mem{{,p}cpy,move}. Return
|
8696 |
|
|
NULL_TREE if no simplification can be made.
|
8697 |
|
|
If ENDP is 0, return DEST (like memcpy).
|
8698 |
|
|
If ENDP is 1, return DEST+LEN (like mempcpy).
|
8699 |
|
|
If ENDP is 2, return DEST+LEN-1 (like stpcpy).
|
8700 |
|
|
If ENDP is 3, return DEST, additionally *SRC and *DEST may overlap
|
8701 |
|
|
(memmove). */
|
8702 |
|
|
|
8703 |
|
|
static tree
|
8704 |
|
|
fold_builtin_memory_op (location_t loc, tree dest, tree src,
|
8705 |
|
|
tree len, tree type, bool ignore, int endp)
|
8706 |
|
|
{
|
8707 |
|
|
tree destvar, srcvar, expr;
|
8708 |
|
|
|
8709 |
|
|
if (! validate_arg (dest, POINTER_TYPE)
|
8710 |
|
|
|| ! validate_arg (src, POINTER_TYPE)
|
8711 |
|
|
|| ! validate_arg (len, INTEGER_TYPE))
|
8712 |
|
|
return NULL_TREE;
|
8713 |
|
|
|
8714 |
|
|
/* If the LEN parameter is zero, return DEST. */
|
8715 |
|
|
if (integer_zerop (len))
|
8716 |
|
|
return omit_one_operand_loc (loc, type, dest, src);
|
8717 |
|
|
|
8718 |
|
|
/* If SRC and DEST are the same (and not volatile), return
|
8719 |
|
|
DEST{,+LEN,+LEN-1}. */
|
8720 |
|
|
if (operand_equal_p (src, dest, 0))
|
8721 |
|
|
expr = len;
|
8722 |
|
|
else
|
8723 |
|
|
{
|
8724 |
|
|
tree srctype, desttype;
|
8725 |
|
|
unsigned int src_align, dest_align;
|
8726 |
|
|
tree off0;
|
8727 |
|
|
|
8728 |
|
|
if (endp == 3)
|
8729 |
|
|
{
|
8730 |
|
|
src_align = get_pointer_alignment (src);
|
8731 |
|
|
dest_align = get_pointer_alignment (dest);
|
8732 |
|
|
|
8733 |
|
|
/* Both DEST and SRC must be pointer types.
|
8734 |
|
|
??? This is what old code did. Is the testing for pointer types
|
8735 |
|
|
really mandatory?
|
8736 |
|
|
|
8737 |
|
|
If either SRC is readonly or length is 1, we can use memcpy. */
|
8738 |
|
|
if (!dest_align || !src_align)
|
8739 |
|
|
return NULL_TREE;
|
8740 |
|
|
if (readonly_data_expr (src)
|
8741 |
|
|
|| (host_integerp (len, 1)
|
8742 |
|
|
&& (MIN (src_align, dest_align) / BITS_PER_UNIT
|
8743 |
|
|
>= (unsigned HOST_WIDE_INT) tree_low_cst (len, 1))))
|
8744 |
|
|
{
|
8745 |
|
|
tree fn = builtin_decl_implicit (BUILT_IN_MEMCPY);
|
8746 |
|
|
if (!fn)
|
8747 |
|
|
return NULL_TREE;
|
8748 |
|
|
return build_call_expr_loc (loc, fn, 3, dest, src, len);
|
8749 |
|
|
}
|
8750 |
|
|
|
8751 |
|
|
/* If *src and *dest can't overlap, optimize into memcpy as well. */
|
8752 |
|
|
if (TREE_CODE (src) == ADDR_EXPR
|
8753 |
|
|
&& TREE_CODE (dest) == ADDR_EXPR)
|
8754 |
|
|
{
|
8755 |
|
|
tree src_base, dest_base, fn;
|
8756 |
|
|
HOST_WIDE_INT src_offset = 0, dest_offset = 0;
|
8757 |
|
|
HOST_WIDE_INT size = -1;
|
8758 |
|
|
HOST_WIDE_INT maxsize = -1;
|
8759 |
|
|
|
8760 |
|
|
srcvar = TREE_OPERAND (src, 0);
|
8761 |
|
|
src_base = get_ref_base_and_extent (srcvar, &src_offset,
|
8762 |
|
|
&size, &maxsize);
|
8763 |
|
|
destvar = TREE_OPERAND (dest, 0);
|
8764 |
|
|
dest_base = get_ref_base_and_extent (destvar, &dest_offset,
|
8765 |
|
|
&size, &maxsize);
|
8766 |
|
|
if (host_integerp (len, 1))
|
8767 |
|
|
maxsize = tree_low_cst (len, 1);
|
8768 |
|
|
else
|
8769 |
|
|
maxsize = -1;
|
8770 |
|
|
src_offset /= BITS_PER_UNIT;
|
8771 |
|
|
dest_offset /= BITS_PER_UNIT;
|
8772 |
|
|
if (SSA_VAR_P (src_base)
|
8773 |
|
|
&& SSA_VAR_P (dest_base))
|
8774 |
|
|
{
|
8775 |
|
|
if (operand_equal_p (src_base, dest_base, 0)
|
8776 |
|
|
&& ranges_overlap_p (src_offset, maxsize,
|
8777 |
|
|
dest_offset, maxsize))
|
8778 |
|
|
return NULL_TREE;
|
8779 |
|
|
}
|
8780 |
|
|
else if (TREE_CODE (src_base) == MEM_REF
|
8781 |
|
|
&& TREE_CODE (dest_base) == MEM_REF)
|
8782 |
|
|
{
|
8783 |
|
|
double_int off;
|
8784 |
|
|
if (! operand_equal_p (TREE_OPERAND (src_base, 0),
|
8785 |
|
|
TREE_OPERAND (dest_base, 0), 0))
|
8786 |
|
|
return NULL_TREE;
|
8787 |
|
|
off = double_int_add (mem_ref_offset (src_base),
|
8788 |
|
|
shwi_to_double_int (src_offset));
|
8789 |
|
|
if (!double_int_fits_in_shwi_p (off))
|
8790 |
|
|
return NULL_TREE;
|
8791 |
|
|
src_offset = off.low;
|
8792 |
|
|
off = double_int_add (mem_ref_offset (dest_base),
|
8793 |
|
|
shwi_to_double_int (dest_offset));
|
8794 |
|
|
if (!double_int_fits_in_shwi_p (off))
|
8795 |
|
|
return NULL_TREE;
|
8796 |
|
|
dest_offset = off.low;
|
8797 |
|
|
if (ranges_overlap_p (src_offset, maxsize,
|
8798 |
|
|
dest_offset, maxsize))
|
8799 |
|
|
return NULL_TREE;
|
8800 |
|
|
}
|
8801 |
|
|
else
|
8802 |
|
|
return NULL_TREE;
|
8803 |
|
|
|
8804 |
|
|
fn = builtin_decl_implicit (BUILT_IN_MEMCPY);
|
8805 |
|
|
if (!fn)
|
8806 |
|
|
return NULL_TREE;
|
8807 |
|
|
return build_call_expr_loc (loc, fn, 3, dest, src, len);
|
8808 |
|
|
}
|
8809 |
|
|
|
8810 |
|
|
/* If the destination and source do not alias optimize into
|
8811 |
|
|
memcpy as well. */
|
8812 |
|
|
if ((is_gimple_min_invariant (dest)
|
8813 |
|
|
|| TREE_CODE (dest) == SSA_NAME)
|
8814 |
|
|
&& (is_gimple_min_invariant (src)
|
8815 |
|
|
|| TREE_CODE (src) == SSA_NAME))
|
8816 |
|
|
{
|
8817 |
|
|
ao_ref destr, srcr;
|
8818 |
|
|
ao_ref_init_from_ptr_and_size (&destr, dest, len);
|
8819 |
|
|
ao_ref_init_from_ptr_and_size (&srcr, src, len);
|
8820 |
|
|
if (!refs_may_alias_p_1 (&destr, &srcr, false))
|
8821 |
|
|
{
|
8822 |
|
|
tree fn;
|
8823 |
|
|
fn = builtin_decl_implicit (BUILT_IN_MEMCPY);
|
8824 |
|
|
if (!fn)
|
8825 |
|
|
return NULL_TREE;
|
8826 |
|
|
return build_call_expr_loc (loc, fn, 3, dest, src, len);
|
8827 |
|
|
}
|
8828 |
|
|
}
|
8829 |
|
|
|
8830 |
|
|
return NULL_TREE;
|
8831 |
|
|
}
|
8832 |
|
|
|
8833 |
|
|
if (!host_integerp (len, 0))
|
8834 |
|
|
return NULL_TREE;
|
8835 |
|
|
/* FIXME:
|
8836 |
|
|
This logic lose for arguments like (type *)malloc (sizeof (type)),
|
8837 |
|
|
since we strip the casts of up to VOID return value from malloc.
|
8838 |
|
|
Perhaps we ought to inherit type from non-VOID argument here? */
|
8839 |
|
|
STRIP_NOPS (src);
|
8840 |
|
|
STRIP_NOPS (dest);
|
8841 |
|
|
if (!POINTER_TYPE_P (TREE_TYPE (src))
|
8842 |
|
|
|| !POINTER_TYPE_P (TREE_TYPE (dest)))
|
8843 |
|
|
return NULL_TREE;
|
8844 |
|
|
/* As we fold (void *)(p + CST) to (void *)p + CST undo this here. */
|
8845 |
|
|
if (TREE_CODE (src) == POINTER_PLUS_EXPR)
|
8846 |
|
|
{
|
8847 |
|
|
tree tem = TREE_OPERAND (src, 0);
|
8848 |
|
|
STRIP_NOPS (tem);
|
8849 |
|
|
if (tem != TREE_OPERAND (src, 0))
|
8850 |
|
|
src = build1 (NOP_EXPR, TREE_TYPE (tem), src);
|
8851 |
|
|
}
|
8852 |
|
|
if (TREE_CODE (dest) == POINTER_PLUS_EXPR)
|
8853 |
|
|
{
|
8854 |
|
|
tree tem = TREE_OPERAND (dest, 0);
|
8855 |
|
|
STRIP_NOPS (tem);
|
8856 |
|
|
if (tem != TREE_OPERAND (dest, 0))
|
8857 |
|
|
dest = build1 (NOP_EXPR, TREE_TYPE (tem), dest);
|
8858 |
|
|
}
|
8859 |
|
|
srctype = TREE_TYPE (TREE_TYPE (src));
|
8860 |
|
|
if (TREE_CODE (srctype) == ARRAY_TYPE
|
8861 |
|
|
&& !tree_int_cst_equal (TYPE_SIZE_UNIT (srctype), len))
|
8862 |
|
|
{
|
8863 |
|
|
srctype = TREE_TYPE (srctype);
|
8864 |
|
|
STRIP_NOPS (src);
|
8865 |
|
|
src = build1 (NOP_EXPR, build_pointer_type (srctype), src);
|
8866 |
|
|
}
|
8867 |
|
|
desttype = TREE_TYPE (TREE_TYPE (dest));
|
8868 |
|
|
if (TREE_CODE (desttype) == ARRAY_TYPE
|
8869 |
|
|
&& !tree_int_cst_equal (TYPE_SIZE_UNIT (desttype), len))
|
8870 |
|
|
{
|
8871 |
|
|
desttype = TREE_TYPE (desttype);
|
8872 |
|
|
STRIP_NOPS (dest);
|
8873 |
|
|
dest = build1 (NOP_EXPR, build_pointer_type (desttype), dest);
|
8874 |
|
|
}
|
8875 |
|
|
if (TREE_ADDRESSABLE (srctype)
|
8876 |
|
|
|| TREE_ADDRESSABLE (desttype))
|
8877 |
|
|
return NULL_TREE;
|
8878 |
|
|
|
8879 |
|
|
src_align = get_pointer_alignment (src);
|
8880 |
|
|
dest_align = get_pointer_alignment (dest);
|
8881 |
|
|
if (dest_align < TYPE_ALIGN (desttype)
|
8882 |
|
|
|| src_align < TYPE_ALIGN (srctype))
|
8883 |
|
|
return NULL_TREE;
|
8884 |
|
|
|
8885 |
|
|
if (!ignore)
|
8886 |
|
|
dest = builtin_save_expr (dest);
|
8887 |
|
|
|
8888 |
|
|
/* Build accesses at offset zero with a ref-all character type. */
|
8889 |
|
|
off0 = build_int_cst (build_pointer_type_for_mode (char_type_node,
|
8890 |
|
|
ptr_mode, true), 0);
|
8891 |
|
|
|
8892 |
|
|
destvar = dest;
|
8893 |
|
|
STRIP_NOPS (destvar);
|
8894 |
|
|
if (TREE_CODE (destvar) == ADDR_EXPR
|
8895 |
|
|
&& var_decl_component_p (TREE_OPERAND (destvar, 0))
|
8896 |
|
|
&& tree_int_cst_equal (TYPE_SIZE_UNIT (desttype), len))
|
8897 |
|
|
destvar = fold_build2 (MEM_REF, desttype, destvar, off0);
|
8898 |
|
|
else
|
8899 |
|
|
destvar = NULL_TREE;
|
8900 |
|
|
|
8901 |
|
|
srcvar = src;
|
8902 |
|
|
STRIP_NOPS (srcvar);
|
8903 |
|
|
if (TREE_CODE (srcvar) == ADDR_EXPR
|
8904 |
|
|
&& var_decl_component_p (TREE_OPERAND (srcvar, 0))
|
8905 |
|
|
&& tree_int_cst_equal (TYPE_SIZE_UNIT (srctype), len))
|
8906 |
|
|
{
|
8907 |
|
|
if (!destvar
|
8908 |
|
|
|| src_align >= TYPE_ALIGN (desttype))
|
8909 |
|
|
srcvar = fold_build2 (MEM_REF, destvar ? desttype : srctype,
|
8910 |
|
|
srcvar, off0);
|
8911 |
|
|
else if (!STRICT_ALIGNMENT)
|
8912 |
|
|
{
|
8913 |
|
|
srctype = build_aligned_type (TYPE_MAIN_VARIANT (desttype),
|
8914 |
|
|
src_align);
|
8915 |
|
|
srcvar = fold_build2 (MEM_REF, srctype, srcvar, off0);
|
8916 |
|
|
}
|
8917 |
|
|
else
|
8918 |
|
|
srcvar = NULL_TREE;
|
8919 |
|
|
}
|
8920 |
|
|
else
|
8921 |
|
|
srcvar = NULL_TREE;
|
8922 |
|
|
|
8923 |
|
|
if (srcvar == NULL_TREE && destvar == NULL_TREE)
|
8924 |
|
|
return NULL_TREE;
|
8925 |
|
|
|
8926 |
|
|
if (srcvar == NULL_TREE)
|
8927 |
|
|
{
|
8928 |
|
|
STRIP_NOPS (src);
|
8929 |
|
|
if (src_align >= TYPE_ALIGN (desttype))
|
8930 |
|
|
srcvar = fold_build2 (MEM_REF, desttype, src, off0);
|
8931 |
|
|
else
|
8932 |
|
|
{
|
8933 |
|
|
if (STRICT_ALIGNMENT)
|
8934 |
|
|
return NULL_TREE;
|
8935 |
|
|
srctype = build_aligned_type (TYPE_MAIN_VARIANT (desttype),
|
8936 |
|
|
src_align);
|
8937 |
|
|
srcvar = fold_build2 (MEM_REF, srctype, src, off0);
|
8938 |
|
|
}
|
8939 |
|
|
}
|
8940 |
|
|
else if (destvar == NULL_TREE)
|
8941 |
|
|
{
|
8942 |
|
|
STRIP_NOPS (dest);
|
8943 |
|
|
if (dest_align >= TYPE_ALIGN (srctype))
|
8944 |
|
|
destvar = fold_build2 (MEM_REF, srctype, dest, off0);
|
8945 |
|
|
else
|
8946 |
|
|
{
|
8947 |
|
|
if (STRICT_ALIGNMENT)
|
8948 |
|
|
return NULL_TREE;
|
8949 |
|
|
desttype = build_aligned_type (TYPE_MAIN_VARIANT (srctype),
|
8950 |
|
|
dest_align);
|
8951 |
|
|
destvar = fold_build2 (MEM_REF, desttype, dest, off0);
|
8952 |
|
|
}
|
8953 |
|
|
}
|
8954 |
|
|
|
8955 |
|
|
expr = build2 (MODIFY_EXPR, TREE_TYPE (destvar), destvar, srcvar);
|
8956 |
|
|
}
|
8957 |
|
|
|
8958 |
|
|
if (ignore)
|
8959 |
|
|
return expr;
|
8960 |
|
|
|
8961 |
|
|
if (endp == 0 || endp == 3)
|
8962 |
|
|
return omit_one_operand_loc (loc, type, dest, expr);
|
8963 |
|
|
|
8964 |
|
|
if (expr == len)
|
8965 |
|
|
expr = NULL_TREE;
|
8966 |
|
|
|
8967 |
|
|
if (endp == 2)
|
8968 |
|
|
len = fold_build2_loc (loc, MINUS_EXPR, TREE_TYPE (len), len,
|
8969 |
|
|
ssize_int (1));
|
8970 |
|
|
|
8971 |
|
|
dest = fold_build_pointer_plus_loc (loc, dest, len);
|
8972 |
|
|
dest = fold_convert_loc (loc, type, dest);
|
8973 |
|
|
if (expr)
|
8974 |
|
|
dest = omit_one_operand_loc (loc, type, dest, expr);
|
8975 |
|
|
return dest;
|
8976 |
|
|
}
|
8977 |
|
|
|
8978 |
|
|
/* Fold function call to builtin strcpy with arguments DEST and SRC.
|
8979 |
|
|
If LEN is not NULL, it represents the length of the string to be
|
8980 |
|
|
copied. Return NULL_TREE if no simplification can be made. */
|
8981 |
|
|
|
8982 |
|
|
tree
|
8983 |
|
|
fold_builtin_strcpy (location_t loc, tree fndecl, tree dest, tree src, tree len)
|
8984 |
|
|
{
|
8985 |
|
|
tree fn;
|
8986 |
|
|
|
8987 |
|
|
if (!validate_arg (dest, POINTER_TYPE)
|
8988 |
|
|
|| !validate_arg (src, POINTER_TYPE))
|
8989 |
|
|
return NULL_TREE;
|
8990 |
|
|
|
8991 |
|
|
/* If SRC and DEST are the same (and not volatile), return DEST. */
|
8992 |
|
|
if (operand_equal_p (src, dest, 0))
|
8993 |
|
|
return fold_convert_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)), dest);
|
8994 |
|
|
|
8995 |
|
|
if (optimize_function_for_size_p (cfun))
|
8996 |
|
|
return NULL_TREE;
|
8997 |
|
|
|
8998 |
|
|
fn = builtin_decl_implicit (BUILT_IN_MEMCPY);
|
8999 |
|
|
if (!fn)
|
9000 |
|
|
return NULL_TREE;
|
9001 |
|
|
|
9002 |
|
|
if (!len)
|
9003 |
|
|
{
|
9004 |
|
|
len = c_strlen (src, 1);
|
9005 |
|
|
if (! len || TREE_SIDE_EFFECTS (len))
|
9006 |
|
|
return NULL_TREE;
|
9007 |
|
|
}
|
9008 |
|
|
|
9009 |
|
|
len = fold_convert_loc (loc, size_type_node, len);
|
9010 |
|
|
len = size_binop_loc (loc, PLUS_EXPR, len, build_int_cst (size_type_node, 1));
|
9011 |
|
|
return fold_convert_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)),
|
9012 |
|
|
build_call_expr_loc (loc, fn, 3, dest, src, len));
|
9013 |
|
|
}
|
9014 |
|
|
|
9015 |
|
|
/* Fold function call to builtin stpcpy with arguments DEST and SRC.
|
9016 |
|
|
Return NULL_TREE if no simplification can be made. */
|
9017 |
|
|
|
9018 |
|
|
static tree
|
9019 |
|
|
fold_builtin_stpcpy (location_t loc, tree fndecl, tree dest, tree src)
|
9020 |
|
|
{
|
9021 |
|
|
tree fn, len, lenp1, call, type;
|
9022 |
|
|
|
9023 |
|
|
if (!validate_arg (dest, POINTER_TYPE)
|
9024 |
|
|
|| !validate_arg (src, POINTER_TYPE))
|
9025 |
|
|
return NULL_TREE;
|
9026 |
|
|
|
9027 |
|
|
len = c_strlen (src, 1);
|
9028 |
|
|
if (!len
|
9029 |
|
|
|| TREE_CODE (len) != INTEGER_CST)
|
9030 |
|
|
return NULL_TREE;
|
9031 |
|
|
|
9032 |
|
|
if (optimize_function_for_size_p (cfun)
|
9033 |
|
|
/* If length is zero it's small enough. */
|
9034 |
|
|
&& !integer_zerop (len))
|
9035 |
|
|
return NULL_TREE;
|
9036 |
|
|
|
9037 |
|
|
fn = builtin_decl_implicit (BUILT_IN_MEMCPY);
|
9038 |
|
|
if (!fn)
|
9039 |
|
|
return NULL_TREE;
|
9040 |
|
|
|
9041 |
|
|
lenp1 = size_binop_loc (loc, PLUS_EXPR,
|
9042 |
|
|
fold_convert_loc (loc, size_type_node, len),
|
9043 |
|
|
build_int_cst (size_type_node, 1));
|
9044 |
|
|
/* We use dest twice in building our expression. Save it from
|
9045 |
|
|
multiple expansions. */
|
9046 |
|
|
dest = builtin_save_expr (dest);
|
9047 |
|
|
call = build_call_expr_loc (loc, fn, 3, dest, src, lenp1);
|
9048 |
|
|
|
9049 |
|
|
type = TREE_TYPE (TREE_TYPE (fndecl));
|
9050 |
|
|
dest = fold_build_pointer_plus_loc (loc, dest, len);
|
9051 |
|
|
dest = fold_convert_loc (loc, type, dest);
|
9052 |
|
|
dest = omit_one_operand_loc (loc, type, dest, call);
|
9053 |
|
|
return dest;
|
9054 |
|
|
}
|
9055 |
|
|
|
9056 |
|
|
/* Fold function call to builtin strncpy with arguments DEST, SRC, and LEN.
|
9057 |
|
|
If SLEN is not NULL, it represents the length of the source string.
|
9058 |
|
|
Return NULL_TREE if no simplification can be made. */
|
9059 |
|
|
|
9060 |
|
|
tree
|
9061 |
|
|
fold_builtin_strncpy (location_t loc, tree fndecl, tree dest,
|
9062 |
|
|
tree src, tree len, tree slen)
|
9063 |
|
|
{
|
9064 |
|
|
tree fn;
|
9065 |
|
|
|
9066 |
|
|
if (!validate_arg (dest, POINTER_TYPE)
|
9067 |
|
|
|| !validate_arg (src, POINTER_TYPE)
|
9068 |
|
|
|| !validate_arg (len, INTEGER_TYPE))
|
9069 |
|
|
return NULL_TREE;
|
9070 |
|
|
|
9071 |
|
|
/* If the LEN parameter is zero, return DEST. */
|
9072 |
|
|
if (integer_zerop (len))
|
9073 |
|
|
return omit_one_operand_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)), dest, src);
|
9074 |
|
|
|
9075 |
|
|
/* We can't compare slen with len as constants below if len is not a
|
9076 |
|
|
constant. */
|
9077 |
|
|
if (len == 0 || TREE_CODE (len) != INTEGER_CST)
|
9078 |
|
|
return NULL_TREE;
|
9079 |
|
|
|
9080 |
|
|
if (!slen)
|
9081 |
|
|
slen = c_strlen (src, 1);
|
9082 |
|
|
|
9083 |
|
|
/* Now, we must be passed a constant src ptr parameter. */
|
9084 |
|
|
if (slen == 0 || TREE_CODE (slen) != INTEGER_CST)
|
9085 |
|
|
return NULL_TREE;
|
9086 |
|
|
|
9087 |
|
|
slen = size_binop_loc (loc, PLUS_EXPR, slen, ssize_int (1));
|
9088 |
|
|
|
9089 |
|
|
/* We do not support simplification of this case, though we do
|
9090 |
|
|
support it when expanding trees into RTL. */
|
9091 |
|
|
/* FIXME: generate a call to __builtin_memset. */
|
9092 |
|
|
if (tree_int_cst_lt (slen, len))
|
9093 |
|
|
return NULL_TREE;
|
9094 |
|
|
|
9095 |
|
|
/* OK transform into builtin memcpy. */
|
9096 |
|
|
fn = builtin_decl_implicit (BUILT_IN_MEMCPY);
|
9097 |
|
|
if (!fn)
|
9098 |
|
|
return NULL_TREE;
|
9099 |
|
|
|
9100 |
|
|
len = fold_convert_loc (loc, size_type_node, len);
|
9101 |
|
|
return fold_convert_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)),
|
9102 |
|
|
build_call_expr_loc (loc, fn, 3, dest, src, len));
|
9103 |
|
|
}
|
9104 |
|
|
|
9105 |
|
|
/* Fold function call to builtin memchr. ARG1, ARG2 and LEN are the
|
9106 |
|
|
arguments to the call, and TYPE is its return type.
|
9107 |
|
|
Return NULL_TREE if no simplification can be made. */
|
9108 |
|
|
|
9109 |
|
|
static tree
|
9110 |
|
|
fold_builtin_memchr (location_t loc, tree arg1, tree arg2, tree len, tree type)
|
9111 |
|
|
{
|
9112 |
|
|
if (!validate_arg (arg1, POINTER_TYPE)
|
9113 |
|
|
|| !validate_arg (arg2, INTEGER_TYPE)
|
9114 |
|
|
|| !validate_arg (len, INTEGER_TYPE))
|
9115 |
|
|
return NULL_TREE;
|
9116 |
|
|
else
|
9117 |
|
|
{
|
9118 |
|
|
const char *p1;
|
9119 |
|
|
|
9120 |
|
|
if (TREE_CODE (arg2) != INTEGER_CST
|
9121 |
|
|
|| !host_integerp (len, 1))
|
9122 |
|
|
return NULL_TREE;
|
9123 |
|
|
|
9124 |
|
|
p1 = c_getstr (arg1);
|
9125 |
|
|
if (p1 && compare_tree_int (len, strlen (p1) + 1) <= 0)
|
9126 |
|
|
{
|
9127 |
|
|
char c;
|
9128 |
|
|
const char *r;
|
9129 |
|
|
tree tem;
|
9130 |
|
|
|
9131 |
|
|
if (target_char_cast (arg2, &c))
|
9132 |
|
|
return NULL_TREE;
|
9133 |
|
|
|
9134 |
|
|
r = (const char *) memchr (p1, c, tree_low_cst (len, 1));
|
9135 |
|
|
|
9136 |
|
|
if (r == NULL)
|
9137 |
|
|
return build_int_cst (TREE_TYPE (arg1), 0);
|
9138 |
|
|
|
9139 |
|
|
tem = fold_build_pointer_plus_hwi_loc (loc, arg1, r - p1);
|
9140 |
|
|
return fold_convert_loc (loc, type, tem);
|
9141 |
|
|
}
|
9142 |
|
|
return NULL_TREE;
|
9143 |
|
|
}
|
9144 |
|
|
}
|
9145 |
|
|
|
9146 |
|
|
/* Fold function call to builtin memcmp with arguments ARG1 and ARG2.
|
9147 |
|
|
Return NULL_TREE if no simplification can be made. */
|
9148 |
|
|
|
9149 |
|
|
static tree
|
9150 |
|
|
fold_builtin_memcmp (location_t loc, tree arg1, tree arg2, tree len)
|
9151 |
|
|
{
|
9152 |
|
|
const char *p1, *p2;
|
9153 |
|
|
|
9154 |
|
|
if (!validate_arg (arg1, POINTER_TYPE)
|
9155 |
|
|
|| !validate_arg (arg2, POINTER_TYPE)
|
9156 |
|
|
|| !validate_arg (len, INTEGER_TYPE))
|
9157 |
|
|
return NULL_TREE;
|
9158 |
|
|
|
9159 |
|
|
/* If the LEN parameter is zero, return zero. */
|
9160 |
|
|
if (integer_zerop (len))
|
9161 |
|
|
return omit_two_operands_loc (loc, integer_type_node, integer_zero_node,
|
9162 |
|
|
arg1, arg2);
|
9163 |
|
|
|
9164 |
|
|
/* If ARG1 and ARG2 are the same (and not volatile), return zero. */
|
9165 |
|
|
if (operand_equal_p (arg1, arg2, 0))
|
9166 |
|
|
return omit_one_operand_loc (loc, integer_type_node, integer_zero_node, len);
|
9167 |
|
|
|
9168 |
|
|
p1 = c_getstr (arg1);
|
9169 |
|
|
p2 = c_getstr (arg2);
|
9170 |
|
|
|
9171 |
|
|
/* If all arguments are constant, and the value of len is not greater
|
9172 |
|
|
than the lengths of arg1 and arg2, evaluate at compile-time. */
|
9173 |
|
|
if (host_integerp (len, 1) && p1 && p2
|
9174 |
|
|
&& compare_tree_int (len, strlen (p1) + 1) <= 0
|
9175 |
|
|
&& compare_tree_int (len, strlen (p2) + 1) <= 0)
|
9176 |
|
|
{
|
9177 |
|
|
const int r = memcmp (p1, p2, tree_low_cst (len, 1));
|
9178 |
|
|
|
9179 |
|
|
if (r > 0)
|
9180 |
|
|
return integer_one_node;
|
9181 |
|
|
else if (r < 0)
|
9182 |
|
|
return integer_minus_one_node;
|
9183 |
|
|
else
|
9184 |
|
|
return integer_zero_node;
|
9185 |
|
|
}
|
9186 |
|
|
|
9187 |
|
|
/* If len parameter is one, return an expression corresponding to
|
9188 |
|
|
(*(const unsigned char*)arg1 - (const unsigned char*)arg2). */
|
9189 |
|
|
if (host_integerp (len, 1) && tree_low_cst (len, 1) == 1)
|
9190 |
|
|
{
|
9191 |
|
|
tree cst_uchar_node = build_type_variant (unsigned_char_type_node, 1, 0);
|
9192 |
|
|
tree cst_uchar_ptr_node
|
9193 |
|
|
= build_pointer_type_for_mode (cst_uchar_node, ptr_mode, true);
|
9194 |
|
|
|
9195 |
|
|
tree ind1
|
9196 |
|
|
= fold_convert_loc (loc, integer_type_node,
|
9197 |
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
9198 |
|
|
fold_convert_loc (loc,
|
9199 |
|
|
cst_uchar_ptr_node,
|
9200 |
|
|
arg1)));
|
9201 |
|
|
tree ind2
|
9202 |
|
|
= fold_convert_loc (loc, integer_type_node,
|
9203 |
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
9204 |
|
|
fold_convert_loc (loc,
|
9205 |
|
|
cst_uchar_ptr_node,
|
9206 |
|
|
arg2)));
|
9207 |
|
|
return fold_build2_loc (loc, MINUS_EXPR, integer_type_node, ind1, ind2);
|
9208 |
|
|
}
|
9209 |
|
|
|
9210 |
|
|
return NULL_TREE;
|
9211 |
|
|
}
|
9212 |
|
|
|
9213 |
|
|
/* Fold function call to builtin strcmp with arguments ARG1 and ARG2.
|
9214 |
|
|
Return NULL_TREE if no simplification can be made. */
|
9215 |
|
|
|
9216 |
|
|
static tree
|
9217 |
|
|
fold_builtin_strcmp (location_t loc, tree arg1, tree arg2)
|
9218 |
|
|
{
|
9219 |
|
|
const char *p1, *p2;
|
9220 |
|
|
|
9221 |
|
|
if (!validate_arg (arg1, POINTER_TYPE)
|
9222 |
|
|
|| !validate_arg (arg2, POINTER_TYPE))
|
9223 |
|
|
return NULL_TREE;
|
9224 |
|
|
|
9225 |
|
|
/* If ARG1 and ARG2 are the same (and not volatile), return zero. */
|
9226 |
|
|
if (operand_equal_p (arg1, arg2, 0))
|
9227 |
|
|
return integer_zero_node;
|
9228 |
|
|
|
9229 |
|
|
p1 = c_getstr (arg1);
|
9230 |
|
|
p2 = c_getstr (arg2);
|
9231 |
|
|
|
9232 |
|
|
if (p1 && p2)
|
9233 |
|
|
{
|
9234 |
|
|
const int i = strcmp (p1, p2);
|
9235 |
|
|
if (i < 0)
|
9236 |
|
|
return integer_minus_one_node;
|
9237 |
|
|
else if (i > 0)
|
9238 |
|
|
return integer_one_node;
|
9239 |
|
|
else
|
9240 |
|
|
return integer_zero_node;
|
9241 |
|
|
}
|
9242 |
|
|
|
9243 |
|
|
/* If the second arg is "", return *(const unsigned char*)arg1. */
|
9244 |
|
|
if (p2 && *p2 == '\0')
|
9245 |
|
|
{
|
9246 |
|
|
tree cst_uchar_node = build_type_variant (unsigned_char_type_node, 1, 0);
|
9247 |
|
|
tree cst_uchar_ptr_node
|
9248 |
|
|
= build_pointer_type_for_mode (cst_uchar_node, ptr_mode, true);
|
9249 |
|
|
|
9250 |
|
|
return fold_convert_loc (loc, integer_type_node,
|
9251 |
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
9252 |
|
|
fold_convert_loc (loc,
|
9253 |
|
|
cst_uchar_ptr_node,
|
9254 |
|
|
arg1)));
|
9255 |
|
|
}
|
9256 |
|
|
|
9257 |
|
|
/* If the first arg is "", return -*(const unsigned char*)arg2. */
|
9258 |
|
|
if (p1 && *p1 == '\0')
|
9259 |
|
|
{
|
9260 |
|
|
tree cst_uchar_node = build_type_variant (unsigned_char_type_node, 1, 0);
|
9261 |
|
|
tree cst_uchar_ptr_node
|
9262 |
|
|
= build_pointer_type_for_mode (cst_uchar_node, ptr_mode, true);
|
9263 |
|
|
|
9264 |
|
|
tree temp
|
9265 |
|
|
= fold_convert_loc (loc, integer_type_node,
|
9266 |
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
9267 |
|
|
fold_convert_loc (loc,
|
9268 |
|
|
cst_uchar_ptr_node,
|
9269 |
|
|
arg2)));
|
9270 |
|
|
return fold_build1_loc (loc, NEGATE_EXPR, integer_type_node, temp);
|
9271 |
|
|
}
|
9272 |
|
|
|
9273 |
|
|
return NULL_TREE;
|
9274 |
|
|
}
|
9275 |
|
|
|
9276 |
|
|
/* Fold function call to builtin strncmp with arguments ARG1, ARG2, and LEN.
|
9277 |
|
|
Return NULL_TREE if no simplification can be made. */
|
9278 |
|
|
|
9279 |
|
|
static tree
|
9280 |
|
|
fold_builtin_strncmp (location_t loc, tree arg1, tree arg2, tree len)
|
9281 |
|
|
{
|
9282 |
|
|
const char *p1, *p2;
|
9283 |
|
|
|
9284 |
|
|
if (!validate_arg (arg1, POINTER_TYPE)
|
9285 |
|
|
|| !validate_arg (arg2, POINTER_TYPE)
|
9286 |
|
|
|| !validate_arg (len, INTEGER_TYPE))
|
9287 |
|
|
return NULL_TREE;
|
9288 |
|
|
|
9289 |
|
|
/* If the LEN parameter is zero, return zero. */
|
9290 |
|
|
if (integer_zerop (len))
|
9291 |
|
|
return omit_two_operands_loc (loc, integer_type_node, integer_zero_node,
|
9292 |
|
|
arg1, arg2);
|
9293 |
|
|
|
9294 |
|
|
/* If ARG1 and ARG2 are the same (and not volatile), return zero. */
|
9295 |
|
|
if (operand_equal_p (arg1, arg2, 0))
|
9296 |
|
|
return omit_one_operand_loc (loc, integer_type_node, integer_zero_node, len);
|
9297 |
|
|
|
9298 |
|
|
p1 = c_getstr (arg1);
|
9299 |
|
|
p2 = c_getstr (arg2);
|
9300 |
|
|
|
9301 |
|
|
if (host_integerp (len, 1) && p1 && p2)
|
9302 |
|
|
{
|
9303 |
|
|
const int i = strncmp (p1, p2, tree_low_cst (len, 1));
|
9304 |
|
|
if (i > 0)
|
9305 |
|
|
return integer_one_node;
|
9306 |
|
|
else if (i < 0)
|
9307 |
|
|
return integer_minus_one_node;
|
9308 |
|
|
else
|
9309 |
|
|
return integer_zero_node;
|
9310 |
|
|
}
|
9311 |
|
|
|
9312 |
|
|
/* If the second arg is "", and the length is greater than zero,
|
9313 |
|
|
return *(const unsigned char*)arg1. */
|
9314 |
|
|
if (p2 && *p2 == '\0'
|
9315 |
|
|
&& TREE_CODE (len) == INTEGER_CST
|
9316 |
|
|
&& tree_int_cst_sgn (len) == 1)
|
9317 |
|
|
{
|
9318 |
|
|
tree cst_uchar_node = build_type_variant (unsigned_char_type_node, 1, 0);
|
9319 |
|
|
tree cst_uchar_ptr_node
|
9320 |
|
|
= build_pointer_type_for_mode (cst_uchar_node, ptr_mode, true);
|
9321 |
|
|
|
9322 |
|
|
return fold_convert_loc (loc, integer_type_node,
|
9323 |
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
9324 |
|
|
fold_convert_loc (loc,
|
9325 |
|
|
cst_uchar_ptr_node,
|
9326 |
|
|
arg1)));
|
9327 |
|
|
}
|
9328 |
|
|
|
9329 |
|
|
/* If the first arg is "", and the length is greater than zero,
|
9330 |
|
|
return -*(const unsigned char*)arg2. */
|
9331 |
|
|
if (p1 && *p1 == '\0'
|
9332 |
|
|
&& TREE_CODE (len) == INTEGER_CST
|
9333 |
|
|
&& tree_int_cst_sgn (len) == 1)
|
9334 |
|
|
{
|
9335 |
|
|
tree cst_uchar_node = build_type_variant (unsigned_char_type_node, 1, 0);
|
9336 |
|
|
tree cst_uchar_ptr_node
|
9337 |
|
|
= build_pointer_type_for_mode (cst_uchar_node, ptr_mode, true);
|
9338 |
|
|
|
9339 |
|
|
tree temp = fold_convert_loc (loc, integer_type_node,
|
9340 |
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
9341 |
|
|
fold_convert_loc (loc,
|
9342 |
|
|
cst_uchar_ptr_node,
|
9343 |
|
|
arg2)));
|
9344 |
|
|
return fold_build1_loc (loc, NEGATE_EXPR, integer_type_node, temp);
|
9345 |
|
|
}
|
9346 |
|
|
|
9347 |
|
|
/* If len parameter is one, return an expression corresponding to
|
9348 |
|
|
(*(const unsigned char*)arg1 - (const unsigned char*)arg2). */
|
9349 |
|
|
if (host_integerp (len, 1) && tree_low_cst (len, 1) == 1)
|
9350 |
|
|
{
|
9351 |
|
|
tree cst_uchar_node = build_type_variant (unsigned_char_type_node, 1, 0);
|
9352 |
|
|
tree cst_uchar_ptr_node
|
9353 |
|
|
= build_pointer_type_for_mode (cst_uchar_node, ptr_mode, true);
|
9354 |
|
|
|
9355 |
|
|
tree ind1 = fold_convert_loc (loc, integer_type_node,
|
9356 |
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
9357 |
|
|
fold_convert_loc (loc,
|
9358 |
|
|
cst_uchar_ptr_node,
|
9359 |
|
|
arg1)));
|
9360 |
|
|
tree ind2 = fold_convert_loc (loc, integer_type_node,
|
9361 |
|
|
build1 (INDIRECT_REF, cst_uchar_node,
|
9362 |
|
|
fold_convert_loc (loc,
|
9363 |
|
|
cst_uchar_ptr_node,
|
9364 |
|
|
arg2)));
|
9365 |
|
|
return fold_build2_loc (loc, MINUS_EXPR, integer_type_node, ind1, ind2);
|
9366 |
|
|
}
|
9367 |
|
|
|
9368 |
|
|
return NULL_TREE;
|
9369 |
|
|
}
|
9370 |
|
|
|
9371 |
|
|
/* Fold function call to builtin signbit, signbitf or signbitl with argument
|
9372 |
|
|
ARG. Return NULL_TREE if no simplification can be made. */
|
9373 |
|
|
|
9374 |
|
|
static tree
|
9375 |
|
|
fold_builtin_signbit (location_t loc, tree arg, tree type)
|
9376 |
|
|
{
|
9377 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
9378 |
|
|
return NULL_TREE;
|
9379 |
|
|
|
9380 |
|
|
/* If ARG is a compile-time constant, determine the result. */
|
9381 |
|
|
if (TREE_CODE (arg) == REAL_CST
|
9382 |
|
|
&& !TREE_OVERFLOW (arg))
|
9383 |
|
|
{
|
9384 |
|
|
REAL_VALUE_TYPE c;
|
9385 |
|
|
|
9386 |
|
|
c = TREE_REAL_CST (arg);
|
9387 |
|
|
return (REAL_VALUE_NEGATIVE (c)
|
9388 |
|
|
? build_one_cst (type)
|
9389 |
|
|
: build_zero_cst (type));
|
9390 |
|
|
}
|
9391 |
|
|
|
9392 |
|
|
/* If ARG is non-negative, the result is always zero. */
|
9393 |
|
|
if (tree_expr_nonnegative_p (arg))
|
9394 |
|
|
return omit_one_operand_loc (loc, type, integer_zero_node, arg);
|
9395 |
|
|
|
9396 |
|
|
/* If ARG's format doesn't have signed zeros, return "arg < 0.0". */
|
9397 |
|
|
if (!HONOR_SIGNED_ZEROS (TYPE_MODE (TREE_TYPE (arg))))
|
9398 |
|
|
return fold_convert (type,
|
9399 |
|
|
fold_build2_loc (loc, LT_EXPR, boolean_type_node, arg,
|
9400 |
|
|
build_real (TREE_TYPE (arg), dconst0)));
|
9401 |
|
|
|
9402 |
|
|
return NULL_TREE;
|
9403 |
|
|
}
|
9404 |
|
|
|
9405 |
|
|
/* Fold function call to builtin copysign, copysignf or copysignl with
|
9406 |
|
|
arguments ARG1 and ARG2. Return NULL_TREE if no simplification can
|
9407 |
|
|
be made. */
|
9408 |
|
|
|
9409 |
|
|
static tree
|
9410 |
|
|
fold_builtin_copysign (location_t loc, tree fndecl,
|
9411 |
|
|
tree arg1, tree arg2, tree type)
|
9412 |
|
|
{
|
9413 |
|
|
tree tem;
|
9414 |
|
|
|
9415 |
|
|
if (!validate_arg (arg1, REAL_TYPE)
|
9416 |
|
|
|| !validate_arg (arg2, REAL_TYPE))
|
9417 |
|
|
return NULL_TREE;
|
9418 |
|
|
|
9419 |
|
|
/* copysign(X,X) is X. */
|
9420 |
|
|
if (operand_equal_p (arg1, arg2, 0))
|
9421 |
|
|
return fold_convert_loc (loc, type, arg1);
|
9422 |
|
|
|
9423 |
|
|
/* If ARG1 and ARG2 are compile-time constants, determine the result. */
|
9424 |
|
|
if (TREE_CODE (arg1) == REAL_CST
|
9425 |
|
|
&& TREE_CODE (arg2) == REAL_CST
|
9426 |
|
|
&& !TREE_OVERFLOW (arg1)
|
9427 |
|
|
&& !TREE_OVERFLOW (arg2))
|
9428 |
|
|
{
|
9429 |
|
|
REAL_VALUE_TYPE c1, c2;
|
9430 |
|
|
|
9431 |
|
|
c1 = TREE_REAL_CST (arg1);
|
9432 |
|
|
c2 = TREE_REAL_CST (arg2);
|
9433 |
|
|
/* c1.sign := c2.sign. */
|
9434 |
|
|
real_copysign (&c1, &c2);
|
9435 |
|
|
return build_real (type, c1);
|
9436 |
|
|
}
|
9437 |
|
|
|
9438 |
|
|
/* copysign(X, Y) is fabs(X) when Y is always non-negative.
|
9439 |
|
|
Remember to evaluate Y for side-effects. */
|
9440 |
|
|
if (tree_expr_nonnegative_p (arg2))
|
9441 |
|
|
return omit_one_operand_loc (loc, type,
|
9442 |
|
|
fold_build1_loc (loc, ABS_EXPR, type, arg1),
|
9443 |
|
|
arg2);
|
9444 |
|
|
|
9445 |
|
|
/* Strip sign changing operations for the first argument. */
|
9446 |
|
|
tem = fold_strip_sign_ops (arg1);
|
9447 |
|
|
if (tem)
|
9448 |
|
|
return build_call_expr_loc (loc, fndecl, 2, tem, arg2);
|
9449 |
|
|
|
9450 |
|
|
return NULL_TREE;
|
9451 |
|
|
}
|
9452 |
|
|
|
9453 |
|
|
/* Fold a call to builtin isascii with argument ARG. */
|
9454 |
|
|
|
9455 |
|
|
static tree
|
9456 |
|
|
fold_builtin_isascii (location_t loc, tree arg)
|
9457 |
|
|
{
|
9458 |
|
|
if (!validate_arg (arg, INTEGER_TYPE))
|
9459 |
|
|
return NULL_TREE;
|
9460 |
|
|
else
|
9461 |
|
|
{
|
9462 |
|
|
/* Transform isascii(c) -> ((c & ~0x7f) == 0). */
|
9463 |
|
|
arg = fold_build2 (BIT_AND_EXPR, integer_type_node, arg,
|
9464 |
|
|
build_int_cst (integer_type_node,
|
9465 |
|
|
~ (unsigned HOST_WIDE_INT) 0x7f));
|
9466 |
|
|
return fold_build2_loc (loc, EQ_EXPR, integer_type_node,
|
9467 |
|
|
arg, integer_zero_node);
|
9468 |
|
|
}
|
9469 |
|
|
}
|
9470 |
|
|
|
9471 |
|
|
/* Fold a call to builtin toascii with argument ARG. */
|
9472 |
|
|
|
9473 |
|
|
static tree
|
9474 |
|
|
fold_builtin_toascii (location_t loc, tree arg)
|
9475 |
|
|
{
|
9476 |
|
|
if (!validate_arg (arg, INTEGER_TYPE))
|
9477 |
|
|
return NULL_TREE;
|
9478 |
|
|
|
9479 |
|
|
/* Transform toascii(c) -> (c & 0x7f). */
|
9480 |
|
|
return fold_build2_loc (loc, BIT_AND_EXPR, integer_type_node, arg,
|
9481 |
|
|
build_int_cst (integer_type_node, 0x7f));
|
9482 |
|
|
}
|
9483 |
|
|
|
9484 |
|
|
/* Fold a call to builtin isdigit with argument ARG. */
|
9485 |
|
|
|
9486 |
|
|
static tree
|
9487 |
|
|
fold_builtin_isdigit (location_t loc, tree arg)
|
9488 |
|
|
{
|
9489 |
|
|
if (!validate_arg (arg, INTEGER_TYPE))
|
9490 |
|
|
return NULL_TREE;
|
9491 |
|
|
else
|
9492 |
|
|
{
|
9493 |
|
|
/* Transform isdigit(c) -> (unsigned)(c) - '0' <= 9. */
|
9494 |
|
|
/* According to the C standard, isdigit is unaffected by locale.
|
9495 |
|
|
However, it definitely is affected by the target character set. */
|
9496 |
|
|
unsigned HOST_WIDE_INT target_digit0
|
9497 |
|
|
= lang_hooks.to_target_charset ('0');
|
9498 |
|
|
|
9499 |
|
|
if (target_digit0 == 0)
|
9500 |
|
|
return NULL_TREE;
|
9501 |
|
|
|
9502 |
|
|
arg = fold_convert_loc (loc, unsigned_type_node, arg);
|
9503 |
|
|
arg = fold_build2 (MINUS_EXPR, unsigned_type_node, arg,
|
9504 |
|
|
build_int_cst (unsigned_type_node, target_digit0));
|
9505 |
|
|
return fold_build2_loc (loc, LE_EXPR, integer_type_node, arg,
|
9506 |
|
|
build_int_cst (unsigned_type_node, 9));
|
9507 |
|
|
}
|
9508 |
|
|
}
|
9509 |
|
|
|
9510 |
|
|
/* Fold a call to fabs, fabsf or fabsl with argument ARG. */
|
9511 |
|
|
|
9512 |
|
|
static tree
|
9513 |
|
|
fold_builtin_fabs (location_t loc, tree arg, tree type)
|
9514 |
|
|
{
|
9515 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
9516 |
|
|
return NULL_TREE;
|
9517 |
|
|
|
9518 |
|
|
arg = fold_convert_loc (loc, type, arg);
|
9519 |
|
|
if (TREE_CODE (arg) == REAL_CST)
|
9520 |
|
|
return fold_abs_const (arg, type);
|
9521 |
|
|
return fold_build1_loc (loc, ABS_EXPR, type, arg);
|
9522 |
|
|
}
|
9523 |
|
|
|
9524 |
|
|
/* Fold a call to abs, labs, llabs or imaxabs with argument ARG. */
|
9525 |
|
|
|
9526 |
|
|
static tree
|
9527 |
|
|
fold_builtin_abs (location_t loc, tree arg, tree type)
|
9528 |
|
|
{
|
9529 |
|
|
if (!validate_arg (arg, INTEGER_TYPE))
|
9530 |
|
|
return NULL_TREE;
|
9531 |
|
|
|
9532 |
|
|
arg = fold_convert_loc (loc, type, arg);
|
9533 |
|
|
if (TREE_CODE (arg) == INTEGER_CST)
|
9534 |
|
|
return fold_abs_const (arg, type);
|
9535 |
|
|
return fold_build1_loc (loc, ABS_EXPR, type, arg);
|
9536 |
|
|
}
|
9537 |
|
|
|
9538 |
|
|
/* Fold a fma operation with arguments ARG[012]. */
|
9539 |
|
|
|
9540 |
|
|
tree
|
9541 |
|
|
fold_fma (location_t loc ATTRIBUTE_UNUSED,
|
9542 |
|
|
tree type, tree arg0, tree arg1, tree arg2)
|
9543 |
|
|
{
|
9544 |
|
|
if (TREE_CODE (arg0) == REAL_CST
|
9545 |
|
|
&& TREE_CODE (arg1) == REAL_CST
|
9546 |
|
|
&& TREE_CODE (arg2) == REAL_CST)
|
9547 |
|
|
return do_mpfr_arg3 (arg0, arg1, arg2, type, mpfr_fma);
|
9548 |
|
|
|
9549 |
|
|
return NULL_TREE;
|
9550 |
|
|
}
|
9551 |
|
|
|
9552 |
|
|
/* Fold a call to fma, fmaf, or fmal with arguments ARG[012]. */
|
9553 |
|
|
|
9554 |
|
|
static tree
|
9555 |
|
|
fold_builtin_fma (location_t loc, tree arg0, tree arg1, tree arg2, tree type)
|
9556 |
|
|
{
|
9557 |
|
|
if (validate_arg (arg0, REAL_TYPE)
|
9558 |
|
|
&& validate_arg(arg1, REAL_TYPE)
|
9559 |
|
|
&& validate_arg(arg2, REAL_TYPE))
|
9560 |
|
|
{
|
9561 |
|
|
tree tem = fold_fma (loc, type, arg0, arg1, arg2);
|
9562 |
|
|
if (tem)
|
9563 |
|
|
return tem;
|
9564 |
|
|
|
9565 |
|
|
/* ??? Only expand to FMA_EXPR if it's directly supported. */
|
9566 |
|
|
if (optab_handler (fma_optab, TYPE_MODE (type)) != CODE_FOR_nothing)
|
9567 |
|
|
return fold_build3_loc (loc, FMA_EXPR, type, arg0, arg1, arg2);
|
9568 |
|
|
}
|
9569 |
|
|
return NULL_TREE;
|
9570 |
|
|
}
|
9571 |
|
|
|
9572 |
|
|
/* Fold a call to builtin fmin or fmax. */
|
9573 |
|
|
|
9574 |
|
|
static tree
|
9575 |
|
|
fold_builtin_fmin_fmax (location_t loc, tree arg0, tree arg1,
|
9576 |
|
|
tree type, bool max)
|
9577 |
|
|
{
|
9578 |
|
|
if (validate_arg (arg0, REAL_TYPE) && validate_arg (arg1, REAL_TYPE))
|
9579 |
|
|
{
|
9580 |
|
|
/* Calculate the result when the argument is a constant. */
|
9581 |
|
|
tree res = do_mpfr_arg2 (arg0, arg1, type, (max ? mpfr_max : mpfr_min));
|
9582 |
|
|
|
9583 |
|
|
if (res)
|
9584 |
|
|
return res;
|
9585 |
|
|
|
9586 |
|
|
/* If either argument is NaN, return the other one. Avoid the
|
9587 |
|
|
transformation if we get (and honor) a signalling NaN. Using
|
9588 |
|
|
omit_one_operand() ensures we create a non-lvalue. */
|
9589 |
|
|
if (TREE_CODE (arg0) == REAL_CST
|
9590 |
|
|
&& real_isnan (&TREE_REAL_CST (arg0))
|
9591 |
|
|
&& (! HONOR_SNANS (TYPE_MODE (TREE_TYPE (arg0)))
|
9592 |
|
|
|| ! TREE_REAL_CST (arg0).signalling))
|
9593 |
|
|
return omit_one_operand_loc (loc, type, arg1, arg0);
|
9594 |
|
|
if (TREE_CODE (arg1) == REAL_CST
|
9595 |
|
|
&& real_isnan (&TREE_REAL_CST (arg1))
|
9596 |
|
|
&& (! HONOR_SNANS (TYPE_MODE (TREE_TYPE (arg1)))
|
9597 |
|
|
|| ! TREE_REAL_CST (arg1).signalling))
|
9598 |
|
|
return omit_one_operand_loc (loc, type, arg0, arg1);
|
9599 |
|
|
|
9600 |
|
|
/* Transform fmin/fmax(x,x) -> x. */
|
9601 |
|
|
if (operand_equal_p (arg0, arg1, OEP_PURE_SAME))
|
9602 |
|
|
return omit_one_operand_loc (loc, type, arg0, arg1);
|
9603 |
|
|
|
9604 |
|
|
/* Convert fmin/fmax to MIN_EXPR/MAX_EXPR. C99 requires these
|
9605 |
|
|
functions to return the numeric arg if the other one is NaN.
|
9606 |
|
|
These tree codes don't honor that, so only transform if
|
9607 |
|
|
-ffinite-math-only is set. C99 doesn't require -0.0 to be
|
9608 |
|
|
handled, so we don't have to worry about it either. */
|
9609 |
|
|
if (flag_finite_math_only)
|
9610 |
|
|
return fold_build2_loc (loc, (max ? MAX_EXPR : MIN_EXPR), type,
|
9611 |
|
|
fold_convert_loc (loc, type, arg0),
|
9612 |
|
|
fold_convert_loc (loc, type, arg1));
|
9613 |
|
|
}
|
9614 |
|
|
return NULL_TREE;
|
9615 |
|
|
}
|
9616 |
|
|
|
9617 |
|
|
/* Fold a call to builtin carg(a+bi) -> atan2(b,a). */
|
9618 |
|
|
|
9619 |
|
|
static tree
|
9620 |
|
|
fold_builtin_carg (location_t loc, tree arg, tree type)
|
9621 |
|
|
{
|
9622 |
|
|
if (validate_arg (arg, COMPLEX_TYPE)
|
9623 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg))) == REAL_TYPE)
|
9624 |
|
|
{
|
9625 |
|
|
tree atan2_fn = mathfn_built_in (type, BUILT_IN_ATAN2);
|
9626 |
|
|
|
9627 |
|
|
if (atan2_fn)
|
9628 |
|
|
{
|
9629 |
|
|
tree new_arg = builtin_save_expr (arg);
|
9630 |
|
|
tree r_arg = fold_build1_loc (loc, REALPART_EXPR, type, new_arg);
|
9631 |
|
|
tree i_arg = fold_build1_loc (loc, IMAGPART_EXPR, type, new_arg);
|
9632 |
|
|
return build_call_expr_loc (loc, atan2_fn, 2, i_arg, r_arg);
|
9633 |
|
|
}
|
9634 |
|
|
}
|
9635 |
|
|
|
9636 |
|
|
return NULL_TREE;
|
9637 |
|
|
}
|
9638 |
|
|
|
9639 |
|
|
/* Fold a call to builtin logb/ilogb. */
|
9640 |
|
|
|
9641 |
|
|
static tree
|
9642 |
|
|
fold_builtin_logb (location_t loc, tree arg, tree rettype)
|
9643 |
|
|
{
|
9644 |
|
|
if (! validate_arg (arg, REAL_TYPE))
|
9645 |
|
|
return NULL_TREE;
|
9646 |
|
|
|
9647 |
|
|
STRIP_NOPS (arg);
|
9648 |
|
|
|
9649 |
|
|
if (TREE_CODE (arg) == REAL_CST && ! TREE_OVERFLOW (arg))
|
9650 |
|
|
{
|
9651 |
|
|
const REAL_VALUE_TYPE *const value = TREE_REAL_CST_PTR (arg);
|
9652 |
|
|
|
9653 |
|
|
switch (value->cl)
|
9654 |
|
|
{
|
9655 |
|
|
case rvc_nan:
|
9656 |
|
|
case rvc_inf:
|
9657 |
|
|
/* If arg is Inf or NaN and we're logb, return it. */
|
9658 |
|
|
if (TREE_CODE (rettype) == REAL_TYPE)
|
9659 |
|
|
return fold_convert_loc (loc, rettype, arg);
|
9660 |
|
|
/* Fall through... */
|
9661 |
|
|
case rvc_zero:
|
9662 |
|
|
/* Zero may set errno and/or raise an exception for logb, also
|
9663 |
|
|
for ilogb we don't know FP_ILOGB0. */
|
9664 |
|
|
return NULL_TREE;
|
9665 |
|
|
case rvc_normal:
|
9666 |
|
|
/* For normal numbers, proceed iff radix == 2. In GCC,
|
9667 |
|
|
normalized significands are in the range [0.5, 1.0). We
|
9668 |
|
|
want the exponent as if they were [1.0, 2.0) so get the
|
9669 |
|
|
exponent and subtract 1. */
|
9670 |
|
|
if (REAL_MODE_FORMAT (TYPE_MODE (TREE_TYPE (arg)))->b == 2)
|
9671 |
|
|
return fold_convert_loc (loc, rettype,
|
9672 |
|
|
build_int_cst (integer_type_node,
|
9673 |
|
|
REAL_EXP (value)-1));
|
9674 |
|
|
break;
|
9675 |
|
|
}
|
9676 |
|
|
}
|
9677 |
|
|
|
9678 |
|
|
return NULL_TREE;
|
9679 |
|
|
}
|
9680 |
|
|
|
9681 |
|
|
/* Fold a call to builtin significand, if radix == 2. */
|
9682 |
|
|
|
9683 |
|
|
static tree
|
9684 |
|
|
fold_builtin_significand (location_t loc, tree arg, tree rettype)
|
9685 |
|
|
{
|
9686 |
|
|
if (! validate_arg (arg, REAL_TYPE))
|
9687 |
|
|
return NULL_TREE;
|
9688 |
|
|
|
9689 |
|
|
STRIP_NOPS (arg);
|
9690 |
|
|
|
9691 |
|
|
if (TREE_CODE (arg) == REAL_CST && ! TREE_OVERFLOW (arg))
|
9692 |
|
|
{
|
9693 |
|
|
const REAL_VALUE_TYPE *const value = TREE_REAL_CST_PTR (arg);
|
9694 |
|
|
|
9695 |
|
|
switch (value->cl)
|
9696 |
|
|
{
|
9697 |
|
|
case rvc_zero:
|
9698 |
|
|
case rvc_nan:
|
9699 |
|
|
case rvc_inf:
|
9700 |
|
|
/* If arg is +-0, +-Inf or +-NaN, then return it. */
|
9701 |
|
|
return fold_convert_loc (loc, rettype, arg);
|
9702 |
|
|
case rvc_normal:
|
9703 |
|
|
/* For normal numbers, proceed iff radix == 2. */
|
9704 |
|
|
if (REAL_MODE_FORMAT (TYPE_MODE (TREE_TYPE (arg)))->b == 2)
|
9705 |
|
|
{
|
9706 |
|
|
REAL_VALUE_TYPE result = *value;
|
9707 |
|
|
/* In GCC, normalized significands are in the range [0.5,
|
9708 |
|
|
1.0). We want them to be [1.0, 2.0) so set the
|
9709 |
|
|
exponent to 1. */
|
9710 |
|
|
SET_REAL_EXP (&result, 1);
|
9711 |
|
|
return build_real (rettype, result);
|
9712 |
|
|
}
|
9713 |
|
|
break;
|
9714 |
|
|
}
|
9715 |
|
|
}
|
9716 |
|
|
|
9717 |
|
|
return NULL_TREE;
|
9718 |
|
|
}
|
9719 |
|
|
|
9720 |
|
|
/* Fold a call to builtin frexp, we can assume the base is 2. */
|
9721 |
|
|
|
9722 |
|
|
static tree
|
9723 |
|
|
fold_builtin_frexp (location_t loc, tree arg0, tree arg1, tree rettype)
|
9724 |
|
|
{
|
9725 |
|
|
if (! validate_arg (arg0, REAL_TYPE) || ! validate_arg (arg1, POINTER_TYPE))
|
9726 |
|
|
return NULL_TREE;
|
9727 |
|
|
|
9728 |
|
|
STRIP_NOPS (arg0);
|
9729 |
|
|
|
9730 |
|
|
if (!(TREE_CODE (arg0) == REAL_CST && ! TREE_OVERFLOW (arg0)))
|
9731 |
|
|
return NULL_TREE;
|
9732 |
|
|
|
9733 |
|
|
arg1 = build_fold_indirect_ref_loc (loc, arg1);
|
9734 |
|
|
|
9735 |
|
|
/* Proceed if a valid pointer type was passed in. */
|
9736 |
|
|
if (TYPE_MAIN_VARIANT (TREE_TYPE (arg1)) == integer_type_node)
|
9737 |
|
|
{
|
9738 |
|
|
const REAL_VALUE_TYPE *const value = TREE_REAL_CST_PTR (arg0);
|
9739 |
|
|
tree frac, exp;
|
9740 |
|
|
|
9741 |
|
|
switch (value->cl)
|
9742 |
|
|
{
|
9743 |
|
|
case rvc_zero:
|
9744 |
|
|
/* For +-0, return (*exp = 0, +-0). */
|
9745 |
|
|
exp = integer_zero_node;
|
9746 |
|
|
frac = arg0;
|
9747 |
|
|
break;
|
9748 |
|
|
case rvc_nan:
|
9749 |
|
|
case rvc_inf:
|
9750 |
|
|
/* For +-NaN or +-Inf, *exp is unspecified, return arg0. */
|
9751 |
|
|
return omit_one_operand_loc (loc, rettype, arg0, arg1);
|
9752 |
|
|
case rvc_normal:
|
9753 |
|
|
{
|
9754 |
|
|
/* Since the frexp function always expects base 2, and in
|
9755 |
|
|
GCC normalized significands are already in the range
|
9756 |
|
|
[0.5, 1.0), we have exactly what frexp wants. */
|
9757 |
|
|
REAL_VALUE_TYPE frac_rvt = *value;
|
9758 |
|
|
SET_REAL_EXP (&frac_rvt, 0);
|
9759 |
|
|
frac = build_real (rettype, frac_rvt);
|
9760 |
|
|
exp = build_int_cst (integer_type_node, REAL_EXP (value));
|
9761 |
|
|
}
|
9762 |
|
|
break;
|
9763 |
|
|
default:
|
9764 |
|
|
gcc_unreachable ();
|
9765 |
|
|
}
|
9766 |
|
|
|
9767 |
|
|
/* Create the COMPOUND_EXPR (*arg1 = trunc, frac). */
|
9768 |
|
|
arg1 = fold_build2_loc (loc, MODIFY_EXPR, rettype, arg1, exp);
|
9769 |
|
|
TREE_SIDE_EFFECTS (arg1) = 1;
|
9770 |
|
|
return fold_build2_loc (loc, COMPOUND_EXPR, rettype, arg1, frac);
|
9771 |
|
|
}
|
9772 |
|
|
|
9773 |
|
|
return NULL_TREE;
|
9774 |
|
|
}
|
9775 |
|
|
|
9776 |
|
|
/* Fold a call to builtin ldexp or scalbn/scalbln. If LDEXP is true
|
9777 |
|
|
then we can assume the base is two. If it's false, then we have to
|
9778 |
|
|
check the mode of the TYPE parameter in certain cases. */
|
9779 |
|
|
|
9780 |
|
|
static tree
|
9781 |
|
|
fold_builtin_load_exponent (location_t loc, tree arg0, tree arg1,
|
9782 |
|
|
tree type, bool ldexp)
|
9783 |
|
|
{
|
9784 |
|
|
if (validate_arg (arg0, REAL_TYPE) && validate_arg (arg1, INTEGER_TYPE))
|
9785 |
|
|
{
|
9786 |
|
|
STRIP_NOPS (arg0);
|
9787 |
|
|
STRIP_NOPS (arg1);
|
9788 |
|
|
|
9789 |
|
|
/* If arg0 is 0, Inf or NaN, or if arg1 is 0, then return arg0. */
|
9790 |
|
|
if (real_zerop (arg0) || integer_zerop (arg1)
|
9791 |
|
|
|| (TREE_CODE (arg0) == REAL_CST
|
9792 |
|
|
&& !real_isfinite (&TREE_REAL_CST (arg0))))
|
9793 |
|
|
return omit_one_operand_loc (loc, type, arg0, arg1);
|
9794 |
|
|
|
9795 |
|
|
/* If both arguments are constant, then try to evaluate it. */
|
9796 |
|
|
if ((ldexp || REAL_MODE_FORMAT (TYPE_MODE (type))->b == 2)
|
9797 |
|
|
&& TREE_CODE (arg0) == REAL_CST && !TREE_OVERFLOW (arg0)
|
9798 |
|
|
&& host_integerp (arg1, 0))
|
9799 |
|
|
{
|
9800 |
|
|
/* Bound the maximum adjustment to twice the range of the
|
9801 |
|
|
mode's valid exponents. Use abs to ensure the range is
|
9802 |
|
|
positive as a sanity check. */
|
9803 |
|
|
const long max_exp_adj = 2 *
|
9804 |
|
|
labs (REAL_MODE_FORMAT (TYPE_MODE (type))->emax
|
9805 |
|
|
- REAL_MODE_FORMAT (TYPE_MODE (type))->emin);
|
9806 |
|
|
|
9807 |
|
|
/* Get the user-requested adjustment. */
|
9808 |
|
|
const HOST_WIDE_INT req_exp_adj = tree_low_cst (arg1, 0);
|
9809 |
|
|
|
9810 |
|
|
/* The requested adjustment must be inside this range. This
|
9811 |
|
|
is a preliminary cap to avoid things like overflow, we
|
9812 |
|
|
may still fail to compute the result for other reasons. */
|
9813 |
|
|
if (-max_exp_adj < req_exp_adj && req_exp_adj < max_exp_adj)
|
9814 |
|
|
{
|
9815 |
|
|
REAL_VALUE_TYPE initial_result;
|
9816 |
|
|
|
9817 |
|
|
real_ldexp (&initial_result, &TREE_REAL_CST (arg0), req_exp_adj);
|
9818 |
|
|
|
9819 |
|
|
/* Ensure we didn't overflow. */
|
9820 |
|
|
if (! real_isinf (&initial_result))
|
9821 |
|
|
{
|
9822 |
|
|
const REAL_VALUE_TYPE trunc_result
|
9823 |
|
|
= real_value_truncate (TYPE_MODE (type), initial_result);
|
9824 |
|
|
|
9825 |
|
|
/* Only proceed if the target mode can hold the
|
9826 |
|
|
resulting value. */
|
9827 |
|
|
if (REAL_VALUES_EQUAL (initial_result, trunc_result))
|
9828 |
|
|
return build_real (type, trunc_result);
|
9829 |
|
|
}
|
9830 |
|
|
}
|
9831 |
|
|
}
|
9832 |
|
|
}
|
9833 |
|
|
|
9834 |
|
|
return NULL_TREE;
|
9835 |
|
|
}
|
9836 |
|
|
|
9837 |
|
|
/* Fold a call to builtin modf. */
|
9838 |
|
|
|
9839 |
|
|
static tree
|
9840 |
|
|
fold_builtin_modf (location_t loc, tree arg0, tree arg1, tree rettype)
|
9841 |
|
|
{
|
9842 |
|
|
if (! validate_arg (arg0, REAL_TYPE) || ! validate_arg (arg1, POINTER_TYPE))
|
9843 |
|
|
return NULL_TREE;
|
9844 |
|
|
|
9845 |
|
|
STRIP_NOPS (arg0);
|
9846 |
|
|
|
9847 |
|
|
if (!(TREE_CODE (arg0) == REAL_CST && ! TREE_OVERFLOW (arg0)))
|
9848 |
|
|
return NULL_TREE;
|
9849 |
|
|
|
9850 |
|
|
arg1 = build_fold_indirect_ref_loc (loc, arg1);
|
9851 |
|
|
|
9852 |
|
|
/* Proceed if a valid pointer type was passed in. */
|
9853 |
|
|
if (TYPE_MAIN_VARIANT (TREE_TYPE (arg1)) == TYPE_MAIN_VARIANT (rettype))
|
9854 |
|
|
{
|
9855 |
|
|
const REAL_VALUE_TYPE *const value = TREE_REAL_CST_PTR (arg0);
|
9856 |
|
|
REAL_VALUE_TYPE trunc, frac;
|
9857 |
|
|
|
9858 |
|
|
switch (value->cl)
|
9859 |
|
|
{
|
9860 |
|
|
case rvc_nan:
|
9861 |
|
|
case rvc_zero:
|
9862 |
|
|
/* For +-NaN or +-0, return (*arg1 = arg0, arg0). */
|
9863 |
|
|
trunc = frac = *value;
|
9864 |
|
|
break;
|
9865 |
|
|
case rvc_inf:
|
9866 |
|
|
/* For +-Inf, return (*arg1 = arg0, +-0). */
|
9867 |
|
|
frac = dconst0;
|
9868 |
|
|
frac.sign = value->sign;
|
9869 |
|
|
trunc = *value;
|
9870 |
|
|
break;
|
9871 |
|
|
case rvc_normal:
|
9872 |
|
|
/* Return (*arg1 = trunc(arg0), arg0-trunc(arg0)). */
|
9873 |
|
|
real_trunc (&trunc, VOIDmode, value);
|
9874 |
|
|
real_arithmetic (&frac, MINUS_EXPR, value, &trunc);
|
9875 |
|
|
/* If the original number was negative and already
|
9876 |
|
|
integral, then the fractional part is -0.0. */
|
9877 |
|
|
if (value->sign && frac.cl == rvc_zero)
|
9878 |
|
|
frac.sign = value->sign;
|
9879 |
|
|
break;
|
9880 |
|
|
}
|
9881 |
|
|
|
9882 |
|
|
/* Create the COMPOUND_EXPR (*arg1 = trunc, frac). */
|
9883 |
|
|
arg1 = fold_build2_loc (loc, MODIFY_EXPR, rettype, arg1,
|
9884 |
|
|
build_real (rettype, trunc));
|
9885 |
|
|
TREE_SIDE_EFFECTS (arg1) = 1;
|
9886 |
|
|
return fold_build2_loc (loc, COMPOUND_EXPR, rettype, arg1,
|
9887 |
|
|
build_real (rettype, frac));
|
9888 |
|
|
}
|
9889 |
|
|
|
9890 |
|
|
return NULL_TREE;
|
9891 |
|
|
}
|
9892 |
|
|
|
9893 |
|
|
/* Given a location LOC, an interclass builtin function decl FNDECL
|
9894 |
|
|
and its single argument ARG, return an folded expression computing
|
9895 |
|
|
the same, or NULL_TREE if we either couldn't or didn't want to fold
|
9896 |
|
|
(the latter happen if there's an RTL instruction available). */
|
9897 |
|
|
|
9898 |
|
|
static tree
|
9899 |
|
|
fold_builtin_interclass_mathfn (location_t loc, tree fndecl, tree arg)
|
9900 |
|
|
{
|
9901 |
|
|
enum machine_mode mode;
|
9902 |
|
|
|
9903 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
9904 |
|
|
return NULL_TREE;
|
9905 |
|
|
|
9906 |
|
|
if (interclass_mathfn_icode (arg, fndecl) != CODE_FOR_nothing)
|
9907 |
|
|
return NULL_TREE;
|
9908 |
|
|
|
9909 |
|
|
mode = TYPE_MODE (TREE_TYPE (arg));
|
9910 |
|
|
|
9911 |
|
|
/* If there is no optab, try generic code. */
|
9912 |
|
|
switch (DECL_FUNCTION_CODE (fndecl))
|
9913 |
|
|
{
|
9914 |
|
|
tree result;
|
9915 |
|
|
|
9916 |
|
|
CASE_FLT_FN (BUILT_IN_ISINF):
|
9917 |
|
|
{
|
9918 |
|
|
/* isinf(x) -> isgreater(fabs(x),DBL_MAX). */
|
9919 |
|
|
tree const isgr_fn = builtin_decl_explicit (BUILT_IN_ISGREATER);
|
9920 |
|
|
tree const type = TREE_TYPE (arg);
|
9921 |
|
|
REAL_VALUE_TYPE r;
|
9922 |
|
|
char buf[128];
|
9923 |
|
|
|
9924 |
|
|
get_max_float (REAL_MODE_FORMAT (mode), buf, sizeof (buf));
|
9925 |
|
|
real_from_string (&r, buf);
|
9926 |
|
|
result = build_call_expr (isgr_fn, 2,
|
9927 |
|
|
fold_build1_loc (loc, ABS_EXPR, type, arg),
|
9928 |
|
|
build_real (type, r));
|
9929 |
|
|
return result;
|
9930 |
|
|
}
|
9931 |
|
|
CASE_FLT_FN (BUILT_IN_FINITE):
|
9932 |
|
|
case BUILT_IN_ISFINITE:
|
9933 |
|
|
{
|
9934 |
|
|
/* isfinite(x) -> islessequal(fabs(x),DBL_MAX). */
|
9935 |
|
|
tree const isle_fn = builtin_decl_explicit (BUILT_IN_ISLESSEQUAL);
|
9936 |
|
|
tree const type = TREE_TYPE (arg);
|
9937 |
|
|
REAL_VALUE_TYPE r;
|
9938 |
|
|
char buf[128];
|
9939 |
|
|
|
9940 |
|
|
get_max_float (REAL_MODE_FORMAT (mode), buf, sizeof (buf));
|
9941 |
|
|
real_from_string (&r, buf);
|
9942 |
|
|
result = build_call_expr (isle_fn, 2,
|
9943 |
|
|
fold_build1_loc (loc, ABS_EXPR, type, arg),
|
9944 |
|
|
build_real (type, r));
|
9945 |
|
|
/*result = fold_build2_loc (loc, UNGT_EXPR,
|
9946 |
|
|
TREE_TYPE (TREE_TYPE (fndecl)),
|
9947 |
|
|
fold_build1_loc (loc, ABS_EXPR, type, arg),
|
9948 |
|
|
build_real (type, r));
|
9949 |
|
|
result = fold_build1_loc (loc, TRUTH_NOT_EXPR,
|
9950 |
|
|
TREE_TYPE (TREE_TYPE (fndecl)),
|
9951 |
|
|
result);*/
|
9952 |
|
|
return result;
|
9953 |
|
|
}
|
9954 |
|
|
case BUILT_IN_ISNORMAL:
|
9955 |
|
|
{
|
9956 |
|
|
/* isnormal(x) -> isgreaterequal(fabs(x),DBL_MIN) &
|
9957 |
|
|
islessequal(fabs(x),DBL_MAX). */
|
9958 |
|
|
tree const isle_fn = builtin_decl_explicit (BUILT_IN_ISLESSEQUAL);
|
9959 |
|
|
tree const isge_fn = builtin_decl_explicit (BUILT_IN_ISGREATEREQUAL);
|
9960 |
|
|
tree const type = TREE_TYPE (arg);
|
9961 |
|
|
REAL_VALUE_TYPE rmax, rmin;
|
9962 |
|
|
char buf[128];
|
9963 |
|
|
|
9964 |
|
|
get_max_float (REAL_MODE_FORMAT (mode), buf, sizeof (buf));
|
9965 |
|
|
real_from_string (&rmax, buf);
|
9966 |
|
|
sprintf (buf, "0x1p%d", REAL_MODE_FORMAT (mode)->emin - 1);
|
9967 |
|
|
real_from_string (&rmin, buf);
|
9968 |
|
|
arg = builtin_save_expr (fold_build1_loc (loc, ABS_EXPR, type, arg));
|
9969 |
|
|
result = build_call_expr (isle_fn, 2, arg,
|
9970 |
|
|
build_real (type, rmax));
|
9971 |
|
|
result = fold_build2 (BIT_AND_EXPR, integer_type_node, result,
|
9972 |
|
|
build_call_expr (isge_fn, 2, arg,
|
9973 |
|
|
build_real (type, rmin)));
|
9974 |
|
|
return result;
|
9975 |
|
|
}
|
9976 |
|
|
default:
|
9977 |
|
|
break;
|
9978 |
|
|
}
|
9979 |
|
|
|
9980 |
|
|
return NULL_TREE;
|
9981 |
|
|
}
|
9982 |
|
|
|
9983 |
|
|
/* Fold a call to __builtin_isnan(), __builtin_isinf, __builtin_finite.
|
9984 |
|
|
ARG is the argument for the call. */
|
9985 |
|
|
|
9986 |
|
|
static tree
|
9987 |
|
|
fold_builtin_classify (location_t loc, tree fndecl, tree arg, int builtin_index)
|
9988 |
|
|
{
|
9989 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
9990 |
|
|
REAL_VALUE_TYPE r;
|
9991 |
|
|
|
9992 |
|
|
if (!validate_arg (arg, REAL_TYPE))
|
9993 |
|
|
return NULL_TREE;
|
9994 |
|
|
|
9995 |
|
|
switch (builtin_index)
|
9996 |
|
|
{
|
9997 |
|
|
case BUILT_IN_ISINF:
|
9998 |
|
|
if (!HONOR_INFINITIES (TYPE_MODE (TREE_TYPE (arg))))
|
9999 |
|
|
return omit_one_operand_loc (loc, type, integer_zero_node, arg);
|
10000 |
|
|
|
10001 |
|
|
if (TREE_CODE (arg) == REAL_CST)
|
10002 |
|
|
{
|
10003 |
|
|
r = TREE_REAL_CST (arg);
|
10004 |
|
|
if (real_isinf (&r))
|
10005 |
|
|
return real_compare (GT_EXPR, &r, &dconst0)
|
10006 |
|
|
? integer_one_node : integer_minus_one_node;
|
10007 |
|
|
else
|
10008 |
|
|
return integer_zero_node;
|
10009 |
|
|
}
|
10010 |
|
|
|
10011 |
|
|
return NULL_TREE;
|
10012 |
|
|
|
10013 |
|
|
case BUILT_IN_ISINF_SIGN:
|
10014 |
|
|
{
|
10015 |
|
|
/* isinf_sign(x) -> isinf(x) ? (signbit(x) ? -1 : 1) : 0 */
|
10016 |
|
|
/* In a boolean context, GCC will fold the inner COND_EXPR to
|
10017 |
|
|
1. So e.g. "if (isinf_sign(x))" would be folded to just
|
10018 |
|
|
"if (isinf(x) ? 1 : 0)" which becomes "if (isinf(x))". */
|
10019 |
|
|
tree signbit_fn = mathfn_built_in_1 (TREE_TYPE (arg), BUILT_IN_SIGNBIT, 0);
|
10020 |
|
|
tree isinf_fn = builtin_decl_explicit (BUILT_IN_ISINF);
|
10021 |
|
|
tree tmp = NULL_TREE;
|
10022 |
|
|
|
10023 |
|
|
arg = builtin_save_expr (arg);
|
10024 |
|
|
|
10025 |
|
|
if (signbit_fn && isinf_fn)
|
10026 |
|
|
{
|
10027 |
|
|
tree signbit_call = build_call_expr_loc (loc, signbit_fn, 1, arg);
|
10028 |
|
|
tree isinf_call = build_call_expr_loc (loc, isinf_fn, 1, arg);
|
10029 |
|
|
|
10030 |
|
|
signbit_call = fold_build2_loc (loc, NE_EXPR, integer_type_node,
|
10031 |
|
|
signbit_call, integer_zero_node);
|
10032 |
|
|
isinf_call = fold_build2_loc (loc, NE_EXPR, integer_type_node,
|
10033 |
|
|
isinf_call, integer_zero_node);
|
10034 |
|
|
|
10035 |
|
|
tmp = fold_build3_loc (loc, COND_EXPR, integer_type_node, signbit_call,
|
10036 |
|
|
integer_minus_one_node, integer_one_node);
|
10037 |
|
|
tmp = fold_build3_loc (loc, COND_EXPR, integer_type_node,
|
10038 |
|
|
isinf_call, tmp,
|
10039 |
|
|
integer_zero_node);
|
10040 |
|
|
}
|
10041 |
|
|
|
10042 |
|
|
return tmp;
|
10043 |
|
|
}
|
10044 |
|
|
|
10045 |
|
|
case BUILT_IN_ISFINITE:
|
10046 |
|
|
if (!HONOR_NANS (TYPE_MODE (TREE_TYPE (arg)))
|
10047 |
|
|
&& !HONOR_INFINITIES (TYPE_MODE (TREE_TYPE (arg))))
|
10048 |
|
|
return omit_one_operand_loc (loc, type, integer_one_node, arg);
|
10049 |
|
|
|
10050 |
|
|
if (TREE_CODE (arg) == REAL_CST)
|
10051 |
|
|
{
|
10052 |
|
|
r = TREE_REAL_CST (arg);
|
10053 |
|
|
return real_isfinite (&r) ? integer_one_node : integer_zero_node;
|
10054 |
|
|
}
|
10055 |
|
|
|
10056 |
|
|
return NULL_TREE;
|
10057 |
|
|
|
10058 |
|
|
case BUILT_IN_ISNAN:
|
10059 |
|
|
if (!HONOR_NANS (TYPE_MODE (TREE_TYPE (arg))))
|
10060 |
|
|
return omit_one_operand_loc (loc, type, integer_zero_node, arg);
|
10061 |
|
|
|
10062 |
|
|
if (TREE_CODE (arg) == REAL_CST)
|
10063 |
|
|
{
|
10064 |
|
|
r = TREE_REAL_CST (arg);
|
10065 |
|
|
return real_isnan (&r) ? integer_one_node : integer_zero_node;
|
10066 |
|
|
}
|
10067 |
|
|
|
10068 |
|
|
arg = builtin_save_expr (arg);
|
10069 |
|
|
return fold_build2_loc (loc, UNORDERED_EXPR, type, arg, arg);
|
10070 |
|
|
|
10071 |
|
|
default:
|
10072 |
|
|
gcc_unreachable ();
|
10073 |
|
|
}
|
10074 |
|
|
}
|
10075 |
|
|
|
10076 |
|
|
/* Fold a call to __builtin_fpclassify(int, int, int, int, int, ...).
|
10077 |
|
|
This builtin will generate code to return the appropriate floating
|
10078 |
|
|
point classification depending on the value of the floating point
|
10079 |
|
|
number passed in. The possible return values must be supplied as
|
10080 |
|
|
int arguments to the call in the following order: FP_NAN, FP_INFINITE,
|
10081 |
|
|
FP_NORMAL, FP_SUBNORMAL and FP_ZERO. The ellipses is for exactly
|
10082 |
|
|
one floating point argument which is "type generic". */
|
10083 |
|
|
|
10084 |
|
|
static tree
|
10085 |
|
|
fold_builtin_fpclassify (location_t loc, tree exp)
|
10086 |
|
|
{
|
10087 |
|
|
tree fp_nan, fp_infinite, fp_normal, fp_subnormal, fp_zero,
|
10088 |
|
|
arg, type, res, tmp;
|
10089 |
|
|
enum machine_mode mode;
|
10090 |
|
|
REAL_VALUE_TYPE r;
|
10091 |
|
|
char buf[128];
|
10092 |
|
|
|
10093 |
|
|
/* Verify the required arguments in the original call. */
|
10094 |
|
|
if (!validate_arglist (exp, INTEGER_TYPE, INTEGER_TYPE,
|
10095 |
|
|
INTEGER_TYPE, INTEGER_TYPE,
|
10096 |
|
|
INTEGER_TYPE, REAL_TYPE, VOID_TYPE))
|
10097 |
|
|
return NULL_TREE;
|
10098 |
|
|
|
10099 |
|
|
fp_nan = CALL_EXPR_ARG (exp, 0);
|
10100 |
|
|
fp_infinite = CALL_EXPR_ARG (exp, 1);
|
10101 |
|
|
fp_normal = CALL_EXPR_ARG (exp, 2);
|
10102 |
|
|
fp_subnormal = CALL_EXPR_ARG (exp, 3);
|
10103 |
|
|
fp_zero = CALL_EXPR_ARG (exp, 4);
|
10104 |
|
|
arg = CALL_EXPR_ARG (exp, 5);
|
10105 |
|
|
type = TREE_TYPE (arg);
|
10106 |
|
|
mode = TYPE_MODE (type);
|
10107 |
|
|
arg = builtin_save_expr (fold_build1_loc (loc, ABS_EXPR, type, arg));
|
10108 |
|
|
|
10109 |
|
|
/* fpclassify(x) ->
|
10110 |
|
|
isnan(x) ? FP_NAN :
|
10111 |
|
|
(fabs(x) == Inf ? FP_INFINITE :
|
10112 |
|
|
(fabs(x) >= DBL_MIN ? FP_NORMAL :
|
10113 |
|
|
(x == 0 ? FP_ZERO : FP_SUBNORMAL))). */
|
10114 |
|
|
|
10115 |
|
|
tmp = fold_build2_loc (loc, EQ_EXPR, integer_type_node, arg,
|
10116 |
|
|
build_real (type, dconst0));
|
10117 |
|
|
res = fold_build3_loc (loc, COND_EXPR, integer_type_node,
|
10118 |
|
|
tmp, fp_zero, fp_subnormal);
|
10119 |
|
|
|
10120 |
|
|
sprintf (buf, "0x1p%d", REAL_MODE_FORMAT (mode)->emin - 1);
|
10121 |
|
|
real_from_string (&r, buf);
|
10122 |
|
|
tmp = fold_build2_loc (loc, GE_EXPR, integer_type_node,
|
10123 |
|
|
arg, build_real (type, r));
|
10124 |
|
|
res = fold_build3_loc (loc, COND_EXPR, integer_type_node, tmp, fp_normal, res);
|
10125 |
|
|
|
10126 |
|
|
if (HONOR_INFINITIES (mode))
|
10127 |
|
|
{
|
10128 |
|
|
real_inf (&r);
|
10129 |
|
|
tmp = fold_build2_loc (loc, EQ_EXPR, integer_type_node, arg,
|
10130 |
|
|
build_real (type, r));
|
10131 |
|
|
res = fold_build3_loc (loc, COND_EXPR, integer_type_node, tmp,
|
10132 |
|
|
fp_infinite, res);
|
10133 |
|
|
}
|
10134 |
|
|
|
10135 |
|
|
if (HONOR_NANS (mode))
|
10136 |
|
|
{
|
10137 |
|
|
tmp = fold_build2_loc (loc, ORDERED_EXPR, integer_type_node, arg, arg);
|
10138 |
|
|
res = fold_build3_loc (loc, COND_EXPR, integer_type_node, tmp, res, fp_nan);
|
10139 |
|
|
}
|
10140 |
|
|
|
10141 |
|
|
return res;
|
10142 |
|
|
}
|
10143 |
|
|
|
10144 |
|
|
/* Fold a call to an unordered comparison function such as
|
10145 |
|
|
__builtin_isgreater(). FNDECL is the FUNCTION_DECL for the function
|
10146 |
|
|
being called and ARG0 and ARG1 are the arguments for the call.
|
10147 |
|
|
UNORDERED_CODE and ORDERED_CODE are comparison codes that give
|
10148 |
|
|
the opposite of the desired result. UNORDERED_CODE is used
|
10149 |
|
|
for modes that can hold NaNs and ORDERED_CODE is used for
|
10150 |
|
|
the rest. */
|
10151 |
|
|
|
10152 |
|
|
static tree
|
10153 |
|
|
fold_builtin_unordered_cmp (location_t loc, tree fndecl, tree arg0, tree arg1,
|
10154 |
|
|
enum tree_code unordered_code,
|
10155 |
|
|
enum tree_code ordered_code)
|
10156 |
|
|
{
|
10157 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
10158 |
|
|
enum tree_code code;
|
10159 |
|
|
tree type0, type1;
|
10160 |
|
|
enum tree_code code0, code1;
|
10161 |
|
|
tree cmp_type = NULL_TREE;
|
10162 |
|
|
|
10163 |
|
|
type0 = TREE_TYPE (arg0);
|
10164 |
|
|
type1 = TREE_TYPE (arg1);
|
10165 |
|
|
|
10166 |
|
|
code0 = TREE_CODE (type0);
|
10167 |
|
|
code1 = TREE_CODE (type1);
|
10168 |
|
|
|
10169 |
|
|
if (code0 == REAL_TYPE && code1 == REAL_TYPE)
|
10170 |
|
|
/* Choose the wider of two real types. */
|
10171 |
|
|
cmp_type = TYPE_PRECISION (type0) >= TYPE_PRECISION (type1)
|
10172 |
|
|
? type0 : type1;
|
10173 |
|
|
else if (code0 == REAL_TYPE && code1 == INTEGER_TYPE)
|
10174 |
|
|
cmp_type = type0;
|
10175 |
|
|
else if (code0 == INTEGER_TYPE && code1 == REAL_TYPE)
|
10176 |
|
|
cmp_type = type1;
|
10177 |
|
|
|
10178 |
|
|
arg0 = fold_convert_loc (loc, cmp_type, arg0);
|
10179 |
|
|
arg1 = fold_convert_loc (loc, cmp_type, arg1);
|
10180 |
|
|
|
10181 |
|
|
if (unordered_code == UNORDERED_EXPR)
|
10182 |
|
|
{
|
10183 |
|
|
if (!HONOR_NANS (TYPE_MODE (TREE_TYPE (arg0))))
|
10184 |
|
|
return omit_two_operands_loc (loc, type, integer_zero_node, arg0, arg1);
|
10185 |
|
|
return fold_build2_loc (loc, UNORDERED_EXPR, type, arg0, arg1);
|
10186 |
|
|
}
|
10187 |
|
|
|
10188 |
|
|
code = HONOR_NANS (TYPE_MODE (TREE_TYPE (arg0))) ? unordered_code
|
10189 |
|
|
: ordered_code;
|
10190 |
|
|
return fold_build1_loc (loc, TRUTH_NOT_EXPR, type,
|
10191 |
|
|
fold_build2_loc (loc, code, type, arg0, arg1));
|
10192 |
|
|
}
|
10193 |
|
|
|
10194 |
|
|
/* Fold a call to built-in function FNDECL with 0 arguments.
|
10195 |
|
|
IGNORE is true if the result of the function call is ignored. This
|
10196 |
|
|
function returns NULL_TREE if no simplification was possible. */
|
10197 |
|
|
|
10198 |
|
|
static tree
|
10199 |
|
|
fold_builtin_0 (location_t loc, tree fndecl, bool ignore ATTRIBUTE_UNUSED)
|
10200 |
|
|
{
|
10201 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
10202 |
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
10203 |
|
|
switch (fcode)
|
10204 |
|
|
{
|
10205 |
|
|
CASE_FLT_FN (BUILT_IN_INF):
|
10206 |
|
|
case BUILT_IN_INFD32:
|
10207 |
|
|
case BUILT_IN_INFD64:
|
10208 |
|
|
case BUILT_IN_INFD128:
|
10209 |
|
|
return fold_builtin_inf (loc, type, true);
|
10210 |
|
|
|
10211 |
|
|
CASE_FLT_FN (BUILT_IN_HUGE_VAL):
|
10212 |
|
|
return fold_builtin_inf (loc, type, false);
|
10213 |
|
|
|
10214 |
|
|
case BUILT_IN_CLASSIFY_TYPE:
|
10215 |
|
|
return fold_builtin_classify_type (NULL_TREE);
|
10216 |
|
|
|
10217 |
|
|
default:
|
10218 |
|
|
break;
|
10219 |
|
|
}
|
10220 |
|
|
return NULL_TREE;
|
10221 |
|
|
}
|
10222 |
|
|
|
10223 |
|
|
/* Fold a call to built-in function FNDECL with 1 argument, ARG0.
|
10224 |
|
|
IGNORE is true if the result of the function call is ignored. This
|
10225 |
|
|
function returns NULL_TREE if no simplification was possible. */
|
10226 |
|
|
|
10227 |
|
|
static tree
|
10228 |
|
|
fold_builtin_1 (location_t loc, tree fndecl, tree arg0, bool ignore)
|
10229 |
|
|
{
|
10230 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
10231 |
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
10232 |
|
|
switch (fcode)
|
10233 |
|
|
{
|
10234 |
|
|
case BUILT_IN_CONSTANT_P:
|
10235 |
|
|
{
|
10236 |
|
|
tree val = fold_builtin_constant_p (arg0);
|
10237 |
|
|
|
10238 |
|
|
/* Gimplification will pull the CALL_EXPR for the builtin out of
|
10239 |
|
|
an if condition. When not optimizing, we'll not CSE it back.
|
10240 |
|
|
To avoid link error types of regressions, return false now. */
|
10241 |
|
|
if (!val && !optimize)
|
10242 |
|
|
val = integer_zero_node;
|
10243 |
|
|
|
10244 |
|
|
return val;
|
10245 |
|
|
}
|
10246 |
|
|
|
10247 |
|
|
case BUILT_IN_CLASSIFY_TYPE:
|
10248 |
|
|
return fold_builtin_classify_type (arg0);
|
10249 |
|
|
|
10250 |
|
|
case BUILT_IN_STRLEN:
|
10251 |
|
|
return fold_builtin_strlen (loc, type, arg0);
|
10252 |
|
|
|
10253 |
|
|
CASE_FLT_FN (BUILT_IN_FABS):
|
10254 |
|
|
return fold_builtin_fabs (loc, arg0, type);
|
10255 |
|
|
|
10256 |
|
|
case BUILT_IN_ABS:
|
10257 |
|
|
case BUILT_IN_LABS:
|
10258 |
|
|
case BUILT_IN_LLABS:
|
10259 |
|
|
case BUILT_IN_IMAXABS:
|
10260 |
|
|
return fold_builtin_abs (loc, arg0, type);
|
10261 |
|
|
|
10262 |
|
|
CASE_FLT_FN (BUILT_IN_CONJ):
|
10263 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10264 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10265 |
|
|
return fold_build1_loc (loc, CONJ_EXPR, type, arg0);
|
10266 |
|
|
break;
|
10267 |
|
|
|
10268 |
|
|
CASE_FLT_FN (BUILT_IN_CREAL):
|
10269 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10270 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10271 |
|
|
return non_lvalue_loc (loc, fold_build1_loc (loc, REALPART_EXPR, type, arg0));;
|
10272 |
|
|
break;
|
10273 |
|
|
|
10274 |
|
|
CASE_FLT_FN (BUILT_IN_CIMAG):
|
10275 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10276 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10277 |
|
|
return non_lvalue_loc (loc, fold_build1_loc (loc, IMAGPART_EXPR, type, arg0));
|
10278 |
|
|
break;
|
10279 |
|
|
|
10280 |
|
|
CASE_FLT_FN (BUILT_IN_CCOS):
|
10281 |
|
|
return fold_builtin_ccos(loc, arg0, type, fndecl, /*hyper=*/ false);
|
10282 |
|
|
|
10283 |
|
|
CASE_FLT_FN (BUILT_IN_CCOSH):
|
10284 |
|
|
return fold_builtin_ccos(loc, arg0, type, fndecl, /*hyper=*/ true);
|
10285 |
|
|
|
10286 |
|
|
CASE_FLT_FN (BUILT_IN_CPROJ):
|
10287 |
|
|
return fold_builtin_cproj(loc, arg0, type);
|
10288 |
|
|
|
10289 |
|
|
CASE_FLT_FN (BUILT_IN_CSIN):
|
10290 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10291 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10292 |
|
|
return do_mpc_arg1 (arg0, type, mpc_sin);
|
10293 |
|
|
break;
|
10294 |
|
|
|
10295 |
|
|
CASE_FLT_FN (BUILT_IN_CSINH):
|
10296 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10297 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10298 |
|
|
return do_mpc_arg1 (arg0, type, mpc_sinh);
|
10299 |
|
|
break;
|
10300 |
|
|
|
10301 |
|
|
CASE_FLT_FN (BUILT_IN_CTAN):
|
10302 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10303 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10304 |
|
|
return do_mpc_arg1 (arg0, type, mpc_tan);
|
10305 |
|
|
break;
|
10306 |
|
|
|
10307 |
|
|
CASE_FLT_FN (BUILT_IN_CTANH):
|
10308 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10309 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10310 |
|
|
return do_mpc_arg1 (arg0, type, mpc_tanh);
|
10311 |
|
|
break;
|
10312 |
|
|
|
10313 |
|
|
CASE_FLT_FN (BUILT_IN_CLOG):
|
10314 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10315 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10316 |
|
|
return do_mpc_arg1 (arg0, type, mpc_log);
|
10317 |
|
|
break;
|
10318 |
|
|
|
10319 |
|
|
CASE_FLT_FN (BUILT_IN_CSQRT):
|
10320 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10321 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10322 |
|
|
return do_mpc_arg1 (arg0, type, mpc_sqrt);
|
10323 |
|
|
break;
|
10324 |
|
|
|
10325 |
|
|
CASE_FLT_FN (BUILT_IN_CASIN):
|
10326 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10327 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10328 |
|
|
return do_mpc_arg1 (arg0, type, mpc_asin);
|
10329 |
|
|
break;
|
10330 |
|
|
|
10331 |
|
|
CASE_FLT_FN (BUILT_IN_CACOS):
|
10332 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10333 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10334 |
|
|
return do_mpc_arg1 (arg0, type, mpc_acos);
|
10335 |
|
|
break;
|
10336 |
|
|
|
10337 |
|
|
CASE_FLT_FN (BUILT_IN_CATAN):
|
10338 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10339 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10340 |
|
|
return do_mpc_arg1 (arg0, type, mpc_atan);
|
10341 |
|
|
break;
|
10342 |
|
|
|
10343 |
|
|
CASE_FLT_FN (BUILT_IN_CASINH):
|
10344 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10345 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10346 |
|
|
return do_mpc_arg1 (arg0, type, mpc_asinh);
|
10347 |
|
|
break;
|
10348 |
|
|
|
10349 |
|
|
CASE_FLT_FN (BUILT_IN_CACOSH):
|
10350 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10351 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10352 |
|
|
return do_mpc_arg1 (arg0, type, mpc_acosh);
|
10353 |
|
|
break;
|
10354 |
|
|
|
10355 |
|
|
CASE_FLT_FN (BUILT_IN_CATANH):
|
10356 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10357 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE)
|
10358 |
|
|
return do_mpc_arg1 (arg0, type, mpc_atanh);
|
10359 |
|
|
break;
|
10360 |
|
|
|
10361 |
|
|
CASE_FLT_FN (BUILT_IN_CABS):
|
10362 |
|
|
return fold_builtin_cabs (loc, arg0, type, fndecl);
|
10363 |
|
|
|
10364 |
|
|
CASE_FLT_FN (BUILT_IN_CARG):
|
10365 |
|
|
return fold_builtin_carg (loc, arg0, type);
|
10366 |
|
|
|
10367 |
|
|
CASE_FLT_FN (BUILT_IN_SQRT):
|
10368 |
|
|
return fold_builtin_sqrt (loc, arg0, type);
|
10369 |
|
|
|
10370 |
|
|
CASE_FLT_FN (BUILT_IN_CBRT):
|
10371 |
|
|
return fold_builtin_cbrt (loc, arg0, type);
|
10372 |
|
|
|
10373 |
|
|
CASE_FLT_FN (BUILT_IN_ASIN):
|
10374 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10375 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_asin,
|
10376 |
|
|
&dconstm1, &dconst1, true);
|
10377 |
|
|
break;
|
10378 |
|
|
|
10379 |
|
|
CASE_FLT_FN (BUILT_IN_ACOS):
|
10380 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10381 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_acos,
|
10382 |
|
|
&dconstm1, &dconst1, true);
|
10383 |
|
|
break;
|
10384 |
|
|
|
10385 |
|
|
CASE_FLT_FN (BUILT_IN_ATAN):
|
10386 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10387 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_atan, NULL, NULL, 0);
|
10388 |
|
|
break;
|
10389 |
|
|
|
10390 |
|
|
CASE_FLT_FN (BUILT_IN_ASINH):
|
10391 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10392 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_asinh, NULL, NULL, 0);
|
10393 |
|
|
break;
|
10394 |
|
|
|
10395 |
|
|
CASE_FLT_FN (BUILT_IN_ACOSH):
|
10396 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10397 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_acosh,
|
10398 |
|
|
&dconst1, NULL, true);
|
10399 |
|
|
break;
|
10400 |
|
|
|
10401 |
|
|
CASE_FLT_FN (BUILT_IN_ATANH):
|
10402 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10403 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_atanh,
|
10404 |
|
|
&dconstm1, &dconst1, false);
|
10405 |
|
|
break;
|
10406 |
|
|
|
10407 |
|
|
CASE_FLT_FN (BUILT_IN_SIN):
|
10408 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10409 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_sin, NULL, NULL, 0);
|
10410 |
|
|
break;
|
10411 |
|
|
|
10412 |
|
|
CASE_FLT_FN (BUILT_IN_COS):
|
10413 |
|
|
return fold_builtin_cos (loc, arg0, type, fndecl);
|
10414 |
|
|
|
10415 |
|
|
CASE_FLT_FN (BUILT_IN_TAN):
|
10416 |
|
|
return fold_builtin_tan (arg0, type);
|
10417 |
|
|
|
10418 |
|
|
CASE_FLT_FN (BUILT_IN_CEXP):
|
10419 |
|
|
return fold_builtin_cexp (loc, arg0, type);
|
10420 |
|
|
|
10421 |
|
|
CASE_FLT_FN (BUILT_IN_CEXPI):
|
10422 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10423 |
|
|
return do_mpfr_sincos (arg0, NULL_TREE, NULL_TREE);
|
10424 |
|
|
break;
|
10425 |
|
|
|
10426 |
|
|
CASE_FLT_FN (BUILT_IN_SINH):
|
10427 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10428 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_sinh, NULL, NULL, 0);
|
10429 |
|
|
break;
|
10430 |
|
|
|
10431 |
|
|
CASE_FLT_FN (BUILT_IN_COSH):
|
10432 |
|
|
return fold_builtin_cosh (loc, arg0, type, fndecl);
|
10433 |
|
|
|
10434 |
|
|
CASE_FLT_FN (BUILT_IN_TANH):
|
10435 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10436 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_tanh, NULL, NULL, 0);
|
10437 |
|
|
break;
|
10438 |
|
|
|
10439 |
|
|
CASE_FLT_FN (BUILT_IN_ERF):
|
10440 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10441 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_erf, NULL, NULL, 0);
|
10442 |
|
|
break;
|
10443 |
|
|
|
10444 |
|
|
CASE_FLT_FN (BUILT_IN_ERFC):
|
10445 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10446 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_erfc, NULL, NULL, 0);
|
10447 |
|
|
break;
|
10448 |
|
|
|
10449 |
|
|
CASE_FLT_FN (BUILT_IN_TGAMMA):
|
10450 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10451 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_gamma, NULL, NULL, 0);
|
10452 |
|
|
break;
|
10453 |
|
|
|
10454 |
|
|
CASE_FLT_FN (BUILT_IN_EXP):
|
10455 |
|
|
return fold_builtin_exponent (loc, fndecl, arg0, mpfr_exp);
|
10456 |
|
|
|
10457 |
|
|
CASE_FLT_FN (BUILT_IN_EXP2):
|
10458 |
|
|
return fold_builtin_exponent (loc, fndecl, arg0, mpfr_exp2);
|
10459 |
|
|
|
10460 |
|
|
CASE_FLT_FN (BUILT_IN_EXP10):
|
10461 |
|
|
CASE_FLT_FN (BUILT_IN_POW10):
|
10462 |
|
|
return fold_builtin_exponent (loc, fndecl, arg0, mpfr_exp10);
|
10463 |
|
|
|
10464 |
|
|
CASE_FLT_FN (BUILT_IN_EXPM1):
|
10465 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10466 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_expm1, NULL, NULL, 0);
|
10467 |
|
|
break;
|
10468 |
|
|
|
10469 |
|
|
CASE_FLT_FN (BUILT_IN_LOG):
|
10470 |
|
|
return fold_builtin_logarithm (loc, fndecl, arg0, mpfr_log);
|
10471 |
|
|
|
10472 |
|
|
CASE_FLT_FN (BUILT_IN_LOG2):
|
10473 |
|
|
return fold_builtin_logarithm (loc, fndecl, arg0, mpfr_log2);
|
10474 |
|
|
|
10475 |
|
|
CASE_FLT_FN (BUILT_IN_LOG10):
|
10476 |
|
|
return fold_builtin_logarithm (loc, fndecl, arg0, mpfr_log10);
|
10477 |
|
|
|
10478 |
|
|
CASE_FLT_FN (BUILT_IN_LOG1P):
|
10479 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10480 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_log1p,
|
10481 |
|
|
&dconstm1, NULL, false);
|
10482 |
|
|
break;
|
10483 |
|
|
|
10484 |
|
|
CASE_FLT_FN (BUILT_IN_J0):
|
10485 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10486 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_j0,
|
10487 |
|
|
NULL, NULL, 0);
|
10488 |
|
|
break;
|
10489 |
|
|
|
10490 |
|
|
CASE_FLT_FN (BUILT_IN_J1):
|
10491 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10492 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_j1,
|
10493 |
|
|
NULL, NULL, 0);
|
10494 |
|
|
break;
|
10495 |
|
|
|
10496 |
|
|
CASE_FLT_FN (BUILT_IN_Y0):
|
10497 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10498 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_y0,
|
10499 |
|
|
&dconst0, NULL, false);
|
10500 |
|
|
break;
|
10501 |
|
|
|
10502 |
|
|
CASE_FLT_FN (BUILT_IN_Y1):
|
10503 |
|
|
if (validate_arg (arg0, REAL_TYPE))
|
10504 |
|
|
return do_mpfr_arg1 (arg0, type, mpfr_y1,
|
10505 |
|
|
&dconst0, NULL, false);
|
10506 |
|
|
break;
|
10507 |
|
|
|
10508 |
|
|
CASE_FLT_FN (BUILT_IN_NAN):
|
10509 |
|
|
case BUILT_IN_NAND32:
|
10510 |
|
|
case BUILT_IN_NAND64:
|
10511 |
|
|
case BUILT_IN_NAND128:
|
10512 |
|
|
return fold_builtin_nan (arg0, type, true);
|
10513 |
|
|
|
10514 |
|
|
CASE_FLT_FN (BUILT_IN_NANS):
|
10515 |
|
|
return fold_builtin_nan (arg0, type, false);
|
10516 |
|
|
|
10517 |
|
|
CASE_FLT_FN (BUILT_IN_FLOOR):
|
10518 |
|
|
return fold_builtin_floor (loc, fndecl, arg0);
|
10519 |
|
|
|
10520 |
|
|
CASE_FLT_FN (BUILT_IN_CEIL):
|
10521 |
|
|
return fold_builtin_ceil (loc, fndecl, arg0);
|
10522 |
|
|
|
10523 |
|
|
CASE_FLT_FN (BUILT_IN_TRUNC):
|
10524 |
|
|
return fold_builtin_trunc (loc, fndecl, arg0);
|
10525 |
|
|
|
10526 |
|
|
CASE_FLT_FN (BUILT_IN_ROUND):
|
10527 |
|
|
return fold_builtin_round (loc, fndecl, arg0);
|
10528 |
|
|
|
10529 |
|
|
CASE_FLT_FN (BUILT_IN_NEARBYINT):
|
10530 |
|
|
CASE_FLT_FN (BUILT_IN_RINT):
|
10531 |
|
|
return fold_trunc_transparent_mathfn (loc, fndecl, arg0);
|
10532 |
|
|
|
10533 |
|
|
CASE_FLT_FN (BUILT_IN_ICEIL):
|
10534 |
|
|
CASE_FLT_FN (BUILT_IN_LCEIL):
|
10535 |
|
|
CASE_FLT_FN (BUILT_IN_LLCEIL):
|
10536 |
|
|
CASE_FLT_FN (BUILT_IN_LFLOOR):
|
10537 |
|
|
CASE_FLT_FN (BUILT_IN_IFLOOR):
|
10538 |
|
|
CASE_FLT_FN (BUILT_IN_LLFLOOR):
|
10539 |
|
|
CASE_FLT_FN (BUILT_IN_IROUND):
|
10540 |
|
|
CASE_FLT_FN (BUILT_IN_LROUND):
|
10541 |
|
|
CASE_FLT_FN (BUILT_IN_LLROUND):
|
10542 |
|
|
return fold_builtin_int_roundingfn (loc, fndecl, arg0);
|
10543 |
|
|
|
10544 |
|
|
CASE_FLT_FN (BUILT_IN_IRINT):
|
10545 |
|
|
CASE_FLT_FN (BUILT_IN_LRINT):
|
10546 |
|
|
CASE_FLT_FN (BUILT_IN_LLRINT):
|
10547 |
|
|
return fold_fixed_mathfn (loc, fndecl, arg0);
|
10548 |
|
|
|
10549 |
|
|
case BUILT_IN_BSWAP32:
|
10550 |
|
|
case BUILT_IN_BSWAP64:
|
10551 |
|
|
return fold_builtin_bswap (fndecl, arg0);
|
10552 |
|
|
|
10553 |
|
|
CASE_INT_FN (BUILT_IN_FFS):
|
10554 |
|
|
CASE_INT_FN (BUILT_IN_CLZ):
|
10555 |
|
|
CASE_INT_FN (BUILT_IN_CTZ):
|
10556 |
|
|
CASE_INT_FN (BUILT_IN_CLRSB):
|
10557 |
|
|
CASE_INT_FN (BUILT_IN_POPCOUNT):
|
10558 |
|
|
CASE_INT_FN (BUILT_IN_PARITY):
|
10559 |
|
|
return fold_builtin_bitop (fndecl, arg0);
|
10560 |
|
|
|
10561 |
|
|
CASE_FLT_FN (BUILT_IN_SIGNBIT):
|
10562 |
|
|
return fold_builtin_signbit (loc, arg0, type);
|
10563 |
|
|
|
10564 |
|
|
CASE_FLT_FN (BUILT_IN_SIGNIFICAND):
|
10565 |
|
|
return fold_builtin_significand (loc, arg0, type);
|
10566 |
|
|
|
10567 |
|
|
CASE_FLT_FN (BUILT_IN_ILOGB):
|
10568 |
|
|
CASE_FLT_FN (BUILT_IN_LOGB):
|
10569 |
|
|
return fold_builtin_logb (loc, arg0, type);
|
10570 |
|
|
|
10571 |
|
|
case BUILT_IN_ISASCII:
|
10572 |
|
|
return fold_builtin_isascii (loc, arg0);
|
10573 |
|
|
|
10574 |
|
|
case BUILT_IN_TOASCII:
|
10575 |
|
|
return fold_builtin_toascii (loc, arg0);
|
10576 |
|
|
|
10577 |
|
|
case BUILT_IN_ISDIGIT:
|
10578 |
|
|
return fold_builtin_isdigit (loc, arg0);
|
10579 |
|
|
|
10580 |
|
|
CASE_FLT_FN (BUILT_IN_FINITE):
|
10581 |
|
|
case BUILT_IN_FINITED32:
|
10582 |
|
|
case BUILT_IN_FINITED64:
|
10583 |
|
|
case BUILT_IN_FINITED128:
|
10584 |
|
|
case BUILT_IN_ISFINITE:
|
10585 |
|
|
{
|
10586 |
|
|
tree ret = fold_builtin_classify (loc, fndecl, arg0, BUILT_IN_ISFINITE);
|
10587 |
|
|
if (ret)
|
10588 |
|
|
return ret;
|
10589 |
|
|
return fold_builtin_interclass_mathfn (loc, fndecl, arg0);
|
10590 |
|
|
}
|
10591 |
|
|
|
10592 |
|
|
CASE_FLT_FN (BUILT_IN_ISINF):
|
10593 |
|
|
case BUILT_IN_ISINFD32:
|
10594 |
|
|
case BUILT_IN_ISINFD64:
|
10595 |
|
|
case BUILT_IN_ISINFD128:
|
10596 |
|
|
{
|
10597 |
|
|
tree ret = fold_builtin_classify (loc, fndecl, arg0, BUILT_IN_ISINF);
|
10598 |
|
|
if (ret)
|
10599 |
|
|
return ret;
|
10600 |
|
|
return fold_builtin_interclass_mathfn (loc, fndecl, arg0);
|
10601 |
|
|
}
|
10602 |
|
|
|
10603 |
|
|
case BUILT_IN_ISNORMAL:
|
10604 |
|
|
return fold_builtin_interclass_mathfn (loc, fndecl, arg0);
|
10605 |
|
|
|
10606 |
|
|
case BUILT_IN_ISINF_SIGN:
|
10607 |
|
|
return fold_builtin_classify (loc, fndecl, arg0, BUILT_IN_ISINF_SIGN);
|
10608 |
|
|
|
10609 |
|
|
CASE_FLT_FN (BUILT_IN_ISNAN):
|
10610 |
|
|
case BUILT_IN_ISNAND32:
|
10611 |
|
|
case BUILT_IN_ISNAND64:
|
10612 |
|
|
case BUILT_IN_ISNAND128:
|
10613 |
|
|
return fold_builtin_classify (loc, fndecl, arg0, BUILT_IN_ISNAN);
|
10614 |
|
|
|
10615 |
|
|
case BUILT_IN_PRINTF:
|
10616 |
|
|
case BUILT_IN_PRINTF_UNLOCKED:
|
10617 |
|
|
case BUILT_IN_VPRINTF:
|
10618 |
|
|
return fold_builtin_printf (loc, fndecl, arg0, NULL_TREE, ignore, fcode);
|
10619 |
|
|
|
10620 |
|
|
case BUILT_IN_FREE:
|
10621 |
|
|
if (integer_zerop (arg0))
|
10622 |
|
|
return build_empty_stmt (loc);
|
10623 |
|
|
break;
|
10624 |
|
|
|
10625 |
|
|
default:
|
10626 |
|
|
break;
|
10627 |
|
|
}
|
10628 |
|
|
|
10629 |
|
|
return NULL_TREE;
|
10630 |
|
|
|
10631 |
|
|
}
|
10632 |
|
|
|
10633 |
|
|
/* Fold a call to built-in function FNDECL with 2 arguments, ARG0 and ARG1.
|
10634 |
|
|
IGNORE is true if the result of the function call is ignored. This
|
10635 |
|
|
function returns NULL_TREE if no simplification was possible. */
|
10636 |
|
|
|
10637 |
|
|
static tree
|
10638 |
|
|
fold_builtin_2 (location_t loc, tree fndecl, tree arg0, tree arg1, bool ignore)
|
10639 |
|
|
{
|
10640 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
10641 |
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
10642 |
|
|
|
10643 |
|
|
switch (fcode)
|
10644 |
|
|
{
|
10645 |
|
|
CASE_FLT_FN (BUILT_IN_JN):
|
10646 |
|
|
if (validate_arg (arg0, INTEGER_TYPE)
|
10647 |
|
|
&& validate_arg (arg1, REAL_TYPE))
|
10648 |
|
|
return do_mpfr_bessel_n (arg0, arg1, type, mpfr_jn, NULL, 0);
|
10649 |
|
|
break;
|
10650 |
|
|
|
10651 |
|
|
CASE_FLT_FN (BUILT_IN_YN):
|
10652 |
|
|
if (validate_arg (arg0, INTEGER_TYPE)
|
10653 |
|
|
&& validate_arg (arg1, REAL_TYPE))
|
10654 |
|
|
return do_mpfr_bessel_n (arg0, arg1, type, mpfr_yn,
|
10655 |
|
|
&dconst0, false);
|
10656 |
|
|
break;
|
10657 |
|
|
|
10658 |
|
|
CASE_FLT_FN (BUILT_IN_DREM):
|
10659 |
|
|
CASE_FLT_FN (BUILT_IN_REMAINDER):
|
10660 |
|
|
if (validate_arg (arg0, REAL_TYPE)
|
10661 |
|
|
&& validate_arg(arg1, REAL_TYPE))
|
10662 |
|
|
return do_mpfr_arg2 (arg0, arg1, type, mpfr_remainder);
|
10663 |
|
|
break;
|
10664 |
|
|
|
10665 |
|
|
CASE_FLT_FN_REENT (BUILT_IN_GAMMA): /* GAMMA_R */
|
10666 |
|
|
CASE_FLT_FN_REENT (BUILT_IN_LGAMMA): /* LGAMMA_R */
|
10667 |
|
|
if (validate_arg (arg0, REAL_TYPE)
|
10668 |
|
|
&& validate_arg(arg1, POINTER_TYPE))
|
10669 |
|
|
return do_mpfr_lgamma_r (arg0, arg1, type);
|
10670 |
|
|
break;
|
10671 |
|
|
|
10672 |
|
|
CASE_FLT_FN (BUILT_IN_ATAN2):
|
10673 |
|
|
if (validate_arg (arg0, REAL_TYPE)
|
10674 |
|
|
&& validate_arg(arg1, REAL_TYPE))
|
10675 |
|
|
return do_mpfr_arg2 (arg0, arg1, type, mpfr_atan2);
|
10676 |
|
|
break;
|
10677 |
|
|
|
10678 |
|
|
CASE_FLT_FN (BUILT_IN_FDIM):
|
10679 |
|
|
if (validate_arg (arg0, REAL_TYPE)
|
10680 |
|
|
&& validate_arg(arg1, REAL_TYPE))
|
10681 |
|
|
return do_mpfr_arg2 (arg0, arg1, type, mpfr_dim);
|
10682 |
|
|
break;
|
10683 |
|
|
|
10684 |
|
|
CASE_FLT_FN (BUILT_IN_HYPOT):
|
10685 |
|
|
return fold_builtin_hypot (loc, fndecl, arg0, arg1, type);
|
10686 |
|
|
|
10687 |
|
|
CASE_FLT_FN (BUILT_IN_CPOW):
|
10688 |
|
|
if (validate_arg (arg0, COMPLEX_TYPE)
|
10689 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE
|
10690 |
|
|
&& validate_arg (arg1, COMPLEX_TYPE)
|
10691 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg1))) == REAL_TYPE)
|
10692 |
|
|
return do_mpc_arg2 (arg0, arg1, type, /*do_nonfinite=*/ 0, mpc_pow);
|
10693 |
|
|
break;
|
10694 |
|
|
|
10695 |
|
|
CASE_FLT_FN (BUILT_IN_LDEXP):
|
10696 |
|
|
return fold_builtin_load_exponent (loc, arg0, arg1, type, /*ldexp=*/true);
|
10697 |
|
|
CASE_FLT_FN (BUILT_IN_SCALBN):
|
10698 |
|
|
CASE_FLT_FN (BUILT_IN_SCALBLN):
|
10699 |
|
|
return fold_builtin_load_exponent (loc, arg0, arg1,
|
10700 |
|
|
type, /*ldexp=*/false);
|
10701 |
|
|
|
10702 |
|
|
CASE_FLT_FN (BUILT_IN_FREXP):
|
10703 |
|
|
return fold_builtin_frexp (loc, arg0, arg1, type);
|
10704 |
|
|
|
10705 |
|
|
CASE_FLT_FN (BUILT_IN_MODF):
|
10706 |
|
|
return fold_builtin_modf (loc, arg0, arg1, type);
|
10707 |
|
|
|
10708 |
|
|
case BUILT_IN_BZERO:
|
10709 |
|
|
return fold_builtin_bzero (loc, arg0, arg1, ignore);
|
10710 |
|
|
|
10711 |
|
|
case BUILT_IN_FPUTS:
|
10712 |
|
|
return fold_builtin_fputs (loc, arg0, arg1, ignore, false, NULL_TREE);
|
10713 |
|
|
|
10714 |
|
|
case BUILT_IN_FPUTS_UNLOCKED:
|
10715 |
|
|
return fold_builtin_fputs (loc, arg0, arg1, ignore, true, NULL_TREE);
|
10716 |
|
|
|
10717 |
|
|
case BUILT_IN_STRSTR:
|
10718 |
|
|
return fold_builtin_strstr (loc, arg0, arg1, type);
|
10719 |
|
|
|
10720 |
|
|
case BUILT_IN_STRCAT:
|
10721 |
|
|
return fold_builtin_strcat (loc, arg0, arg1);
|
10722 |
|
|
|
10723 |
|
|
case BUILT_IN_STRSPN:
|
10724 |
|
|
return fold_builtin_strspn (loc, arg0, arg1);
|
10725 |
|
|
|
10726 |
|
|
case BUILT_IN_STRCSPN:
|
10727 |
|
|
return fold_builtin_strcspn (loc, arg0, arg1);
|
10728 |
|
|
|
10729 |
|
|
case BUILT_IN_STRCHR:
|
10730 |
|
|
case BUILT_IN_INDEX:
|
10731 |
|
|
return fold_builtin_strchr (loc, arg0, arg1, type);
|
10732 |
|
|
|
10733 |
|
|
case BUILT_IN_STRRCHR:
|
10734 |
|
|
case BUILT_IN_RINDEX:
|
10735 |
|
|
return fold_builtin_strrchr (loc, arg0, arg1, type);
|
10736 |
|
|
|
10737 |
|
|
case BUILT_IN_STRCPY:
|
10738 |
|
|
return fold_builtin_strcpy (loc, fndecl, arg0, arg1, NULL_TREE);
|
10739 |
|
|
|
10740 |
|
|
case BUILT_IN_STPCPY:
|
10741 |
|
|
if (ignore)
|
10742 |
|
|
{
|
10743 |
|
|
tree fn = builtin_decl_implicit (BUILT_IN_STRCPY);
|
10744 |
|
|
if (!fn)
|
10745 |
|
|
break;
|
10746 |
|
|
|
10747 |
|
|
return build_call_expr_loc (loc, fn, 2, arg0, arg1);
|
10748 |
|
|
}
|
10749 |
|
|
else
|
10750 |
|
|
return fold_builtin_stpcpy (loc, fndecl, arg0, arg1);
|
10751 |
|
|
break;
|
10752 |
|
|
|
10753 |
|
|
case BUILT_IN_STRCMP:
|
10754 |
|
|
return fold_builtin_strcmp (loc, arg0, arg1);
|
10755 |
|
|
|
10756 |
|
|
case BUILT_IN_STRPBRK:
|
10757 |
|
|
return fold_builtin_strpbrk (loc, arg0, arg1, type);
|
10758 |
|
|
|
10759 |
|
|
case BUILT_IN_EXPECT:
|
10760 |
|
|
return fold_builtin_expect (loc, arg0, arg1);
|
10761 |
|
|
|
10762 |
|
|
CASE_FLT_FN (BUILT_IN_POW):
|
10763 |
|
|
return fold_builtin_pow (loc, fndecl, arg0, arg1, type);
|
10764 |
|
|
|
10765 |
|
|
CASE_FLT_FN (BUILT_IN_POWI):
|
10766 |
|
|
return fold_builtin_powi (loc, fndecl, arg0, arg1, type);
|
10767 |
|
|
|
10768 |
|
|
CASE_FLT_FN (BUILT_IN_COPYSIGN):
|
10769 |
|
|
return fold_builtin_copysign (loc, fndecl, arg0, arg1, type);
|
10770 |
|
|
|
10771 |
|
|
CASE_FLT_FN (BUILT_IN_FMIN):
|
10772 |
|
|
return fold_builtin_fmin_fmax (loc, arg0, arg1, type, /*max=*/false);
|
10773 |
|
|
|
10774 |
|
|
CASE_FLT_FN (BUILT_IN_FMAX):
|
10775 |
|
|
return fold_builtin_fmin_fmax (loc, arg0, arg1, type, /*max=*/true);
|
10776 |
|
|
|
10777 |
|
|
case BUILT_IN_ISGREATER:
|
10778 |
|
|
return fold_builtin_unordered_cmp (loc, fndecl,
|
10779 |
|
|
arg0, arg1, UNLE_EXPR, LE_EXPR);
|
10780 |
|
|
case BUILT_IN_ISGREATEREQUAL:
|
10781 |
|
|
return fold_builtin_unordered_cmp (loc, fndecl,
|
10782 |
|
|
arg0, arg1, UNLT_EXPR, LT_EXPR);
|
10783 |
|
|
case BUILT_IN_ISLESS:
|
10784 |
|
|
return fold_builtin_unordered_cmp (loc, fndecl,
|
10785 |
|
|
arg0, arg1, UNGE_EXPR, GE_EXPR);
|
10786 |
|
|
case BUILT_IN_ISLESSEQUAL:
|
10787 |
|
|
return fold_builtin_unordered_cmp (loc, fndecl,
|
10788 |
|
|
arg0, arg1, UNGT_EXPR, GT_EXPR);
|
10789 |
|
|
case BUILT_IN_ISLESSGREATER:
|
10790 |
|
|
return fold_builtin_unordered_cmp (loc, fndecl,
|
10791 |
|
|
arg0, arg1, UNEQ_EXPR, EQ_EXPR);
|
10792 |
|
|
case BUILT_IN_ISUNORDERED:
|
10793 |
|
|
return fold_builtin_unordered_cmp (loc, fndecl,
|
10794 |
|
|
arg0, arg1, UNORDERED_EXPR,
|
10795 |
|
|
NOP_EXPR);
|
10796 |
|
|
|
10797 |
|
|
/* We do the folding for va_start in the expander. */
|
10798 |
|
|
case BUILT_IN_VA_START:
|
10799 |
|
|
break;
|
10800 |
|
|
|
10801 |
|
|
case BUILT_IN_SPRINTF:
|
10802 |
|
|
return fold_builtin_sprintf (loc, arg0, arg1, NULL_TREE, ignore);
|
10803 |
|
|
|
10804 |
|
|
case BUILT_IN_OBJECT_SIZE:
|
10805 |
|
|
return fold_builtin_object_size (arg0, arg1);
|
10806 |
|
|
|
10807 |
|
|
case BUILT_IN_PRINTF:
|
10808 |
|
|
case BUILT_IN_PRINTF_UNLOCKED:
|
10809 |
|
|
case BUILT_IN_VPRINTF:
|
10810 |
|
|
return fold_builtin_printf (loc, fndecl, arg0, arg1, ignore, fcode);
|
10811 |
|
|
|
10812 |
|
|
case BUILT_IN_PRINTF_CHK:
|
10813 |
|
|
case BUILT_IN_VPRINTF_CHK:
|
10814 |
|
|
if (!validate_arg (arg0, INTEGER_TYPE)
|
10815 |
|
|
|| TREE_SIDE_EFFECTS (arg0))
|
10816 |
|
|
return NULL_TREE;
|
10817 |
|
|
else
|
10818 |
|
|
return fold_builtin_printf (loc, fndecl,
|
10819 |
|
|
arg1, NULL_TREE, ignore, fcode);
|
10820 |
|
|
break;
|
10821 |
|
|
|
10822 |
|
|
case BUILT_IN_FPRINTF:
|
10823 |
|
|
case BUILT_IN_FPRINTF_UNLOCKED:
|
10824 |
|
|
case BUILT_IN_VFPRINTF:
|
10825 |
|
|
return fold_builtin_fprintf (loc, fndecl, arg0, arg1, NULL_TREE,
|
10826 |
|
|
ignore, fcode);
|
10827 |
|
|
|
10828 |
|
|
case BUILT_IN_ATOMIC_ALWAYS_LOCK_FREE:
|
10829 |
|
|
return fold_builtin_atomic_always_lock_free (arg0, arg1);
|
10830 |
|
|
|
10831 |
|
|
case BUILT_IN_ATOMIC_IS_LOCK_FREE:
|
10832 |
|
|
return fold_builtin_atomic_is_lock_free (arg0, arg1);
|
10833 |
|
|
|
10834 |
|
|
default:
|
10835 |
|
|
break;
|
10836 |
|
|
}
|
10837 |
|
|
return NULL_TREE;
|
10838 |
|
|
}
|
10839 |
|
|
|
10840 |
|
|
/* Fold a call to built-in function FNDECL with 3 arguments, ARG0, ARG1,
|
10841 |
|
|
and ARG2. IGNORE is true if the result of the function call is ignored.
|
10842 |
|
|
This function returns NULL_TREE if no simplification was possible. */
|
10843 |
|
|
|
10844 |
|
|
static tree
|
10845 |
|
|
fold_builtin_3 (location_t loc, tree fndecl,
|
10846 |
|
|
tree arg0, tree arg1, tree arg2, bool ignore)
|
10847 |
|
|
{
|
10848 |
|
|
tree type = TREE_TYPE (TREE_TYPE (fndecl));
|
10849 |
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
10850 |
|
|
switch (fcode)
|
10851 |
|
|
{
|
10852 |
|
|
|
10853 |
|
|
CASE_FLT_FN (BUILT_IN_SINCOS):
|
10854 |
|
|
return fold_builtin_sincos (loc, arg0, arg1, arg2);
|
10855 |
|
|
|
10856 |
|
|
CASE_FLT_FN (BUILT_IN_FMA):
|
10857 |
|
|
return fold_builtin_fma (loc, arg0, arg1, arg2, type);
|
10858 |
|
|
break;
|
10859 |
|
|
|
10860 |
|
|
CASE_FLT_FN (BUILT_IN_REMQUO):
|
10861 |
|
|
if (validate_arg (arg0, REAL_TYPE)
|
10862 |
|
|
&& validate_arg(arg1, REAL_TYPE)
|
10863 |
|
|
&& validate_arg(arg2, POINTER_TYPE))
|
10864 |
|
|
return do_mpfr_remquo (arg0, arg1, arg2);
|
10865 |
|
|
break;
|
10866 |
|
|
|
10867 |
|
|
case BUILT_IN_MEMSET:
|
10868 |
|
|
return fold_builtin_memset (loc, arg0, arg1, arg2, type, ignore);
|
10869 |
|
|
|
10870 |
|
|
case BUILT_IN_BCOPY:
|
10871 |
|
|
return fold_builtin_memory_op (loc, arg1, arg0, arg2,
|
10872 |
|
|
void_type_node, true, /*endp=*/3);
|
10873 |
|
|
|
10874 |
|
|
case BUILT_IN_MEMCPY:
|
10875 |
|
|
return fold_builtin_memory_op (loc, arg0, arg1, arg2,
|
10876 |
|
|
type, ignore, /*endp=*/0);
|
10877 |
|
|
|
10878 |
|
|
case BUILT_IN_MEMPCPY:
|
10879 |
|
|
return fold_builtin_memory_op (loc, arg0, arg1, arg2,
|
10880 |
|
|
type, ignore, /*endp=*/1);
|
10881 |
|
|
|
10882 |
|
|
case BUILT_IN_MEMMOVE:
|
10883 |
|
|
return fold_builtin_memory_op (loc, arg0, arg1, arg2,
|
10884 |
|
|
type, ignore, /*endp=*/3);
|
10885 |
|
|
|
10886 |
|
|
case BUILT_IN_STRNCAT:
|
10887 |
|
|
return fold_builtin_strncat (loc, arg0, arg1, arg2);
|
10888 |
|
|
|
10889 |
|
|
case BUILT_IN_STRNCPY:
|
10890 |
|
|
return fold_builtin_strncpy (loc, fndecl, arg0, arg1, arg2, NULL_TREE);
|
10891 |
|
|
|
10892 |
|
|
case BUILT_IN_STRNCMP:
|
10893 |
|
|
return fold_builtin_strncmp (loc, arg0, arg1, arg2);
|
10894 |
|
|
|
10895 |
|
|
case BUILT_IN_MEMCHR:
|
10896 |
|
|
return fold_builtin_memchr (loc, arg0, arg1, arg2, type);
|
10897 |
|
|
|
10898 |
|
|
case BUILT_IN_BCMP:
|
10899 |
|
|
case BUILT_IN_MEMCMP:
|
10900 |
|
|
return fold_builtin_memcmp (loc, arg0, arg1, arg2);;
|
10901 |
|
|
|
10902 |
|
|
case BUILT_IN_SPRINTF:
|
10903 |
|
|
return fold_builtin_sprintf (loc, arg0, arg1, arg2, ignore);
|
10904 |
|
|
|
10905 |
|
|
case BUILT_IN_SNPRINTF:
|
10906 |
|
|
return fold_builtin_snprintf (loc, arg0, arg1, arg2, NULL_TREE, ignore);
|
10907 |
|
|
|
10908 |
|
|
case BUILT_IN_STRCPY_CHK:
|
10909 |
|
|
case BUILT_IN_STPCPY_CHK:
|
10910 |
|
|
return fold_builtin_stxcpy_chk (loc, fndecl, arg0, arg1, arg2, NULL_TREE,
|
10911 |
|
|
ignore, fcode);
|
10912 |
|
|
|
10913 |
|
|
case BUILT_IN_STRCAT_CHK:
|
10914 |
|
|
return fold_builtin_strcat_chk (loc, fndecl, arg0, arg1, arg2);
|
10915 |
|
|
|
10916 |
|
|
case BUILT_IN_PRINTF_CHK:
|
10917 |
|
|
case BUILT_IN_VPRINTF_CHK:
|
10918 |
|
|
if (!validate_arg (arg0, INTEGER_TYPE)
|
10919 |
|
|
|| TREE_SIDE_EFFECTS (arg0))
|
10920 |
|
|
return NULL_TREE;
|
10921 |
|
|
else
|
10922 |
|
|
return fold_builtin_printf (loc, fndecl, arg1, arg2, ignore, fcode);
|
10923 |
|
|
break;
|
10924 |
|
|
|
10925 |
|
|
case BUILT_IN_FPRINTF:
|
10926 |
|
|
case BUILT_IN_FPRINTF_UNLOCKED:
|
10927 |
|
|
case BUILT_IN_VFPRINTF:
|
10928 |
|
|
return fold_builtin_fprintf (loc, fndecl, arg0, arg1, arg2,
|
10929 |
|
|
ignore, fcode);
|
10930 |
|
|
|
10931 |
|
|
case BUILT_IN_FPRINTF_CHK:
|
10932 |
|
|
case BUILT_IN_VFPRINTF_CHK:
|
10933 |
|
|
if (!validate_arg (arg1, INTEGER_TYPE)
|
10934 |
|
|
|| TREE_SIDE_EFFECTS (arg1))
|
10935 |
|
|
return NULL_TREE;
|
10936 |
|
|
else
|
10937 |
|
|
return fold_builtin_fprintf (loc, fndecl, arg0, arg2, NULL_TREE,
|
10938 |
|
|
ignore, fcode);
|
10939 |
|
|
|
10940 |
|
|
default:
|
10941 |
|
|
break;
|
10942 |
|
|
}
|
10943 |
|
|
return NULL_TREE;
|
10944 |
|
|
}
|
10945 |
|
|
|
10946 |
|
|
/* Fold a call to built-in function FNDECL with 4 arguments, ARG0, ARG1,
|
10947 |
|
|
ARG2, and ARG3. IGNORE is true if the result of the function call is
|
10948 |
|
|
ignored. This function returns NULL_TREE if no simplification was
|
10949 |
|
|
possible. */
|
10950 |
|
|
|
10951 |
|
|
static tree
|
10952 |
|
|
fold_builtin_4 (location_t loc, tree fndecl,
|
10953 |
|
|
tree arg0, tree arg1, tree arg2, tree arg3, bool ignore)
|
10954 |
|
|
{
|
10955 |
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
10956 |
|
|
|
10957 |
|
|
switch (fcode)
|
10958 |
|
|
{
|
10959 |
|
|
case BUILT_IN_MEMCPY_CHK:
|
10960 |
|
|
case BUILT_IN_MEMPCPY_CHK:
|
10961 |
|
|
case BUILT_IN_MEMMOVE_CHK:
|
10962 |
|
|
case BUILT_IN_MEMSET_CHK:
|
10963 |
|
|
return fold_builtin_memory_chk (loc, fndecl, arg0, arg1, arg2, arg3,
|
10964 |
|
|
NULL_TREE, ignore,
|
10965 |
|
|
DECL_FUNCTION_CODE (fndecl));
|
10966 |
|
|
|
10967 |
|
|
case BUILT_IN_STRNCPY_CHK:
|
10968 |
|
|
case BUILT_IN_STPNCPY_CHK:
|
10969 |
|
|
return fold_builtin_stxncpy_chk (loc, arg0, arg1, arg2, arg3, NULL_TREE,
|
10970 |
|
|
ignore, fcode);
|
10971 |
|
|
|
10972 |
|
|
case BUILT_IN_STRNCAT_CHK:
|
10973 |
|
|
return fold_builtin_strncat_chk (loc, fndecl, arg0, arg1, arg2, arg3);
|
10974 |
|
|
|
10975 |
|
|
case BUILT_IN_SNPRINTF:
|
10976 |
|
|
return fold_builtin_snprintf (loc, arg0, arg1, arg2, arg3, ignore);
|
10977 |
|
|
|
10978 |
|
|
case BUILT_IN_FPRINTF_CHK:
|
10979 |
|
|
case BUILT_IN_VFPRINTF_CHK:
|
10980 |
|
|
if (!validate_arg (arg1, INTEGER_TYPE)
|
10981 |
|
|
|| TREE_SIDE_EFFECTS (arg1))
|
10982 |
|
|
return NULL_TREE;
|
10983 |
|
|
else
|
10984 |
|
|
return fold_builtin_fprintf (loc, fndecl, arg0, arg2, arg3,
|
10985 |
|
|
ignore, fcode);
|
10986 |
|
|
break;
|
10987 |
|
|
|
10988 |
|
|
default:
|
10989 |
|
|
break;
|
10990 |
|
|
}
|
10991 |
|
|
return NULL_TREE;
|
10992 |
|
|
}
|
10993 |
|
|
|
10994 |
|
|
/* Fold a call to built-in function FNDECL. ARGS is an array of NARGS
|
10995 |
|
|
arguments, where NARGS <= 4. IGNORE is true if the result of the
|
10996 |
|
|
function call is ignored. This function returns NULL_TREE if no
|
10997 |
|
|
simplification was possible. Note that this only folds builtins with
|
10998 |
|
|
fixed argument patterns. Foldings that do varargs-to-varargs
|
10999 |
|
|
transformations, or that match calls with more than 4 arguments,
|
11000 |
|
|
need to be handled with fold_builtin_varargs instead. */
|
11001 |
|
|
|
11002 |
|
|
#define MAX_ARGS_TO_FOLD_BUILTIN 4
|
11003 |
|
|
|
11004 |
|
|
static tree
|
11005 |
|
|
fold_builtin_n (location_t loc, tree fndecl, tree *args, int nargs, bool ignore)
|
11006 |
|
|
{
|
11007 |
|
|
tree ret = NULL_TREE;
|
11008 |
|
|
|
11009 |
|
|
switch (nargs)
|
11010 |
|
|
{
|
11011 |
|
|
case 0:
|
11012 |
|
|
ret = fold_builtin_0 (loc, fndecl, ignore);
|
11013 |
|
|
break;
|
11014 |
|
|
case 1:
|
11015 |
|
|
ret = fold_builtin_1 (loc, fndecl, args[0], ignore);
|
11016 |
|
|
break;
|
11017 |
|
|
case 2:
|
11018 |
|
|
ret = fold_builtin_2 (loc, fndecl, args[0], args[1], ignore);
|
11019 |
|
|
break;
|
11020 |
|
|
case 3:
|
11021 |
|
|
ret = fold_builtin_3 (loc, fndecl, args[0], args[1], args[2], ignore);
|
11022 |
|
|
break;
|
11023 |
|
|
case 4:
|
11024 |
|
|
ret = fold_builtin_4 (loc, fndecl, args[0], args[1], args[2], args[3],
|
11025 |
|
|
ignore);
|
11026 |
|
|
break;
|
11027 |
|
|
default:
|
11028 |
|
|
break;
|
11029 |
|
|
}
|
11030 |
|
|
if (ret)
|
11031 |
|
|
{
|
11032 |
|
|
ret = build1 (NOP_EXPR, TREE_TYPE (ret), ret);
|
11033 |
|
|
SET_EXPR_LOCATION (ret, loc);
|
11034 |
|
|
TREE_NO_WARNING (ret) = 1;
|
11035 |
|
|
return ret;
|
11036 |
|
|
}
|
11037 |
|
|
return NULL_TREE;
|
11038 |
|
|
}
|
11039 |
|
|
|
11040 |
|
|
/* Builtins with folding operations that operate on "..." arguments
|
11041 |
|
|
need special handling; we need to store the arguments in a convenient
|
11042 |
|
|
data structure before attempting any folding. Fortunately there are
|
11043 |
|
|
only a few builtins that fall into this category. FNDECL is the
|
11044 |
|
|
function, EXP is the CALL_EXPR for the call, and IGNORE is true if the
|
11045 |
|
|
result of the function call is ignored. */
|
11046 |
|
|
|
11047 |
|
|
static tree
|
11048 |
|
|
fold_builtin_varargs (location_t loc, tree fndecl, tree exp,
|
11049 |
|
|
bool ignore ATTRIBUTE_UNUSED)
|
11050 |
|
|
{
|
11051 |
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
11052 |
|
|
tree ret = NULL_TREE;
|
11053 |
|
|
|
11054 |
|
|
switch (fcode)
|
11055 |
|
|
{
|
11056 |
|
|
case BUILT_IN_SPRINTF_CHK:
|
11057 |
|
|
case BUILT_IN_VSPRINTF_CHK:
|
11058 |
|
|
ret = fold_builtin_sprintf_chk (loc, exp, fcode);
|
11059 |
|
|
break;
|
11060 |
|
|
|
11061 |
|
|
case BUILT_IN_SNPRINTF_CHK:
|
11062 |
|
|
case BUILT_IN_VSNPRINTF_CHK:
|
11063 |
|
|
ret = fold_builtin_snprintf_chk (loc, exp, NULL_TREE, fcode);
|
11064 |
|
|
break;
|
11065 |
|
|
|
11066 |
|
|
case BUILT_IN_FPCLASSIFY:
|
11067 |
|
|
ret = fold_builtin_fpclassify (loc, exp);
|
11068 |
|
|
break;
|
11069 |
|
|
|
11070 |
|
|
default:
|
11071 |
|
|
break;
|
11072 |
|
|
}
|
11073 |
|
|
if (ret)
|
11074 |
|
|
{
|
11075 |
|
|
ret = build1 (NOP_EXPR, TREE_TYPE (ret), ret);
|
11076 |
|
|
SET_EXPR_LOCATION (ret, loc);
|
11077 |
|
|
TREE_NO_WARNING (ret) = 1;
|
11078 |
|
|
return ret;
|
11079 |
|
|
}
|
11080 |
|
|
return NULL_TREE;
|
11081 |
|
|
}
|
11082 |
|
|
|
11083 |
|
|
/* Return true if FNDECL shouldn't be folded right now.
|
11084 |
|
|
If a built-in function has an inline attribute always_inline
|
11085 |
|
|
wrapper, defer folding it after always_inline functions have
|
11086 |
|
|
been inlined, otherwise e.g. -D_FORTIFY_SOURCE checking
|
11087 |
|
|
might not be performed. */
|
11088 |
|
|
|
11089 |
|
|
bool
|
11090 |
|
|
avoid_folding_inline_builtin (tree fndecl)
|
11091 |
|
|
{
|
11092 |
|
|
return (DECL_DECLARED_INLINE_P (fndecl)
|
11093 |
|
|
&& DECL_DISREGARD_INLINE_LIMITS (fndecl)
|
11094 |
|
|
&& cfun
|
11095 |
|
|
&& !cfun->always_inline_functions_inlined
|
11096 |
|
|
&& lookup_attribute ("always_inline", DECL_ATTRIBUTES (fndecl)));
|
11097 |
|
|
}
|
11098 |
|
|
|
11099 |
|
|
/* A wrapper function for builtin folding that prevents warnings for
|
11100 |
|
|
"statement without effect" and the like, caused by removing the
|
11101 |
|
|
call node earlier than the warning is generated. */
|
11102 |
|
|
|
11103 |
|
|
tree
|
11104 |
|
|
fold_call_expr (location_t loc, tree exp, bool ignore)
|
11105 |
|
|
{
|
11106 |
|
|
tree ret = NULL_TREE;
|
11107 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
11108 |
|
|
if (fndecl
|
11109 |
|
|
&& TREE_CODE (fndecl) == FUNCTION_DECL
|
11110 |
|
|
&& DECL_BUILT_IN (fndecl)
|
11111 |
|
|
/* If CALL_EXPR_VA_ARG_PACK is set, the arguments aren't finalized
|
11112 |
|
|
yet. Defer folding until we see all the arguments
|
11113 |
|
|
(after inlining). */
|
11114 |
|
|
&& !CALL_EXPR_VA_ARG_PACK (exp))
|
11115 |
|
|
{
|
11116 |
|
|
int nargs = call_expr_nargs (exp);
|
11117 |
|
|
|
11118 |
|
|
/* Before gimplification CALL_EXPR_VA_ARG_PACK is not set, but
|
11119 |
|
|
instead last argument is __builtin_va_arg_pack (). Defer folding
|
11120 |
|
|
even in that case, until arguments are finalized. */
|
11121 |
|
|
if (nargs && TREE_CODE (CALL_EXPR_ARG (exp, nargs - 1)) == CALL_EXPR)
|
11122 |
|
|
{
|
11123 |
|
|
tree fndecl2 = get_callee_fndecl (CALL_EXPR_ARG (exp, nargs - 1));
|
11124 |
|
|
if (fndecl2
|
11125 |
|
|
&& TREE_CODE (fndecl2) == FUNCTION_DECL
|
11126 |
|
|
&& DECL_BUILT_IN_CLASS (fndecl2) == BUILT_IN_NORMAL
|
11127 |
|
|
&& DECL_FUNCTION_CODE (fndecl2) == BUILT_IN_VA_ARG_PACK)
|
11128 |
|
|
return NULL_TREE;
|
11129 |
|
|
}
|
11130 |
|
|
|
11131 |
|
|
if (avoid_folding_inline_builtin (fndecl))
|
11132 |
|
|
return NULL_TREE;
|
11133 |
|
|
|
11134 |
|
|
if (DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_MD)
|
11135 |
|
|
return targetm.fold_builtin (fndecl, call_expr_nargs (exp),
|
11136 |
|
|
CALL_EXPR_ARGP (exp), ignore);
|
11137 |
|
|
else
|
11138 |
|
|
{
|
11139 |
|
|
if (nargs <= MAX_ARGS_TO_FOLD_BUILTIN)
|
11140 |
|
|
{
|
11141 |
|
|
tree *args = CALL_EXPR_ARGP (exp);
|
11142 |
|
|
ret = fold_builtin_n (loc, fndecl, args, nargs, ignore);
|
11143 |
|
|
}
|
11144 |
|
|
if (!ret)
|
11145 |
|
|
ret = fold_builtin_varargs (loc, fndecl, exp, ignore);
|
11146 |
|
|
if (ret)
|
11147 |
|
|
return ret;
|
11148 |
|
|
}
|
11149 |
|
|
}
|
11150 |
|
|
return NULL_TREE;
|
11151 |
|
|
}
|
11152 |
|
|
|
11153 |
|
|
/* Conveniently construct a function call expression. FNDECL names the
|
11154 |
|
|
function to be called and N arguments are passed in the array
|
11155 |
|
|
ARGARRAY. */
|
11156 |
|
|
|
11157 |
|
|
tree
|
11158 |
|
|
build_call_expr_loc_array (location_t loc, tree fndecl, int n, tree *argarray)
|
11159 |
|
|
{
|
11160 |
|
|
tree fntype = TREE_TYPE (fndecl);
|
11161 |
|
|
tree fn = build1 (ADDR_EXPR, build_pointer_type (fntype), fndecl);
|
11162 |
|
|
|
11163 |
|
|
return fold_builtin_call_array (loc, TREE_TYPE (fntype), fn, n, argarray);
|
11164 |
|
|
}
|
11165 |
|
|
|
11166 |
|
|
/* Conveniently construct a function call expression. FNDECL names the
|
11167 |
|
|
function to be called and the arguments are passed in the vector
|
11168 |
|
|
VEC. */
|
11169 |
|
|
|
11170 |
|
|
tree
|
11171 |
|
|
build_call_expr_loc_vec (location_t loc, tree fndecl, VEC(tree,gc) *vec)
|
11172 |
|
|
{
|
11173 |
|
|
return build_call_expr_loc_array (loc, fndecl, VEC_length (tree, vec),
|
11174 |
|
|
VEC_address (tree, vec));
|
11175 |
|
|
}
|
11176 |
|
|
|
11177 |
|
|
|
11178 |
|
|
/* Conveniently construct a function call expression. FNDECL names the
|
11179 |
|
|
function to be called, N is the number of arguments, and the "..."
|
11180 |
|
|
parameters are the argument expressions. */
|
11181 |
|
|
|
11182 |
|
|
tree
|
11183 |
|
|
build_call_expr_loc (location_t loc, tree fndecl, int n, ...)
|
11184 |
|
|
{
|
11185 |
|
|
va_list ap;
|
11186 |
|
|
tree *argarray = XALLOCAVEC (tree, n);
|
11187 |
|
|
int i;
|
11188 |
|
|
|
11189 |
|
|
va_start (ap, n);
|
11190 |
|
|
for (i = 0; i < n; i++)
|
11191 |
|
|
argarray[i] = va_arg (ap, tree);
|
11192 |
|
|
va_end (ap);
|
11193 |
|
|
return build_call_expr_loc_array (loc, fndecl, n, argarray);
|
11194 |
|
|
}
|
11195 |
|
|
|
11196 |
|
|
/* Like build_call_expr_loc (UNKNOWN_LOCATION, ...). Duplicated because
|
11197 |
|
|
varargs macros aren't supported by all bootstrap compilers. */
|
11198 |
|
|
|
11199 |
|
|
tree
|
11200 |
|
|
build_call_expr (tree fndecl, int n, ...)
|
11201 |
|
|
{
|
11202 |
|
|
va_list ap;
|
11203 |
|
|
tree *argarray = XALLOCAVEC (tree, n);
|
11204 |
|
|
int i;
|
11205 |
|
|
|
11206 |
|
|
va_start (ap, n);
|
11207 |
|
|
for (i = 0; i < n; i++)
|
11208 |
|
|
argarray[i] = va_arg (ap, tree);
|
11209 |
|
|
va_end (ap);
|
11210 |
|
|
return build_call_expr_loc_array (UNKNOWN_LOCATION, fndecl, n, argarray);
|
11211 |
|
|
}
|
11212 |
|
|
|
11213 |
|
|
/* Construct a CALL_EXPR with type TYPE with FN as the function expression.
|
11214 |
|
|
N arguments are passed in the array ARGARRAY. */
|
11215 |
|
|
|
11216 |
|
|
tree
|
11217 |
|
|
fold_builtin_call_array (location_t loc, tree type,
|
11218 |
|
|
tree fn,
|
11219 |
|
|
int n,
|
11220 |
|
|
tree *argarray)
|
11221 |
|
|
{
|
11222 |
|
|
tree ret = NULL_TREE;
|
11223 |
|
|
tree exp;
|
11224 |
|
|
|
11225 |
|
|
if (TREE_CODE (fn) == ADDR_EXPR)
|
11226 |
|
|
{
|
11227 |
|
|
tree fndecl = TREE_OPERAND (fn, 0);
|
11228 |
|
|
if (TREE_CODE (fndecl) == FUNCTION_DECL
|
11229 |
|
|
&& DECL_BUILT_IN (fndecl))
|
11230 |
|
|
{
|
11231 |
|
|
/* If last argument is __builtin_va_arg_pack (), arguments to this
|
11232 |
|
|
function are not finalized yet. Defer folding until they are. */
|
11233 |
|
|
if (n && TREE_CODE (argarray[n - 1]) == CALL_EXPR)
|
11234 |
|
|
{
|
11235 |
|
|
tree fndecl2 = get_callee_fndecl (argarray[n - 1]);
|
11236 |
|
|
if (fndecl2
|
11237 |
|
|
&& TREE_CODE (fndecl2) == FUNCTION_DECL
|
11238 |
|
|
&& DECL_BUILT_IN_CLASS (fndecl2) == BUILT_IN_NORMAL
|
11239 |
|
|
&& DECL_FUNCTION_CODE (fndecl2) == BUILT_IN_VA_ARG_PACK)
|
11240 |
|
|
return build_call_array_loc (loc, type, fn, n, argarray);
|
11241 |
|
|
}
|
11242 |
|
|
if (avoid_folding_inline_builtin (fndecl))
|
11243 |
|
|
return build_call_array_loc (loc, type, fn, n, argarray);
|
11244 |
|
|
if (DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_MD)
|
11245 |
|
|
{
|
11246 |
|
|
ret = targetm.fold_builtin (fndecl, n, argarray, false);
|
11247 |
|
|
if (ret)
|
11248 |
|
|
return ret;
|
11249 |
|
|
|
11250 |
|
|
return build_call_array_loc (loc, type, fn, n, argarray);
|
11251 |
|
|
}
|
11252 |
|
|
else if (n <= MAX_ARGS_TO_FOLD_BUILTIN)
|
11253 |
|
|
{
|
11254 |
|
|
/* First try the transformations that don't require consing up
|
11255 |
|
|
an exp. */
|
11256 |
|
|
ret = fold_builtin_n (loc, fndecl, argarray, n, false);
|
11257 |
|
|
if (ret)
|
11258 |
|
|
return ret;
|
11259 |
|
|
}
|
11260 |
|
|
|
11261 |
|
|
/* If we got this far, we need to build an exp. */
|
11262 |
|
|
exp = build_call_array_loc (loc, type, fn, n, argarray);
|
11263 |
|
|
ret = fold_builtin_varargs (loc, fndecl, exp, false);
|
11264 |
|
|
return ret ? ret : exp;
|
11265 |
|
|
}
|
11266 |
|
|
}
|
11267 |
|
|
|
11268 |
|
|
return build_call_array_loc (loc, type, fn, n, argarray);
|
11269 |
|
|
}
|
11270 |
|
|
|
11271 |
|
|
/* Construct a new CALL_EXPR to FNDECL using the tail of the argument
|
11272 |
|
|
list ARGS along with N new arguments in NEWARGS. SKIP is the number
|
11273 |
|
|
of arguments in ARGS to be omitted. OLDNARGS is the number of
|
11274 |
|
|
elements in ARGS. */
|
11275 |
|
|
|
11276 |
|
|
static tree
|
11277 |
|
|
rewrite_call_expr_valist (location_t loc, int oldnargs, tree *args,
|
11278 |
|
|
int skip, tree fndecl, int n, va_list newargs)
|
11279 |
|
|
{
|
11280 |
|
|
int nargs = oldnargs - skip + n;
|
11281 |
|
|
tree *buffer;
|
11282 |
|
|
|
11283 |
|
|
if (n > 0)
|
11284 |
|
|
{
|
11285 |
|
|
int i, j;
|
11286 |
|
|
|
11287 |
|
|
buffer = XALLOCAVEC (tree, nargs);
|
11288 |
|
|
for (i = 0; i < n; i++)
|
11289 |
|
|
buffer[i] = va_arg (newargs, tree);
|
11290 |
|
|
for (j = skip; j < oldnargs; j++, i++)
|
11291 |
|
|
buffer[i] = args[j];
|
11292 |
|
|
}
|
11293 |
|
|
else
|
11294 |
|
|
buffer = args + skip;
|
11295 |
|
|
|
11296 |
|
|
return build_call_expr_loc_array (loc, fndecl, nargs, buffer);
|
11297 |
|
|
}
|
11298 |
|
|
|
11299 |
|
|
/* Construct a new CALL_EXPR to FNDECL using the tail of the argument
|
11300 |
|
|
list ARGS along with N new arguments specified as the "..."
|
11301 |
|
|
parameters. SKIP is the number of arguments in ARGS to be omitted.
|
11302 |
|
|
OLDNARGS is the number of elements in ARGS. */
|
11303 |
|
|
|
11304 |
|
|
static tree
|
11305 |
|
|
rewrite_call_expr_array (location_t loc, int oldnargs, tree *args,
|
11306 |
|
|
int skip, tree fndecl, int n, ...)
|
11307 |
|
|
{
|
11308 |
|
|
va_list ap;
|
11309 |
|
|
tree t;
|
11310 |
|
|
|
11311 |
|
|
va_start (ap, n);
|
11312 |
|
|
t = rewrite_call_expr_valist (loc, oldnargs, args, skip, fndecl, n, ap);
|
11313 |
|
|
va_end (ap);
|
11314 |
|
|
|
11315 |
|
|
return t;
|
11316 |
|
|
}
|
11317 |
|
|
|
11318 |
|
|
/* Construct a new CALL_EXPR using the tail of the argument list of EXP
|
11319 |
|
|
along with N new arguments specified as the "..." parameters. SKIP
|
11320 |
|
|
is the number of arguments in EXP to be omitted. This function is used
|
11321 |
|
|
to do varargs-to-varargs transformations. */
|
11322 |
|
|
|
11323 |
|
|
static tree
|
11324 |
|
|
rewrite_call_expr (location_t loc, tree exp, int skip, tree fndecl, int n, ...)
|
11325 |
|
|
{
|
11326 |
|
|
va_list ap;
|
11327 |
|
|
tree t;
|
11328 |
|
|
|
11329 |
|
|
va_start (ap, n);
|
11330 |
|
|
t = rewrite_call_expr_valist (loc, call_expr_nargs (exp),
|
11331 |
|
|
CALL_EXPR_ARGP (exp), skip, fndecl, n, ap);
|
11332 |
|
|
va_end (ap);
|
11333 |
|
|
|
11334 |
|
|
return t;
|
11335 |
|
|
}
|
11336 |
|
|
|
11337 |
|
|
/* Validate a single argument ARG against a tree code CODE representing
|
11338 |
|
|
a type. */
|
11339 |
|
|
|
11340 |
|
|
static bool
|
11341 |
|
|
validate_arg (const_tree arg, enum tree_code code)
|
11342 |
|
|
{
|
11343 |
|
|
if (!arg)
|
11344 |
|
|
return false;
|
11345 |
|
|
else if (code == POINTER_TYPE)
|
11346 |
|
|
return POINTER_TYPE_P (TREE_TYPE (arg));
|
11347 |
|
|
else if (code == INTEGER_TYPE)
|
11348 |
|
|
return INTEGRAL_TYPE_P (TREE_TYPE (arg));
|
11349 |
|
|
return code == TREE_CODE (TREE_TYPE (arg));
|
11350 |
|
|
}
|
11351 |
|
|
|
11352 |
|
|
/* This function validates the types of a function call argument list
|
11353 |
|
|
against a specified list of tree_codes. If the last specifier is a 0,
|
11354 |
|
|
that represents an ellipses, otherwise the last specifier must be a
|
11355 |
|
|
VOID_TYPE.
|
11356 |
|
|
|
11357 |
|
|
This is the GIMPLE version of validate_arglist. Eventually we want to
|
11358 |
|
|
completely convert builtins.c to work from GIMPLEs and the tree based
|
11359 |
|
|
validate_arglist will then be removed. */
|
11360 |
|
|
|
11361 |
|
|
bool
|
11362 |
|
|
validate_gimple_arglist (const_gimple call, ...)
|
11363 |
|
|
{
|
11364 |
|
|
enum tree_code code;
|
11365 |
|
|
bool res = 0;
|
11366 |
|
|
va_list ap;
|
11367 |
|
|
const_tree arg;
|
11368 |
|
|
size_t i;
|
11369 |
|
|
|
11370 |
|
|
va_start (ap, call);
|
11371 |
|
|
i = 0;
|
11372 |
|
|
|
11373 |
|
|
do
|
11374 |
|
|
{
|
11375 |
|
|
code = (enum tree_code) va_arg (ap, int);
|
11376 |
|
|
switch (code)
|
11377 |
|
|
{
|
11378 |
|
|
case 0:
|
11379 |
|
|
/* This signifies an ellipses, any further arguments are all ok. */
|
11380 |
|
|
res = true;
|
11381 |
|
|
goto end;
|
11382 |
|
|
case VOID_TYPE:
|
11383 |
|
|
/* This signifies an endlink, if no arguments remain, return
|
11384 |
|
|
true, otherwise return false. */
|
11385 |
|
|
res = (i == gimple_call_num_args (call));
|
11386 |
|
|
goto end;
|
11387 |
|
|
default:
|
11388 |
|
|
/* If no parameters remain or the parameter's code does not
|
11389 |
|
|
match the specified code, return false. Otherwise continue
|
11390 |
|
|
checking any remaining arguments. */
|
11391 |
|
|
arg = gimple_call_arg (call, i++);
|
11392 |
|
|
if (!validate_arg (arg, code))
|
11393 |
|
|
goto end;
|
11394 |
|
|
break;
|
11395 |
|
|
}
|
11396 |
|
|
}
|
11397 |
|
|
while (1);
|
11398 |
|
|
|
11399 |
|
|
/* We need gotos here since we can only have one VA_CLOSE in a
|
11400 |
|
|
function. */
|
11401 |
|
|
end: ;
|
11402 |
|
|
va_end (ap);
|
11403 |
|
|
|
11404 |
|
|
return res;
|
11405 |
|
|
}
|
11406 |
|
|
|
11407 |
|
|
/* This function validates the types of a function call argument list
|
11408 |
|
|
against a specified list of tree_codes. If the last specifier is a 0,
|
11409 |
|
|
that represents an ellipses, otherwise the last specifier must be a
|
11410 |
|
|
VOID_TYPE. */
|
11411 |
|
|
|
11412 |
|
|
bool
|
11413 |
|
|
validate_arglist (const_tree callexpr, ...)
|
11414 |
|
|
{
|
11415 |
|
|
enum tree_code code;
|
11416 |
|
|
bool res = 0;
|
11417 |
|
|
va_list ap;
|
11418 |
|
|
const_call_expr_arg_iterator iter;
|
11419 |
|
|
const_tree arg;
|
11420 |
|
|
|
11421 |
|
|
va_start (ap, callexpr);
|
11422 |
|
|
init_const_call_expr_arg_iterator (callexpr, &iter);
|
11423 |
|
|
|
11424 |
|
|
do
|
11425 |
|
|
{
|
11426 |
|
|
code = (enum tree_code) va_arg (ap, int);
|
11427 |
|
|
switch (code)
|
11428 |
|
|
{
|
11429 |
|
|
case 0:
|
11430 |
|
|
/* This signifies an ellipses, any further arguments are all ok. */
|
11431 |
|
|
res = true;
|
11432 |
|
|
goto end;
|
11433 |
|
|
case VOID_TYPE:
|
11434 |
|
|
/* This signifies an endlink, if no arguments remain, return
|
11435 |
|
|
true, otherwise return false. */
|
11436 |
|
|
res = !more_const_call_expr_args_p (&iter);
|
11437 |
|
|
goto end;
|
11438 |
|
|
default:
|
11439 |
|
|
/* If no parameters remain or the parameter's code does not
|
11440 |
|
|
match the specified code, return false. Otherwise continue
|
11441 |
|
|
checking any remaining arguments. */
|
11442 |
|
|
arg = next_const_call_expr_arg (&iter);
|
11443 |
|
|
if (!validate_arg (arg, code))
|
11444 |
|
|
goto end;
|
11445 |
|
|
break;
|
11446 |
|
|
}
|
11447 |
|
|
}
|
11448 |
|
|
while (1);
|
11449 |
|
|
|
11450 |
|
|
/* We need gotos here since we can only have one VA_CLOSE in a
|
11451 |
|
|
function. */
|
11452 |
|
|
end: ;
|
11453 |
|
|
va_end (ap);
|
11454 |
|
|
|
11455 |
|
|
return res;
|
11456 |
|
|
}
|
11457 |
|
|
|
11458 |
|
|
/* Default target-specific builtin expander that does nothing. */
|
11459 |
|
|
|
11460 |
|
|
rtx
|
11461 |
|
|
default_expand_builtin (tree exp ATTRIBUTE_UNUSED,
|
11462 |
|
|
rtx target ATTRIBUTE_UNUSED,
|
11463 |
|
|
rtx subtarget ATTRIBUTE_UNUSED,
|
11464 |
|
|
enum machine_mode mode ATTRIBUTE_UNUSED,
|
11465 |
|
|
int ignore ATTRIBUTE_UNUSED)
|
11466 |
|
|
{
|
11467 |
|
|
return NULL_RTX;
|
11468 |
|
|
}
|
11469 |
|
|
|
11470 |
|
|
/* Returns true is EXP represents data that would potentially reside
|
11471 |
|
|
in a readonly section. */
|
11472 |
|
|
|
11473 |
|
|
static bool
|
11474 |
|
|
readonly_data_expr (tree exp)
|
11475 |
|
|
{
|
11476 |
|
|
STRIP_NOPS (exp);
|
11477 |
|
|
|
11478 |
|
|
if (TREE_CODE (exp) != ADDR_EXPR)
|
11479 |
|
|
return false;
|
11480 |
|
|
|
11481 |
|
|
exp = get_base_address (TREE_OPERAND (exp, 0));
|
11482 |
|
|
if (!exp)
|
11483 |
|
|
return false;
|
11484 |
|
|
|
11485 |
|
|
/* Make sure we call decl_readonly_section only for trees it
|
11486 |
|
|
can handle (since it returns true for everything it doesn't
|
11487 |
|
|
understand). */
|
11488 |
|
|
if (TREE_CODE (exp) == STRING_CST
|
11489 |
|
|
|| TREE_CODE (exp) == CONSTRUCTOR
|
11490 |
|
|
|| (TREE_CODE (exp) == VAR_DECL && TREE_STATIC (exp)))
|
11491 |
|
|
return decl_readonly_section (exp, 0);
|
11492 |
|
|
else
|
11493 |
|
|
return false;
|
11494 |
|
|
}
|
11495 |
|
|
|
11496 |
|
|
/* Simplify a call to the strstr builtin. S1 and S2 are the arguments
|
11497 |
|
|
to the call, and TYPE is its return type.
|
11498 |
|
|
|
11499 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
11500 |
|
|
simplified form of the call as a tree.
|
11501 |
|
|
|
11502 |
|
|
The simplified form may be a constant or other expression which
|
11503 |
|
|
computes the same value, but in a more efficient manner (including
|
11504 |
|
|
calls to other builtin functions).
|
11505 |
|
|
|
11506 |
|
|
The call may contain arguments which need to be evaluated, but
|
11507 |
|
|
which are not useful to determine the result of the call. In
|
11508 |
|
|
this case we return a chain of COMPOUND_EXPRs. The LHS of each
|
11509 |
|
|
COMPOUND_EXPR will be an argument which must be evaluated.
|
11510 |
|
|
COMPOUND_EXPRs are chained through their RHS. The RHS of the last
|
11511 |
|
|
COMPOUND_EXPR in the chain will contain the tree for the simplified
|
11512 |
|
|
form of the builtin function call. */
|
11513 |
|
|
|
11514 |
|
|
static tree
|
11515 |
|
|
fold_builtin_strstr (location_t loc, tree s1, tree s2, tree type)
|
11516 |
|
|
{
|
11517 |
|
|
if (!validate_arg (s1, POINTER_TYPE)
|
11518 |
|
|
|| !validate_arg (s2, POINTER_TYPE))
|
11519 |
|
|
return NULL_TREE;
|
11520 |
|
|
else
|
11521 |
|
|
{
|
11522 |
|
|
tree fn;
|
11523 |
|
|
const char *p1, *p2;
|
11524 |
|
|
|
11525 |
|
|
p2 = c_getstr (s2);
|
11526 |
|
|
if (p2 == NULL)
|
11527 |
|
|
return NULL_TREE;
|
11528 |
|
|
|
11529 |
|
|
p1 = c_getstr (s1);
|
11530 |
|
|
if (p1 != NULL)
|
11531 |
|
|
{
|
11532 |
|
|
const char *r = strstr (p1, p2);
|
11533 |
|
|
tree tem;
|
11534 |
|
|
|
11535 |
|
|
if (r == NULL)
|
11536 |
|
|
return build_int_cst (TREE_TYPE (s1), 0);
|
11537 |
|
|
|
11538 |
|
|
/* Return an offset into the constant string argument. */
|
11539 |
|
|
tem = fold_build_pointer_plus_hwi_loc (loc, s1, r - p1);
|
11540 |
|
|
return fold_convert_loc (loc, type, tem);
|
11541 |
|
|
}
|
11542 |
|
|
|
11543 |
|
|
/* The argument is const char *, and the result is char *, so we need
|
11544 |
|
|
a type conversion here to avoid a warning. */
|
11545 |
|
|
if (p2[0] == '\0')
|
11546 |
|
|
return fold_convert_loc (loc, type, s1);
|
11547 |
|
|
|
11548 |
|
|
if (p2[1] != '\0')
|
11549 |
|
|
return NULL_TREE;
|
11550 |
|
|
|
11551 |
|
|
fn = builtin_decl_implicit (BUILT_IN_STRCHR);
|
11552 |
|
|
if (!fn)
|
11553 |
|
|
return NULL_TREE;
|
11554 |
|
|
|
11555 |
|
|
/* New argument list transforming strstr(s1, s2) to
|
11556 |
|
|
strchr(s1, s2[0]). */
|
11557 |
|
|
return build_call_expr_loc (loc, fn, 2, s1,
|
11558 |
|
|
build_int_cst (integer_type_node, p2[0]));
|
11559 |
|
|
}
|
11560 |
|
|
}
|
11561 |
|
|
|
11562 |
|
|
/* Simplify a call to the strchr builtin. S1 and S2 are the arguments to
|
11563 |
|
|
the call, and TYPE is its return type.
|
11564 |
|
|
|
11565 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
11566 |
|
|
simplified form of the call as a tree.
|
11567 |
|
|
|
11568 |
|
|
The simplified form may be a constant or other expression which
|
11569 |
|
|
computes the same value, but in a more efficient manner (including
|
11570 |
|
|
calls to other builtin functions).
|
11571 |
|
|
|
11572 |
|
|
The call may contain arguments which need to be evaluated, but
|
11573 |
|
|
which are not useful to determine the result of the call. In
|
11574 |
|
|
this case we return a chain of COMPOUND_EXPRs. The LHS of each
|
11575 |
|
|
COMPOUND_EXPR will be an argument which must be evaluated.
|
11576 |
|
|
COMPOUND_EXPRs are chained through their RHS. The RHS of the last
|
11577 |
|
|
COMPOUND_EXPR in the chain will contain the tree for the simplified
|
11578 |
|
|
form of the builtin function call. */
|
11579 |
|
|
|
11580 |
|
|
static tree
|
11581 |
|
|
fold_builtin_strchr (location_t loc, tree s1, tree s2, tree type)
|
11582 |
|
|
{
|
11583 |
|
|
if (!validate_arg (s1, POINTER_TYPE)
|
11584 |
|
|
|| !validate_arg (s2, INTEGER_TYPE))
|
11585 |
|
|
return NULL_TREE;
|
11586 |
|
|
else
|
11587 |
|
|
{
|
11588 |
|
|
const char *p1;
|
11589 |
|
|
|
11590 |
|
|
if (TREE_CODE (s2) != INTEGER_CST)
|
11591 |
|
|
return NULL_TREE;
|
11592 |
|
|
|
11593 |
|
|
p1 = c_getstr (s1);
|
11594 |
|
|
if (p1 != NULL)
|
11595 |
|
|
{
|
11596 |
|
|
char c;
|
11597 |
|
|
const char *r;
|
11598 |
|
|
tree tem;
|
11599 |
|
|
|
11600 |
|
|
if (target_char_cast (s2, &c))
|
11601 |
|
|
return NULL_TREE;
|
11602 |
|
|
|
11603 |
|
|
r = strchr (p1, c);
|
11604 |
|
|
|
11605 |
|
|
if (r == NULL)
|
11606 |
|
|
return build_int_cst (TREE_TYPE (s1), 0);
|
11607 |
|
|
|
11608 |
|
|
/* Return an offset into the constant string argument. */
|
11609 |
|
|
tem = fold_build_pointer_plus_hwi_loc (loc, s1, r - p1);
|
11610 |
|
|
return fold_convert_loc (loc, type, tem);
|
11611 |
|
|
}
|
11612 |
|
|
return NULL_TREE;
|
11613 |
|
|
}
|
11614 |
|
|
}
|
11615 |
|
|
|
11616 |
|
|
/* Simplify a call to the strrchr builtin. S1 and S2 are the arguments to
|
11617 |
|
|
the call, and TYPE is its return type.
|
11618 |
|
|
|
11619 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
11620 |
|
|
simplified form of the call as a tree.
|
11621 |
|
|
|
11622 |
|
|
The simplified form may be a constant or other expression which
|
11623 |
|
|
computes the same value, but in a more efficient manner (including
|
11624 |
|
|
calls to other builtin functions).
|
11625 |
|
|
|
11626 |
|
|
The call may contain arguments which need to be evaluated, but
|
11627 |
|
|
which are not useful to determine the result of the call. In
|
11628 |
|
|
this case we return a chain of COMPOUND_EXPRs. The LHS of each
|
11629 |
|
|
COMPOUND_EXPR will be an argument which must be evaluated.
|
11630 |
|
|
COMPOUND_EXPRs are chained through their RHS. The RHS of the last
|
11631 |
|
|
COMPOUND_EXPR in the chain will contain the tree for the simplified
|
11632 |
|
|
form of the builtin function call. */
|
11633 |
|
|
|
11634 |
|
|
static tree
|
11635 |
|
|
fold_builtin_strrchr (location_t loc, tree s1, tree s2, tree type)
|
11636 |
|
|
{
|
11637 |
|
|
if (!validate_arg (s1, POINTER_TYPE)
|
11638 |
|
|
|| !validate_arg (s2, INTEGER_TYPE))
|
11639 |
|
|
return NULL_TREE;
|
11640 |
|
|
else
|
11641 |
|
|
{
|
11642 |
|
|
tree fn;
|
11643 |
|
|
const char *p1;
|
11644 |
|
|
|
11645 |
|
|
if (TREE_CODE (s2) != INTEGER_CST)
|
11646 |
|
|
return NULL_TREE;
|
11647 |
|
|
|
11648 |
|
|
p1 = c_getstr (s1);
|
11649 |
|
|
if (p1 != NULL)
|
11650 |
|
|
{
|
11651 |
|
|
char c;
|
11652 |
|
|
const char *r;
|
11653 |
|
|
tree tem;
|
11654 |
|
|
|
11655 |
|
|
if (target_char_cast (s2, &c))
|
11656 |
|
|
return NULL_TREE;
|
11657 |
|
|
|
11658 |
|
|
r = strrchr (p1, c);
|
11659 |
|
|
|
11660 |
|
|
if (r == NULL)
|
11661 |
|
|
return build_int_cst (TREE_TYPE (s1), 0);
|
11662 |
|
|
|
11663 |
|
|
/* Return an offset into the constant string argument. */
|
11664 |
|
|
tem = fold_build_pointer_plus_hwi_loc (loc, s1, r - p1);
|
11665 |
|
|
return fold_convert_loc (loc, type, tem);
|
11666 |
|
|
}
|
11667 |
|
|
|
11668 |
|
|
if (! integer_zerop (s2))
|
11669 |
|
|
return NULL_TREE;
|
11670 |
|
|
|
11671 |
|
|
fn = builtin_decl_implicit (BUILT_IN_STRCHR);
|
11672 |
|
|
if (!fn)
|
11673 |
|
|
return NULL_TREE;
|
11674 |
|
|
|
11675 |
|
|
/* Transform strrchr(s1, '\0') to strchr(s1, '\0'). */
|
11676 |
|
|
return build_call_expr_loc (loc, fn, 2, s1, s2);
|
11677 |
|
|
}
|
11678 |
|
|
}
|
11679 |
|
|
|
11680 |
|
|
/* Simplify a call to the strpbrk builtin. S1 and S2 are the arguments
|
11681 |
|
|
to the call, and TYPE is its return type.
|
11682 |
|
|
|
11683 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
11684 |
|
|
simplified form of the call as a tree.
|
11685 |
|
|
|
11686 |
|
|
The simplified form may be a constant or other expression which
|
11687 |
|
|
computes the same value, but in a more efficient manner (including
|
11688 |
|
|
calls to other builtin functions).
|
11689 |
|
|
|
11690 |
|
|
The call may contain arguments which need to be evaluated, but
|
11691 |
|
|
which are not useful to determine the result of the call. In
|
11692 |
|
|
this case we return a chain of COMPOUND_EXPRs. The LHS of each
|
11693 |
|
|
COMPOUND_EXPR will be an argument which must be evaluated.
|
11694 |
|
|
COMPOUND_EXPRs are chained through their RHS. The RHS of the last
|
11695 |
|
|
COMPOUND_EXPR in the chain will contain the tree for the simplified
|
11696 |
|
|
form of the builtin function call. */
|
11697 |
|
|
|
11698 |
|
|
static tree
|
11699 |
|
|
fold_builtin_strpbrk (location_t loc, tree s1, tree s2, tree type)
|
11700 |
|
|
{
|
11701 |
|
|
if (!validate_arg (s1, POINTER_TYPE)
|
11702 |
|
|
|| !validate_arg (s2, POINTER_TYPE))
|
11703 |
|
|
return NULL_TREE;
|
11704 |
|
|
else
|
11705 |
|
|
{
|
11706 |
|
|
tree fn;
|
11707 |
|
|
const char *p1, *p2;
|
11708 |
|
|
|
11709 |
|
|
p2 = c_getstr (s2);
|
11710 |
|
|
if (p2 == NULL)
|
11711 |
|
|
return NULL_TREE;
|
11712 |
|
|
|
11713 |
|
|
p1 = c_getstr (s1);
|
11714 |
|
|
if (p1 != NULL)
|
11715 |
|
|
{
|
11716 |
|
|
const char *r = strpbrk (p1, p2);
|
11717 |
|
|
tree tem;
|
11718 |
|
|
|
11719 |
|
|
if (r == NULL)
|
11720 |
|
|
return build_int_cst (TREE_TYPE (s1), 0);
|
11721 |
|
|
|
11722 |
|
|
/* Return an offset into the constant string argument. */
|
11723 |
|
|
tem = fold_build_pointer_plus_hwi_loc (loc, s1, r - p1);
|
11724 |
|
|
return fold_convert_loc (loc, type, tem);
|
11725 |
|
|
}
|
11726 |
|
|
|
11727 |
|
|
if (p2[0] == '\0')
|
11728 |
|
|
/* strpbrk(x, "") == NULL.
|
11729 |
|
|
Evaluate and ignore s1 in case it had side-effects. */
|
11730 |
|
|
return omit_one_operand_loc (loc, TREE_TYPE (s1), integer_zero_node, s1);
|
11731 |
|
|
|
11732 |
|
|
if (p2[1] != '\0')
|
11733 |
|
|
return NULL_TREE; /* Really call strpbrk. */
|
11734 |
|
|
|
11735 |
|
|
fn = builtin_decl_implicit (BUILT_IN_STRCHR);
|
11736 |
|
|
if (!fn)
|
11737 |
|
|
return NULL_TREE;
|
11738 |
|
|
|
11739 |
|
|
/* New argument list transforming strpbrk(s1, s2) to
|
11740 |
|
|
strchr(s1, s2[0]). */
|
11741 |
|
|
return build_call_expr_loc (loc, fn, 2, s1,
|
11742 |
|
|
build_int_cst (integer_type_node, p2[0]));
|
11743 |
|
|
}
|
11744 |
|
|
}
|
11745 |
|
|
|
11746 |
|
|
/* Simplify a call to the strcat builtin. DST and SRC are the arguments
|
11747 |
|
|
to the call.
|
11748 |
|
|
|
11749 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
11750 |
|
|
simplified form of the call as a tree.
|
11751 |
|
|
|
11752 |
|
|
The simplified form may be a constant or other expression which
|
11753 |
|
|
computes the same value, but in a more efficient manner (including
|
11754 |
|
|
calls to other builtin functions).
|
11755 |
|
|
|
11756 |
|
|
The call may contain arguments which need to be evaluated, but
|
11757 |
|
|
which are not useful to determine the result of the call. In
|
11758 |
|
|
this case we return a chain of COMPOUND_EXPRs. The LHS of each
|
11759 |
|
|
COMPOUND_EXPR will be an argument which must be evaluated.
|
11760 |
|
|
COMPOUND_EXPRs are chained through their RHS. The RHS of the last
|
11761 |
|
|
COMPOUND_EXPR in the chain will contain the tree for the simplified
|
11762 |
|
|
form of the builtin function call. */
|
11763 |
|
|
|
11764 |
|
|
static tree
|
11765 |
|
|
fold_builtin_strcat (location_t loc ATTRIBUTE_UNUSED, tree dst, tree src)
|
11766 |
|
|
{
|
11767 |
|
|
if (!validate_arg (dst, POINTER_TYPE)
|
11768 |
|
|
|| !validate_arg (src, POINTER_TYPE))
|
11769 |
|
|
return NULL_TREE;
|
11770 |
|
|
else
|
11771 |
|
|
{
|
11772 |
|
|
const char *p = c_getstr (src);
|
11773 |
|
|
|
11774 |
|
|
/* If the string length is zero, return the dst parameter. */
|
11775 |
|
|
if (p && *p == '\0')
|
11776 |
|
|
return dst;
|
11777 |
|
|
|
11778 |
|
|
if (optimize_insn_for_speed_p ())
|
11779 |
|
|
{
|
11780 |
|
|
/* See if we can store by pieces into (dst + strlen(dst)). */
|
11781 |
|
|
tree newdst, call;
|
11782 |
|
|
tree strlen_fn = builtin_decl_implicit (BUILT_IN_STRLEN);
|
11783 |
|
|
tree strcpy_fn = builtin_decl_implicit (BUILT_IN_STRCPY);
|
11784 |
|
|
|
11785 |
|
|
if (!strlen_fn || !strcpy_fn)
|
11786 |
|
|
return NULL_TREE;
|
11787 |
|
|
|
11788 |
|
|
/* If we don't have a movstr we don't want to emit an strcpy
|
11789 |
|
|
call. We have to do that if the length of the source string
|
11790 |
|
|
isn't computable (in that case we can use memcpy probably
|
11791 |
|
|
later expanding to a sequence of mov instructions). If we
|
11792 |
|
|
have movstr instructions we can emit strcpy calls. */
|
11793 |
|
|
if (!HAVE_movstr)
|
11794 |
|
|
{
|
11795 |
|
|
tree len = c_strlen (src, 1);
|
11796 |
|
|
if (! len || TREE_SIDE_EFFECTS (len))
|
11797 |
|
|
return NULL_TREE;
|
11798 |
|
|
}
|
11799 |
|
|
|
11800 |
|
|
/* Stabilize the argument list. */
|
11801 |
|
|
dst = builtin_save_expr (dst);
|
11802 |
|
|
|
11803 |
|
|
/* Create strlen (dst). */
|
11804 |
|
|
newdst = build_call_expr_loc (loc, strlen_fn, 1, dst);
|
11805 |
|
|
/* Create (dst p+ strlen (dst)). */
|
11806 |
|
|
|
11807 |
|
|
newdst = fold_build_pointer_plus_loc (loc, dst, newdst);
|
11808 |
|
|
newdst = builtin_save_expr (newdst);
|
11809 |
|
|
|
11810 |
|
|
call = build_call_expr_loc (loc, strcpy_fn, 2, newdst, src);
|
11811 |
|
|
return build2 (COMPOUND_EXPR, TREE_TYPE (dst), call, dst);
|
11812 |
|
|
}
|
11813 |
|
|
return NULL_TREE;
|
11814 |
|
|
}
|
11815 |
|
|
}
|
11816 |
|
|
|
11817 |
|
|
/* Simplify a call to the strncat builtin. DST, SRC, and LEN are the
|
11818 |
|
|
arguments to the call.
|
11819 |
|
|
|
11820 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
11821 |
|
|
simplified form of the call as a tree.
|
11822 |
|
|
|
11823 |
|
|
The simplified form may be a constant or other expression which
|
11824 |
|
|
computes the same value, but in a more efficient manner (including
|
11825 |
|
|
calls to other builtin functions).
|
11826 |
|
|
|
11827 |
|
|
The call may contain arguments which need to be evaluated, but
|
11828 |
|
|
which are not useful to determine the result of the call. In
|
11829 |
|
|
this case we return a chain of COMPOUND_EXPRs. The LHS of each
|
11830 |
|
|
COMPOUND_EXPR will be an argument which must be evaluated.
|
11831 |
|
|
COMPOUND_EXPRs are chained through their RHS. The RHS of the last
|
11832 |
|
|
COMPOUND_EXPR in the chain will contain the tree for the simplified
|
11833 |
|
|
form of the builtin function call. */
|
11834 |
|
|
|
11835 |
|
|
static tree
|
11836 |
|
|
fold_builtin_strncat (location_t loc, tree dst, tree src, tree len)
|
11837 |
|
|
{
|
11838 |
|
|
if (!validate_arg (dst, POINTER_TYPE)
|
11839 |
|
|
|| !validate_arg (src, POINTER_TYPE)
|
11840 |
|
|
|| !validate_arg (len, INTEGER_TYPE))
|
11841 |
|
|
return NULL_TREE;
|
11842 |
|
|
else
|
11843 |
|
|
{
|
11844 |
|
|
const char *p = c_getstr (src);
|
11845 |
|
|
|
11846 |
|
|
/* If the requested length is zero, or the src parameter string
|
11847 |
|
|
length is zero, return the dst parameter. */
|
11848 |
|
|
if (integer_zerop (len) || (p && *p == '\0'))
|
11849 |
|
|
return omit_two_operands_loc (loc, TREE_TYPE (dst), dst, src, len);
|
11850 |
|
|
|
11851 |
|
|
/* If the requested len is greater than or equal to the string
|
11852 |
|
|
length, call strcat. */
|
11853 |
|
|
if (TREE_CODE (len) == INTEGER_CST && p
|
11854 |
|
|
&& compare_tree_int (len, strlen (p)) >= 0)
|
11855 |
|
|
{
|
11856 |
|
|
tree fn = builtin_decl_implicit (BUILT_IN_STRCAT);
|
11857 |
|
|
|
11858 |
|
|
/* If the replacement _DECL isn't initialized, don't do the
|
11859 |
|
|
transformation. */
|
11860 |
|
|
if (!fn)
|
11861 |
|
|
return NULL_TREE;
|
11862 |
|
|
|
11863 |
|
|
return build_call_expr_loc (loc, fn, 2, dst, src);
|
11864 |
|
|
}
|
11865 |
|
|
return NULL_TREE;
|
11866 |
|
|
}
|
11867 |
|
|
}
|
11868 |
|
|
|
11869 |
|
|
/* Simplify a call to the strspn builtin. S1 and S2 are the arguments
|
11870 |
|
|
to the call.
|
11871 |
|
|
|
11872 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
11873 |
|
|
simplified form of the call as a tree.
|
11874 |
|
|
|
11875 |
|
|
The simplified form may be a constant or other expression which
|
11876 |
|
|
computes the same value, but in a more efficient manner (including
|
11877 |
|
|
calls to other builtin functions).
|
11878 |
|
|
|
11879 |
|
|
The call may contain arguments which need to be evaluated, but
|
11880 |
|
|
which are not useful to determine the result of the call. In
|
11881 |
|
|
this case we return a chain of COMPOUND_EXPRs. The LHS of each
|
11882 |
|
|
COMPOUND_EXPR will be an argument which must be evaluated.
|
11883 |
|
|
COMPOUND_EXPRs are chained through their RHS. The RHS of the last
|
11884 |
|
|
COMPOUND_EXPR in the chain will contain the tree for the simplified
|
11885 |
|
|
form of the builtin function call. */
|
11886 |
|
|
|
11887 |
|
|
static tree
|
11888 |
|
|
fold_builtin_strspn (location_t loc, tree s1, tree s2)
|
11889 |
|
|
{
|
11890 |
|
|
if (!validate_arg (s1, POINTER_TYPE)
|
11891 |
|
|
|| !validate_arg (s2, POINTER_TYPE))
|
11892 |
|
|
return NULL_TREE;
|
11893 |
|
|
else
|
11894 |
|
|
{
|
11895 |
|
|
const char *p1 = c_getstr (s1), *p2 = c_getstr (s2);
|
11896 |
|
|
|
11897 |
|
|
/* If both arguments are constants, evaluate at compile-time. */
|
11898 |
|
|
if (p1 && p2)
|
11899 |
|
|
{
|
11900 |
|
|
const size_t r = strspn (p1, p2);
|
11901 |
|
|
return size_int (r);
|
11902 |
|
|
}
|
11903 |
|
|
|
11904 |
|
|
/* If either argument is "", return NULL_TREE. */
|
11905 |
|
|
if ((p1 && *p1 == '\0') || (p2 && *p2 == '\0'))
|
11906 |
|
|
/* Evaluate and ignore both arguments in case either one has
|
11907 |
|
|
side-effects. */
|
11908 |
|
|
return omit_two_operands_loc (loc, size_type_node, size_zero_node,
|
11909 |
|
|
s1, s2);
|
11910 |
|
|
return NULL_TREE;
|
11911 |
|
|
}
|
11912 |
|
|
}
|
11913 |
|
|
|
11914 |
|
|
/* Simplify a call to the strcspn builtin. S1 and S2 are the arguments
|
11915 |
|
|
to the call.
|
11916 |
|
|
|
11917 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
11918 |
|
|
simplified form of the call as a tree.
|
11919 |
|
|
|
11920 |
|
|
The simplified form may be a constant or other expression which
|
11921 |
|
|
computes the same value, but in a more efficient manner (including
|
11922 |
|
|
calls to other builtin functions).
|
11923 |
|
|
|
11924 |
|
|
The call may contain arguments which need to be evaluated, but
|
11925 |
|
|
which are not useful to determine the result of the call. In
|
11926 |
|
|
this case we return a chain of COMPOUND_EXPRs. The LHS of each
|
11927 |
|
|
COMPOUND_EXPR will be an argument which must be evaluated.
|
11928 |
|
|
COMPOUND_EXPRs are chained through their RHS. The RHS of the last
|
11929 |
|
|
COMPOUND_EXPR in the chain will contain the tree for the simplified
|
11930 |
|
|
form of the builtin function call. */
|
11931 |
|
|
|
11932 |
|
|
static tree
|
11933 |
|
|
fold_builtin_strcspn (location_t loc, tree s1, tree s2)
|
11934 |
|
|
{
|
11935 |
|
|
if (!validate_arg (s1, POINTER_TYPE)
|
11936 |
|
|
|| !validate_arg (s2, POINTER_TYPE))
|
11937 |
|
|
return NULL_TREE;
|
11938 |
|
|
else
|
11939 |
|
|
{
|
11940 |
|
|
const char *p1 = c_getstr (s1), *p2 = c_getstr (s2);
|
11941 |
|
|
|
11942 |
|
|
/* If both arguments are constants, evaluate at compile-time. */
|
11943 |
|
|
if (p1 && p2)
|
11944 |
|
|
{
|
11945 |
|
|
const size_t r = strcspn (p1, p2);
|
11946 |
|
|
return size_int (r);
|
11947 |
|
|
}
|
11948 |
|
|
|
11949 |
|
|
/* If the first argument is "", return NULL_TREE. */
|
11950 |
|
|
if (p1 && *p1 == '\0')
|
11951 |
|
|
{
|
11952 |
|
|
/* Evaluate and ignore argument s2 in case it has
|
11953 |
|
|
side-effects. */
|
11954 |
|
|
return omit_one_operand_loc (loc, size_type_node,
|
11955 |
|
|
size_zero_node, s2);
|
11956 |
|
|
}
|
11957 |
|
|
|
11958 |
|
|
/* If the second argument is "", return __builtin_strlen(s1). */
|
11959 |
|
|
if (p2 && *p2 == '\0')
|
11960 |
|
|
{
|
11961 |
|
|
tree fn = builtin_decl_implicit (BUILT_IN_STRLEN);
|
11962 |
|
|
|
11963 |
|
|
/* If the replacement _DECL isn't initialized, don't do the
|
11964 |
|
|
transformation. */
|
11965 |
|
|
if (!fn)
|
11966 |
|
|
return NULL_TREE;
|
11967 |
|
|
|
11968 |
|
|
return build_call_expr_loc (loc, fn, 1, s1);
|
11969 |
|
|
}
|
11970 |
|
|
return NULL_TREE;
|
11971 |
|
|
}
|
11972 |
|
|
}
|
11973 |
|
|
|
11974 |
|
|
/* Fold a call to the fputs builtin. ARG0 and ARG1 are the arguments
|
11975 |
|
|
to the call. IGNORE is true if the value returned
|
11976 |
|
|
by the builtin will be ignored. UNLOCKED is true is true if this
|
11977 |
|
|
actually a call to fputs_unlocked. If LEN in non-NULL, it represents
|
11978 |
|
|
the known length of the string. Return NULL_TREE if no simplification
|
11979 |
|
|
was possible. */
|
11980 |
|
|
|
11981 |
|
|
tree
|
11982 |
|
|
fold_builtin_fputs (location_t loc, tree arg0, tree arg1,
|
11983 |
|
|
bool ignore, bool unlocked, tree len)
|
11984 |
|
|
{
|
11985 |
|
|
/* If we're using an unlocked function, assume the other unlocked
|
11986 |
|
|
functions exist explicitly. */
|
11987 |
|
|
tree const fn_fputc = (unlocked
|
11988 |
|
|
? builtin_decl_explicit (BUILT_IN_FPUTC_UNLOCKED)
|
11989 |
|
|
: builtin_decl_implicit (BUILT_IN_FPUTC));
|
11990 |
|
|
tree const fn_fwrite = (unlocked
|
11991 |
|
|
? builtin_decl_explicit (BUILT_IN_FWRITE_UNLOCKED)
|
11992 |
|
|
: builtin_decl_implicit (BUILT_IN_FWRITE));
|
11993 |
|
|
|
11994 |
|
|
/* If the return value is used, don't do the transformation. */
|
11995 |
|
|
if (!ignore)
|
11996 |
|
|
return NULL_TREE;
|
11997 |
|
|
|
11998 |
|
|
/* Verify the arguments in the original call. */
|
11999 |
|
|
if (!validate_arg (arg0, POINTER_TYPE)
|
12000 |
|
|
|| !validate_arg (arg1, POINTER_TYPE))
|
12001 |
|
|
return NULL_TREE;
|
12002 |
|
|
|
12003 |
|
|
if (! len)
|
12004 |
|
|
len = c_strlen (arg0, 0);
|
12005 |
|
|
|
12006 |
|
|
/* Get the length of the string passed to fputs. If the length
|
12007 |
|
|
can't be determined, punt. */
|
12008 |
|
|
if (!len
|
12009 |
|
|
|| TREE_CODE (len) != INTEGER_CST)
|
12010 |
|
|
return NULL_TREE;
|
12011 |
|
|
|
12012 |
|
|
switch (compare_tree_int (len, 1))
|
12013 |
|
|
{
|
12014 |
|
|
case -1: /* length is 0, delete the call entirely . */
|
12015 |
|
|
return omit_one_operand_loc (loc, integer_type_node,
|
12016 |
|
|
integer_zero_node, arg1);;
|
12017 |
|
|
|
12018 |
|
|
case 0: /* length is 1, call fputc. */
|
12019 |
|
|
{
|
12020 |
|
|
const char *p = c_getstr (arg0);
|
12021 |
|
|
|
12022 |
|
|
if (p != NULL)
|
12023 |
|
|
{
|
12024 |
|
|
if (fn_fputc)
|
12025 |
|
|
return build_call_expr_loc (loc, fn_fputc, 2,
|
12026 |
|
|
build_int_cst
|
12027 |
|
|
(integer_type_node, p[0]), arg1);
|
12028 |
|
|
else
|
12029 |
|
|
return NULL_TREE;
|
12030 |
|
|
}
|
12031 |
|
|
}
|
12032 |
|
|
/* FALLTHROUGH */
|
12033 |
|
|
case 1: /* length is greater than 1, call fwrite. */
|
12034 |
|
|
{
|
12035 |
|
|
/* If optimizing for size keep fputs. */
|
12036 |
|
|
if (optimize_function_for_size_p (cfun))
|
12037 |
|
|
return NULL_TREE;
|
12038 |
|
|
/* New argument list transforming fputs(string, stream) to
|
12039 |
|
|
fwrite(string, 1, len, stream). */
|
12040 |
|
|
if (fn_fwrite)
|
12041 |
|
|
return build_call_expr_loc (loc, fn_fwrite, 4, arg0,
|
12042 |
|
|
size_one_node, len, arg1);
|
12043 |
|
|
else
|
12044 |
|
|
return NULL_TREE;
|
12045 |
|
|
}
|
12046 |
|
|
default:
|
12047 |
|
|
gcc_unreachable ();
|
12048 |
|
|
}
|
12049 |
|
|
return NULL_TREE;
|
12050 |
|
|
}
|
12051 |
|
|
|
12052 |
|
|
/* Fold the next_arg or va_start call EXP. Returns true if there was an error
|
12053 |
|
|
produced. False otherwise. This is done so that we don't output the error
|
12054 |
|
|
or warning twice or three times. */
|
12055 |
|
|
|
12056 |
|
|
bool
|
12057 |
|
|
fold_builtin_next_arg (tree exp, bool va_start_p)
|
12058 |
|
|
{
|
12059 |
|
|
tree fntype = TREE_TYPE (current_function_decl);
|
12060 |
|
|
int nargs = call_expr_nargs (exp);
|
12061 |
|
|
tree arg;
|
12062 |
|
|
|
12063 |
|
|
if (!stdarg_p (fntype))
|
12064 |
|
|
{
|
12065 |
|
|
error ("%<va_start%> used in function with fixed args");
|
12066 |
|
|
return true;
|
12067 |
|
|
}
|
12068 |
|
|
|
12069 |
|
|
if (va_start_p)
|
12070 |
|
|
{
|
12071 |
|
|
if (va_start_p && (nargs != 2))
|
12072 |
|
|
{
|
12073 |
|
|
error ("wrong number of arguments to function %<va_start%>");
|
12074 |
|
|
return true;
|
12075 |
|
|
}
|
12076 |
|
|
arg = CALL_EXPR_ARG (exp, 1);
|
12077 |
|
|
}
|
12078 |
|
|
/* We use __builtin_va_start (ap, 0, 0) or __builtin_next_arg (0, 0)
|
12079 |
|
|
when we checked the arguments and if needed issued a warning. */
|
12080 |
|
|
else
|
12081 |
|
|
{
|
12082 |
|
|
if (nargs == 0)
|
12083 |
|
|
{
|
12084 |
|
|
/* Evidently an out of date version of <stdarg.h>; can't validate
|
12085 |
|
|
va_start's second argument, but can still work as intended. */
|
12086 |
|
|
warning (0, "%<__builtin_next_arg%> called without an argument");
|
12087 |
|
|
return true;
|
12088 |
|
|
}
|
12089 |
|
|
else if (nargs > 1)
|
12090 |
|
|
{
|
12091 |
|
|
error ("wrong number of arguments to function %<__builtin_next_arg%>");
|
12092 |
|
|
return true;
|
12093 |
|
|
}
|
12094 |
|
|
arg = CALL_EXPR_ARG (exp, 0);
|
12095 |
|
|
}
|
12096 |
|
|
|
12097 |
|
|
if (TREE_CODE (arg) == SSA_NAME)
|
12098 |
|
|
arg = SSA_NAME_VAR (arg);
|
12099 |
|
|
|
12100 |
|
|
/* We destructively modify the call to be __builtin_va_start (ap, 0)
|
12101 |
|
|
or __builtin_next_arg (0) the first time we see it, after checking
|
12102 |
|
|
the arguments and if needed issuing a warning. */
|
12103 |
|
|
if (!integer_zerop (arg))
|
12104 |
|
|
{
|
12105 |
|
|
tree last_parm = tree_last (DECL_ARGUMENTS (current_function_decl));
|
12106 |
|
|
|
12107 |
|
|
/* Strip off all nops for the sake of the comparison. This
|
12108 |
|
|
is not quite the same as STRIP_NOPS. It does more.
|
12109 |
|
|
We must also strip off INDIRECT_EXPR for C++ reference
|
12110 |
|
|
parameters. */
|
12111 |
|
|
while (CONVERT_EXPR_P (arg)
|
12112 |
|
|
|| TREE_CODE (arg) == INDIRECT_REF)
|
12113 |
|
|
arg = TREE_OPERAND (arg, 0);
|
12114 |
|
|
if (arg != last_parm)
|
12115 |
|
|
{
|
12116 |
|
|
/* FIXME: Sometimes with the tree optimizers we can get the
|
12117 |
|
|
not the last argument even though the user used the last
|
12118 |
|
|
argument. We just warn and set the arg to be the last
|
12119 |
|
|
argument so that we will get wrong-code because of
|
12120 |
|
|
it. */
|
12121 |
|
|
warning (0, "second parameter of %<va_start%> not last named argument");
|
12122 |
|
|
}
|
12123 |
|
|
|
12124 |
|
|
/* Undefined by C99 7.15.1.4p4 (va_start):
|
12125 |
|
|
"If the parameter parmN is declared with the register storage
|
12126 |
|
|
class, with a function or array type, or with a type that is
|
12127 |
|
|
not compatible with the type that results after application of
|
12128 |
|
|
the default argument promotions, the behavior is undefined."
|
12129 |
|
|
*/
|
12130 |
|
|
else if (DECL_REGISTER (arg))
|
12131 |
|
|
warning (0, "undefined behaviour when second parameter of "
|
12132 |
|
|
"%<va_start%> is declared with %<register%> storage");
|
12133 |
|
|
|
12134 |
|
|
/* We want to verify the second parameter just once before the tree
|
12135 |
|
|
optimizers are run and then avoid keeping it in the tree,
|
12136 |
|
|
as otherwise we could warn even for correct code like:
|
12137 |
|
|
void foo (int i, ...)
|
12138 |
|
|
{ va_list ap; i++; va_start (ap, i); va_end (ap); } */
|
12139 |
|
|
if (va_start_p)
|
12140 |
|
|
CALL_EXPR_ARG (exp, 1) = integer_zero_node;
|
12141 |
|
|
else
|
12142 |
|
|
CALL_EXPR_ARG (exp, 0) = integer_zero_node;
|
12143 |
|
|
}
|
12144 |
|
|
return false;
|
12145 |
|
|
}
|
12146 |
|
|
|
12147 |
|
|
|
12148 |
|
|
/* Simplify a call to the sprintf builtin with arguments DEST, FMT, and ORIG.
|
12149 |
|
|
ORIG may be null if this is a 2-argument call. We don't attempt to
|
12150 |
|
|
simplify calls with more than 3 arguments.
|
12151 |
|
|
|
12152 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
12153 |
|
|
simplified form of the call as a tree. If IGNORED is true, it means that
|
12154 |
|
|
the caller does not use the returned value of the function. */
|
12155 |
|
|
|
12156 |
|
|
static tree
|
12157 |
|
|
fold_builtin_sprintf (location_t loc, tree dest, tree fmt,
|
12158 |
|
|
tree orig, int ignored)
|
12159 |
|
|
{
|
12160 |
|
|
tree call, retval;
|
12161 |
|
|
const char *fmt_str = NULL;
|
12162 |
|
|
|
12163 |
|
|
/* Verify the required arguments in the original call. We deal with two
|
12164 |
|
|
types of sprintf() calls: 'sprintf (str, fmt)' and
|
12165 |
|
|
'sprintf (dest, "%s", orig)'. */
|
12166 |
|
|
if (!validate_arg (dest, POINTER_TYPE)
|
12167 |
|
|
|| !validate_arg (fmt, POINTER_TYPE))
|
12168 |
|
|
return NULL_TREE;
|
12169 |
|
|
if (orig && !validate_arg (orig, POINTER_TYPE))
|
12170 |
|
|
return NULL_TREE;
|
12171 |
|
|
|
12172 |
|
|
/* Check whether the format is a literal string constant. */
|
12173 |
|
|
fmt_str = c_getstr (fmt);
|
12174 |
|
|
if (fmt_str == NULL)
|
12175 |
|
|
return NULL_TREE;
|
12176 |
|
|
|
12177 |
|
|
call = NULL_TREE;
|
12178 |
|
|
retval = NULL_TREE;
|
12179 |
|
|
|
12180 |
|
|
if (!init_target_chars ())
|
12181 |
|
|
return NULL_TREE;
|
12182 |
|
|
|
12183 |
|
|
/* If the format doesn't contain % args or %%, use strcpy. */
|
12184 |
|
|
if (strchr (fmt_str, target_percent) == NULL)
|
12185 |
|
|
{
|
12186 |
|
|
tree fn = builtin_decl_implicit (BUILT_IN_STRCPY);
|
12187 |
|
|
|
12188 |
|
|
if (!fn)
|
12189 |
|
|
return NULL_TREE;
|
12190 |
|
|
|
12191 |
|
|
/* Don't optimize sprintf (buf, "abc", ptr++). */
|
12192 |
|
|
if (orig)
|
12193 |
|
|
return NULL_TREE;
|
12194 |
|
|
|
12195 |
|
|
/* Convert sprintf (str, fmt) into strcpy (str, fmt) when
|
12196 |
|
|
'format' is known to contain no % formats. */
|
12197 |
|
|
call = build_call_expr_loc (loc, fn, 2, dest, fmt);
|
12198 |
|
|
if (!ignored)
|
12199 |
|
|
retval = build_int_cst (integer_type_node, strlen (fmt_str));
|
12200 |
|
|
}
|
12201 |
|
|
|
12202 |
|
|
/* If the format is "%s", use strcpy if the result isn't used. */
|
12203 |
|
|
else if (fmt_str && strcmp (fmt_str, target_percent_s) == 0)
|
12204 |
|
|
{
|
12205 |
|
|
tree fn;
|
12206 |
|
|
fn = builtin_decl_implicit (BUILT_IN_STRCPY);
|
12207 |
|
|
|
12208 |
|
|
if (!fn)
|
12209 |
|
|
return NULL_TREE;
|
12210 |
|
|
|
12211 |
|
|
/* Don't crash on sprintf (str1, "%s"). */
|
12212 |
|
|
if (!orig)
|
12213 |
|
|
return NULL_TREE;
|
12214 |
|
|
|
12215 |
|
|
/* Convert sprintf (str1, "%s", str2) into strcpy (str1, str2). */
|
12216 |
|
|
if (!ignored)
|
12217 |
|
|
{
|
12218 |
|
|
retval = c_strlen (orig, 1);
|
12219 |
|
|
if (!retval || TREE_CODE (retval) != INTEGER_CST)
|
12220 |
|
|
return NULL_TREE;
|
12221 |
|
|
}
|
12222 |
|
|
call = build_call_expr_loc (loc, fn, 2, dest, orig);
|
12223 |
|
|
}
|
12224 |
|
|
|
12225 |
|
|
if (call && retval)
|
12226 |
|
|
{
|
12227 |
|
|
retval = fold_convert_loc
|
12228 |
|
|
(loc, TREE_TYPE (TREE_TYPE (builtin_decl_implicit (BUILT_IN_SPRINTF))),
|
12229 |
|
|
retval);
|
12230 |
|
|
return build2 (COMPOUND_EXPR, TREE_TYPE (retval), call, retval);
|
12231 |
|
|
}
|
12232 |
|
|
else
|
12233 |
|
|
return call;
|
12234 |
|
|
}
|
12235 |
|
|
|
12236 |
|
|
/* Simplify a call to the snprintf builtin with arguments DEST, DESTSIZE,
|
12237 |
|
|
FMT, and ORIG. ORIG may be null if this is a 3-argument call. We don't
|
12238 |
|
|
attempt to simplify calls with more than 4 arguments.
|
12239 |
|
|
|
12240 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
12241 |
|
|
simplified form of the call as a tree. If IGNORED is true, it means that
|
12242 |
|
|
the caller does not use the returned value of the function. */
|
12243 |
|
|
|
12244 |
|
|
static tree
|
12245 |
|
|
fold_builtin_snprintf (location_t loc, tree dest, tree destsize, tree fmt,
|
12246 |
|
|
tree orig, int ignored)
|
12247 |
|
|
{
|
12248 |
|
|
tree call, retval;
|
12249 |
|
|
const char *fmt_str = NULL;
|
12250 |
|
|
unsigned HOST_WIDE_INT destlen;
|
12251 |
|
|
|
12252 |
|
|
/* Verify the required arguments in the original call. We deal with two
|
12253 |
|
|
types of snprintf() calls: 'snprintf (str, cst, fmt)' and
|
12254 |
|
|
'snprintf (dest, cst, "%s", orig)'. */
|
12255 |
|
|
if (!validate_arg (dest, POINTER_TYPE)
|
12256 |
|
|
|| !validate_arg (destsize, INTEGER_TYPE)
|
12257 |
|
|
|| !validate_arg (fmt, POINTER_TYPE))
|
12258 |
|
|
return NULL_TREE;
|
12259 |
|
|
if (orig && !validate_arg (orig, POINTER_TYPE))
|
12260 |
|
|
return NULL_TREE;
|
12261 |
|
|
|
12262 |
|
|
if (!host_integerp (destsize, 1))
|
12263 |
|
|
return NULL_TREE;
|
12264 |
|
|
|
12265 |
|
|
/* Check whether the format is a literal string constant. */
|
12266 |
|
|
fmt_str = c_getstr (fmt);
|
12267 |
|
|
if (fmt_str == NULL)
|
12268 |
|
|
return NULL_TREE;
|
12269 |
|
|
|
12270 |
|
|
call = NULL_TREE;
|
12271 |
|
|
retval = NULL_TREE;
|
12272 |
|
|
|
12273 |
|
|
if (!init_target_chars ())
|
12274 |
|
|
return NULL_TREE;
|
12275 |
|
|
|
12276 |
|
|
destlen = tree_low_cst (destsize, 1);
|
12277 |
|
|
|
12278 |
|
|
/* If the format doesn't contain % args or %%, use strcpy. */
|
12279 |
|
|
if (strchr (fmt_str, target_percent) == NULL)
|
12280 |
|
|
{
|
12281 |
|
|
tree fn = builtin_decl_implicit (BUILT_IN_STRCPY);
|
12282 |
|
|
size_t len = strlen (fmt_str);
|
12283 |
|
|
|
12284 |
|
|
/* Don't optimize snprintf (buf, 4, "abc", ptr++). */
|
12285 |
|
|
if (orig)
|
12286 |
|
|
return NULL_TREE;
|
12287 |
|
|
|
12288 |
|
|
/* We could expand this as
|
12289 |
|
|
memcpy (str, fmt, cst - 1); str[cst - 1] = '\0';
|
12290 |
|
|
or to
|
12291 |
|
|
memcpy (str, fmt_with_nul_at_cstm1, cst);
|
12292 |
|
|
but in the former case that might increase code size
|
12293 |
|
|
and in the latter case grow .rodata section too much.
|
12294 |
|
|
So punt for now. */
|
12295 |
|
|
if (len >= destlen)
|
12296 |
|
|
return NULL_TREE;
|
12297 |
|
|
|
12298 |
|
|
if (!fn)
|
12299 |
|
|
return NULL_TREE;
|
12300 |
|
|
|
12301 |
|
|
/* Convert snprintf (str, cst, fmt) into strcpy (str, fmt) when
|
12302 |
|
|
'format' is known to contain no % formats and
|
12303 |
|
|
strlen (fmt) < cst. */
|
12304 |
|
|
call = build_call_expr_loc (loc, fn, 2, dest, fmt);
|
12305 |
|
|
|
12306 |
|
|
if (!ignored)
|
12307 |
|
|
retval = build_int_cst (integer_type_node, strlen (fmt_str));
|
12308 |
|
|
}
|
12309 |
|
|
|
12310 |
|
|
/* If the format is "%s", use strcpy if the result isn't used. */
|
12311 |
|
|
else if (fmt_str && strcmp (fmt_str, target_percent_s) == 0)
|
12312 |
|
|
{
|
12313 |
|
|
tree fn = builtin_decl_implicit (BUILT_IN_STRCPY);
|
12314 |
|
|
unsigned HOST_WIDE_INT origlen;
|
12315 |
|
|
|
12316 |
|
|
/* Don't crash on snprintf (str1, cst, "%s"). */
|
12317 |
|
|
if (!orig)
|
12318 |
|
|
return NULL_TREE;
|
12319 |
|
|
|
12320 |
|
|
retval = c_strlen (orig, 1);
|
12321 |
|
|
if (!retval || !host_integerp (retval, 1))
|
12322 |
|
|
return NULL_TREE;
|
12323 |
|
|
|
12324 |
|
|
origlen = tree_low_cst (retval, 1);
|
12325 |
|
|
/* We could expand this as
|
12326 |
|
|
memcpy (str1, str2, cst - 1); str1[cst - 1] = '\0';
|
12327 |
|
|
or to
|
12328 |
|
|
memcpy (str1, str2_with_nul_at_cstm1, cst);
|
12329 |
|
|
but in the former case that might increase code size
|
12330 |
|
|
and in the latter case grow .rodata section too much.
|
12331 |
|
|
So punt for now. */
|
12332 |
|
|
if (origlen >= destlen)
|
12333 |
|
|
return NULL_TREE;
|
12334 |
|
|
|
12335 |
|
|
/* Convert snprintf (str1, cst, "%s", str2) into
|
12336 |
|
|
strcpy (str1, str2) if strlen (str2) < cst. */
|
12337 |
|
|
if (!fn)
|
12338 |
|
|
return NULL_TREE;
|
12339 |
|
|
|
12340 |
|
|
call = build_call_expr_loc (loc, fn, 2, dest, orig);
|
12341 |
|
|
|
12342 |
|
|
if (ignored)
|
12343 |
|
|
retval = NULL_TREE;
|
12344 |
|
|
}
|
12345 |
|
|
|
12346 |
|
|
if (call && retval)
|
12347 |
|
|
{
|
12348 |
|
|
tree fn = builtin_decl_explicit (BUILT_IN_SNPRINTF);
|
12349 |
|
|
retval = fold_convert_loc (loc, TREE_TYPE (TREE_TYPE (fn)), retval);
|
12350 |
|
|
return build2 (COMPOUND_EXPR, TREE_TYPE (retval), call, retval);
|
12351 |
|
|
}
|
12352 |
|
|
else
|
12353 |
|
|
return call;
|
12354 |
|
|
}
|
12355 |
|
|
|
12356 |
|
|
/* Expand a call EXP to __builtin_object_size. */
|
12357 |
|
|
|
12358 |
|
|
rtx
|
12359 |
|
|
expand_builtin_object_size (tree exp)
|
12360 |
|
|
{
|
12361 |
|
|
tree ost;
|
12362 |
|
|
int object_size_type;
|
12363 |
|
|
tree fndecl = get_callee_fndecl (exp);
|
12364 |
|
|
|
12365 |
|
|
if (!validate_arglist (exp, POINTER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
12366 |
|
|
{
|
12367 |
|
|
error ("%Kfirst argument of %D must be a pointer, second integer constant",
|
12368 |
|
|
exp, fndecl);
|
12369 |
|
|
expand_builtin_trap ();
|
12370 |
|
|
return const0_rtx;
|
12371 |
|
|
}
|
12372 |
|
|
|
12373 |
|
|
ost = CALL_EXPR_ARG (exp, 1);
|
12374 |
|
|
STRIP_NOPS (ost);
|
12375 |
|
|
|
12376 |
|
|
if (TREE_CODE (ost) != INTEGER_CST
|
12377 |
|
|
|| tree_int_cst_sgn (ost) < 0
|
12378 |
|
|
|| compare_tree_int (ost, 3) > 0)
|
12379 |
|
|
{
|
12380 |
|
|
error ("%Klast argument of %D is not integer constant between 0 and 3",
|
12381 |
|
|
exp, fndecl);
|
12382 |
|
|
expand_builtin_trap ();
|
12383 |
|
|
return const0_rtx;
|
12384 |
|
|
}
|
12385 |
|
|
|
12386 |
|
|
object_size_type = tree_low_cst (ost, 0);
|
12387 |
|
|
|
12388 |
|
|
return object_size_type < 2 ? constm1_rtx : const0_rtx;
|
12389 |
|
|
}
|
12390 |
|
|
|
12391 |
|
|
/* Expand EXP, a call to the __mem{cpy,pcpy,move,set}_chk builtin.
|
12392 |
|
|
FCODE is the BUILT_IN_* to use.
|
12393 |
|
|
Return NULL_RTX if we failed; the caller should emit a normal call,
|
12394 |
|
|
otherwise try to get the result in TARGET, if convenient (and in
|
12395 |
|
|
mode MODE if that's convenient). */
|
12396 |
|
|
|
12397 |
|
|
static rtx
|
12398 |
|
|
expand_builtin_memory_chk (tree exp, rtx target, enum machine_mode mode,
|
12399 |
|
|
enum built_in_function fcode)
|
12400 |
|
|
{
|
12401 |
|
|
tree dest, src, len, size;
|
12402 |
|
|
|
12403 |
|
|
if (!validate_arglist (exp,
|
12404 |
|
|
POINTER_TYPE,
|
12405 |
|
|
fcode == BUILT_IN_MEMSET_CHK
|
12406 |
|
|
? INTEGER_TYPE : POINTER_TYPE,
|
12407 |
|
|
INTEGER_TYPE, INTEGER_TYPE, VOID_TYPE))
|
12408 |
|
|
return NULL_RTX;
|
12409 |
|
|
|
12410 |
|
|
dest = CALL_EXPR_ARG (exp, 0);
|
12411 |
|
|
src = CALL_EXPR_ARG (exp, 1);
|
12412 |
|
|
len = CALL_EXPR_ARG (exp, 2);
|
12413 |
|
|
size = CALL_EXPR_ARG (exp, 3);
|
12414 |
|
|
|
12415 |
|
|
if (! host_integerp (size, 1))
|
12416 |
|
|
return NULL_RTX;
|
12417 |
|
|
|
12418 |
|
|
if (host_integerp (len, 1) || integer_all_onesp (size))
|
12419 |
|
|
{
|
12420 |
|
|
tree fn;
|
12421 |
|
|
|
12422 |
|
|
if (! integer_all_onesp (size) && tree_int_cst_lt (size, len))
|
12423 |
|
|
{
|
12424 |
|
|
warning_at (tree_nonartificial_location (exp),
|
12425 |
|
|
0, "%Kcall to %D will always overflow destination buffer",
|
12426 |
|
|
exp, get_callee_fndecl (exp));
|
12427 |
|
|
return NULL_RTX;
|
12428 |
|
|
}
|
12429 |
|
|
|
12430 |
|
|
fn = NULL_TREE;
|
12431 |
|
|
/* If __builtin_mem{cpy,pcpy,move,set}_chk is used, assume
|
12432 |
|
|
mem{cpy,pcpy,move,set} is available. */
|
12433 |
|
|
switch (fcode)
|
12434 |
|
|
{
|
12435 |
|
|
case BUILT_IN_MEMCPY_CHK:
|
12436 |
|
|
fn = builtin_decl_explicit (BUILT_IN_MEMCPY);
|
12437 |
|
|
break;
|
12438 |
|
|
case BUILT_IN_MEMPCPY_CHK:
|
12439 |
|
|
fn = builtin_decl_explicit (BUILT_IN_MEMPCPY);
|
12440 |
|
|
break;
|
12441 |
|
|
case BUILT_IN_MEMMOVE_CHK:
|
12442 |
|
|
fn = builtin_decl_explicit (BUILT_IN_MEMMOVE);
|
12443 |
|
|
break;
|
12444 |
|
|
case BUILT_IN_MEMSET_CHK:
|
12445 |
|
|
fn = builtin_decl_explicit (BUILT_IN_MEMSET);
|
12446 |
|
|
break;
|
12447 |
|
|
default:
|
12448 |
|
|
break;
|
12449 |
|
|
}
|
12450 |
|
|
|
12451 |
|
|
if (! fn)
|
12452 |
|
|
return NULL_RTX;
|
12453 |
|
|
|
12454 |
|
|
fn = build_call_nofold_loc (EXPR_LOCATION (exp), fn, 3, dest, src, len);
|
12455 |
|
|
gcc_assert (TREE_CODE (fn) == CALL_EXPR);
|
12456 |
|
|
CALL_EXPR_TAILCALL (fn) = CALL_EXPR_TAILCALL (exp);
|
12457 |
|
|
return expand_expr (fn, target, mode, EXPAND_NORMAL);
|
12458 |
|
|
}
|
12459 |
|
|
else if (fcode == BUILT_IN_MEMSET_CHK)
|
12460 |
|
|
return NULL_RTX;
|
12461 |
|
|
else
|
12462 |
|
|
{
|
12463 |
|
|
unsigned int dest_align = get_pointer_alignment (dest);
|
12464 |
|
|
|
12465 |
|
|
/* If DEST is not a pointer type, call the normal function. */
|
12466 |
|
|
if (dest_align == 0)
|
12467 |
|
|
return NULL_RTX;
|
12468 |
|
|
|
12469 |
|
|
/* If SRC and DEST are the same (and not volatile), do nothing. */
|
12470 |
|
|
if (operand_equal_p (src, dest, 0))
|
12471 |
|
|
{
|
12472 |
|
|
tree expr;
|
12473 |
|
|
|
12474 |
|
|
if (fcode != BUILT_IN_MEMPCPY_CHK)
|
12475 |
|
|
{
|
12476 |
|
|
/* Evaluate and ignore LEN in case it has side-effects. */
|
12477 |
|
|
expand_expr (len, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
12478 |
|
|
return expand_expr (dest, target, mode, EXPAND_NORMAL);
|
12479 |
|
|
}
|
12480 |
|
|
|
12481 |
|
|
expr = fold_build_pointer_plus (dest, len);
|
12482 |
|
|
return expand_expr (expr, target, mode, EXPAND_NORMAL);
|
12483 |
|
|
}
|
12484 |
|
|
|
12485 |
|
|
/* __memmove_chk special case. */
|
12486 |
|
|
if (fcode == BUILT_IN_MEMMOVE_CHK)
|
12487 |
|
|
{
|
12488 |
|
|
unsigned int src_align = get_pointer_alignment (src);
|
12489 |
|
|
|
12490 |
|
|
if (src_align == 0)
|
12491 |
|
|
return NULL_RTX;
|
12492 |
|
|
|
12493 |
|
|
/* If src is categorized for a readonly section we can use
|
12494 |
|
|
normal __memcpy_chk. */
|
12495 |
|
|
if (readonly_data_expr (src))
|
12496 |
|
|
{
|
12497 |
|
|
tree fn = builtin_decl_explicit (BUILT_IN_MEMCPY_CHK);
|
12498 |
|
|
if (!fn)
|
12499 |
|
|
return NULL_RTX;
|
12500 |
|
|
fn = build_call_nofold_loc (EXPR_LOCATION (exp), fn, 4,
|
12501 |
|
|
dest, src, len, size);
|
12502 |
|
|
gcc_assert (TREE_CODE (fn) == CALL_EXPR);
|
12503 |
|
|
CALL_EXPR_TAILCALL (fn) = CALL_EXPR_TAILCALL (exp);
|
12504 |
|
|
return expand_expr (fn, target, mode, EXPAND_NORMAL);
|
12505 |
|
|
}
|
12506 |
|
|
}
|
12507 |
|
|
return NULL_RTX;
|
12508 |
|
|
}
|
12509 |
|
|
}
|
12510 |
|
|
|
12511 |
|
|
/* Emit warning if a buffer overflow is detected at compile time. */
|
12512 |
|
|
|
12513 |
|
|
static void
|
12514 |
|
|
maybe_emit_chk_warning (tree exp, enum built_in_function fcode)
|
12515 |
|
|
{
|
12516 |
|
|
int is_strlen = 0;
|
12517 |
|
|
tree len, size;
|
12518 |
|
|
location_t loc = tree_nonartificial_location (exp);
|
12519 |
|
|
|
12520 |
|
|
switch (fcode)
|
12521 |
|
|
{
|
12522 |
|
|
case BUILT_IN_STRCPY_CHK:
|
12523 |
|
|
case BUILT_IN_STPCPY_CHK:
|
12524 |
|
|
/* For __strcat_chk the warning will be emitted only if overflowing
|
12525 |
|
|
by at least strlen (dest) + 1 bytes. */
|
12526 |
|
|
case BUILT_IN_STRCAT_CHK:
|
12527 |
|
|
len = CALL_EXPR_ARG (exp, 1);
|
12528 |
|
|
size = CALL_EXPR_ARG (exp, 2);
|
12529 |
|
|
is_strlen = 1;
|
12530 |
|
|
break;
|
12531 |
|
|
case BUILT_IN_STRNCAT_CHK:
|
12532 |
|
|
case BUILT_IN_STRNCPY_CHK:
|
12533 |
|
|
case BUILT_IN_STPNCPY_CHK:
|
12534 |
|
|
len = CALL_EXPR_ARG (exp, 2);
|
12535 |
|
|
size = CALL_EXPR_ARG (exp, 3);
|
12536 |
|
|
break;
|
12537 |
|
|
case BUILT_IN_SNPRINTF_CHK:
|
12538 |
|
|
case BUILT_IN_VSNPRINTF_CHK:
|
12539 |
|
|
len = CALL_EXPR_ARG (exp, 1);
|
12540 |
|
|
size = CALL_EXPR_ARG (exp, 3);
|
12541 |
|
|
break;
|
12542 |
|
|
default:
|
12543 |
|
|
gcc_unreachable ();
|
12544 |
|
|
}
|
12545 |
|
|
|
12546 |
|
|
if (!len || !size)
|
12547 |
|
|
return;
|
12548 |
|
|
|
12549 |
|
|
if (! host_integerp (size, 1) || integer_all_onesp (size))
|
12550 |
|
|
return;
|
12551 |
|
|
|
12552 |
|
|
if (is_strlen)
|
12553 |
|
|
{
|
12554 |
|
|
len = c_strlen (len, 1);
|
12555 |
|
|
if (! len || ! host_integerp (len, 1) || tree_int_cst_lt (len, size))
|
12556 |
|
|
return;
|
12557 |
|
|
}
|
12558 |
|
|
else if (fcode == BUILT_IN_STRNCAT_CHK)
|
12559 |
|
|
{
|
12560 |
|
|
tree src = CALL_EXPR_ARG (exp, 1);
|
12561 |
|
|
if (! src || ! host_integerp (len, 1) || tree_int_cst_lt (len, size))
|
12562 |
|
|
return;
|
12563 |
|
|
src = c_strlen (src, 1);
|
12564 |
|
|
if (! src || ! host_integerp (src, 1))
|
12565 |
|
|
{
|
12566 |
|
|
warning_at (loc, 0, "%Kcall to %D might overflow destination buffer",
|
12567 |
|
|
exp, get_callee_fndecl (exp));
|
12568 |
|
|
return;
|
12569 |
|
|
}
|
12570 |
|
|
else if (tree_int_cst_lt (src, size))
|
12571 |
|
|
return;
|
12572 |
|
|
}
|
12573 |
|
|
else if (! host_integerp (len, 1) || ! tree_int_cst_lt (size, len))
|
12574 |
|
|
return;
|
12575 |
|
|
|
12576 |
|
|
warning_at (loc, 0, "%Kcall to %D will always overflow destination buffer",
|
12577 |
|
|
exp, get_callee_fndecl (exp));
|
12578 |
|
|
}
|
12579 |
|
|
|
12580 |
|
|
/* Emit warning if a buffer overflow is detected at compile time
|
12581 |
|
|
in __sprintf_chk/__vsprintf_chk calls. */
|
12582 |
|
|
|
12583 |
|
|
static void
|
12584 |
|
|
maybe_emit_sprintf_chk_warning (tree exp, enum built_in_function fcode)
|
12585 |
|
|
{
|
12586 |
|
|
tree size, len, fmt;
|
12587 |
|
|
const char *fmt_str;
|
12588 |
|
|
int nargs = call_expr_nargs (exp);
|
12589 |
|
|
|
12590 |
|
|
/* Verify the required arguments in the original call. */
|
12591 |
|
|
|
12592 |
|
|
if (nargs < 4)
|
12593 |
|
|
return;
|
12594 |
|
|
size = CALL_EXPR_ARG (exp, 2);
|
12595 |
|
|
fmt = CALL_EXPR_ARG (exp, 3);
|
12596 |
|
|
|
12597 |
|
|
if (! host_integerp (size, 1) || integer_all_onesp (size))
|
12598 |
|
|
return;
|
12599 |
|
|
|
12600 |
|
|
/* Check whether the format is a literal string constant. */
|
12601 |
|
|
fmt_str = c_getstr (fmt);
|
12602 |
|
|
if (fmt_str == NULL)
|
12603 |
|
|
return;
|
12604 |
|
|
|
12605 |
|
|
if (!init_target_chars ())
|
12606 |
|
|
return;
|
12607 |
|
|
|
12608 |
|
|
/* If the format doesn't contain % args or %%, we know its size. */
|
12609 |
|
|
if (strchr (fmt_str, target_percent) == 0)
|
12610 |
|
|
len = build_int_cstu (size_type_node, strlen (fmt_str));
|
12611 |
|
|
/* If the format is "%s" and first ... argument is a string literal,
|
12612 |
|
|
we know it too. */
|
12613 |
|
|
else if (fcode == BUILT_IN_SPRINTF_CHK
|
12614 |
|
|
&& strcmp (fmt_str, target_percent_s) == 0)
|
12615 |
|
|
{
|
12616 |
|
|
tree arg;
|
12617 |
|
|
|
12618 |
|
|
if (nargs < 5)
|
12619 |
|
|
return;
|
12620 |
|
|
arg = CALL_EXPR_ARG (exp, 4);
|
12621 |
|
|
if (! POINTER_TYPE_P (TREE_TYPE (arg)))
|
12622 |
|
|
return;
|
12623 |
|
|
|
12624 |
|
|
len = c_strlen (arg, 1);
|
12625 |
|
|
if (!len || ! host_integerp (len, 1))
|
12626 |
|
|
return;
|
12627 |
|
|
}
|
12628 |
|
|
else
|
12629 |
|
|
return;
|
12630 |
|
|
|
12631 |
|
|
if (! tree_int_cst_lt (len, size))
|
12632 |
|
|
warning_at (tree_nonartificial_location (exp),
|
12633 |
|
|
0, "%Kcall to %D will always overflow destination buffer",
|
12634 |
|
|
exp, get_callee_fndecl (exp));
|
12635 |
|
|
}
|
12636 |
|
|
|
12637 |
|
|
/* Emit warning if a free is called with address of a variable. */
|
12638 |
|
|
|
12639 |
|
|
static void
|
12640 |
|
|
maybe_emit_free_warning (tree exp)
|
12641 |
|
|
{
|
12642 |
|
|
tree arg = CALL_EXPR_ARG (exp, 0);
|
12643 |
|
|
|
12644 |
|
|
STRIP_NOPS (arg);
|
12645 |
|
|
if (TREE_CODE (arg) != ADDR_EXPR)
|
12646 |
|
|
return;
|
12647 |
|
|
|
12648 |
|
|
arg = get_base_address (TREE_OPERAND (arg, 0));
|
12649 |
|
|
if (arg == NULL || INDIRECT_REF_P (arg) || TREE_CODE (arg) == MEM_REF)
|
12650 |
|
|
return;
|
12651 |
|
|
|
12652 |
|
|
if (SSA_VAR_P (arg))
|
12653 |
|
|
warning_at (tree_nonartificial_location (exp), OPT_Wfree_nonheap_object,
|
12654 |
|
|
"%Kattempt to free a non-heap object %qD", exp, arg);
|
12655 |
|
|
else
|
12656 |
|
|
warning_at (tree_nonartificial_location (exp), OPT_Wfree_nonheap_object,
|
12657 |
|
|
"%Kattempt to free a non-heap object", exp);
|
12658 |
|
|
}
|
12659 |
|
|
|
12660 |
|
|
/* Fold a call to __builtin_object_size with arguments PTR and OST,
|
12661 |
|
|
if possible. */
|
12662 |
|
|
|
12663 |
|
|
tree
|
12664 |
|
|
fold_builtin_object_size (tree ptr, tree ost)
|
12665 |
|
|
{
|
12666 |
|
|
unsigned HOST_WIDE_INT bytes;
|
12667 |
|
|
int object_size_type;
|
12668 |
|
|
|
12669 |
|
|
if (!validate_arg (ptr, POINTER_TYPE)
|
12670 |
|
|
|| !validate_arg (ost, INTEGER_TYPE))
|
12671 |
|
|
return NULL_TREE;
|
12672 |
|
|
|
12673 |
|
|
STRIP_NOPS (ost);
|
12674 |
|
|
|
12675 |
|
|
if (TREE_CODE (ost) != INTEGER_CST
|
12676 |
|
|
|| tree_int_cst_sgn (ost) < 0
|
12677 |
|
|
|| compare_tree_int (ost, 3) > 0)
|
12678 |
|
|
return NULL_TREE;
|
12679 |
|
|
|
12680 |
|
|
object_size_type = tree_low_cst (ost, 0);
|
12681 |
|
|
|
12682 |
|
|
/* __builtin_object_size doesn't evaluate side-effects in its arguments;
|
12683 |
|
|
if there are any side-effects, it returns (size_t) -1 for types 0 and 1
|
12684 |
|
|
and (size_t) 0 for types 2 and 3. */
|
12685 |
|
|
if (TREE_SIDE_EFFECTS (ptr))
|
12686 |
|
|
return build_int_cst_type (size_type_node, object_size_type < 2 ? -1 : 0);
|
12687 |
|
|
|
12688 |
|
|
if (TREE_CODE (ptr) == ADDR_EXPR)
|
12689 |
|
|
{
|
12690 |
|
|
bytes = compute_builtin_object_size (ptr, object_size_type);
|
12691 |
|
|
if (double_int_fits_to_tree_p (size_type_node,
|
12692 |
|
|
uhwi_to_double_int (bytes)))
|
12693 |
|
|
return build_int_cstu (size_type_node, bytes);
|
12694 |
|
|
}
|
12695 |
|
|
else if (TREE_CODE (ptr) == SSA_NAME)
|
12696 |
|
|
{
|
12697 |
|
|
/* If object size is not known yet, delay folding until
|
12698 |
|
|
later. Maybe subsequent passes will help determining
|
12699 |
|
|
it. */
|
12700 |
|
|
bytes = compute_builtin_object_size (ptr, object_size_type);
|
12701 |
|
|
if (bytes != (unsigned HOST_WIDE_INT) (object_size_type < 2 ? -1 : 0)
|
12702 |
|
|
&& double_int_fits_to_tree_p (size_type_node,
|
12703 |
|
|
uhwi_to_double_int (bytes)))
|
12704 |
|
|
return build_int_cstu (size_type_node, bytes);
|
12705 |
|
|
}
|
12706 |
|
|
|
12707 |
|
|
return NULL_TREE;
|
12708 |
|
|
}
|
12709 |
|
|
|
12710 |
|
|
/* Fold a call to the __mem{cpy,pcpy,move,set}_chk builtin.
|
12711 |
|
|
DEST, SRC, LEN, and SIZE are the arguments to the call.
|
12712 |
|
|
IGNORE is true, if return value can be ignored. FCODE is the BUILT_IN_*
|
12713 |
|
|
code of the builtin. If MAXLEN is not NULL, it is maximum length
|
12714 |
|
|
passed as third argument. */
|
12715 |
|
|
|
12716 |
|
|
tree
|
12717 |
|
|
fold_builtin_memory_chk (location_t loc, tree fndecl,
|
12718 |
|
|
tree dest, tree src, tree len, tree size,
|
12719 |
|
|
tree maxlen, bool ignore,
|
12720 |
|
|
enum built_in_function fcode)
|
12721 |
|
|
{
|
12722 |
|
|
tree fn;
|
12723 |
|
|
|
12724 |
|
|
if (!validate_arg (dest, POINTER_TYPE)
|
12725 |
|
|
|| !validate_arg (src,
|
12726 |
|
|
(fcode == BUILT_IN_MEMSET_CHK
|
12727 |
|
|
? INTEGER_TYPE : POINTER_TYPE))
|
12728 |
|
|
|| !validate_arg (len, INTEGER_TYPE)
|
12729 |
|
|
|| !validate_arg (size, INTEGER_TYPE))
|
12730 |
|
|
return NULL_TREE;
|
12731 |
|
|
|
12732 |
|
|
/* If SRC and DEST are the same (and not volatile), return DEST
|
12733 |
|
|
(resp. DEST+LEN for __mempcpy_chk). */
|
12734 |
|
|
if (fcode != BUILT_IN_MEMSET_CHK && operand_equal_p (src, dest, 0))
|
12735 |
|
|
{
|
12736 |
|
|
if (fcode != BUILT_IN_MEMPCPY_CHK)
|
12737 |
|
|
return omit_one_operand_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)),
|
12738 |
|
|
dest, len);
|
12739 |
|
|
else
|
12740 |
|
|
{
|
12741 |
|
|
tree temp = fold_build_pointer_plus_loc (loc, dest, len);
|
12742 |
|
|
return fold_convert_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)), temp);
|
12743 |
|
|
}
|
12744 |
|
|
}
|
12745 |
|
|
|
12746 |
|
|
if (! host_integerp (size, 1))
|
12747 |
|
|
return NULL_TREE;
|
12748 |
|
|
|
12749 |
|
|
if (! integer_all_onesp (size))
|
12750 |
|
|
{
|
12751 |
|
|
if (! host_integerp (len, 1))
|
12752 |
|
|
{
|
12753 |
|
|
/* If LEN is not constant, try MAXLEN too.
|
12754 |
|
|
For MAXLEN only allow optimizing into non-_ocs function
|
12755 |
|
|
if SIZE is >= MAXLEN, never convert to __ocs_fail (). */
|
12756 |
|
|
if (maxlen == NULL_TREE || ! host_integerp (maxlen, 1))
|
12757 |
|
|
{
|
12758 |
|
|
if (fcode == BUILT_IN_MEMPCPY_CHK && ignore)
|
12759 |
|
|
{
|
12760 |
|
|
/* (void) __mempcpy_chk () can be optimized into
|
12761 |
|
|
(void) __memcpy_chk (). */
|
12762 |
|
|
fn = builtin_decl_explicit (BUILT_IN_MEMCPY_CHK);
|
12763 |
|
|
if (!fn)
|
12764 |
|
|
return NULL_TREE;
|
12765 |
|
|
|
12766 |
|
|
return build_call_expr_loc (loc, fn, 4, dest, src, len, size);
|
12767 |
|
|
}
|
12768 |
|
|
return NULL_TREE;
|
12769 |
|
|
}
|
12770 |
|
|
}
|
12771 |
|
|
else
|
12772 |
|
|
maxlen = len;
|
12773 |
|
|
|
12774 |
|
|
if (tree_int_cst_lt (size, maxlen))
|
12775 |
|
|
return NULL_TREE;
|
12776 |
|
|
}
|
12777 |
|
|
|
12778 |
|
|
fn = NULL_TREE;
|
12779 |
|
|
/* If __builtin_mem{cpy,pcpy,move,set}_chk is used, assume
|
12780 |
|
|
mem{cpy,pcpy,move,set} is available. */
|
12781 |
|
|
switch (fcode)
|
12782 |
|
|
{
|
12783 |
|
|
case BUILT_IN_MEMCPY_CHK:
|
12784 |
|
|
fn = builtin_decl_explicit (BUILT_IN_MEMCPY);
|
12785 |
|
|
break;
|
12786 |
|
|
case BUILT_IN_MEMPCPY_CHK:
|
12787 |
|
|
fn = builtin_decl_explicit (BUILT_IN_MEMPCPY);
|
12788 |
|
|
break;
|
12789 |
|
|
case BUILT_IN_MEMMOVE_CHK:
|
12790 |
|
|
fn = builtin_decl_explicit (BUILT_IN_MEMMOVE);
|
12791 |
|
|
break;
|
12792 |
|
|
case BUILT_IN_MEMSET_CHK:
|
12793 |
|
|
fn = builtin_decl_explicit (BUILT_IN_MEMSET);
|
12794 |
|
|
break;
|
12795 |
|
|
default:
|
12796 |
|
|
break;
|
12797 |
|
|
}
|
12798 |
|
|
|
12799 |
|
|
if (!fn)
|
12800 |
|
|
return NULL_TREE;
|
12801 |
|
|
|
12802 |
|
|
return build_call_expr_loc (loc, fn, 3, dest, src, len);
|
12803 |
|
|
}
|
12804 |
|
|
|
12805 |
|
|
/* Fold a call to the __st[rp]cpy_chk builtin.
|
12806 |
|
|
DEST, SRC, and SIZE are the arguments to the call.
|
12807 |
|
|
IGNORE is true if return value can be ignored. FCODE is the BUILT_IN_*
|
12808 |
|
|
code of the builtin. If MAXLEN is not NULL, it is maximum length of
|
12809 |
|
|
strings passed as second argument. */
|
12810 |
|
|
|
12811 |
|
|
tree
|
12812 |
|
|
fold_builtin_stxcpy_chk (location_t loc, tree fndecl, tree dest,
|
12813 |
|
|
tree src, tree size,
|
12814 |
|
|
tree maxlen, bool ignore,
|
12815 |
|
|
enum built_in_function fcode)
|
12816 |
|
|
{
|
12817 |
|
|
tree len, fn;
|
12818 |
|
|
|
12819 |
|
|
if (!validate_arg (dest, POINTER_TYPE)
|
12820 |
|
|
|| !validate_arg (src, POINTER_TYPE)
|
12821 |
|
|
|| !validate_arg (size, INTEGER_TYPE))
|
12822 |
|
|
return NULL_TREE;
|
12823 |
|
|
|
12824 |
|
|
/* If SRC and DEST are the same (and not volatile), return DEST. */
|
12825 |
|
|
if (fcode == BUILT_IN_STRCPY_CHK && operand_equal_p (src, dest, 0))
|
12826 |
|
|
return fold_convert_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)), dest);
|
12827 |
|
|
|
12828 |
|
|
if (! host_integerp (size, 1))
|
12829 |
|
|
return NULL_TREE;
|
12830 |
|
|
|
12831 |
|
|
if (! integer_all_onesp (size))
|
12832 |
|
|
{
|
12833 |
|
|
len = c_strlen (src, 1);
|
12834 |
|
|
if (! len || ! host_integerp (len, 1))
|
12835 |
|
|
{
|
12836 |
|
|
/* If LEN is not constant, try MAXLEN too.
|
12837 |
|
|
For MAXLEN only allow optimizing into non-_ocs function
|
12838 |
|
|
if SIZE is >= MAXLEN, never convert to __ocs_fail (). */
|
12839 |
|
|
if (maxlen == NULL_TREE || ! host_integerp (maxlen, 1))
|
12840 |
|
|
{
|
12841 |
|
|
if (fcode == BUILT_IN_STPCPY_CHK)
|
12842 |
|
|
{
|
12843 |
|
|
if (! ignore)
|
12844 |
|
|
return NULL_TREE;
|
12845 |
|
|
|
12846 |
|
|
/* If return value of __stpcpy_chk is ignored,
|
12847 |
|
|
optimize into __strcpy_chk. */
|
12848 |
|
|
fn = builtin_decl_explicit (BUILT_IN_STRCPY_CHK);
|
12849 |
|
|
if (!fn)
|
12850 |
|
|
return NULL_TREE;
|
12851 |
|
|
|
12852 |
|
|
return build_call_expr_loc (loc, fn, 3, dest, src, size);
|
12853 |
|
|
}
|
12854 |
|
|
|
12855 |
|
|
if (! len || TREE_SIDE_EFFECTS (len))
|
12856 |
|
|
return NULL_TREE;
|
12857 |
|
|
|
12858 |
|
|
/* If c_strlen returned something, but not a constant,
|
12859 |
|
|
transform __strcpy_chk into __memcpy_chk. */
|
12860 |
|
|
fn = builtin_decl_explicit (BUILT_IN_MEMCPY_CHK);
|
12861 |
|
|
if (!fn)
|
12862 |
|
|
return NULL_TREE;
|
12863 |
|
|
|
12864 |
|
|
len = fold_convert_loc (loc, size_type_node, len);
|
12865 |
|
|
len = size_binop_loc (loc, PLUS_EXPR, len,
|
12866 |
|
|
build_int_cst (size_type_node, 1));
|
12867 |
|
|
return fold_convert_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)),
|
12868 |
|
|
build_call_expr_loc (loc, fn, 4,
|
12869 |
|
|
dest, src, len, size));
|
12870 |
|
|
}
|
12871 |
|
|
}
|
12872 |
|
|
else
|
12873 |
|
|
maxlen = len;
|
12874 |
|
|
|
12875 |
|
|
if (! tree_int_cst_lt (maxlen, size))
|
12876 |
|
|
return NULL_TREE;
|
12877 |
|
|
}
|
12878 |
|
|
|
12879 |
|
|
/* If __builtin_st{r,p}cpy_chk is used, assume st{r,p}cpy is available. */
|
12880 |
|
|
fn = builtin_decl_explicit (fcode == BUILT_IN_STPCPY_CHK
|
12881 |
|
|
? BUILT_IN_STPCPY : BUILT_IN_STRCPY);
|
12882 |
|
|
if (!fn)
|
12883 |
|
|
return NULL_TREE;
|
12884 |
|
|
|
12885 |
|
|
return build_call_expr_loc (loc, fn, 2, dest, src);
|
12886 |
|
|
}
|
12887 |
|
|
|
12888 |
|
|
/* Fold a call to the __st{r,p}ncpy_chk builtin. DEST, SRC, LEN, and SIZE
|
12889 |
|
|
are the arguments to the call. If MAXLEN is not NULL, it is maximum
|
12890 |
|
|
length passed as third argument. IGNORE is true if return value can be
|
12891 |
|
|
ignored. FCODE is the BUILT_IN_* code of the builtin. */
|
12892 |
|
|
|
12893 |
|
|
tree
|
12894 |
|
|
fold_builtin_stxncpy_chk (location_t loc, tree dest, tree src,
|
12895 |
|
|
tree len, tree size, tree maxlen, bool ignore,
|
12896 |
|
|
enum built_in_function fcode)
|
12897 |
|
|
{
|
12898 |
|
|
tree fn;
|
12899 |
|
|
|
12900 |
|
|
if (!validate_arg (dest, POINTER_TYPE)
|
12901 |
|
|
|| !validate_arg (src, POINTER_TYPE)
|
12902 |
|
|
|| !validate_arg (len, INTEGER_TYPE)
|
12903 |
|
|
|| !validate_arg (size, INTEGER_TYPE))
|
12904 |
|
|
return NULL_TREE;
|
12905 |
|
|
|
12906 |
|
|
if (fcode == BUILT_IN_STPNCPY_CHK && ignore)
|
12907 |
|
|
{
|
12908 |
|
|
/* If return value of __stpncpy_chk is ignored,
|
12909 |
|
|
optimize into __strncpy_chk. */
|
12910 |
|
|
fn = builtin_decl_explicit (BUILT_IN_STRNCPY_CHK);
|
12911 |
|
|
if (fn)
|
12912 |
|
|
return build_call_expr_loc (loc, fn, 4, dest, src, len, size);
|
12913 |
|
|
}
|
12914 |
|
|
|
12915 |
|
|
if (! host_integerp (size, 1))
|
12916 |
|
|
return NULL_TREE;
|
12917 |
|
|
|
12918 |
|
|
if (! integer_all_onesp (size))
|
12919 |
|
|
{
|
12920 |
|
|
if (! host_integerp (len, 1))
|
12921 |
|
|
{
|
12922 |
|
|
/* If LEN is not constant, try MAXLEN too.
|
12923 |
|
|
For MAXLEN only allow optimizing into non-_ocs function
|
12924 |
|
|
if SIZE is >= MAXLEN, never convert to __ocs_fail (). */
|
12925 |
|
|
if (maxlen == NULL_TREE || ! host_integerp (maxlen, 1))
|
12926 |
|
|
return NULL_TREE;
|
12927 |
|
|
}
|
12928 |
|
|
else
|
12929 |
|
|
maxlen = len;
|
12930 |
|
|
|
12931 |
|
|
if (tree_int_cst_lt (size, maxlen))
|
12932 |
|
|
return NULL_TREE;
|
12933 |
|
|
}
|
12934 |
|
|
|
12935 |
|
|
/* If __builtin_st{r,p}ncpy_chk is used, assume st{r,p}ncpy is available. */
|
12936 |
|
|
fn = builtin_decl_explicit (fcode == BUILT_IN_STPNCPY_CHK
|
12937 |
|
|
? BUILT_IN_STPNCPY : BUILT_IN_STRNCPY);
|
12938 |
|
|
if (!fn)
|
12939 |
|
|
return NULL_TREE;
|
12940 |
|
|
|
12941 |
|
|
return build_call_expr_loc (loc, fn, 3, dest, src, len);
|
12942 |
|
|
}
|
12943 |
|
|
|
12944 |
|
|
/* Fold a call to the __strcat_chk builtin FNDECL. DEST, SRC, and SIZE
|
12945 |
|
|
are the arguments to the call. */
|
12946 |
|
|
|
12947 |
|
|
static tree
|
12948 |
|
|
fold_builtin_strcat_chk (location_t loc, tree fndecl, tree dest,
|
12949 |
|
|
tree src, tree size)
|
12950 |
|
|
{
|
12951 |
|
|
tree fn;
|
12952 |
|
|
const char *p;
|
12953 |
|
|
|
12954 |
|
|
if (!validate_arg (dest, POINTER_TYPE)
|
12955 |
|
|
|| !validate_arg (src, POINTER_TYPE)
|
12956 |
|
|
|| !validate_arg (size, INTEGER_TYPE))
|
12957 |
|
|
return NULL_TREE;
|
12958 |
|
|
|
12959 |
|
|
p = c_getstr (src);
|
12960 |
|
|
/* If the SRC parameter is "", return DEST. */
|
12961 |
|
|
if (p && *p == '\0')
|
12962 |
|
|
return omit_one_operand_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)), dest, src);
|
12963 |
|
|
|
12964 |
|
|
if (! host_integerp (size, 1) || ! integer_all_onesp (size))
|
12965 |
|
|
return NULL_TREE;
|
12966 |
|
|
|
12967 |
|
|
/* If __builtin_strcat_chk is used, assume strcat is available. */
|
12968 |
|
|
fn = builtin_decl_explicit (BUILT_IN_STRCAT);
|
12969 |
|
|
if (!fn)
|
12970 |
|
|
return NULL_TREE;
|
12971 |
|
|
|
12972 |
|
|
return build_call_expr_loc (loc, fn, 2, dest, src);
|
12973 |
|
|
}
|
12974 |
|
|
|
12975 |
|
|
/* Fold a call to the __strncat_chk builtin with arguments DEST, SRC,
|
12976 |
|
|
LEN, and SIZE. */
|
12977 |
|
|
|
12978 |
|
|
static tree
|
12979 |
|
|
fold_builtin_strncat_chk (location_t loc, tree fndecl,
|
12980 |
|
|
tree dest, tree src, tree len, tree size)
|
12981 |
|
|
{
|
12982 |
|
|
tree fn;
|
12983 |
|
|
const char *p;
|
12984 |
|
|
|
12985 |
|
|
if (!validate_arg (dest, POINTER_TYPE)
|
12986 |
|
|
|| !validate_arg (src, POINTER_TYPE)
|
12987 |
|
|
|| !validate_arg (size, INTEGER_TYPE)
|
12988 |
|
|
|| !validate_arg (size, INTEGER_TYPE))
|
12989 |
|
|
return NULL_TREE;
|
12990 |
|
|
|
12991 |
|
|
p = c_getstr (src);
|
12992 |
|
|
/* If the SRC parameter is "" or if LEN is 0, return DEST. */
|
12993 |
|
|
if (p && *p == '\0')
|
12994 |
|
|
return omit_one_operand_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)), dest, len);
|
12995 |
|
|
else if (integer_zerop (len))
|
12996 |
|
|
return omit_one_operand_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)), dest, src);
|
12997 |
|
|
|
12998 |
|
|
if (! host_integerp (size, 1))
|
12999 |
|
|
return NULL_TREE;
|
13000 |
|
|
|
13001 |
|
|
if (! integer_all_onesp (size))
|
13002 |
|
|
{
|
13003 |
|
|
tree src_len = c_strlen (src, 1);
|
13004 |
|
|
if (src_len
|
13005 |
|
|
&& host_integerp (src_len, 1)
|
13006 |
|
|
&& host_integerp (len, 1)
|
13007 |
|
|
&& ! tree_int_cst_lt (len, src_len))
|
13008 |
|
|
{
|
13009 |
|
|
/* If LEN >= strlen (SRC), optimize into __strcat_chk. */
|
13010 |
|
|
fn = builtin_decl_explicit (BUILT_IN_STRCAT_CHK);
|
13011 |
|
|
if (!fn)
|
13012 |
|
|
return NULL_TREE;
|
13013 |
|
|
|
13014 |
|
|
return build_call_expr_loc (loc, fn, 3, dest, src, size);
|
13015 |
|
|
}
|
13016 |
|
|
return NULL_TREE;
|
13017 |
|
|
}
|
13018 |
|
|
|
13019 |
|
|
/* If __builtin_strncat_chk is used, assume strncat is available. */
|
13020 |
|
|
fn = builtin_decl_explicit (BUILT_IN_STRNCAT);
|
13021 |
|
|
if (!fn)
|
13022 |
|
|
return NULL_TREE;
|
13023 |
|
|
|
13024 |
|
|
return build_call_expr_loc (loc, fn, 3, dest, src, len);
|
13025 |
|
|
}
|
13026 |
|
|
|
13027 |
|
|
/* Fold a call EXP to __{,v}sprintf_chk having NARGS passed as ARGS.
|
13028 |
|
|
Return NULL_TREE if a normal call should be emitted rather than
|
13029 |
|
|
expanding the function inline. FCODE is either BUILT_IN_SPRINTF_CHK
|
13030 |
|
|
or BUILT_IN_VSPRINTF_CHK. */
|
13031 |
|
|
|
13032 |
|
|
static tree
|
13033 |
|
|
fold_builtin_sprintf_chk_1 (location_t loc, int nargs, tree *args,
|
13034 |
|
|
enum built_in_function fcode)
|
13035 |
|
|
{
|
13036 |
|
|
tree dest, size, len, fn, fmt, flag;
|
13037 |
|
|
const char *fmt_str;
|
13038 |
|
|
|
13039 |
|
|
/* Verify the required arguments in the original call. */
|
13040 |
|
|
if (nargs < 4)
|
13041 |
|
|
return NULL_TREE;
|
13042 |
|
|
dest = args[0];
|
13043 |
|
|
if (!validate_arg (dest, POINTER_TYPE))
|
13044 |
|
|
return NULL_TREE;
|
13045 |
|
|
flag = args[1];
|
13046 |
|
|
if (!validate_arg (flag, INTEGER_TYPE))
|
13047 |
|
|
return NULL_TREE;
|
13048 |
|
|
size = args[2];
|
13049 |
|
|
if (!validate_arg (size, INTEGER_TYPE))
|
13050 |
|
|
return NULL_TREE;
|
13051 |
|
|
fmt = args[3];
|
13052 |
|
|
if (!validate_arg (fmt, POINTER_TYPE))
|
13053 |
|
|
return NULL_TREE;
|
13054 |
|
|
|
13055 |
|
|
if (! host_integerp (size, 1))
|
13056 |
|
|
return NULL_TREE;
|
13057 |
|
|
|
13058 |
|
|
len = NULL_TREE;
|
13059 |
|
|
|
13060 |
|
|
if (!init_target_chars ())
|
13061 |
|
|
return NULL_TREE;
|
13062 |
|
|
|
13063 |
|
|
/* Check whether the format is a literal string constant. */
|
13064 |
|
|
fmt_str = c_getstr (fmt);
|
13065 |
|
|
if (fmt_str != NULL)
|
13066 |
|
|
{
|
13067 |
|
|
/* If the format doesn't contain % args or %%, we know the size. */
|
13068 |
|
|
if (strchr (fmt_str, target_percent) == 0)
|
13069 |
|
|
{
|
13070 |
|
|
if (fcode != BUILT_IN_SPRINTF_CHK || nargs == 4)
|
13071 |
|
|
len = build_int_cstu (size_type_node, strlen (fmt_str));
|
13072 |
|
|
}
|
13073 |
|
|
/* If the format is "%s" and first ... argument is a string literal,
|
13074 |
|
|
we know the size too. */
|
13075 |
|
|
else if (fcode == BUILT_IN_SPRINTF_CHK
|
13076 |
|
|
&& strcmp (fmt_str, target_percent_s) == 0)
|
13077 |
|
|
{
|
13078 |
|
|
tree arg;
|
13079 |
|
|
|
13080 |
|
|
if (nargs == 5)
|
13081 |
|
|
{
|
13082 |
|
|
arg = args[4];
|
13083 |
|
|
if (validate_arg (arg, POINTER_TYPE))
|
13084 |
|
|
{
|
13085 |
|
|
len = c_strlen (arg, 1);
|
13086 |
|
|
if (! len || ! host_integerp (len, 1))
|
13087 |
|
|
len = NULL_TREE;
|
13088 |
|
|
}
|
13089 |
|
|
}
|
13090 |
|
|
}
|
13091 |
|
|
}
|
13092 |
|
|
|
13093 |
|
|
if (! integer_all_onesp (size))
|
13094 |
|
|
{
|
13095 |
|
|
if (! len || ! tree_int_cst_lt (len, size))
|
13096 |
|
|
return NULL_TREE;
|
13097 |
|
|
}
|
13098 |
|
|
|
13099 |
|
|
/* Only convert __{,v}sprintf_chk to {,v}sprintf if flag is 0
|
13100 |
|
|
or if format doesn't contain % chars or is "%s". */
|
13101 |
|
|
if (! integer_zerop (flag))
|
13102 |
|
|
{
|
13103 |
|
|
if (fmt_str == NULL)
|
13104 |
|
|
return NULL_TREE;
|
13105 |
|
|
if (strchr (fmt_str, target_percent) != NULL
|
13106 |
|
|
&& strcmp (fmt_str, target_percent_s))
|
13107 |
|
|
return NULL_TREE;
|
13108 |
|
|
}
|
13109 |
|
|
|
13110 |
|
|
/* If __builtin_{,v}sprintf_chk is used, assume {,v}sprintf is available. */
|
13111 |
|
|
fn = builtin_decl_explicit (fcode == BUILT_IN_VSPRINTF_CHK
|
13112 |
|
|
? BUILT_IN_VSPRINTF : BUILT_IN_SPRINTF);
|
13113 |
|
|
if (!fn)
|
13114 |
|
|
return NULL_TREE;
|
13115 |
|
|
|
13116 |
|
|
return rewrite_call_expr_array (loc, nargs, args, 4, fn, 2, dest, fmt);
|
13117 |
|
|
}
|
13118 |
|
|
|
13119 |
|
|
/* Fold a call EXP to __{,v}sprintf_chk. Return NULL_TREE if
|
13120 |
|
|
a normal call should be emitted rather than expanding the function
|
13121 |
|
|
inline. FCODE is either BUILT_IN_SPRINTF_CHK or BUILT_IN_VSPRINTF_CHK. */
|
13122 |
|
|
|
13123 |
|
|
static tree
|
13124 |
|
|
fold_builtin_sprintf_chk (location_t loc, tree exp,
|
13125 |
|
|
enum built_in_function fcode)
|
13126 |
|
|
{
|
13127 |
|
|
return fold_builtin_sprintf_chk_1 (loc, call_expr_nargs (exp),
|
13128 |
|
|
CALL_EXPR_ARGP (exp), fcode);
|
13129 |
|
|
}
|
13130 |
|
|
|
13131 |
|
|
/* Fold a call EXP to {,v}snprintf having NARGS passed as ARGS. Return
|
13132 |
|
|
NULL_TREE if a normal call should be emitted rather than expanding
|
13133 |
|
|
the function inline. FCODE is either BUILT_IN_SNPRINTF_CHK or
|
13134 |
|
|
BUILT_IN_VSNPRINTF_CHK. If MAXLEN is not NULL, it is maximum length
|
13135 |
|
|
passed as second argument. */
|
13136 |
|
|
|
13137 |
|
|
static tree
|
13138 |
|
|
fold_builtin_snprintf_chk_1 (location_t loc, int nargs, tree *args,
|
13139 |
|
|
tree maxlen, enum built_in_function fcode)
|
13140 |
|
|
{
|
13141 |
|
|
tree dest, size, len, fn, fmt, flag;
|
13142 |
|
|
const char *fmt_str;
|
13143 |
|
|
|
13144 |
|
|
/* Verify the required arguments in the original call. */
|
13145 |
|
|
if (nargs < 5)
|
13146 |
|
|
return NULL_TREE;
|
13147 |
|
|
dest = args[0];
|
13148 |
|
|
if (!validate_arg (dest, POINTER_TYPE))
|
13149 |
|
|
return NULL_TREE;
|
13150 |
|
|
len = args[1];
|
13151 |
|
|
if (!validate_arg (len, INTEGER_TYPE))
|
13152 |
|
|
return NULL_TREE;
|
13153 |
|
|
flag = args[2];
|
13154 |
|
|
if (!validate_arg (flag, INTEGER_TYPE))
|
13155 |
|
|
return NULL_TREE;
|
13156 |
|
|
size = args[3];
|
13157 |
|
|
if (!validate_arg (size, INTEGER_TYPE))
|
13158 |
|
|
return NULL_TREE;
|
13159 |
|
|
fmt = args[4];
|
13160 |
|
|
if (!validate_arg (fmt, POINTER_TYPE))
|
13161 |
|
|
return NULL_TREE;
|
13162 |
|
|
|
13163 |
|
|
if (! host_integerp (size, 1))
|
13164 |
|
|
return NULL_TREE;
|
13165 |
|
|
|
13166 |
|
|
if (! integer_all_onesp (size))
|
13167 |
|
|
{
|
13168 |
|
|
if (! host_integerp (len, 1))
|
13169 |
|
|
{
|
13170 |
|
|
/* If LEN is not constant, try MAXLEN too.
|
13171 |
|
|
For MAXLEN only allow optimizing into non-_ocs function
|
13172 |
|
|
if SIZE is >= MAXLEN, never convert to __ocs_fail (). */
|
13173 |
|
|
if (maxlen == NULL_TREE || ! host_integerp (maxlen, 1))
|
13174 |
|
|
return NULL_TREE;
|
13175 |
|
|
}
|
13176 |
|
|
else
|
13177 |
|
|
maxlen = len;
|
13178 |
|
|
|
13179 |
|
|
if (tree_int_cst_lt (size, maxlen))
|
13180 |
|
|
return NULL_TREE;
|
13181 |
|
|
}
|
13182 |
|
|
|
13183 |
|
|
if (!init_target_chars ())
|
13184 |
|
|
return NULL_TREE;
|
13185 |
|
|
|
13186 |
|
|
/* Only convert __{,v}snprintf_chk to {,v}snprintf if flag is 0
|
13187 |
|
|
or if format doesn't contain % chars or is "%s". */
|
13188 |
|
|
if (! integer_zerop (flag))
|
13189 |
|
|
{
|
13190 |
|
|
fmt_str = c_getstr (fmt);
|
13191 |
|
|
if (fmt_str == NULL)
|
13192 |
|
|
return NULL_TREE;
|
13193 |
|
|
if (strchr (fmt_str, target_percent) != NULL
|
13194 |
|
|
&& strcmp (fmt_str, target_percent_s))
|
13195 |
|
|
return NULL_TREE;
|
13196 |
|
|
}
|
13197 |
|
|
|
13198 |
|
|
/* If __builtin_{,v}snprintf_chk is used, assume {,v}snprintf is
|
13199 |
|
|
available. */
|
13200 |
|
|
fn = builtin_decl_explicit (fcode == BUILT_IN_VSNPRINTF_CHK
|
13201 |
|
|
? BUILT_IN_VSNPRINTF : BUILT_IN_SNPRINTF);
|
13202 |
|
|
if (!fn)
|
13203 |
|
|
return NULL_TREE;
|
13204 |
|
|
|
13205 |
|
|
return rewrite_call_expr_array (loc, nargs, args, 5, fn, 3, dest, len, fmt);
|
13206 |
|
|
}
|
13207 |
|
|
|
13208 |
|
|
/* Fold a call EXP to {,v}snprintf. Return NULL_TREE if
|
13209 |
|
|
a normal call should be emitted rather than expanding the function
|
13210 |
|
|
inline. FCODE is either BUILT_IN_SNPRINTF_CHK or
|
13211 |
|
|
BUILT_IN_VSNPRINTF_CHK. If MAXLEN is not NULL, it is maximum length
|
13212 |
|
|
passed as second argument. */
|
13213 |
|
|
|
13214 |
|
|
tree
|
13215 |
|
|
fold_builtin_snprintf_chk (location_t loc, tree exp, tree maxlen,
|
13216 |
|
|
enum built_in_function fcode)
|
13217 |
|
|
{
|
13218 |
|
|
return fold_builtin_snprintf_chk_1 (loc, call_expr_nargs (exp),
|
13219 |
|
|
CALL_EXPR_ARGP (exp), maxlen, fcode);
|
13220 |
|
|
}
|
13221 |
|
|
|
13222 |
|
|
/* Fold a call to the {,v}printf{,_unlocked} and __{,v}printf_chk builtins.
|
13223 |
|
|
FMT and ARG are the arguments to the call; we don't fold cases with
|
13224 |
|
|
more than 2 arguments, and ARG may be null if this is a 1-argument case.
|
13225 |
|
|
|
13226 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
13227 |
|
|
simplified form of the call as a tree. FCODE is the BUILT_IN_*
|
13228 |
|
|
code of the function to be simplified. */
|
13229 |
|
|
|
13230 |
|
|
static tree
|
13231 |
|
|
fold_builtin_printf (location_t loc, tree fndecl, tree fmt,
|
13232 |
|
|
tree arg, bool ignore,
|
13233 |
|
|
enum built_in_function fcode)
|
13234 |
|
|
{
|
13235 |
|
|
tree fn_putchar, fn_puts, newarg, call = NULL_TREE;
|
13236 |
|
|
const char *fmt_str = NULL;
|
13237 |
|
|
|
13238 |
|
|
/* If the return value is used, don't do the transformation. */
|
13239 |
|
|
if (! ignore)
|
13240 |
|
|
return NULL_TREE;
|
13241 |
|
|
|
13242 |
|
|
/* Verify the required arguments in the original call. */
|
13243 |
|
|
if (!validate_arg (fmt, POINTER_TYPE))
|
13244 |
|
|
return NULL_TREE;
|
13245 |
|
|
|
13246 |
|
|
/* Check whether the format is a literal string constant. */
|
13247 |
|
|
fmt_str = c_getstr (fmt);
|
13248 |
|
|
if (fmt_str == NULL)
|
13249 |
|
|
return NULL_TREE;
|
13250 |
|
|
|
13251 |
|
|
if (fcode == BUILT_IN_PRINTF_UNLOCKED)
|
13252 |
|
|
{
|
13253 |
|
|
/* If we're using an unlocked function, assume the other
|
13254 |
|
|
unlocked functions exist explicitly. */
|
13255 |
|
|
fn_putchar = builtin_decl_explicit (BUILT_IN_PUTCHAR_UNLOCKED);
|
13256 |
|
|
fn_puts = builtin_decl_explicit (BUILT_IN_PUTS_UNLOCKED);
|
13257 |
|
|
}
|
13258 |
|
|
else
|
13259 |
|
|
{
|
13260 |
|
|
fn_putchar = builtin_decl_implicit (BUILT_IN_PUTCHAR);
|
13261 |
|
|
fn_puts = builtin_decl_implicit (BUILT_IN_PUTS);
|
13262 |
|
|
}
|
13263 |
|
|
|
13264 |
|
|
if (!init_target_chars ())
|
13265 |
|
|
return NULL_TREE;
|
13266 |
|
|
|
13267 |
|
|
if (strcmp (fmt_str, target_percent_s) == 0
|
13268 |
|
|
|| strchr (fmt_str, target_percent) == NULL)
|
13269 |
|
|
{
|
13270 |
|
|
const char *str;
|
13271 |
|
|
|
13272 |
|
|
if (strcmp (fmt_str, target_percent_s) == 0)
|
13273 |
|
|
{
|
13274 |
|
|
if (fcode == BUILT_IN_VPRINTF || fcode == BUILT_IN_VPRINTF_CHK)
|
13275 |
|
|
return NULL_TREE;
|
13276 |
|
|
|
13277 |
|
|
if (!arg || !validate_arg (arg, POINTER_TYPE))
|
13278 |
|
|
return NULL_TREE;
|
13279 |
|
|
|
13280 |
|
|
str = c_getstr (arg);
|
13281 |
|
|
if (str == NULL)
|
13282 |
|
|
return NULL_TREE;
|
13283 |
|
|
}
|
13284 |
|
|
else
|
13285 |
|
|
{
|
13286 |
|
|
/* The format specifier doesn't contain any '%' characters. */
|
13287 |
|
|
if (fcode != BUILT_IN_VPRINTF && fcode != BUILT_IN_VPRINTF_CHK
|
13288 |
|
|
&& arg)
|
13289 |
|
|
return NULL_TREE;
|
13290 |
|
|
str = fmt_str;
|
13291 |
|
|
}
|
13292 |
|
|
|
13293 |
|
|
/* If the string was "", printf does nothing. */
|
13294 |
|
|
if (str[0] == '\0')
|
13295 |
|
|
return build_int_cst (TREE_TYPE (TREE_TYPE (fndecl)), 0);
|
13296 |
|
|
|
13297 |
|
|
/* If the string has length of 1, call putchar. */
|
13298 |
|
|
if (str[1] == '\0')
|
13299 |
|
|
{
|
13300 |
|
|
/* Given printf("c"), (where c is any one character,)
|
13301 |
|
|
convert "c"[0] to an int and pass that to the replacement
|
13302 |
|
|
function. */
|
13303 |
|
|
newarg = build_int_cst (integer_type_node, str[0]);
|
13304 |
|
|
if (fn_putchar)
|
13305 |
|
|
call = build_call_expr_loc (loc, fn_putchar, 1, newarg);
|
13306 |
|
|
}
|
13307 |
|
|
else
|
13308 |
|
|
{
|
13309 |
|
|
/* If the string was "string\n", call puts("string"). */
|
13310 |
|
|
size_t len = strlen (str);
|
13311 |
|
|
if ((unsigned char)str[len - 1] == target_newline
|
13312 |
|
|
&& (size_t) (int) len == len
|
13313 |
|
|
&& (int) len > 0)
|
13314 |
|
|
{
|
13315 |
|
|
char *newstr;
|
13316 |
|
|
tree offset_node, string_cst;
|
13317 |
|
|
|
13318 |
|
|
/* Create a NUL-terminated string that's one char shorter
|
13319 |
|
|
than the original, stripping off the trailing '\n'. */
|
13320 |
|
|
newarg = build_string_literal (len, str);
|
13321 |
|
|
string_cst = string_constant (newarg, &offset_node);
|
13322 |
|
|
gcc_checking_assert (string_cst
|
13323 |
|
|
&& (TREE_STRING_LENGTH (string_cst)
|
13324 |
|
|
== (int) len)
|
13325 |
|
|
&& integer_zerop (offset_node)
|
13326 |
|
|
&& (unsigned char)
|
13327 |
|
|
TREE_STRING_POINTER (string_cst)[len - 1]
|
13328 |
|
|
== target_newline);
|
13329 |
|
|
/* build_string_literal creates a new STRING_CST,
|
13330 |
|
|
modify it in place to avoid double copying. */
|
13331 |
|
|
newstr = CONST_CAST (char *, TREE_STRING_POINTER (string_cst));
|
13332 |
|
|
newstr[len - 1] = '\0';
|
13333 |
|
|
if (fn_puts)
|
13334 |
|
|
call = build_call_expr_loc (loc, fn_puts, 1, newarg);
|
13335 |
|
|
}
|
13336 |
|
|
else
|
13337 |
|
|
/* We'd like to arrange to call fputs(string,stdout) here,
|
13338 |
|
|
but we need stdout and don't have a way to get it yet. */
|
13339 |
|
|
return NULL_TREE;
|
13340 |
|
|
}
|
13341 |
|
|
}
|
13342 |
|
|
|
13343 |
|
|
/* The other optimizations can be done only on the non-va_list variants. */
|
13344 |
|
|
else if (fcode == BUILT_IN_VPRINTF || fcode == BUILT_IN_VPRINTF_CHK)
|
13345 |
|
|
return NULL_TREE;
|
13346 |
|
|
|
13347 |
|
|
/* If the format specifier was "%s\n", call __builtin_puts(arg). */
|
13348 |
|
|
else if (strcmp (fmt_str, target_percent_s_newline) == 0)
|
13349 |
|
|
{
|
13350 |
|
|
if (!arg || !validate_arg (arg, POINTER_TYPE))
|
13351 |
|
|
return NULL_TREE;
|
13352 |
|
|
if (fn_puts)
|
13353 |
|
|
call = build_call_expr_loc (loc, fn_puts, 1, arg);
|
13354 |
|
|
}
|
13355 |
|
|
|
13356 |
|
|
/* If the format specifier was "%c", call __builtin_putchar(arg). */
|
13357 |
|
|
else if (strcmp (fmt_str, target_percent_c) == 0)
|
13358 |
|
|
{
|
13359 |
|
|
if (!arg || !validate_arg (arg, INTEGER_TYPE))
|
13360 |
|
|
return NULL_TREE;
|
13361 |
|
|
if (fn_putchar)
|
13362 |
|
|
call = build_call_expr_loc (loc, fn_putchar, 1, arg);
|
13363 |
|
|
}
|
13364 |
|
|
|
13365 |
|
|
if (!call)
|
13366 |
|
|
return NULL_TREE;
|
13367 |
|
|
|
13368 |
|
|
return fold_convert_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)), call);
|
13369 |
|
|
}
|
13370 |
|
|
|
13371 |
|
|
/* Fold a call to the {,v}fprintf{,_unlocked} and __{,v}printf_chk builtins.
|
13372 |
|
|
FP, FMT, and ARG are the arguments to the call. We don't fold calls with
|
13373 |
|
|
more than 3 arguments, and ARG may be null in the 2-argument case.
|
13374 |
|
|
|
13375 |
|
|
Return NULL_TREE if no simplification was possible, otherwise return the
|
13376 |
|
|
simplified form of the call as a tree. FCODE is the BUILT_IN_*
|
13377 |
|
|
code of the function to be simplified. */
|
13378 |
|
|
|
13379 |
|
|
static tree
|
13380 |
|
|
fold_builtin_fprintf (location_t loc, tree fndecl, tree fp,
|
13381 |
|
|
tree fmt, tree arg, bool ignore,
|
13382 |
|
|
enum built_in_function fcode)
|
13383 |
|
|
{
|
13384 |
|
|
tree fn_fputc, fn_fputs, call = NULL_TREE;
|
13385 |
|
|
const char *fmt_str = NULL;
|
13386 |
|
|
|
13387 |
|
|
/* If the return value is used, don't do the transformation. */
|
13388 |
|
|
if (! ignore)
|
13389 |
|
|
return NULL_TREE;
|
13390 |
|
|
|
13391 |
|
|
/* Verify the required arguments in the original call. */
|
13392 |
|
|
if (!validate_arg (fp, POINTER_TYPE))
|
13393 |
|
|
return NULL_TREE;
|
13394 |
|
|
if (!validate_arg (fmt, POINTER_TYPE))
|
13395 |
|
|
return NULL_TREE;
|
13396 |
|
|
|
13397 |
|
|
/* Check whether the format is a literal string constant. */
|
13398 |
|
|
fmt_str = c_getstr (fmt);
|
13399 |
|
|
if (fmt_str == NULL)
|
13400 |
|
|
return NULL_TREE;
|
13401 |
|
|
|
13402 |
|
|
if (fcode == BUILT_IN_FPRINTF_UNLOCKED)
|
13403 |
|
|
{
|
13404 |
|
|
/* If we're using an unlocked function, assume the other
|
13405 |
|
|
unlocked functions exist explicitly. */
|
13406 |
|
|
fn_fputc = builtin_decl_explicit (BUILT_IN_FPUTC_UNLOCKED);
|
13407 |
|
|
fn_fputs = builtin_decl_explicit (BUILT_IN_FPUTS_UNLOCKED);
|
13408 |
|
|
}
|
13409 |
|
|
else
|
13410 |
|
|
{
|
13411 |
|
|
fn_fputc = builtin_decl_implicit (BUILT_IN_FPUTC);
|
13412 |
|
|
fn_fputs = builtin_decl_implicit (BUILT_IN_FPUTS);
|
13413 |
|
|
}
|
13414 |
|
|
|
13415 |
|
|
if (!init_target_chars ())
|
13416 |
|
|
return NULL_TREE;
|
13417 |
|
|
|
13418 |
|
|
/* If the format doesn't contain % args or %%, use strcpy. */
|
13419 |
|
|
if (strchr (fmt_str, target_percent) == NULL)
|
13420 |
|
|
{
|
13421 |
|
|
if (fcode != BUILT_IN_VFPRINTF && fcode != BUILT_IN_VFPRINTF_CHK
|
13422 |
|
|
&& arg)
|
13423 |
|
|
return NULL_TREE;
|
13424 |
|
|
|
13425 |
|
|
/* If the format specifier was "", fprintf does nothing. */
|
13426 |
|
|
if (fmt_str[0] == '\0')
|
13427 |
|
|
{
|
13428 |
|
|
/* If FP has side-effects, just wait until gimplification is
|
13429 |
|
|
done. */
|
13430 |
|
|
if (TREE_SIDE_EFFECTS (fp))
|
13431 |
|
|
return NULL_TREE;
|
13432 |
|
|
|
13433 |
|
|
return build_int_cst (TREE_TYPE (TREE_TYPE (fndecl)), 0);
|
13434 |
|
|
}
|
13435 |
|
|
|
13436 |
|
|
/* When "string" doesn't contain %, replace all cases of
|
13437 |
|
|
fprintf (fp, string) with fputs (string, fp). The fputs
|
13438 |
|
|
builtin will take care of special cases like length == 1. */
|
13439 |
|
|
if (fn_fputs)
|
13440 |
|
|
call = build_call_expr_loc (loc, fn_fputs, 2, fmt, fp);
|
13441 |
|
|
}
|
13442 |
|
|
|
13443 |
|
|
/* The other optimizations can be done only on the non-va_list variants. */
|
13444 |
|
|
else if (fcode == BUILT_IN_VFPRINTF || fcode == BUILT_IN_VFPRINTF_CHK)
|
13445 |
|
|
return NULL_TREE;
|
13446 |
|
|
|
13447 |
|
|
/* If the format specifier was "%s", call __builtin_fputs (arg, fp). */
|
13448 |
|
|
else if (strcmp (fmt_str, target_percent_s) == 0)
|
13449 |
|
|
{
|
13450 |
|
|
if (!arg || !validate_arg (arg, POINTER_TYPE))
|
13451 |
|
|
return NULL_TREE;
|
13452 |
|
|
if (fn_fputs)
|
13453 |
|
|
call = build_call_expr_loc (loc, fn_fputs, 2, arg, fp);
|
13454 |
|
|
}
|
13455 |
|
|
|
13456 |
|
|
/* If the format specifier was "%c", call __builtin_fputc (arg, fp). */
|
13457 |
|
|
else if (strcmp (fmt_str, target_percent_c) == 0)
|
13458 |
|
|
{
|
13459 |
|
|
if (!arg || !validate_arg (arg, INTEGER_TYPE))
|
13460 |
|
|
return NULL_TREE;
|
13461 |
|
|
if (fn_fputc)
|
13462 |
|
|
call = build_call_expr_loc (loc, fn_fputc, 2, arg, fp);
|
13463 |
|
|
}
|
13464 |
|
|
|
13465 |
|
|
if (!call)
|
13466 |
|
|
return NULL_TREE;
|
13467 |
|
|
return fold_convert_loc (loc, TREE_TYPE (TREE_TYPE (fndecl)), call);
|
13468 |
|
|
}
|
13469 |
|
|
|
13470 |
|
|
/* Initialize format string characters in the target charset. */
|
13471 |
|
|
|
13472 |
|
|
static bool
|
13473 |
|
|
init_target_chars (void)
|
13474 |
|
|
{
|
13475 |
|
|
static bool init;
|
13476 |
|
|
if (!init)
|
13477 |
|
|
{
|
13478 |
|
|
target_newline = lang_hooks.to_target_charset ('\n');
|
13479 |
|
|
target_percent = lang_hooks.to_target_charset ('%');
|
13480 |
|
|
target_c = lang_hooks.to_target_charset ('c');
|
13481 |
|
|
target_s = lang_hooks.to_target_charset ('s');
|
13482 |
|
|
if (target_newline == 0 || target_percent == 0 || target_c == 0
|
13483 |
|
|
|| target_s == 0)
|
13484 |
|
|
return false;
|
13485 |
|
|
|
13486 |
|
|
target_percent_c[0] = target_percent;
|
13487 |
|
|
target_percent_c[1] = target_c;
|
13488 |
|
|
target_percent_c[2] = '\0';
|
13489 |
|
|
|
13490 |
|
|
target_percent_s[0] = target_percent;
|
13491 |
|
|
target_percent_s[1] = target_s;
|
13492 |
|
|
target_percent_s[2] = '\0';
|
13493 |
|
|
|
13494 |
|
|
target_percent_s_newline[0] = target_percent;
|
13495 |
|
|
target_percent_s_newline[1] = target_s;
|
13496 |
|
|
target_percent_s_newline[2] = target_newline;
|
13497 |
|
|
target_percent_s_newline[3] = '\0';
|
13498 |
|
|
|
13499 |
|
|
init = true;
|
13500 |
|
|
}
|
13501 |
|
|
return true;
|
13502 |
|
|
}
|
13503 |
|
|
|
13504 |
|
|
/* Helper function for do_mpfr_arg*(). Ensure M is a normal number
|
13505 |
|
|
and no overflow/underflow occurred. INEXACT is true if M was not
|
13506 |
|
|
exactly calculated. TYPE is the tree type for the result. This
|
13507 |
|
|
function assumes that you cleared the MPFR flags and then
|
13508 |
|
|
calculated M to see if anything subsequently set a flag prior to
|
13509 |
|
|
entering this function. Return NULL_TREE if any checks fail. */
|
13510 |
|
|
|
13511 |
|
|
static tree
|
13512 |
|
|
do_mpfr_ckconv (mpfr_srcptr m, tree type, int inexact)
|
13513 |
|
|
{
|
13514 |
|
|
/* Proceed iff we get a normal number, i.e. not NaN or Inf and no
|
13515 |
|
|
overflow/underflow occurred. If -frounding-math, proceed iff the
|
13516 |
|
|
result of calling FUNC was exact. */
|
13517 |
|
|
if (mpfr_number_p (m) && !mpfr_overflow_p () && !mpfr_underflow_p ()
|
13518 |
|
|
&& (!flag_rounding_math || !inexact))
|
13519 |
|
|
{
|
13520 |
|
|
REAL_VALUE_TYPE rr;
|
13521 |
|
|
|
13522 |
|
|
real_from_mpfr (&rr, m, type, GMP_RNDN);
|
13523 |
|
|
/* Proceed iff GCC's REAL_VALUE_TYPE can hold the MPFR value,
|
13524 |
|
|
check for overflow/underflow. If the REAL_VALUE_TYPE is zero
|
13525 |
|
|
but the mpft_t is not, then we underflowed in the
|
13526 |
|
|
conversion. */
|
13527 |
|
|
if (real_isfinite (&rr)
|
13528 |
|
|
&& (rr.cl == rvc_zero) == (mpfr_zero_p (m) != 0))
|
13529 |
|
|
{
|
13530 |
|
|
REAL_VALUE_TYPE rmode;
|
13531 |
|
|
|
13532 |
|
|
real_convert (&rmode, TYPE_MODE (type), &rr);
|
13533 |
|
|
/* Proceed iff the specified mode can hold the value. */
|
13534 |
|
|
if (real_identical (&rmode, &rr))
|
13535 |
|
|
return build_real (type, rmode);
|
13536 |
|
|
}
|
13537 |
|
|
}
|
13538 |
|
|
return NULL_TREE;
|
13539 |
|
|
}
|
13540 |
|
|
|
13541 |
|
|
/* Helper function for do_mpc_arg*(). Ensure M is a normal complex
|
13542 |
|
|
number and no overflow/underflow occurred. INEXACT is true if M
|
13543 |
|
|
was not exactly calculated. TYPE is the tree type for the result.
|
13544 |
|
|
This function assumes that you cleared the MPFR flags and then
|
13545 |
|
|
calculated M to see if anything subsequently set a flag prior to
|
13546 |
|
|
entering this function. Return NULL_TREE if any checks fail, if
|
13547 |
|
|
FORCE_CONVERT is true, then bypass the checks. */
|
13548 |
|
|
|
13549 |
|
|
static tree
|
13550 |
|
|
do_mpc_ckconv (mpc_srcptr m, tree type, int inexact, int force_convert)
|
13551 |
|
|
{
|
13552 |
|
|
/* Proceed iff we get a normal number, i.e. not NaN or Inf and no
|
13553 |
|
|
overflow/underflow occurred. If -frounding-math, proceed iff the
|
13554 |
|
|
result of calling FUNC was exact. */
|
13555 |
|
|
if (force_convert
|
13556 |
|
|
|| (mpfr_number_p (mpc_realref (m)) && mpfr_number_p (mpc_imagref (m))
|
13557 |
|
|
&& !mpfr_overflow_p () && !mpfr_underflow_p ()
|
13558 |
|
|
&& (!flag_rounding_math || !inexact)))
|
13559 |
|
|
{
|
13560 |
|
|
REAL_VALUE_TYPE re, im;
|
13561 |
|
|
|
13562 |
|
|
real_from_mpfr (&re, mpc_realref (m), TREE_TYPE (type), GMP_RNDN);
|
13563 |
|
|
real_from_mpfr (&im, mpc_imagref (m), TREE_TYPE (type), GMP_RNDN);
|
13564 |
|
|
/* Proceed iff GCC's REAL_VALUE_TYPE can hold the MPFR values,
|
13565 |
|
|
check for overflow/underflow. If the REAL_VALUE_TYPE is zero
|
13566 |
|
|
but the mpft_t is not, then we underflowed in the
|
13567 |
|
|
conversion. */
|
13568 |
|
|
if (force_convert
|
13569 |
|
|
|| (real_isfinite (&re) && real_isfinite (&im)
|
13570 |
|
|
&& (re.cl == rvc_zero) == (mpfr_zero_p (mpc_realref (m)) != 0)
|
13571 |
|
|
&& (im.cl == rvc_zero) == (mpfr_zero_p (mpc_imagref (m)) != 0)))
|
13572 |
|
|
{
|
13573 |
|
|
REAL_VALUE_TYPE re_mode, im_mode;
|
13574 |
|
|
|
13575 |
|
|
real_convert (&re_mode, TYPE_MODE (TREE_TYPE (type)), &re);
|
13576 |
|
|
real_convert (&im_mode, TYPE_MODE (TREE_TYPE (type)), &im);
|
13577 |
|
|
/* Proceed iff the specified mode can hold the value. */
|
13578 |
|
|
if (force_convert
|
13579 |
|
|
|| (real_identical (&re_mode, &re)
|
13580 |
|
|
&& real_identical (&im_mode, &im)))
|
13581 |
|
|
return build_complex (type, build_real (TREE_TYPE (type), re_mode),
|
13582 |
|
|
build_real (TREE_TYPE (type), im_mode));
|
13583 |
|
|
}
|
13584 |
|
|
}
|
13585 |
|
|
return NULL_TREE;
|
13586 |
|
|
}
|
13587 |
|
|
|
13588 |
|
|
/* If argument ARG is a REAL_CST, call the one-argument mpfr function
|
13589 |
|
|
FUNC on it and return the resulting value as a tree with type TYPE.
|
13590 |
|
|
If MIN and/or MAX are not NULL, then the supplied ARG must be
|
13591 |
|
|
within those bounds. If INCLUSIVE is true, then MIN/MAX are
|
13592 |
|
|
acceptable values, otherwise they are not. The mpfr precision is
|
13593 |
|
|
set to the precision of TYPE. We assume that function FUNC returns
|
13594 |
|
|
zero if the result could be calculated exactly within the requested
|
13595 |
|
|
precision. */
|
13596 |
|
|
|
13597 |
|
|
static tree
|
13598 |
|
|
do_mpfr_arg1 (tree arg, tree type, int (*func)(mpfr_ptr, mpfr_srcptr, mp_rnd_t),
|
13599 |
|
|
const REAL_VALUE_TYPE *min, const REAL_VALUE_TYPE *max,
|
13600 |
|
|
bool inclusive)
|
13601 |
|
|
{
|
13602 |
|
|
tree result = NULL_TREE;
|
13603 |
|
|
|
13604 |
|
|
STRIP_NOPS (arg);
|
13605 |
|
|
|
13606 |
|
|
/* To proceed, MPFR must exactly represent the target floating point
|
13607 |
|
|
format, which only happens when the target base equals two. */
|
13608 |
|
|
if (REAL_MODE_FORMAT (TYPE_MODE (type))->b == 2
|
13609 |
|
|
&& TREE_CODE (arg) == REAL_CST && !TREE_OVERFLOW (arg))
|
13610 |
|
|
{
|
13611 |
|
|
const REAL_VALUE_TYPE *const ra = &TREE_REAL_CST (arg);
|
13612 |
|
|
|
13613 |
|
|
if (real_isfinite (ra)
|
13614 |
|
|
&& (!min || real_compare (inclusive ? GE_EXPR: GT_EXPR , ra, min))
|
13615 |
|
|
&& (!max || real_compare (inclusive ? LE_EXPR: LT_EXPR , ra, max)))
|
13616 |
|
|
{
|
13617 |
|
|
const struct real_format *fmt = REAL_MODE_FORMAT (TYPE_MODE (type));
|
13618 |
|
|
const int prec = fmt->p;
|
13619 |
|
|
const mp_rnd_t rnd = fmt->round_towards_zero? GMP_RNDZ : GMP_RNDN;
|
13620 |
|
|
int inexact;
|
13621 |
|
|
mpfr_t m;
|
13622 |
|
|
|
13623 |
|
|
mpfr_init2 (m, prec);
|
13624 |
|
|
mpfr_from_real (m, ra, GMP_RNDN);
|
13625 |
|
|
mpfr_clear_flags ();
|
13626 |
|
|
inexact = func (m, m, rnd);
|
13627 |
|
|
result = do_mpfr_ckconv (m, type, inexact);
|
13628 |
|
|
mpfr_clear (m);
|
13629 |
|
|
}
|
13630 |
|
|
}
|
13631 |
|
|
|
13632 |
|
|
return result;
|
13633 |
|
|
}
|
13634 |
|
|
|
13635 |
|
|
/* If argument ARG is a REAL_CST, call the two-argument mpfr function
|
13636 |
|
|
FUNC on it and return the resulting value as a tree with type TYPE.
|
13637 |
|
|
The mpfr precision is set to the precision of TYPE. We assume that
|
13638 |
|
|
function FUNC returns zero if the result could be calculated
|
13639 |
|
|
exactly within the requested precision. */
|
13640 |
|
|
|
13641 |
|
|
static tree
|
13642 |
|
|
do_mpfr_arg2 (tree arg1, tree arg2, tree type,
|
13643 |
|
|
int (*func)(mpfr_ptr, mpfr_srcptr, mpfr_srcptr, mp_rnd_t))
|
13644 |
|
|
{
|
13645 |
|
|
tree result = NULL_TREE;
|
13646 |
|
|
|
13647 |
|
|
STRIP_NOPS (arg1);
|
13648 |
|
|
STRIP_NOPS (arg2);
|
13649 |
|
|
|
13650 |
|
|
/* To proceed, MPFR must exactly represent the target floating point
|
13651 |
|
|
format, which only happens when the target base equals two. */
|
13652 |
|
|
if (REAL_MODE_FORMAT (TYPE_MODE (type))->b == 2
|
13653 |
|
|
&& TREE_CODE (arg1) == REAL_CST && !TREE_OVERFLOW (arg1)
|
13654 |
|
|
&& TREE_CODE (arg2) == REAL_CST && !TREE_OVERFLOW (arg2))
|
13655 |
|
|
{
|
13656 |
|
|
const REAL_VALUE_TYPE *const ra1 = &TREE_REAL_CST (arg1);
|
13657 |
|
|
const REAL_VALUE_TYPE *const ra2 = &TREE_REAL_CST (arg2);
|
13658 |
|
|
|
13659 |
|
|
if (real_isfinite (ra1) && real_isfinite (ra2))
|
13660 |
|
|
{
|
13661 |
|
|
const struct real_format *fmt = REAL_MODE_FORMAT (TYPE_MODE (type));
|
13662 |
|
|
const int prec = fmt->p;
|
13663 |
|
|
const mp_rnd_t rnd = fmt->round_towards_zero? GMP_RNDZ : GMP_RNDN;
|
13664 |
|
|
int inexact;
|
13665 |
|
|
mpfr_t m1, m2;
|
13666 |
|
|
|
13667 |
|
|
mpfr_inits2 (prec, m1, m2, NULL);
|
13668 |
|
|
mpfr_from_real (m1, ra1, GMP_RNDN);
|
13669 |
|
|
mpfr_from_real (m2, ra2, GMP_RNDN);
|
13670 |
|
|
mpfr_clear_flags ();
|
13671 |
|
|
inexact = func (m1, m1, m2, rnd);
|
13672 |
|
|
result = do_mpfr_ckconv (m1, type, inexact);
|
13673 |
|
|
mpfr_clears (m1, m2, NULL);
|
13674 |
|
|
}
|
13675 |
|
|
}
|
13676 |
|
|
|
13677 |
|
|
return result;
|
13678 |
|
|
}
|
13679 |
|
|
|
13680 |
|
|
/* If argument ARG is a REAL_CST, call the three-argument mpfr function
|
13681 |
|
|
FUNC on it and return the resulting value as a tree with type TYPE.
|
13682 |
|
|
The mpfr precision is set to the precision of TYPE. We assume that
|
13683 |
|
|
function FUNC returns zero if the result could be calculated
|
13684 |
|
|
exactly within the requested precision. */
|
13685 |
|
|
|
13686 |
|
|
static tree
|
13687 |
|
|
do_mpfr_arg3 (tree arg1, tree arg2, tree arg3, tree type,
|
13688 |
|
|
int (*func)(mpfr_ptr, mpfr_srcptr, mpfr_srcptr, mpfr_srcptr, mp_rnd_t))
|
13689 |
|
|
{
|
13690 |
|
|
tree result = NULL_TREE;
|
13691 |
|
|
|
13692 |
|
|
STRIP_NOPS (arg1);
|
13693 |
|
|
STRIP_NOPS (arg2);
|
13694 |
|
|
STRIP_NOPS (arg3);
|
13695 |
|
|
|
13696 |
|
|
/* To proceed, MPFR must exactly represent the target floating point
|
13697 |
|
|
format, which only happens when the target base equals two. */
|
13698 |
|
|
if (REAL_MODE_FORMAT (TYPE_MODE (type))->b == 2
|
13699 |
|
|
&& TREE_CODE (arg1) == REAL_CST && !TREE_OVERFLOW (arg1)
|
13700 |
|
|
&& TREE_CODE (arg2) == REAL_CST && !TREE_OVERFLOW (arg2)
|
13701 |
|
|
&& TREE_CODE (arg3) == REAL_CST && !TREE_OVERFLOW (arg3))
|
13702 |
|
|
{
|
13703 |
|
|
const REAL_VALUE_TYPE *const ra1 = &TREE_REAL_CST (arg1);
|
13704 |
|
|
const REAL_VALUE_TYPE *const ra2 = &TREE_REAL_CST (arg2);
|
13705 |
|
|
const REAL_VALUE_TYPE *const ra3 = &TREE_REAL_CST (arg3);
|
13706 |
|
|
|
13707 |
|
|
if (real_isfinite (ra1) && real_isfinite (ra2) && real_isfinite (ra3))
|
13708 |
|
|
{
|
13709 |
|
|
const struct real_format *fmt = REAL_MODE_FORMAT (TYPE_MODE (type));
|
13710 |
|
|
const int prec = fmt->p;
|
13711 |
|
|
const mp_rnd_t rnd = fmt->round_towards_zero? GMP_RNDZ : GMP_RNDN;
|
13712 |
|
|
int inexact;
|
13713 |
|
|
mpfr_t m1, m2, m3;
|
13714 |
|
|
|
13715 |
|
|
mpfr_inits2 (prec, m1, m2, m3, NULL);
|
13716 |
|
|
mpfr_from_real (m1, ra1, GMP_RNDN);
|
13717 |
|
|
mpfr_from_real (m2, ra2, GMP_RNDN);
|
13718 |
|
|
mpfr_from_real (m3, ra3, GMP_RNDN);
|
13719 |
|
|
mpfr_clear_flags ();
|
13720 |
|
|
inexact = func (m1, m1, m2, m3, rnd);
|
13721 |
|
|
result = do_mpfr_ckconv (m1, type, inexact);
|
13722 |
|
|
mpfr_clears (m1, m2, m3, NULL);
|
13723 |
|
|
}
|
13724 |
|
|
}
|
13725 |
|
|
|
13726 |
|
|
return result;
|
13727 |
|
|
}
|
13728 |
|
|
|
13729 |
|
|
/* If argument ARG is a REAL_CST, call mpfr_sin_cos() on it and set
|
13730 |
|
|
the pointers *(ARG_SINP) and *(ARG_COSP) to the resulting values.
|
13731 |
|
|
If ARG_SINP and ARG_COSP are NULL then the result is returned
|
13732 |
|
|
as a complex value.
|
13733 |
|
|
The type is taken from the type of ARG and is used for setting the
|
13734 |
|
|
precision of the calculation and results. */
|
13735 |
|
|
|
13736 |
|
|
static tree
|
13737 |
|
|
do_mpfr_sincos (tree arg, tree arg_sinp, tree arg_cosp)
|
13738 |
|
|
{
|
13739 |
|
|
tree const type = TREE_TYPE (arg);
|
13740 |
|
|
tree result = NULL_TREE;
|
13741 |
|
|
|
13742 |
|
|
STRIP_NOPS (arg);
|
13743 |
|
|
|
13744 |
|
|
/* To proceed, MPFR must exactly represent the target floating point
|
13745 |
|
|
format, which only happens when the target base equals two. */
|
13746 |
|
|
if (REAL_MODE_FORMAT (TYPE_MODE (type))->b == 2
|
13747 |
|
|
&& TREE_CODE (arg) == REAL_CST
|
13748 |
|
|
&& !TREE_OVERFLOW (arg))
|
13749 |
|
|
{
|
13750 |
|
|
const REAL_VALUE_TYPE *const ra = &TREE_REAL_CST (arg);
|
13751 |
|
|
|
13752 |
|
|
if (real_isfinite (ra))
|
13753 |
|
|
{
|
13754 |
|
|
const struct real_format *fmt = REAL_MODE_FORMAT (TYPE_MODE (type));
|
13755 |
|
|
const int prec = fmt->p;
|
13756 |
|
|
const mp_rnd_t rnd = fmt->round_towards_zero? GMP_RNDZ : GMP_RNDN;
|
13757 |
|
|
tree result_s, result_c;
|
13758 |
|
|
int inexact;
|
13759 |
|
|
mpfr_t m, ms, mc;
|
13760 |
|
|
|
13761 |
|
|
mpfr_inits2 (prec, m, ms, mc, NULL);
|
13762 |
|
|
mpfr_from_real (m, ra, GMP_RNDN);
|
13763 |
|
|
mpfr_clear_flags ();
|
13764 |
|
|
inexact = mpfr_sin_cos (ms, mc, m, rnd);
|
13765 |
|
|
result_s = do_mpfr_ckconv (ms, type, inexact);
|
13766 |
|
|
result_c = do_mpfr_ckconv (mc, type, inexact);
|
13767 |
|
|
mpfr_clears (m, ms, mc, NULL);
|
13768 |
|
|
if (result_s && result_c)
|
13769 |
|
|
{
|
13770 |
|
|
/* If we are to return in a complex value do so. */
|
13771 |
|
|
if (!arg_sinp && !arg_cosp)
|
13772 |
|
|
return build_complex (build_complex_type (type),
|
13773 |
|
|
result_c, result_s);
|
13774 |
|
|
|
13775 |
|
|
/* Dereference the sin/cos pointer arguments. */
|
13776 |
|
|
arg_sinp = build_fold_indirect_ref (arg_sinp);
|
13777 |
|
|
arg_cosp = build_fold_indirect_ref (arg_cosp);
|
13778 |
|
|
/* Proceed if valid pointer type were passed in. */
|
13779 |
|
|
if (TYPE_MAIN_VARIANT (TREE_TYPE (arg_sinp)) == TYPE_MAIN_VARIANT (type)
|
13780 |
|
|
&& TYPE_MAIN_VARIANT (TREE_TYPE (arg_cosp)) == TYPE_MAIN_VARIANT (type))
|
13781 |
|
|
{
|
13782 |
|
|
/* Set the values. */
|
13783 |
|
|
result_s = fold_build2 (MODIFY_EXPR, type, arg_sinp,
|
13784 |
|
|
result_s);
|
13785 |
|
|
TREE_SIDE_EFFECTS (result_s) = 1;
|
13786 |
|
|
result_c = fold_build2 (MODIFY_EXPR, type, arg_cosp,
|
13787 |
|
|
result_c);
|
13788 |
|
|
TREE_SIDE_EFFECTS (result_c) = 1;
|
13789 |
|
|
/* Combine the assignments into a compound expr. */
|
13790 |
|
|
result = non_lvalue (fold_build2 (COMPOUND_EXPR, type,
|
13791 |
|
|
result_s, result_c));
|
13792 |
|
|
}
|
13793 |
|
|
}
|
13794 |
|
|
}
|
13795 |
|
|
}
|
13796 |
|
|
return result;
|
13797 |
|
|
}
|
13798 |
|
|
|
13799 |
|
|
/* If argument ARG1 is an INTEGER_CST and ARG2 is a REAL_CST, call the
|
13800 |
|
|
two-argument mpfr order N Bessel function FUNC on them and return
|
13801 |
|
|
the resulting value as a tree with type TYPE. The mpfr precision
|
13802 |
|
|
is set to the precision of TYPE. We assume that function FUNC
|
13803 |
|
|
returns zero if the result could be calculated exactly within the
|
13804 |
|
|
requested precision. */
|
13805 |
|
|
static tree
|
13806 |
|
|
do_mpfr_bessel_n (tree arg1, tree arg2, tree type,
|
13807 |
|
|
int (*func)(mpfr_ptr, long, mpfr_srcptr, mp_rnd_t),
|
13808 |
|
|
const REAL_VALUE_TYPE *min, bool inclusive)
|
13809 |
|
|
{
|
13810 |
|
|
tree result = NULL_TREE;
|
13811 |
|
|
|
13812 |
|
|
STRIP_NOPS (arg1);
|
13813 |
|
|
STRIP_NOPS (arg2);
|
13814 |
|
|
|
13815 |
|
|
/* To proceed, MPFR must exactly represent the target floating point
|
13816 |
|
|
format, which only happens when the target base equals two. */
|
13817 |
|
|
if (REAL_MODE_FORMAT (TYPE_MODE (type))->b == 2
|
13818 |
|
|
&& host_integerp (arg1, 0)
|
13819 |
|
|
&& TREE_CODE (arg2) == REAL_CST && !TREE_OVERFLOW (arg2))
|
13820 |
|
|
{
|
13821 |
|
|
const HOST_WIDE_INT n = tree_low_cst(arg1, 0);
|
13822 |
|
|
const REAL_VALUE_TYPE *const ra = &TREE_REAL_CST (arg2);
|
13823 |
|
|
|
13824 |
|
|
if (n == (long)n
|
13825 |
|
|
&& real_isfinite (ra)
|
13826 |
|
|
&& (!min || real_compare (inclusive ? GE_EXPR: GT_EXPR , ra, min)))
|
13827 |
|
|
{
|
13828 |
|
|
const struct real_format *fmt = REAL_MODE_FORMAT (TYPE_MODE (type));
|
13829 |
|
|
const int prec = fmt->p;
|
13830 |
|
|
const mp_rnd_t rnd = fmt->round_towards_zero? GMP_RNDZ : GMP_RNDN;
|
13831 |
|
|
int inexact;
|
13832 |
|
|
mpfr_t m;
|
13833 |
|
|
|
13834 |
|
|
mpfr_init2 (m, prec);
|
13835 |
|
|
mpfr_from_real (m, ra, GMP_RNDN);
|
13836 |
|
|
mpfr_clear_flags ();
|
13837 |
|
|
inexact = func (m, n, m, rnd);
|
13838 |
|
|
result = do_mpfr_ckconv (m, type, inexact);
|
13839 |
|
|
mpfr_clear (m);
|
13840 |
|
|
}
|
13841 |
|
|
}
|
13842 |
|
|
|
13843 |
|
|
return result;
|
13844 |
|
|
}
|
13845 |
|
|
|
13846 |
|
|
/* If arguments ARG0 and ARG1 are REAL_CSTs, call mpfr_remquo() to set
|
13847 |
|
|
the pointer *(ARG_QUO) and return the result. The type is taken
|
13848 |
|
|
from the type of ARG0 and is used for setting the precision of the
|
13849 |
|
|
calculation and results. */
|
13850 |
|
|
|
13851 |
|
|
static tree
|
13852 |
|
|
do_mpfr_remquo (tree arg0, tree arg1, tree arg_quo)
|
13853 |
|
|
{
|
13854 |
|
|
tree const type = TREE_TYPE (arg0);
|
13855 |
|
|
tree result = NULL_TREE;
|
13856 |
|
|
|
13857 |
|
|
STRIP_NOPS (arg0);
|
13858 |
|
|
STRIP_NOPS (arg1);
|
13859 |
|
|
|
13860 |
|
|
/* To proceed, MPFR must exactly represent the target floating point
|
13861 |
|
|
format, which only happens when the target base equals two. */
|
13862 |
|
|
if (REAL_MODE_FORMAT (TYPE_MODE (type))->b == 2
|
13863 |
|
|
&& TREE_CODE (arg0) == REAL_CST && !TREE_OVERFLOW (arg0)
|
13864 |
|
|
&& TREE_CODE (arg1) == REAL_CST && !TREE_OVERFLOW (arg1))
|
13865 |
|
|
{
|
13866 |
|
|
const REAL_VALUE_TYPE *const ra0 = TREE_REAL_CST_PTR (arg0);
|
13867 |
|
|
const REAL_VALUE_TYPE *const ra1 = TREE_REAL_CST_PTR (arg1);
|
13868 |
|
|
|
13869 |
|
|
if (real_isfinite (ra0) && real_isfinite (ra1))
|
13870 |
|
|
{
|
13871 |
|
|
const struct real_format *fmt = REAL_MODE_FORMAT (TYPE_MODE (type));
|
13872 |
|
|
const int prec = fmt->p;
|
13873 |
|
|
const mp_rnd_t rnd = fmt->round_towards_zero? GMP_RNDZ : GMP_RNDN;
|
13874 |
|
|
tree result_rem;
|
13875 |
|
|
long integer_quo;
|
13876 |
|
|
mpfr_t m0, m1;
|
13877 |
|
|
|
13878 |
|
|
mpfr_inits2 (prec, m0, m1, NULL);
|
13879 |
|
|
mpfr_from_real (m0, ra0, GMP_RNDN);
|
13880 |
|
|
mpfr_from_real (m1, ra1, GMP_RNDN);
|
13881 |
|
|
mpfr_clear_flags ();
|
13882 |
|
|
mpfr_remquo (m0, &integer_quo, m0, m1, rnd);
|
13883 |
|
|
/* Remquo is independent of the rounding mode, so pass
|
13884 |
|
|
inexact=0 to do_mpfr_ckconv(). */
|
13885 |
|
|
result_rem = do_mpfr_ckconv (m0, type, /*inexact=*/ 0);
|
13886 |
|
|
mpfr_clears (m0, m1, NULL);
|
13887 |
|
|
if (result_rem)
|
13888 |
|
|
{
|
13889 |
|
|
/* MPFR calculates quo in the host's long so it may
|
13890 |
|
|
return more bits in quo than the target int can hold
|
13891 |
|
|
if sizeof(host long) > sizeof(target int). This can
|
13892 |
|
|
happen even for native compilers in LP64 mode. In
|
13893 |
|
|
these cases, modulo the quo value with the largest
|
13894 |
|
|
number that the target int can hold while leaving one
|
13895 |
|
|
bit for the sign. */
|
13896 |
|
|
if (sizeof (integer_quo) * CHAR_BIT > INT_TYPE_SIZE)
|
13897 |
|
|
integer_quo %= (long)(1UL << (INT_TYPE_SIZE - 1));
|
13898 |
|
|
|
13899 |
|
|
/* Dereference the quo pointer argument. */
|
13900 |
|
|
arg_quo = build_fold_indirect_ref (arg_quo);
|
13901 |
|
|
/* Proceed iff a valid pointer type was passed in. */
|
13902 |
|
|
if (TYPE_MAIN_VARIANT (TREE_TYPE (arg_quo)) == integer_type_node)
|
13903 |
|
|
{
|
13904 |
|
|
/* Set the value. */
|
13905 |
|
|
tree result_quo
|
13906 |
|
|
= fold_build2 (MODIFY_EXPR, TREE_TYPE (arg_quo), arg_quo,
|
13907 |
|
|
build_int_cst (TREE_TYPE (arg_quo),
|
13908 |
|
|
integer_quo));
|
13909 |
|
|
TREE_SIDE_EFFECTS (result_quo) = 1;
|
13910 |
|
|
/* Combine the quo assignment with the rem. */
|
13911 |
|
|
result = non_lvalue (fold_build2 (COMPOUND_EXPR, type,
|
13912 |
|
|
result_quo, result_rem));
|
13913 |
|
|
}
|
13914 |
|
|
}
|
13915 |
|
|
}
|
13916 |
|
|
}
|
13917 |
|
|
return result;
|
13918 |
|
|
}
|
13919 |
|
|
|
13920 |
|
|
/* If ARG is a REAL_CST, call mpfr_lgamma() on it and return the
|
13921 |
|
|
resulting value as a tree with type TYPE. The mpfr precision is
|
13922 |
|
|
set to the precision of TYPE. We assume that this mpfr function
|
13923 |
|
|
returns zero if the result could be calculated exactly within the
|
13924 |
|
|
requested precision. In addition, the integer pointer represented
|
13925 |
|
|
by ARG_SG will be dereferenced and set to the appropriate signgam
|
13926 |
|
|
(-1,1) value. */
|
13927 |
|
|
|
13928 |
|
|
static tree
|
13929 |
|
|
do_mpfr_lgamma_r (tree arg, tree arg_sg, tree type)
|
13930 |
|
|
{
|
13931 |
|
|
tree result = NULL_TREE;
|
13932 |
|
|
|
13933 |
|
|
STRIP_NOPS (arg);
|
13934 |
|
|
|
13935 |
|
|
/* To proceed, MPFR must exactly represent the target floating point
|
13936 |
|
|
format, which only happens when the target base equals two. Also
|
13937 |
|
|
verify ARG is a constant and that ARG_SG is an int pointer. */
|
13938 |
|
|
if (REAL_MODE_FORMAT (TYPE_MODE (type))->b == 2
|
13939 |
|
|
&& TREE_CODE (arg) == REAL_CST && !TREE_OVERFLOW (arg)
|
13940 |
|
|
&& TREE_CODE (TREE_TYPE (arg_sg)) == POINTER_TYPE
|
13941 |
|
|
&& TYPE_MAIN_VARIANT (TREE_TYPE (TREE_TYPE (arg_sg))) == integer_type_node)
|
13942 |
|
|
{
|
13943 |
|
|
const REAL_VALUE_TYPE *const ra = TREE_REAL_CST_PTR (arg);
|
13944 |
|
|
|
13945 |
|
|
/* In addition to NaN and Inf, the argument cannot be zero or a
|
13946 |
|
|
negative integer. */
|
13947 |
|
|
if (real_isfinite (ra)
|
13948 |
|
|
&& ra->cl != rvc_zero
|
13949 |
|
|
&& !(real_isneg(ra) && real_isinteger(ra, TYPE_MODE (type))))
|
13950 |
|
|
{
|
13951 |
|
|
const struct real_format *fmt = REAL_MODE_FORMAT (TYPE_MODE (type));
|
13952 |
|
|
const int prec = fmt->p;
|
13953 |
|
|
const mp_rnd_t rnd = fmt->round_towards_zero? GMP_RNDZ : GMP_RNDN;
|
13954 |
|
|
int inexact, sg;
|
13955 |
|
|
mpfr_t m;
|
13956 |
|
|
tree result_lg;
|
13957 |
|
|
|
13958 |
|
|
mpfr_init2 (m, prec);
|
13959 |
|
|
mpfr_from_real (m, ra, GMP_RNDN);
|
13960 |
|
|
mpfr_clear_flags ();
|
13961 |
|
|
inexact = mpfr_lgamma (m, &sg, m, rnd);
|
13962 |
|
|
result_lg = do_mpfr_ckconv (m, type, inexact);
|
13963 |
|
|
mpfr_clear (m);
|
13964 |
|
|
if (result_lg)
|
13965 |
|
|
{
|
13966 |
|
|
tree result_sg;
|
13967 |
|
|
|
13968 |
|
|
/* Dereference the arg_sg pointer argument. */
|
13969 |
|
|
arg_sg = build_fold_indirect_ref (arg_sg);
|
13970 |
|
|
/* Assign the signgam value into *arg_sg. */
|
13971 |
|
|
result_sg = fold_build2 (MODIFY_EXPR,
|
13972 |
|
|
TREE_TYPE (arg_sg), arg_sg,
|
13973 |
|
|
build_int_cst (TREE_TYPE (arg_sg), sg));
|
13974 |
|
|
TREE_SIDE_EFFECTS (result_sg) = 1;
|
13975 |
|
|
/* Combine the signgam assignment with the lgamma result. */
|
13976 |
|
|
result = non_lvalue (fold_build2 (COMPOUND_EXPR, type,
|
13977 |
|
|
result_sg, result_lg));
|
13978 |
|
|
}
|
13979 |
|
|
}
|
13980 |
|
|
}
|
13981 |
|
|
|
13982 |
|
|
return result;
|
13983 |
|
|
}
|
13984 |
|
|
|
13985 |
|
|
/* If argument ARG is a COMPLEX_CST, call the one-argument mpc
|
13986 |
|
|
function FUNC on it and return the resulting value as a tree with
|
13987 |
|
|
type TYPE. The mpfr precision is set to the precision of TYPE. We
|
13988 |
|
|
assume that function FUNC returns zero if the result could be
|
13989 |
|
|
calculated exactly within the requested precision. */
|
13990 |
|
|
|
13991 |
|
|
static tree
|
13992 |
|
|
do_mpc_arg1 (tree arg, tree type, int (*func)(mpc_ptr, mpc_srcptr, mpc_rnd_t))
|
13993 |
|
|
{
|
13994 |
|
|
tree result = NULL_TREE;
|
13995 |
|
|
|
13996 |
|
|
STRIP_NOPS (arg);
|
13997 |
|
|
|
13998 |
|
|
/* To proceed, MPFR must exactly represent the target floating point
|
13999 |
|
|
format, which only happens when the target base equals two. */
|
14000 |
|
|
if (TREE_CODE (arg) == COMPLEX_CST && !TREE_OVERFLOW (arg)
|
14001 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg))) == REAL_TYPE
|
14002 |
|
|
&& REAL_MODE_FORMAT (TYPE_MODE (TREE_TYPE (TREE_TYPE (arg))))->b == 2)
|
14003 |
|
|
{
|
14004 |
|
|
const REAL_VALUE_TYPE *const re = TREE_REAL_CST_PTR (TREE_REALPART (arg));
|
14005 |
|
|
const REAL_VALUE_TYPE *const im = TREE_REAL_CST_PTR (TREE_IMAGPART (arg));
|
14006 |
|
|
|
14007 |
|
|
if (real_isfinite (re) && real_isfinite (im))
|
14008 |
|
|
{
|
14009 |
|
|
const struct real_format *const fmt =
|
14010 |
|
|
REAL_MODE_FORMAT (TYPE_MODE (TREE_TYPE (type)));
|
14011 |
|
|
const int prec = fmt->p;
|
14012 |
|
|
const mp_rnd_t rnd = fmt->round_towards_zero ? GMP_RNDZ : GMP_RNDN;
|
14013 |
|
|
const mpc_rnd_t crnd = fmt->round_towards_zero ? MPC_RNDZZ : MPC_RNDNN;
|
14014 |
|
|
int inexact;
|
14015 |
|
|
mpc_t m;
|
14016 |
|
|
|
14017 |
|
|
mpc_init2 (m, prec);
|
14018 |
|
|
mpfr_from_real (mpc_realref(m), re, rnd);
|
14019 |
|
|
mpfr_from_real (mpc_imagref(m), im, rnd);
|
14020 |
|
|
mpfr_clear_flags ();
|
14021 |
|
|
inexact = func (m, m, crnd);
|
14022 |
|
|
result = do_mpc_ckconv (m, type, inexact, /*force_convert=*/ 0);
|
14023 |
|
|
mpc_clear (m);
|
14024 |
|
|
}
|
14025 |
|
|
}
|
14026 |
|
|
|
14027 |
|
|
return result;
|
14028 |
|
|
}
|
14029 |
|
|
|
14030 |
|
|
/* If arguments ARG0 and ARG1 are a COMPLEX_CST, call the two-argument
|
14031 |
|
|
mpc function FUNC on it and return the resulting value as a tree
|
14032 |
|
|
with type TYPE. The mpfr precision is set to the precision of
|
14033 |
|
|
TYPE. We assume that function FUNC returns zero if the result
|
14034 |
|
|
could be calculated exactly within the requested precision. If
|
14035 |
|
|
DO_NONFINITE is true, then fold expressions containing Inf or NaN
|
14036 |
|
|
in the arguments and/or results. */
|
14037 |
|
|
|
14038 |
|
|
tree
|
14039 |
|
|
do_mpc_arg2 (tree arg0, tree arg1, tree type, int do_nonfinite,
|
14040 |
|
|
int (*func)(mpc_ptr, mpc_srcptr, mpc_srcptr, mpc_rnd_t))
|
14041 |
|
|
{
|
14042 |
|
|
tree result = NULL_TREE;
|
14043 |
|
|
|
14044 |
|
|
STRIP_NOPS (arg0);
|
14045 |
|
|
STRIP_NOPS (arg1);
|
14046 |
|
|
|
14047 |
|
|
/* To proceed, MPFR must exactly represent the target floating point
|
14048 |
|
|
format, which only happens when the target base equals two. */
|
14049 |
|
|
if (TREE_CODE (arg0) == COMPLEX_CST && !TREE_OVERFLOW (arg0)
|
14050 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg0))) == REAL_TYPE
|
14051 |
|
|
&& TREE_CODE (arg1) == COMPLEX_CST && !TREE_OVERFLOW (arg1)
|
14052 |
|
|
&& TREE_CODE (TREE_TYPE (TREE_TYPE (arg1))) == REAL_TYPE
|
14053 |
|
|
&& REAL_MODE_FORMAT (TYPE_MODE (TREE_TYPE (TREE_TYPE (arg0))))->b == 2)
|
14054 |
|
|
{
|
14055 |
|
|
const REAL_VALUE_TYPE *const re0 = TREE_REAL_CST_PTR (TREE_REALPART (arg0));
|
14056 |
|
|
const REAL_VALUE_TYPE *const im0 = TREE_REAL_CST_PTR (TREE_IMAGPART (arg0));
|
14057 |
|
|
const REAL_VALUE_TYPE *const re1 = TREE_REAL_CST_PTR (TREE_REALPART (arg1));
|
14058 |
|
|
const REAL_VALUE_TYPE *const im1 = TREE_REAL_CST_PTR (TREE_IMAGPART (arg1));
|
14059 |
|
|
|
14060 |
|
|
if (do_nonfinite
|
14061 |
|
|
|| (real_isfinite (re0) && real_isfinite (im0)
|
14062 |
|
|
&& real_isfinite (re1) && real_isfinite (im1)))
|
14063 |
|
|
{
|
14064 |
|
|
const struct real_format *const fmt =
|
14065 |
|
|
REAL_MODE_FORMAT (TYPE_MODE (TREE_TYPE (type)));
|
14066 |
|
|
const int prec = fmt->p;
|
14067 |
|
|
const mp_rnd_t rnd = fmt->round_towards_zero ? GMP_RNDZ : GMP_RNDN;
|
14068 |
|
|
const mpc_rnd_t crnd = fmt->round_towards_zero ? MPC_RNDZZ : MPC_RNDNN;
|
14069 |
|
|
int inexact;
|
14070 |
|
|
mpc_t m0, m1;
|
14071 |
|
|
|
14072 |
|
|
mpc_init2 (m0, prec);
|
14073 |
|
|
mpc_init2 (m1, prec);
|
14074 |
|
|
mpfr_from_real (mpc_realref(m0), re0, rnd);
|
14075 |
|
|
mpfr_from_real (mpc_imagref(m0), im0, rnd);
|
14076 |
|
|
mpfr_from_real (mpc_realref(m1), re1, rnd);
|
14077 |
|
|
mpfr_from_real (mpc_imagref(m1), im1, rnd);
|
14078 |
|
|
mpfr_clear_flags ();
|
14079 |
|
|
inexact = func (m0, m0, m1, crnd);
|
14080 |
|
|
result = do_mpc_ckconv (m0, type, inexact, do_nonfinite);
|
14081 |
|
|
mpc_clear (m0);
|
14082 |
|
|
mpc_clear (m1);
|
14083 |
|
|
}
|
14084 |
|
|
}
|
14085 |
|
|
|
14086 |
|
|
return result;
|
14087 |
|
|
}
|
14088 |
|
|
|
14089 |
|
|
/* Fold a call STMT to __{,v}sprintf_chk. Return NULL_TREE if
|
14090 |
|
|
a normal call should be emitted rather than expanding the function
|
14091 |
|
|
inline. FCODE is either BUILT_IN_SPRINTF_CHK or BUILT_IN_VSPRINTF_CHK. */
|
14092 |
|
|
|
14093 |
|
|
static tree
|
14094 |
|
|
gimple_fold_builtin_sprintf_chk (gimple stmt, enum built_in_function fcode)
|
14095 |
|
|
{
|
14096 |
|
|
int nargs = gimple_call_num_args (stmt);
|
14097 |
|
|
|
14098 |
|
|
return fold_builtin_sprintf_chk_1 (gimple_location (stmt), nargs,
|
14099 |
|
|
(nargs > 0
|
14100 |
|
|
? gimple_call_arg_ptr (stmt, 0)
|
14101 |
|
|
: &error_mark_node), fcode);
|
14102 |
|
|
}
|
14103 |
|
|
|
14104 |
|
|
/* Fold a call STMT to {,v}snprintf. Return NULL_TREE if
|
14105 |
|
|
a normal call should be emitted rather than expanding the function
|
14106 |
|
|
inline. FCODE is either BUILT_IN_SNPRINTF_CHK or
|
14107 |
|
|
BUILT_IN_VSNPRINTF_CHK. If MAXLEN is not NULL, it is maximum length
|
14108 |
|
|
passed as second argument. */
|
14109 |
|
|
|
14110 |
|
|
tree
|
14111 |
|
|
gimple_fold_builtin_snprintf_chk (gimple stmt, tree maxlen,
|
14112 |
|
|
enum built_in_function fcode)
|
14113 |
|
|
{
|
14114 |
|
|
int nargs = gimple_call_num_args (stmt);
|
14115 |
|
|
|
14116 |
|
|
return fold_builtin_snprintf_chk_1 (gimple_location (stmt), nargs,
|
14117 |
|
|
(nargs > 0
|
14118 |
|
|
? gimple_call_arg_ptr (stmt, 0)
|
14119 |
|
|
: &error_mark_node), maxlen, fcode);
|
14120 |
|
|
}
|
14121 |
|
|
|
14122 |
|
|
/* Builtins with folding operations that operate on "..." arguments
|
14123 |
|
|
need special handling; we need to store the arguments in a convenient
|
14124 |
|
|
data structure before attempting any folding. Fortunately there are
|
14125 |
|
|
only a few builtins that fall into this category. FNDECL is the
|
14126 |
|
|
function, EXP is the CALL_EXPR for the call, and IGNORE is true if the
|
14127 |
|
|
result of the function call is ignored. */
|
14128 |
|
|
|
14129 |
|
|
static tree
|
14130 |
|
|
gimple_fold_builtin_varargs (tree fndecl, gimple stmt,
|
14131 |
|
|
bool ignore ATTRIBUTE_UNUSED)
|
14132 |
|
|
{
|
14133 |
|
|
enum built_in_function fcode = DECL_FUNCTION_CODE (fndecl);
|
14134 |
|
|
tree ret = NULL_TREE;
|
14135 |
|
|
|
14136 |
|
|
switch (fcode)
|
14137 |
|
|
{
|
14138 |
|
|
case BUILT_IN_SPRINTF_CHK:
|
14139 |
|
|
case BUILT_IN_VSPRINTF_CHK:
|
14140 |
|
|
ret = gimple_fold_builtin_sprintf_chk (stmt, fcode);
|
14141 |
|
|
break;
|
14142 |
|
|
|
14143 |
|
|
case BUILT_IN_SNPRINTF_CHK:
|
14144 |
|
|
case BUILT_IN_VSNPRINTF_CHK:
|
14145 |
|
|
ret = gimple_fold_builtin_snprintf_chk (stmt, NULL_TREE, fcode);
|
14146 |
|
|
|
14147 |
|
|
default:
|
14148 |
|
|
break;
|
14149 |
|
|
}
|
14150 |
|
|
if (ret)
|
14151 |
|
|
{
|
14152 |
|
|
ret = build1 (NOP_EXPR, TREE_TYPE (ret), ret);
|
14153 |
|
|
TREE_NO_WARNING (ret) = 1;
|
14154 |
|
|
return ret;
|
14155 |
|
|
}
|
14156 |
|
|
return NULL_TREE;
|
14157 |
|
|
}
|
14158 |
|
|
|
14159 |
|
|
/* A wrapper function for builtin folding that prevents warnings for
|
14160 |
|
|
"statement without effect" and the like, caused by removing the
|
14161 |
|
|
call node earlier than the warning is generated. */
|
14162 |
|
|
|
14163 |
|
|
tree
|
14164 |
|
|
fold_call_stmt (gimple stmt, bool ignore)
|
14165 |
|
|
{
|
14166 |
|
|
tree ret = NULL_TREE;
|
14167 |
|
|
tree fndecl = gimple_call_fndecl (stmt);
|
14168 |
|
|
location_t loc = gimple_location (stmt);
|
14169 |
|
|
if (fndecl
|
14170 |
|
|
&& TREE_CODE (fndecl) == FUNCTION_DECL
|
14171 |
|
|
&& DECL_BUILT_IN (fndecl)
|
14172 |
|
|
&& !gimple_call_va_arg_pack_p (stmt))
|
14173 |
|
|
{
|
14174 |
|
|
int nargs = gimple_call_num_args (stmt);
|
14175 |
|
|
tree *args = (nargs > 0
|
14176 |
|
|
? gimple_call_arg_ptr (stmt, 0)
|
14177 |
|
|
: &error_mark_node);
|
14178 |
|
|
|
14179 |
|
|
if (avoid_folding_inline_builtin (fndecl))
|
14180 |
|
|
return NULL_TREE;
|
14181 |
|
|
if (DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_MD)
|
14182 |
|
|
{
|
14183 |
|
|
return targetm.fold_builtin (fndecl, nargs, args, ignore);
|
14184 |
|
|
}
|
14185 |
|
|
else
|
14186 |
|
|
{
|
14187 |
|
|
if (nargs <= MAX_ARGS_TO_FOLD_BUILTIN)
|
14188 |
|
|
ret = fold_builtin_n (loc, fndecl, args, nargs, ignore);
|
14189 |
|
|
if (!ret)
|
14190 |
|
|
ret = gimple_fold_builtin_varargs (fndecl, stmt, ignore);
|
14191 |
|
|
if (ret)
|
14192 |
|
|
{
|
14193 |
|
|
/* Propagate location information from original call to
|
14194 |
|
|
expansion of builtin. Otherwise things like
|
14195 |
|
|
maybe_emit_chk_warning, that operate on the expansion
|
14196 |
|
|
of a builtin, will use the wrong location information. */
|
14197 |
|
|
if (gimple_has_location (stmt))
|
14198 |
|
|
{
|
14199 |
|
|
tree realret = ret;
|
14200 |
|
|
if (TREE_CODE (ret) == NOP_EXPR)
|
14201 |
|
|
realret = TREE_OPERAND (ret, 0);
|
14202 |
|
|
if (CAN_HAVE_LOCATION_P (realret)
|
14203 |
|
|
&& !EXPR_HAS_LOCATION (realret))
|
14204 |
|
|
SET_EXPR_LOCATION (realret, loc);
|
14205 |
|
|
return realret;
|
14206 |
|
|
}
|
14207 |
|
|
return ret;
|
14208 |
|
|
}
|
14209 |
|
|
}
|
14210 |
|
|
}
|
14211 |
|
|
return NULL_TREE;
|
14212 |
|
|
}
|
14213 |
|
|
|
14214 |
|
|
/* Look up the function in builtin_decl that corresponds to DECL
|
14215 |
|
|
and set ASMSPEC as its user assembler name. DECL must be a
|
14216 |
|
|
function decl that declares a builtin. */
|
14217 |
|
|
|
14218 |
|
|
void
|
14219 |
|
|
set_builtin_user_assembler_name (tree decl, const char *asmspec)
|
14220 |
|
|
{
|
14221 |
|
|
tree builtin;
|
14222 |
|
|
gcc_assert (TREE_CODE (decl) == FUNCTION_DECL
|
14223 |
|
|
&& DECL_BUILT_IN_CLASS (decl) == BUILT_IN_NORMAL
|
14224 |
|
|
&& asmspec != 0);
|
14225 |
|
|
|
14226 |
|
|
builtin = builtin_decl_explicit (DECL_FUNCTION_CODE (decl));
|
14227 |
|
|
set_user_assembler_name (builtin, asmspec);
|
14228 |
|
|
switch (DECL_FUNCTION_CODE (decl))
|
14229 |
|
|
{
|
14230 |
|
|
case BUILT_IN_MEMCPY:
|
14231 |
|
|
init_block_move_fn (asmspec);
|
14232 |
|
|
memcpy_libfunc = set_user_assembler_libfunc ("memcpy", asmspec);
|
14233 |
|
|
break;
|
14234 |
|
|
case BUILT_IN_MEMSET:
|
14235 |
|
|
init_block_clear_fn (asmspec);
|
14236 |
|
|
memset_libfunc = set_user_assembler_libfunc ("memset", asmspec);
|
14237 |
|
|
break;
|
14238 |
|
|
case BUILT_IN_MEMMOVE:
|
14239 |
|
|
memmove_libfunc = set_user_assembler_libfunc ("memmove", asmspec);
|
14240 |
|
|
break;
|
14241 |
|
|
case BUILT_IN_MEMCMP:
|
14242 |
|
|
memcmp_libfunc = set_user_assembler_libfunc ("memcmp", asmspec);
|
14243 |
|
|
break;
|
14244 |
|
|
case BUILT_IN_ABORT:
|
14245 |
|
|
abort_libfunc = set_user_assembler_libfunc ("abort", asmspec);
|
14246 |
|
|
break;
|
14247 |
|
|
case BUILT_IN_FFS:
|
14248 |
|
|
if (INT_TYPE_SIZE < BITS_PER_WORD)
|
14249 |
|
|
{
|
14250 |
|
|
set_user_assembler_libfunc ("ffs", asmspec);
|
14251 |
|
|
set_optab_libfunc (ffs_optab, mode_for_size (INT_TYPE_SIZE,
|
14252 |
|
|
MODE_INT, 0), "ffs");
|
14253 |
|
|
}
|
14254 |
|
|
break;
|
14255 |
|
|
default:
|
14256 |
|
|
break;
|
14257 |
|
|
}
|
14258 |
|
|
}
|
14259 |
|
|
|
14260 |
|
|
/* Return true if DECL is a builtin that expands to a constant or similarly
|
14261 |
|
|
simple code. */
|
14262 |
|
|
bool
|
14263 |
|
|
is_simple_builtin (tree decl)
|
14264 |
|
|
{
|
14265 |
|
|
if (decl && DECL_BUILT_IN_CLASS (decl) == BUILT_IN_NORMAL)
|
14266 |
|
|
switch (DECL_FUNCTION_CODE (decl))
|
14267 |
|
|
{
|
14268 |
|
|
/* Builtins that expand to constants. */
|
14269 |
|
|
case BUILT_IN_CONSTANT_P:
|
14270 |
|
|
case BUILT_IN_EXPECT:
|
14271 |
|
|
case BUILT_IN_OBJECT_SIZE:
|
14272 |
|
|
case BUILT_IN_UNREACHABLE:
|
14273 |
|
|
/* Simple register moves or loads from stack. */
|
14274 |
|
|
case BUILT_IN_ASSUME_ALIGNED:
|
14275 |
|
|
case BUILT_IN_RETURN_ADDRESS:
|
14276 |
|
|
case BUILT_IN_EXTRACT_RETURN_ADDR:
|
14277 |
|
|
case BUILT_IN_FROB_RETURN_ADDR:
|
14278 |
|
|
case BUILT_IN_RETURN:
|
14279 |
|
|
case BUILT_IN_AGGREGATE_INCOMING_ADDRESS:
|
14280 |
|
|
case BUILT_IN_FRAME_ADDRESS:
|
14281 |
|
|
case BUILT_IN_VA_END:
|
14282 |
|
|
case BUILT_IN_STACK_SAVE:
|
14283 |
|
|
case BUILT_IN_STACK_RESTORE:
|
14284 |
|
|
/* Exception state returns or moves registers around. */
|
14285 |
|
|
case BUILT_IN_EH_FILTER:
|
14286 |
|
|
case BUILT_IN_EH_POINTER:
|
14287 |
|
|
case BUILT_IN_EH_COPY_VALUES:
|
14288 |
|
|
return true;
|
14289 |
|
|
|
14290 |
|
|
default:
|
14291 |
|
|
return false;
|
14292 |
|
|
}
|
14293 |
|
|
|
14294 |
|
|
return false;
|
14295 |
|
|
}
|
14296 |
|
|
|
14297 |
|
|
/* Return true if DECL is a builtin that is not expensive, i.e., they are
|
14298 |
|
|
most probably expanded inline into reasonably simple code. This is a
|
14299 |
|
|
superset of is_simple_builtin. */
|
14300 |
|
|
bool
|
14301 |
|
|
is_inexpensive_builtin (tree decl)
|
14302 |
|
|
{
|
14303 |
|
|
if (!decl)
|
14304 |
|
|
return false;
|
14305 |
|
|
else if (DECL_BUILT_IN_CLASS (decl) == BUILT_IN_MD)
|
14306 |
|
|
return true;
|
14307 |
|
|
else if (DECL_BUILT_IN_CLASS (decl) == BUILT_IN_NORMAL)
|
14308 |
|
|
switch (DECL_FUNCTION_CODE (decl))
|
14309 |
|
|
{
|
14310 |
|
|
case BUILT_IN_ABS:
|
14311 |
|
|
case BUILT_IN_ALLOCA:
|
14312 |
|
|
case BUILT_IN_ALLOCA_WITH_ALIGN:
|
14313 |
|
|
case BUILT_IN_BSWAP32:
|
14314 |
|
|
case BUILT_IN_BSWAP64:
|
14315 |
|
|
case BUILT_IN_CLZ:
|
14316 |
|
|
case BUILT_IN_CLZIMAX:
|
14317 |
|
|
case BUILT_IN_CLZL:
|
14318 |
|
|
case BUILT_IN_CLZLL:
|
14319 |
|
|
case BUILT_IN_CTZ:
|
14320 |
|
|
case BUILT_IN_CTZIMAX:
|
14321 |
|
|
case BUILT_IN_CTZL:
|
14322 |
|
|
case BUILT_IN_CTZLL:
|
14323 |
|
|
case BUILT_IN_FFS:
|
14324 |
|
|
case BUILT_IN_FFSIMAX:
|
14325 |
|
|
case BUILT_IN_FFSL:
|
14326 |
|
|
case BUILT_IN_FFSLL:
|
14327 |
|
|
case BUILT_IN_IMAXABS:
|
14328 |
|
|
case BUILT_IN_FINITE:
|
14329 |
|
|
case BUILT_IN_FINITEF:
|
14330 |
|
|
case BUILT_IN_FINITEL:
|
14331 |
|
|
case BUILT_IN_FINITED32:
|
14332 |
|
|
case BUILT_IN_FINITED64:
|
14333 |
|
|
case BUILT_IN_FINITED128:
|
14334 |
|
|
case BUILT_IN_FPCLASSIFY:
|
14335 |
|
|
case BUILT_IN_ISFINITE:
|
14336 |
|
|
case BUILT_IN_ISINF_SIGN:
|
14337 |
|
|
case BUILT_IN_ISINF:
|
14338 |
|
|
case BUILT_IN_ISINFF:
|
14339 |
|
|
case BUILT_IN_ISINFL:
|
14340 |
|
|
case BUILT_IN_ISINFD32:
|
14341 |
|
|
case BUILT_IN_ISINFD64:
|
14342 |
|
|
case BUILT_IN_ISINFD128:
|
14343 |
|
|
case BUILT_IN_ISNAN:
|
14344 |
|
|
case BUILT_IN_ISNANF:
|
14345 |
|
|
case BUILT_IN_ISNANL:
|
14346 |
|
|
case BUILT_IN_ISNAND32:
|
14347 |
|
|
case BUILT_IN_ISNAND64:
|
14348 |
|
|
case BUILT_IN_ISNAND128:
|
14349 |
|
|
case BUILT_IN_ISNORMAL:
|
14350 |
|
|
case BUILT_IN_ISGREATER:
|
14351 |
|
|
case BUILT_IN_ISGREATEREQUAL:
|
14352 |
|
|
case BUILT_IN_ISLESS:
|
14353 |
|
|
case BUILT_IN_ISLESSEQUAL:
|
14354 |
|
|
case BUILT_IN_ISLESSGREATER:
|
14355 |
|
|
case BUILT_IN_ISUNORDERED:
|
14356 |
|
|
case BUILT_IN_VA_ARG_PACK:
|
14357 |
|
|
case BUILT_IN_VA_ARG_PACK_LEN:
|
14358 |
|
|
case BUILT_IN_VA_COPY:
|
14359 |
|
|
case BUILT_IN_TRAP:
|
14360 |
|
|
case BUILT_IN_SAVEREGS:
|
14361 |
|
|
case BUILT_IN_POPCOUNTL:
|
14362 |
|
|
case BUILT_IN_POPCOUNTLL:
|
14363 |
|
|
case BUILT_IN_POPCOUNTIMAX:
|
14364 |
|
|
case BUILT_IN_POPCOUNT:
|
14365 |
|
|
case BUILT_IN_PARITYL:
|
14366 |
|
|
case BUILT_IN_PARITYLL:
|
14367 |
|
|
case BUILT_IN_PARITYIMAX:
|
14368 |
|
|
case BUILT_IN_PARITY:
|
14369 |
|
|
case BUILT_IN_LABS:
|
14370 |
|
|
case BUILT_IN_LLABS:
|
14371 |
|
|
case BUILT_IN_PREFETCH:
|
14372 |
|
|
return true;
|
14373 |
|
|
|
14374 |
|
|
default:
|
14375 |
|
|
return is_simple_builtin (decl);
|
14376 |
|
|
}
|
14377 |
|
|
|
14378 |
|
|
return false;
|
14379 |
|
|
}
|