1 |
38 |
julius |
/* Subroutines for manipulating rtx's in semantically interesting ways.
|
2 |
|
|
Copyright (C) 1987, 1991, 1994, 1995, 1996, 1997, 1998,
|
3 |
|
|
1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007
|
4 |
|
|
Free Software Foundation, Inc.
|
5 |
|
|
|
6 |
|
|
This file is part of GCC.
|
7 |
|
|
|
8 |
|
|
GCC is free software; you can redistribute it and/or modify it under
|
9 |
|
|
the terms of the GNU General Public License as published by the Free
|
10 |
|
|
Software Foundation; either version 3, or (at your option) any later
|
11 |
|
|
version.
|
12 |
|
|
|
13 |
|
|
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
|
14 |
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
15 |
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
16 |
|
|
for more details.
|
17 |
|
|
|
18 |
|
|
You should have received a copy of the GNU General Public License
|
19 |
|
|
along with GCC; see the file COPYING3. If not see
|
20 |
|
|
<http://www.gnu.org/licenses/>. */
|
21 |
|
|
|
22 |
|
|
|
23 |
|
|
#include "config.h"
|
24 |
|
|
#include "system.h"
|
25 |
|
|
#include "coretypes.h"
|
26 |
|
|
#include "tm.h"
|
27 |
|
|
#include "toplev.h"
|
28 |
|
|
#include "rtl.h"
|
29 |
|
|
#include "tree.h"
|
30 |
|
|
#include "tm_p.h"
|
31 |
|
|
#include "flags.h"
|
32 |
|
|
#include "function.h"
|
33 |
|
|
#include "expr.h"
|
34 |
|
|
#include "optabs.h"
|
35 |
|
|
#include "hard-reg-set.h"
|
36 |
|
|
#include "insn-config.h"
|
37 |
|
|
#include "ggc.h"
|
38 |
|
|
#include "recog.h"
|
39 |
|
|
#include "langhooks.h"
|
40 |
|
|
#include "target.h"
|
41 |
|
|
#include "output.h"
|
42 |
|
|
|
43 |
|
|
static rtx break_out_memory_refs (rtx);
|
44 |
|
|
static void emit_stack_probe (rtx);
|
45 |
|
|
|
46 |
|
|
|
47 |
|
|
/* Truncate and perhaps sign-extend C as appropriate for MODE. */
|
48 |
|
|
|
49 |
|
|
HOST_WIDE_INT
|
50 |
|
|
trunc_int_for_mode (HOST_WIDE_INT c, enum machine_mode mode)
|
51 |
|
|
{
|
52 |
|
|
int width = GET_MODE_BITSIZE (mode);
|
53 |
|
|
|
54 |
|
|
/* You want to truncate to a _what_? */
|
55 |
|
|
gcc_assert (SCALAR_INT_MODE_P (mode));
|
56 |
|
|
|
57 |
|
|
/* Canonicalize BImode to 0 and STORE_FLAG_VALUE. */
|
58 |
|
|
if (mode == BImode)
|
59 |
|
|
return c & 1 ? STORE_FLAG_VALUE : 0;
|
60 |
|
|
|
61 |
|
|
/* Sign-extend for the requested mode. */
|
62 |
|
|
|
63 |
|
|
if (width < HOST_BITS_PER_WIDE_INT)
|
64 |
|
|
{
|
65 |
|
|
HOST_WIDE_INT sign = 1;
|
66 |
|
|
sign <<= width - 1;
|
67 |
|
|
c &= (sign << 1) - 1;
|
68 |
|
|
c ^= sign;
|
69 |
|
|
c -= sign;
|
70 |
|
|
}
|
71 |
|
|
|
72 |
|
|
return c;
|
73 |
|
|
}
|
74 |
|
|
|
75 |
|
|
/* Return an rtx for the sum of X and the integer C. */
|
76 |
|
|
|
77 |
|
|
rtx
|
78 |
|
|
plus_constant (rtx x, HOST_WIDE_INT c)
|
79 |
|
|
{
|
80 |
|
|
RTX_CODE code;
|
81 |
|
|
rtx y;
|
82 |
|
|
enum machine_mode mode;
|
83 |
|
|
rtx tem;
|
84 |
|
|
int all_constant = 0;
|
85 |
|
|
|
86 |
|
|
if (c == 0)
|
87 |
|
|
return x;
|
88 |
|
|
|
89 |
|
|
restart:
|
90 |
|
|
|
91 |
|
|
code = GET_CODE (x);
|
92 |
|
|
mode = GET_MODE (x);
|
93 |
|
|
y = x;
|
94 |
|
|
|
95 |
|
|
switch (code)
|
96 |
|
|
{
|
97 |
|
|
case CONST_INT:
|
98 |
|
|
return GEN_INT (INTVAL (x) + c);
|
99 |
|
|
|
100 |
|
|
case CONST_DOUBLE:
|
101 |
|
|
{
|
102 |
|
|
unsigned HOST_WIDE_INT l1 = CONST_DOUBLE_LOW (x);
|
103 |
|
|
HOST_WIDE_INT h1 = CONST_DOUBLE_HIGH (x);
|
104 |
|
|
unsigned HOST_WIDE_INT l2 = c;
|
105 |
|
|
HOST_WIDE_INT h2 = c < 0 ? ~0 : 0;
|
106 |
|
|
unsigned HOST_WIDE_INT lv;
|
107 |
|
|
HOST_WIDE_INT hv;
|
108 |
|
|
|
109 |
|
|
add_double (l1, h1, l2, h2, &lv, &hv);
|
110 |
|
|
|
111 |
|
|
return immed_double_const (lv, hv, VOIDmode);
|
112 |
|
|
}
|
113 |
|
|
|
114 |
|
|
case MEM:
|
115 |
|
|
/* If this is a reference to the constant pool, try replacing it with
|
116 |
|
|
a reference to a new constant. If the resulting address isn't
|
117 |
|
|
valid, don't return it because we have no way to validize it. */
|
118 |
|
|
if (GET_CODE (XEXP (x, 0)) == SYMBOL_REF
|
119 |
|
|
&& CONSTANT_POOL_ADDRESS_P (XEXP (x, 0)))
|
120 |
|
|
{
|
121 |
|
|
tem
|
122 |
|
|
= force_const_mem (GET_MODE (x),
|
123 |
|
|
plus_constant (get_pool_constant (XEXP (x, 0)),
|
124 |
|
|
c));
|
125 |
|
|
if (memory_address_p (GET_MODE (tem), XEXP (tem, 0)))
|
126 |
|
|
return tem;
|
127 |
|
|
}
|
128 |
|
|
break;
|
129 |
|
|
|
130 |
|
|
case CONST:
|
131 |
|
|
/* If adding to something entirely constant, set a flag
|
132 |
|
|
so that we can add a CONST around the result. */
|
133 |
|
|
x = XEXP (x, 0);
|
134 |
|
|
all_constant = 1;
|
135 |
|
|
goto restart;
|
136 |
|
|
|
137 |
|
|
case SYMBOL_REF:
|
138 |
|
|
case LABEL_REF:
|
139 |
|
|
all_constant = 1;
|
140 |
|
|
break;
|
141 |
|
|
|
142 |
|
|
case PLUS:
|
143 |
|
|
/* The interesting case is adding the integer to a sum.
|
144 |
|
|
Look for constant term in the sum and combine
|
145 |
|
|
with C. For an integer constant term, we make a combined
|
146 |
|
|
integer. For a constant term that is not an explicit integer,
|
147 |
|
|
we cannot really combine, but group them together anyway.
|
148 |
|
|
|
149 |
|
|
Restart or use a recursive call in case the remaining operand is
|
150 |
|
|
something that we handle specially, such as a SYMBOL_REF.
|
151 |
|
|
|
152 |
|
|
We may not immediately return from the recursive call here, lest
|
153 |
|
|
all_constant gets lost. */
|
154 |
|
|
|
155 |
|
|
if (GET_CODE (XEXP (x, 1)) == CONST_INT)
|
156 |
|
|
{
|
157 |
|
|
c += INTVAL (XEXP (x, 1));
|
158 |
|
|
|
159 |
|
|
if (GET_MODE (x) != VOIDmode)
|
160 |
|
|
c = trunc_int_for_mode (c, GET_MODE (x));
|
161 |
|
|
|
162 |
|
|
x = XEXP (x, 0);
|
163 |
|
|
goto restart;
|
164 |
|
|
}
|
165 |
|
|
else if (CONSTANT_P (XEXP (x, 1)))
|
166 |
|
|
{
|
167 |
|
|
x = gen_rtx_PLUS (mode, XEXP (x, 0), plus_constant (XEXP (x, 1), c));
|
168 |
|
|
c = 0;
|
169 |
|
|
}
|
170 |
|
|
else if (find_constant_term_loc (&y))
|
171 |
|
|
{
|
172 |
|
|
/* We need to be careful since X may be shared and we can't
|
173 |
|
|
modify it in place. */
|
174 |
|
|
rtx copy = copy_rtx (x);
|
175 |
|
|
rtx *const_loc = find_constant_term_loc (©);
|
176 |
|
|
|
177 |
|
|
*const_loc = plus_constant (*const_loc, c);
|
178 |
|
|
x = copy;
|
179 |
|
|
c = 0;
|
180 |
|
|
}
|
181 |
|
|
break;
|
182 |
|
|
|
183 |
|
|
default:
|
184 |
|
|
break;
|
185 |
|
|
}
|
186 |
|
|
|
187 |
|
|
if (c != 0)
|
188 |
|
|
x = gen_rtx_PLUS (mode, x, GEN_INT (c));
|
189 |
|
|
|
190 |
|
|
if (GET_CODE (x) == SYMBOL_REF || GET_CODE (x) == LABEL_REF)
|
191 |
|
|
return x;
|
192 |
|
|
else if (all_constant)
|
193 |
|
|
return gen_rtx_CONST (mode, x);
|
194 |
|
|
else
|
195 |
|
|
return x;
|
196 |
|
|
}
|
197 |
|
|
|
198 |
|
|
/* If X is a sum, return a new sum like X but lacking any constant terms.
|
199 |
|
|
Add all the removed constant terms into *CONSTPTR.
|
200 |
|
|
X itself is not altered. The result != X if and only if
|
201 |
|
|
it is not isomorphic to X. */
|
202 |
|
|
|
203 |
|
|
rtx
|
204 |
|
|
eliminate_constant_term (rtx x, rtx *constptr)
|
205 |
|
|
{
|
206 |
|
|
rtx x0, x1;
|
207 |
|
|
rtx tem;
|
208 |
|
|
|
209 |
|
|
if (GET_CODE (x) != PLUS)
|
210 |
|
|
return x;
|
211 |
|
|
|
212 |
|
|
/* First handle constants appearing at this level explicitly. */
|
213 |
|
|
if (GET_CODE (XEXP (x, 1)) == CONST_INT
|
214 |
|
|
&& 0 != (tem = simplify_binary_operation (PLUS, GET_MODE (x), *constptr,
|
215 |
|
|
XEXP (x, 1)))
|
216 |
|
|
&& GET_CODE (tem) == CONST_INT)
|
217 |
|
|
{
|
218 |
|
|
*constptr = tem;
|
219 |
|
|
return eliminate_constant_term (XEXP (x, 0), constptr);
|
220 |
|
|
}
|
221 |
|
|
|
222 |
|
|
tem = const0_rtx;
|
223 |
|
|
x0 = eliminate_constant_term (XEXP (x, 0), &tem);
|
224 |
|
|
x1 = eliminate_constant_term (XEXP (x, 1), &tem);
|
225 |
|
|
if ((x1 != XEXP (x, 1) || x0 != XEXP (x, 0))
|
226 |
|
|
&& 0 != (tem = simplify_binary_operation (PLUS, GET_MODE (x),
|
227 |
|
|
*constptr, tem))
|
228 |
|
|
&& GET_CODE (tem) == CONST_INT)
|
229 |
|
|
{
|
230 |
|
|
*constptr = tem;
|
231 |
|
|
return gen_rtx_PLUS (GET_MODE (x), x0, x1);
|
232 |
|
|
}
|
233 |
|
|
|
234 |
|
|
return x;
|
235 |
|
|
}
|
236 |
|
|
|
237 |
|
|
/* Return an rtx for the size in bytes of the value of EXP. */
|
238 |
|
|
|
239 |
|
|
rtx
|
240 |
|
|
expr_size (tree exp)
|
241 |
|
|
{
|
242 |
|
|
tree size;
|
243 |
|
|
|
244 |
|
|
if (TREE_CODE (exp) == WITH_SIZE_EXPR)
|
245 |
|
|
size = TREE_OPERAND (exp, 1);
|
246 |
|
|
else
|
247 |
|
|
size = SUBSTITUTE_PLACEHOLDER_IN_EXPR (lang_hooks.expr_size (exp), exp);
|
248 |
|
|
|
249 |
|
|
return expand_expr (size, NULL_RTX, TYPE_MODE (sizetype), 0);
|
250 |
|
|
}
|
251 |
|
|
|
252 |
|
|
/* Return a wide integer for the size in bytes of the value of EXP, or -1
|
253 |
|
|
if the size can vary or is larger than an integer. */
|
254 |
|
|
|
255 |
|
|
HOST_WIDE_INT
|
256 |
|
|
int_expr_size (tree exp)
|
257 |
|
|
{
|
258 |
|
|
tree size;
|
259 |
|
|
|
260 |
|
|
if (TREE_CODE (exp) == WITH_SIZE_EXPR)
|
261 |
|
|
size = TREE_OPERAND (exp, 1);
|
262 |
|
|
else
|
263 |
|
|
size = lang_hooks.expr_size (exp);
|
264 |
|
|
|
265 |
|
|
if (size == 0 || !host_integerp (size, 0))
|
266 |
|
|
return -1;
|
267 |
|
|
|
268 |
|
|
return tree_low_cst (size, 0);
|
269 |
|
|
}
|
270 |
|
|
|
271 |
|
|
/* Return a copy of X in which all memory references
|
272 |
|
|
and all constants that involve symbol refs
|
273 |
|
|
have been replaced with new temporary registers.
|
274 |
|
|
Also emit code to load the memory locations and constants
|
275 |
|
|
into those registers.
|
276 |
|
|
|
277 |
|
|
If X contains no such constants or memory references,
|
278 |
|
|
X itself (not a copy) is returned.
|
279 |
|
|
|
280 |
|
|
If a constant is found in the address that is not a legitimate constant
|
281 |
|
|
in an insn, it is left alone in the hope that it might be valid in the
|
282 |
|
|
address.
|
283 |
|
|
|
284 |
|
|
X may contain no arithmetic except addition, subtraction and multiplication.
|
285 |
|
|
Values returned by expand_expr with 1 for sum_ok fit this constraint. */
|
286 |
|
|
|
287 |
|
|
static rtx
|
288 |
|
|
break_out_memory_refs (rtx x)
|
289 |
|
|
{
|
290 |
|
|
if (MEM_P (x)
|
291 |
|
|
|| (CONSTANT_P (x) && CONSTANT_ADDRESS_P (x)
|
292 |
|
|
&& GET_MODE (x) != VOIDmode))
|
293 |
|
|
x = force_reg (GET_MODE (x), x);
|
294 |
|
|
else if (GET_CODE (x) == PLUS || GET_CODE (x) == MINUS
|
295 |
|
|
|| GET_CODE (x) == MULT)
|
296 |
|
|
{
|
297 |
|
|
rtx op0 = break_out_memory_refs (XEXP (x, 0));
|
298 |
|
|
rtx op1 = break_out_memory_refs (XEXP (x, 1));
|
299 |
|
|
|
300 |
|
|
if (op0 != XEXP (x, 0) || op1 != XEXP (x, 1))
|
301 |
|
|
x = gen_rtx_fmt_ee (GET_CODE (x), Pmode, op0, op1);
|
302 |
|
|
}
|
303 |
|
|
|
304 |
|
|
return x;
|
305 |
|
|
}
|
306 |
|
|
|
307 |
|
|
/* Given X, a memory address in ptr_mode, convert it to an address
|
308 |
|
|
in Pmode, or vice versa (TO_MODE says which way). We take advantage of
|
309 |
|
|
the fact that pointers are not allowed to overflow by commuting arithmetic
|
310 |
|
|
operations over conversions so that address arithmetic insns can be
|
311 |
|
|
used. */
|
312 |
|
|
|
313 |
|
|
rtx
|
314 |
|
|
convert_memory_address (enum machine_mode to_mode ATTRIBUTE_UNUSED,
|
315 |
|
|
rtx x)
|
316 |
|
|
{
|
317 |
|
|
#ifndef POINTERS_EXTEND_UNSIGNED
|
318 |
|
|
gcc_assert (GET_MODE (x) == to_mode || GET_MODE (x) == VOIDmode);
|
319 |
|
|
return x;
|
320 |
|
|
#else /* defined(POINTERS_EXTEND_UNSIGNED) */
|
321 |
|
|
enum machine_mode from_mode;
|
322 |
|
|
rtx temp;
|
323 |
|
|
enum rtx_code code;
|
324 |
|
|
|
325 |
|
|
/* If X already has the right mode, just return it. */
|
326 |
|
|
if (GET_MODE (x) == to_mode)
|
327 |
|
|
return x;
|
328 |
|
|
|
329 |
|
|
from_mode = to_mode == ptr_mode ? Pmode : ptr_mode;
|
330 |
|
|
|
331 |
|
|
/* Here we handle some special cases. If none of them apply, fall through
|
332 |
|
|
to the default case. */
|
333 |
|
|
switch (GET_CODE (x))
|
334 |
|
|
{
|
335 |
|
|
case CONST_INT:
|
336 |
|
|
case CONST_DOUBLE:
|
337 |
|
|
if (GET_MODE_SIZE (to_mode) < GET_MODE_SIZE (from_mode))
|
338 |
|
|
code = TRUNCATE;
|
339 |
|
|
else if (POINTERS_EXTEND_UNSIGNED < 0)
|
340 |
|
|
break;
|
341 |
|
|
else if (POINTERS_EXTEND_UNSIGNED > 0)
|
342 |
|
|
code = ZERO_EXTEND;
|
343 |
|
|
else
|
344 |
|
|
code = SIGN_EXTEND;
|
345 |
|
|
temp = simplify_unary_operation (code, to_mode, x, from_mode);
|
346 |
|
|
if (temp)
|
347 |
|
|
return temp;
|
348 |
|
|
break;
|
349 |
|
|
|
350 |
|
|
case SUBREG:
|
351 |
|
|
if ((SUBREG_PROMOTED_VAR_P (x) || REG_POINTER (SUBREG_REG (x)))
|
352 |
|
|
&& GET_MODE (SUBREG_REG (x)) == to_mode)
|
353 |
|
|
return SUBREG_REG (x);
|
354 |
|
|
break;
|
355 |
|
|
|
356 |
|
|
case LABEL_REF:
|
357 |
|
|
temp = gen_rtx_LABEL_REF (to_mode, XEXP (x, 0));
|
358 |
|
|
LABEL_REF_NONLOCAL_P (temp) = LABEL_REF_NONLOCAL_P (x);
|
359 |
|
|
return temp;
|
360 |
|
|
break;
|
361 |
|
|
|
362 |
|
|
case SYMBOL_REF:
|
363 |
|
|
temp = shallow_copy_rtx (x);
|
364 |
|
|
PUT_MODE (temp, to_mode);
|
365 |
|
|
return temp;
|
366 |
|
|
break;
|
367 |
|
|
|
368 |
|
|
case CONST:
|
369 |
|
|
return gen_rtx_CONST (to_mode,
|
370 |
|
|
convert_memory_address (to_mode, XEXP (x, 0)));
|
371 |
|
|
break;
|
372 |
|
|
|
373 |
|
|
case PLUS:
|
374 |
|
|
case MULT:
|
375 |
|
|
/* For addition we can safely permute the conversion and addition
|
376 |
|
|
operation if one operand is a constant and converting the constant
|
377 |
|
|
does not change it. We can always safely permute them if we are
|
378 |
|
|
making the address narrower. */
|
379 |
|
|
if (GET_MODE_SIZE (to_mode) < GET_MODE_SIZE (from_mode)
|
380 |
|
|
|| (GET_CODE (x) == PLUS
|
381 |
|
|
&& GET_CODE (XEXP (x, 1)) == CONST_INT
|
382 |
|
|
&& XEXP (x, 1) == convert_memory_address (to_mode, XEXP (x, 1))))
|
383 |
|
|
return gen_rtx_fmt_ee (GET_CODE (x), to_mode,
|
384 |
|
|
convert_memory_address (to_mode, XEXP (x, 0)),
|
385 |
|
|
XEXP (x, 1));
|
386 |
|
|
break;
|
387 |
|
|
|
388 |
|
|
default:
|
389 |
|
|
break;
|
390 |
|
|
}
|
391 |
|
|
|
392 |
|
|
return convert_modes (to_mode, from_mode,
|
393 |
|
|
x, POINTERS_EXTEND_UNSIGNED);
|
394 |
|
|
#endif /* defined(POINTERS_EXTEND_UNSIGNED) */
|
395 |
|
|
}
|
396 |
|
|
|
397 |
|
|
/* Return something equivalent to X but valid as a memory address
|
398 |
|
|
for something of mode MODE. When X is not itself valid, this
|
399 |
|
|
works by copying X or subexpressions of it into registers. */
|
400 |
|
|
|
401 |
|
|
rtx
|
402 |
|
|
memory_address (enum machine_mode mode, rtx x)
|
403 |
|
|
{
|
404 |
|
|
rtx oldx = x;
|
405 |
|
|
|
406 |
|
|
x = convert_memory_address (Pmode, x);
|
407 |
|
|
|
408 |
|
|
/* By passing constant addresses through registers
|
409 |
|
|
we get a chance to cse them. */
|
410 |
|
|
if (! cse_not_expected && CONSTANT_P (x) && CONSTANT_ADDRESS_P (x))
|
411 |
|
|
x = force_reg (Pmode, x);
|
412 |
|
|
|
413 |
|
|
/* We get better cse by rejecting indirect addressing at this stage.
|
414 |
|
|
Let the combiner create indirect addresses where appropriate.
|
415 |
|
|
For now, generate the code so that the subexpressions useful to share
|
416 |
|
|
are visible. But not if cse won't be done! */
|
417 |
|
|
else
|
418 |
|
|
{
|
419 |
|
|
if (! cse_not_expected && !REG_P (x))
|
420 |
|
|
x = break_out_memory_refs (x);
|
421 |
|
|
|
422 |
|
|
/* At this point, any valid address is accepted. */
|
423 |
|
|
if (memory_address_p (mode, x))
|
424 |
|
|
goto win;
|
425 |
|
|
|
426 |
|
|
/* If it was valid before but breaking out memory refs invalidated it,
|
427 |
|
|
use it the old way. */
|
428 |
|
|
if (memory_address_p (mode, oldx))
|
429 |
|
|
goto win2;
|
430 |
|
|
|
431 |
|
|
/* Perform machine-dependent transformations on X
|
432 |
|
|
in certain cases. This is not necessary since the code
|
433 |
|
|
below can handle all possible cases, but machine-dependent
|
434 |
|
|
transformations can make better code. */
|
435 |
|
|
LEGITIMIZE_ADDRESS (x, oldx, mode, win);
|
436 |
|
|
|
437 |
|
|
/* PLUS and MULT can appear in special ways
|
438 |
|
|
as the result of attempts to make an address usable for indexing.
|
439 |
|
|
Usually they are dealt with by calling force_operand, below.
|
440 |
|
|
But a sum containing constant terms is special
|
441 |
|
|
if removing them makes the sum a valid address:
|
442 |
|
|
then we generate that address in a register
|
443 |
|
|
and index off of it. We do this because it often makes
|
444 |
|
|
shorter code, and because the addresses thus generated
|
445 |
|
|
in registers often become common subexpressions. */
|
446 |
|
|
if (GET_CODE (x) == PLUS)
|
447 |
|
|
{
|
448 |
|
|
rtx constant_term = const0_rtx;
|
449 |
|
|
rtx y = eliminate_constant_term (x, &constant_term);
|
450 |
|
|
if (constant_term == const0_rtx
|
451 |
|
|
|| ! memory_address_p (mode, y))
|
452 |
|
|
x = force_operand (x, NULL_RTX);
|
453 |
|
|
else
|
454 |
|
|
{
|
455 |
|
|
y = gen_rtx_PLUS (GET_MODE (x), copy_to_reg (y), constant_term);
|
456 |
|
|
if (! memory_address_p (mode, y))
|
457 |
|
|
x = force_operand (x, NULL_RTX);
|
458 |
|
|
else
|
459 |
|
|
x = y;
|
460 |
|
|
}
|
461 |
|
|
}
|
462 |
|
|
|
463 |
|
|
else if (GET_CODE (x) == MULT || GET_CODE (x) == MINUS)
|
464 |
|
|
x = force_operand (x, NULL_RTX);
|
465 |
|
|
|
466 |
|
|
/* If we have a register that's an invalid address,
|
467 |
|
|
it must be a hard reg of the wrong class. Copy it to a pseudo. */
|
468 |
|
|
else if (REG_P (x))
|
469 |
|
|
x = copy_to_reg (x);
|
470 |
|
|
|
471 |
|
|
/* Last resort: copy the value to a register, since
|
472 |
|
|
the register is a valid address. */
|
473 |
|
|
else
|
474 |
|
|
x = force_reg (Pmode, x);
|
475 |
|
|
|
476 |
|
|
goto done;
|
477 |
|
|
|
478 |
|
|
win2:
|
479 |
|
|
x = oldx;
|
480 |
|
|
win:
|
481 |
|
|
if (flag_force_addr && ! cse_not_expected && !REG_P (x))
|
482 |
|
|
{
|
483 |
|
|
x = force_operand (x, NULL_RTX);
|
484 |
|
|
x = force_reg (Pmode, x);
|
485 |
|
|
}
|
486 |
|
|
}
|
487 |
|
|
|
488 |
|
|
done:
|
489 |
|
|
|
490 |
|
|
/* If we didn't change the address, we are done. Otherwise, mark
|
491 |
|
|
a reg as a pointer if we have REG or REG + CONST_INT. */
|
492 |
|
|
if (oldx == x)
|
493 |
|
|
return x;
|
494 |
|
|
else if (REG_P (x))
|
495 |
|
|
mark_reg_pointer (x, BITS_PER_UNIT);
|
496 |
|
|
else if (GET_CODE (x) == PLUS
|
497 |
|
|
&& REG_P (XEXP (x, 0))
|
498 |
|
|
&& GET_CODE (XEXP (x, 1)) == CONST_INT)
|
499 |
|
|
mark_reg_pointer (XEXP (x, 0), BITS_PER_UNIT);
|
500 |
|
|
|
501 |
|
|
/* OLDX may have been the address on a temporary. Update the address
|
502 |
|
|
to indicate that X is now used. */
|
503 |
|
|
update_temp_slot_address (oldx, x);
|
504 |
|
|
|
505 |
|
|
return x;
|
506 |
|
|
}
|
507 |
|
|
|
508 |
|
|
/* Like `memory_address' but pretend `flag_force_addr' is 0. */
|
509 |
|
|
|
510 |
|
|
rtx
|
511 |
|
|
memory_address_noforce (enum machine_mode mode, rtx x)
|
512 |
|
|
{
|
513 |
|
|
int ambient_force_addr = flag_force_addr;
|
514 |
|
|
rtx val;
|
515 |
|
|
|
516 |
|
|
flag_force_addr = 0;
|
517 |
|
|
val = memory_address (mode, x);
|
518 |
|
|
flag_force_addr = ambient_force_addr;
|
519 |
|
|
return val;
|
520 |
|
|
}
|
521 |
|
|
|
522 |
|
|
/* Convert a mem ref into one with a valid memory address.
|
523 |
|
|
Pass through anything else unchanged. */
|
524 |
|
|
|
525 |
|
|
rtx
|
526 |
|
|
validize_mem (rtx ref)
|
527 |
|
|
{
|
528 |
|
|
if (!MEM_P (ref))
|
529 |
|
|
return ref;
|
530 |
|
|
ref = use_anchored_address (ref);
|
531 |
|
|
if (! (flag_force_addr && CONSTANT_ADDRESS_P (XEXP (ref, 0)))
|
532 |
|
|
&& memory_address_p (GET_MODE (ref), XEXP (ref, 0)))
|
533 |
|
|
return ref;
|
534 |
|
|
|
535 |
|
|
/* Don't alter REF itself, since that is probably a stack slot. */
|
536 |
|
|
return replace_equiv_address (ref, XEXP (ref, 0));
|
537 |
|
|
}
|
538 |
|
|
|
539 |
|
|
/* If X is a memory reference to a member of an object block, try rewriting
|
540 |
|
|
it to use an anchor instead. Return the new memory reference on success
|
541 |
|
|
and the old one on failure. */
|
542 |
|
|
|
543 |
|
|
rtx
|
544 |
|
|
use_anchored_address (rtx x)
|
545 |
|
|
{
|
546 |
|
|
rtx base;
|
547 |
|
|
HOST_WIDE_INT offset;
|
548 |
|
|
|
549 |
|
|
if (!flag_section_anchors)
|
550 |
|
|
return x;
|
551 |
|
|
|
552 |
|
|
if (!MEM_P (x))
|
553 |
|
|
return x;
|
554 |
|
|
|
555 |
|
|
/* Split the address into a base and offset. */
|
556 |
|
|
base = XEXP (x, 0);
|
557 |
|
|
offset = 0;
|
558 |
|
|
if (GET_CODE (base) == CONST
|
559 |
|
|
&& GET_CODE (XEXP (base, 0)) == PLUS
|
560 |
|
|
&& GET_CODE (XEXP (XEXP (base, 0), 1)) == CONST_INT)
|
561 |
|
|
{
|
562 |
|
|
offset += INTVAL (XEXP (XEXP (base, 0), 1));
|
563 |
|
|
base = XEXP (XEXP (base, 0), 0);
|
564 |
|
|
}
|
565 |
|
|
|
566 |
|
|
/* Check whether BASE is suitable for anchors. */
|
567 |
|
|
if (GET_CODE (base) != SYMBOL_REF
|
568 |
|
|
|| !SYMBOL_REF_HAS_BLOCK_INFO_P (base)
|
569 |
|
|
|| SYMBOL_REF_ANCHOR_P (base)
|
570 |
|
|
|| SYMBOL_REF_BLOCK (base) == NULL
|
571 |
|
|
|| !targetm.use_anchors_for_symbol_p (base))
|
572 |
|
|
return x;
|
573 |
|
|
|
574 |
|
|
/* Decide where BASE is going to be. */
|
575 |
|
|
place_block_symbol (base);
|
576 |
|
|
|
577 |
|
|
/* Get the anchor we need to use. */
|
578 |
|
|
offset += SYMBOL_REF_BLOCK_OFFSET (base);
|
579 |
|
|
base = get_section_anchor (SYMBOL_REF_BLOCK (base), offset,
|
580 |
|
|
SYMBOL_REF_TLS_MODEL (base));
|
581 |
|
|
|
582 |
|
|
/* Work out the offset from the anchor. */
|
583 |
|
|
offset -= SYMBOL_REF_BLOCK_OFFSET (base);
|
584 |
|
|
|
585 |
|
|
/* If we're going to run a CSE pass, force the anchor into a register.
|
586 |
|
|
We will then be able to reuse registers for several accesses, if the
|
587 |
|
|
target costs say that that's worthwhile. */
|
588 |
|
|
if (!cse_not_expected)
|
589 |
|
|
base = force_reg (GET_MODE (base), base);
|
590 |
|
|
|
591 |
|
|
return replace_equiv_address (x, plus_constant (base, offset));
|
592 |
|
|
}
|
593 |
|
|
|
594 |
|
|
/* Copy the value or contents of X to a new temp reg and return that reg. */
|
595 |
|
|
|
596 |
|
|
rtx
|
597 |
|
|
copy_to_reg (rtx x)
|
598 |
|
|
{
|
599 |
|
|
rtx temp = gen_reg_rtx (GET_MODE (x));
|
600 |
|
|
|
601 |
|
|
/* If not an operand, must be an address with PLUS and MULT so
|
602 |
|
|
do the computation. */
|
603 |
|
|
if (! general_operand (x, VOIDmode))
|
604 |
|
|
x = force_operand (x, temp);
|
605 |
|
|
|
606 |
|
|
if (x != temp)
|
607 |
|
|
emit_move_insn (temp, x);
|
608 |
|
|
|
609 |
|
|
return temp;
|
610 |
|
|
}
|
611 |
|
|
|
612 |
|
|
/* Like copy_to_reg but always give the new register mode Pmode
|
613 |
|
|
in case X is a constant. */
|
614 |
|
|
|
615 |
|
|
rtx
|
616 |
|
|
copy_addr_to_reg (rtx x)
|
617 |
|
|
{
|
618 |
|
|
return copy_to_mode_reg (Pmode, x);
|
619 |
|
|
}
|
620 |
|
|
|
621 |
|
|
/* Like copy_to_reg but always give the new register mode MODE
|
622 |
|
|
in case X is a constant. */
|
623 |
|
|
|
624 |
|
|
rtx
|
625 |
|
|
copy_to_mode_reg (enum machine_mode mode, rtx x)
|
626 |
|
|
{
|
627 |
|
|
rtx temp = gen_reg_rtx (mode);
|
628 |
|
|
|
629 |
|
|
/* If not an operand, must be an address with PLUS and MULT so
|
630 |
|
|
do the computation. */
|
631 |
|
|
if (! general_operand (x, VOIDmode))
|
632 |
|
|
x = force_operand (x, temp);
|
633 |
|
|
|
634 |
|
|
gcc_assert (GET_MODE (x) == mode || GET_MODE (x) == VOIDmode);
|
635 |
|
|
if (x != temp)
|
636 |
|
|
emit_move_insn (temp, x);
|
637 |
|
|
return temp;
|
638 |
|
|
}
|
639 |
|
|
|
640 |
|
|
/* Load X into a register if it is not already one.
|
641 |
|
|
Use mode MODE for the register.
|
642 |
|
|
X should be valid for mode MODE, but it may be a constant which
|
643 |
|
|
is valid for all integer modes; that's why caller must specify MODE.
|
644 |
|
|
|
645 |
|
|
The caller must not alter the value in the register we return,
|
646 |
|
|
since we mark it as a "constant" register. */
|
647 |
|
|
|
648 |
|
|
rtx
|
649 |
|
|
force_reg (enum machine_mode mode, rtx x)
|
650 |
|
|
{
|
651 |
|
|
rtx temp, insn, set;
|
652 |
|
|
|
653 |
|
|
if (REG_P (x))
|
654 |
|
|
return x;
|
655 |
|
|
|
656 |
|
|
if (general_operand (x, mode))
|
657 |
|
|
{
|
658 |
|
|
temp = gen_reg_rtx (mode);
|
659 |
|
|
insn = emit_move_insn (temp, x);
|
660 |
|
|
}
|
661 |
|
|
else
|
662 |
|
|
{
|
663 |
|
|
temp = force_operand (x, NULL_RTX);
|
664 |
|
|
if (REG_P (temp))
|
665 |
|
|
insn = get_last_insn ();
|
666 |
|
|
else
|
667 |
|
|
{
|
668 |
|
|
rtx temp2 = gen_reg_rtx (mode);
|
669 |
|
|
insn = emit_move_insn (temp2, temp);
|
670 |
|
|
temp = temp2;
|
671 |
|
|
}
|
672 |
|
|
}
|
673 |
|
|
|
674 |
|
|
/* Let optimizers know that TEMP's value never changes
|
675 |
|
|
and that X can be substituted for it. Don't get confused
|
676 |
|
|
if INSN set something else (such as a SUBREG of TEMP). */
|
677 |
|
|
if (CONSTANT_P (x)
|
678 |
|
|
&& (set = single_set (insn)) != 0
|
679 |
|
|
&& SET_DEST (set) == temp
|
680 |
|
|
&& ! rtx_equal_p (x, SET_SRC (set)))
|
681 |
|
|
set_unique_reg_note (insn, REG_EQUAL, x);
|
682 |
|
|
|
683 |
|
|
/* Let optimizers know that TEMP is a pointer, and if so, the
|
684 |
|
|
known alignment of that pointer. */
|
685 |
|
|
{
|
686 |
|
|
unsigned align = 0;
|
687 |
|
|
if (GET_CODE (x) == SYMBOL_REF)
|
688 |
|
|
{
|
689 |
|
|
align = BITS_PER_UNIT;
|
690 |
|
|
if (SYMBOL_REF_DECL (x) && DECL_P (SYMBOL_REF_DECL (x)))
|
691 |
|
|
align = DECL_ALIGN (SYMBOL_REF_DECL (x));
|
692 |
|
|
}
|
693 |
|
|
else if (GET_CODE (x) == LABEL_REF)
|
694 |
|
|
align = BITS_PER_UNIT;
|
695 |
|
|
else if (GET_CODE (x) == CONST
|
696 |
|
|
&& GET_CODE (XEXP (x, 0)) == PLUS
|
697 |
|
|
&& GET_CODE (XEXP (XEXP (x, 0), 0)) == SYMBOL_REF
|
698 |
|
|
&& GET_CODE (XEXP (XEXP (x, 0), 1)) == CONST_INT)
|
699 |
|
|
{
|
700 |
|
|
rtx s = XEXP (XEXP (x, 0), 0);
|
701 |
|
|
rtx c = XEXP (XEXP (x, 0), 1);
|
702 |
|
|
unsigned sa, ca;
|
703 |
|
|
|
704 |
|
|
sa = BITS_PER_UNIT;
|
705 |
|
|
if (SYMBOL_REF_DECL (s) && DECL_P (SYMBOL_REF_DECL (s)))
|
706 |
|
|
sa = DECL_ALIGN (SYMBOL_REF_DECL (s));
|
707 |
|
|
|
708 |
|
|
ca = exact_log2 (INTVAL (c) & -INTVAL (c)) * BITS_PER_UNIT;
|
709 |
|
|
|
710 |
|
|
align = MIN (sa, ca);
|
711 |
|
|
}
|
712 |
|
|
else if (MEM_P (x) && MEM_POINTER (x))
|
713 |
|
|
align = MEM_ALIGN (x);
|
714 |
|
|
|
715 |
|
|
if (align)
|
716 |
|
|
mark_reg_pointer (temp, align);
|
717 |
|
|
}
|
718 |
|
|
|
719 |
|
|
return temp;
|
720 |
|
|
}
|
721 |
|
|
|
722 |
|
|
/* If X is a memory ref, copy its contents to a new temp reg and return
|
723 |
|
|
that reg. Otherwise, return X. */
|
724 |
|
|
|
725 |
|
|
rtx
|
726 |
|
|
force_not_mem (rtx x)
|
727 |
|
|
{
|
728 |
|
|
rtx temp;
|
729 |
|
|
|
730 |
|
|
if (!MEM_P (x) || GET_MODE (x) == BLKmode)
|
731 |
|
|
return x;
|
732 |
|
|
|
733 |
|
|
temp = gen_reg_rtx (GET_MODE (x));
|
734 |
|
|
|
735 |
|
|
if (MEM_POINTER (x))
|
736 |
|
|
REG_POINTER (temp) = 1;
|
737 |
|
|
|
738 |
|
|
emit_move_insn (temp, x);
|
739 |
|
|
return temp;
|
740 |
|
|
}
|
741 |
|
|
|
742 |
|
|
/* Copy X to TARGET (if it's nonzero and a reg)
|
743 |
|
|
or to a new temp reg and return that reg.
|
744 |
|
|
MODE is the mode to use for X in case it is a constant. */
|
745 |
|
|
|
746 |
|
|
rtx
|
747 |
|
|
copy_to_suggested_reg (rtx x, rtx target, enum machine_mode mode)
|
748 |
|
|
{
|
749 |
|
|
rtx temp;
|
750 |
|
|
|
751 |
|
|
if (target && REG_P (target))
|
752 |
|
|
temp = target;
|
753 |
|
|
else
|
754 |
|
|
temp = gen_reg_rtx (mode);
|
755 |
|
|
|
756 |
|
|
emit_move_insn (temp, x);
|
757 |
|
|
return temp;
|
758 |
|
|
}
|
759 |
|
|
|
760 |
|
|
/* Return the mode to use to store a scalar of TYPE and MODE.
|
761 |
|
|
PUNSIGNEDP points to the signedness of the type and may be adjusted
|
762 |
|
|
to show what signedness to use on extension operations.
|
763 |
|
|
|
764 |
|
|
FOR_CALL is nonzero if this call is promoting args for a call. */
|
765 |
|
|
|
766 |
|
|
#if defined(PROMOTE_MODE) && !defined(PROMOTE_FUNCTION_MODE)
|
767 |
|
|
#define PROMOTE_FUNCTION_MODE PROMOTE_MODE
|
768 |
|
|
#endif
|
769 |
|
|
|
770 |
|
|
enum machine_mode
|
771 |
|
|
promote_mode (tree type, enum machine_mode mode, int *punsignedp,
|
772 |
|
|
int for_call ATTRIBUTE_UNUSED)
|
773 |
|
|
{
|
774 |
|
|
enum tree_code code = TREE_CODE (type);
|
775 |
|
|
int unsignedp = *punsignedp;
|
776 |
|
|
|
777 |
|
|
#ifndef PROMOTE_MODE
|
778 |
|
|
if (! for_call)
|
779 |
|
|
return mode;
|
780 |
|
|
#endif
|
781 |
|
|
|
782 |
|
|
switch (code)
|
783 |
|
|
{
|
784 |
|
|
#ifdef PROMOTE_FUNCTION_MODE
|
785 |
|
|
case INTEGER_TYPE: case ENUMERAL_TYPE: case BOOLEAN_TYPE:
|
786 |
|
|
case REAL_TYPE: case OFFSET_TYPE:
|
787 |
|
|
#ifdef PROMOTE_MODE
|
788 |
|
|
if (for_call)
|
789 |
|
|
{
|
790 |
|
|
#endif
|
791 |
|
|
PROMOTE_FUNCTION_MODE (mode, unsignedp, type);
|
792 |
|
|
#ifdef PROMOTE_MODE
|
793 |
|
|
}
|
794 |
|
|
else
|
795 |
|
|
{
|
796 |
|
|
PROMOTE_MODE (mode, unsignedp, type);
|
797 |
|
|
}
|
798 |
|
|
#endif
|
799 |
|
|
break;
|
800 |
|
|
#endif
|
801 |
|
|
|
802 |
|
|
#ifdef POINTERS_EXTEND_UNSIGNED
|
803 |
|
|
case REFERENCE_TYPE:
|
804 |
|
|
case POINTER_TYPE:
|
805 |
|
|
mode = Pmode;
|
806 |
|
|
unsignedp = POINTERS_EXTEND_UNSIGNED;
|
807 |
|
|
break;
|
808 |
|
|
#endif
|
809 |
|
|
|
810 |
|
|
default:
|
811 |
|
|
break;
|
812 |
|
|
}
|
813 |
|
|
|
814 |
|
|
*punsignedp = unsignedp;
|
815 |
|
|
return mode;
|
816 |
|
|
}
|
817 |
|
|
|
818 |
|
|
/* Adjust the stack pointer by ADJUST (an rtx for a number of bytes).
|
819 |
|
|
This pops when ADJUST is positive. ADJUST need not be constant. */
|
820 |
|
|
|
821 |
|
|
void
|
822 |
|
|
adjust_stack (rtx adjust)
|
823 |
|
|
{
|
824 |
|
|
rtx temp;
|
825 |
|
|
|
826 |
|
|
if (adjust == const0_rtx)
|
827 |
|
|
return;
|
828 |
|
|
|
829 |
|
|
/* We expect all variable sized adjustments to be multiple of
|
830 |
|
|
PREFERRED_STACK_BOUNDARY. */
|
831 |
|
|
if (GET_CODE (adjust) == CONST_INT)
|
832 |
|
|
stack_pointer_delta -= INTVAL (adjust);
|
833 |
|
|
|
834 |
|
|
temp = expand_binop (Pmode,
|
835 |
|
|
#ifdef STACK_GROWS_DOWNWARD
|
836 |
|
|
add_optab,
|
837 |
|
|
#else
|
838 |
|
|
sub_optab,
|
839 |
|
|
#endif
|
840 |
|
|
stack_pointer_rtx, adjust, stack_pointer_rtx, 0,
|
841 |
|
|
OPTAB_LIB_WIDEN);
|
842 |
|
|
|
843 |
|
|
if (temp != stack_pointer_rtx)
|
844 |
|
|
emit_move_insn (stack_pointer_rtx, temp);
|
845 |
|
|
}
|
846 |
|
|
|
847 |
|
|
/* Adjust the stack pointer by minus ADJUST (an rtx for a number of bytes).
|
848 |
|
|
This pushes when ADJUST is positive. ADJUST need not be constant. */
|
849 |
|
|
|
850 |
|
|
void
|
851 |
|
|
anti_adjust_stack (rtx adjust)
|
852 |
|
|
{
|
853 |
|
|
rtx temp;
|
854 |
|
|
|
855 |
|
|
if (adjust == const0_rtx)
|
856 |
|
|
return;
|
857 |
|
|
|
858 |
|
|
/* We expect all variable sized adjustments to be multiple of
|
859 |
|
|
PREFERRED_STACK_BOUNDARY. */
|
860 |
|
|
if (GET_CODE (adjust) == CONST_INT)
|
861 |
|
|
stack_pointer_delta += INTVAL (adjust);
|
862 |
|
|
|
863 |
|
|
temp = expand_binop (Pmode,
|
864 |
|
|
#ifdef STACK_GROWS_DOWNWARD
|
865 |
|
|
sub_optab,
|
866 |
|
|
#else
|
867 |
|
|
add_optab,
|
868 |
|
|
#endif
|
869 |
|
|
stack_pointer_rtx, adjust, stack_pointer_rtx, 0,
|
870 |
|
|
OPTAB_LIB_WIDEN);
|
871 |
|
|
|
872 |
|
|
if (temp != stack_pointer_rtx)
|
873 |
|
|
emit_move_insn (stack_pointer_rtx, temp);
|
874 |
|
|
}
|
875 |
|
|
|
876 |
|
|
/* Round the size of a block to be pushed up to the boundary required
|
877 |
|
|
by this machine. SIZE is the desired size, which need not be constant. */
|
878 |
|
|
|
879 |
|
|
static rtx
|
880 |
|
|
round_push (rtx size)
|
881 |
|
|
{
|
882 |
|
|
int align = PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT;
|
883 |
|
|
|
884 |
|
|
if (align == 1)
|
885 |
|
|
return size;
|
886 |
|
|
|
887 |
|
|
if (GET_CODE (size) == CONST_INT)
|
888 |
|
|
{
|
889 |
|
|
HOST_WIDE_INT new = (INTVAL (size) + align - 1) / align * align;
|
890 |
|
|
|
891 |
|
|
if (INTVAL (size) != new)
|
892 |
|
|
size = GEN_INT (new);
|
893 |
|
|
}
|
894 |
|
|
else
|
895 |
|
|
{
|
896 |
|
|
/* CEIL_DIV_EXPR needs to worry about the addition overflowing,
|
897 |
|
|
but we know it can't. So add ourselves and then do
|
898 |
|
|
TRUNC_DIV_EXPR. */
|
899 |
|
|
size = expand_binop (Pmode, add_optab, size, GEN_INT (align - 1),
|
900 |
|
|
NULL_RTX, 1, OPTAB_LIB_WIDEN);
|
901 |
|
|
size = expand_divmod (0, TRUNC_DIV_EXPR, Pmode, size, GEN_INT (align),
|
902 |
|
|
NULL_RTX, 1);
|
903 |
|
|
size = expand_mult (Pmode, size, GEN_INT (align), NULL_RTX, 1);
|
904 |
|
|
}
|
905 |
|
|
|
906 |
|
|
return size;
|
907 |
|
|
}
|
908 |
|
|
|
909 |
|
|
/* Save the stack pointer for the purpose in SAVE_LEVEL. PSAVE is a pointer
|
910 |
|
|
to a previously-created save area. If no save area has been allocated,
|
911 |
|
|
this function will allocate one. If a save area is specified, it
|
912 |
|
|
must be of the proper mode.
|
913 |
|
|
|
914 |
|
|
The insns are emitted after insn AFTER, if nonzero, otherwise the insns
|
915 |
|
|
are emitted at the current position. */
|
916 |
|
|
|
917 |
|
|
void
|
918 |
|
|
emit_stack_save (enum save_level save_level, rtx *psave, rtx after)
|
919 |
|
|
{
|
920 |
|
|
rtx sa = *psave;
|
921 |
|
|
/* The default is that we use a move insn and save in a Pmode object. */
|
922 |
|
|
rtx (*fcn) (rtx, rtx) = gen_move_insn;
|
923 |
|
|
enum machine_mode mode = STACK_SAVEAREA_MODE (save_level);
|
924 |
|
|
|
925 |
|
|
/* See if this machine has anything special to do for this kind of save. */
|
926 |
|
|
switch (save_level)
|
927 |
|
|
{
|
928 |
|
|
#ifdef HAVE_save_stack_block
|
929 |
|
|
case SAVE_BLOCK:
|
930 |
|
|
if (HAVE_save_stack_block)
|
931 |
|
|
fcn = gen_save_stack_block;
|
932 |
|
|
break;
|
933 |
|
|
#endif
|
934 |
|
|
#ifdef HAVE_save_stack_function
|
935 |
|
|
case SAVE_FUNCTION:
|
936 |
|
|
if (HAVE_save_stack_function)
|
937 |
|
|
fcn = gen_save_stack_function;
|
938 |
|
|
break;
|
939 |
|
|
#endif
|
940 |
|
|
#ifdef HAVE_save_stack_nonlocal
|
941 |
|
|
case SAVE_NONLOCAL:
|
942 |
|
|
if (HAVE_save_stack_nonlocal)
|
943 |
|
|
fcn = gen_save_stack_nonlocal;
|
944 |
|
|
break;
|
945 |
|
|
#endif
|
946 |
|
|
default:
|
947 |
|
|
break;
|
948 |
|
|
}
|
949 |
|
|
|
950 |
|
|
/* If there is no save area and we have to allocate one, do so. Otherwise
|
951 |
|
|
verify the save area is the proper mode. */
|
952 |
|
|
|
953 |
|
|
if (sa == 0)
|
954 |
|
|
{
|
955 |
|
|
if (mode != VOIDmode)
|
956 |
|
|
{
|
957 |
|
|
if (save_level == SAVE_NONLOCAL)
|
958 |
|
|
*psave = sa = assign_stack_local (mode, GET_MODE_SIZE (mode), 0);
|
959 |
|
|
else
|
960 |
|
|
*psave = sa = gen_reg_rtx (mode);
|
961 |
|
|
}
|
962 |
|
|
}
|
963 |
|
|
|
964 |
|
|
if (after)
|
965 |
|
|
{
|
966 |
|
|
rtx seq;
|
967 |
|
|
|
968 |
|
|
start_sequence ();
|
969 |
|
|
do_pending_stack_adjust ();
|
970 |
|
|
/* We must validize inside the sequence, to ensure that any instructions
|
971 |
|
|
created by the validize call also get moved to the right place. */
|
972 |
|
|
if (sa != 0)
|
973 |
|
|
sa = validize_mem (sa);
|
974 |
|
|
emit_insn (fcn (sa, stack_pointer_rtx));
|
975 |
|
|
seq = get_insns ();
|
976 |
|
|
end_sequence ();
|
977 |
|
|
emit_insn_after (seq, after);
|
978 |
|
|
}
|
979 |
|
|
else
|
980 |
|
|
{
|
981 |
|
|
do_pending_stack_adjust ();
|
982 |
|
|
if (sa != 0)
|
983 |
|
|
sa = validize_mem (sa);
|
984 |
|
|
emit_insn (fcn (sa, stack_pointer_rtx));
|
985 |
|
|
}
|
986 |
|
|
}
|
987 |
|
|
|
988 |
|
|
/* Restore the stack pointer for the purpose in SAVE_LEVEL. SA is the save
|
989 |
|
|
area made by emit_stack_save. If it is zero, we have nothing to do.
|
990 |
|
|
|
991 |
|
|
Put any emitted insns after insn AFTER, if nonzero, otherwise at
|
992 |
|
|
current position. */
|
993 |
|
|
|
994 |
|
|
void
|
995 |
|
|
emit_stack_restore (enum save_level save_level, rtx sa, rtx after)
|
996 |
|
|
{
|
997 |
|
|
/* The default is that we use a move insn. */
|
998 |
|
|
rtx (*fcn) (rtx, rtx) = gen_move_insn;
|
999 |
|
|
|
1000 |
|
|
/* See if this machine has anything special to do for this kind of save. */
|
1001 |
|
|
switch (save_level)
|
1002 |
|
|
{
|
1003 |
|
|
#ifdef HAVE_restore_stack_block
|
1004 |
|
|
case SAVE_BLOCK:
|
1005 |
|
|
if (HAVE_restore_stack_block)
|
1006 |
|
|
fcn = gen_restore_stack_block;
|
1007 |
|
|
break;
|
1008 |
|
|
#endif
|
1009 |
|
|
#ifdef HAVE_restore_stack_function
|
1010 |
|
|
case SAVE_FUNCTION:
|
1011 |
|
|
if (HAVE_restore_stack_function)
|
1012 |
|
|
fcn = gen_restore_stack_function;
|
1013 |
|
|
break;
|
1014 |
|
|
#endif
|
1015 |
|
|
#ifdef HAVE_restore_stack_nonlocal
|
1016 |
|
|
case SAVE_NONLOCAL:
|
1017 |
|
|
if (HAVE_restore_stack_nonlocal)
|
1018 |
|
|
fcn = gen_restore_stack_nonlocal;
|
1019 |
|
|
break;
|
1020 |
|
|
#endif
|
1021 |
|
|
default:
|
1022 |
|
|
break;
|
1023 |
|
|
}
|
1024 |
|
|
|
1025 |
|
|
if (sa != 0)
|
1026 |
|
|
{
|
1027 |
|
|
sa = validize_mem (sa);
|
1028 |
|
|
/* These clobbers prevent the scheduler from moving
|
1029 |
|
|
references to variable arrays below the code
|
1030 |
|
|
that deletes (pops) the arrays. */
|
1031 |
|
|
emit_insn (gen_rtx_CLOBBER (VOIDmode,
|
1032 |
|
|
gen_rtx_MEM (BLKmode,
|
1033 |
|
|
gen_rtx_SCRATCH (VOIDmode))));
|
1034 |
|
|
emit_insn (gen_rtx_CLOBBER (VOIDmode,
|
1035 |
|
|
gen_rtx_MEM (BLKmode, stack_pointer_rtx)));
|
1036 |
|
|
}
|
1037 |
|
|
|
1038 |
|
|
discard_pending_stack_adjust ();
|
1039 |
|
|
|
1040 |
|
|
if (after)
|
1041 |
|
|
{
|
1042 |
|
|
rtx seq;
|
1043 |
|
|
|
1044 |
|
|
start_sequence ();
|
1045 |
|
|
emit_insn (fcn (stack_pointer_rtx, sa));
|
1046 |
|
|
seq = get_insns ();
|
1047 |
|
|
end_sequence ();
|
1048 |
|
|
emit_insn_after (seq, after);
|
1049 |
|
|
}
|
1050 |
|
|
else
|
1051 |
|
|
emit_insn (fcn (stack_pointer_rtx, sa));
|
1052 |
|
|
}
|
1053 |
|
|
|
1054 |
|
|
/* Invoke emit_stack_save on the nonlocal_goto_save_area for the current
|
1055 |
|
|
function. This function should be called whenever we allocate or
|
1056 |
|
|
deallocate dynamic stack space. */
|
1057 |
|
|
|
1058 |
|
|
void
|
1059 |
|
|
update_nonlocal_goto_save_area (void)
|
1060 |
|
|
{
|
1061 |
|
|
tree t_save;
|
1062 |
|
|
rtx r_save;
|
1063 |
|
|
|
1064 |
|
|
/* The nonlocal_goto_save_area object is an array of N pointers. The
|
1065 |
|
|
first one is used for the frame pointer save; the rest are sized by
|
1066 |
|
|
STACK_SAVEAREA_MODE. Create a reference to array index 1, the first
|
1067 |
|
|
of the stack save area slots. */
|
1068 |
|
|
t_save = build4 (ARRAY_REF, ptr_type_node, cfun->nonlocal_goto_save_area,
|
1069 |
|
|
integer_one_node, NULL_TREE, NULL_TREE);
|
1070 |
|
|
r_save = expand_expr (t_save, NULL_RTX, VOIDmode, EXPAND_WRITE);
|
1071 |
|
|
|
1072 |
|
|
emit_stack_save (SAVE_NONLOCAL, &r_save, NULL_RTX);
|
1073 |
|
|
}
|
1074 |
|
|
|
1075 |
|
|
/* Return an rtx representing the address of an area of memory dynamically
|
1076 |
|
|
pushed on the stack. This region of memory is always aligned to
|
1077 |
|
|
a multiple of BIGGEST_ALIGNMENT.
|
1078 |
|
|
|
1079 |
|
|
Any required stack pointer alignment is preserved.
|
1080 |
|
|
|
1081 |
|
|
SIZE is an rtx representing the size of the area.
|
1082 |
|
|
TARGET is a place in which the address can be placed.
|
1083 |
|
|
|
1084 |
|
|
KNOWN_ALIGN is the alignment (in bits) that we know SIZE has. */
|
1085 |
|
|
|
1086 |
|
|
rtx
|
1087 |
|
|
allocate_dynamic_stack_space (rtx size, rtx target, int known_align)
|
1088 |
|
|
{
|
1089 |
|
|
/* If we're asking for zero bytes, it doesn't matter what we point
|
1090 |
|
|
to since we can't dereference it. But return a reasonable
|
1091 |
|
|
address anyway. */
|
1092 |
|
|
if (size == const0_rtx)
|
1093 |
|
|
return virtual_stack_dynamic_rtx;
|
1094 |
|
|
|
1095 |
|
|
/* Otherwise, show we're calling alloca or equivalent. */
|
1096 |
|
|
current_function_calls_alloca = 1;
|
1097 |
|
|
|
1098 |
|
|
/* Ensure the size is in the proper mode. */
|
1099 |
|
|
if (GET_MODE (size) != VOIDmode && GET_MODE (size) != Pmode)
|
1100 |
|
|
size = convert_to_mode (Pmode, size, 1);
|
1101 |
|
|
|
1102 |
|
|
/* We can't attempt to minimize alignment necessary, because we don't
|
1103 |
|
|
know the final value of preferred_stack_boundary yet while executing
|
1104 |
|
|
this code. */
|
1105 |
|
|
cfun->preferred_stack_boundary = PREFERRED_STACK_BOUNDARY;
|
1106 |
|
|
|
1107 |
|
|
/* We will need to ensure that the address we return is aligned to
|
1108 |
|
|
BIGGEST_ALIGNMENT. If STACK_DYNAMIC_OFFSET is defined, we don't
|
1109 |
|
|
always know its final value at this point in the compilation (it
|
1110 |
|
|
might depend on the size of the outgoing parameter lists, for
|
1111 |
|
|
example), so we must align the value to be returned in that case.
|
1112 |
|
|
(Note that STACK_DYNAMIC_OFFSET will have a default nonzero value if
|
1113 |
|
|
STACK_POINTER_OFFSET or ACCUMULATE_OUTGOING_ARGS are defined).
|
1114 |
|
|
We must also do an alignment operation on the returned value if
|
1115 |
|
|
the stack pointer alignment is less strict that BIGGEST_ALIGNMENT.
|
1116 |
|
|
|
1117 |
|
|
If we have to align, we must leave space in SIZE for the hole
|
1118 |
|
|
that might result from the alignment operation. */
|
1119 |
|
|
|
1120 |
|
|
#if defined (STACK_DYNAMIC_OFFSET) || defined (STACK_POINTER_OFFSET)
|
1121 |
|
|
#define MUST_ALIGN 1
|
1122 |
|
|
#else
|
1123 |
|
|
#define MUST_ALIGN (PREFERRED_STACK_BOUNDARY < BIGGEST_ALIGNMENT)
|
1124 |
|
|
#endif
|
1125 |
|
|
|
1126 |
|
|
if (MUST_ALIGN)
|
1127 |
|
|
size
|
1128 |
|
|
= force_operand (plus_constant (size,
|
1129 |
|
|
BIGGEST_ALIGNMENT / BITS_PER_UNIT - 1),
|
1130 |
|
|
NULL_RTX);
|
1131 |
|
|
|
1132 |
|
|
#ifdef SETJMP_VIA_SAVE_AREA
|
1133 |
|
|
/* If setjmp restores regs from a save area in the stack frame,
|
1134 |
|
|
avoid clobbering the reg save area. Note that the offset of
|
1135 |
|
|
virtual_incoming_args_rtx includes the preallocated stack args space.
|
1136 |
|
|
It would be no problem to clobber that, but it's on the wrong side
|
1137 |
|
|
of the old save area.
|
1138 |
|
|
|
1139 |
|
|
What used to happen is that, since we did not know for sure
|
1140 |
|
|
whether setjmp() was invoked until after RTL generation, we
|
1141 |
|
|
would use reg notes to store the "optimized" size and fix things
|
1142 |
|
|
up later. These days we know this information before we ever
|
1143 |
|
|
start building RTL so the reg notes are unnecessary. */
|
1144 |
|
|
if (!current_function_calls_setjmp)
|
1145 |
|
|
{
|
1146 |
|
|
int align = PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT;
|
1147 |
|
|
|
1148 |
|
|
/* ??? Code below assumes that the save area needs maximal
|
1149 |
|
|
alignment. This constraint may be too strong. */
|
1150 |
|
|
gcc_assert (PREFERRED_STACK_BOUNDARY == BIGGEST_ALIGNMENT);
|
1151 |
|
|
|
1152 |
|
|
if (GET_CODE (size) == CONST_INT)
|
1153 |
|
|
{
|
1154 |
|
|
HOST_WIDE_INT new = INTVAL (size) / align * align;
|
1155 |
|
|
|
1156 |
|
|
if (INTVAL (size) != new)
|
1157 |
|
|
size = GEN_INT (new);
|
1158 |
|
|
}
|
1159 |
|
|
else
|
1160 |
|
|
{
|
1161 |
|
|
/* Since we know overflow is not possible, we avoid using
|
1162 |
|
|
CEIL_DIV_EXPR and use TRUNC_DIV_EXPR instead. */
|
1163 |
|
|
size = expand_divmod (0, TRUNC_DIV_EXPR, Pmode, size,
|
1164 |
|
|
GEN_INT (align), NULL_RTX, 1);
|
1165 |
|
|
size = expand_mult (Pmode, size,
|
1166 |
|
|
GEN_INT (align), NULL_RTX, 1);
|
1167 |
|
|
}
|
1168 |
|
|
}
|
1169 |
|
|
else
|
1170 |
|
|
{
|
1171 |
|
|
rtx dynamic_offset
|
1172 |
|
|
= expand_binop (Pmode, sub_optab, virtual_stack_dynamic_rtx,
|
1173 |
|
|
stack_pointer_rtx, NULL_RTX, 1, OPTAB_LIB_WIDEN);
|
1174 |
|
|
|
1175 |
|
|
size = expand_binop (Pmode, add_optab, size, dynamic_offset,
|
1176 |
|
|
NULL_RTX, 1, OPTAB_LIB_WIDEN);
|
1177 |
|
|
}
|
1178 |
|
|
#endif /* SETJMP_VIA_SAVE_AREA */
|
1179 |
|
|
|
1180 |
|
|
/* Round the size to a multiple of the required stack alignment.
|
1181 |
|
|
Since the stack if presumed to be rounded before this allocation,
|
1182 |
|
|
this will maintain the required alignment.
|
1183 |
|
|
|
1184 |
|
|
If the stack grows downward, we could save an insn by subtracting
|
1185 |
|
|
SIZE from the stack pointer and then aligning the stack pointer.
|
1186 |
|
|
The problem with this is that the stack pointer may be unaligned
|
1187 |
|
|
between the execution of the subtraction and alignment insns and
|
1188 |
|
|
some machines do not allow this. Even on those that do, some
|
1189 |
|
|
signal handlers malfunction if a signal should occur between those
|
1190 |
|
|
insns. Since this is an extremely rare event, we have no reliable
|
1191 |
|
|
way of knowing which systems have this problem. So we avoid even
|
1192 |
|
|
momentarily mis-aligning the stack. */
|
1193 |
|
|
|
1194 |
|
|
/* If we added a variable amount to SIZE,
|
1195 |
|
|
we can no longer assume it is aligned. */
|
1196 |
|
|
#if !defined (SETJMP_VIA_SAVE_AREA)
|
1197 |
|
|
if (MUST_ALIGN || known_align % PREFERRED_STACK_BOUNDARY != 0)
|
1198 |
|
|
#endif
|
1199 |
|
|
size = round_push (size);
|
1200 |
|
|
|
1201 |
|
|
do_pending_stack_adjust ();
|
1202 |
|
|
|
1203 |
|
|
/* We ought to be called always on the toplevel and stack ought to be aligned
|
1204 |
|
|
properly. */
|
1205 |
|
|
gcc_assert (!(stack_pointer_delta
|
1206 |
|
|
% (PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT)));
|
1207 |
|
|
|
1208 |
|
|
/* If needed, check that we have the required amount of stack. Take into
|
1209 |
|
|
account what has already been checked. */
|
1210 |
|
|
if (flag_stack_check && ! STACK_CHECK_BUILTIN)
|
1211 |
|
|
probe_stack_range (STACK_CHECK_MAX_FRAME_SIZE + STACK_CHECK_PROTECT, size);
|
1212 |
|
|
|
1213 |
|
|
/* Don't use a TARGET that isn't a pseudo or is the wrong mode. */
|
1214 |
|
|
if (target == 0 || !REG_P (target)
|
1215 |
|
|
|| REGNO (target) < FIRST_PSEUDO_REGISTER
|
1216 |
|
|
|| GET_MODE (target) != Pmode)
|
1217 |
|
|
target = gen_reg_rtx (Pmode);
|
1218 |
|
|
|
1219 |
|
|
mark_reg_pointer (target, known_align);
|
1220 |
|
|
|
1221 |
|
|
/* Perform the required allocation from the stack. Some systems do
|
1222 |
|
|
this differently than simply incrementing/decrementing from the
|
1223 |
|
|
stack pointer, such as acquiring the space by calling malloc(). */
|
1224 |
|
|
#ifdef HAVE_allocate_stack
|
1225 |
|
|
if (HAVE_allocate_stack)
|
1226 |
|
|
{
|
1227 |
|
|
enum machine_mode mode = STACK_SIZE_MODE;
|
1228 |
|
|
insn_operand_predicate_fn pred;
|
1229 |
|
|
|
1230 |
|
|
/* We don't have to check against the predicate for operand 0 since
|
1231 |
|
|
TARGET is known to be a pseudo of the proper mode, which must
|
1232 |
|
|
be valid for the operand. For operand 1, convert to the
|
1233 |
|
|
proper mode and validate. */
|
1234 |
|
|
if (mode == VOIDmode)
|
1235 |
|
|
mode = insn_data[(int) CODE_FOR_allocate_stack].operand[1].mode;
|
1236 |
|
|
|
1237 |
|
|
pred = insn_data[(int) CODE_FOR_allocate_stack].operand[1].predicate;
|
1238 |
|
|
if (pred && ! ((*pred) (size, mode)))
|
1239 |
|
|
size = copy_to_mode_reg (mode, convert_to_mode (mode, size, 1));
|
1240 |
|
|
|
1241 |
|
|
emit_insn (gen_allocate_stack (target, size));
|
1242 |
|
|
}
|
1243 |
|
|
else
|
1244 |
|
|
#endif
|
1245 |
|
|
{
|
1246 |
|
|
#ifndef STACK_GROWS_DOWNWARD
|
1247 |
|
|
emit_move_insn (target, virtual_stack_dynamic_rtx);
|
1248 |
|
|
#endif
|
1249 |
|
|
|
1250 |
|
|
/* Check stack bounds if necessary. */
|
1251 |
|
|
if (current_function_limit_stack)
|
1252 |
|
|
{
|
1253 |
|
|
rtx available;
|
1254 |
|
|
rtx space_available = gen_label_rtx ();
|
1255 |
|
|
#ifdef STACK_GROWS_DOWNWARD
|
1256 |
|
|
available = expand_binop (Pmode, sub_optab,
|
1257 |
|
|
stack_pointer_rtx, stack_limit_rtx,
|
1258 |
|
|
NULL_RTX, 1, OPTAB_WIDEN);
|
1259 |
|
|
#else
|
1260 |
|
|
available = expand_binop (Pmode, sub_optab,
|
1261 |
|
|
stack_limit_rtx, stack_pointer_rtx,
|
1262 |
|
|
NULL_RTX, 1, OPTAB_WIDEN);
|
1263 |
|
|
#endif
|
1264 |
|
|
emit_cmp_and_jump_insns (available, size, GEU, NULL_RTX, Pmode, 1,
|
1265 |
|
|
space_available);
|
1266 |
|
|
#ifdef HAVE_trap
|
1267 |
|
|
if (HAVE_trap)
|
1268 |
|
|
emit_insn (gen_trap ());
|
1269 |
|
|
else
|
1270 |
|
|
#endif
|
1271 |
|
|
error ("stack limits not supported on this target");
|
1272 |
|
|
emit_barrier ();
|
1273 |
|
|
emit_label (space_available);
|
1274 |
|
|
}
|
1275 |
|
|
|
1276 |
|
|
anti_adjust_stack (size);
|
1277 |
|
|
|
1278 |
|
|
#ifdef STACK_GROWS_DOWNWARD
|
1279 |
|
|
emit_move_insn (target, virtual_stack_dynamic_rtx);
|
1280 |
|
|
#endif
|
1281 |
|
|
}
|
1282 |
|
|
|
1283 |
|
|
if (MUST_ALIGN)
|
1284 |
|
|
{
|
1285 |
|
|
/* CEIL_DIV_EXPR needs to worry about the addition overflowing,
|
1286 |
|
|
but we know it can't. So add ourselves and then do
|
1287 |
|
|
TRUNC_DIV_EXPR. */
|
1288 |
|
|
target = expand_binop (Pmode, add_optab, target,
|
1289 |
|
|
GEN_INT (BIGGEST_ALIGNMENT / BITS_PER_UNIT - 1),
|
1290 |
|
|
NULL_RTX, 1, OPTAB_LIB_WIDEN);
|
1291 |
|
|
target = expand_divmod (0, TRUNC_DIV_EXPR, Pmode, target,
|
1292 |
|
|
GEN_INT (BIGGEST_ALIGNMENT / BITS_PER_UNIT),
|
1293 |
|
|
NULL_RTX, 1);
|
1294 |
|
|
target = expand_mult (Pmode, target,
|
1295 |
|
|
GEN_INT (BIGGEST_ALIGNMENT / BITS_PER_UNIT),
|
1296 |
|
|
NULL_RTX, 1);
|
1297 |
|
|
}
|
1298 |
|
|
|
1299 |
|
|
/* Record the new stack level for nonlocal gotos. */
|
1300 |
|
|
if (cfun->nonlocal_goto_save_area != 0)
|
1301 |
|
|
update_nonlocal_goto_save_area ();
|
1302 |
|
|
|
1303 |
|
|
return target;
|
1304 |
|
|
}
|
1305 |
|
|
|
1306 |
|
|
/* A front end may want to override GCC's stack checking by providing a
|
1307 |
|
|
run-time routine to call to check the stack, so provide a mechanism for
|
1308 |
|
|
calling that routine. */
|
1309 |
|
|
|
1310 |
|
|
static GTY(()) rtx stack_check_libfunc;
|
1311 |
|
|
|
1312 |
|
|
void
|
1313 |
|
|
set_stack_check_libfunc (rtx libfunc)
|
1314 |
|
|
{
|
1315 |
|
|
stack_check_libfunc = libfunc;
|
1316 |
|
|
}
|
1317 |
|
|
|
1318 |
|
|
/* Emit one stack probe at ADDRESS, an address within the stack. */
|
1319 |
|
|
|
1320 |
|
|
static void
|
1321 |
|
|
emit_stack_probe (rtx address)
|
1322 |
|
|
{
|
1323 |
|
|
rtx memref = gen_rtx_MEM (word_mode, address);
|
1324 |
|
|
|
1325 |
|
|
MEM_VOLATILE_P (memref) = 1;
|
1326 |
|
|
|
1327 |
|
|
if (STACK_CHECK_PROBE_LOAD)
|
1328 |
|
|
emit_move_insn (gen_reg_rtx (word_mode), memref);
|
1329 |
|
|
else
|
1330 |
|
|
emit_move_insn (memref, const0_rtx);
|
1331 |
|
|
}
|
1332 |
|
|
|
1333 |
|
|
/* Probe a range of stack addresses from FIRST to FIRST+SIZE, inclusive.
|
1334 |
|
|
FIRST is a constant and size is a Pmode RTX. These are offsets from the
|
1335 |
|
|
current stack pointer. STACK_GROWS_DOWNWARD says whether to add or
|
1336 |
|
|
subtract from the stack. If SIZE is constant, this is done
|
1337 |
|
|
with a fixed number of probes. Otherwise, we must make a loop. */
|
1338 |
|
|
|
1339 |
|
|
#ifdef STACK_GROWS_DOWNWARD
|
1340 |
|
|
#define STACK_GROW_OP MINUS
|
1341 |
|
|
#else
|
1342 |
|
|
#define STACK_GROW_OP PLUS
|
1343 |
|
|
#endif
|
1344 |
|
|
|
1345 |
|
|
void
|
1346 |
|
|
probe_stack_range (HOST_WIDE_INT first, rtx size)
|
1347 |
|
|
{
|
1348 |
|
|
/* First ensure SIZE is Pmode. */
|
1349 |
|
|
if (GET_MODE (size) != VOIDmode && GET_MODE (size) != Pmode)
|
1350 |
|
|
size = convert_to_mode (Pmode, size, 1);
|
1351 |
|
|
|
1352 |
|
|
/* Next see if the front end has set up a function for us to call to
|
1353 |
|
|
check the stack. */
|
1354 |
|
|
if (stack_check_libfunc != 0)
|
1355 |
|
|
{
|
1356 |
|
|
rtx addr = memory_address (QImode,
|
1357 |
|
|
gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
|
1358 |
|
|
stack_pointer_rtx,
|
1359 |
|
|
plus_constant (size, first)));
|
1360 |
|
|
|
1361 |
|
|
addr = convert_memory_address (ptr_mode, addr);
|
1362 |
|
|
emit_library_call (stack_check_libfunc, LCT_NORMAL, VOIDmode, 1, addr,
|
1363 |
|
|
ptr_mode);
|
1364 |
|
|
}
|
1365 |
|
|
|
1366 |
|
|
/* Next see if we have an insn to check the stack. Use it if so. */
|
1367 |
|
|
#ifdef HAVE_check_stack
|
1368 |
|
|
else if (HAVE_check_stack)
|
1369 |
|
|
{
|
1370 |
|
|
insn_operand_predicate_fn pred;
|
1371 |
|
|
rtx last_addr
|
1372 |
|
|
= force_operand (gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
|
1373 |
|
|
stack_pointer_rtx,
|
1374 |
|
|
plus_constant (size, first)),
|
1375 |
|
|
NULL_RTX);
|
1376 |
|
|
|
1377 |
|
|
pred = insn_data[(int) CODE_FOR_check_stack].operand[0].predicate;
|
1378 |
|
|
if (pred && ! ((*pred) (last_addr, Pmode)))
|
1379 |
|
|
last_addr = copy_to_mode_reg (Pmode, last_addr);
|
1380 |
|
|
|
1381 |
|
|
emit_insn (gen_check_stack (last_addr));
|
1382 |
|
|
}
|
1383 |
|
|
#endif
|
1384 |
|
|
|
1385 |
|
|
/* If we have to generate explicit probes, see if we have a constant
|
1386 |
|
|
small number of them to generate. If so, that's the easy case. */
|
1387 |
|
|
else if (GET_CODE (size) == CONST_INT
|
1388 |
|
|
&& INTVAL (size) < 10 * STACK_CHECK_PROBE_INTERVAL)
|
1389 |
|
|
{
|
1390 |
|
|
HOST_WIDE_INT offset;
|
1391 |
|
|
|
1392 |
|
|
/* Start probing at FIRST + N * STACK_CHECK_PROBE_INTERVAL
|
1393 |
|
|
for values of N from 1 until it exceeds LAST. If only one
|
1394 |
|
|
probe is needed, this will not generate any code. Then probe
|
1395 |
|
|
at LAST. */
|
1396 |
|
|
for (offset = first + STACK_CHECK_PROBE_INTERVAL;
|
1397 |
|
|
offset < INTVAL (size);
|
1398 |
|
|
offset = offset + STACK_CHECK_PROBE_INTERVAL)
|
1399 |
|
|
emit_stack_probe (gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
|
1400 |
|
|
stack_pointer_rtx,
|
1401 |
|
|
GEN_INT (offset)));
|
1402 |
|
|
|
1403 |
|
|
emit_stack_probe (gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
|
1404 |
|
|
stack_pointer_rtx,
|
1405 |
|
|
plus_constant (size, first)));
|
1406 |
|
|
}
|
1407 |
|
|
|
1408 |
|
|
/* In the variable case, do the same as above, but in a loop. We emit loop
|
1409 |
|
|
notes so that loop optimization can be done. */
|
1410 |
|
|
else
|
1411 |
|
|
{
|
1412 |
|
|
rtx test_addr
|
1413 |
|
|
= force_operand (gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
|
1414 |
|
|
stack_pointer_rtx,
|
1415 |
|
|
GEN_INT (first + STACK_CHECK_PROBE_INTERVAL)),
|
1416 |
|
|
NULL_RTX);
|
1417 |
|
|
rtx last_addr
|
1418 |
|
|
= force_operand (gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
|
1419 |
|
|
stack_pointer_rtx,
|
1420 |
|
|
plus_constant (size, first)),
|
1421 |
|
|
NULL_RTX);
|
1422 |
|
|
rtx incr = GEN_INT (STACK_CHECK_PROBE_INTERVAL);
|
1423 |
|
|
rtx loop_lab = gen_label_rtx ();
|
1424 |
|
|
rtx test_lab = gen_label_rtx ();
|
1425 |
|
|
rtx end_lab = gen_label_rtx ();
|
1426 |
|
|
rtx temp;
|
1427 |
|
|
|
1428 |
|
|
if (!REG_P (test_addr)
|
1429 |
|
|
|| REGNO (test_addr) < FIRST_PSEUDO_REGISTER)
|
1430 |
|
|
test_addr = force_reg (Pmode, test_addr);
|
1431 |
|
|
|
1432 |
|
|
emit_jump (test_lab);
|
1433 |
|
|
|
1434 |
|
|
emit_label (loop_lab);
|
1435 |
|
|
emit_stack_probe (test_addr);
|
1436 |
|
|
|
1437 |
|
|
#ifdef STACK_GROWS_DOWNWARD
|
1438 |
|
|
#define CMP_OPCODE GTU
|
1439 |
|
|
temp = expand_binop (Pmode, sub_optab, test_addr, incr, test_addr,
|
1440 |
|
|
1, OPTAB_WIDEN);
|
1441 |
|
|
#else
|
1442 |
|
|
#define CMP_OPCODE LTU
|
1443 |
|
|
temp = expand_binop (Pmode, add_optab, test_addr, incr, test_addr,
|
1444 |
|
|
1, OPTAB_WIDEN);
|
1445 |
|
|
#endif
|
1446 |
|
|
|
1447 |
|
|
gcc_assert (temp == test_addr);
|
1448 |
|
|
|
1449 |
|
|
emit_label (test_lab);
|
1450 |
|
|
emit_cmp_and_jump_insns (test_addr, last_addr, CMP_OPCODE,
|
1451 |
|
|
NULL_RTX, Pmode, 1, loop_lab);
|
1452 |
|
|
emit_jump (end_lab);
|
1453 |
|
|
emit_label (end_lab);
|
1454 |
|
|
|
1455 |
|
|
emit_stack_probe (last_addr);
|
1456 |
|
|
}
|
1457 |
|
|
}
|
1458 |
|
|
|
1459 |
|
|
/* Return an rtx representing the register or memory location
|
1460 |
|
|
in which a scalar value of data type VALTYPE
|
1461 |
|
|
was returned by a function call to function FUNC.
|
1462 |
|
|
FUNC is a FUNCTION_DECL, FNTYPE a FUNCTION_TYPE node if the precise
|
1463 |
|
|
function is known, otherwise 0.
|
1464 |
|
|
OUTGOING is 1 if on a machine with register windows this function
|
1465 |
|
|
should return the register in which the function will put its result
|
1466 |
|
|
and 0 otherwise. */
|
1467 |
|
|
|
1468 |
|
|
rtx
|
1469 |
|
|
hard_function_value (tree valtype, tree func, tree fntype,
|
1470 |
|
|
int outgoing ATTRIBUTE_UNUSED)
|
1471 |
|
|
{
|
1472 |
|
|
rtx val;
|
1473 |
|
|
|
1474 |
|
|
val = targetm.calls.function_value (valtype, func ? func : fntype, outgoing);
|
1475 |
|
|
|
1476 |
|
|
if (REG_P (val)
|
1477 |
|
|
&& GET_MODE (val) == BLKmode)
|
1478 |
|
|
{
|
1479 |
|
|
unsigned HOST_WIDE_INT bytes = int_size_in_bytes (valtype);
|
1480 |
|
|
enum machine_mode tmpmode;
|
1481 |
|
|
|
1482 |
|
|
/* int_size_in_bytes can return -1. We don't need a check here
|
1483 |
|
|
since the value of bytes will then be large enough that no
|
1484 |
|
|
mode will match anyway. */
|
1485 |
|
|
|
1486 |
|
|
for (tmpmode = GET_CLASS_NARROWEST_MODE (MODE_INT);
|
1487 |
|
|
tmpmode != VOIDmode;
|
1488 |
|
|
tmpmode = GET_MODE_WIDER_MODE (tmpmode))
|
1489 |
|
|
{
|
1490 |
|
|
/* Have we found a large enough mode? */
|
1491 |
|
|
if (GET_MODE_SIZE (tmpmode) >= bytes)
|
1492 |
|
|
break;
|
1493 |
|
|
}
|
1494 |
|
|
|
1495 |
|
|
/* No suitable mode found. */
|
1496 |
|
|
gcc_assert (tmpmode != VOIDmode);
|
1497 |
|
|
|
1498 |
|
|
PUT_MODE (val, tmpmode);
|
1499 |
|
|
}
|
1500 |
|
|
return val;
|
1501 |
|
|
}
|
1502 |
|
|
|
1503 |
|
|
/* Return an rtx representing the register or memory location
|
1504 |
|
|
in which a scalar value of mode MODE was returned by a library call. */
|
1505 |
|
|
|
1506 |
|
|
rtx
|
1507 |
|
|
hard_libcall_value (enum machine_mode mode)
|
1508 |
|
|
{
|
1509 |
|
|
return LIBCALL_VALUE (mode);
|
1510 |
|
|
}
|
1511 |
|
|
|
1512 |
|
|
/* Look up the tree code for a given rtx code
|
1513 |
|
|
to provide the arithmetic operation for REAL_ARITHMETIC.
|
1514 |
|
|
The function returns an int because the caller may not know
|
1515 |
|
|
what `enum tree_code' means. */
|
1516 |
|
|
|
1517 |
|
|
int
|
1518 |
|
|
rtx_to_tree_code (enum rtx_code code)
|
1519 |
|
|
{
|
1520 |
|
|
enum tree_code tcode;
|
1521 |
|
|
|
1522 |
|
|
switch (code)
|
1523 |
|
|
{
|
1524 |
|
|
case PLUS:
|
1525 |
|
|
tcode = PLUS_EXPR;
|
1526 |
|
|
break;
|
1527 |
|
|
case MINUS:
|
1528 |
|
|
tcode = MINUS_EXPR;
|
1529 |
|
|
break;
|
1530 |
|
|
case MULT:
|
1531 |
|
|
tcode = MULT_EXPR;
|
1532 |
|
|
break;
|
1533 |
|
|
case DIV:
|
1534 |
|
|
tcode = RDIV_EXPR;
|
1535 |
|
|
break;
|
1536 |
|
|
case SMIN:
|
1537 |
|
|
tcode = MIN_EXPR;
|
1538 |
|
|
break;
|
1539 |
|
|
case SMAX:
|
1540 |
|
|
tcode = MAX_EXPR;
|
1541 |
|
|
break;
|
1542 |
|
|
default:
|
1543 |
|
|
tcode = LAST_AND_UNUSED_TREE_CODE;
|
1544 |
|
|
break;
|
1545 |
|
|
}
|
1546 |
|
|
return ((int) tcode);
|
1547 |
|
|
}
|
1548 |
|
|
|
1549 |
|
|
#include "gt-explow.h"
|