1 |
38 |
julius |
/* Tree inlining.
|
2 |
|
|
Copyright 2001, 2002, 2003, 2004, 2005, 2006, 2007
|
3 |
|
|
Free Software Foundation, Inc.
|
4 |
|
|
Contributed by Alexandre Oliva <aoliva@redhat.com>
|
5 |
|
|
|
6 |
|
|
This file is part of GCC.
|
7 |
|
|
|
8 |
|
|
GCC is free software; you can redistribute it and/or modify
|
9 |
|
|
it under the terms of the GNU General Public License as published by
|
10 |
|
|
the Free Software Foundation; either version 3, or (at your option)
|
11 |
|
|
any later version.
|
12 |
|
|
|
13 |
|
|
GCC is distributed in the hope that it will be useful,
|
14 |
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
15 |
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
16 |
|
|
GNU General Public License for more details.
|
17 |
|
|
|
18 |
|
|
You should have received a copy of the GNU General Public License
|
19 |
|
|
along with GCC; see the file COPYING3. If not see
|
20 |
|
|
<http://www.gnu.org/licenses/>. */
|
21 |
|
|
|
22 |
|
|
#include "config.h"
|
23 |
|
|
#include "system.h"
|
24 |
|
|
#include "coretypes.h"
|
25 |
|
|
#include "tm.h"
|
26 |
|
|
#include "toplev.h"
|
27 |
|
|
#include "tree.h"
|
28 |
|
|
#include "tree-inline.h"
|
29 |
|
|
#include "rtl.h"
|
30 |
|
|
#include "expr.h"
|
31 |
|
|
#include "flags.h"
|
32 |
|
|
#include "params.h"
|
33 |
|
|
#include "input.h"
|
34 |
|
|
#include "insn-config.h"
|
35 |
|
|
#include "varray.h"
|
36 |
|
|
#include "hashtab.h"
|
37 |
|
|
#include "langhooks.h"
|
38 |
|
|
#include "basic-block.h"
|
39 |
|
|
#include "tree-iterator.h"
|
40 |
|
|
#include "cgraph.h"
|
41 |
|
|
#include "intl.h"
|
42 |
|
|
#include "tree-mudflap.h"
|
43 |
|
|
#include "tree-flow.h"
|
44 |
|
|
#include "function.h"
|
45 |
|
|
#include "ggc.h"
|
46 |
|
|
#include "tree-flow.h"
|
47 |
|
|
#include "diagnostic.h"
|
48 |
|
|
#include "except.h"
|
49 |
|
|
#include "debug.h"
|
50 |
|
|
#include "pointer-set.h"
|
51 |
|
|
#include "ipa-prop.h"
|
52 |
|
|
|
53 |
|
|
/* I'm not real happy about this, but we need to handle gimple and
|
54 |
|
|
non-gimple trees. */
|
55 |
|
|
#include "tree-gimple.h"
|
56 |
|
|
|
57 |
|
|
/* Inlining, Cloning, Versioning, Parallelization
|
58 |
|
|
|
59 |
|
|
Inlining: a function body is duplicated, but the PARM_DECLs are
|
60 |
|
|
remapped into VAR_DECLs, and non-void RETURN_EXPRs become
|
61 |
|
|
MODIFY_EXPRs that store to a dedicated returned-value variable.
|
62 |
|
|
The duplicated eh_region info of the copy will later be appended
|
63 |
|
|
to the info for the caller; the eh_region info in copied throwing
|
64 |
|
|
statements and RESX_EXPRs is adjusted accordingly.
|
65 |
|
|
|
66 |
|
|
Cloning: (only in C++) We have one body for a con/de/structor, and
|
67 |
|
|
multiple function decls, each with a unique parameter list.
|
68 |
|
|
Duplicate the body, using the given splay tree; some parameters
|
69 |
|
|
will become constants (like 0 or 1).
|
70 |
|
|
|
71 |
|
|
Versioning: a function body is duplicated and the result is a new
|
72 |
|
|
function rather than into blocks of an existing function as with
|
73 |
|
|
inlining. Some parameters will become constants.
|
74 |
|
|
|
75 |
|
|
Parallelization: a region of a function is duplicated resulting in
|
76 |
|
|
a new function. Variables may be replaced with complex expressions
|
77 |
|
|
to enable shared variable semantics.
|
78 |
|
|
|
79 |
|
|
All of these will simultaneously lookup any callgraph edges. If
|
80 |
|
|
we're going to inline the duplicated function body, and the given
|
81 |
|
|
function has some cloned callgraph nodes (one for each place this
|
82 |
|
|
function will be inlined) those callgraph edges will be duplicated.
|
83 |
|
|
If we're cloning the body, those callgraph edges will be
|
84 |
|
|
updated to point into the new body. (Note that the original
|
85 |
|
|
callgraph node and edge list will not be altered.)
|
86 |
|
|
|
87 |
|
|
See the CALL_EXPR handling case in copy_body_r (). */
|
88 |
|
|
|
89 |
|
|
/* 0 if we should not perform inlining.
|
90 |
|
|
1 if we should expand functions calls inline at the tree level.
|
91 |
|
|
2 if we should consider *all* functions to be inline
|
92 |
|
|
candidates. */
|
93 |
|
|
|
94 |
|
|
int flag_inline_trees = 0;
|
95 |
|
|
|
96 |
|
|
/* To Do:
|
97 |
|
|
|
98 |
|
|
o In order to make inlining-on-trees work, we pessimized
|
99 |
|
|
function-local static constants. In particular, they are now
|
100 |
|
|
always output, even when not addressed. Fix this by treating
|
101 |
|
|
function-local static constants just like global static
|
102 |
|
|
constants; the back-end already knows not to output them if they
|
103 |
|
|
are not needed.
|
104 |
|
|
|
105 |
|
|
o Provide heuristics to clamp inlining of recursive template
|
106 |
|
|
calls? */
|
107 |
|
|
|
108 |
|
|
/* Prototypes. */
|
109 |
|
|
|
110 |
|
|
static tree declare_return_variable (copy_body_data *, tree, tree, tree *);
|
111 |
|
|
static tree copy_generic_body (copy_body_data *);
|
112 |
|
|
static bool inlinable_function_p (tree);
|
113 |
|
|
static void remap_block (tree *, copy_body_data *);
|
114 |
|
|
static tree remap_decls (tree, copy_body_data *);
|
115 |
|
|
static void copy_bind_expr (tree *, int *, copy_body_data *);
|
116 |
|
|
static tree mark_local_for_remap_r (tree *, int *, void *);
|
117 |
|
|
static void unsave_expr_1 (tree);
|
118 |
|
|
static tree unsave_r (tree *, int *, void *);
|
119 |
|
|
static void declare_inline_vars (tree, tree);
|
120 |
|
|
static void remap_save_expr (tree *, void *, int *);
|
121 |
|
|
static void add_lexical_block (tree current_block, tree new_block);
|
122 |
|
|
static tree copy_decl_to_var (tree, copy_body_data *);
|
123 |
|
|
static tree copy_result_decl_to_var (tree, copy_body_data *);
|
124 |
|
|
static tree copy_decl_no_change (tree, copy_body_data *);
|
125 |
|
|
static tree copy_decl_maybe_to_var (tree, copy_body_data *);
|
126 |
|
|
|
127 |
|
|
/* Insert a tree->tree mapping for ID. Despite the name suggests
|
128 |
|
|
that the trees should be variables, it is used for more than that. */
|
129 |
|
|
|
130 |
|
|
void
|
131 |
|
|
insert_decl_map (copy_body_data *id, tree key, tree value)
|
132 |
|
|
{
|
133 |
|
|
splay_tree_insert (id->decl_map, (splay_tree_key) key,
|
134 |
|
|
(splay_tree_value) value);
|
135 |
|
|
|
136 |
|
|
/* Always insert an identity map as well. If we see this same new
|
137 |
|
|
node again, we won't want to duplicate it a second time. */
|
138 |
|
|
if (key != value)
|
139 |
|
|
splay_tree_insert (id->decl_map, (splay_tree_key) value,
|
140 |
|
|
(splay_tree_value) value);
|
141 |
|
|
}
|
142 |
|
|
|
143 |
|
|
/* Remap DECL during the copying of the BLOCK tree for the function. */
|
144 |
|
|
|
145 |
|
|
tree
|
146 |
|
|
remap_decl (tree decl, copy_body_data *id)
|
147 |
|
|
{
|
148 |
|
|
splay_tree_node n;
|
149 |
|
|
tree fn;
|
150 |
|
|
|
151 |
|
|
/* We only remap local variables in the current function. */
|
152 |
|
|
fn = id->src_fn;
|
153 |
|
|
|
154 |
|
|
/* See if we have remapped this declaration. */
|
155 |
|
|
|
156 |
|
|
n = splay_tree_lookup (id->decl_map, (splay_tree_key) decl);
|
157 |
|
|
|
158 |
|
|
/* If we didn't already have an equivalent for this declaration,
|
159 |
|
|
create one now. */
|
160 |
|
|
if (!n)
|
161 |
|
|
{
|
162 |
|
|
/* Make a copy of the variable or label. */
|
163 |
|
|
tree t = id->copy_decl (decl, id);
|
164 |
|
|
|
165 |
|
|
/* Remember it, so that if we encounter this local entity again
|
166 |
|
|
we can reuse this copy. Do this early because remap_type may
|
167 |
|
|
need this decl for TYPE_STUB_DECL. */
|
168 |
|
|
insert_decl_map (id, decl, t);
|
169 |
|
|
|
170 |
|
|
if (!DECL_P (t))
|
171 |
|
|
return t;
|
172 |
|
|
|
173 |
|
|
/* Remap types, if necessary. */
|
174 |
|
|
TREE_TYPE (t) = remap_type (TREE_TYPE (t), id);
|
175 |
|
|
if (TREE_CODE (t) == TYPE_DECL)
|
176 |
|
|
DECL_ORIGINAL_TYPE (t) = remap_type (DECL_ORIGINAL_TYPE (t), id);
|
177 |
|
|
|
178 |
|
|
/* Remap sizes as necessary. */
|
179 |
|
|
walk_tree (&DECL_SIZE (t), copy_body_r, id, NULL);
|
180 |
|
|
walk_tree (&DECL_SIZE_UNIT (t), copy_body_r, id, NULL);
|
181 |
|
|
|
182 |
|
|
/* If fields, do likewise for offset and qualifier. */
|
183 |
|
|
if (TREE_CODE (t) == FIELD_DECL)
|
184 |
|
|
{
|
185 |
|
|
walk_tree (&DECL_FIELD_OFFSET (t), copy_body_r, id, NULL);
|
186 |
|
|
if (TREE_CODE (DECL_CONTEXT (t)) == QUAL_UNION_TYPE)
|
187 |
|
|
walk_tree (&DECL_QUALIFIER (t), copy_body_r, id, NULL);
|
188 |
|
|
}
|
189 |
|
|
|
190 |
|
|
return t;
|
191 |
|
|
}
|
192 |
|
|
|
193 |
|
|
return unshare_expr ((tree) n->value);
|
194 |
|
|
}
|
195 |
|
|
|
196 |
|
|
static tree
|
197 |
|
|
remap_type_1 (tree type, copy_body_data *id)
|
198 |
|
|
{
|
199 |
|
|
splay_tree_node node;
|
200 |
|
|
tree new, t;
|
201 |
|
|
|
202 |
|
|
if (type == NULL)
|
203 |
|
|
return type;
|
204 |
|
|
|
205 |
|
|
/* See if we have remapped this type. */
|
206 |
|
|
node = splay_tree_lookup (id->decl_map, (splay_tree_key) type);
|
207 |
|
|
if (node)
|
208 |
|
|
return (tree) node->value;
|
209 |
|
|
|
210 |
|
|
/* The type only needs remapping if it's variably modified. */
|
211 |
|
|
if (! variably_modified_type_p (type, id->src_fn))
|
212 |
|
|
{
|
213 |
|
|
insert_decl_map (id, type, type);
|
214 |
|
|
return type;
|
215 |
|
|
}
|
216 |
|
|
|
217 |
|
|
/* We do need a copy. build and register it now. If this is a pointer or
|
218 |
|
|
reference type, remap the designated type and make a new pointer or
|
219 |
|
|
reference type. */
|
220 |
|
|
if (TREE_CODE (type) == POINTER_TYPE)
|
221 |
|
|
{
|
222 |
|
|
new = build_pointer_type_for_mode (remap_type (TREE_TYPE (type), id),
|
223 |
|
|
TYPE_MODE (type),
|
224 |
|
|
TYPE_REF_CAN_ALIAS_ALL (type));
|
225 |
|
|
insert_decl_map (id, type, new);
|
226 |
|
|
return new;
|
227 |
|
|
}
|
228 |
|
|
else if (TREE_CODE (type) == REFERENCE_TYPE)
|
229 |
|
|
{
|
230 |
|
|
new = build_reference_type_for_mode (remap_type (TREE_TYPE (type), id),
|
231 |
|
|
TYPE_MODE (type),
|
232 |
|
|
TYPE_REF_CAN_ALIAS_ALL (type));
|
233 |
|
|
insert_decl_map (id, type, new);
|
234 |
|
|
return new;
|
235 |
|
|
}
|
236 |
|
|
else
|
237 |
|
|
new = copy_node (type);
|
238 |
|
|
|
239 |
|
|
insert_decl_map (id, type, new);
|
240 |
|
|
|
241 |
|
|
/* This is a new type, not a copy of an old type. Need to reassociate
|
242 |
|
|
variants. We can handle everything except the main variant lazily. */
|
243 |
|
|
t = TYPE_MAIN_VARIANT (type);
|
244 |
|
|
if (type != t)
|
245 |
|
|
{
|
246 |
|
|
t = remap_type (t, id);
|
247 |
|
|
TYPE_MAIN_VARIANT (new) = t;
|
248 |
|
|
TYPE_NEXT_VARIANT (new) = TYPE_MAIN_VARIANT (t);
|
249 |
|
|
TYPE_NEXT_VARIANT (t) = new;
|
250 |
|
|
}
|
251 |
|
|
else
|
252 |
|
|
{
|
253 |
|
|
TYPE_MAIN_VARIANT (new) = new;
|
254 |
|
|
TYPE_NEXT_VARIANT (new) = NULL;
|
255 |
|
|
}
|
256 |
|
|
|
257 |
|
|
if (TYPE_STUB_DECL (type))
|
258 |
|
|
TYPE_STUB_DECL (new) = remap_decl (TYPE_STUB_DECL (type), id);
|
259 |
|
|
|
260 |
|
|
/* Lazily create pointer and reference types. */
|
261 |
|
|
TYPE_POINTER_TO (new) = NULL;
|
262 |
|
|
TYPE_REFERENCE_TO (new) = NULL;
|
263 |
|
|
|
264 |
|
|
switch (TREE_CODE (new))
|
265 |
|
|
{
|
266 |
|
|
case INTEGER_TYPE:
|
267 |
|
|
case REAL_TYPE:
|
268 |
|
|
case ENUMERAL_TYPE:
|
269 |
|
|
case BOOLEAN_TYPE:
|
270 |
|
|
t = TYPE_MIN_VALUE (new);
|
271 |
|
|
if (t && TREE_CODE (t) != INTEGER_CST)
|
272 |
|
|
walk_tree (&TYPE_MIN_VALUE (new), copy_body_r, id, NULL);
|
273 |
|
|
|
274 |
|
|
t = TYPE_MAX_VALUE (new);
|
275 |
|
|
if (t && TREE_CODE (t) != INTEGER_CST)
|
276 |
|
|
walk_tree (&TYPE_MAX_VALUE (new), copy_body_r, id, NULL);
|
277 |
|
|
return new;
|
278 |
|
|
|
279 |
|
|
case FUNCTION_TYPE:
|
280 |
|
|
TREE_TYPE (new) = remap_type (TREE_TYPE (new), id);
|
281 |
|
|
walk_tree (&TYPE_ARG_TYPES (new), copy_body_r, id, NULL);
|
282 |
|
|
return new;
|
283 |
|
|
|
284 |
|
|
case ARRAY_TYPE:
|
285 |
|
|
TREE_TYPE (new) = remap_type (TREE_TYPE (new), id);
|
286 |
|
|
TYPE_DOMAIN (new) = remap_type (TYPE_DOMAIN (new), id);
|
287 |
|
|
break;
|
288 |
|
|
|
289 |
|
|
case RECORD_TYPE:
|
290 |
|
|
case UNION_TYPE:
|
291 |
|
|
case QUAL_UNION_TYPE:
|
292 |
|
|
{
|
293 |
|
|
tree f, nf = NULL;
|
294 |
|
|
|
295 |
|
|
for (f = TYPE_FIELDS (new); f ; f = TREE_CHAIN (f))
|
296 |
|
|
{
|
297 |
|
|
t = remap_decl (f, id);
|
298 |
|
|
DECL_CONTEXT (t) = new;
|
299 |
|
|
TREE_CHAIN (t) = nf;
|
300 |
|
|
nf = t;
|
301 |
|
|
}
|
302 |
|
|
TYPE_FIELDS (new) = nreverse (nf);
|
303 |
|
|
}
|
304 |
|
|
break;
|
305 |
|
|
|
306 |
|
|
case OFFSET_TYPE:
|
307 |
|
|
default:
|
308 |
|
|
/* Shouldn't have been thought variable sized. */
|
309 |
|
|
gcc_unreachable ();
|
310 |
|
|
}
|
311 |
|
|
|
312 |
|
|
walk_tree (&TYPE_SIZE (new), copy_body_r, id, NULL);
|
313 |
|
|
walk_tree (&TYPE_SIZE_UNIT (new), copy_body_r, id, NULL);
|
314 |
|
|
|
315 |
|
|
return new;
|
316 |
|
|
}
|
317 |
|
|
|
318 |
|
|
tree
|
319 |
|
|
remap_type (tree type, copy_body_data *id)
|
320 |
|
|
{
|
321 |
|
|
splay_tree_node node;
|
322 |
|
|
|
323 |
|
|
if (type == NULL)
|
324 |
|
|
return type;
|
325 |
|
|
|
326 |
|
|
/* See if we have remapped this type. */
|
327 |
|
|
node = splay_tree_lookup (id->decl_map, (splay_tree_key) type);
|
328 |
|
|
if (node)
|
329 |
|
|
return (tree) node->value;
|
330 |
|
|
|
331 |
|
|
/* The type only needs remapping if it's variably modified. */
|
332 |
|
|
if (! variably_modified_type_p (type, id->src_fn))
|
333 |
|
|
{
|
334 |
|
|
insert_decl_map (id, type, type);
|
335 |
|
|
return type;
|
336 |
|
|
}
|
337 |
|
|
|
338 |
|
|
return remap_type_1 (type, id);
|
339 |
|
|
}
|
340 |
|
|
|
341 |
|
|
static tree
|
342 |
|
|
remap_decls (tree decls, copy_body_data *id)
|
343 |
|
|
{
|
344 |
|
|
tree old_var;
|
345 |
|
|
tree new_decls = NULL_TREE;
|
346 |
|
|
|
347 |
|
|
/* Remap its variables. */
|
348 |
|
|
for (old_var = decls; old_var; old_var = TREE_CHAIN (old_var))
|
349 |
|
|
{
|
350 |
|
|
tree new_var;
|
351 |
|
|
|
352 |
|
|
/* We can not chain the local static declarations into the unexpanded_var_list
|
353 |
|
|
as we can't duplicate them or break one decl rule. Go ahead and link
|
354 |
|
|
them into unexpanded_var_list. */
|
355 |
|
|
if (!lang_hooks.tree_inlining.auto_var_in_fn_p (old_var, id->src_fn)
|
356 |
|
|
&& !DECL_EXTERNAL (old_var))
|
357 |
|
|
{
|
358 |
|
|
cfun->unexpanded_var_list = tree_cons (NULL_TREE, old_var,
|
359 |
|
|
cfun->unexpanded_var_list);
|
360 |
|
|
continue;
|
361 |
|
|
}
|
362 |
|
|
|
363 |
|
|
/* Remap the variable. */
|
364 |
|
|
new_var = remap_decl (old_var, id);
|
365 |
|
|
|
366 |
|
|
/* If we didn't remap this variable, so we can't mess with its
|
367 |
|
|
TREE_CHAIN. If we remapped this variable to the return slot, it's
|
368 |
|
|
already declared somewhere else, so don't declare it here. */
|
369 |
|
|
if (!new_var || new_var == id->retvar)
|
370 |
|
|
;
|
371 |
|
|
else
|
372 |
|
|
{
|
373 |
|
|
gcc_assert (DECL_P (new_var));
|
374 |
|
|
TREE_CHAIN (new_var) = new_decls;
|
375 |
|
|
new_decls = new_var;
|
376 |
|
|
}
|
377 |
|
|
}
|
378 |
|
|
|
379 |
|
|
return nreverse (new_decls);
|
380 |
|
|
}
|
381 |
|
|
|
382 |
|
|
/* Copy the BLOCK to contain remapped versions of the variables
|
383 |
|
|
therein. And hook the new block into the block-tree. */
|
384 |
|
|
|
385 |
|
|
static void
|
386 |
|
|
remap_block (tree *block, copy_body_data *id)
|
387 |
|
|
{
|
388 |
|
|
tree old_block;
|
389 |
|
|
tree new_block;
|
390 |
|
|
tree fn;
|
391 |
|
|
|
392 |
|
|
/* Make the new block. */
|
393 |
|
|
old_block = *block;
|
394 |
|
|
new_block = make_node (BLOCK);
|
395 |
|
|
TREE_USED (new_block) = TREE_USED (old_block);
|
396 |
|
|
BLOCK_ABSTRACT_ORIGIN (new_block) = old_block;
|
397 |
|
|
BLOCK_SOURCE_LOCATION (new_block) = BLOCK_SOURCE_LOCATION (old_block);
|
398 |
|
|
*block = new_block;
|
399 |
|
|
|
400 |
|
|
/* Remap its variables. */
|
401 |
|
|
BLOCK_VARS (new_block) = remap_decls (BLOCK_VARS (old_block), id);
|
402 |
|
|
|
403 |
|
|
fn = id->dst_fn;
|
404 |
|
|
|
405 |
|
|
if (id->transform_lang_insert_block)
|
406 |
|
|
lang_hooks.decls.insert_block (new_block);
|
407 |
|
|
|
408 |
|
|
/* Remember the remapped block. */
|
409 |
|
|
insert_decl_map (id, old_block, new_block);
|
410 |
|
|
}
|
411 |
|
|
|
412 |
|
|
/* Copy the whole block tree and root it in id->block. */
|
413 |
|
|
static tree
|
414 |
|
|
remap_blocks (tree block, copy_body_data *id)
|
415 |
|
|
{
|
416 |
|
|
tree t;
|
417 |
|
|
tree new = block;
|
418 |
|
|
|
419 |
|
|
if (!block)
|
420 |
|
|
return NULL;
|
421 |
|
|
|
422 |
|
|
remap_block (&new, id);
|
423 |
|
|
gcc_assert (new != block);
|
424 |
|
|
for (t = BLOCK_SUBBLOCKS (block); t ; t = BLOCK_CHAIN (t))
|
425 |
|
|
add_lexical_block (new, remap_blocks (t, id));
|
426 |
|
|
return new;
|
427 |
|
|
}
|
428 |
|
|
|
429 |
|
|
static void
|
430 |
|
|
copy_statement_list (tree *tp)
|
431 |
|
|
{
|
432 |
|
|
tree_stmt_iterator oi, ni;
|
433 |
|
|
tree new;
|
434 |
|
|
|
435 |
|
|
new = alloc_stmt_list ();
|
436 |
|
|
ni = tsi_start (new);
|
437 |
|
|
oi = tsi_start (*tp);
|
438 |
|
|
*tp = new;
|
439 |
|
|
|
440 |
|
|
for (; !tsi_end_p (oi); tsi_next (&oi))
|
441 |
|
|
tsi_link_after (&ni, tsi_stmt (oi), TSI_NEW_STMT);
|
442 |
|
|
}
|
443 |
|
|
|
444 |
|
|
static void
|
445 |
|
|
copy_bind_expr (tree *tp, int *walk_subtrees, copy_body_data *id)
|
446 |
|
|
{
|
447 |
|
|
tree block = BIND_EXPR_BLOCK (*tp);
|
448 |
|
|
/* Copy (and replace) the statement. */
|
449 |
|
|
copy_tree_r (tp, walk_subtrees, NULL);
|
450 |
|
|
if (block)
|
451 |
|
|
{
|
452 |
|
|
remap_block (&block, id);
|
453 |
|
|
BIND_EXPR_BLOCK (*tp) = block;
|
454 |
|
|
}
|
455 |
|
|
|
456 |
|
|
if (BIND_EXPR_VARS (*tp))
|
457 |
|
|
/* This will remap a lot of the same decls again, but this should be
|
458 |
|
|
harmless. */
|
459 |
|
|
BIND_EXPR_VARS (*tp) = remap_decls (BIND_EXPR_VARS (*tp), id);
|
460 |
|
|
}
|
461 |
|
|
|
462 |
|
|
/* Called from copy_body_id via walk_tree. DATA is really an
|
463 |
|
|
`copy_body_data *'. */
|
464 |
|
|
|
465 |
|
|
tree
|
466 |
|
|
copy_body_r (tree *tp, int *walk_subtrees, void *data)
|
467 |
|
|
{
|
468 |
|
|
copy_body_data *id = (copy_body_data *) data;
|
469 |
|
|
tree fn = id->src_fn;
|
470 |
|
|
tree new_block;
|
471 |
|
|
|
472 |
|
|
/* Begin by recognizing trees that we'll completely rewrite for the
|
473 |
|
|
inlining context. Our output for these trees is completely
|
474 |
|
|
different from out input (e.g. RETURN_EXPR is deleted, and morphs
|
475 |
|
|
into an edge). Further down, we'll handle trees that get
|
476 |
|
|
duplicated and/or tweaked. */
|
477 |
|
|
|
478 |
|
|
/* When requested, RETURN_EXPRs should be transformed to just the
|
479 |
|
|
contained MODIFY_EXPR. The branch semantics of the return will
|
480 |
|
|
be handled elsewhere by manipulating the CFG rather than a statement. */
|
481 |
|
|
if (TREE_CODE (*tp) == RETURN_EXPR && id->transform_return_to_modify)
|
482 |
|
|
{
|
483 |
|
|
tree assignment = TREE_OPERAND (*tp, 0);
|
484 |
|
|
|
485 |
|
|
/* If we're returning something, just turn that into an
|
486 |
|
|
assignment into the equivalent of the original RESULT_DECL.
|
487 |
|
|
If the "assignment" is just the result decl, the result
|
488 |
|
|
decl has already been set (e.g. a recent "foo (&result_decl,
|
489 |
|
|
...)"); just toss the entire RETURN_EXPR. */
|
490 |
|
|
if (assignment && TREE_CODE (assignment) == MODIFY_EXPR)
|
491 |
|
|
{
|
492 |
|
|
/* Replace the RETURN_EXPR with (a copy of) the
|
493 |
|
|
MODIFY_EXPR hanging underneath. */
|
494 |
|
|
*tp = copy_node (assignment);
|
495 |
|
|
}
|
496 |
|
|
else /* Else the RETURN_EXPR returns no value. */
|
497 |
|
|
{
|
498 |
|
|
*tp = NULL;
|
499 |
|
|
return (tree) (void *)1;
|
500 |
|
|
}
|
501 |
|
|
}
|
502 |
|
|
|
503 |
|
|
/* Local variables and labels need to be replaced by equivalent
|
504 |
|
|
variables. We don't want to copy static variables; there's only
|
505 |
|
|
one of those, no matter how many times we inline the containing
|
506 |
|
|
function. Similarly for globals from an outer function. */
|
507 |
|
|
else if (lang_hooks.tree_inlining.auto_var_in_fn_p (*tp, fn))
|
508 |
|
|
{
|
509 |
|
|
tree new_decl;
|
510 |
|
|
|
511 |
|
|
/* Remap the declaration. */
|
512 |
|
|
new_decl = remap_decl (*tp, id);
|
513 |
|
|
gcc_assert (new_decl);
|
514 |
|
|
/* Replace this variable with the copy. */
|
515 |
|
|
STRIP_TYPE_NOPS (new_decl);
|
516 |
|
|
*tp = new_decl;
|
517 |
|
|
*walk_subtrees = 0;
|
518 |
|
|
}
|
519 |
|
|
else if (TREE_CODE (*tp) == STATEMENT_LIST)
|
520 |
|
|
copy_statement_list (tp);
|
521 |
|
|
else if (TREE_CODE (*tp) == SAVE_EXPR)
|
522 |
|
|
remap_save_expr (tp, id->decl_map, walk_subtrees);
|
523 |
|
|
else if (TREE_CODE (*tp) == LABEL_DECL
|
524 |
|
|
&& (! DECL_CONTEXT (*tp)
|
525 |
|
|
|| decl_function_context (*tp) == id->src_fn))
|
526 |
|
|
/* These may need to be remapped for EH handling. */
|
527 |
|
|
*tp = remap_decl (*tp, id);
|
528 |
|
|
else if (TREE_CODE (*tp) == BIND_EXPR)
|
529 |
|
|
copy_bind_expr (tp, walk_subtrees, id);
|
530 |
|
|
/* Types may need remapping as well. */
|
531 |
|
|
else if (TYPE_P (*tp))
|
532 |
|
|
*tp = remap_type (*tp, id);
|
533 |
|
|
|
534 |
|
|
/* If this is a constant, we have to copy the node iff the type will be
|
535 |
|
|
remapped. copy_tree_r will not copy a constant. */
|
536 |
|
|
else if (CONSTANT_CLASS_P (*tp))
|
537 |
|
|
{
|
538 |
|
|
tree new_type = remap_type (TREE_TYPE (*tp), id);
|
539 |
|
|
|
540 |
|
|
if (new_type == TREE_TYPE (*tp))
|
541 |
|
|
*walk_subtrees = 0;
|
542 |
|
|
|
543 |
|
|
else if (TREE_CODE (*tp) == INTEGER_CST)
|
544 |
|
|
*tp = build_int_cst_wide (new_type, TREE_INT_CST_LOW (*tp),
|
545 |
|
|
TREE_INT_CST_HIGH (*tp));
|
546 |
|
|
else
|
547 |
|
|
{
|
548 |
|
|
*tp = copy_node (*tp);
|
549 |
|
|
TREE_TYPE (*tp) = new_type;
|
550 |
|
|
}
|
551 |
|
|
}
|
552 |
|
|
|
553 |
|
|
/* Otherwise, just copy the node. Note that copy_tree_r already
|
554 |
|
|
knows not to copy VAR_DECLs, etc., so this is safe. */
|
555 |
|
|
else
|
556 |
|
|
{
|
557 |
|
|
/* Here we handle trees that are not completely rewritten.
|
558 |
|
|
First we detect some inlining-induced bogosities for
|
559 |
|
|
discarding. */
|
560 |
|
|
if (TREE_CODE (*tp) == MODIFY_EXPR
|
561 |
|
|
&& TREE_OPERAND (*tp, 0) == TREE_OPERAND (*tp, 1)
|
562 |
|
|
&& (lang_hooks.tree_inlining.auto_var_in_fn_p
|
563 |
|
|
(TREE_OPERAND (*tp, 0), fn)))
|
564 |
|
|
{
|
565 |
|
|
/* Some assignments VAR = VAR; don't generate any rtl code
|
566 |
|
|
and thus don't count as variable modification. Avoid
|
567 |
|
|
keeping bogosities like 0 = 0. */
|
568 |
|
|
tree decl = TREE_OPERAND (*tp, 0), value;
|
569 |
|
|
splay_tree_node n;
|
570 |
|
|
|
571 |
|
|
n = splay_tree_lookup (id->decl_map, (splay_tree_key) decl);
|
572 |
|
|
if (n)
|
573 |
|
|
{
|
574 |
|
|
value = (tree) n->value;
|
575 |
|
|
STRIP_TYPE_NOPS (value);
|
576 |
|
|
if (TREE_CONSTANT (value) || TREE_READONLY_DECL_P (value))
|
577 |
|
|
{
|
578 |
|
|
*tp = build_empty_stmt ();
|
579 |
|
|
return copy_body_r (tp, walk_subtrees, data);
|
580 |
|
|
}
|
581 |
|
|
}
|
582 |
|
|
}
|
583 |
|
|
else if (TREE_CODE (*tp) == INDIRECT_REF)
|
584 |
|
|
{
|
585 |
|
|
/* Get rid of *& from inline substitutions that can happen when a
|
586 |
|
|
pointer argument is an ADDR_EXPR. */
|
587 |
|
|
tree decl = TREE_OPERAND (*tp, 0);
|
588 |
|
|
splay_tree_node n;
|
589 |
|
|
|
590 |
|
|
n = splay_tree_lookup (id->decl_map, (splay_tree_key) decl);
|
591 |
|
|
if (n)
|
592 |
|
|
{
|
593 |
|
|
tree new;
|
594 |
|
|
tree old;
|
595 |
|
|
/* If we happen to get an ADDR_EXPR in n->value, strip
|
596 |
|
|
it manually here as we'll eventually get ADDR_EXPRs
|
597 |
|
|
which lie about their types pointed to. In this case
|
598 |
|
|
build_fold_indirect_ref wouldn't strip the INDIRECT_REF,
|
599 |
|
|
but we absolutely rely on that. As fold_indirect_ref
|
600 |
|
|
does other useful transformations, try that first, though. */
|
601 |
|
|
tree type = TREE_TYPE (TREE_TYPE ((tree)n->value));
|
602 |
|
|
new = unshare_expr ((tree)n->value);
|
603 |
|
|
old = *tp;
|
604 |
|
|
*tp = fold_indirect_ref_1 (type, new);
|
605 |
|
|
if (! *tp)
|
606 |
|
|
{
|
607 |
|
|
if (TREE_CODE (new) == ADDR_EXPR)
|
608 |
|
|
*tp = TREE_OPERAND (new, 0);
|
609 |
|
|
else
|
610 |
|
|
{
|
611 |
|
|
*tp = build1 (INDIRECT_REF, type, new);
|
612 |
|
|
TREE_THIS_VOLATILE (*tp) = TREE_THIS_VOLATILE (old);
|
613 |
|
|
}
|
614 |
|
|
}
|
615 |
|
|
*walk_subtrees = 0;
|
616 |
|
|
return NULL;
|
617 |
|
|
}
|
618 |
|
|
}
|
619 |
|
|
|
620 |
|
|
/* Here is the "usual case". Copy this tree node, and then
|
621 |
|
|
tweak some special cases. */
|
622 |
|
|
copy_tree_r (tp, walk_subtrees, NULL);
|
623 |
|
|
|
624 |
|
|
/* If EXPR has block defined, map it to newly constructed block.
|
625 |
|
|
When inlining we want EXPRs without block appear in the block
|
626 |
|
|
of function call. */
|
627 |
|
|
if (IS_EXPR_CODE_CLASS (TREE_CODE_CLASS (TREE_CODE (*tp))))
|
628 |
|
|
{
|
629 |
|
|
new_block = id->block;
|
630 |
|
|
if (TREE_BLOCK (*tp))
|
631 |
|
|
{
|
632 |
|
|
splay_tree_node n;
|
633 |
|
|
n = splay_tree_lookup (id->decl_map,
|
634 |
|
|
(splay_tree_key) TREE_BLOCK (*tp));
|
635 |
|
|
gcc_assert (n);
|
636 |
|
|
new_block = (tree) n->value;
|
637 |
|
|
}
|
638 |
|
|
TREE_BLOCK (*tp) = new_block;
|
639 |
|
|
}
|
640 |
|
|
|
641 |
|
|
if (TREE_CODE (*tp) == RESX_EXPR && id->eh_region_offset)
|
642 |
|
|
TREE_OPERAND (*tp, 0) =
|
643 |
|
|
build_int_cst
|
644 |
|
|
(NULL_TREE,
|
645 |
|
|
id->eh_region_offset + TREE_INT_CST_LOW (TREE_OPERAND (*tp, 0)));
|
646 |
|
|
|
647 |
|
|
if (TREE_CODE (*tp) != OMP_CLAUSE)
|
648 |
|
|
TREE_TYPE (*tp) = remap_type (TREE_TYPE (*tp), id);
|
649 |
|
|
|
650 |
|
|
/* The copied TARGET_EXPR has never been expanded, even if the
|
651 |
|
|
original node was expanded already. */
|
652 |
|
|
if (TREE_CODE (*tp) == TARGET_EXPR && TREE_OPERAND (*tp, 3))
|
653 |
|
|
{
|
654 |
|
|
TREE_OPERAND (*tp, 1) = TREE_OPERAND (*tp, 3);
|
655 |
|
|
TREE_OPERAND (*tp, 3) = NULL_TREE;
|
656 |
|
|
}
|
657 |
|
|
|
658 |
|
|
/* Variable substitution need not be simple. In particular, the
|
659 |
|
|
INDIRECT_REF substitution above. Make sure that TREE_CONSTANT
|
660 |
|
|
and friends are up-to-date. */
|
661 |
|
|
else if (TREE_CODE (*tp) == ADDR_EXPR)
|
662 |
|
|
{
|
663 |
|
|
walk_tree (&TREE_OPERAND (*tp, 0), copy_body_r, id, NULL);
|
664 |
|
|
/* Handle the case where we substituted an INDIRECT_REF
|
665 |
|
|
into the operand of the ADDR_EXPR. */
|
666 |
|
|
if (TREE_CODE (TREE_OPERAND (*tp, 0)) == INDIRECT_REF)
|
667 |
|
|
*tp = TREE_OPERAND (TREE_OPERAND (*tp, 0), 0);
|
668 |
|
|
else
|
669 |
|
|
recompute_tree_invariant_for_addr_expr (*tp);
|
670 |
|
|
*walk_subtrees = 0;
|
671 |
|
|
}
|
672 |
|
|
}
|
673 |
|
|
|
674 |
|
|
/* Keep iterating. */
|
675 |
|
|
return NULL_TREE;
|
676 |
|
|
}
|
677 |
|
|
|
678 |
|
|
/* Copy basic block, scale profile accordingly. Edges will be taken care of
|
679 |
|
|
later */
|
680 |
|
|
|
681 |
|
|
static basic_block
|
682 |
|
|
copy_bb (copy_body_data *id, basic_block bb, int frequency_scale, int count_scale)
|
683 |
|
|
{
|
684 |
|
|
block_stmt_iterator bsi, copy_bsi;
|
685 |
|
|
basic_block copy_basic_block;
|
686 |
|
|
|
687 |
|
|
/* create_basic_block() will append every new block to
|
688 |
|
|
basic_block_info automatically. */
|
689 |
|
|
copy_basic_block = create_basic_block (NULL, (void *) 0,
|
690 |
|
|
(basic_block) bb->prev_bb->aux);
|
691 |
|
|
copy_basic_block->count = bb->count * count_scale / REG_BR_PROB_BASE;
|
692 |
|
|
copy_basic_block->frequency = (bb->frequency
|
693 |
|
|
* frequency_scale / REG_BR_PROB_BASE);
|
694 |
|
|
copy_bsi = bsi_start (copy_basic_block);
|
695 |
|
|
|
696 |
|
|
for (bsi = bsi_start (bb);
|
697 |
|
|
!bsi_end_p (bsi); bsi_next (&bsi))
|
698 |
|
|
{
|
699 |
|
|
tree stmt = bsi_stmt (bsi);
|
700 |
|
|
tree orig_stmt = stmt;
|
701 |
|
|
|
702 |
|
|
walk_tree (&stmt, copy_body_r, id, NULL);
|
703 |
|
|
|
704 |
|
|
/* RETURN_EXPR might be removed,
|
705 |
|
|
this is signalled by making stmt pointer NULL. */
|
706 |
|
|
if (stmt)
|
707 |
|
|
{
|
708 |
|
|
tree call, decl;
|
709 |
|
|
|
710 |
|
|
/* With return slot optimization we can end up with
|
711 |
|
|
non-gimple (foo *)&this->m, fix that here. */
|
712 |
|
|
if (TREE_CODE (stmt) == MODIFY_EXPR
|
713 |
|
|
&& TREE_CODE (TREE_OPERAND (stmt, 1)) == NOP_EXPR
|
714 |
|
|
&& !is_gimple_val (TREE_OPERAND (TREE_OPERAND (stmt, 1), 0)))
|
715 |
|
|
gimplify_stmt (&stmt);
|
716 |
|
|
|
717 |
|
|
bsi_insert_after (©_bsi, stmt, BSI_NEW_STMT);
|
718 |
|
|
call = get_call_expr_in (stmt);
|
719 |
|
|
/* We're duplicating a CALL_EXPR. Find any corresponding
|
720 |
|
|
callgraph edges and update or duplicate them. */
|
721 |
|
|
if (call && (decl = get_callee_fndecl (call)))
|
722 |
|
|
{
|
723 |
|
|
struct cgraph_node *node;
|
724 |
|
|
struct cgraph_edge *edge;
|
725 |
|
|
|
726 |
|
|
switch (id->transform_call_graph_edges)
|
727 |
|
|
{
|
728 |
|
|
case CB_CGE_DUPLICATE:
|
729 |
|
|
edge = cgraph_edge (id->src_node, orig_stmt);
|
730 |
|
|
if (edge)
|
731 |
|
|
cgraph_clone_edge (edge, id->dst_node, stmt,
|
732 |
|
|
REG_BR_PROB_BASE, 1, true);
|
733 |
|
|
break;
|
734 |
|
|
|
735 |
|
|
case CB_CGE_MOVE_CLONES:
|
736 |
|
|
for (node = id->dst_node->next_clone;
|
737 |
|
|
node;
|
738 |
|
|
node = node->next_clone)
|
739 |
|
|
{
|
740 |
|
|
edge = cgraph_edge (node, orig_stmt);
|
741 |
|
|
gcc_assert (edge);
|
742 |
|
|
cgraph_set_call_stmt (edge, stmt);
|
743 |
|
|
}
|
744 |
|
|
/* FALLTHRU */
|
745 |
|
|
|
746 |
|
|
case CB_CGE_MOVE:
|
747 |
|
|
edge = cgraph_edge (id->dst_node, orig_stmt);
|
748 |
|
|
if (edge)
|
749 |
|
|
cgraph_set_call_stmt (edge, stmt);
|
750 |
|
|
break;
|
751 |
|
|
|
752 |
|
|
default:
|
753 |
|
|
gcc_unreachable ();
|
754 |
|
|
}
|
755 |
|
|
}
|
756 |
|
|
/* If you think we can abort here, you are wrong.
|
757 |
|
|
There is no region 0 in tree land. */
|
758 |
|
|
gcc_assert (lookup_stmt_eh_region_fn (id->src_cfun, orig_stmt)
|
759 |
|
|
!= 0);
|
760 |
|
|
|
761 |
|
|
if (tree_could_throw_p (stmt))
|
762 |
|
|
{
|
763 |
|
|
int region = lookup_stmt_eh_region_fn (id->src_cfun, orig_stmt);
|
764 |
|
|
/* Add an entry for the copied tree in the EH hashtable.
|
765 |
|
|
When cloning or versioning, use the hashtable in
|
766 |
|
|
cfun, and just copy the EH number. When inlining, use the
|
767 |
|
|
hashtable in the caller, and adjust the region number. */
|
768 |
|
|
if (region > 0)
|
769 |
|
|
add_stmt_to_eh_region (stmt, region + id->eh_region_offset);
|
770 |
|
|
|
771 |
|
|
/* If this tree doesn't have a region associated with it,
|
772 |
|
|
and there is a "current region,"
|
773 |
|
|
then associate this tree with the current region
|
774 |
|
|
and add edges associated with this region. */
|
775 |
|
|
if ((lookup_stmt_eh_region_fn (id->src_cfun,
|
776 |
|
|
orig_stmt) <= 0
|
777 |
|
|
&& id->eh_region > 0)
|
778 |
|
|
&& tree_could_throw_p (stmt))
|
779 |
|
|
add_stmt_to_eh_region (stmt, id->eh_region);
|
780 |
|
|
}
|
781 |
|
|
}
|
782 |
|
|
}
|
783 |
|
|
return copy_basic_block;
|
784 |
|
|
}
|
785 |
|
|
|
786 |
|
|
/* Copy edges from BB into its copy constructed earlier, scale profile
|
787 |
|
|
accordingly. Edges will be taken care of later. Assume aux
|
788 |
|
|
pointers to point to the copies of each BB. */
|
789 |
|
|
static void
|
790 |
|
|
copy_edges_for_bb (basic_block bb, int count_scale)
|
791 |
|
|
{
|
792 |
|
|
basic_block new_bb = (basic_block) bb->aux;
|
793 |
|
|
edge_iterator ei;
|
794 |
|
|
edge old_edge;
|
795 |
|
|
block_stmt_iterator bsi;
|
796 |
|
|
int flags;
|
797 |
|
|
|
798 |
|
|
/* Use the indices from the original blocks to create edges for the
|
799 |
|
|
new ones. */
|
800 |
|
|
FOR_EACH_EDGE (old_edge, ei, bb->succs)
|
801 |
|
|
if (!(old_edge->flags & EDGE_EH))
|
802 |
|
|
{
|
803 |
|
|
edge new;
|
804 |
|
|
|
805 |
|
|
flags = old_edge->flags;
|
806 |
|
|
|
807 |
|
|
/* Return edges do get a FALLTHRU flag when the get inlined. */
|
808 |
|
|
if (old_edge->dest->index == EXIT_BLOCK && !old_edge->flags
|
809 |
|
|
&& old_edge->dest->aux != EXIT_BLOCK_PTR)
|
810 |
|
|
flags |= EDGE_FALLTHRU;
|
811 |
|
|
new = make_edge (new_bb, (basic_block) old_edge->dest->aux, flags);
|
812 |
|
|
new->count = old_edge->count * count_scale / REG_BR_PROB_BASE;
|
813 |
|
|
new->probability = old_edge->probability;
|
814 |
|
|
}
|
815 |
|
|
|
816 |
|
|
if (bb->index == ENTRY_BLOCK || bb->index == EXIT_BLOCK)
|
817 |
|
|
return;
|
818 |
|
|
|
819 |
|
|
for (bsi = bsi_start (new_bb); !bsi_end_p (bsi);)
|
820 |
|
|
{
|
821 |
|
|
tree copy_stmt;
|
822 |
|
|
|
823 |
|
|
copy_stmt = bsi_stmt (bsi);
|
824 |
|
|
update_stmt (copy_stmt);
|
825 |
|
|
/* Do this before the possible split_block. */
|
826 |
|
|
bsi_next (&bsi);
|
827 |
|
|
|
828 |
|
|
/* If this tree could throw an exception, there are two
|
829 |
|
|
cases where we need to add abnormal edge(s): the
|
830 |
|
|
tree wasn't in a region and there is a "current
|
831 |
|
|
region" in the caller; or the original tree had
|
832 |
|
|
EH edges. In both cases split the block after the tree,
|
833 |
|
|
and add abnormal edge(s) as needed; we need both
|
834 |
|
|
those from the callee and the caller.
|
835 |
|
|
We check whether the copy can throw, because the const
|
836 |
|
|
propagation can change an INDIRECT_REF which throws
|
837 |
|
|
into a COMPONENT_REF which doesn't. If the copy
|
838 |
|
|
can throw, the original could also throw. */
|
839 |
|
|
|
840 |
|
|
if (tree_can_throw_internal (copy_stmt))
|
841 |
|
|
{
|
842 |
|
|
if (!bsi_end_p (bsi))
|
843 |
|
|
/* Note that bb's predecessor edges aren't necessarily
|
844 |
|
|
right at this point; split_block doesn't care. */
|
845 |
|
|
{
|
846 |
|
|
edge e = split_block (new_bb, copy_stmt);
|
847 |
|
|
new_bb = e->dest;
|
848 |
|
|
bsi = bsi_start (new_bb);
|
849 |
|
|
}
|
850 |
|
|
|
851 |
|
|
make_eh_edges (copy_stmt);
|
852 |
|
|
}
|
853 |
|
|
}
|
854 |
|
|
}
|
855 |
|
|
|
856 |
|
|
/* Wrapper for remap_decl so it can be used as a callback. */
|
857 |
|
|
static tree
|
858 |
|
|
remap_decl_1 (tree decl, void *data)
|
859 |
|
|
{
|
860 |
|
|
return remap_decl (decl, (copy_body_data *) data);
|
861 |
|
|
}
|
862 |
|
|
|
863 |
|
|
/* Make a copy of the body of FN so that it can be inserted inline in
|
864 |
|
|
another function. Walks FN via CFG, returns new fndecl. */
|
865 |
|
|
|
866 |
|
|
static tree
|
867 |
|
|
copy_cfg_body (copy_body_data * id, gcov_type count, int frequency,
|
868 |
|
|
basic_block entry_block_map, basic_block exit_block_map)
|
869 |
|
|
{
|
870 |
|
|
tree callee_fndecl = id->src_fn;
|
871 |
|
|
/* Original cfun for the callee, doesn't change. */
|
872 |
|
|
struct function *src_cfun = DECL_STRUCT_FUNCTION (callee_fndecl);
|
873 |
|
|
/* Copy, built by this function. */
|
874 |
|
|
struct function *new_cfun;
|
875 |
|
|
/* Place to copy from; when a copy of the function was saved off earlier,
|
876 |
|
|
use that instead of the main copy. */
|
877 |
|
|
struct function *cfun_to_copy =
|
878 |
|
|
(struct function *) ggc_alloc_cleared (sizeof (struct function));
|
879 |
|
|
basic_block bb;
|
880 |
|
|
tree new_fndecl = NULL;
|
881 |
|
|
int count_scale, frequency_scale;
|
882 |
|
|
|
883 |
|
|
if (ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->count)
|
884 |
|
|
count_scale = (REG_BR_PROB_BASE * count
|
885 |
|
|
/ ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->count);
|
886 |
|
|
else
|
887 |
|
|
count_scale = 1;
|
888 |
|
|
|
889 |
|
|
if (ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->frequency)
|
890 |
|
|
frequency_scale = (REG_BR_PROB_BASE * frequency
|
891 |
|
|
/
|
892 |
|
|
ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->frequency);
|
893 |
|
|
else
|
894 |
|
|
frequency_scale = count_scale;
|
895 |
|
|
|
896 |
|
|
/* Register specific tree functions. */
|
897 |
|
|
tree_register_cfg_hooks ();
|
898 |
|
|
|
899 |
|
|
/* Must have a CFG here at this point. */
|
900 |
|
|
gcc_assert (ENTRY_BLOCK_PTR_FOR_FUNCTION
|
901 |
|
|
(DECL_STRUCT_FUNCTION (callee_fndecl)));
|
902 |
|
|
|
903 |
|
|
*cfun_to_copy = *DECL_STRUCT_FUNCTION (callee_fndecl);
|
904 |
|
|
|
905 |
|
|
id->src_cfun = cfun_to_copy;
|
906 |
|
|
|
907 |
|
|
/* If requested, create new basic_block_info and label_to_block_maps.
|
908 |
|
|
Otherwise, insert our new blocks and labels into the existing cfg. */
|
909 |
|
|
if (id->transform_new_cfg)
|
910 |
|
|
{
|
911 |
|
|
new_cfun =
|
912 |
|
|
(struct function *) ggc_alloc_cleared (sizeof (struct function));
|
913 |
|
|
*new_cfun = *DECL_STRUCT_FUNCTION (callee_fndecl);
|
914 |
|
|
new_cfun->cfg = NULL;
|
915 |
|
|
new_cfun->decl = new_fndecl = copy_node (callee_fndecl);
|
916 |
|
|
new_cfun->ib_boundaries_block = NULL;
|
917 |
|
|
DECL_STRUCT_FUNCTION (new_fndecl) = new_cfun;
|
918 |
|
|
push_cfun (new_cfun);
|
919 |
|
|
init_empty_tree_cfg ();
|
920 |
|
|
|
921 |
|
|
ENTRY_BLOCK_PTR->count =
|
922 |
|
|
(ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->count * count_scale /
|
923 |
|
|
REG_BR_PROB_BASE);
|
924 |
|
|
ENTRY_BLOCK_PTR->frequency =
|
925 |
|
|
(ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->frequency *
|
926 |
|
|
frequency_scale / REG_BR_PROB_BASE);
|
927 |
|
|
EXIT_BLOCK_PTR->count =
|
928 |
|
|
(EXIT_BLOCK_PTR_FOR_FUNCTION (src_cfun)->count * count_scale /
|
929 |
|
|
REG_BR_PROB_BASE);
|
930 |
|
|
EXIT_BLOCK_PTR->frequency =
|
931 |
|
|
(EXIT_BLOCK_PTR_FOR_FUNCTION (src_cfun)->frequency *
|
932 |
|
|
frequency_scale / REG_BR_PROB_BASE);
|
933 |
|
|
|
934 |
|
|
entry_block_map = ENTRY_BLOCK_PTR;
|
935 |
|
|
exit_block_map = EXIT_BLOCK_PTR;
|
936 |
|
|
}
|
937 |
|
|
|
938 |
|
|
ENTRY_BLOCK_PTR_FOR_FUNCTION (cfun_to_copy)->aux = entry_block_map;
|
939 |
|
|
EXIT_BLOCK_PTR_FOR_FUNCTION (cfun_to_copy)->aux = exit_block_map;
|
940 |
|
|
|
941 |
|
|
/* Duplicate any exception-handling regions. */
|
942 |
|
|
if (cfun->eh)
|
943 |
|
|
{
|
944 |
|
|
if (id->transform_new_cfg)
|
945 |
|
|
init_eh_for_function ();
|
946 |
|
|
id->eh_region_offset
|
947 |
|
|
= duplicate_eh_regions (cfun_to_copy, remap_decl_1, id,
|
948 |
|
|
0, id->eh_region);
|
949 |
|
|
}
|
950 |
|
|
/* Use aux pointers to map the original blocks to copy. */
|
951 |
|
|
FOR_EACH_BB_FN (bb, cfun_to_copy)
|
952 |
|
|
bb->aux = copy_bb (id, bb, frequency_scale, count_scale);
|
953 |
|
|
/* Now that we've duplicated the blocks, duplicate their edges. */
|
954 |
|
|
FOR_ALL_BB_FN (bb, cfun_to_copy)
|
955 |
|
|
copy_edges_for_bb (bb, count_scale);
|
956 |
|
|
FOR_ALL_BB_FN (bb, cfun_to_copy)
|
957 |
|
|
bb->aux = NULL;
|
958 |
|
|
|
959 |
|
|
if (id->transform_new_cfg)
|
960 |
|
|
pop_cfun ();
|
961 |
|
|
|
962 |
|
|
return new_fndecl;
|
963 |
|
|
}
|
964 |
|
|
|
965 |
|
|
/* Make a copy of the body of FN so that it can be inserted inline in
|
966 |
|
|
another function. */
|
967 |
|
|
|
968 |
|
|
static tree
|
969 |
|
|
copy_generic_body (copy_body_data *id)
|
970 |
|
|
{
|
971 |
|
|
tree body;
|
972 |
|
|
tree fndecl = id->src_fn;
|
973 |
|
|
|
974 |
|
|
body = DECL_SAVED_TREE (fndecl);
|
975 |
|
|
walk_tree (&body, copy_body_r, id, NULL);
|
976 |
|
|
|
977 |
|
|
return body;
|
978 |
|
|
}
|
979 |
|
|
|
980 |
|
|
static tree
|
981 |
|
|
copy_body (copy_body_data *id, gcov_type count, int frequency,
|
982 |
|
|
basic_block entry_block_map, basic_block exit_block_map)
|
983 |
|
|
{
|
984 |
|
|
tree fndecl = id->src_fn;
|
985 |
|
|
tree body;
|
986 |
|
|
|
987 |
|
|
/* If this body has a CFG, walk CFG and copy. */
|
988 |
|
|
gcc_assert (ENTRY_BLOCK_PTR_FOR_FUNCTION (DECL_STRUCT_FUNCTION (fndecl)));
|
989 |
|
|
body = copy_cfg_body (id, count, frequency, entry_block_map, exit_block_map);
|
990 |
|
|
|
991 |
|
|
return body;
|
992 |
|
|
}
|
993 |
|
|
|
994 |
|
|
/* Return true if VALUE is an ADDR_EXPR of an automatic variable
|
995 |
|
|
defined in function FN, or of a data member thereof. */
|
996 |
|
|
|
997 |
|
|
static bool
|
998 |
|
|
self_inlining_addr_expr (tree value, tree fn)
|
999 |
|
|
{
|
1000 |
|
|
tree var;
|
1001 |
|
|
|
1002 |
|
|
if (TREE_CODE (value) != ADDR_EXPR)
|
1003 |
|
|
return false;
|
1004 |
|
|
|
1005 |
|
|
var = get_base_address (TREE_OPERAND (value, 0));
|
1006 |
|
|
|
1007 |
|
|
return var && lang_hooks.tree_inlining.auto_var_in_fn_p (var, fn);
|
1008 |
|
|
}
|
1009 |
|
|
|
1010 |
|
|
static void
|
1011 |
|
|
setup_one_parameter (copy_body_data *id, tree p, tree value, tree fn,
|
1012 |
|
|
basic_block bb, tree *vars)
|
1013 |
|
|
{
|
1014 |
|
|
tree init_stmt;
|
1015 |
|
|
tree var;
|
1016 |
|
|
tree var_sub;
|
1017 |
|
|
|
1018 |
|
|
/* If the parameter is never assigned to, we may not need to
|
1019 |
|
|
create a new variable here at all. Instead, we may be able
|
1020 |
|
|
to just use the argument value. */
|
1021 |
|
|
if (TREE_READONLY (p)
|
1022 |
|
|
&& !TREE_ADDRESSABLE (p)
|
1023 |
|
|
&& value && !TREE_SIDE_EFFECTS (value))
|
1024 |
|
|
{
|
1025 |
|
|
/* We may produce non-gimple trees by adding NOPs or introduce
|
1026 |
|
|
invalid sharing when operand is not really constant.
|
1027 |
|
|
It is not big deal to prohibit constant propagation here as
|
1028 |
|
|
we will constant propagate in DOM1 pass anyway. */
|
1029 |
|
|
if (is_gimple_min_invariant (value)
|
1030 |
|
|
&& lang_hooks.types_compatible_p (TREE_TYPE (value), TREE_TYPE (p))
|
1031 |
|
|
/* We have to be very careful about ADDR_EXPR. Make sure
|
1032 |
|
|
the base variable isn't a local variable of the inlined
|
1033 |
|
|
function, e.g., when doing recursive inlining, direct or
|
1034 |
|
|
mutually-recursive or whatever, which is why we don't
|
1035 |
|
|
just test whether fn == current_function_decl. */
|
1036 |
|
|
&& ! self_inlining_addr_expr (value, fn))
|
1037 |
|
|
{
|
1038 |
|
|
insert_decl_map (id, p, value);
|
1039 |
|
|
return;
|
1040 |
|
|
}
|
1041 |
|
|
}
|
1042 |
|
|
|
1043 |
|
|
/* Make an equivalent VAR_DECL. Note that we must NOT remap the type
|
1044 |
|
|
here since the type of this decl must be visible to the calling
|
1045 |
|
|
function. */
|
1046 |
|
|
var = copy_decl_to_var (p, id);
|
1047 |
|
|
|
1048 |
|
|
/* See if the frontend wants to pass this by invisible reference. If
|
1049 |
|
|
so, our new VAR_DECL will have REFERENCE_TYPE, and we need to
|
1050 |
|
|
replace uses of the PARM_DECL with dereferences. */
|
1051 |
|
|
if (TREE_TYPE (var) != TREE_TYPE (p)
|
1052 |
|
|
&& POINTER_TYPE_P (TREE_TYPE (var))
|
1053 |
|
|
&& TREE_TYPE (TREE_TYPE (var)) == TREE_TYPE (p))
|
1054 |
|
|
{
|
1055 |
|
|
insert_decl_map (id, var, var);
|
1056 |
|
|
var_sub = build_fold_indirect_ref (var);
|
1057 |
|
|
}
|
1058 |
|
|
else
|
1059 |
|
|
var_sub = var;
|
1060 |
|
|
|
1061 |
|
|
/* Register the VAR_DECL as the equivalent for the PARM_DECL;
|
1062 |
|
|
that way, when the PARM_DECL is encountered, it will be
|
1063 |
|
|
automatically replaced by the VAR_DECL. */
|
1064 |
|
|
insert_decl_map (id, p, var_sub);
|
1065 |
|
|
|
1066 |
|
|
/* Declare this new variable. */
|
1067 |
|
|
TREE_CHAIN (var) = *vars;
|
1068 |
|
|
*vars = var;
|
1069 |
|
|
|
1070 |
|
|
/* Make gimplifier happy about this variable. */
|
1071 |
|
|
DECL_SEEN_IN_BIND_EXPR_P (var) = 1;
|
1072 |
|
|
|
1073 |
|
|
/* Even if P was TREE_READONLY, the new VAR should not be.
|
1074 |
|
|
In the original code, we would have constructed a
|
1075 |
|
|
temporary, and then the function body would have never
|
1076 |
|
|
changed the value of P. However, now, we will be
|
1077 |
|
|
constructing VAR directly. The constructor body may
|
1078 |
|
|
change its value multiple times as it is being
|
1079 |
|
|
constructed. Therefore, it must not be TREE_READONLY;
|
1080 |
|
|
the back-end assumes that TREE_READONLY variable is
|
1081 |
|
|
assigned to only once. */
|
1082 |
|
|
if (TYPE_NEEDS_CONSTRUCTING (TREE_TYPE (p)))
|
1083 |
|
|
TREE_READONLY (var) = 0;
|
1084 |
|
|
|
1085 |
|
|
/* Initialize this VAR_DECL from the equivalent argument. Convert
|
1086 |
|
|
the argument to the proper type in case it was promoted. */
|
1087 |
|
|
if (value)
|
1088 |
|
|
{
|
1089 |
|
|
tree rhs = fold_convert (TREE_TYPE (var), value);
|
1090 |
|
|
block_stmt_iterator bsi = bsi_last (bb);
|
1091 |
|
|
|
1092 |
|
|
if (rhs == error_mark_node)
|
1093 |
|
|
return;
|
1094 |
|
|
|
1095 |
|
|
STRIP_USELESS_TYPE_CONVERSION (rhs);
|
1096 |
|
|
|
1097 |
|
|
/* We want to use MODIFY_EXPR, not INIT_EXPR here so that we
|
1098 |
|
|
keep our trees in gimple form. */
|
1099 |
|
|
init_stmt = build2 (MODIFY_EXPR, TREE_TYPE (var), var, rhs);
|
1100 |
|
|
|
1101 |
|
|
/* If we did not create a gimple value and we did not create a gimple
|
1102 |
|
|
cast of a gimple value, then we will need to gimplify INIT_STMTS
|
1103 |
|
|
at the end. Note that is_gimple_cast only checks the outer
|
1104 |
|
|
tree code, not its operand. Thus the explicit check that its
|
1105 |
|
|
operand is a gimple value. */
|
1106 |
|
|
if (!is_gimple_val (rhs)
|
1107 |
|
|
&& (!is_gimple_cast (rhs)
|
1108 |
|
|
|| !is_gimple_val (TREE_OPERAND (rhs, 0))))
|
1109 |
|
|
gimplify_stmt (&init_stmt);
|
1110 |
|
|
|
1111 |
|
|
/* If VAR represents a zero-sized variable, it's possible that the
|
1112 |
|
|
assignment statment may result in no gimple statements. */
|
1113 |
|
|
if (init_stmt)
|
1114 |
|
|
bsi_insert_after (&bsi, init_stmt, BSI_NEW_STMT);
|
1115 |
|
|
}
|
1116 |
|
|
}
|
1117 |
|
|
|
1118 |
|
|
/* Generate code to initialize the parameters of the function at the
|
1119 |
|
|
top of the stack in ID from the ARGS (presented as a TREE_LIST). */
|
1120 |
|
|
|
1121 |
|
|
static void
|
1122 |
|
|
initialize_inlined_parameters (copy_body_data *id, tree args, tree static_chain,
|
1123 |
|
|
tree fn, basic_block bb)
|
1124 |
|
|
{
|
1125 |
|
|
tree parms;
|
1126 |
|
|
tree a;
|
1127 |
|
|
tree p;
|
1128 |
|
|
tree vars = NULL_TREE;
|
1129 |
|
|
int argnum = 0;
|
1130 |
|
|
|
1131 |
|
|
/* Figure out what the parameters are. */
|
1132 |
|
|
parms = DECL_ARGUMENTS (fn);
|
1133 |
|
|
|
1134 |
|
|
/* Loop through the parameter declarations, replacing each with an
|
1135 |
|
|
equivalent VAR_DECL, appropriately initialized. */
|
1136 |
|
|
for (p = parms, a = args; p;
|
1137 |
|
|
a = a ? TREE_CHAIN (a) : a, p = TREE_CHAIN (p))
|
1138 |
|
|
{
|
1139 |
|
|
tree value;
|
1140 |
|
|
|
1141 |
|
|
++argnum;
|
1142 |
|
|
|
1143 |
|
|
/* Find the initializer. */
|
1144 |
|
|
value = lang_hooks.tree_inlining.convert_parm_for_inlining
|
1145 |
|
|
(p, a ? TREE_VALUE (a) : NULL_TREE, fn, argnum);
|
1146 |
|
|
|
1147 |
|
|
setup_one_parameter (id, p, value, fn, bb, &vars);
|
1148 |
|
|
}
|
1149 |
|
|
|
1150 |
|
|
/* Initialize the static chain. */
|
1151 |
|
|
p = DECL_STRUCT_FUNCTION (fn)->static_chain_decl;
|
1152 |
|
|
gcc_assert (fn != current_function_decl);
|
1153 |
|
|
if (p)
|
1154 |
|
|
{
|
1155 |
|
|
/* No static chain? Seems like a bug in tree-nested.c. */
|
1156 |
|
|
gcc_assert (static_chain);
|
1157 |
|
|
|
1158 |
|
|
setup_one_parameter (id, p, static_chain, fn, bb, &vars);
|
1159 |
|
|
}
|
1160 |
|
|
|
1161 |
|
|
declare_inline_vars (id->block, vars);
|
1162 |
|
|
}
|
1163 |
|
|
|
1164 |
|
|
/* Declare a return variable to replace the RESULT_DECL for the
|
1165 |
|
|
function we are calling. An appropriate DECL_STMT is returned.
|
1166 |
|
|
The USE_STMT is filled to contain a use of the declaration to
|
1167 |
|
|
indicate the return value of the function.
|
1168 |
|
|
|
1169 |
|
|
RETURN_SLOT_ADDR, if non-null, was a fake parameter that
|
1170 |
|
|
took the address of the result. MODIFY_DEST, if non-null, was the LHS of
|
1171 |
|
|
the MODIFY_EXPR to which this call is the RHS.
|
1172 |
|
|
|
1173 |
|
|
The return value is a (possibly null) value that is the result of the
|
1174 |
|
|
function as seen by the callee. *USE_P is a (possibly null) value that
|
1175 |
|
|
holds the result as seen by the caller. */
|
1176 |
|
|
|
1177 |
|
|
static tree
|
1178 |
|
|
declare_return_variable (copy_body_data *id, tree return_slot_addr,
|
1179 |
|
|
tree modify_dest, tree *use_p)
|
1180 |
|
|
{
|
1181 |
|
|
tree callee = id->src_fn;
|
1182 |
|
|
tree caller = id->dst_fn;
|
1183 |
|
|
tree result = DECL_RESULT (callee);
|
1184 |
|
|
tree callee_type = TREE_TYPE (result);
|
1185 |
|
|
tree caller_type = TREE_TYPE (TREE_TYPE (callee));
|
1186 |
|
|
tree var, use;
|
1187 |
|
|
|
1188 |
|
|
/* We don't need to do anything for functions that don't return
|
1189 |
|
|
anything. */
|
1190 |
|
|
if (!result || VOID_TYPE_P (callee_type))
|
1191 |
|
|
{
|
1192 |
|
|
*use_p = NULL_TREE;
|
1193 |
|
|
return NULL_TREE;
|
1194 |
|
|
}
|
1195 |
|
|
|
1196 |
|
|
/* If there was a return slot, then the return value is the
|
1197 |
|
|
dereferenced address of that object. */
|
1198 |
|
|
if (return_slot_addr)
|
1199 |
|
|
{
|
1200 |
|
|
/* The front end shouldn't have used both return_slot_addr and
|
1201 |
|
|
a modify expression. */
|
1202 |
|
|
gcc_assert (!modify_dest);
|
1203 |
|
|
if (DECL_BY_REFERENCE (result))
|
1204 |
|
|
var = return_slot_addr;
|
1205 |
|
|
else
|
1206 |
|
|
var = build_fold_indirect_ref (return_slot_addr);
|
1207 |
|
|
if (TREE_CODE (TREE_TYPE (result)) == COMPLEX_TYPE
|
1208 |
|
|
&& !DECL_COMPLEX_GIMPLE_REG_P (result)
|
1209 |
|
|
&& DECL_P (var))
|
1210 |
|
|
DECL_COMPLEX_GIMPLE_REG_P (var) = 0;
|
1211 |
|
|
use = NULL;
|
1212 |
|
|
goto done;
|
1213 |
|
|
}
|
1214 |
|
|
|
1215 |
|
|
/* All types requiring non-trivial constructors should have been handled. */
|
1216 |
|
|
gcc_assert (!TREE_ADDRESSABLE (callee_type));
|
1217 |
|
|
|
1218 |
|
|
/* Attempt to avoid creating a new temporary variable. */
|
1219 |
|
|
if (modify_dest)
|
1220 |
|
|
{
|
1221 |
|
|
bool use_it = false;
|
1222 |
|
|
|
1223 |
|
|
/* We can't use MODIFY_DEST if there's type promotion involved. */
|
1224 |
|
|
if (!lang_hooks.types_compatible_p (caller_type, callee_type))
|
1225 |
|
|
use_it = false;
|
1226 |
|
|
|
1227 |
|
|
/* ??? If we're assigning to a variable sized type, then we must
|
1228 |
|
|
reuse the destination variable, because we've no good way to
|
1229 |
|
|
create variable sized temporaries at this point. */
|
1230 |
|
|
else if (TREE_CODE (TYPE_SIZE_UNIT (caller_type)) != INTEGER_CST)
|
1231 |
|
|
use_it = true;
|
1232 |
|
|
|
1233 |
|
|
/* If the callee cannot possibly modify MODIFY_DEST, then we can
|
1234 |
|
|
reuse it as the result of the call directly. Don't do this if
|
1235 |
|
|
it would promote MODIFY_DEST to addressable. */
|
1236 |
|
|
else if (TREE_ADDRESSABLE (result))
|
1237 |
|
|
use_it = false;
|
1238 |
|
|
else
|
1239 |
|
|
{
|
1240 |
|
|
tree base_m = get_base_address (modify_dest);
|
1241 |
|
|
|
1242 |
|
|
/* If the base isn't a decl, then it's a pointer, and we don't
|
1243 |
|
|
know where that's going to go. */
|
1244 |
|
|
if (!DECL_P (base_m))
|
1245 |
|
|
use_it = false;
|
1246 |
|
|
else if (is_global_var (base_m))
|
1247 |
|
|
use_it = false;
|
1248 |
|
|
else if (TREE_CODE (TREE_TYPE (result)) == COMPLEX_TYPE
|
1249 |
|
|
&& !DECL_COMPLEX_GIMPLE_REG_P (result)
|
1250 |
|
|
&& DECL_COMPLEX_GIMPLE_REG_P (base_m))
|
1251 |
|
|
use_it = false;
|
1252 |
|
|
else if (!TREE_ADDRESSABLE (base_m))
|
1253 |
|
|
use_it = true;
|
1254 |
|
|
}
|
1255 |
|
|
|
1256 |
|
|
if (use_it)
|
1257 |
|
|
{
|
1258 |
|
|
var = modify_dest;
|
1259 |
|
|
use = NULL;
|
1260 |
|
|
goto done;
|
1261 |
|
|
}
|
1262 |
|
|
}
|
1263 |
|
|
|
1264 |
|
|
gcc_assert (TREE_CODE (TYPE_SIZE_UNIT (callee_type)) == INTEGER_CST);
|
1265 |
|
|
|
1266 |
|
|
var = copy_result_decl_to_var (result, id);
|
1267 |
|
|
|
1268 |
|
|
DECL_SEEN_IN_BIND_EXPR_P (var) = 1;
|
1269 |
|
|
DECL_STRUCT_FUNCTION (caller)->unexpanded_var_list
|
1270 |
|
|
= tree_cons (NULL_TREE, var,
|
1271 |
|
|
DECL_STRUCT_FUNCTION (caller)->unexpanded_var_list);
|
1272 |
|
|
|
1273 |
|
|
/* Do not have the rest of GCC warn about this variable as it should
|
1274 |
|
|
not be visible to the user. */
|
1275 |
|
|
TREE_NO_WARNING (var) = 1;
|
1276 |
|
|
|
1277 |
|
|
declare_inline_vars (id->block, var);
|
1278 |
|
|
|
1279 |
|
|
/* Build the use expr. If the return type of the function was
|
1280 |
|
|
promoted, convert it back to the expected type. */
|
1281 |
|
|
use = var;
|
1282 |
|
|
if (!lang_hooks.types_compatible_p (TREE_TYPE (var), caller_type))
|
1283 |
|
|
use = fold_convert (caller_type, var);
|
1284 |
|
|
|
1285 |
|
|
STRIP_USELESS_TYPE_CONVERSION (use);
|
1286 |
|
|
|
1287 |
|
|
if (DECL_BY_REFERENCE (result))
|
1288 |
|
|
var = build_fold_addr_expr (var);
|
1289 |
|
|
|
1290 |
|
|
done:
|
1291 |
|
|
/* Register the VAR_DECL as the equivalent for the RESULT_DECL; that
|
1292 |
|
|
way, when the RESULT_DECL is encountered, it will be
|
1293 |
|
|
automatically replaced by the VAR_DECL. */
|
1294 |
|
|
insert_decl_map (id, result, var);
|
1295 |
|
|
|
1296 |
|
|
/* Remember this so we can ignore it in remap_decls. */
|
1297 |
|
|
id->retvar = var;
|
1298 |
|
|
|
1299 |
|
|
*use_p = use;
|
1300 |
|
|
return var;
|
1301 |
|
|
}
|
1302 |
|
|
|
1303 |
|
|
/* Returns nonzero if a function can be inlined as a tree. */
|
1304 |
|
|
|
1305 |
|
|
bool
|
1306 |
|
|
tree_inlinable_function_p (tree fn)
|
1307 |
|
|
{
|
1308 |
|
|
return inlinable_function_p (fn);
|
1309 |
|
|
}
|
1310 |
|
|
|
1311 |
|
|
static const char *inline_forbidden_reason;
|
1312 |
|
|
|
1313 |
|
|
static tree
|
1314 |
|
|
inline_forbidden_p_1 (tree *nodep, int *walk_subtrees ATTRIBUTE_UNUSED,
|
1315 |
|
|
void *fnp)
|
1316 |
|
|
{
|
1317 |
|
|
tree node = *nodep;
|
1318 |
|
|
tree fn = (tree) fnp;
|
1319 |
|
|
tree t;
|
1320 |
|
|
|
1321 |
|
|
switch (TREE_CODE (node))
|
1322 |
|
|
{
|
1323 |
|
|
case CALL_EXPR:
|
1324 |
|
|
/* Refuse to inline alloca call unless user explicitly forced so as
|
1325 |
|
|
this may change program's memory overhead drastically when the
|
1326 |
|
|
function using alloca is called in loop. In GCC present in
|
1327 |
|
|
SPEC2000 inlining into schedule_block cause it to require 2GB of
|
1328 |
|
|
RAM instead of 256MB. */
|
1329 |
|
|
if (alloca_call_p (node)
|
1330 |
|
|
&& !lookup_attribute ("always_inline", DECL_ATTRIBUTES (fn)))
|
1331 |
|
|
{
|
1332 |
|
|
inline_forbidden_reason
|
1333 |
|
|
= G_("function %q+F can never be inlined because it uses "
|
1334 |
|
|
"alloca (override using the always_inline attribute)");
|
1335 |
|
|
return node;
|
1336 |
|
|
}
|
1337 |
|
|
t = get_callee_fndecl (node);
|
1338 |
|
|
if (! t)
|
1339 |
|
|
break;
|
1340 |
|
|
|
1341 |
|
|
/* We cannot inline functions that call setjmp. */
|
1342 |
|
|
if (setjmp_call_p (t))
|
1343 |
|
|
{
|
1344 |
|
|
inline_forbidden_reason
|
1345 |
|
|
= G_("function %q+F can never be inlined because it uses setjmp");
|
1346 |
|
|
return node;
|
1347 |
|
|
}
|
1348 |
|
|
|
1349 |
|
|
if (DECL_BUILT_IN_CLASS (t) == BUILT_IN_NORMAL)
|
1350 |
|
|
switch (DECL_FUNCTION_CODE (t))
|
1351 |
|
|
{
|
1352 |
|
|
/* We cannot inline functions that take a variable number of
|
1353 |
|
|
arguments. */
|
1354 |
|
|
case BUILT_IN_VA_START:
|
1355 |
|
|
case BUILT_IN_STDARG_START:
|
1356 |
|
|
case BUILT_IN_NEXT_ARG:
|
1357 |
|
|
case BUILT_IN_VA_END:
|
1358 |
|
|
inline_forbidden_reason
|
1359 |
|
|
= G_("function %q+F can never be inlined because it "
|
1360 |
|
|
"uses variable argument lists");
|
1361 |
|
|
return node;
|
1362 |
|
|
|
1363 |
|
|
case BUILT_IN_LONGJMP:
|
1364 |
|
|
/* We can't inline functions that call __builtin_longjmp at
|
1365 |
|
|
all. The non-local goto machinery really requires the
|
1366 |
|
|
destination be in a different function. If we allow the
|
1367 |
|
|
function calling __builtin_longjmp to be inlined into the
|
1368 |
|
|
function calling __builtin_setjmp, Things will Go Awry. */
|
1369 |
|
|
inline_forbidden_reason
|
1370 |
|
|
= G_("function %q+F can never be inlined because "
|
1371 |
|
|
"it uses setjmp-longjmp exception handling");
|
1372 |
|
|
return node;
|
1373 |
|
|
|
1374 |
|
|
case BUILT_IN_NONLOCAL_GOTO:
|
1375 |
|
|
/* Similarly. */
|
1376 |
|
|
inline_forbidden_reason
|
1377 |
|
|
= G_("function %q+F can never be inlined because "
|
1378 |
|
|
"it uses non-local goto");
|
1379 |
|
|
return node;
|
1380 |
|
|
|
1381 |
|
|
case BUILT_IN_RETURN:
|
1382 |
|
|
case BUILT_IN_APPLY_ARGS:
|
1383 |
|
|
/* If a __builtin_apply_args caller would be inlined,
|
1384 |
|
|
it would be saving arguments of the function it has
|
1385 |
|
|
been inlined into. Similarly __builtin_return would
|
1386 |
|
|
return from the function the inline has been inlined into. */
|
1387 |
|
|
inline_forbidden_reason
|
1388 |
|
|
= G_("function %q+F can never be inlined because "
|
1389 |
|
|
"it uses __builtin_return or __builtin_apply_args");
|
1390 |
|
|
return node;
|
1391 |
|
|
|
1392 |
|
|
default:
|
1393 |
|
|
break;
|
1394 |
|
|
}
|
1395 |
|
|
break;
|
1396 |
|
|
|
1397 |
|
|
case GOTO_EXPR:
|
1398 |
|
|
t = TREE_OPERAND (node, 0);
|
1399 |
|
|
|
1400 |
|
|
/* We will not inline a function which uses computed goto. The
|
1401 |
|
|
addresses of its local labels, which may be tucked into
|
1402 |
|
|
global storage, are of course not constant across
|
1403 |
|
|
instantiations, which causes unexpected behavior. */
|
1404 |
|
|
if (TREE_CODE (t) != LABEL_DECL)
|
1405 |
|
|
{
|
1406 |
|
|
inline_forbidden_reason
|
1407 |
|
|
= G_("function %q+F can never be inlined "
|
1408 |
|
|
"because it contains a computed goto");
|
1409 |
|
|
return node;
|
1410 |
|
|
}
|
1411 |
|
|
break;
|
1412 |
|
|
|
1413 |
|
|
case LABEL_EXPR:
|
1414 |
|
|
t = TREE_OPERAND (node, 0);
|
1415 |
|
|
if (DECL_NONLOCAL (t))
|
1416 |
|
|
{
|
1417 |
|
|
/* We cannot inline a function that receives a non-local goto
|
1418 |
|
|
because we cannot remap the destination label used in the
|
1419 |
|
|
function that is performing the non-local goto. */
|
1420 |
|
|
inline_forbidden_reason
|
1421 |
|
|
= G_("function %q+F can never be inlined "
|
1422 |
|
|
"because it receives a non-local goto");
|
1423 |
|
|
return node;
|
1424 |
|
|
}
|
1425 |
|
|
break;
|
1426 |
|
|
|
1427 |
|
|
case RECORD_TYPE:
|
1428 |
|
|
case UNION_TYPE:
|
1429 |
|
|
/* We cannot inline a function of the form
|
1430 |
|
|
|
1431 |
|
|
void F (int i) { struct S { int ar[i]; } s; }
|
1432 |
|
|
|
1433 |
|
|
Attempting to do so produces a catch-22.
|
1434 |
|
|
If walk_tree examines the TYPE_FIELDS chain of RECORD_TYPE/
|
1435 |
|
|
UNION_TYPE nodes, then it goes into infinite recursion on a
|
1436 |
|
|
structure containing a pointer to its own type. If it doesn't,
|
1437 |
|
|
then the type node for S doesn't get adjusted properly when
|
1438 |
|
|
F is inlined.
|
1439 |
|
|
|
1440 |
|
|
??? This is likely no longer true, but it's too late in the 4.0
|
1441 |
|
|
cycle to try to find out. This should be checked for 4.1. */
|
1442 |
|
|
for (t = TYPE_FIELDS (node); t; t = TREE_CHAIN (t))
|
1443 |
|
|
if (variably_modified_type_p (TREE_TYPE (t), NULL))
|
1444 |
|
|
{
|
1445 |
|
|
inline_forbidden_reason
|
1446 |
|
|
= G_("function %q+F can never be inlined "
|
1447 |
|
|
"because it uses variable sized variables");
|
1448 |
|
|
return node;
|
1449 |
|
|
}
|
1450 |
|
|
|
1451 |
|
|
default:
|
1452 |
|
|
break;
|
1453 |
|
|
}
|
1454 |
|
|
|
1455 |
|
|
return NULL_TREE;
|
1456 |
|
|
}
|
1457 |
|
|
|
1458 |
|
|
/* Return subexpression representing possible alloca call, if any. */
|
1459 |
|
|
static tree
|
1460 |
|
|
inline_forbidden_p (tree fndecl)
|
1461 |
|
|
{
|
1462 |
|
|
location_t saved_loc = input_location;
|
1463 |
|
|
block_stmt_iterator bsi;
|
1464 |
|
|
basic_block bb;
|
1465 |
|
|
tree ret = NULL_TREE;
|
1466 |
|
|
|
1467 |
|
|
FOR_EACH_BB_FN (bb, DECL_STRUCT_FUNCTION (fndecl))
|
1468 |
|
|
for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi))
|
1469 |
|
|
{
|
1470 |
|
|
ret = walk_tree_without_duplicates (bsi_stmt_ptr (bsi),
|
1471 |
|
|
inline_forbidden_p_1, fndecl);
|
1472 |
|
|
if (ret)
|
1473 |
|
|
goto egress;
|
1474 |
|
|
}
|
1475 |
|
|
|
1476 |
|
|
egress:
|
1477 |
|
|
input_location = saved_loc;
|
1478 |
|
|
return ret;
|
1479 |
|
|
}
|
1480 |
|
|
|
1481 |
|
|
/* Returns nonzero if FN is a function that does not have any
|
1482 |
|
|
fundamental inline blocking properties. */
|
1483 |
|
|
|
1484 |
|
|
static bool
|
1485 |
|
|
inlinable_function_p (tree fn)
|
1486 |
|
|
{
|
1487 |
|
|
bool inlinable = true;
|
1488 |
|
|
|
1489 |
|
|
/* If we've already decided this function shouldn't be inlined,
|
1490 |
|
|
there's no need to check again. */
|
1491 |
|
|
if (DECL_UNINLINABLE (fn))
|
1492 |
|
|
return false;
|
1493 |
|
|
|
1494 |
|
|
/* See if there is any language-specific reason it cannot be
|
1495 |
|
|
inlined. (It is important that this hook be called early because
|
1496 |
|
|
in C++ it may result in template instantiation.)
|
1497 |
|
|
If the function is not inlinable for language-specific reasons,
|
1498 |
|
|
it is left up to the langhook to explain why. */
|
1499 |
|
|
inlinable = !lang_hooks.tree_inlining.cannot_inline_tree_fn (&fn);
|
1500 |
|
|
|
1501 |
|
|
/* If we don't have the function body available, we can't inline it.
|
1502 |
|
|
However, this should not be recorded since we also get here for
|
1503 |
|
|
forward declared inline functions. Therefore, return at once. */
|
1504 |
|
|
if (!DECL_SAVED_TREE (fn))
|
1505 |
|
|
return false;
|
1506 |
|
|
|
1507 |
|
|
/* If we're not inlining at all, then we cannot inline this function. */
|
1508 |
|
|
else if (!flag_inline_trees)
|
1509 |
|
|
inlinable = false;
|
1510 |
|
|
|
1511 |
|
|
/* Only try to inline functions if DECL_INLINE is set. This should be
|
1512 |
|
|
true for all functions declared `inline', and for all other functions
|
1513 |
|
|
as well with -finline-functions.
|
1514 |
|
|
|
1515 |
|
|
Don't think of disregarding DECL_INLINE when flag_inline_trees == 2;
|
1516 |
|
|
it's the front-end that must set DECL_INLINE in this case, because
|
1517 |
|
|
dwarf2out loses if a function that does not have DECL_INLINE set is
|
1518 |
|
|
inlined anyway. That is why we have both DECL_INLINE and
|
1519 |
|
|
DECL_DECLARED_INLINE_P. */
|
1520 |
|
|
/* FIXME: When flag_inline_trees dies, the check for flag_unit_at_a_time
|
1521 |
|
|
here should be redundant. */
|
1522 |
|
|
else if (!DECL_INLINE (fn) && !flag_unit_at_a_time)
|
1523 |
|
|
inlinable = false;
|
1524 |
|
|
|
1525 |
|
|
else if (inline_forbidden_p (fn))
|
1526 |
|
|
{
|
1527 |
|
|
/* See if we should warn about uninlinable functions. Previously,
|
1528 |
|
|
some of these warnings would be issued while trying to expand
|
1529 |
|
|
the function inline, but that would cause multiple warnings
|
1530 |
|
|
about functions that would for example call alloca. But since
|
1531 |
|
|
this a property of the function, just one warning is enough.
|
1532 |
|
|
As a bonus we can now give more details about the reason why a
|
1533 |
|
|
function is not inlinable.
|
1534 |
|
|
We only warn for functions declared `inline' by the user. */
|
1535 |
|
|
bool do_warning = (warn_inline
|
1536 |
|
|
&& DECL_INLINE (fn)
|
1537 |
|
|
&& DECL_DECLARED_INLINE_P (fn)
|
1538 |
|
|
&& !DECL_IN_SYSTEM_HEADER (fn));
|
1539 |
|
|
|
1540 |
|
|
if (lookup_attribute ("always_inline", DECL_ATTRIBUTES (fn)))
|
1541 |
|
|
sorry (inline_forbidden_reason, fn);
|
1542 |
|
|
else if (do_warning)
|
1543 |
|
|
warning (OPT_Winline, inline_forbidden_reason, fn);
|
1544 |
|
|
|
1545 |
|
|
inlinable = false;
|
1546 |
|
|
}
|
1547 |
|
|
|
1548 |
|
|
/* Squirrel away the result so that we don't have to check again. */
|
1549 |
|
|
DECL_UNINLINABLE (fn) = !inlinable;
|
1550 |
|
|
|
1551 |
|
|
return inlinable;
|
1552 |
|
|
}
|
1553 |
|
|
|
1554 |
|
|
/* Estimate the cost of a memory move. Use machine dependent
|
1555 |
|
|
word size and take possible memcpy call into account. */
|
1556 |
|
|
|
1557 |
|
|
int
|
1558 |
|
|
estimate_move_cost (tree type)
|
1559 |
|
|
{
|
1560 |
|
|
HOST_WIDE_INT size;
|
1561 |
|
|
|
1562 |
|
|
size = int_size_in_bytes (type);
|
1563 |
|
|
|
1564 |
|
|
if (size < 0 || size > MOVE_MAX_PIECES * MOVE_RATIO)
|
1565 |
|
|
/* Cost of a memcpy call, 3 arguments and the call. */
|
1566 |
|
|
return 4;
|
1567 |
|
|
else
|
1568 |
|
|
return ((size + MOVE_MAX_PIECES - 1) / MOVE_MAX_PIECES);
|
1569 |
|
|
}
|
1570 |
|
|
|
1571 |
|
|
/* Used by estimate_num_insns. Estimate number of instructions seen
|
1572 |
|
|
by given statement. */
|
1573 |
|
|
|
1574 |
|
|
static tree
|
1575 |
|
|
estimate_num_insns_1 (tree *tp, int *walk_subtrees, void *data)
|
1576 |
|
|
{
|
1577 |
|
|
int *count = (int *) data;
|
1578 |
|
|
tree x = *tp;
|
1579 |
|
|
|
1580 |
|
|
if (IS_TYPE_OR_DECL_P (x))
|
1581 |
|
|
{
|
1582 |
|
|
*walk_subtrees = 0;
|
1583 |
|
|
return NULL;
|
1584 |
|
|
}
|
1585 |
|
|
/* Assume that constants and references counts nothing. These should
|
1586 |
|
|
be majorized by amount of operations among them we count later
|
1587 |
|
|
and are common target of CSE and similar optimizations. */
|
1588 |
|
|
else if (CONSTANT_CLASS_P (x) || REFERENCE_CLASS_P (x))
|
1589 |
|
|
return NULL;
|
1590 |
|
|
|
1591 |
|
|
switch (TREE_CODE (x))
|
1592 |
|
|
{
|
1593 |
|
|
/* Containers have no cost. */
|
1594 |
|
|
case TREE_LIST:
|
1595 |
|
|
case TREE_VEC:
|
1596 |
|
|
case BLOCK:
|
1597 |
|
|
case COMPONENT_REF:
|
1598 |
|
|
case BIT_FIELD_REF:
|
1599 |
|
|
case INDIRECT_REF:
|
1600 |
|
|
case ALIGN_INDIRECT_REF:
|
1601 |
|
|
case MISALIGNED_INDIRECT_REF:
|
1602 |
|
|
case ARRAY_REF:
|
1603 |
|
|
case ARRAY_RANGE_REF:
|
1604 |
|
|
case OBJ_TYPE_REF:
|
1605 |
|
|
case EXC_PTR_EXPR: /* ??? */
|
1606 |
|
|
case FILTER_EXPR: /* ??? */
|
1607 |
|
|
case COMPOUND_EXPR:
|
1608 |
|
|
case BIND_EXPR:
|
1609 |
|
|
case WITH_CLEANUP_EXPR:
|
1610 |
|
|
case NOP_EXPR:
|
1611 |
|
|
case VIEW_CONVERT_EXPR:
|
1612 |
|
|
case SAVE_EXPR:
|
1613 |
|
|
case ADDR_EXPR:
|
1614 |
|
|
case COMPLEX_EXPR:
|
1615 |
|
|
case RANGE_EXPR:
|
1616 |
|
|
case CASE_LABEL_EXPR:
|
1617 |
|
|
case SSA_NAME:
|
1618 |
|
|
case CATCH_EXPR:
|
1619 |
|
|
case EH_FILTER_EXPR:
|
1620 |
|
|
case STATEMENT_LIST:
|
1621 |
|
|
case ERROR_MARK:
|
1622 |
|
|
case NON_LVALUE_EXPR:
|
1623 |
|
|
case FDESC_EXPR:
|
1624 |
|
|
case VA_ARG_EXPR:
|
1625 |
|
|
case TRY_CATCH_EXPR:
|
1626 |
|
|
case TRY_FINALLY_EXPR:
|
1627 |
|
|
case LABEL_EXPR:
|
1628 |
|
|
case GOTO_EXPR:
|
1629 |
|
|
case RETURN_EXPR:
|
1630 |
|
|
case EXIT_EXPR:
|
1631 |
|
|
case LOOP_EXPR:
|
1632 |
|
|
case PHI_NODE:
|
1633 |
|
|
case WITH_SIZE_EXPR:
|
1634 |
|
|
case OMP_CLAUSE:
|
1635 |
|
|
case OMP_RETURN:
|
1636 |
|
|
case OMP_CONTINUE:
|
1637 |
|
|
break;
|
1638 |
|
|
|
1639 |
|
|
/* We don't account constants for now. Assume that the cost is amortized
|
1640 |
|
|
by operations that do use them. We may re-consider this decision once
|
1641 |
|
|
we are able to optimize the tree before estimating its size and break
|
1642 |
|
|
out static initializers. */
|
1643 |
|
|
case IDENTIFIER_NODE:
|
1644 |
|
|
case INTEGER_CST:
|
1645 |
|
|
case REAL_CST:
|
1646 |
|
|
case COMPLEX_CST:
|
1647 |
|
|
case VECTOR_CST:
|
1648 |
|
|
case STRING_CST:
|
1649 |
|
|
*walk_subtrees = 0;
|
1650 |
|
|
return NULL;
|
1651 |
|
|
|
1652 |
|
|
/* Try to estimate the cost of assignments. We have three cases to
|
1653 |
|
|
deal with:
|
1654 |
|
|
1) Simple assignments to registers;
|
1655 |
|
|
2) Stores to things that must live in memory. This includes
|
1656 |
|
|
"normal" stores to scalars, but also assignments of large
|
1657 |
|
|
structures, or constructors of big arrays;
|
1658 |
|
|
3) TARGET_EXPRs.
|
1659 |
|
|
|
1660 |
|
|
Let us look at the first two cases, assuming we have "a = b + C":
|
1661 |
|
|
<modify_expr <var_decl "a"> <plus_expr <var_decl "b"> <constant C>>
|
1662 |
|
|
If "a" is a GIMPLE register, the assignment to it is free on almost
|
1663 |
|
|
any target, because "a" usually ends up in a real register. Hence
|
1664 |
|
|
the only cost of this expression comes from the PLUS_EXPR, and we
|
1665 |
|
|
can ignore the MODIFY_EXPR.
|
1666 |
|
|
If "a" is not a GIMPLE register, the assignment to "a" will most
|
1667 |
|
|
likely be a real store, so the cost of the MODIFY_EXPR is the cost
|
1668 |
|
|
of moving something into "a", which we compute using the function
|
1669 |
|
|
estimate_move_cost.
|
1670 |
|
|
|
1671 |
|
|
The third case deals with TARGET_EXPRs, for which the semantics are
|
1672 |
|
|
that a temporary is assigned, unless the TARGET_EXPR itself is being
|
1673 |
|
|
assigned to something else. In the latter case we do not need the
|
1674 |
|
|
temporary. E.g. in <modify_expr <var_decl "a"> <target_expr>>, the
|
1675 |
|
|
MODIFY_EXPR is free. */
|
1676 |
|
|
case INIT_EXPR:
|
1677 |
|
|
case MODIFY_EXPR:
|
1678 |
|
|
/* Is the right and side a TARGET_EXPR? */
|
1679 |
|
|
if (TREE_CODE (TREE_OPERAND (x, 1)) == TARGET_EXPR)
|
1680 |
|
|
break;
|
1681 |
|
|
/* ... fall through ... */
|
1682 |
|
|
|
1683 |
|
|
case TARGET_EXPR:
|
1684 |
|
|
x = TREE_OPERAND (x, 0);
|
1685 |
|
|
/* Is this an assignments to a register? */
|
1686 |
|
|
if (is_gimple_reg (x))
|
1687 |
|
|
break;
|
1688 |
|
|
/* Otherwise it's a store, so fall through to compute the move cost. */
|
1689 |
|
|
|
1690 |
|
|
case CONSTRUCTOR:
|
1691 |
|
|
*count += estimate_move_cost (TREE_TYPE (x));
|
1692 |
|
|
break;
|
1693 |
|
|
|
1694 |
|
|
/* Assign cost of 1 to usual operations.
|
1695 |
|
|
??? We may consider mapping RTL costs to this. */
|
1696 |
|
|
case COND_EXPR:
|
1697 |
|
|
case VEC_COND_EXPR:
|
1698 |
|
|
|
1699 |
|
|
case PLUS_EXPR:
|
1700 |
|
|
case MINUS_EXPR:
|
1701 |
|
|
case MULT_EXPR:
|
1702 |
|
|
|
1703 |
|
|
case FIX_TRUNC_EXPR:
|
1704 |
|
|
case FIX_CEIL_EXPR:
|
1705 |
|
|
case FIX_FLOOR_EXPR:
|
1706 |
|
|
case FIX_ROUND_EXPR:
|
1707 |
|
|
|
1708 |
|
|
case NEGATE_EXPR:
|
1709 |
|
|
case FLOAT_EXPR:
|
1710 |
|
|
case MIN_EXPR:
|
1711 |
|
|
case MAX_EXPR:
|
1712 |
|
|
case ABS_EXPR:
|
1713 |
|
|
|
1714 |
|
|
case LSHIFT_EXPR:
|
1715 |
|
|
case RSHIFT_EXPR:
|
1716 |
|
|
case LROTATE_EXPR:
|
1717 |
|
|
case RROTATE_EXPR:
|
1718 |
|
|
case VEC_LSHIFT_EXPR:
|
1719 |
|
|
case VEC_RSHIFT_EXPR:
|
1720 |
|
|
|
1721 |
|
|
case BIT_IOR_EXPR:
|
1722 |
|
|
case BIT_XOR_EXPR:
|
1723 |
|
|
case BIT_AND_EXPR:
|
1724 |
|
|
case BIT_NOT_EXPR:
|
1725 |
|
|
|
1726 |
|
|
case TRUTH_ANDIF_EXPR:
|
1727 |
|
|
case TRUTH_ORIF_EXPR:
|
1728 |
|
|
case TRUTH_AND_EXPR:
|
1729 |
|
|
case TRUTH_OR_EXPR:
|
1730 |
|
|
case TRUTH_XOR_EXPR:
|
1731 |
|
|
case TRUTH_NOT_EXPR:
|
1732 |
|
|
|
1733 |
|
|
case LT_EXPR:
|
1734 |
|
|
case LE_EXPR:
|
1735 |
|
|
case GT_EXPR:
|
1736 |
|
|
case GE_EXPR:
|
1737 |
|
|
case EQ_EXPR:
|
1738 |
|
|
case NE_EXPR:
|
1739 |
|
|
case ORDERED_EXPR:
|
1740 |
|
|
case UNORDERED_EXPR:
|
1741 |
|
|
|
1742 |
|
|
case UNLT_EXPR:
|
1743 |
|
|
case UNLE_EXPR:
|
1744 |
|
|
case UNGT_EXPR:
|
1745 |
|
|
case UNGE_EXPR:
|
1746 |
|
|
case UNEQ_EXPR:
|
1747 |
|
|
case LTGT_EXPR:
|
1748 |
|
|
|
1749 |
|
|
case CONVERT_EXPR:
|
1750 |
|
|
|
1751 |
|
|
case CONJ_EXPR:
|
1752 |
|
|
|
1753 |
|
|
case PREDECREMENT_EXPR:
|
1754 |
|
|
case PREINCREMENT_EXPR:
|
1755 |
|
|
case POSTDECREMENT_EXPR:
|
1756 |
|
|
case POSTINCREMENT_EXPR:
|
1757 |
|
|
|
1758 |
|
|
case SWITCH_EXPR:
|
1759 |
|
|
|
1760 |
|
|
case ASM_EXPR:
|
1761 |
|
|
|
1762 |
|
|
case REALIGN_LOAD_EXPR:
|
1763 |
|
|
|
1764 |
|
|
case REDUC_MAX_EXPR:
|
1765 |
|
|
case REDUC_MIN_EXPR:
|
1766 |
|
|
case REDUC_PLUS_EXPR:
|
1767 |
|
|
case WIDEN_SUM_EXPR:
|
1768 |
|
|
case DOT_PROD_EXPR:
|
1769 |
|
|
|
1770 |
|
|
case WIDEN_MULT_EXPR:
|
1771 |
|
|
|
1772 |
|
|
case RESX_EXPR:
|
1773 |
|
|
*count += 1;
|
1774 |
|
|
break;
|
1775 |
|
|
|
1776 |
|
|
/* Few special cases of expensive operations. This is useful
|
1777 |
|
|
to avoid inlining on functions having too many of these. */
|
1778 |
|
|
case TRUNC_DIV_EXPR:
|
1779 |
|
|
case CEIL_DIV_EXPR:
|
1780 |
|
|
case FLOOR_DIV_EXPR:
|
1781 |
|
|
case ROUND_DIV_EXPR:
|
1782 |
|
|
case EXACT_DIV_EXPR:
|
1783 |
|
|
case TRUNC_MOD_EXPR:
|
1784 |
|
|
case CEIL_MOD_EXPR:
|
1785 |
|
|
case FLOOR_MOD_EXPR:
|
1786 |
|
|
case ROUND_MOD_EXPR:
|
1787 |
|
|
case RDIV_EXPR:
|
1788 |
|
|
*count += 10;
|
1789 |
|
|
break;
|
1790 |
|
|
case CALL_EXPR:
|
1791 |
|
|
{
|
1792 |
|
|
tree decl = get_callee_fndecl (x);
|
1793 |
|
|
tree arg;
|
1794 |
|
|
|
1795 |
|
|
if (decl && DECL_BUILT_IN_CLASS (decl) == BUILT_IN_NORMAL)
|
1796 |
|
|
switch (DECL_FUNCTION_CODE (decl))
|
1797 |
|
|
{
|
1798 |
|
|
case BUILT_IN_CONSTANT_P:
|
1799 |
|
|
*walk_subtrees = 0;
|
1800 |
|
|
return NULL_TREE;
|
1801 |
|
|
case BUILT_IN_EXPECT:
|
1802 |
|
|
return NULL_TREE;
|
1803 |
|
|
default:
|
1804 |
|
|
break;
|
1805 |
|
|
}
|
1806 |
|
|
|
1807 |
|
|
/* Our cost must be kept in sync with cgraph_estimate_size_after_inlining
|
1808 |
|
|
that does use function declaration to figure out the arguments. */
|
1809 |
|
|
if (!decl)
|
1810 |
|
|
{
|
1811 |
|
|
for (arg = TREE_OPERAND (x, 1); arg; arg = TREE_CHAIN (arg))
|
1812 |
|
|
*count += estimate_move_cost (TREE_TYPE (TREE_VALUE (arg)));
|
1813 |
|
|
}
|
1814 |
|
|
else
|
1815 |
|
|
{
|
1816 |
|
|
for (arg = DECL_ARGUMENTS (decl); arg; arg = TREE_CHAIN (arg))
|
1817 |
|
|
*count += estimate_move_cost (TREE_TYPE (arg));
|
1818 |
|
|
}
|
1819 |
|
|
|
1820 |
|
|
*count += PARAM_VALUE (PARAM_INLINE_CALL_COST);
|
1821 |
|
|
break;
|
1822 |
|
|
}
|
1823 |
|
|
|
1824 |
|
|
case OMP_PARALLEL:
|
1825 |
|
|
case OMP_FOR:
|
1826 |
|
|
case OMP_SECTIONS:
|
1827 |
|
|
case OMP_SINGLE:
|
1828 |
|
|
case OMP_SECTION:
|
1829 |
|
|
case OMP_MASTER:
|
1830 |
|
|
case OMP_ORDERED:
|
1831 |
|
|
case OMP_CRITICAL:
|
1832 |
|
|
case OMP_ATOMIC:
|
1833 |
|
|
/* OpenMP directives are generally very expensive. */
|
1834 |
|
|
*count += 40;
|
1835 |
|
|
break;
|
1836 |
|
|
|
1837 |
|
|
default:
|
1838 |
|
|
gcc_unreachable ();
|
1839 |
|
|
}
|
1840 |
|
|
return NULL;
|
1841 |
|
|
}
|
1842 |
|
|
|
1843 |
|
|
/* Estimate number of instructions that will be created by expanding EXPR. */
|
1844 |
|
|
|
1845 |
|
|
int
|
1846 |
|
|
estimate_num_insns (tree expr)
|
1847 |
|
|
{
|
1848 |
|
|
int num = 0;
|
1849 |
|
|
struct pointer_set_t *visited_nodes;
|
1850 |
|
|
basic_block bb;
|
1851 |
|
|
block_stmt_iterator bsi;
|
1852 |
|
|
struct function *my_function;
|
1853 |
|
|
|
1854 |
|
|
/* If we're given an entire function, walk the CFG. */
|
1855 |
|
|
if (TREE_CODE (expr) == FUNCTION_DECL)
|
1856 |
|
|
{
|
1857 |
|
|
my_function = DECL_STRUCT_FUNCTION (expr);
|
1858 |
|
|
gcc_assert (my_function && my_function->cfg);
|
1859 |
|
|
visited_nodes = pointer_set_create ();
|
1860 |
|
|
FOR_EACH_BB_FN (bb, my_function)
|
1861 |
|
|
{
|
1862 |
|
|
for (bsi = bsi_start (bb);
|
1863 |
|
|
!bsi_end_p (bsi);
|
1864 |
|
|
bsi_next (&bsi))
|
1865 |
|
|
{
|
1866 |
|
|
walk_tree (bsi_stmt_ptr (bsi), estimate_num_insns_1,
|
1867 |
|
|
&num, visited_nodes);
|
1868 |
|
|
}
|
1869 |
|
|
}
|
1870 |
|
|
pointer_set_destroy (visited_nodes);
|
1871 |
|
|
}
|
1872 |
|
|
else
|
1873 |
|
|
walk_tree_without_duplicates (&expr, estimate_num_insns_1, &num);
|
1874 |
|
|
|
1875 |
|
|
return num;
|
1876 |
|
|
}
|
1877 |
|
|
|
1878 |
|
|
typedef struct function *function_p;
|
1879 |
|
|
|
1880 |
|
|
DEF_VEC_P(function_p);
|
1881 |
|
|
DEF_VEC_ALLOC_P(function_p,heap);
|
1882 |
|
|
|
1883 |
|
|
/* Initialized with NOGC, making this poisonous to the garbage collector. */
|
1884 |
|
|
static VEC(function_p,heap) *cfun_stack;
|
1885 |
|
|
|
1886 |
|
|
void
|
1887 |
|
|
push_cfun (struct function *new_cfun)
|
1888 |
|
|
{
|
1889 |
|
|
VEC_safe_push (function_p, heap, cfun_stack, cfun);
|
1890 |
|
|
cfun = new_cfun;
|
1891 |
|
|
}
|
1892 |
|
|
|
1893 |
|
|
void
|
1894 |
|
|
pop_cfun (void)
|
1895 |
|
|
{
|
1896 |
|
|
cfun = VEC_pop (function_p, cfun_stack);
|
1897 |
|
|
}
|
1898 |
|
|
|
1899 |
|
|
/* Install new lexical TREE_BLOCK underneath 'current_block'. */
|
1900 |
|
|
static void
|
1901 |
|
|
add_lexical_block (tree current_block, tree new_block)
|
1902 |
|
|
{
|
1903 |
|
|
tree *blk_p;
|
1904 |
|
|
|
1905 |
|
|
/* Walk to the last sub-block. */
|
1906 |
|
|
for (blk_p = &BLOCK_SUBBLOCKS (current_block);
|
1907 |
|
|
*blk_p;
|
1908 |
|
|
blk_p = &TREE_CHAIN (*blk_p))
|
1909 |
|
|
;
|
1910 |
|
|
*blk_p = new_block;
|
1911 |
|
|
BLOCK_SUPERCONTEXT (new_block) = current_block;
|
1912 |
|
|
}
|
1913 |
|
|
|
1914 |
|
|
/* If *TP is a CALL_EXPR, replace it with its inline expansion. */
|
1915 |
|
|
|
1916 |
|
|
static bool
|
1917 |
|
|
expand_call_inline (basic_block bb, tree stmt, tree *tp, void *data)
|
1918 |
|
|
{
|
1919 |
|
|
copy_body_data *id;
|
1920 |
|
|
tree t;
|
1921 |
|
|
tree use_retvar;
|
1922 |
|
|
tree fn;
|
1923 |
|
|
splay_tree st;
|
1924 |
|
|
tree args;
|
1925 |
|
|
tree return_slot_addr;
|
1926 |
|
|
tree modify_dest;
|
1927 |
|
|
location_t saved_location;
|
1928 |
|
|
struct cgraph_edge *cg_edge;
|
1929 |
|
|
const char *reason;
|
1930 |
|
|
basic_block return_block;
|
1931 |
|
|
edge e;
|
1932 |
|
|
block_stmt_iterator bsi, stmt_bsi;
|
1933 |
|
|
bool successfully_inlined = FALSE;
|
1934 |
|
|
bool purge_dead_abnormal_edges;
|
1935 |
|
|
tree t_step;
|
1936 |
|
|
tree var;
|
1937 |
|
|
|
1938 |
|
|
/* See what we've got. */
|
1939 |
|
|
id = (copy_body_data *) data;
|
1940 |
|
|
t = *tp;
|
1941 |
|
|
|
1942 |
|
|
/* Set input_location here so we get the right instantiation context
|
1943 |
|
|
if we call instantiate_decl from inlinable_function_p. */
|
1944 |
|
|
saved_location = input_location;
|
1945 |
|
|
if (EXPR_HAS_LOCATION (t))
|
1946 |
|
|
input_location = EXPR_LOCATION (t);
|
1947 |
|
|
|
1948 |
|
|
/* From here on, we're only interested in CALL_EXPRs. */
|
1949 |
|
|
if (TREE_CODE (t) != CALL_EXPR)
|
1950 |
|
|
goto egress;
|
1951 |
|
|
|
1952 |
|
|
/* First, see if we can figure out what function is being called.
|
1953 |
|
|
If we cannot, then there is no hope of inlining the function. */
|
1954 |
|
|
fn = get_callee_fndecl (t);
|
1955 |
|
|
if (!fn)
|
1956 |
|
|
goto egress;
|
1957 |
|
|
|
1958 |
|
|
/* Turn forward declarations into real ones. */
|
1959 |
|
|
fn = cgraph_node (fn)->decl;
|
1960 |
|
|
|
1961 |
|
|
/* If fn is a declaration of a function in a nested scope that was
|
1962 |
|
|
globally declared inline, we don't set its DECL_INITIAL.
|
1963 |
|
|
However, we can't blindly follow DECL_ABSTRACT_ORIGIN because the
|
1964 |
|
|
C++ front-end uses it for cdtors to refer to their internal
|
1965 |
|
|
declarations, that are not real functions. Fortunately those
|
1966 |
|
|
don't have trees to be saved, so we can tell by checking their
|
1967 |
|
|
DECL_SAVED_TREE. */
|
1968 |
|
|
if (! DECL_INITIAL (fn)
|
1969 |
|
|
&& DECL_ABSTRACT_ORIGIN (fn)
|
1970 |
|
|
&& DECL_SAVED_TREE (DECL_ABSTRACT_ORIGIN (fn)))
|
1971 |
|
|
fn = DECL_ABSTRACT_ORIGIN (fn);
|
1972 |
|
|
|
1973 |
|
|
/* Objective C and fortran still calls tree_rest_of_compilation directly.
|
1974 |
|
|
Kill this check once this is fixed. */
|
1975 |
|
|
if (!id->dst_node->analyzed)
|
1976 |
|
|
goto egress;
|
1977 |
|
|
|
1978 |
|
|
cg_edge = cgraph_edge (id->dst_node, stmt);
|
1979 |
|
|
|
1980 |
|
|
/* Constant propagation on argument done during previous inlining
|
1981 |
|
|
may create new direct call. Produce an edge for it. */
|
1982 |
|
|
if (!cg_edge)
|
1983 |
|
|
{
|
1984 |
|
|
struct cgraph_node *dest = cgraph_node (fn);
|
1985 |
|
|
|
1986 |
|
|
/* We have missing edge in the callgraph. This can happen in one case
|
1987 |
|
|
where previous inlining turned indirect call into direct call by
|
1988 |
|
|
constant propagating arguments. In all other cases we hit a bug
|
1989 |
|
|
(incorrect node sharing is most common reason for missing edges. */
|
1990 |
|
|
gcc_assert (dest->needed || !flag_unit_at_a_time);
|
1991 |
|
|
cgraph_create_edge (id->dst_node, dest, stmt,
|
1992 |
|
|
bb->count, bb->loop_depth)->inline_failed
|
1993 |
|
|
= N_("originally indirect function call not considered for inlining");
|
1994 |
|
|
goto egress;
|
1995 |
|
|
}
|
1996 |
|
|
|
1997 |
|
|
/* Don't try to inline functions that are not well-suited to
|
1998 |
|
|
inlining. */
|
1999 |
|
|
if (!cgraph_inline_p (cg_edge, &reason))
|
2000 |
|
|
{
|
2001 |
|
|
if (lookup_attribute ("always_inline", DECL_ATTRIBUTES (fn))
|
2002 |
|
|
/* Avoid warnings during early inline pass. */
|
2003 |
|
|
&& (!flag_unit_at_a_time || cgraph_global_info_ready))
|
2004 |
|
|
{
|
2005 |
|
|
sorry ("inlining failed in call to %q+F: %s", fn, reason);
|
2006 |
|
|
sorry ("called from here");
|
2007 |
|
|
}
|
2008 |
|
|
else if (warn_inline && DECL_DECLARED_INLINE_P (fn)
|
2009 |
|
|
&& !DECL_IN_SYSTEM_HEADER (fn)
|
2010 |
|
|
&& strlen (reason)
|
2011 |
|
|
&& !lookup_attribute ("noinline", DECL_ATTRIBUTES (fn))
|
2012 |
|
|
/* Avoid warnings during early inline pass. */
|
2013 |
|
|
&& (!flag_unit_at_a_time || cgraph_global_info_ready))
|
2014 |
|
|
{
|
2015 |
|
|
warning (OPT_Winline, "inlining failed in call to %q+F: %s",
|
2016 |
|
|
fn, reason);
|
2017 |
|
|
warning (OPT_Winline, "called from here");
|
2018 |
|
|
}
|
2019 |
|
|
goto egress;
|
2020 |
|
|
}
|
2021 |
|
|
fn = cg_edge->callee->decl;
|
2022 |
|
|
|
2023 |
|
|
#ifdef ENABLE_CHECKING
|
2024 |
|
|
if (cg_edge->callee->decl != id->dst_node->decl)
|
2025 |
|
|
verify_cgraph_node (cg_edge->callee);
|
2026 |
|
|
#endif
|
2027 |
|
|
|
2028 |
|
|
/* We will be inlining this callee. */
|
2029 |
|
|
id->eh_region = lookup_stmt_eh_region (stmt);
|
2030 |
|
|
|
2031 |
|
|
/* Split the block holding the CALL_EXPR. */
|
2032 |
|
|
e = split_block (bb, stmt);
|
2033 |
|
|
bb = e->src;
|
2034 |
|
|
return_block = e->dest;
|
2035 |
|
|
remove_edge (e);
|
2036 |
|
|
|
2037 |
|
|
/* split_block splits after the statement; work around this by
|
2038 |
|
|
moving the call into the second block manually. Not pretty,
|
2039 |
|
|
but seems easier than doing the CFG manipulation by hand
|
2040 |
|
|
when the CALL_EXPR is in the last statement of BB. */
|
2041 |
|
|
stmt_bsi = bsi_last (bb);
|
2042 |
|
|
bsi_remove (&stmt_bsi, false);
|
2043 |
|
|
|
2044 |
|
|
/* If the CALL_EXPR was in the last statement of BB, it may have
|
2045 |
|
|
been the source of abnormal edges. In this case, schedule
|
2046 |
|
|
the removal of dead abnormal edges. */
|
2047 |
|
|
bsi = bsi_start (return_block);
|
2048 |
|
|
if (bsi_end_p (bsi))
|
2049 |
|
|
{
|
2050 |
|
|
bsi_insert_after (&bsi, stmt, BSI_NEW_STMT);
|
2051 |
|
|
purge_dead_abnormal_edges = true;
|
2052 |
|
|
}
|
2053 |
|
|
else
|
2054 |
|
|
{
|
2055 |
|
|
bsi_insert_before (&bsi, stmt, BSI_NEW_STMT);
|
2056 |
|
|
purge_dead_abnormal_edges = false;
|
2057 |
|
|
}
|
2058 |
|
|
|
2059 |
|
|
stmt_bsi = bsi_start (return_block);
|
2060 |
|
|
|
2061 |
|
|
/* Build a block containing code to initialize the arguments, the
|
2062 |
|
|
actual inline expansion of the body, and a label for the return
|
2063 |
|
|
statements within the function to jump to. The type of the
|
2064 |
|
|
statement expression is the return type of the function call. */
|
2065 |
|
|
id->block = make_node (BLOCK);
|
2066 |
|
|
BLOCK_ABSTRACT_ORIGIN (id->block) = fn;
|
2067 |
|
|
BLOCK_SOURCE_LOCATION (id->block) = input_location;
|
2068 |
|
|
add_lexical_block (TREE_BLOCK (stmt), id->block);
|
2069 |
|
|
|
2070 |
|
|
/* Local declarations will be replaced by their equivalents in this
|
2071 |
|
|
map. */
|
2072 |
|
|
st = id->decl_map;
|
2073 |
|
|
id->decl_map = splay_tree_new (splay_tree_compare_pointers,
|
2074 |
|
|
NULL, NULL);
|
2075 |
|
|
|
2076 |
|
|
/* Initialize the parameters. */
|
2077 |
|
|
args = TREE_OPERAND (t, 1);
|
2078 |
|
|
|
2079 |
|
|
/* Record the function we are about to inline. */
|
2080 |
|
|
id->src_fn = fn;
|
2081 |
|
|
id->src_node = cg_edge->callee;
|
2082 |
|
|
|
2083 |
|
|
initialize_inlined_parameters (id, args, TREE_OPERAND (t, 2), fn, bb);
|
2084 |
|
|
|
2085 |
|
|
if (DECL_INITIAL (fn))
|
2086 |
|
|
add_lexical_block (id->block, remap_blocks (DECL_INITIAL (fn), id));
|
2087 |
|
|
|
2088 |
|
|
/* Return statements in the function body will be replaced by jumps
|
2089 |
|
|
to the RET_LABEL. */
|
2090 |
|
|
|
2091 |
|
|
gcc_assert (DECL_INITIAL (fn));
|
2092 |
|
|
gcc_assert (TREE_CODE (DECL_INITIAL (fn)) == BLOCK);
|
2093 |
|
|
|
2094 |
|
|
/* Find the lhs to which the result of this call is assigned. */
|
2095 |
|
|
return_slot_addr = NULL;
|
2096 |
|
|
if (TREE_CODE (stmt) == MODIFY_EXPR)
|
2097 |
|
|
{
|
2098 |
|
|
modify_dest = TREE_OPERAND (stmt, 0);
|
2099 |
|
|
|
2100 |
|
|
/* The function which we are inlining might not return a value,
|
2101 |
|
|
in which case we should issue a warning that the function
|
2102 |
|
|
does not return a value. In that case the optimizers will
|
2103 |
|
|
see that the variable to which the value is assigned was not
|
2104 |
|
|
initialized. We do not want to issue a warning about that
|
2105 |
|
|
uninitialized variable. */
|
2106 |
|
|
if (DECL_P (modify_dest))
|
2107 |
|
|
TREE_NO_WARNING (modify_dest) = 1;
|
2108 |
|
|
if (CALL_EXPR_RETURN_SLOT_OPT (t))
|
2109 |
|
|
{
|
2110 |
|
|
return_slot_addr = build_fold_addr_expr (modify_dest);
|
2111 |
|
|
STRIP_USELESS_TYPE_CONVERSION (return_slot_addr);
|
2112 |
|
|
modify_dest = NULL;
|
2113 |
|
|
}
|
2114 |
|
|
}
|
2115 |
|
|
else
|
2116 |
|
|
modify_dest = NULL;
|
2117 |
|
|
|
2118 |
|
|
/* Declare the return variable for the function. */
|
2119 |
|
|
declare_return_variable (id, return_slot_addr,
|
2120 |
|
|
modify_dest, &use_retvar);
|
2121 |
|
|
|
2122 |
|
|
/* This is it. Duplicate the callee body. Assume callee is
|
2123 |
|
|
pre-gimplified. Note that we must not alter the caller
|
2124 |
|
|
function in any way before this point, as this CALL_EXPR may be
|
2125 |
|
|
a self-referential call; if we're calling ourselves, we need to
|
2126 |
|
|
duplicate our body before altering anything. */
|
2127 |
|
|
copy_body (id, bb->count, bb->frequency, bb, return_block);
|
2128 |
|
|
|
2129 |
|
|
/* Add local vars in this inlined callee to caller. */
|
2130 |
|
|
t_step = id->src_cfun->unexpanded_var_list;
|
2131 |
|
|
for (; t_step; t_step = TREE_CHAIN (t_step))
|
2132 |
|
|
{
|
2133 |
|
|
var = TREE_VALUE (t_step);
|
2134 |
|
|
if (TREE_STATIC (var) && !TREE_ASM_WRITTEN (var))
|
2135 |
|
|
cfun->unexpanded_var_list = tree_cons (NULL_TREE, var,
|
2136 |
|
|
cfun->unexpanded_var_list);
|
2137 |
|
|
else
|
2138 |
|
|
cfun->unexpanded_var_list = tree_cons (NULL_TREE, remap_decl (var, id),
|
2139 |
|
|
cfun->unexpanded_var_list);
|
2140 |
|
|
}
|
2141 |
|
|
|
2142 |
|
|
/* Clean up. */
|
2143 |
|
|
splay_tree_delete (id->decl_map);
|
2144 |
|
|
id->decl_map = st;
|
2145 |
|
|
|
2146 |
|
|
/* If the inlined function returns a result that we care about,
|
2147 |
|
|
clobber the CALL_EXPR with a reference to the return variable. */
|
2148 |
|
|
if (use_retvar && (TREE_CODE (bsi_stmt (stmt_bsi)) != CALL_EXPR))
|
2149 |
|
|
{
|
2150 |
|
|
*tp = use_retvar;
|
2151 |
|
|
maybe_clean_or_replace_eh_stmt (stmt, stmt);
|
2152 |
|
|
}
|
2153 |
|
|
else
|
2154 |
|
|
/* We're modifying a TSI owned by gimple_expand_calls_inline();
|
2155 |
|
|
tsi_delink() will leave the iterator in a sane state. */
|
2156 |
|
|
bsi_remove (&stmt_bsi, true);
|
2157 |
|
|
|
2158 |
|
|
if (purge_dead_abnormal_edges)
|
2159 |
|
|
tree_purge_dead_abnormal_call_edges (return_block);
|
2160 |
|
|
|
2161 |
|
|
/* If the value of the new expression is ignored, that's OK. We
|
2162 |
|
|
don't warn about this for CALL_EXPRs, so we shouldn't warn about
|
2163 |
|
|
the equivalent inlined version either. */
|
2164 |
|
|
TREE_USED (*tp) = 1;
|
2165 |
|
|
|
2166 |
|
|
/* Output the inlining info for this abstract function, since it has been
|
2167 |
|
|
inlined. If we don't do this now, we can lose the information about the
|
2168 |
|
|
variables in the function when the blocks get blown away as soon as we
|
2169 |
|
|
remove the cgraph node. */
|
2170 |
|
|
(*debug_hooks->outlining_inline_function) (cg_edge->callee->decl);
|
2171 |
|
|
|
2172 |
|
|
/* Update callgraph if needed. */
|
2173 |
|
|
cgraph_remove_node (cg_edge->callee);
|
2174 |
|
|
|
2175 |
|
|
id->block = NULL_TREE;
|
2176 |
|
|
successfully_inlined = TRUE;
|
2177 |
|
|
|
2178 |
|
|
egress:
|
2179 |
|
|
input_location = saved_location;
|
2180 |
|
|
return successfully_inlined;
|
2181 |
|
|
}
|
2182 |
|
|
|
2183 |
|
|
/* Expand call statements reachable from STMT_P.
|
2184 |
|
|
We can only have CALL_EXPRs as the "toplevel" tree code or nested
|
2185 |
|
|
in a MODIFY_EXPR. See tree-gimple.c:get_call_expr_in(). We can
|
2186 |
|
|
unfortunately not use that function here because we need a pointer
|
2187 |
|
|
to the CALL_EXPR, not the tree itself. */
|
2188 |
|
|
|
2189 |
|
|
static bool
|
2190 |
|
|
gimple_expand_calls_inline (basic_block bb, copy_body_data *id)
|
2191 |
|
|
{
|
2192 |
|
|
block_stmt_iterator bsi;
|
2193 |
|
|
|
2194 |
|
|
/* Register specific tree functions. */
|
2195 |
|
|
tree_register_cfg_hooks ();
|
2196 |
|
|
for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi))
|
2197 |
|
|
{
|
2198 |
|
|
tree *expr_p = bsi_stmt_ptr (bsi);
|
2199 |
|
|
tree stmt = *expr_p;
|
2200 |
|
|
|
2201 |
|
|
if (TREE_CODE (*expr_p) == MODIFY_EXPR)
|
2202 |
|
|
expr_p = &TREE_OPERAND (*expr_p, 1);
|
2203 |
|
|
if (TREE_CODE (*expr_p) == WITH_SIZE_EXPR)
|
2204 |
|
|
expr_p = &TREE_OPERAND (*expr_p, 0);
|
2205 |
|
|
if (TREE_CODE (*expr_p) == CALL_EXPR)
|
2206 |
|
|
if (expand_call_inline (bb, stmt, expr_p, id))
|
2207 |
|
|
return true;
|
2208 |
|
|
}
|
2209 |
|
|
return false;
|
2210 |
|
|
}
|
2211 |
|
|
|
2212 |
|
|
/* Expand calls to inline functions in the body of FN. */
|
2213 |
|
|
|
2214 |
|
|
void
|
2215 |
|
|
optimize_inline_calls (tree fn)
|
2216 |
|
|
{
|
2217 |
|
|
copy_body_data id;
|
2218 |
|
|
tree prev_fn;
|
2219 |
|
|
basic_block bb;
|
2220 |
|
|
/* There is no point in performing inlining if errors have already
|
2221 |
|
|
occurred -- and we might crash if we try to inline invalid
|
2222 |
|
|
code. */
|
2223 |
|
|
if (errorcount || sorrycount)
|
2224 |
|
|
return;
|
2225 |
|
|
|
2226 |
|
|
/* Clear out ID. */
|
2227 |
|
|
memset (&id, 0, sizeof (id));
|
2228 |
|
|
|
2229 |
|
|
id.src_node = id.dst_node = cgraph_node (fn);
|
2230 |
|
|
id.dst_fn = fn;
|
2231 |
|
|
/* Or any functions that aren't finished yet. */
|
2232 |
|
|
prev_fn = NULL_TREE;
|
2233 |
|
|
if (current_function_decl)
|
2234 |
|
|
{
|
2235 |
|
|
id.dst_fn = current_function_decl;
|
2236 |
|
|
prev_fn = current_function_decl;
|
2237 |
|
|
}
|
2238 |
|
|
|
2239 |
|
|
id.copy_decl = copy_decl_maybe_to_var;
|
2240 |
|
|
id.transform_call_graph_edges = CB_CGE_DUPLICATE;
|
2241 |
|
|
id.transform_new_cfg = false;
|
2242 |
|
|
id.transform_return_to_modify = true;
|
2243 |
|
|
id.transform_lang_insert_block = false;
|
2244 |
|
|
|
2245 |
|
|
push_gimplify_context ();
|
2246 |
|
|
|
2247 |
|
|
/* Reach the trees by walking over the CFG, and note the
|
2248 |
|
|
enclosing basic-blocks in the call edges. */
|
2249 |
|
|
/* We walk the blocks going forward, because inlined function bodies
|
2250 |
|
|
will split id->current_basic_block, and the new blocks will
|
2251 |
|
|
follow it; we'll trudge through them, processing their CALL_EXPRs
|
2252 |
|
|
along the way. */
|
2253 |
|
|
FOR_EACH_BB (bb)
|
2254 |
|
|
gimple_expand_calls_inline (bb, &id);
|
2255 |
|
|
|
2256 |
|
|
pop_gimplify_context (NULL);
|
2257 |
|
|
/* Renumber the (code) basic_blocks consecutively. */
|
2258 |
|
|
compact_blocks ();
|
2259 |
|
|
/* Renumber the lexical scoping (non-code) blocks consecutively. */
|
2260 |
|
|
number_blocks (fn);
|
2261 |
|
|
|
2262 |
|
|
#ifdef ENABLE_CHECKING
|
2263 |
|
|
{
|
2264 |
|
|
struct cgraph_edge *e;
|
2265 |
|
|
|
2266 |
|
|
verify_cgraph_node (id.dst_node);
|
2267 |
|
|
|
2268 |
|
|
/* Double check that we inlined everything we are supposed to inline. */
|
2269 |
|
|
for (e = id.dst_node->callees; e; e = e->next_callee)
|
2270 |
|
|
gcc_assert (e->inline_failed);
|
2271 |
|
|
}
|
2272 |
|
|
#endif
|
2273 |
|
|
/* We need to rescale frequencies again to peak at REG_BR_PROB_BASE
|
2274 |
|
|
as inlining loops might increase the maximum. */
|
2275 |
|
|
if (ENTRY_BLOCK_PTR->count)
|
2276 |
|
|
counts_to_freqs ();
|
2277 |
|
|
fold_cond_expr_cond ();
|
2278 |
|
|
}
|
2279 |
|
|
|
2280 |
|
|
/* FN is a function that has a complete body, and CLONE is a function whose
|
2281 |
|
|
body is to be set to a copy of FN, mapping argument declarations according
|
2282 |
|
|
to the ARG_MAP splay_tree. */
|
2283 |
|
|
|
2284 |
|
|
void
|
2285 |
|
|
clone_body (tree clone, tree fn, void *arg_map)
|
2286 |
|
|
{
|
2287 |
|
|
copy_body_data id;
|
2288 |
|
|
|
2289 |
|
|
/* Clone the body, as if we were making an inline call. But, remap the
|
2290 |
|
|
parameters in the callee to the parameters of caller. */
|
2291 |
|
|
memset (&id, 0, sizeof (id));
|
2292 |
|
|
id.src_fn = fn;
|
2293 |
|
|
id.dst_fn = clone;
|
2294 |
|
|
id.src_cfun = DECL_STRUCT_FUNCTION (fn);
|
2295 |
|
|
id.decl_map = (splay_tree)arg_map;
|
2296 |
|
|
|
2297 |
|
|
id.copy_decl = copy_decl_no_change;
|
2298 |
|
|
id.transform_call_graph_edges = CB_CGE_DUPLICATE;
|
2299 |
|
|
id.transform_new_cfg = true;
|
2300 |
|
|
id.transform_return_to_modify = false;
|
2301 |
|
|
id.transform_lang_insert_block = true;
|
2302 |
|
|
|
2303 |
|
|
/* We're not inside any EH region. */
|
2304 |
|
|
id.eh_region = -1;
|
2305 |
|
|
|
2306 |
|
|
/* Actually copy the body. */
|
2307 |
|
|
append_to_statement_list_force (copy_generic_body (&id), &DECL_SAVED_TREE (clone));
|
2308 |
|
|
}
|
2309 |
|
|
|
2310 |
|
|
/* Passed to walk_tree. Copies the node pointed to, if appropriate. */
|
2311 |
|
|
|
2312 |
|
|
tree
|
2313 |
|
|
copy_tree_r (tree *tp, int *walk_subtrees, void *data ATTRIBUTE_UNUSED)
|
2314 |
|
|
{
|
2315 |
|
|
enum tree_code code = TREE_CODE (*tp);
|
2316 |
|
|
|
2317 |
|
|
/* We make copies of most nodes. */
|
2318 |
|
|
if (IS_EXPR_CODE_CLASS (TREE_CODE_CLASS (code))
|
2319 |
|
|
|| code == TREE_LIST
|
2320 |
|
|
|| code == TREE_VEC
|
2321 |
|
|
|| code == TYPE_DECL
|
2322 |
|
|
|| code == OMP_CLAUSE)
|
2323 |
|
|
{
|
2324 |
|
|
/* Because the chain gets clobbered when we make a copy, we save it
|
2325 |
|
|
here. */
|
2326 |
|
|
tree chain = TREE_CHAIN (*tp);
|
2327 |
|
|
tree new;
|
2328 |
|
|
|
2329 |
|
|
/* Copy the node. */
|
2330 |
|
|
new = copy_node (*tp);
|
2331 |
|
|
|
2332 |
|
|
/* Propagate mudflap marked-ness. */
|
2333 |
|
|
if (flag_mudflap && mf_marked_p (*tp))
|
2334 |
|
|
mf_mark (new);
|
2335 |
|
|
|
2336 |
|
|
*tp = new;
|
2337 |
|
|
|
2338 |
|
|
/* Now, restore the chain, if appropriate. That will cause
|
2339 |
|
|
walk_tree to walk into the chain as well. */
|
2340 |
|
|
if (code == PARM_DECL
|
2341 |
|
|
|| code == TREE_LIST
|
2342 |
|
|
|| code == OMP_CLAUSE)
|
2343 |
|
|
TREE_CHAIN (*tp) = chain;
|
2344 |
|
|
|
2345 |
|
|
/* For now, we don't update BLOCKs when we make copies. So, we
|
2346 |
|
|
have to nullify all BIND_EXPRs. */
|
2347 |
|
|
if (TREE_CODE (*tp) == BIND_EXPR)
|
2348 |
|
|
BIND_EXPR_BLOCK (*tp) = NULL_TREE;
|
2349 |
|
|
}
|
2350 |
|
|
else if (code == CONSTRUCTOR)
|
2351 |
|
|
{
|
2352 |
|
|
/* CONSTRUCTOR nodes need special handling because
|
2353 |
|
|
we need to duplicate the vector of elements. */
|
2354 |
|
|
tree new;
|
2355 |
|
|
|
2356 |
|
|
new = copy_node (*tp);
|
2357 |
|
|
|
2358 |
|
|
/* Propagate mudflap marked-ness. */
|
2359 |
|
|
if (flag_mudflap && mf_marked_p (*tp))
|
2360 |
|
|
mf_mark (new);
|
2361 |
|
|
|
2362 |
|
|
CONSTRUCTOR_ELTS (new) = VEC_copy (constructor_elt, gc,
|
2363 |
|
|
CONSTRUCTOR_ELTS (*tp));
|
2364 |
|
|
*tp = new;
|
2365 |
|
|
}
|
2366 |
|
|
else if (TREE_CODE_CLASS (code) == tcc_type)
|
2367 |
|
|
*walk_subtrees = 0;
|
2368 |
|
|
else if (TREE_CODE_CLASS (code) == tcc_declaration)
|
2369 |
|
|
*walk_subtrees = 0;
|
2370 |
|
|
else if (TREE_CODE_CLASS (code) == tcc_constant)
|
2371 |
|
|
*walk_subtrees = 0;
|
2372 |
|
|
else
|
2373 |
|
|
gcc_assert (code != STATEMENT_LIST);
|
2374 |
|
|
return NULL_TREE;
|
2375 |
|
|
}
|
2376 |
|
|
|
2377 |
|
|
/* The SAVE_EXPR pointed to by TP is being copied. If ST contains
|
2378 |
|
|
information indicating to what new SAVE_EXPR this one should be mapped,
|
2379 |
|
|
use that one. Otherwise, create a new node and enter it in ST. FN is
|
2380 |
|
|
the function into which the copy will be placed. */
|
2381 |
|
|
|
2382 |
|
|
static void
|
2383 |
|
|
remap_save_expr (tree *tp, void *st_, int *walk_subtrees)
|
2384 |
|
|
{
|
2385 |
|
|
splay_tree st = (splay_tree) st_;
|
2386 |
|
|
splay_tree_node n;
|
2387 |
|
|
tree t;
|
2388 |
|
|
|
2389 |
|
|
/* See if we already encountered this SAVE_EXPR. */
|
2390 |
|
|
n = splay_tree_lookup (st, (splay_tree_key) *tp);
|
2391 |
|
|
|
2392 |
|
|
/* If we didn't already remap this SAVE_EXPR, do so now. */
|
2393 |
|
|
if (!n)
|
2394 |
|
|
{
|
2395 |
|
|
t = copy_node (*tp);
|
2396 |
|
|
|
2397 |
|
|
/* Remember this SAVE_EXPR. */
|
2398 |
|
|
splay_tree_insert (st, (splay_tree_key) *tp, (splay_tree_value) t);
|
2399 |
|
|
/* Make sure we don't remap an already-remapped SAVE_EXPR. */
|
2400 |
|
|
splay_tree_insert (st, (splay_tree_key) t, (splay_tree_value) t);
|
2401 |
|
|
}
|
2402 |
|
|
else
|
2403 |
|
|
{
|
2404 |
|
|
/* We've already walked into this SAVE_EXPR; don't do it again. */
|
2405 |
|
|
*walk_subtrees = 0;
|
2406 |
|
|
t = (tree) n->value;
|
2407 |
|
|
}
|
2408 |
|
|
|
2409 |
|
|
/* Replace this SAVE_EXPR with the copy. */
|
2410 |
|
|
*tp = t;
|
2411 |
|
|
}
|
2412 |
|
|
|
2413 |
|
|
/* Called via walk_tree. If *TP points to a DECL_STMT for a local label,
|
2414 |
|
|
copies the declaration and enters it in the splay_tree in DATA (which is
|
2415 |
|
|
really an `copy_body_data *'). */
|
2416 |
|
|
|
2417 |
|
|
static tree
|
2418 |
|
|
mark_local_for_remap_r (tree *tp, int *walk_subtrees ATTRIBUTE_UNUSED,
|
2419 |
|
|
void *data)
|
2420 |
|
|
{
|
2421 |
|
|
copy_body_data *id = (copy_body_data *) data;
|
2422 |
|
|
|
2423 |
|
|
/* Don't walk into types. */
|
2424 |
|
|
if (TYPE_P (*tp))
|
2425 |
|
|
*walk_subtrees = 0;
|
2426 |
|
|
|
2427 |
|
|
else if (TREE_CODE (*tp) == LABEL_EXPR)
|
2428 |
|
|
{
|
2429 |
|
|
tree decl = TREE_OPERAND (*tp, 0);
|
2430 |
|
|
|
2431 |
|
|
/* Copy the decl and remember the copy. */
|
2432 |
|
|
insert_decl_map (id, decl, id->copy_decl (decl, id));
|
2433 |
|
|
}
|
2434 |
|
|
|
2435 |
|
|
return NULL_TREE;
|
2436 |
|
|
}
|
2437 |
|
|
|
2438 |
|
|
/* Perform any modifications to EXPR required when it is unsaved. Does
|
2439 |
|
|
not recurse into EXPR's subtrees. */
|
2440 |
|
|
|
2441 |
|
|
static void
|
2442 |
|
|
unsave_expr_1 (tree expr)
|
2443 |
|
|
{
|
2444 |
|
|
switch (TREE_CODE (expr))
|
2445 |
|
|
{
|
2446 |
|
|
case TARGET_EXPR:
|
2447 |
|
|
/* Don't mess with a TARGET_EXPR that hasn't been expanded.
|
2448 |
|
|
It's OK for this to happen if it was part of a subtree that
|
2449 |
|
|
isn't immediately expanded, such as operand 2 of another
|
2450 |
|
|
TARGET_EXPR. */
|
2451 |
|
|
if (TREE_OPERAND (expr, 1))
|
2452 |
|
|
break;
|
2453 |
|
|
|
2454 |
|
|
TREE_OPERAND (expr, 1) = TREE_OPERAND (expr, 3);
|
2455 |
|
|
TREE_OPERAND (expr, 3) = NULL_TREE;
|
2456 |
|
|
break;
|
2457 |
|
|
|
2458 |
|
|
default:
|
2459 |
|
|
break;
|
2460 |
|
|
}
|
2461 |
|
|
}
|
2462 |
|
|
|
2463 |
|
|
/* Called via walk_tree when an expression is unsaved. Using the
|
2464 |
|
|
splay_tree pointed to by ST (which is really a `splay_tree'),
|
2465 |
|
|
remaps all local declarations to appropriate replacements. */
|
2466 |
|
|
|
2467 |
|
|
static tree
|
2468 |
|
|
unsave_r (tree *tp, int *walk_subtrees, void *data)
|
2469 |
|
|
{
|
2470 |
|
|
copy_body_data *id = (copy_body_data *) data;
|
2471 |
|
|
splay_tree st = id->decl_map;
|
2472 |
|
|
splay_tree_node n;
|
2473 |
|
|
|
2474 |
|
|
/* Only a local declaration (variable or label). */
|
2475 |
|
|
if ((TREE_CODE (*tp) == VAR_DECL && !TREE_STATIC (*tp))
|
2476 |
|
|
|| TREE_CODE (*tp) == LABEL_DECL)
|
2477 |
|
|
{
|
2478 |
|
|
/* Lookup the declaration. */
|
2479 |
|
|
n = splay_tree_lookup (st, (splay_tree_key) *tp);
|
2480 |
|
|
|
2481 |
|
|
/* If it's there, remap it. */
|
2482 |
|
|
if (n)
|
2483 |
|
|
*tp = (tree) n->value;
|
2484 |
|
|
}
|
2485 |
|
|
|
2486 |
|
|
else if (TREE_CODE (*tp) == STATEMENT_LIST)
|
2487 |
|
|
copy_statement_list (tp);
|
2488 |
|
|
else if (TREE_CODE (*tp) == BIND_EXPR)
|
2489 |
|
|
copy_bind_expr (tp, walk_subtrees, id);
|
2490 |
|
|
else if (TREE_CODE (*tp) == SAVE_EXPR)
|
2491 |
|
|
remap_save_expr (tp, st, walk_subtrees);
|
2492 |
|
|
else
|
2493 |
|
|
{
|
2494 |
|
|
copy_tree_r (tp, walk_subtrees, NULL);
|
2495 |
|
|
|
2496 |
|
|
/* Do whatever unsaving is required. */
|
2497 |
|
|
unsave_expr_1 (*tp);
|
2498 |
|
|
}
|
2499 |
|
|
|
2500 |
|
|
/* Keep iterating. */
|
2501 |
|
|
return NULL_TREE;
|
2502 |
|
|
}
|
2503 |
|
|
|
2504 |
|
|
/* Copies everything in EXPR and replaces variables, labels
|
2505 |
|
|
and SAVE_EXPRs local to EXPR. */
|
2506 |
|
|
|
2507 |
|
|
tree
|
2508 |
|
|
unsave_expr_now (tree expr)
|
2509 |
|
|
{
|
2510 |
|
|
copy_body_data id;
|
2511 |
|
|
|
2512 |
|
|
/* There's nothing to do for NULL_TREE. */
|
2513 |
|
|
if (expr == 0)
|
2514 |
|
|
return expr;
|
2515 |
|
|
|
2516 |
|
|
/* Set up ID. */
|
2517 |
|
|
memset (&id, 0, sizeof (id));
|
2518 |
|
|
id.src_fn = current_function_decl;
|
2519 |
|
|
id.dst_fn = current_function_decl;
|
2520 |
|
|
id.decl_map = splay_tree_new (splay_tree_compare_pointers, NULL, NULL);
|
2521 |
|
|
|
2522 |
|
|
id.copy_decl = copy_decl_no_change;
|
2523 |
|
|
id.transform_call_graph_edges = CB_CGE_DUPLICATE;
|
2524 |
|
|
id.transform_new_cfg = false;
|
2525 |
|
|
id.transform_return_to_modify = false;
|
2526 |
|
|
id.transform_lang_insert_block = false;
|
2527 |
|
|
|
2528 |
|
|
/* Walk the tree once to find local labels. */
|
2529 |
|
|
walk_tree_without_duplicates (&expr, mark_local_for_remap_r, &id);
|
2530 |
|
|
|
2531 |
|
|
/* Walk the tree again, copying, remapping, and unsaving. */
|
2532 |
|
|
walk_tree (&expr, unsave_r, &id, NULL);
|
2533 |
|
|
|
2534 |
|
|
/* Clean up. */
|
2535 |
|
|
splay_tree_delete (id.decl_map);
|
2536 |
|
|
|
2537 |
|
|
return expr;
|
2538 |
|
|
}
|
2539 |
|
|
|
2540 |
|
|
/* Allow someone to determine if SEARCH is a child of TOP from gdb. */
|
2541 |
|
|
|
2542 |
|
|
static tree
|
2543 |
|
|
debug_find_tree_1 (tree *tp, int *walk_subtrees ATTRIBUTE_UNUSED, void *data)
|
2544 |
|
|
{
|
2545 |
|
|
if (*tp == data)
|
2546 |
|
|
return (tree) data;
|
2547 |
|
|
else
|
2548 |
|
|
return NULL;
|
2549 |
|
|
}
|
2550 |
|
|
|
2551 |
|
|
bool
|
2552 |
|
|
debug_find_tree (tree top, tree search)
|
2553 |
|
|
{
|
2554 |
|
|
return walk_tree_without_duplicates (&top, debug_find_tree_1, search) != 0;
|
2555 |
|
|
}
|
2556 |
|
|
|
2557 |
|
|
|
2558 |
|
|
/* Declare the variables created by the inliner. Add all the variables in
|
2559 |
|
|
VARS to BIND_EXPR. */
|
2560 |
|
|
|
2561 |
|
|
static void
|
2562 |
|
|
declare_inline_vars (tree block, tree vars)
|
2563 |
|
|
{
|
2564 |
|
|
tree t;
|
2565 |
|
|
for (t = vars; t; t = TREE_CHAIN (t))
|
2566 |
|
|
{
|
2567 |
|
|
DECL_SEEN_IN_BIND_EXPR_P (t) = 1;
|
2568 |
|
|
gcc_assert (!TREE_STATIC (t) && !TREE_ASM_WRITTEN (t));
|
2569 |
|
|
cfun->unexpanded_var_list =
|
2570 |
|
|
tree_cons (NULL_TREE, t,
|
2571 |
|
|
cfun->unexpanded_var_list);
|
2572 |
|
|
}
|
2573 |
|
|
|
2574 |
|
|
if (block)
|
2575 |
|
|
BLOCK_VARS (block) = chainon (BLOCK_VARS (block), vars);
|
2576 |
|
|
}
|
2577 |
|
|
|
2578 |
|
|
|
2579 |
|
|
/* Copy NODE (which must be a DECL). The DECL originally was in the FROM_FN,
|
2580 |
|
|
but now it will be in the TO_FN. PARM_TO_VAR means enable PARM_DECL to
|
2581 |
|
|
VAR_DECL translation. */
|
2582 |
|
|
|
2583 |
|
|
static tree
|
2584 |
|
|
copy_decl_for_dup_finish (copy_body_data *id, tree decl, tree copy)
|
2585 |
|
|
{
|
2586 |
|
|
/* Don't generate debug information for the copy if we wouldn't have
|
2587 |
|
|
generated it for the copy either. */
|
2588 |
|
|
DECL_ARTIFICIAL (copy) = DECL_ARTIFICIAL (decl);
|
2589 |
|
|
DECL_IGNORED_P (copy) = DECL_IGNORED_P (decl);
|
2590 |
|
|
|
2591 |
|
|
/* Set the DECL_ABSTRACT_ORIGIN so the debugging routines know what
|
2592 |
|
|
declaration inspired this copy. */
|
2593 |
|
|
DECL_ABSTRACT_ORIGIN (copy) = DECL_ORIGIN (decl);
|
2594 |
|
|
|
2595 |
|
|
/* The new variable/label has no RTL, yet. */
|
2596 |
|
|
if (CODE_CONTAINS_STRUCT (TREE_CODE (copy), TS_DECL_WRTL)
|
2597 |
|
|
&& !TREE_STATIC (copy) && !DECL_EXTERNAL (copy))
|
2598 |
|
|
SET_DECL_RTL (copy, NULL_RTX);
|
2599 |
|
|
|
2600 |
|
|
/* These args would always appear unused, if not for this. */
|
2601 |
|
|
TREE_USED (copy) = 1;
|
2602 |
|
|
|
2603 |
|
|
/* Set the context for the new declaration. */
|
2604 |
|
|
if (!DECL_CONTEXT (decl))
|
2605 |
|
|
/* Globals stay global. */
|
2606 |
|
|
;
|
2607 |
|
|
else if (DECL_CONTEXT (decl) != id->src_fn)
|
2608 |
|
|
/* Things that weren't in the scope of the function we're inlining
|
2609 |
|
|
from aren't in the scope we're inlining to, either. */
|
2610 |
|
|
;
|
2611 |
|
|
else if (TREE_STATIC (decl))
|
2612 |
|
|
/* Function-scoped static variables should stay in the original
|
2613 |
|
|
function. */
|
2614 |
|
|
;
|
2615 |
|
|
else
|
2616 |
|
|
/* Ordinary automatic local variables are now in the scope of the
|
2617 |
|
|
new function. */
|
2618 |
|
|
DECL_CONTEXT (copy) = id->dst_fn;
|
2619 |
|
|
|
2620 |
|
|
return copy;
|
2621 |
|
|
}
|
2622 |
|
|
|
2623 |
|
|
static tree
|
2624 |
|
|
copy_decl_to_var (tree decl, copy_body_data *id)
|
2625 |
|
|
{
|
2626 |
|
|
tree copy, type;
|
2627 |
|
|
|
2628 |
|
|
gcc_assert (TREE_CODE (decl) == PARM_DECL
|
2629 |
|
|
|| TREE_CODE (decl) == RESULT_DECL);
|
2630 |
|
|
|
2631 |
|
|
type = TREE_TYPE (decl);
|
2632 |
|
|
|
2633 |
|
|
copy = build_decl (VAR_DECL, DECL_NAME (decl), type);
|
2634 |
|
|
TREE_ADDRESSABLE (copy) = TREE_ADDRESSABLE (decl);
|
2635 |
|
|
TREE_READONLY (copy) = TREE_READONLY (decl);
|
2636 |
|
|
TREE_THIS_VOLATILE (copy) = TREE_THIS_VOLATILE (decl);
|
2637 |
|
|
DECL_COMPLEX_GIMPLE_REG_P (copy) = DECL_COMPLEX_GIMPLE_REG_P (decl);
|
2638 |
|
|
|
2639 |
|
|
return copy_decl_for_dup_finish (id, decl, copy);
|
2640 |
|
|
}
|
2641 |
|
|
|
2642 |
|
|
/* Like copy_decl_to_var, but create a return slot object instead of a
|
2643 |
|
|
pointer variable for return by invisible reference. */
|
2644 |
|
|
|
2645 |
|
|
static tree
|
2646 |
|
|
copy_result_decl_to_var (tree decl, copy_body_data *id)
|
2647 |
|
|
{
|
2648 |
|
|
tree copy, type;
|
2649 |
|
|
|
2650 |
|
|
gcc_assert (TREE_CODE (decl) == PARM_DECL
|
2651 |
|
|
|| TREE_CODE (decl) == RESULT_DECL);
|
2652 |
|
|
|
2653 |
|
|
type = TREE_TYPE (decl);
|
2654 |
|
|
if (DECL_BY_REFERENCE (decl))
|
2655 |
|
|
type = TREE_TYPE (type);
|
2656 |
|
|
|
2657 |
|
|
copy = build_decl (VAR_DECL, DECL_NAME (decl), type);
|
2658 |
|
|
TREE_READONLY (copy) = TREE_READONLY (decl);
|
2659 |
|
|
TREE_THIS_VOLATILE (copy) = TREE_THIS_VOLATILE (decl);
|
2660 |
|
|
if (!DECL_BY_REFERENCE (decl))
|
2661 |
|
|
{
|
2662 |
|
|
TREE_ADDRESSABLE (copy) = TREE_ADDRESSABLE (decl);
|
2663 |
|
|
DECL_COMPLEX_GIMPLE_REG_P (copy) = DECL_COMPLEX_GIMPLE_REG_P (decl);
|
2664 |
|
|
}
|
2665 |
|
|
|
2666 |
|
|
return copy_decl_for_dup_finish (id, decl, copy);
|
2667 |
|
|
}
|
2668 |
|
|
|
2669 |
|
|
|
2670 |
|
|
static tree
|
2671 |
|
|
copy_decl_no_change (tree decl, copy_body_data *id)
|
2672 |
|
|
{
|
2673 |
|
|
tree copy;
|
2674 |
|
|
|
2675 |
|
|
copy = copy_node (decl);
|
2676 |
|
|
|
2677 |
|
|
/* The COPY is not abstract; it will be generated in DST_FN. */
|
2678 |
|
|
DECL_ABSTRACT (copy) = 0;
|
2679 |
|
|
lang_hooks.dup_lang_specific_decl (copy);
|
2680 |
|
|
|
2681 |
|
|
/* TREE_ADDRESSABLE isn't used to indicate that a label's address has
|
2682 |
|
|
been taken; it's for internal bookkeeping in expand_goto_internal. */
|
2683 |
|
|
if (TREE_CODE (copy) == LABEL_DECL)
|
2684 |
|
|
{
|
2685 |
|
|
TREE_ADDRESSABLE (copy) = 0;
|
2686 |
|
|
LABEL_DECL_UID (copy) = -1;
|
2687 |
|
|
}
|
2688 |
|
|
|
2689 |
|
|
return copy_decl_for_dup_finish (id, decl, copy);
|
2690 |
|
|
}
|
2691 |
|
|
|
2692 |
|
|
static tree
|
2693 |
|
|
copy_decl_maybe_to_var (tree decl, copy_body_data *id)
|
2694 |
|
|
{
|
2695 |
|
|
if (TREE_CODE (decl) == PARM_DECL || TREE_CODE (decl) == RESULT_DECL)
|
2696 |
|
|
return copy_decl_to_var (decl, id);
|
2697 |
|
|
else
|
2698 |
|
|
return copy_decl_no_change (decl, id);
|
2699 |
|
|
}
|
2700 |
|
|
|
2701 |
|
|
/* Return a copy of the function's argument tree. */
|
2702 |
|
|
static tree
|
2703 |
|
|
copy_arguments_for_versioning (tree orig_parm, copy_body_data * id)
|
2704 |
|
|
{
|
2705 |
|
|
tree *arg_copy, *parg;
|
2706 |
|
|
|
2707 |
|
|
arg_copy = &orig_parm;
|
2708 |
|
|
for (parg = arg_copy; *parg; parg = &TREE_CHAIN (*parg))
|
2709 |
|
|
{
|
2710 |
|
|
tree new = remap_decl (*parg, id);
|
2711 |
|
|
lang_hooks.dup_lang_specific_decl (new);
|
2712 |
|
|
TREE_CHAIN (new) = TREE_CHAIN (*parg);
|
2713 |
|
|
*parg = new;
|
2714 |
|
|
}
|
2715 |
|
|
return orig_parm;
|
2716 |
|
|
}
|
2717 |
|
|
|
2718 |
|
|
/* Return a copy of the function's static chain. */
|
2719 |
|
|
static tree
|
2720 |
|
|
copy_static_chain (tree static_chain, copy_body_data * id)
|
2721 |
|
|
{
|
2722 |
|
|
tree *chain_copy, *pvar;
|
2723 |
|
|
|
2724 |
|
|
chain_copy = &static_chain;
|
2725 |
|
|
for (pvar = chain_copy; *pvar; pvar = &TREE_CHAIN (*pvar))
|
2726 |
|
|
{
|
2727 |
|
|
tree new = remap_decl (*pvar, id);
|
2728 |
|
|
lang_hooks.dup_lang_specific_decl (new);
|
2729 |
|
|
TREE_CHAIN (new) = TREE_CHAIN (*pvar);
|
2730 |
|
|
*pvar = new;
|
2731 |
|
|
}
|
2732 |
|
|
return static_chain;
|
2733 |
|
|
}
|
2734 |
|
|
|
2735 |
|
|
/* Return true if the function is allowed to be versioned.
|
2736 |
|
|
This is a guard for the versioning functionality. */
|
2737 |
|
|
bool
|
2738 |
|
|
tree_versionable_function_p (tree fndecl)
|
2739 |
|
|
{
|
2740 |
|
|
if (fndecl == NULL_TREE)
|
2741 |
|
|
return false;
|
2742 |
|
|
/* ??? There are cases where a function is
|
2743 |
|
|
uninlinable but can be versioned. */
|
2744 |
|
|
if (!tree_inlinable_function_p (fndecl))
|
2745 |
|
|
return false;
|
2746 |
|
|
|
2747 |
|
|
return true;
|
2748 |
|
|
}
|
2749 |
|
|
|
2750 |
|
|
/* Create a copy of a function's tree.
|
2751 |
|
|
OLD_DECL and NEW_DECL are FUNCTION_DECL tree nodes
|
2752 |
|
|
of the original function and the new copied function
|
2753 |
|
|
respectively. In case we want to replace a DECL
|
2754 |
|
|
tree with another tree while duplicating the function's
|
2755 |
|
|
body, TREE_MAP represents the mapping between these
|
2756 |
|
|
trees. If UPDATE_CLONES is set, the call_stmt fields
|
2757 |
|
|
of edges of clones of the function will be updated. */
|
2758 |
|
|
void
|
2759 |
|
|
tree_function_versioning (tree old_decl, tree new_decl, varray_type tree_map,
|
2760 |
|
|
bool update_clones)
|
2761 |
|
|
{
|
2762 |
|
|
struct cgraph_node *old_version_node;
|
2763 |
|
|
struct cgraph_node *new_version_node;
|
2764 |
|
|
copy_body_data id;
|
2765 |
|
|
tree p, new_fndecl;
|
2766 |
|
|
unsigned i;
|
2767 |
|
|
struct ipa_replace_map *replace_info;
|
2768 |
|
|
basic_block old_entry_block;
|
2769 |
|
|
tree t_step;
|
2770 |
|
|
|
2771 |
|
|
gcc_assert (TREE_CODE (old_decl) == FUNCTION_DECL
|
2772 |
|
|
&& TREE_CODE (new_decl) == FUNCTION_DECL);
|
2773 |
|
|
DECL_POSSIBLY_INLINED (old_decl) = 1;
|
2774 |
|
|
|
2775 |
|
|
old_version_node = cgraph_node (old_decl);
|
2776 |
|
|
new_version_node = cgraph_node (new_decl);
|
2777 |
|
|
|
2778 |
|
|
allocate_struct_function (new_decl);
|
2779 |
|
|
/* Cfun points to the new allocated function struct at this point. */
|
2780 |
|
|
cfun->function_end_locus = DECL_SOURCE_LOCATION (new_decl);
|
2781 |
|
|
|
2782 |
|
|
DECL_ARTIFICIAL (new_decl) = 1;
|
2783 |
|
|
DECL_ABSTRACT_ORIGIN (new_decl) = DECL_ORIGIN (old_decl);
|
2784 |
|
|
|
2785 |
|
|
/* Generate a new name for the new version. */
|
2786 |
|
|
if (!update_clones)
|
2787 |
|
|
DECL_NAME (new_decl) = create_tmp_var_name (NULL);
|
2788 |
|
|
/* Create a new SYMBOL_REF rtx for the new name. */
|
2789 |
|
|
if (DECL_RTL (old_decl) != NULL)
|
2790 |
|
|
{
|
2791 |
|
|
SET_DECL_RTL (new_decl, copy_rtx (DECL_RTL (old_decl)));
|
2792 |
|
|
XEXP (DECL_RTL (new_decl), 0) =
|
2793 |
|
|
gen_rtx_SYMBOL_REF (GET_MODE (XEXP (DECL_RTL (old_decl), 0)),
|
2794 |
|
|
IDENTIFIER_POINTER (DECL_NAME (new_decl)));
|
2795 |
|
|
}
|
2796 |
|
|
|
2797 |
|
|
/* Prepare the data structures for the tree copy. */
|
2798 |
|
|
memset (&id, 0, sizeof (id));
|
2799 |
|
|
|
2800 |
|
|
id.decl_map = splay_tree_new (splay_tree_compare_pointers, NULL, NULL);
|
2801 |
|
|
id.src_fn = old_decl;
|
2802 |
|
|
id.dst_fn = new_decl;
|
2803 |
|
|
id.src_node = old_version_node;
|
2804 |
|
|
id.dst_node = new_version_node;
|
2805 |
|
|
id.src_cfun = DECL_STRUCT_FUNCTION (old_decl);
|
2806 |
|
|
|
2807 |
|
|
id.copy_decl = copy_decl_no_change;
|
2808 |
|
|
id.transform_call_graph_edges
|
2809 |
|
|
= update_clones ? CB_CGE_MOVE_CLONES : CB_CGE_MOVE;
|
2810 |
|
|
id.transform_new_cfg = true;
|
2811 |
|
|
id.transform_return_to_modify = false;
|
2812 |
|
|
id.transform_lang_insert_block = false;
|
2813 |
|
|
|
2814 |
|
|
current_function_decl = new_decl;
|
2815 |
|
|
|
2816 |
|
|
/* Copy the function's static chain. */
|
2817 |
|
|
p = DECL_STRUCT_FUNCTION (old_decl)->static_chain_decl;
|
2818 |
|
|
if (p)
|
2819 |
|
|
DECL_STRUCT_FUNCTION (new_decl)->static_chain_decl =
|
2820 |
|
|
copy_static_chain (DECL_STRUCT_FUNCTION (old_decl)->static_chain_decl,
|
2821 |
|
|
&id);
|
2822 |
|
|
/* Copy the function's arguments. */
|
2823 |
|
|
if (DECL_ARGUMENTS (old_decl) != NULL_TREE)
|
2824 |
|
|
DECL_ARGUMENTS (new_decl) =
|
2825 |
|
|
copy_arguments_for_versioning (DECL_ARGUMENTS (old_decl), &id);
|
2826 |
|
|
|
2827 |
|
|
/* If there's a tree_map, prepare for substitution. */
|
2828 |
|
|
if (tree_map)
|
2829 |
|
|
for (i = 0; i < VARRAY_ACTIVE_SIZE (tree_map); i++)
|
2830 |
|
|
{
|
2831 |
|
|
replace_info = VARRAY_GENERIC_PTR (tree_map, i);
|
2832 |
|
|
if (replace_info->replace_p)
|
2833 |
|
|
insert_decl_map (&id, replace_info->old_tree,
|
2834 |
|
|
replace_info->new_tree);
|
2835 |
|
|
}
|
2836 |
|
|
|
2837 |
|
|
DECL_INITIAL (new_decl) = remap_blocks (DECL_INITIAL (id.src_fn), &id);
|
2838 |
|
|
|
2839 |
|
|
/* Renumber the lexical scoping (non-code) blocks consecutively. */
|
2840 |
|
|
number_blocks (id.dst_fn);
|
2841 |
|
|
|
2842 |
|
|
if (DECL_STRUCT_FUNCTION (old_decl)->unexpanded_var_list != NULL_TREE)
|
2843 |
|
|
/* Add local vars. */
|
2844 |
|
|
for (t_step = DECL_STRUCT_FUNCTION (old_decl)->unexpanded_var_list;
|
2845 |
|
|
t_step; t_step = TREE_CHAIN (t_step))
|
2846 |
|
|
{
|
2847 |
|
|
tree var = TREE_VALUE (t_step);
|
2848 |
|
|
if (TREE_STATIC (var) && !TREE_ASM_WRITTEN (var))
|
2849 |
|
|
cfun->unexpanded_var_list = tree_cons (NULL_TREE, var,
|
2850 |
|
|
cfun->unexpanded_var_list);
|
2851 |
|
|
else
|
2852 |
|
|
cfun->unexpanded_var_list =
|
2853 |
|
|
tree_cons (NULL_TREE, remap_decl (var, &id),
|
2854 |
|
|
cfun->unexpanded_var_list);
|
2855 |
|
|
}
|
2856 |
|
|
|
2857 |
|
|
/* Copy the Function's body. */
|
2858 |
|
|
old_entry_block = ENTRY_BLOCK_PTR_FOR_FUNCTION
|
2859 |
|
|
(DECL_STRUCT_FUNCTION (old_decl));
|
2860 |
|
|
new_fndecl = copy_body (&id,
|
2861 |
|
|
old_entry_block->count,
|
2862 |
|
|
old_entry_block->frequency, NULL, NULL);
|
2863 |
|
|
|
2864 |
|
|
DECL_SAVED_TREE (new_decl) = DECL_SAVED_TREE (new_fndecl);
|
2865 |
|
|
|
2866 |
|
|
DECL_STRUCT_FUNCTION (new_decl)->cfg =
|
2867 |
|
|
DECL_STRUCT_FUNCTION (new_fndecl)->cfg;
|
2868 |
|
|
DECL_STRUCT_FUNCTION (new_decl)->eh = DECL_STRUCT_FUNCTION (new_fndecl)->eh;
|
2869 |
|
|
DECL_STRUCT_FUNCTION (new_decl)->ib_boundaries_block =
|
2870 |
|
|
DECL_STRUCT_FUNCTION (new_fndecl)->ib_boundaries_block;
|
2871 |
|
|
DECL_STRUCT_FUNCTION (new_decl)->last_label_uid =
|
2872 |
|
|
DECL_STRUCT_FUNCTION (new_fndecl)->last_label_uid;
|
2873 |
|
|
|
2874 |
|
|
if (DECL_RESULT (old_decl) != NULL_TREE)
|
2875 |
|
|
{
|
2876 |
|
|
tree *res_decl = &DECL_RESULT (old_decl);
|
2877 |
|
|
DECL_RESULT (new_decl) = remap_decl (*res_decl, &id);
|
2878 |
|
|
lang_hooks.dup_lang_specific_decl (DECL_RESULT (new_decl));
|
2879 |
|
|
}
|
2880 |
|
|
|
2881 |
|
|
current_function_decl = NULL;
|
2882 |
|
|
/* Renumber the lexical scoping (non-code) blocks consecutively. */
|
2883 |
|
|
number_blocks (new_decl);
|
2884 |
|
|
|
2885 |
|
|
/* Clean up. */
|
2886 |
|
|
splay_tree_delete (id.decl_map);
|
2887 |
|
|
fold_cond_expr_cond ();
|
2888 |
|
|
return;
|
2889 |
|
|
}
|
2890 |
|
|
|
2891 |
|
|
/* Duplicate a type, fields and all. */
|
2892 |
|
|
|
2893 |
|
|
tree
|
2894 |
|
|
build_duplicate_type (tree type)
|
2895 |
|
|
{
|
2896 |
|
|
struct copy_body_data id;
|
2897 |
|
|
|
2898 |
|
|
memset (&id, 0, sizeof (id));
|
2899 |
|
|
id.src_fn = current_function_decl;
|
2900 |
|
|
id.dst_fn = current_function_decl;
|
2901 |
|
|
id.src_cfun = cfun;
|
2902 |
|
|
id.decl_map = splay_tree_new (splay_tree_compare_pointers, NULL, NULL);
|
2903 |
|
|
|
2904 |
|
|
type = remap_type_1 (type, &id);
|
2905 |
|
|
|
2906 |
|
|
splay_tree_delete (id.decl_map);
|
2907 |
|
|
|
2908 |
|
|
return type;
|
2909 |
|
|
}
|