OpenCores
URL https://opencores.org/ocsvn/openrisc/openrisc/trunk

Subversion Repositories openrisc

[/] [openrisc/] [trunk/] [gnu-dev/] [or1k-gcc/] [gcc/] [cgraphbuild.c] - Blame information for rev 775

Go to most recent revision | Details | Compare with Previous | View Log

Line No. Rev Author Line
1 684 jeremybenn
/* Callgraph construction.
2
   Copyright (C) 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010
3
   Free Software Foundation, Inc.
4
   Contributed by Jan Hubicka
5
 
6
This file is part of GCC.
7
 
8
GCC is free software; you can redistribute it and/or modify it under
9
the terms of the GNU General Public License as published by the Free
10
Software Foundation; either version 3, or (at your option) any later
11
version.
12
 
13
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
14
WARRANTY; without even the implied warranty of MERCHANTABILITY or
15
FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
16
for more details.
17
 
18
You should have received a copy of the GNU General Public License
19
along with GCC; see the file COPYING3.  If not see
20
<http://www.gnu.org/licenses/>.  */
21
 
22
#include "config.h"
23
#include "system.h"
24
#include "coretypes.h"
25
#include "tm.h"
26
#include "tree.h"
27
#include "tree-flow.h"
28
#include "langhooks.h"
29
#include "pointer-set.h"
30
#include "cgraph.h"
31
#include "intl.h"
32
#include "gimple.h"
33
#include "tree-pass.h"
34
#include "ipa-utils.h"
35
#include "except.h"
36
#include "ipa-inline.h"
37
 
38
/* Context of record_reference.  */
39
struct record_reference_ctx
40
{
41
  bool only_vars;
42
  struct varpool_node *varpool_node;
43
};
44
 
45
/* Walk tree and record all calls and references to functions/variables.
46
   Called via walk_tree: TP is pointer to tree to be examined.
47
   When DATA is non-null, record references to callgraph.
48
   */
49
 
50
static tree
51
record_reference (tree *tp, int *walk_subtrees, void *data)
52
{
53
  tree t = *tp;
54
  tree decl;
55
  struct record_reference_ctx *ctx = (struct record_reference_ctx *)data;
56
 
57
  t = canonicalize_constructor_val (t);
58
  if (!t)
59
    t = *tp;
60
  else if (t != *tp)
61
    *tp = t;
62
 
63
  switch (TREE_CODE (t))
64
    {
65
    case VAR_DECL:
66
    case FUNCTION_DECL:
67
      gcc_unreachable ();
68
      break;
69
 
70
    case FDESC_EXPR:
71
    case ADDR_EXPR:
72
      /* Record dereferences to the functions.  This makes the
73
         functions reachable unconditionally.  */
74
      decl = get_base_var (*tp);
75
      if (TREE_CODE (decl) == FUNCTION_DECL)
76
        {
77
          struct cgraph_node *node = cgraph_get_create_node (decl);
78
          if (!ctx->only_vars)
79
            cgraph_mark_address_taken_node (node);
80
          ipa_record_reference (NULL, ctx->varpool_node, node, NULL,
81
                                IPA_REF_ADDR, NULL);
82
        }
83
 
84
      if (TREE_CODE (decl) == VAR_DECL)
85
        {
86
          struct varpool_node *vnode = varpool_node (decl);
87
          if (lang_hooks.callgraph.analyze_expr)
88
            lang_hooks.callgraph.analyze_expr (&decl, walk_subtrees);
89
          varpool_mark_needed_node (vnode);
90
          ipa_record_reference (NULL, ctx->varpool_node,
91
                                NULL, vnode,
92
                                IPA_REF_ADDR, NULL);
93
        }
94
      *walk_subtrees = 0;
95
      break;
96
 
97
    default:
98
      /* Save some cycles by not walking types and declaration as we
99
         won't find anything useful there anyway.  */
100
      if (IS_TYPE_OR_DECL_P (*tp))
101
        {
102
          *walk_subtrees = 0;
103
          break;
104
        }
105
 
106
      if ((unsigned int) TREE_CODE (t) >= LAST_AND_UNUSED_TREE_CODE)
107
        return lang_hooks.callgraph.analyze_expr (tp, walk_subtrees);
108
      break;
109
    }
110
 
111
  return NULL_TREE;
112
}
113
 
114
/* Record references to typeinfos in the type list LIST.  */
115
 
116
static void
117
record_type_list (struct cgraph_node *node, tree list)
118
{
119
  for (; list; list = TREE_CHAIN (list))
120
    {
121
      tree type = TREE_VALUE (list);
122
 
123
      if (TYPE_P (type))
124
        type = lookup_type_for_runtime (type);
125
      STRIP_NOPS (type);
126
      if (TREE_CODE (type) == ADDR_EXPR)
127
        {
128
          type = TREE_OPERAND (type, 0);
129
          if (TREE_CODE (type) == VAR_DECL)
130
            {
131
              struct varpool_node *vnode = varpool_node (type);
132
              varpool_mark_needed_node (vnode);
133
              ipa_record_reference (node, NULL,
134
                                    NULL, vnode,
135
                                    IPA_REF_ADDR, NULL);
136
            }
137
        }
138
    }
139
}
140
 
141
/* Record all references we will introduce by producing EH tables
142
   for NODE.  */
143
 
144
static void
145
record_eh_tables (struct cgraph_node *node, struct function *fun)
146
{
147
  eh_region i;
148
 
149
  if (DECL_FUNCTION_PERSONALITY (node->decl))
150
    {
151
      struct cgraph_node *per_node;
152
 
153
      per_node = cgraph_get_create_node (DECL_FUNCTION_PERSONALITY (node->decl));
154
      ipa_record_reference (node, NULL, per_node, NULL, IPA_REF_ADDR, NULL);
155
      cgraph_mark_address_taken_node (per_node);
156
    }
157
 
158
  i = fun->eh->region_tree;
159
  if (!i)
160
    return;
161
 
162
  while (1)
163
    {
164
      switch (i->type)
165
        {
166
        case ERT_CLEANUP:
167
        case ERT_MUST_NOT_THROW:
168
          break;
169
 
170
        case ERT_TRY:
171
          {
172
            eh_catch c;
173
            for (c = i->u.eh_try.first_catch; c; c = c->next_catch)
174
              record_type_list (node, c->type_list);
175
          }
176
          break;
177
 
178
        case ERT_ALLOWED_EXCEPTIONS:
179
          record_type_list (node, i->u.allowed.type_list);
180
          break;
181
        }
182
      /* If there are sub-regions, process them.  */
183
      if (i->inner)
184
        i = i->inner;
185
      /* If there are peers, process them.  */
186
      else if (i->next_peer)
187
        i = i->next_peer;
188
      /* Otherwise, step back up the tree to the next peer.  */
189
      else
190
        {
191
          do
192
            {
193
              i = i->outer;
194
              if (i == NULL)
195
                return;
196
            }
197
          while (i->next_peer == NULL);
198
          i = i->next_peer;
199
        }
200
    }
201
}
202
 
203
/* Reset inlining information of all incoming call edges of NODE.  */
204
 
205
void
206
reset_inline_failed (struct cgraph_node *node)
207
{
208
  struct cgraph_edge *e;
209
 
210
  for (e = node->callers; e; e = e->next_caller)
211
    {
212
      e->callee->global.inlined_to = NULL;
213
      initialize_inline_failed (e);
214
    }
215
}
216
 
217
/* Computes the frequency of the call statement so that it can be stored in
218
   cgraph_edge.  BB is the basic block of the call statement.  */
219
int
220
compute_call_stmt_bb_frequency (tree decl, basic_block bb)
221
{
222
  int entry_freq = ENTRY_BLOCK_PTR_FOR_FUNCTION
223
                     (DECL_STRUCT_FUNCTION (decl))->frequency;
224
  int freq = bb->frequency;
225
 
226
  if (profile_status_for_function (DECL_STRUCT_FUNCTION (decl)) == PROFILE_ABSENT)
227
    return CGRAPH_FREQ_BASE;
228
 
229
  if (!entry_freq)
230
    entry_freq = 1, freq++;
231
 
232
  freq = freq * CGRAPH_FREQ_BASE / entry_freq;
233
  if (freq > CGRAPH_FREQ_MAX)
234
    freq = CGRAPH_FREQ_MAX;
235
 
236
  return freq;
237
}
238
 
239
/* Mark address taken in STMT.  */
240
 
241
static bool
242
mark_address (gimple stmt, tree addr, void *data)
243
{
244
  addr = get_base_address (addr);
245
  if (TREE_CODE (addr) == FUNCTION_DECL)
246
    {
247
      struct cgraph_node *node = cgraph_get_create_node (addr);
248
      cgraph_mark_address_taken_node (node);
249
      ipa_record_reference ((struct cgraph_node *)data, NULL,
250
                            node, NULL,
251
                            IPA_REF_ADDR, stmt);
252
    }
253
  else if (addr && TREE_CODE (addr) == VAR_DECL
254
           && (TREE_STATIC (addr) || DECL_EXTERNAL (addr)))
255
    {
256
      struct varpool_node *vnode = varpool_node (addr);
257
      int walk_subtrees;
258
 
259
      if (lang_hooks.callgraph.analyze_expr)
260
        lang_hooks.callgraph.analyze_expr (&addr, &walk_subtrees);
261
      varpool_mark_needed_node (vnode);
262
      ipa_record_reference ((struct cgraph_node *)data, NULL,
263
                            NULL, vnode,
264
                            IPA_REF_ADDR, stmt);
265
    }
266
 
267
  return false;
268
}
269
 
270
/* Mark load of T.  */
271
 
272
static bool
273
mark_load (gimple stmt, tree t, void *data)
274
{
275
  t = get_base_address (t);
276
  if (t && TREE_CODE (t) == FUNCTION_DECL)
277
    {
278
      /* ??? This can happen on platforms with descriptors when these are
279
         directly manipulated in the code.  Pretend that it's an address.  */
280
      struct cgraph_node *node = cgraph_get_create_node (t);
281
      cgraph_mark_address_taken_node (node);
282
      ipa_record_reference ((struct cgraph_node *)data, NULL,
283
                            node, NULL,
284
                            IPA_REF_ADDR, stmt);
285
    }
286
  else if (t && TREE_CODE (t) == VAR_DECL
287
           && (TREE_STATIC (t) || DECL_EXTERNAL (t)))
288
    {
289
      struct varpool_node *vnode = varpool_node (t);
290
      int walk_subtrees;
291
 
292
      if (lang_hooks.callgraph.analyze_expr)
293
        lang_hooks.callgraph.analyze_expr (&t, &walk_subtrees);
294
      varpool_mark_needed_node (vnode);
295
      ipa_record_reference ((struct cgraph_node *)data, NULL,
296
                            NULL, vnode,
297
                            IPA_REF_LOAD, stmt);
298
    }
299
  return false;
300
}
301
 
302
/* Mark store of T.  */
303
 
304
static bool
305
mark_store (gimple stmt, tree t, void *data)
306
{
307
  t = get_base_address (t);
308
  if (t && TREE_CODE (t) == VAR_DECL
309
      && (TREE_STATIC (t) || DECL_EXTERNAL (t)))
310
    {
311
      struct varpool_node *vnode = varpool_node (t);
312
      int walk_subtrees;
313
 
314
      if (lang_hooks.callgraph.analyze_expr)
315
        lang_hooks.callgraph.analyze_expr (&t, &walk_subtrees);
316
      varpool_mark_needed_node (vnode);
317
      ipa_record_reference ((struct cgraph_node *)data, NULL,
318
                            NULL, vnode,
319
                            IPA_REF_STORE, stmt);
320
     }
321
  return false;
322
}
323
 
324
/* Create cgraph edges for function calls.
325
   Also look for functions and variables having addresses taken.  */
326
 
327
static unsigned int
328
build_cgraph_edges (void)
329
{
330
  basic_block bb;
331
  struct cgraph_node *node = cgraph_get_node (current_function_decl);
332
  struct pointer_set_t *visited_nodes = pointer_set_create ();
333
  gimple_stmt_iterator gsi;
334
  tree decl;
335
  unsigned ix;
336
 
337
  /* Create the callgraph edges and record the nodes referenced by the function.
338
     body.  */
339
  FOR_EACH_BB (bb)
340
    {
341
      for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
342
        {
343
          gimple stmt = gsi_stmt (gsi);
344
          tree decl;
345
 
346
          if (is_gimple_call (stmt))
347
            {
348
              int freq = compute_call_stmt_bb_frequency (current_function_decl,
349
                                                         bb);
350
              decl = gimple_call_fndecl (stmt);
351
              if (decl)
352
                cgraph_create_edge (node, cgraph_get_create_node (decl),
353
                                    stmt, bb->count, freq);
354
              else
355
                cgraph_create_indirect_edge (node, stmt,
356
                                             gimple_call_flags (stmt),
357
                                             bb->count, freq);
358
            }
359
          walk_stmt_load_store_addr_ops (stmt, node, mark_load,
360
                                         mark_store, mark_address);
361
          if (gimple_code (stmt) == GIMPLE_OMP_PARALLEL
362
              && gimple_omp_parallel_child_fn (stmt))
363
            {
364
              tree fn = gimple_omp_parallel_child_fn (stmt);
365
              ipa_record_reference (node, NULL, cgraph_get_create_node (fn),
366
                                    NULL, IPA_REF_ADDR, stmt);
367
            }
368
          if (gimple_code (stmt) == GIMPLE_OMP_TASK)
369
            {
370
              tree fn = gimple_omp_task_child_fn (stmt);
371
              if (fn)
372
                ipa_record_reference (node, NULL, cgraph_get_create_node (fn),
373
                                      NULL, IPA_REF_ADDR, stmt);
374
              fn = gimple_omp_task_copy_fn (stmt);
375
              if (fn)
376
                ipa_record_reference (node, NULL, cgraph_get_create_node (fn),
377
                                      NULL, IPA_REF_ADDR, stmt);
378
            }
379
        }
380
      for (gsi = gsi_start (phi_nodes (bb)); !gsi_end_p (gsi); gsi_next (&gsi))
381
        walk_stmt_load_store_addr_ops (gsi_stmt (gsi), node,
382
                                       mark_load, mark_store, mark_address);
383
   }
384
 
385
  /* Look for initializers of constant variables and private statics.  */
386
  FOR_EACH_LOCAL_DECL (cfun, ix, decl)
387
    if (TREE_CODE (decl) == VAR_DECL
388
        && (TREE_STATIC (decl) && !DECL_EXTERNAL (decl)))
389
      varpool_finalize_decl (decl);
390
  record_eh_tables (node, cfun);
391
 
392
  pointer_set_destroy (visited_nodes);
393
  return 0;
394
}
395
 
396
struct gimple_opt_pass pass_build_cgraph_edges =
397
{
398
 {
399
  GIMPLE_PASS,
400
  "*build_cgraph_edges",                        /* name */
401
  NULL,                                 /* gate */
402
  build_cgraph_edges,                   /* execute */
403
  NULL,                                 /* sub */
404
  NULL,                                 /* next */
405
  0,                                     /* static_pass_number */
406
  TV_NONE,                              /* tv_id */
407
  PROP_cfg,                             /* properties_required */
408
  0,                                     /* properties_provided */
409
  0,                                     /* properties_destroyed */
410
  0,                                     /* todo_flags_start */
411
 
412
 }
413
};
414
 
415
/* Record references to functions and other variables present in the
416
   initial value of DECL, a variable.
417
   When ONLY_VARS is true, we mark needed only variables, not functions.  */
418
 
419
void
420
record_references_in_initializer (tree decl, bool only_vars)
421
{
422
  struct pointer_set_t *visited_nodes = pointer_set_create ();
423
  struct varpool_node *node = varpool_node (decl);
424
  struct record_reference_ctx ctx = {false, NULL};
425
 
426
  ctx.varpool_node = node;
427
  ctx.only_vars = only_vars;
428
  walk_tree (&DECL_INITIAL (decl), record_reference,
429
             &ctx, visited_nodes);
430
  pointer_set_destroy (visited_nodes);
431
}
432
 
433
/* Rebuild cgraph edges for current function node.  This needs to be run after
434
   passes that don't update the cgraph.  */
435
 
436
unsigned int
437
rebuild_cgraph_edges (void)
438
{
439
  basic_block bb;
440
  struct cgraph_node *node = cgraph_get_node (current_function_decl);
441
  gimple_stmt_iterator gsi;
442
 
443
  cgraph_node_remove_callees (node);
444
  ipa_remove_all_references (&node->ref_list);
445
 
446
  node->count = ENTRY_BLOCK_PTR->count;
447
 
448
  FOR_EACH_BB (bb)
449
    {
450
      for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
451
        {
452
          gimple stmt = gsi_stmt (gsi);
453
          tree decl;
454
 
455
          if (is_gimple_call (stmt))
456
            {
457
              int freq = compute_call_stmt_bb_frequency (current_function_decl,
458
                                                         bb);
459
              decl = gimple_call_fndecl (stmt);
460
              if (decl)
461
                cgraph_create_edge (node, cgraph_get_create_node (decl), stmt,
462
                                    bb->count, freq);
463
              else
464
                cgraph_create_indirect_edge (node, stmt,
465
                                             gimple_call_flags (stmt),
466
                                             bb->count, freq);
467
            }
468
          walk_stmt_load_store_addr_ops (stmt, node, mark_load,
469
                                         mark_store, mark_address);
470
 
471
        }
472
      for (gsi = gsi_start (phi_nodes (bb)); !gsi_end_p (gsi); gsi_next (&gsi))
473
        walk_stmt_load_store_addr_ops (gsi_stmt (gsi), node,
474
                                       mark_load, mark_store, mark_address);
475
    }
476
  record_eh_tables (node, cfun);
477
  gcc_assert (!node->global.inlined_to);
478
 
479
  return 0;
480
}
481
 
482
/* Rebuild cgraph edges for current function node.  This needs to be run after
483
   passes that don't update the cgraph.  */
484
 
485
void
486
cgraph_rebuild_references (void)
487
{
488
  basic_block bb;
489
  struct cgraph_node *node = cgraph_get_node (current_function_decl);
490
  gimple_stmt_iterator gsi;
491
 
492
  ipa_remove_all_references (&node->ref_list);
493
 
494
  node->count = ENTRY_BLOCK_PTR->count;
495
 
496
  FOR_EACH_BB (bb)
497
    {
498
      for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
499
        {
500
          gimple stmt = gsi_stmt (gsi);
501
 
502
          walk_stmt_load_store_addr_ops (stmt, node, mark_load,
503
                                         mark_store, mark_address);
504
 
505
        }
506
      for (gsi = gsi_start (phi_nodes (bb)); !gsi_end_p (gsi); gsi_next (&gsi))
507
        walk_stmt_load_store_addr_ops (gsi_stmt (gsi), node,
508
                                       mark_load, mark_store, mark_address);
509
    }
510
  record_eh_tables (node, cfun);
511
}
512
 
513
struct gimple_opt_pass pass_rebuild_cgraph_edges =
514
{
515
 {
516
  GIMPLE_PASS,
517
  "*rebuild_cgraph_edges",              /* name */
518
  NULL,                                 /* gate */
519
  rebuild_cgraph_edges,                 /* execute */
520
  NULL,                                 /* sub */
521
  NULL,                                 /* next */
522
  0,                                     /* static_pass_number */
523
  TV_CGRAPH,                            /* tv_id */
524
  PROP_cfg,                             /* properties_required */
525
  0,                                     /* properties_provided */
526
  0,                                     /* properties_destroyed */
527
  0,                                     /* todo_flags_start */
528
  0,                                     /* todo_flags_finish */
529
 }
530
};
531
 
532
 
533
static unsigned int
534
remove_cgraph_callee_edges (void)
535
{
536
  cgraph_node_remove_callees (cgraph_get_node (current_function_decl));
537
  return 0;
538
}
539
 
540
struct gimple_opt_pass pass_remove_cgraph_callee_edges =
541
{
542
 {
543
  GIMPLE_PASS,
544
  "*remove_cgraph_callee_edges",                /* name */
545
  NULL,                                 /* gate */
546
  remove_cgraph_callee_edges,           /* execute */
547
  NULL,                                 /* sub */
548
  NULL,                                 /* next */
549
  0,                                     /* static_pass_number */
550
  TV_NONE,                              /* tv_id */
551
  0,                                     /* properties_required */
552
  0,                                     /* properties_provided */
553
  0,                                     /* properties_destroyed */
554
  0,                                     /* todo_flags_start */
555
  0,                                     /* todo_flags_finish */
556
 }
557
};

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.