URL
https://opencores.org/ocsvn/openrisc_me/openrisc_me/trunk
Subversion Repositories openrisc_me
[/] [openrisc/] [trunk/] [gnu-src/] [gcc-4.2.2/] [gcc/] [tree-ssa-live.c] - Rev 38
Go to most recent revision | Compare with Previous | Blame | View Log
/* Liveness for SSA trees. Copyright (C) 2003, 2004, 2005, 2007 Free Software Foundation, Inc. Contributed by Andrew MacLeod <amacleod@redhat.com> This file is part of GCC. GCC is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3, or (at your option) any later version. GCC is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with GCC; see the file COPYING3. If not see <http://www.gnu.org/licenses/>. */ #include "config.h" #include "system.h" #include "coretypes.h" #include "tm.h" #include "tree.h" #include "flags.h" #include "basic-block.h" #include "function.h" #include "diagnostic.h" #include "bitmap.h" #include "tree-flow.h" #include "tree-gimple.h" #include "tree-inline.h" #include "varray.h" #include "timevar.h" #include "hashtab.h" #include "tree-dump.h" #include "tree-ssa-live.h" #include "toplev.h" #include "vecprim.h" static void live_worklist (tree_live_info_p, int *, int); static tree_live_info_p new_tree_live_info (var_map); static inline void set_if_valid (var_map, bitmap, tree); static inline void add_livein_if_notdef (tree_live_info_p, bitmap, tree, basic_block); static inline void register_ssa_partition (var_map, tree, bool); static inline void add_conflicts_if_valid (tpa_p, conflict_graph, var_map, bitmap, tree); static partition_pair_p find_partition_pair (coalesce_list_p, int, int, bool); /* This is where the mapping from SSA version number to real storage variable is tracked. All SSA versions of the same variable may not ultimately be mapped back to the same real variable. In that instance, we need to detect the live range overlap, and give one of the variable new storage. The vector 'partition_to_var' tracks which partition maps to which variable. Given a VAR, it is sometimes desirable to know which partition that VAR represents. There is an additional field in the variable annotation to track that information. */ /* Create a variable partition map of SIZE, initialize and return it. */ var_map init_var_map (int size) { var_map map; map = (var_map) xmalloc (sizeof (struct _var_map)); map->var_partition = partition_new (size); map->partition_to_var = (tree *)xmalloc (size * sizeof (tree)); memset (map->partition_to_var, 0, size * sizeof (tree)); map->partition_to_compact = NULL; map->compact_to_partition = NULL; map->num_partitions = size; map->partition_size = size; map->ref_count = NULL; return map; } /* Free memory associated with MAP. */ void delete_var_map (var_map map) { free (map->partition_to_var); partition_delete (map->var_partition); if (map->partition_to_compact) free (map->partition_to_compact); if (map->compact_to_partition) free (map->compact_to_partition); if (map->ref_count) free (map->ref_count); free (map); } /* This function will combine the partitions in MAP for VAR1 and VAR2. It Returns the partition which represents the new partition. If the two partitions cannot be combined, NO_PARTITION is returned. */ int var_union (var_map map, tree var1, tree var2) { int p1, p2, p3; tree root_var = NULL_TREE; tree other_var = NULL_TREE; /* This is independent of partition_to_compact. If partition_to_compact is on, then whichever one of these partitions is absorbed will never have a dereference into the partition_to_compact array any more. */ if (TREE_CODE (var1) == SSA_NAME) p1 = partition_find (map->var_partition, SSA_NAME_VERSION (var1)); else { p1 = var_to_partition (map, var1); if (map->compact_to_partition) p1 = map->compact_to_partition[p1]; root_var = var1; } if (TREE_CODE (var2) == SSA_NAME) p2 = partition_find (map->var_partition, SSA_NAME_VERSION (var2)); else { p2 = var_to_partition (map, var2); if (map->compact_to_partition) p2 = map->compact_to_partition[p2]; /* If there is no root_var set, or it's not a user variable, set the root_var to this one. */ if (!root_var || (DECL_P (root_var) && DECL_IGNORED_P (root_var))) { other_var = root_var; root_var = var2; } else other_var = var2; } gcc_assert (p1 != NO_PARTITION); gcc_assert (p2 != NO_PARTITION); if (p1 == p2) p3 = p1; else p3 = partition_union (map->var_partition, p1, p2); if (map->partition_to_compact) p3 = map->partition_to_compact[p3]; if (root_var) change_partition_var (map, root_var, p3); if (other_var) change_partition_var (map, other_var, p3); return p3; } /* Compress the partition numbers in MAP such that they fall in the range 0..(num_partitions-1) instead of wherever they turned out during the partitioning exercise. This removes any references to unused partitions, thereby allowing bitmaps and other vectors to be much denser. Compression type is controlled by FLAGS. This is implemented such that compaction doesn't affect partitioning. Ie., once partitions are created and possibly merged, running one or more different kind of compaction will not affect the partitions themselves. Their index might change, but all the same variables will still be members of the same partition group. This allows work on reduced sets, and no loss of information when a larger set is later desired. In particular, coalescing can work on partitions which have 2 or more definitions, and then 'recompact' later to include all the single definitions for assignment to program variables. */ void compact_var_map (var_map map, int flags) { sbitmap used; int tmp, root, root_i; unsigned int x, limit, count; tree var; root_var_p rv = NULL; limit = map->partition_size; used = sbitmap_alloc (limit); sbitmap_zero (used); /* Already compressed? Abandon the old one. */ if (map->partition_to_compact) { free (map->partition_to_compact); map->partition_to_compact = NULL; } if (map->compact_to_partition) { free (map->compact_to_partition); map->compact_to_partition = NULL; } map->num_partitions = map->partition_size; if (flags & VARMAP_NO_SINGLE_DEFS) rv = root_var_init (map); map->partition_to_compact = (int *)xmalloc (limit * sizeof (int)); memset (map->partition_to_compact, 0xff, (limit * sizeof (int))); /* Find out which partitions are actually referenced. */ count = 0; for (x = 0; x < limit; x++) { tmp = partition_find (map->var_partition, x); if (!TEST_BIT (used, tmp) && map->partition_to_var[tmp] != NULL_TREE) { /* It is referenced, check to see if there is more than one version in the root_var table, if one is available. */ if (rv) { root = root_var_find (rv, tmp); root_i = root_var_first_partition (rv, root); /* If there is only one, don't include this in the compaction. */ if (root_var_next_partition (rv, root_i) == ROOT_VAR_NONE) continue; } SET_BIT (used, tmp); count++; } } /* Build a compacted partitioning. */ if (count != limit) { sbitmap_iterator sbi; map->compact_to_partition = (int *)xmalloc (count * sizeof (int)); count = 0; /* SSA renaming begins at 1, so skip 0 when compacting. */ EXECUTE_IF_SET_IN_SBITMAP (used, 1, x, sbi) { map->partition_to_compact[x] = count; map->compact_to_partition[count] = x; var = map->partition_to_var[x]; if (TREE_CODE (var) != SSA_NAME) change_partition_var (map, var, count); count++; } } else { free (map->partition_to_compact); map->partition_to_compact = NULL; } map->num_partitions = count; if (rv) root_var_delete (rv); sbitmap_free (used); } /* This function is used to change the representative variable in MAP for VAR's partition from an SSA_NAME variable to a regular variable. This allows partitions to be mapped back to real variables. */ void change_partition_var (var_map map, tree var, int part) { var_ann_t ann; gcc_assert (TREE_CODE (var) != SSA_NAME); ann = var_ann (var); ann->out_of_ssa_tag = 1; VAR_ANN_PARTITION (ann) = part; if (map->compact_to_partition) map->partition_to_var[map->compact_to_partition[part]] = var; } static inline void mark_all_vars_used (tree *); /* Helper function for mark_all_vars_used, called via walk_tree. */ static tree mark_all_vars_used_1 (tree *tp, int *walk_subtrees, void *data ATTRIBUTE_UNUSED) { tree t = *tp; if (TREE_CODE (t) == SSA_NAME) t = SSA_NAME_VAR (t); /* Ignore TREE_ORIGINAL for TARGET_MEM_REFS, as well as other fields that do not contain vars. */ if (TREE_CODE (t) == TARGET_MEM_REF) { mark_all_vars_used (&TMR_SYMBOL (t)); mark_all_vars_used (&TMR_BASE (t)); mark_all_vars_used (&TMR_INDEX (t)); *walk_subtrees = 0; return NULL; } /* Only need to mark VAR_DECLS; parameters and return results are not eliminated as unused. */ if (TREE_CODE (t) == VAR_DECL) set_is_used (t); if (IS_TYPE_OR_DECL_P (t)) *walk_subtrees = 0; return NULL; } /* Mark all VAR_DECLS under *EXPR_P as used, so that they won't be eliminated during the tree->rtl conversion process. */ static inline void mark_all_vars_used (tree *expr_p) { walk_tree (expr_p, mark_all_vars_used_1, NULL, NULL); } /* Remove local variables that are not referenced in the IL. */ void remove_unused_locals (void) { basic_block bb; tree t, *cell; /* Assume all locals are unused. */ for (t = cfun->unexpanded_var_list; t; t = TREE_CHAIN (t)) { tree var = TREE_VALUE (t); if (TREE_CODE (var) != FUNCTION_DECL && var_ann (var)) var_ann (var)->used = false; } /* Walk the CFG marking all referenced symbols. */ FOR_EACH_BB (bb) { block_stmt_iterator bsi; tree phi, def; /* Walk the statements. */ for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi)) mark_all_vars_used (bsi_stmt_ptr (bsi)); for (phi = phi_nodes (bb); phi; phi = PHI_CHAIN (phi)) { use_operand_p arg_p; ssa_op_iter i; /* No point processing globals. */ if (is_global_var (SSA_NAME_VAR (PHI_RESULT (phi)))) continue; def = PHI_RESULT (phi); mark_all_vars_used (&def); FOR_EACH_PHI_ARG (arg_p, phi, i, SSA_OP_ALL_USES) { tree arg = USE_FROM_PTR (arg_p); mark_all_vars_used (&arg); } } } /* Remove unmarked vars and clear used flag. */ for (cell = &cfun->unexpanded_var_list; *cell; ) { tree var = TREE_VALUE (*cell); var_ann_t ann; if (TREE_CODE (var) != FUNCTION_DECL && (!(ann = var_ann (var)) || !ann->used)) { *cell = TREE_CHAIN (*cell); continue; } cell = &TREE_CHAIN (*cell); } } /* This function looks through the program and uses FLAGS to determine what SSA versioned variables are given entries in a new partition table. This new partition map is returned. */ var_map create_ssa_var_map (int flags) { block_stmt_iterator bsi; basic_block bb; tree dest, use; tree stmt; var_map map; ssa_op_iter iter; #ifdef ENABLE_CHECKING bitmap used_in_real_ops; bitmap used_in_virtual_ops; #endif map = init_var_map (num_ssa_names + 1); #ifdef ENABLE_CHECKING used_in_real_ops = BITMAP_ALLOC (NULL); used_in_virtual_ops = BITMAP_ALLOC (NULL); #endif if (flags & SSA_VAR_MAP_REF_COUNT) { map->ref_count = (int *)xmalloc (((num_ssa_names + 1) * sizeof (int))); memset (map->ref_count, 0, (num_ssa_names + 1) * sizeof (int)); } FOR_EACH_BB (bb) { tree phi, arg; for (phi = phi_nodes (bb); phi; phi = PHI_CHAIN (phi)) { int i; register_ssa_partition (map, PHI_RESULT (phi), false); for (i = 0; i < PHI_NUM_ARGS (phi); i++) { arg = PHI_ARG_DEF (phi, i); if (TREE_CODE (arg) == SSA_NAME) register_ssa_partition (map, arg, true); mark_all_vars_used (&PHI_ARG_DEF_TREE (phi, i)); } } for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi)) { stmt = bsi_stmt (bsi); /* Register USE and DEF operands in each statement. */ FOR_EACH_SSA_TREE_OPERAND (use , stmt, iter, SSA_OP_USE) { register_ssa_partition (map, use, true); #ifdef ENABLE_CHECKING bitmap_set_bit (used_in_real_ops, DECL_UID (SSA_NAME_VAR (use))); #endif } FOR_EACH_SSA_TREE_OPERAND (dest, stmt, iter, SSA_OP_DEF) { register_ssa_partition (map, dest, false); #ifdef ENABLE_CHECKING bitmap_set_bit (used_in_real_ops, DECL_UID (SSA_NAME_VAR (dest))); #endif } #ifdef ENABLE_CHECKING /* Validate that virtual ops don't get used in funny ways. */ FOR_EACH_SSA_TREE_OPERAND (use, stmt, iter, SSA_OP_VIRTUAL_USES | SSA_OP_VMUSTDEF) { bitmap_set_bit (used_in_virtual_ops, DECL_UID (SSA_NAME_VAR (use))); } #endif /* ENABLE_CHECKING */ mark_all_vars_used (bsi_stmt_ptr (bsi)); } } #if defined ENABLE_CHECKING { unsigned i; bitmap both = BITMAP_ALLOC (NULL); bitmap_and (both, used_in_real_ops, used_in_virtual_ops); if (!bitmap_empty_p (both)) { bitmap_iterator bi; EXECUTE_IF_SET_IN_BITMAP (both, 0, i, bi) fprintf (stderr, "Variable %s used in real and virtual operands\n", get_name (referenced_var (i))); internal_error ("SSA corruption"); } BITMAP_FREE (used_in_real_ops); BITMAP_FREE (used_in_virtual_ops); BITMAP_FREE (both); } #endif return map; } /* Allocate and return a new live range information object base on MAP. */ static tree_live_info_p new_tree_live_info (var_map map) { tree_live_info_p live; unsigned x; live = (tree_live_info_p) xmalloc (sizeof (struct tree_live_info_d)); live->map = map; live->num_blocks = last_basic_block; live->global = BITMAP_ALLOC (NULL); live->livein = (bitmap *)xmalloc (num_var_partitions (map) * sizeof (bitmap)); for (x = 0; x < num_var_partitions (map); x++) live->livein[x] = BITMAP_ALLOC (NULL); /* liveout is deferred until it is actually requested. */ live->liveout = NULL; return live; } /* Free storage for live range info object LIVE. */ void delete_tree_live_info (tree_live_info_p live) { int x; if (live->liveout) { for (x = live->num_blocks - 1; x >= 0; x--) BITMAP_FREE (live->liveout[x]); free (live->liveout); } if (live->livein) { for (x = num_var_partitions (live->map) - 1; x >= 0; x--) BITMAP_FREE (live->livein[x]); free (live->livein); } if (live->global) BITMAP_FREE (live->global); free (live); } /* Using LIVE, fill in all the live-on-entry blocks between the defs and uses for partition I. STACK is a varray used for temporary memory which is passed in rather than being allocated on every call. */ static void live_worklist (tree_live_info_p live, int *stack, int i) { unsigned b; tree var; basic_block def_bb = NULL; edge e; var_map map = live->map; edge_iterator ei; bitmap_iterator bi; int *tos = stack; var = partition_to_var (map, i); if (SSA_NAME_DEF_STMT (var)) def_bb = bb_for_stmt (SSA_NAME_DEF_STMT (var)); EXECUTE_IF_SET_IN_BITMAP (live->livein[i], 0, b, bi) { *tos++ = b; } while (tos != stack) { b = *--tos; FOR_EACH_EDGE (e, ei, BASIC_BLOCK (b)->preds) if (e->src != ENTRY_BLOCK_PTR) { /* Its not live on entry to the block its defined in. */ if (e->src == def_bb) continue; if (!bitmap_bit_p (live->livein[i], e->src->index)) { bitmap_set_bit (live->livein[i], e->src->index); *tos++ = e->src->index; } } } } /* If VAR is in a partition of MAP, set the bit for that partition in VEC. */ static inline void set_if_valid (var_map map, bitmap vec, tree var) { int p = var_to_partition (map, var); if (p != NO_PARTITION) bitmap_set_bit (vec, p); } /* If VAR is in a partition and it isn't defined in DEF_VEC, set the livein and global bit for it in the LIVE object. BB is the block being processed. */ static inline void add_livein_if_notdef (tree_live_info_p live, bitmap def_vec, tree var, basic_block bb) { int p = var_to_partition (live->map, var); if (p == NO_PARTITION || bb == ENTRY_BLOCK_PTR) return; if (!bitmap_bit_p (def_vec, p)) { bitmap_set_bit (live->livein[p], bb->index); bitmap_set_bit (live->global, p); } } /* Given partition map MAP, calculate all the live on entry bitmaps for each basic block. Return a live info object. */ tree_live_info_p calculate_live_on_entry (var_map map) { tree_live_info_p live; unsigned i; basic_block bb; bitmap saw_def; tree phi, var, stmt; tree op; edge e; int *stack; block_stmt_iterator bsi; ssa_op_iter iter; bitmap_iterator bi; #ifdef ENABLE_CHECKING int num; edge_iterator ei; #endif saw_def = BITMAP_ALLOC (NULL); live = new_tree_live_info (map); FOR_EACH_BB (bb) { bitmap_clear (saw_def); for (phi = phi_nodes (bb); phi; phi = PHI_CHAIN (phi)) { for (i = 0; i < (unsigned)PHI_NUM_ARGS (phi); i++) { var = PHI_ARG_DEF (phi, i); if (!phi_ssa_name_p (var)) continue; stmt = SSA_NAME_DEF_STMT (var); e = EDGE_PRED (bb, i); /* Any uses in PHIs which either don't have def's or are not defined in the block from which the def comes, will be live on entry to that block. */ if (!stmt || e->src != bb_for_stmt (stmt)) add_livein_if_notdef (live, saw_def, var, e->src); } } /* Don't mark PHI results as defined until all the PHI nodes have been processed. If the PHI sequence is: a_3 = PHI <a_1, a_2> b_3 = PHI <b_1, a_3> The a_3 referred to in b_3's PHI node is the one incoming on the edge, *not* the PHI node just seen. */ for (phi = phi_nodes (bb); phi; phi = PHI_CHAIN (phi)) { var = PHI_RESULT (phi); set_if_valid (map, saw_def, var); } for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi)) { stmt = bsi_stmt (bsi); FOR_EACH_SSA_TREE_OPERAND (op, stmt, iter, SSA_OP_USE) { add_livein_if_notdef (live, saw_def, op, bb); } FOR_EACH_SSA_TREE_OPERAND (op, stmt, iter, SSA_OP_DEF) { set_if_valid (map, saw_def, op); } } } stack = XNEWVEC (int, last_basic_block); EXECUTE_IF_SET_IN_BITMAP (live->global, 0, i, bi) { live_worklist (live, stack, i); } free (stack); #ifdef ENABLE_CHECKING /* Check for live on entry partitions and report those with a DEF in the program. This will typically mean an optimization has done something wrong. */ bb = ENTRY_BLOCK_PTR; num = 0; FOR_EACH_EDGE (e, ei, bb->succs) { int entry_block = e->dest->index; if (e->dest == EXIT_BLOCK_PTR) continue; for (i = 0; i < (unsigned)num_var_partitions (map); i++) { basic_block tmp; tree d; var = partition_to_var (map, i); stmt = SSA_NAME_DEF_STMT (var); tmp = bb_for_stmt (stmt); d = default_def (SSA_NAME_VAR (var)); if (bitmap_bit_p (live_entry_blocks (live, i), entry_block)) { if (!IS_EMPTY_STMT (stmt)) { num++; print_generic_expr (stderr, var, TDF_SLIM); fprintf (stderr, " is defined "); if (tmp) fprintf (stderr, " in BB%d, ", tmp->index); fprintf (stderr, "by:\n"); print_generic_expr (stderr, stmt, TDF_SLIM); fprintf (stderr, "\nIt is also live-on-entry to entry BB %d", entry_block); fprintf (stderr, " So it appears to have multiple defs.\n"); } else { if (d != var) { num++; print_generic_expr (stderr, var, TDF_SLIM); fprintf (stderr, " is live-on-entry to BB%d ",entry_block); if (d) { fprintf (stderr, " but is not the default def of "); print_generic_expr (stderr, d, TDF_SLIM); fprintf (stderr, "\n"); } else fprintf (stderr, " and there is no default def.\n"); } } } else if (d == var) { /* The only way this var shouldn't be marked live on entry is if it occurs in a PHI argument of the block. */ int z, ok = 0; for (phi = phi_nodes (e->dest); phi && !ok; phi = PHI_CHAIN (phi)) { for (z = 0; z < PHI_NUM_ARGS (phi); z++) if (var == PHI_ARG_DEF (phi, z)) { ok = 1; break; } } if (ok) continue; num++; print_generic_expr (stderr, var, TDF_SLIM); fprintf (stderr, " is not marked live-on-entry to entry BB%d ", entry_block); fprintf (stderr, "but it is a default def so it should be.\n"); } } } gcc_assert (num <= 0); #endif BITMAP_FREE (saw_def); return live; } /* Calculate the live on exit vectors based on the entry info in LIVEINFO. */ void calculate_live_on_exit (tree_live_info_p liveinfo) { unsigned b; unsigned i, x; bitmap *on_exit; basic_block bb; edge e; tree t, phi; bitmap on_entry; var_map map = liveinfo->map; on_exit = (bitmap *)xmalloc (last_basic_block * sizeof (bitmap)); for (x = 0; x < (unsigned)last_basic_block; x++) on_exit[x] = BITMAP_ALLOC (NULL); /* Set all the live-on-exit bits for uses in PHIs. */ FOR_EACH_BB (bb) { for (phi = phi_nodes (bb); phi; phi = PHI_CHAIN (phi)) for (i = 0; i < (unsigned)PHI_NUM_ARGS (phi); i++) { t = PHI_ARG_DEF (phi, i); e = PHI_ARG_EDGE (phi, i); if (!phi_ssa_name_p (t) || e->src == ENTRY_BLOCK_PTR) continue; set_if_valid (map, on_exit[e->src->index], t); } } /* Set live on exit for all predecessors of live on entry's. */ for (i = 0; i < num_var_partitions (map); i++) { bitmap_iterator bi; on_entry = live_entry_blocks (liveinfo, i); EXECUTE_IF_SET_IN_BITMAP (on_entry, 0, b, bi) { edge_iterator ei; FOR_EACH_EDGE (e, ei, BASIC_BLOCK (b)->preds) if (e->src != ENTRY_BLOCK_PTR) bitmap_set_bit (on_exit[e->src->index], i); } } liveinfo->liveout = on_exit; } /* Initialize a tree_partition_associator object using MAP. */ static tpa_p tpa_init (var_map map) { tpa_p tpa; int num_partitions = num_var_partitions (map); int x; if (num_partitions == 0) return NULL; tpa = (tpa_p) xmalloc (sizeof (struct tree_partition_associator_d)); tpa->num_trees = 0; tpa->uncompressed_num = -1; tpa->map = map; tpa->next_partition = (int *)xmalloc (num_partitions * sizeof (int)); memset (tpa->next_partition, TPA_NONE, num_partitions * sizeof (int)); tpa->partition_to_tree_map = (int *)xmalloc (num_partitions * sizeof (int)); memset (tpa->partition_to_tree_map, TPA_NONE, num_partitions * sizeof (int)); x = MAX (40, (num_partitions / 20)); tpa->trees = VEC_alloc (tree, heap, x); tpa->first_partition = VEC_alloc (int, heap, x); return tpa; } /* Remove PARTITION_INDEX from TREE_INDEX's list in the tpa structure TPA. */ void tpa_remove_partition (tpa_p tpa, int tree_index, int partition_index) { int i; i = tpa_first_partition (tpa, tree_index); if (i == partition_index) { VEC_replace (int, tpa->first_partition, tree_index, tpa->next_partition[i]); } else { for ( ; i != TPA_NONE; i = tpa_next_partition (tpa, i)) { if (tpa->next_partition[i] == partition_index) { tpa->next_partition[i] = tpa->next_partition[partition_index]; break; } } } } /* Free the memory used by tree_partition_associator object TPA. */ void tpa_delete (tpa_p tpa) { if (!tpa) return; VEC_free (tree, heap, tpa->trees); VEC_free (int, heap, tpa->first_partition); free (tpa->partition_to_tree_map); free (tpa->next_partition); free (tpa); } /* This function will remove any tree entries from TPA which have only a single element. This will help keep the size of the conflict graph down. The function returns the number of remaining tree lists. */ int tpa_compact (tpa_p tpa) { int last, x, y, first, swap_i; tree swap_t; /* Find the last list which has more than 1 partition. */ for (last = tpa->num_trees - 1; last > 0; last--) { first = tpa_first_partition (tpa, last); if (tpa_next_partition (tpa, first) != NO_PARTITION) break; } x = 0; while (x < last) { first = tpa_first_partition (tpa, x); /* If there is not more than one partition, swap with the current end of the tree list. */ if (tpa_next_partition (tpa, first) == NO_PARTITION) { swap_t = VEC_index (tree, tpa->trees, last); swap_i = VEC_index (int, tpa->first_partition, last); /* Update the last entry. Since it is known to only have one partition, there is nothing else to update. */ VEC_replace (tree, tpa->trees, last, VEC_index (tree, tpa->trees, x)); VEC_replace (int, tpa->first_partition, last, VEC_index (int, tpa->first_partition, x)); tpa->partition_to_tree_map[tpa_first_partition (tpa, last)] = last; /* Since this list is known to have more than one partition, update the list owner entries. */ VEC_replace (tree, tpa->trees, x, swap_t); VEC_replace (int, tpa->first_partition, x, swap_i); for (y = tpa_first_partition (tpa, x); y != NO_PARTITION; y = tpa_next_partition (tpa, y)) tpa->partition_to_tree_map[y] = x; /* Ensure last is a list with more than one partition. */ last--; for (; last > x; last--) { first = tpa_first_partition (tpa, last); if (tpa_next_partition (tpa, first) != NO_PARTITION) break; } } x++; } first = tpa_first_partition (tpa, x); if (tpa_next_partition (tpa, first) != NO_PARTITION) x++; tpa->uncompressed_num = tpa->num_trees; tpa->num_trees = x; return last; } /* Initialize a root_var object with SSA partitions from MAP which are based on each root variable. */ root_var_p root_var_init (var_map map) { root_var_p rv; int num_partitions = num_var_partitions (map); int x, p; tree t; var_ann_t ann; sbitmap seen; rv = tpa_init (map); if (!rv) return NULL; seen = sbitmap_alloc (num_partitions); sbitmap_zero (seen); /* Start at the end and work towards the front. This will provide a list that is ordered from smallest to largest. */ for (x = num_partitions - 1; x >= 0; x--) { t = partition_to_var (map, x); /* The var map may not be compacted yet, so check for NULL. */ if (!t) continue; p = var_to_partition (map, t); gcc_assert (p != NO_PARTITION); /* Make sure we only put coalesced partitions into the list once. */ if (TEST_BIT (seen, p)) continue; SET_BIT (seen, p); if (TREE_CODE (t) == SSA_NAME) t = SSA_NAME_VAR (t); ann = var_ann (t); if (ann->root_var_processed) { rv->next_partition[p] = VEC_index (int, rv->first_partition, VAR_ANN_ROOT_INDEX (ann)); VEC_replace (int, rv->first_partition, VAR_ANN_ROOT_INDEX (ann), p); } else { ann->root_var_processed = 1; VAR_ANN_ROOT_INDEX (ann) = rv->num_trees++; VEC_safe_push (tree, heap, rv->trees, t); VEC_safe_push (int, heap, rv->first_partition, p); } rv->partition_to_tree_map[p] = VAR_ANN_ROOT_INDEX (ann); } /* Reset the out_of_ssa_tag flag on each variable for later use. */ for (x = 0; x < rv->num_trees; x++) { t = VEC_index (tree, rv->trees, x); var_ann (t)->root_var_processed = 0; } sbitmap_free (seen); return rv; } /* Initialize a type_var structure which associates all the partitions in MAP of the same type to the type node's index. Volatiles are ignored. */ type_var_p type_var_init (var_map map) { type_var_p tv; int x, y, p; int num_partitions = num_var_partitions (map); tree t; sbitmap seen; tv = tpa_init (map); if (!tv) return NULL; seen = sbitmap_alloc (num_partitions); sbitmap_zero (seen); for (x = num_partitions - 1; x >= 0; x--) { t = partition_to_var (map, x); /* Disallow coalescing of these types of variables. */ if (!t || TREE_THIS_VOLATILE (t) || TREE_CODE (t) == RESULT_DECL || TREE_CODE (t) == PARM_DECL || (DECL_P (t) && (DECL_REGISTER (t) || !DECL_IGNORED_P (t) || DECL_RTL_SET_P (t)))) continue; p = var_to_partition (map, t); gcc_assert (p != NO_PARTITION); /* If partitions have been coalesced, only add the representative for the partition to the list once. */ if (TEST_BIT (seen, p)) continue; SET_BIT (seen, p); t = TREE_TYPE (t); /* Find the list for this type. */ for (y = 0; y < tv->num_trees; y++) if (t == VEC_index (tree, tv->trees, y)) break; if (y == tv->num_trees) { tv->num_trees++; VEC_safe_push (tree, heap, tv->trees, t); VEC_safe_push (int, heap, tv->first_partition, p); } else { tv->next_partition[p] = VEC_index (int, tv->first_partition, y); VEC_replace (int, tv->first_partition, y, p); } tv->partition_to_tree_map[p] = y; } sbitmap_free (seen); return tv; } /* Create a new coalesce list object from MAP and return it. */ coalesce_list_p create_coalesce_list (var_map map) { coalesce_list_p list; list = (coalesce_list_p) xmalloc (sizeof (struct coalesce_list_d)); list->map = map; list->add_mode = true; list->list = (partition_pair_p *) xcalloc (num_var_partitions (map), sizeof (struct partition_pair_d)); return list; } /* Delete coalesce list CL. */ void delete_coalesce_list (coalesce_list_p cl) { free (cl->list); free (cl); } /* Find a matching coalesce pair object in CL for partitions P1 and P2. If one isn't found, return NULL if CREATE is false, otherwise create a new coalesce pair object and return it. */ static partition_pair_p find_partition_pair (coalesce_list_p cl, int p1, int p2, bool create) { partition_pair_p node, tmp; int s; /* Normalize so that p1 is the smaller value. */ if (p2 < p1) { s = p1; p1 = p2; p2 = s; } tmp = NULL; /* The list is sorted such that if we find a value greater than p2, p2 is not in the list. */ for (node = cl->list[p1]; node; node = node->next) { if (node->second_partition == p2) return node; else if (node->second_partition > p2) break; tmp = node; } if (!create) return NULL; node = (partition_pair_p) xmalloc (sizeof (struct partition_pair_d)); node->first_partition = p1; node->second_partition = p2; node->cost = 0; if (tmp != NULL) { node->next = tmp->next; tmp->next = node; } else { /* This is now the first node in the list. */ node->next = cl->list[p1]; cl->list[p1] = node; } return node; } /* Return cost of execution of copy instruction with FREQUENCY possibly on CRITICAL edge and in HOT basic block. */ int coalesce_cost (int frequency, bool hot, bool critical) { /* Base costs on BB frequencies bounded by 1. */ int cost = frequency; if (!cost) cost = 1; if (optimize_size || hot) cost = 1; /* Inserting copy on critical edge costs more than inserting it elsewhere. */ if (critical) cost *= 2; return cost; } /* Add a potential coalesce between P1 and P2 in CL with a cost of VALUE. */ void add_coalesce (coalesce_list_p cl, int p1, int p2, int value) { partition_pair_p node; gcc_assert (cl->add_mode); if (p1 == p2) return; node = find_partition_pair (cl, p1, p2, true); node->cost += value; } /* Comparison function to allow qsort to sort P1 and P2 in descending order. */ static int compare_pairs (const void *p1, const void *p2) { #if 0 partition_pair_p * pp1 = (partition_pair_p *) p1; partition_pair_p * pp2 = (partition_pair_p *) p2; int result; result = (* pp2)->cost - (* pp1)->cost; /* Issue 128204: Cygwin vs Linux host differences: If the costs are the same, use the partition indicies in order to obtain a stable, reproducible sort. Otherwise the ordering will be at the mercy of the host's qsort library function implementation. */ if (result == 0) { result = (* pp2)->first_partition - (* pp1)->first_partition; if (result == 0) result = (* pp2)->second_partition - (* pp1)->second_partition; } return result; #else return (*(partition_pair_p *)p2)->cost - (*(partition_pair_p *)p1)->cost; #endif } /* Prepare CL for removal of preferred pairs. When finished, list element 0 has all the coalesce pairs, sorted in order from most important coalesce to least important. */ void sort_coalesce_list (coalesce_list_p cl) { unsigned x, num, count; partition_pair_p chain, p; partition_pair_p *list; gcc_assert (cl->add_mode); cl->add_mode = false; /* Compact the array of lists to a single list, and count the elements. */ num = 0; chain = NULL; for (x = 0; x < num_var_partitions (cl->map); x++) if (cl->list[x] != NULL) { for (p = cl->list[x]; p->next != NULL; p = p->next) num++; num++; p->next = chain; chain = cl->list[x]; cl->list[x] = NULL; } /* Only call qsort if there are more than 2 items. */ if (num > 2) { list = XNEWVEC (partition_pair_p, num); count = 0; for (p = chain; p != NULL; p = p->next) list[count++] = p; gcc_assert (count == num); qsort (list, count, sizeof (partition_pair_p), compare_pairs); p = list[0]; for (x = 1; x < num; x++) { p->next = list[x]; p = list[x]; } p->next = NULL; cl->list[0] = list[0]; free (list); } else { cl->list[0] = chain; if (num == 2) { /* Simply swap the two elements if they are in the wrong order. */ if (chain->cost < chain->next->cost) { cl->list[0] = chain->next; cl->list[0]->next = chain; chain->next = NULL; } } } } /* Retrieve the best remaining pair to coalesce from CL. Returns the 2 partitions via P1 and P2. Their calculated cost is returned by the function. NO_BEST_COALESCE is returned if the coalesce list is empty. */ static int pop_best_coalesce (coalesce_list_p cl, int *p1, int *p2) { partition_pair_p node; int ret; gcc_assert (!cl->add_mode); node = cl->list[0]; if (!node) return NO_BEST_COALESCE; cl->list[0] = node->next; *p1 = node->first_partition; *p2 = node->second_partition; ret = node->cost; free (node); return ret; } /* If variable VAR is in a partition in MAP, add a conflict in GRAPH between VAR and any other live partitions in VEC which are associated via TPA. Reset the live bit in VEC. */ static inline void add_conflicts_if_valid (tpa_p tpa, conflict_graph graph, var_map map, bitmap vec, tree var) { int p, y, first; p = var_to_partition (map, var); if (p != NO_PARTITION) { bitmap_clear_bit (vec, p); first = tpa_find_tree (tpa, p); /* If find returns nothing, this object isn't interesting. */ if (first == TPA_NONE) return; /* Only add interferences between objects in the same list. */ for (y = tpa_first_partition (tpa, first); y != TPA_NONE; y = tpa_next_partition (tpa, y)) { if (bitmap_bit_p (vec, y)) conflict_graph_add (graph, p, y); } } } /* Return a conflict graph for the information contained in LIVE_INFO. Only conflicts between items in the same TPA list are added. If optional coalesce list CL is passed in, any copies encountered are added. */ conflict_graph build_tree_conflict_graph (tree_live_info_p liveinfo, tpa_p tpa, coalesce_list_p cl) { conflict_graph graph; var_map map; bitmap live; unsigned x, y, i; basic_block bb; int *partition_link, *tpa_nodes; VEC(int,heap) *tpa_to_clear; unsigned l; ssa_op_iter iter; bitmap_iterator bi; map = live_var_map (liveinfo); graph = conflict_graph_new (num_var_partitions (map)); if (tpa_num_trees (tpa) == 0) return graph; live = BITMAP_ALLOC (NULL); partition_link = XCNEWVEC (int, num_var_partitions (map) + 1); tpa_nodes = XCNEWVEC (int, tpa_num_trees (tpa)); tpa_to_clear = VEC_alloc (int, heap, 50); FOR_EACH_BB (bb) { block_stmt_iterator bsi; tree phi; int idx; /* Start with live on exit temporaries. */ bitmap_copy (live, live_on_exit (liveinfo, bb)); for (bsi = bsi_last (bb); !bsi_end_p (bsi); bsi_prev (&bsi)) { bool is_a_copy = false; tree stmt = bsi_stmt (bsi); /* A copy between 2 partitions does not introduce an interference by itself. If they did, you would never be able to coalesce two things which are copied. If the two variables really do conflict, they will conflict elsewhere in the program. This is handled specially here since we may also be interested in copies between real variables and SSA_NAME variables. We may be interested in trying to coalesce SSA_NAME variables with root variables in some cases. */ if (TREE_CODE (stmt) == MODIFY_EXPR) { tree lhs = TREE_OPERAND (stmt, 0); tree rhs = TREE_OPERAND (stmt, 1); int p1, p2; int bit; if (DECL_P (lhs) || TREE_CODE (lhs) == SSA_NAME) p1 = var_to_partition (map, lhs); else p1 = NO_PARTITION; if (DECL_P (rhs) || TREE_CODE (rhs) == SSA_NAME) p2 = var_to_partition (map, rhs); else p2 = NO_PARTITION; if (p1 != NO_PARTITION && p2 != NO_PARTITION) { is_a_copy = true; bit = bitmap_bit_p (live, p2); /* If the RHS is live, make it not live while we add the conflicts, then make it live again. */ if (bit) bitmap_clear_bit (live, p2); add_conflicts_if_valid (tpa, graph, map, live, lhs); if (bit) bitmap_set_bit (live, p2); if (cl) add_coalesce (cl, p1, p2, coalesce_cost (bb->frequency, maybe_hot_bb_p (bb), false)); set_if_valid (map, live, rhs); } } if (!is_a_copy) { tree var; FOR_EACH_SSA_TREE_OPERAND (var, stmt, iter, SSA_OP_DEF) { add_conflicts_if_valid (tpa, graph, map, live, var); } FOR_EACH_SSA_TREE_OPERAND (var, stmt, iter, SSA_OP_USE) { set_if_valid (map, live, var); } } } /* If result of a PHI is unused, then the loops over the statements will not record any conflicts. However, since the PHI node is going to be translated out of SSA form we must record a conflict between the result of the PHI and any variables with are live. Otherwise the out-of-ssa translation may create incorrect code. */ for (phi = phi_nodes (bb); phi; phi = PHI_CHAIN (phi)) { tree result = PHI_RESULT (phi); int p = var_to_partition (map, result); if (p != NO_PARTITION && ! bitmap_bit_p (live, p)) add_conflicts_if_valid (tpa, graph, map, live, result); } /* Anything which is still live at this point interferes. In order to implement this efficiently, only conflicts between partitions which have the same TPA root need be added. TPA roots which have been seen are tracked in 'tpa_nodes'. A nonzero entry points to an index into 'partition_link', which then indexes into itself forming a linked list of partitions sharing a tpa root which have been seen as live up to this point. Since partitions start at index zero, all entries in partition_link are (partition + 1). Conflicts are added between the current partition and any already seen. tpa_clear contains all the tpa_roots processed, and these are the only entries which need to be zero'd out for a clean restart. */ EXECUTE_IF_SET_IN_BITMAP (live, 0, x, bi) { i = tpa_find_tree (tpa, x); if (i != (unsigned)TPA_NONE) { int start = tpa_nodes[i]; /* If start is 0, a new root reference list is being started. Register it to be cleared. */ if (!start) VEC_safe_push (int, heap, tpa_to_clear, i); /* Add interferences to other tpa members seen. */ for (y = start; y != 0; y = partition_link[y]) conflict_graph_add (graph, x, y - 1); tpa_nodes[i] = x + 1; partition_link[x + 1] = start; } } /* Now clear the used tpa root references. */ for (l = 0; VEC_iterate (int, tpa_to_clear, l, idx); l++) tpa_nodes[idx] = 0; VEC_truncate (int, tpa_to_clear, 0); } free (tpa_nodes); free (partition_link); VEC_free (int, heap, tpa_to_clear); BITMAP_FREE (live); return graph; } /* This routine will attempt to coalesce the elements in TPA subject to the conflicts found in GRAPH. If optional coalesce_list CL is provided, only coalesces specified within the coalesce list are attempted. Otherwise an attempt is made to coalesce as many partitions within each TPA grouping as possible. If DEBUG is provided, debug output will be sent there. */ void coalesce_tpa_members (tpa_p tpa, conflict_graph graph, var_map map, coalesce_list_p cl, FILE *debug) { int x, y, z, w; tree var, tmp; /* Attempt to coalesce any items in a coalesce list. */ if (cl) { while (pop_best_coalesce (cl, &x, &y) != NO_BEST_COALESCE) { if (debug) { fprintf (debug, "Coalesce list: (%d)", x); print_generic_expr (debug, partition_to_var (map, x), TDF_SLIM); fprintf (debug, " & (%d)", y); print_generic_expr (debug, partition_to_var (map, y), TDF_SLIM); } w = tpa_find_tree (tpa, x); z = tpa_find_tree (tpa, y); if (w != z || w == TPA_NONE || z == TPA_NONE) { if (debug) { if (w != z) fprintf (debug, ": Fail, Non-matching TPA's\n"); if (w == TPA_NONE) fprintf (debug, ": Fail %d non TPA.\n", x); else fprintf (debug, ": Fail %d non TPA.\n", y); } continue; } var = partition_to_var (map, x); tmp = partition_to_var (map, y); x = var_to_partition (map, var); y = var_to_partition (map, tmp); if (debug) fprintf (debug, " [map: %d, %d] ", x, y); if (x == y) { if (debug) fprintf (debug, ": Already Coalesced.\n"); continue; } if (!conflict_graph_conflict_p (graph, x, y)) { z = var_union (map, var, tmp); if (z == NO_PARTITION) { if (debug) fprintf (debug, ": Unable to perform partition union.\n"); continue; } /* z is the new combined partition. We need to remove the other partition from the list. Set x to be that other partition. */ if (z == x) { conflict_graph_merge_regs (graph, x, y); w = tpa_find_tree (tpa, y); tpa_remove_partition (tpa, w, y); } else { conflict_graph_merge_regs (graph, y, x); w = tpa_find_tree (tpa, x); tpa_remove_partition (tpa, w, x); } if (debug) fprintf (debug, ": Success -> %d\n", z); } else if (debug) fprintf (debug, ": Fail due to conflict\n"); } /* If using a coalesce list, don't try to coalesce anything else. */ return; } for (x = 0; x < tpa_num_trees (tpa); x++) { while (tpa_first_partition (tpa, x) != TPA_NONE) { int p1, p2; /* Coalesce first partition with anything that doesn't conflict. */ y = tpa_first_partition (tpa, x); tpa_remove_partition (tpa, x, y); var = partition_to_var (map, y); /* p1 is the partition representative to which y belongs. */ p1 = var_to_partition (map, var); for (z = tpa_next_partition (tpa, y); z != TPA_NONE; z = tpa_next_partition (tpa, z)) { tmp = partition_to_var (map, z); /* p2 is the partition representative to which z belongs. */ p2 = var_to_partition (map, tmp); if (debug) { fprintf (debug, "Coalesce : "); print_generic_expr (debug, var, TDF_SLIM); fprintf (debug, " &"); print_generic_expr (debug, tmp, TDF_SLIM); fprintf (debug, " (%d ,%d)", p1, p2); } /* If partitions are already merged, don't check for conflict. */ if (tmp == var) { tpa_remove_partition (tpa, x, z); if (debug) fprintf (debug, ": Already coalesced\n"); } else if (!conflict_graph_conflict_p (graph, p1, p2)) { int v; if (tpa_find_tree (tpa, y) == TPA_NONE || tpa_find_tree (tpa, z) == TPA_NONE) { if (debug) fprintf (debug, ": Fail non-TPA member\n"); continue; } if ((v = var_union (map, var, tmp)) == NO_PARTITION) { if (debug) fprintf (debug, ": Fail cannot combine partitions\n"); continue; } tpa_remove_partition (tpa, x, z); if (v == p1) conflict_graph_merge_regs (graph, v, z); else { /* Update the first partition's representative. */ conflict_graph_merge_regs (graph, v, y); p1 = v; } /* The root variable of the partition may be changed now. */ var = partition_to_var (map, p1); if (debug) fprintf (debug, ": Success -> %d\n", v); } else if (debug) fprintf (debug, ": Fail, Conflict\n"); } } } } /* Send debug info for coalesce list CL to file F. */ void dump_coalesce_list (FILE *f, coalesce_list_p cl) { partition_pair_p node; int x, num; tree var; if (cl->add_mode) { fprintf (f, "Coalesce List:\n"); num = num_var_partitions (cl->map); for (x = 0; x < num; x++) { node = cl->list[x]; if (node) { fprintf (f, "["); print_generic_expr (f, partition_to_var (cl->map, x), TDF_SLIM); fprintf (f, "] - "); for ( ; node; node = node->next) { var = partition_to_var (cl->map, node->second_partition); print_generic_expr (f, var, TDF_SLIM); fprintf (f, "(%1d), ", node->cost); } fprintf (f, "\n"); } } } else { fprintf (f, "Sorted Coalesce list:\n"); for (node = cl->list[0]; node; node = node->next) { fprintf (f, "(%d) ", node->cost); var = partition_to_var (cl->map, node->first_partition); print_generic_expr (f, var, TDF_SLIM); fprintf (f, " : "); var = partition_to_var (cl->map, node->second_partition); print_generic_expr (f, var, TDF_SLIM); fprintf (f, "\n"); } } } /* Output tree_partition_associator object TPA to file F.. */ void tpa_dump (FILE *f, tpa_p tpa) { int x, i; if (!tpa) return; for (x = 0; x < tpa_num_trees (tpa); x++) { print_generic_expr (f, tpa_tree (tpa, x), TDF_SLIM); fprintf (f, " : ("); for (i = tpa_first_partition (tpa, x); i != TPA_NONE; i = tpa_next_partition (tpa, i)) { fprintf (f, "(%d)",i); print_generic_expr (f, partition_to_var (tpa->map, i), TDF_SLIM); fprintf (f, " "); #ifdef ENABLE_CHECKING if (tpa_find_tree (tpa, i) != x) fprintf (f, "**find tree incorrectly set** "); #endif } fprintf (f, ")\n"); } fflush (f); } /* Output partition map MAP to file F. */ void dump_var_map (FILE *f, var_map map) { int t; unsigned x, y; int p; fprintf (f, "\nPartition map \n\n"); for (x = 0; x < map->num_partitions; x++) { if (map->compact_to_partition != NULL) p = map->compact_to_partition[x]; else p = x; if (map->partition_to_var[p] == NULL_TREE) continue; t = 0; for (y = 1; y < num_ssa_names; y++) { p = partition_find (map->var_partition, y); if (map->partition_to_compact) p = map->partition_to_compact[p]; if (p == (int)x) { if (t++ == 0) { fprintf(f, "Partition %d (", x); print_generic_expr (f, partition_to_var (map, p), TDF_SLIM); fprintf (f, " - "); } fprintf (f, "%d ", y); } } if (t != 0) fprintf (f, ")\n"); } fprintf (f, "\n"); } /* Output live range info LIVE to file F, controlled by FLAG. */ void dump_live_info (FILE *f, tree_live_info_p live, int flag) { basic_block bb; unsigned i; var_map map = live->map; bitmap_iterator bi; if ((flag & LIVEDUMP_ENTRY) && live->livein) { FOR_EACH_BB (bb) { fprintf (f, "\nLive on entry to BB%d : ", bb->index); for (i = 0; i < num_var_partitions (map); i++) { if (bitmap_bit_p (live_entry_blocks (live, i), bb->index)) { print_generic_expr (f, partition_to_var (map, i), TDF_SLIM); fprintf (f, " "); } } fprintf (f, "\n"); } } if ((flag & LIVEDUMP_EXIT) && live->liveout) { FOR_EACH_BB (bb) { fprintf (f, "\nLive on exit from BB%d : ", bb->index); EXECUTE_IF_SET_IN_BITMAP (live->liveout[bb->index], 0, i, bi) { print_generic_expr (f, partition_to_var (map, i), TDF_SLIM); fprintf (f, " "); } fprintf (f, "\n"); } } } #ifdef ENABLE_CHECKING void register_ssa_partition_check (tree ssa_var) { gcc_assert (TREE_CODE (ssa_var) == SSA_NAME); if (!is_gimple_reg (SSA_NAME_VAR (ssa_var))) { fprintf (stderr, "Illegally registering a virtual SSA name :"); print_generic_expr (stderr, ssa_var, TDF_SLIM); fprintf (stderr, " in the SSA->Normal phase.\n"); internal_error ("SSA corruption"); } } #endif
Go to most recent revision | Compare with Previous | Blame | View Log