| 1 | 684 | jeremybenn | /* Emit RTL for the GCC expander.
 | 
      
         | 2 |  |  |    Copyright (C) 1987, 1988, 1992, 1993, 1994, 1995, 1996, 1997, 1998,
 | 
      
         | 3 |  |  |    1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009,
 | 
      
         | 4 |  |  |    2010, 2011
 | 
      
         | 5 |  |  |    Free Software Foundation, Inc.
 | 
      
         | 6 |  |  |  
 | 
      
         | 7 |  |  | This file is part of GCC.
 | 
      
         | 8 |  |  |  
 | 
      
         | 9 |  |  | GCC is free software; you can redistribute it and/or modify it under
 | 
      
         | 10 |  |  | the terms of the GNU General Public License as published by the Free
 | 
      
         | 11 |  |  | Software Foundation; either version 3, or (at your option) any later
 | 
      
         | 12 |  |  | version.
 | 
      
         | 13 |  |  |  
 | 
      
         | 14 |  |  | GCC is distributed in the hope that it will be useful, but WITHOUT ANY
 | 
      
         | 15 |  |  | WARRANTY; without even the implied warranty of MERCHANTABILITY or
 | 
      
         | 16 |  |  | FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
 | 
      
         | 17 |  |  | for more details.
 | 
      
         | 18 |  |  |  
 | 
      
         | 19 |  |  | You should have received a copy of the GNU General Public License
 | 
      
         | 20 |  |  | along with GCC; see the file COPYING3.  If not see
 | 
      
         | 21 |  |  | <http://www.gnu.org/licenses/>.  */
 | 
      
         | 22 |  |  |  
 | 
      
         | 23 |  |  |  
 | 
      
         | 24 |  |  | /* Middle-to-low level generation of rtx code and insns.
 | 
      
         | 25 |  |  |  
 | 
      
         | 26 |  |  |    This file contains support functions for creating rtl expressions
 | 
      
         | 27 |  |  |    and manipulating them in the doubly-linked chain of insns.
 | 
      
         | 28 |  |  |  
 | 
      
         | 29 |  |  |    The patterns of the insns are created by machine-dependent
 | 
      
         | 30 |  |  |    routines in insn-emit.c, which is generated automatically from
 | 
      
         | 31 |  |  |    the machine description.  These routines make the individual rtx's
 | 
      
         | 32 |  |  |    of the pattern with `gen_rtx_fmt_ee' and others in genrtl.[ch],
 | 
      
         | 33 |  |  |    which are automatically generated from rtl.def; what is machine
 | 
      
         | 34 |  |  |    dependent is the kind of rtx's they make and what arguments they
 | 
      
         | 35 |  |  |    use.  */
 | 
      
         | 36 |  |  |  
 | 
      
         | 37 |  |  | #include "config.h"
 | 
      
         | 38 |  |  | #include "system.h"
 | 
      
         | 39 |  |  | #include "coretypes.h"
 | 
      
         | 40 |  |  | #include "tm.h"
 | 
      
         | 41 |  |  | #include "diagnostic-core.h"
 | 
      
         | 42 |  |  | #include "rtl.h"
 | 
      
         | 43 |  |  | #include "tree.h"
 | 
      
         | 44 |  |  | #include "tm_p.h"
 | 
      
         | 45 |  |  | #include "flags.h"
 | 
      
         | 46 |  |  | #include "function.h"
 | 
      
         | 47 |  |  | #include "expr.h"
 | 
      
         | 48 |  |  | #include "regs.h"
 | 
      
         | 49 |  |  | #include "hard-reg-set.h"
 | 
      
         | 50 |  |  | #include "hashtab.h"
 | 
      
         | 51 |  |  | #include "insn-config.h"
 | 
      
         | 52 |  |  | #include "recog.h"
 | 
      
         | 53 |  |  | #include "bitmap.h"
 | 
      
         | 54 |  |  | #include "basic-block.h"
 | 
      
         | 55 |  |  | #include "ggc.h"
 | 
      
         | 56 |  |  | #include "debug.h"
 | 
      
         | 57 |  |  | #include "langhooks.h"
 | 
      
         | 58 |  |  | #include "tree-pass.h"
 | 
      
         | 59 |  |  | #include "df.h"
 | 
      
         | 60 |  |  | #include "params.h"
 | 
      
         | 61 |  |  | #include "target.h"
 | 
      
         | 62 |  |  | #include "tree-flow.h"
 | 
      
         | 63 |  |  |  
 | 
      
         | 64 |  |  | struct target_rtl default_target_rtl;
 | 
      
         | 65 |  |  | #if SWITCHABLE_TARGET
 | 
      
         | 66 |  |  | struct target_rtl *this_target_rtl = &default_target_rtl;
 | 
      
         | 67 |  |  | #endif
 | 
      
         | 68 |  |  |  
 | 
      
         | 69 |  |  | #define initial_regno_reg_rtx (this_target_rtl->x_initial_regno_reg_rtx)
 | 
      
         | 70 |  |  |  
 | 
      
         | 71 |  |  | /* Commonly used modes.  */
 | 
      
         | 72 |  |  |  
 | 
      
         | 73 |  |  | enum machine_mode byte_mode;    /* Mode whose width is BITS_PER_UNIT.  */
 | 
      
         | 74 |  |  | enum machine_mode word_mode;    /* Mode whose width is BITS_PER_WORD.  */
 | 
      
         | 75 |  |  | enum machine_mode double_mode;  /* Mode whose width is DOUBLE_TYPE_SIZE.  */
 | 
      
         | 76 |  |  | enum machine_mode ptr_mode;     /* Mode whose width is POINTER_SIZE.  */
 | 
      
         | 77 |  |  |  
 | 
      
         | 78 |  |  | /* Datastructures maintained for currently processed function in RTL form.  */
 | 
      
         | 79 |  |  |  
 | 
      
         | 80 |  |  | struct rtl_data x_rtl;
 | 
      
         | 81 |  |  |  
 | 
      
         | 82 |  |  | /* Indexed by pseudo register number, gives the rtx for that pseudo.
 | 
      
         | 83 |  |  |    Allocated in parallel with regno_pointer_align.
 | 
      
         | 84 |  |  |    FIXME: We could put it into emit_status struct, but gengtype is not able to deal
 | 
      
         | 85 |  |  |    with length attribute nested in top level structures.  */
 | 
      
         | 86 |  |  |  
 | 
      
         | 87 |  |  | rtx * regno_reg_rtx;
 | 
      
         | 88 |  |  |  
 | 
      
         | 89 |  |  | /* This is *not* reset after each function.  It gives each CODE_LABEL
 | 
      
         | 90 |  |  |    in the entire compilation a unique label number.  */
 | 
      
         | 91 |  |  |  
 | 
      
         | 92 |  |  | static GTY(()) int label_num = 1;
 | 
      
         | 93 |  |  |  
 | 
      
         | 94 |  |  | /* We record floating-point CONST_DOUBLEs in each floating-point mode for
 | 
      
         | 95 |  |  |    the values of 0, 1, and 2.  For the integer entries and VOIDmode, we
 | 
      
         | 96 |  |  |    record a copy of const[012]_rtx and constm1_rtx.  CONSTM1_RTX
 | 
      
         | 97 |  |  |    is set only for MODE_INT and MODE_VECTOR_INT modes.  */
 | 
      
         | 98 |  |  |  
 | 
      
         | 99 |  |  | rtx const_tiny_rtx[4][(int) MAX_MACHINE_MODE];
 | 
      
         | 100 |  |  |  
 | 
      
         | 101 |  |  | rtx const_true_rtx;
 | 
      
         | 102 |  |  |  
 | 
      
         | 103 |  |  | REAL_VALUE_TYPE dconst0;
 | 
      
         | 104 |  |  | REAL_VALUE_TYPE dconst1;
 | 
      
         | 105 |  |  | REAL_VALUE_TYPE dconst2;
 | 
      
         | 106 |  |  | REAL_VALUE_TYPE dconstm1;
 | 
      
         | 107 |  |  | REAL_VALUE_TYPE dconsthalf;
 | 
      
         | 108 |  |  |  
 | 
      
         | 109 |  |  | /* Record fixed-point constant 0 and 1.  */
 | 
      
         | 110 |  |  | FIXED_VALUE_TYPE fconst0[MAX_FCONST0];
 | 
      
         | 111 |  |  | FIXED_VALUE_TYPE fconst1[MAX_FCONST1];
 | 
      
         | 112 |  |  |  
 | 
      
         | 113 |  |  | /* We make one copy of (const_int C) where C is in
 | 
      
         | 114 |  |  |    [- MAX_SAVED_CONST_INT, MAX_SAVED_CONST_INT]
 | 
      
         | 115 |  |  |    to save space during the compilation and simplify comparisons of
 | 
      
         | 116 |  |  |    integers.  */
 | 
      
         | 117 |  |  |  
 | 
      
         | 118 |  |  | rtx const_int_rtx[MAX_SAVED_CONST_INT * 2 + 1];
 | 
      
         | 119 |  |  |  
 | 
      
         | 120 |  |  | /* A hash table storing CONST_INTs whose absolute value is greater
 | 
      
         | 121 |  |  |    than MAX_SAVED_CONST_INT.  */
 | 
      
         | 122 |  |  |  
 | 
      
         | 123 |  |  | static GTY ((if_marked ("ggc_marked_p"), param_is (struct rtx_def)))
 | 
      
         | 124 |  |  |      htab_t const_int_htab;
 | 
      
         | 125 |  |  |  
 | 
      
         | 126 |  |  | /* A hash table storing memory attribute structures.  */
 | 
      
         | 127 |  |  | static GTY ((if_marked ("ggc_marked_p"), param_is (struct mem_attrs)))
 | 
      
         | 128 |  |  |      htab_t mem_attrs_htab;
 | 
      
         | 129 |  |  |  
 | 
      
         | 130 |  |  | /* A hash table storing register attribute structures.  */
 | 
      
         | 131 |  |  | static GTY ((if_marked ("ggc_marked_p"), param_is (struct reg_attrs)))
 | 
      
         | 132 |  |  |      htab_t reg_attrs_htab;
 | 
      
         | 133 |  |  |  
 | 
      
         | 134 |  |  | /* A hash table storing all CONST_DOUBLEs.  */
 | 
      
         | 135 |  |  | static GTY ((if_marked ("ggc_marked_p"), param_is (struct rtx_def)))
 | 
      
         | 136 |  |  |      htab_t const_double_htab;
 | 
      
         | 137 |  |  |  
 | 
      
         | 138 |  |  | /* A hash table storing all CONST_FIXEDs.  */
 | 
      
         | 139 |  |  | static GTY ((if_marked ("ggc_marked_p"), param_is (struct rtx_def)))
 | 
      
         | 140 |  |  |      htab_t const_fixed_htab;
 | 
      
         | 141 |  |  |  
 | 
      
         | 142 |  |  | #define cur_insn_uid (crtl->emit.x_cur_insn_uid)
 | 
      
         | 143 |  |  | #define cur_debug_insn_uid (crtl->emit.x_cur_debug_insn_uid)
 | 
      
         | 144 |  |  | #define last_location (crtl->emit.x_last_location)
 | 
      
         | 145 |  |  | #define first_label_num (crtl->emit.x_first_label_num)
 | 
      
         | 146 |  |  |  
 | 
      
         | 147 |  |  | static rtx make_call_insn_raw (rtx);
 | 
      
         | 148 |  |  | static rtx change_address_1 (rtx, enum machine_mode, rtx, int);
 | 
      
         | 149 |  |  | static void set_used_decls (tree);
 | 
      
         | 150 |  |  | static void mark_label_nuses (rtx);
 | 
      
         | 151 |  |  | static hashval_t const_int_htab_hash (const void *);
 | 
      
         | 152 |  |  | static int const_int_htab_eq (const void *, const void *);
 | 
      
         | 153 |  |  | static hashval_t const_double_htab_hash (const void *);
 | 
      
         | 154 |  |  | static int const_double_htab_eq (const void *, const void *);
 | 
      
         | 155 |  |  | static rtx lookup_const_double (rtx);
 | 
      
         | 156 |  |  | static hashval_t const_fixed_htab_hash (const void *);
 | 
      
         | 157 |  |  | static int const_fixed_htab_eq (const void *, const void *);
 | 
      
         | 158 |  |  | static rtx lookup_const_fixed (rtx);
 | 
      
         | 159 |  |  | static hashval_t mem_attrs_htab_hash (const void *);
 | 
      
         | 160 |  |  | static int mem_attrs_htab_eq (const void *, const void *);
 | 
      
         | 161 |  |  | static hashval_t reg_attrs_htab_hash (const void *);
 | 
      
         | 162 |  |  | static int reg_attrs_htab_eq (const void *, const void *);
 | 
      
         | 163 |  |  | static reg_attrs *get_reg_attrs (tree, int);
 | 
      
         | 164 |  |  | static rtx gen_const_vector (enum machine_mode, int);
 | 
      
         | 165 |  |  | static void copy_rtx_if_shared_1 (rtx *orig);
 | 
      
         | 166 |  |  |  
 | 
      
         | 167 |  |  | /* Probability of the conditional branch currently proceeded by try_split.
 | 
      
         | 168 |  |  |    Set to -1 otherwise.  */
 | 
      
         | 169 |  |  | int split_branch_probability = -1;
 | 
      
         | 170 |  |  |  
 | 
      
         | 171 |  |  | /* Returns a hash code for X (which is a really a CONST_INT).  */
 | 
      
         | 172 |  |  |  
 | 
      
         | 173 |  |  | static hashval_t
 | 
      
         | 174 |  |  | const_int_htab_hash (const void *x)
 | 
      
         | 175 |  |  | {
 | 
      
         | 176 |  |  |   return (hashval_t) INTVAL ((const_rtx) x);
 | 
      
         | 177 |  |  | }
 | 
      
         | 178 |  |  |  
 | 
      
         | 179 |  |  | /* Returns nonzero if the value represented by X (which is really a
 | 
      
         | 180 |  |  |    CONST_INT) is the same as that given by Y (which is really a
 | 
      
         | 181 |  |  |    HOST_WIDE_INT *).  */
 | 
      
         | 182 |  |  |  
 | 
      
         | 183 |  |  | static int
 | 
      
         | 184 |  |  | const_int_htab_eq (const void *x, const void *y)
 | 
      
         | 185 |  |  | {
 | 
      
         | 186 |  |  |   return (INTVAL ((const_rtx) x) == *((const HOST_WIDE_INT *) y));
 | 
      
         | 187 |  |  | }
 | 
      
         | 188 |  |  |  
 | 
      
         | 189 |  |  | /* Returns a hash code for X (which is really a CONST_DOUBLE).  */
 | 
      
         | 190 |  |  | static hashval_t
 | 
      
         | 191 |  |  | const_double_htab_hash (const void *x)
 | 
      
         | 192 |  |  | {
 | 
      
         | 193 |  |  |   const_rtx const value = (const_rtx) x;
 | 
      
         | 194 |  |  |   hashval_t h;
 | 
      
         | 195 |  |  |  
 | 
      
         | 196 |  |  |   if (GET_MODE (value) == VOIDmode)
 | 
      
         | 197 |  |  |     h = CONST_DOUBLE_LOW (value) ^ CONST_DOUBLE_HIGH (value);
 | 
      
         | 198 |  |  |   else
 | 
      
         | 199 |  |  |     {
 | 
      
         | 200 |  |  |       h = real_hash (CONST_DOUBLE_REAL_VALUE (value));
 | 
      
         | 201 |  |  |       /* MODE is used in the comparison, so it should be in the hash.  */
 | 
      
         | 202 |  |  |       h ^= GET_MODE (value);
 | 
      
         | 203 |  |  |     }
 | 
      
         | 204 |  |  |   return h;
 | 
      
         | 205 |  |  | }
 | 
      
         | 206 |  |  |  
 | 
      
         | 207 |  |  | /* Returns nonzero if the value represented by X (really a ...)
 | 
      
         | 208 |  |  |    is the same as that represented by Y (really a ...) */
 | 
      
         | 209 |  |  | static int
 | 
      
         | 210 |  |  | const_double_htab_eq (const void *x, const void *y)
 | 
      
         | 211 |  |  | {
 | 
      
         | 212 |  |  |   const_rtx const a = (const_rtx)x, b = (const_rtx)y;
 | 
      
         | 213 |  |  |  
 | 
      
         | 214 |  |  |   if (GET_MODE (a) != GET_MODE (b))
 | 
      
         | 215 |  |  |     return 0;
 | 
      
         | 216 |  |  |   if (GET_MODE (a) == VOIDmode)
 | 
      
         | 217 |  |  |     return (CONST_DOUBLE_LOW (a) == CONST_DOUBLE_LOW (b)
 | 
      
         | 218 |  |  |             && CONST_DOUBLE_HIGH (a) == CONST_DOUBLE_HIGH (b));
 | 
      
         | 219 |  |  |   else
 | 
      
         | 220 |  |  |     return real_identical (CONST_DOUBLE_REAL_VALUE (a),
 | 
      
         | 221 |  |  |                            CONST_DOUBLE_REAL_VALUE (b));
 | 
      
         | 222 |  |  | }
 | 
      
         | 223 |  |  |  
 | 
      
         | 224 |  |  | /* Returns a hash code for X (which is really a CONST_FIXED).  */
 | 
      
         | 225 |  |  |  
 | 
      
         | 226 |  |  | static hashval_t
 | 
      
         | 227 |  |  | const_fixed_htab_hash (const void *x)
 | 
      
         | 228 |  |  | {
 | 
      
         | 229 |  |  |   const_rtx const value = (const_rtx) x;
 | 
      
         | 230 |  |  |   hashval_t h;
 | 
      
         | 231 |  |  |  
 | 
      
         | 232 |  |  |   h = fixed_hash (CONST_FIXED_VALUE (value));
 | 
      
         | 233 |  |  |   /* MODE is used in the comparison, so it should be in the hash.  */
 | 
      
         | 234 |  |  |   h ^= GET_MODE (value);
 | 
      
         | 235 |  |  |   return h;
 | 
      
         | 236 |  |  | }
 | 
      
         | 237 |  |  |  
 | 
      
         | 238 |  |  | /* Returns nonzero if the value represented by X (really a ...)
 | 
      
         | 239 |  |  |    is the same as that represented by Y (really a ...).  */
 | 
      
         | 240 |  |  |  
 | 
      
         | 241 |  |  | static int
 | 
      
         | 242 |  |  | const_fixed_htab_eq (const void *x, const void *y)
 | 
      
         | 243 |  |  | {
 | 
      
         | 244 |  |  |   const_rtx const a = (const_rtx) x, b = (const_rtx) y;
 | 
      
         | 245 |  |  |  
 | 
      
         | 246 |  |  |   if (GET_MODE (a) != GET_MODE (b))
 | 
      
         | 247 |  |  |     return 0;
 | 
      
         | 248 |  |  |   return fixed_identical (CONST_FIXED_VALUE (a), CONST_FIXED_VALUE (b));
 | 
      
         | 249 |  |  | }
 | 
      
         | 250 |  |  |  
 | 
      
         | 251 |  |  | /* Returns a hash code for X (which is a really a mem_attrs *).  */
 | 
      
         | 252 |  |  |  
 | 
      
         | 253 |  |  | static hashval_t
 | 
      
         | 254 |  |  | mem_attrs_htab_hash (const void *x)
 | 
      
         | 255 |  |  | {
 | 
      
         | 256 |  |  |   const mem_attrs *const p = (const mem_attrs *) x;
 | 
      
         | 257 |  |  |  
 | 
      
         | 258 |  |  |   return (p->alias ^ (p->align * 1000)
 | 
      
         | 259 |  |  |           ^ (p->addrspace * 4000)
 | 
      
         | 260 |  |  |           ^ ((p->offset_known_p ? p->offset : 0) * 50000)
 | 
      
         | 261 |  |  |           ^ ((p->size_known_p ? p->size : 0) * 2500000)
 | 
      
         | 262 |  |  |           ^ (size_t) iterative_hash_expr (p->expr, 0));
 | 
      
         | 263 |  |  | }
 | 
      
         | 264 |  |  |  
 | 
      
         | 265 |  |  | /* Return true if the given memory attributes are equal.  */
 | 
      
         | 266 |  |  |  
 | 
      
         | 267 |  |  | static bool
 | 
      
         | 268 |  |  | mem_attrs_eq_p (const struct mem_attrs *p, const struct mem_attrs *q)
 | 
      
         | 269 |  |  | {
 | 
      
         | 270 |  |  |   return (p->alias == q->alias
 | 
      
         | 271 |  |  |           && p->offset_known_p == q->offset_known_p
 | 
      
         | 272 |  |  |           && (!p->offset_known_p || p->offset == q->offset)
 | 
      
         | 273 |  |  |           && p->size_known_p == q->size_known_p
 | 
      
         | 274 |  |  |           && (!p->size_known_p || p->size == q->size)
 | 
      
         | 275 |  |  |           && p->align == q->align
 | 
      
         | 276 |  |  |           && p->addrspace == q->addrspace
 | 
      
         | 277 |  |  |           && (p->expr == q->expr
 | 
      
         | 278 |  |  |               || (p->expr != NULL_TREE && q->expr != NULL_TREE
 | 
      
         | 279 |  |  |                   && operand_equal_p (p->expr, q->expr, 0))));
 | 
      
         | 280 |  |  | }
 | 
      
         | 281 |  |  |  
 | 
      
         | 282 |  |  | /* Returns nonzero if the value represented by X (which is really a
 | 
      
         | 283 |  |  |    mem_attrs *) is the same as that given by Y (which is also really a
 | 
      
         | 284 |  |  |    mem_attrs *).  */
 | 
      
         | 285 |  |  |  
 | 
      
         | 286 |  |  | static int
 | 
      
         | 287 |  |  | mem_attrs_htab_eq (const void *x, const void *y)
 | 
      
         | 288 |  |  | {
 | 
      
         | 289 |  |  |   return mem_attrs_eq_p ((const mem_attrs *) x, (const mem_attrs *) y);
 | 
      
         | 290 |  |  | }
 | 
      
         | 291 |  |  |  
 | 
      
         | 292 |  |  | /* Set MEM's memory attributes so that they are the same as ATTRS.  */
 | 
      
         | 293 |  |  |  
 | 
      
         | 294 |  |  | static void
 | 
      
         | 295 |  |  | set_mem_attrs (rtx mem, mem_attrs *attrs)
 | 
      
         | 296 |  |  | {
 | 
      
         | 297 |  |  |   void **slot;
 | 
      
         | 298 |  |  |  
 | 
      
         | 299 |  |  |   /* If everything is the default, we can just clear the attributes.  */
 | 
      
         | 300 |  |  |   if (mem_attrs_eq_p (attrs, mode_mem_attrs[(int) GET_MODE (mem)]))
 | 
      
         | 301 |  |  |     {
 | 
      
         | 302 |  |  |       MEM_ATTRS (mem) = 0;
 | 
      
         | 303 |  |  |       return;
 | 
      
         | 304 |  |  |     }
 | 
      
         | 305 |  |  |  
 | 
      
         | 306 |  |  |   slot = htab_find_slot (mem_attrs_htab, attrs, INSERT);
 | 
      
         | 307 |  |  |   if (*slot == 0)
 | 
      
         | 308 |  |  |     {
 | 
      
         | 309 |  |  |       *slot = ggc_alloc_mem_attrs ();
 | 
      
         | 310 |  |  |       memcpy (*slot, attrs, sizeof (mem_attrs));
 | 
      
         | 311 |  |  |     }
 | 
      
         | 312 |  |  |  
 | 
      
         | 313 |  |  |   MEM_ATTRS (mem) = (mem_attrs *) *slot;
 | 
      
         | 314 |  |  | }
 | 
      
         | 315 |  |  |  
 | 
      
         | 316 |  |  | /* Returns a hash code for X (which is a really a reg_attrs *).  */
 | 
      
         | 317 |  |  |  
 | 
      
         | 318 |  |  | static hashval_t
 | 
      
         | 319 |  |  | reg_attrs_htab_hash (const void *x)
 | 
      
         | 320 |  |  | {
 | 
      
         | 321 |  |  |   const reg_attrs *const p = (const reg_attrs *) x;
 | 
      
         | 322 |  |  |  
 | 
      
         | 323 |  |  |   return ((p->offset * 1000) ^ (intptr_t) p->decl);
 | 
      
         | 324 |  |  | }
 | 
      
         | 325 |  |  |  
 | 
      
         | 326 |  |  | /* Returns nonzero if the value represented by X (which is really a
 | 
      
         | 327 |  |  |    reg_attrs *) is the same as that given by Y (which is also really a
 | 
      
         | 328 |  |  |    reg_attrs *).  */
 | 
      
         | 329 |  |  |  
 | 
      
         | 330 |  |  | static int
 | 
      
         | 331 |  |  | reg_attrs_htab_eq (const void *x, const void *y)
 | 
      
         | 332 |  |  | {
 | 
      
         | 333 |  |  |   const reg_attrs *const p = (const reg_attrs *) x;
 | 
      
         | 334 |  |  |   const reg_attrs *const q = (const reg_attrs *) y;
 | 
      
         | 335 |  |  |  
 | 
      
         | 336 |  |  |   return (p->decl == q->decl && p->offset == q->offset);
 | 
      
         | 337 |  |  | }
 | 
      
         | 338 |  |  | /* Allocate a new reg_attrs structure and insert it into the hash table if
 | 
      
         | 339 |  |  |    one identical to it is not already in the table.  We are doing this for
 | 
      
         | 340 |  |  |    MEM of mode MODE.  */
 | 
      
         | 341 |  |  |  
 | 
      
         | 342 |  |  | static reg_attrs *
 | 
      
         | 343 |  |  | get_reg_attrs (tree decl, int offset)
 | 
      
         | 344 |  |  | {
 | 
      
         | 345 |  |  |   reg_attrs attrs;
 | 
      
         | 346 |  |  |   void **slot;
 | 
      
         | 347 |  |  |  
 | 
      
         | 348 |  |  |   /* If everything is the default, we can just return zero.  */
 | 
      
         | 349 |  |  |   if (decl == 0 && offset == 0)
 | 
      
         | 350 |  |  |     return 0;
 | 
      
         | 351 |  |  |  
 | 
      
         | 352 |  |  |   attrs.decl = decl;
 | 
      
         | 353 |  |  |   attrs.offset = offset;
 | 
      
         | 354 |  |  |  
 | 
      
         | 355 |  |  |   slot = htab_find_slot (reg_attrs_htab, &attrs, INSERT);
 | 
      
         | 356 |  |  |   if (*slot == 0)
 | 
      
         | 357 |  |  |     {
 | 
      
         | 358 |  |  |       *slot = ggc_alloc_reg_attrs ();
 | 
      
         | 359 |  |  |       memcpy (*slot, &attrs, sizeof (reg_attrs));
 | 
      
         | 360 |  |  |     }
 | 
      
         | 361 |  |  |  
 | 
      
         | 362 |  |  |   return (reg_attrs *) *slot;
 | 
      
         | 363 |  |  | }
 | 
      
         | 364 |  |  |  
 | 
      
         | 365 |  |  |  
 | 
      
         | 366 |  |  | #if !HAVE_blockage
 | 
      
         | 367 |  |  | /* Generate an empty ASM_INPUT, which is used to block attempts to schedule
 | 
      
         | 368 |  |  |    across this insn. */
 | 
      
         | 369 |  |  |  
 | 
      
         | 370 |  |  | rtx
 | 
      
         | 371 |  |  | gen_blockage (void)
 | 
      
         | 372 |  |  | {
 | 
      
         | 373 |  |  |   rtx x = gen_rtx_ASM_INPUT (VOIDmode, "");
 | 
      
         | 374 |  |  |   MEM_VOLATILE_P (x) = true;
 | 
      
         | 375 |  |  |   return x;
 | 
      
         | 376 |  |  | }
 | 
      
         | 377 |  |  | #endif
 | 
      
         | 378 |  |  |  
 | 
      
         | 379 |  |  |  
 | 
      
         | 380 |  |  | /* Generate a new REG rtx.  Make sure ORIGINAL_REGNO is set properly, and
 | 
      
         | 381 |  |  |    don't attempt to share with the various global pieces of rtl (such as
 | 
      
         | 382 |  |  |    frame_pointer_rtx).  */
 | 
      
         | 383 |  |  |  
 | 
      
         | 384 |  |  | rtx
 | 
      
         | 385 |  |  | gen_raw_REG (enum machine_mode mode, int regno)
 | 
      
         | 386 |  |  | {
 | 
      
         | 387 |  |  |   rtx x = gen_rtx_raw_REG (mode, regno);
 | 
      
         | 388 |  |  |   ORIGINAL_REGNO (x) = regno;
 | 
      
         | 389 |  |  |   return x;
 | 
      
         | 390 |  |  | }
 | 
      
         | 391 |  |  |  
 | 
      
         | 392 |  |  | /* There are some RTL codes that require special attention; the generation
 | 
      
         | 393 |  |  |    functions do the raw handling.  If you add to this list, modify
 | 
      
         | 394 |  |  |    special_rtx in gengenrtl.c as well.  */
 | 
      
         | 395 |  |  |  
 | 
      
         | 396 |  |  | rtx
 | 
      
         | 397 |  |  | gen_rtx_CONST_INT (enum machine_mode mode ATTRIBUTE_UNUSED, HOST_WIDE_INT arg)
 | 
      
         | 398 |  |  | {
 | 
      
         | 399 |  |  |   void **slot;
 | 
      
         | 400 |  |  |  
 | 
      
         | 401 |  |  |   if (arg >= - MAX_SAVED_CONST_INT && arg <= MAX_SAVED_CONST_INT)
 | 
      
         | 402 |  |  |     return const_int_rtx[arg + MAX_SAVED_CONST_INT];
 | 
      
         | 403 |  |  |  
 | 
      
         | 404 |  |  | #if STORE_FLAG_VALUE != 1 && STORE_FLAG_VALUE != -1
 | 
      
         | 405 |  |  |   if (const_true_rtx && arg == STORE_FLAG_VALUE)
 | 
      
         | 406 |  |  |     return const_true_rtx;
 | 
      
         | 407 |  |  | #endif
 | 
      
         | 408 |  |  |  
 | 
      
         | 409 |  |  |   /* Look up the CONST_INT in the hash table.  */
 | 
      
         | 410 |  |  |   slot = htab_find_slot_with_hash (const_int_htab, &arg,
 | 
      
         | 411 |  |  |                                    (hashval_t) arg, INSERT);
 | 
      
         | 412 |  |  |   if (*slot == 0)
 | 
      
         | 413 |  |  |     *slot = gen_rtx_raw_CONST_INT (VOIDmode, arg);
 | 
      
         | 414 |  |  |  
 | 
      
         | 415 |  |  |   return (rtx) *slot;
 | 
      
         | 416 |  |  | }
 | 
      
         | 417 |  |  |  
 | 
      
         | 418 |  |  | rtx
 | 
      
         | 419 |  |  | gen_int_mode (HOST_WIDE_INT c, enum machine_mode mode)
 | 
      
         | 420 |  |  | {
 | 
      
         | 421 |  |  |   return GEN_INT (trunc_int_for_mode (c, mode));
 | 
      
         | 422 |  |  | }
 | 
      
         | 423 |  |  |  
 | 
      
         | 424 |  |  | /* CONST_DOUBLEs might be created from pairs of integers, or from
 | 
      
         | 425 |  |  |    REAL_VALUE_TYPEs.  Also, their length is known only at run time,
 | 
      
         | 426 |  |  |    so we cannot use gen_rtx_raw_CONST_DOUBLE.  */
 | 
      
         | 427 |  |  |  
 | 
      
         | 428 |  |  | /* Determine whether REAL, a CONST_DOUBLE, already exists in the
 | 
      
         | 429 |  |  |    hash table.  If so, return its counterpart; otherwise add it
 | 
      
         | 430 |  |  |    to the hash table and return it.  */
 | 
      
         | 431 |  |  | static rtx
 | 
      
         | 432 |  |  | lookup_const_double (rtx real)
 | 
      
         | 433 |  |  | {
 | 
      
         | 434 |  |  |   void **slot = htab_find_slot (const_double_htab, real, INSERT);
 | 
      
         | 435 |  |  |   if (*slot == 0)
 | 
      
         | 436 |  |  |     *slot = real;
 | 
      
         | 437 |  |  |  
 | 
      
         | 438 |  |  |   return (rtx) *slot;
 | 
      
         | 439 |  |  | }
 | 
      
         | 440 |  |  |  
 | 
      
         | 441 |  |  | /* Return a CONST_DOUBLE rtx for a floating-point value specified by
 | 
      
         | 442 |  |  |    VALUE in mode MODE.  */
 | 
      
         | 443 |  |  | rtx
 | 
      
         | 444 |  |  | const_double_from_real_value (REAL_VALUE_TYPE value, enum machine_mode mode)
 | 
      
         | 445 |  |  | {
 | 
      
         | 446 |  |  |   rtx real = rtx_alloc (CONST_DOUBLE);
 | 
      
         | 447 |  |  |   PUT_MODE (real, mode);
 | 
      
         | 448 |  |  |  
 | 
      
         | 449 |  |  |   real->u.rv = value;
 | 
      
         | 450 |  |  |  
 | 
      
         | 451 |  |  |   return lookup_const_double (real);
 | 
      
         | 452 |  |  | }
 | 
      
         | 453 |  |  |  
 | 
      
         | 454 |  |  | /* Determine whether FIXED, a CONST_FIXED, already exists in the
 | 
      
         | 455 |  |  |    hash table.  If so, return its counterpart; otherwise add it
 | 
      
         | 456 |  |  |    to the hash table and return it.  */
 | 
      
         | 457 |  |  |  
 | 
      
         | 458 |  |  | static rtx
 | 
      
         | 459 |  |  | lookup_const_fixed (rtx fixed)
 | 
      
         | 460 |  |  | {
 | 
      
         | 461 |  |  |   void **slot = htab_find_slot (const_fixed_htab, fixed, INSERT);
 | 
      
         | 462 |  |  |   if (*slot == 0)
 | 
      
         | 463 |  |  |     *slot = fixed;
 | 
      
         | 464 |  |  |  
 | 
      
         | 465 |  |  |   return (rtx) *slot;
 | 
      
         | 466 |  |  | }
 | 
      
         | 467 |  |  |  
 | 
      
         | 468 |  |  | /* Return a CONST_FIXED rtx for a fixed-point value specified by
 | 
      
         | 469 |  |  |    VALUE in mode MODE.  */
 | 
      
         | 470 |  |  |  
 | 
      
         | 471 |  |  | rtx
 | 
      
         | 472 |  |  | const_fixed_from_fixed_value (FIXED_VALUE_TYPE value, enum machine_mode mode)
 | 
      
         | 473 |  |  | {
 | 
      
         | 474 |  |  |   rtx fixed = rtx_alloc (CONST_FIXED);
 | 
      
         | 475 |  |  |   PUT_MODE (fixed, mode);
 | 
      
         | 476 |  |  |  
 | 
      
         | 477 |  |  |   fixed->u.fv = value;
 | 
      
         | 478 |  |  |  
 | 
      
         | 479 |  |  |   return lookup_const_fixed (fixed);
 | 
      
         | 480 |  |  | }
 | 
      
         | 481 |  |  |  
 | 
      
         | 482 |  |  | /* Constructs double_int from rtx CST.  */
 | 
      
         | 483 |  |  |  
 | 
      
         | 484 |  |  | double_int
 | 
      
         | 485 |  |  | rtx_to_double_int (const_rtx cst)
 | 
      
         | 486 |  |  | {
 | 
      
         | 487 |  |  |   double_int r;
 | 
      
         | 488 |  |  |  
 | 
      
         | 489 |  |  |   if (CONST_INT_P (cst))
 | 
      
         | 490 |  |  |       r = shwi_to_double_int (INTVAL (cst));
 | 
      
         | 491 |  |  |   else if (CONST_DOUBLE_P (cst) && GET_MODE (cst) == VOIDmode)
 | 
      
         | 492 |  |  |     {
 | 
      
         | 493 |  |  |       r.low = CONST_DOUBLE_LOW (cst);
 | 
      
         | 494 |  |  |       r.high = CONST_DOUBLE_HIGH (cst);
 | 
      
         | 495 |  |  |     }
 | 
      
         | 496 |  |  |   else
 | 
      
         | 497 |  |  |     gcc_unreachable ();
 | 
      
         | 498 |  |  |  
 | 
      
         | 499 |  |  |   return r;
 | 
      
         | 500 |  |  | }
 | 
      
         | 501 |  |  |  
 | 
      
         | 502 |  |  |  
 | 
      
         | 503 |  |  | /* Return a CONST_DOUBLE or CONST_INT for a value specified as
 | 
      
         | 504 |  |  |    a double_int.  */
 | 
      
         | 505 |  |  |  
 | 
      
         | 506 |  |  | rtx
 | 
      
         | 507 |  |  | immed_double_int_const (double_int i, enum machine_mode mode)
 | 
      
         | 508 |  |  | {
 | 
      
         | 509 |  |  |   return immed_double_const (i.low, i.high, mode);
 | 
      
         | 510 |  |  | }
 | 
      
         | 511 |  |  |  
 | 
      
         | 512 |  |  | /* Return a CONST_DOUBLE or CONST_INT for a value specified as a pair
 | 
      
         | 513 |  |  |    of ints: I0 is the low-order word and I1 is the high-order word.
 | 
      
         | 514 |  |  |    Do not use this routine for non-integer modes; convert to
 | 
      
         | 515 |  |  |    REAL_VALUE_TYPE and use CONST_DOUBLE_FROM_REAL_VALUE.  */
 | 
      
         | 516 |  |  |  
 | 
      
         | 517 |  |  | rtx
 | 
      
         | 518 |  |  | immed_double_const (HOST_WIDE_INT i0, HOST_WIDE_INT i1, enum machine_mode mode)
 | 
      
         | 519 |  |  | {
 | 
      
         | 520 |  |  |   rtx value;
 | 
      
         | 521 |  |  |   unsigned int i;
 | 
      
         | 522 |  |  |  
 | 
      
         | 523 |  |  |   /* There are the following cases (note that there are no modes with
 | 
      
         | 524 |  |  |      HOST_BITS_PER_WIDE_INT < GET_MODE_BITSIZE (mode) < 2 * HOST_BITS_PER_WIDE_INT):
 | 
      
         | 525 |  |  |  
 | 
      
         | 526 |  |  |      1) If GET_MODE_BITSIZE (mode) <= HOST_BITS_PER_WIDE_INT, then we use
 | 
      
         | 527 |  |  |         gen_int_mode.
 | 
      
         | 528 |  |  |      2) GET_MODE_BITSIZE (mode) == 2 * HOST_BITS_PER_WIDE_INT, but the value of
 | 
      
         | 529 |  |  |         the integer fits into HOST_WIDE_INT anyway (i.e., i1 consists only
 | 
      
         | 530 |  |  |         from copies of the sign bit, and sign of i0 and i1 are the same),  then
 | 
      
         | 531 |  |  |         we return a CONST_INT for i0.
 | 
      
         | 532 |  |  |      3) Otherwise, we create a CONST_DOUBLE for i0 and i1.  */
 | 
      
         | 533 |  |  |   if (mode != VOIDmode)
 | 
      
         | 534 |  |  |     {
 | 
      
         | 535 |  |  |       gcc_assert (GET_MODE_CLASS (mode) == MODE_INT
 | 
      
         | 536 |  |  |                   || GET_MODE_CLASS (mode) == MODE_PARTIAL_INT
 | 
      
         | 537 |  |  |                   /* We can get a 0 for an error mark.  */
 | 
      
         | 538 |  |  |                   || GET_MODE_CLASS (mode) == MODE_VECTOR_INT
 | 
      
         | 539 |  |  |                   || GET_MODE_CLASS (mode) == MODE_VECTOR_FLOAT);
 | 
      
         | 540 |  |  |  
 | 
      
         | 541 |  |  |       if (GET_MODE_BITSIZE (mode) <= HOST_BITS_PER_WIDE_INT)
 | 
      
         | 542 |  |  |         return gen_int_mode (i0, mode);
 | 
      
         | 543 |  |  |  
 | 
      
         | 544 |  |  |       gcc_assert (GET_MODE_BITSIZE (mode) == 2 * HOST_BITS_PER_WIDE_INT);
 | 
      
         | 545 |  |  |     }
 | 
      
         | 546 |  |  |  
 | 
      
         | 547 |  |  |   /* If this integer fits in one word, return a CONST_INT.  */
 | 
      
         | 548 |  |  |   if ((i1 == 0 && i0 >= 0) || (i1 == ~0 && i0 < 0))
 | 
      
         | 549 |  |  |     return GEN_INT (i0);
 | 
      
         | 550 |  |  |  
 | 
      
         | 551 |  |  |   /* We use VOIDmode for integers.  */
 | 
      
         | 552 |  |  |   value = rtx_alloc (CONST_DOUBLE);
 | 
      
         | 553 |  |  |   PUT_MODE (value, VOIDmode);
 | 
      
         | 554 |  |  |  
 | 
      
         | 555 |  |  |   CONST_DOUBLE_LOW (value) = i0;
 | 
      
         | 556 |  |  |   CONST_DOUBLE_HIGH (value) = i1;
 | 
      
         | 557 |  |  |  
 | 
      
         | 558 |  |  |   for (i = 2; i < (sizeof CONST_DOUBLE_FORMAT - 1); i++)
 | 
      
         | 559 |  |  |     XWINT (value, i) = 0;
 | 
      
         | 560 |  |  |  
 | 
      
         | 561 |  |  |   return lookup_const_double (value);
 | 
      
         | 562 |  |  | }
 | 
      
         | 563 |  |  |  
 | 
      
         | 564 |  |  | rtx
 | 
      
         | 565 |  |  | gen_rtx_REG (enum machine_mode mode, unsigned int regno)
 | 
      
         | 566 |  |  | {
 | 
      
         | 567 |  |  |   /* In case the MD file explicitly references the frame pointer, have
 | 
      
         | 568 |  |  |      all such references point to the same frame pointer.  This is
 | 
      
         | 569 |  |  |      used during frame pointer elimination to distinguish the explicit
 | 
      
         | 570 |  |  |      references to these registers from pseudos that happened to be
 | 
      
         | 571 |  |  |      assigned to them.
 | 
      
         | 572 |  |  |  
 | 
      
         | 573 |  |  |      If we have eliminated the frame pointer or arg pointer, we will
 | 
      
         | 574 |  |  |      be using it as a normal register, for example as a spill
 | 
      
         | 575 |  |  |      register.  In such cases, we might be accessing it in a mode that
 | 
      
         | 576 |  |  |      is not Pmode and therefore cannot use the pre-allocated rtx.
 | 
      
         | 577 |  |  |  
 | 
      
         | 578 |  |  |      Also don't do this when we are making new REGs in reload, since
 | 
      
         | 579 |  |  |      we don't want to get confused with the real pointers.  */
 | 
      
         | 580 |  |  |  
 | 
      
         | 581 |  |  |   if (mode == Pmode && !reload_in_progress)
 | 
      
         | 582 |  |  |     {
 | 
      
         | 583 |  |  |       if (regno == FRAME_POINTER_REGNUM
 | 
      
         | 584 |  |  |           && (!reload_completed || frame_pointer_needed))
 | 
      
         | 585 |  |  |         return frame_pointer_rtx;
 | 
      
         | 586 |  |  | #if !HARD_FRAME_POINTER_IS_FRAME_POINTER
 | 
      
         | 587 |  |  |       if (regno == HARD_FRAME_POINTER_REGNUM
 | 
      
         | 588 |  |  |           && (!reload_completed || frame_pointer_needed))
 | 
      
         | 589 |  |  |         return hard_frame_pointer_rtx;
 | 
      
         | 590 |  |  | #endif
 | 
      
         | 591 |  |  | #if FRAME_POINTER_REGNUM != ARG_POINTER_REGNUM && !HARD_FRAME_POINTER_IS_ARG_POINTER
 | 
      
         | 592 |  |  |       if (regno == ARG_POINTER_REGNUM)
 | 
      
         | 593 |  |  |         return arg_pointer_rtx;
 | 
      
         | 594 |  |  | #endif
 | 
      
         | 595 |  |  | #ifdef RETURN_ADDRESS_POINTER_REGNUM
 | 
      
         | 596 |  |  |       if (regno == RETURN_ADDRESS_POINTER_REGNUM)
 | 
      
         | 597 |  |  |         return return_address_pointer_rtx;
 | 
      
         | 598 |  |  | #endif
 | 
      
         | 599 |  |  |       if (regno == (unsigned) PIC_OFFSET_TABLE_REGNUM
 | 
      
         | 600 |  |  |           && PIC_OFFSET_TABLE_REGNUM != INVALID_REGNUM
 | 
      
         | 601 |  |  |           && fixed_regs[PIC_OFFSET_TABLE_REGNUM])
 | 
      
         | 602 |  |  |         return pic_offset_table_rtx;
 | 
      
         | 603 |  |  |       if (regno == STACK_POINTER_REGNUM)
 | 
      
         | 604 |  |  |         return stack_pointer_rtx;
 | 
      
         | 605 |  |  |     }
 | 
      
         | 606 |  |  |  
 | 
      
         | 607 |  |  | #if 0
 | 
      
         | 608 |  |  |   /* If the per-function register table has been set up, try to re-use
 | 
      
         | 609 |  |  |      an existing entry in that table to avoid useless generation of RTL.
 | 
      
         | 610 |  |  |  
 | 
      
         | 611 |  |  |      This code is disabled for now until we can fix the various backends
 | 
      
         | 612 |  |  |      which depend on having non-shared hard registers in some cases.   Long
 | 
      
         | 613 |  |  |      term we want to re-enable this code as it can significantly cut down
 | 
      
         | 614 |  |  |      on the amount of useless RTL that gets generated.
 | 
      
         | 615 |  |  |  
 | 
      
         | 616 |  |  |      We'll also need to fix some code that runs after reload that wants to
 | 
      
         | 617 |  |  |      set ORIGINAL_REGNO.  */
 | 
      
         | 618 |  |  |  
 | 
      
         | 619 |  |  |   if (cfun
 | 
      
         | 620 |  |  |       && cfun->emit
 | 
      
         | 621 |  |  |       && regno_reg_rtx
 | 
      
         | 622 |  |  |       && regno < FIRST_PSEUDO_REGISTER
 | 
      
         | 623 |  |  |       && reg_raw_mode[regno] == mode)
 | 
      
         | 624 |  |  |     return regno_reg_rtx[regno];
 | 
      
         | 625 |  |  | #endif
 | 
      
         | 626 |  |  |  
 | 
      
         | 627 |  |  |   return gen_raw_REG (mode, regno);
 | 
      
         | 628 |  |  | }
 | 
      
         | 629 |  |  |  
 | 
      
         | 630 |  |  | rtx
 | 
      
         | 631 |  |  | gen_rtx_MEM (enum machine_mode mode, rtx addr)
 | 
      
         | 632 |  |  | {
 | 
      
         | 633 |  |  |   rtx rt = gen_rtx_raw_MEM (mode, addr);
 | 
      
         | 634 |  |  |  
 | 
      
         | 635 |  |  |   /* This field is not cleared by the mere allocation of the rtx, so
 | 
      
         | 636 |  |  |      we clear it here.  */
 | 
      
         | 637 |  |  |   MEM_ATTRS (rt) = 0;
 | 
      
         | 638 |  |  |  
 | 
      
         | 639 |  |  |   return rt;
 | 
      
         | 640 |  |  | }
 | 
      
         | 641 |  |  |  
 | 
      
         | 642 |  |  | /* Generate a memory referring to non-trapping constant memory.  */
 | 
      
         | 643 |  |  |  
 | 
      
         | 644 |  |  | rtx
 | 
      
         | 645 |  |  | gen_const_mem (enum machine_mode mode, rtx addr)
 | 
      
         | 646 |  |  | {
 | 
      
         | 647 |  |  |   rtx mem = gen_rtx_MEM (mode, addr);
 | 
      
         | 648 |  |  |   MEM_READONLY_P (mem) = 1;
 | 
      
         | 649 |  |  |   MEM_NOTRAP_P (mem) = 1;
 | 
      
         | 650 |  |  |   return mem;
 | 
      
         | 651 |  |  | }
 | 
      
         | 652 |  |  |  
 | 
      
         | 653 |  |  | /* Generate a MEM referring to fixed portions of the frame, e.g., register
 | 
      
         | 654 |  |  |    save areas.  */
 | 
      
         | 655 |  |  |  
 | 
      
         | 656 |  |  | rtx
 | 
      
         | 657 |  |  | gen_frame_mem (enum machine_mode mode, rtx addr)
 | 
      
         | 658 |  |  | {
 | 
      
         | 659 |  |  |   rtx mem = gen_rtx_MEM (mode, addr);
 | 
      
         | 660 |  |  |   MEM_NOTRAP_P (mem) = 1;
 | 
      
         | 661 |  |  |   set_mem_alias_set (mem, get_frame_alias_set ());
 | 
      
         | 662 |  |  |   return mem;
 | 
      
         | 663 |  |  | }
 | 
      
         | 664 |  |  |  
 | 
      
         | 665 |  |  | /* Generate a MEM referring to a temporary use of the stack, not part
 | 
      
         | 666 |  |  |     of the fixed stack frame.  For example, something which is pushed
 | 
      
         | 667 |  |  |     by a target splitter.  */
 | 
      
         | 668 |  |  | rtx
 | 
      
         | 669 |  |  | gen_tmp_stack_mem (enum machine_mode mode, rtx addr)
 | 
      
         | 670 |  |  | {
 | 
      
         | 671 |  |  |   rtx mem = gen_rtx_MEM (mode, addr);
 | 
      
         | 672 |  |  |   MEM_NOTRAP_P (mem) = 1;
 | 
      
         | 673 |  |  |   if (!cfun->calls_alloca)
 | 
      
         | 674 |  |  |     set_mem_alias_set (mem, get_frame_alias_set ());
 | 
      
         | 675 |  |  |   return mem;
 | 
      
         | 676 |  |  | }
 | 
      
         | 677 |  |  |  
 | 
      
         | 678 |  |  | /* We want to create (subreg:OMODE (obj:IMODE) OFFSET).  Return true if
 | 
      
         | 679 |  |  |    this construct would be valid, and false otherwise.  */
 | 
      
         | 680 |  |  |  
 | 
      
         | 681 |  |  | bool
 | 
      
         | 682 |  |  | validate_subreg (enum machine_mode omode, enum machine_mode imode,
 | 
      
         | 683 |  |  |                  const_rtx reg, unsigned int offset)
 | 
      
         | 684 |  |  | {
 | 
      
         | 685 |  |  |   unsigned int isize = GET_MODE_SIZE (imode);
 | 
      
         | 686 |  |  |   unsigned int osize = GET_MODE_SIZE (omode);
 | 
      
         | 687 |  |  |  
 | 
      
         | 688 |  |  |   /* All subregs must be aligned.  */
 | 
      
         | 689 |  |  |   if (offset % osize != 0)
 | 
      
         | 690 |  |  |     return false;
 | 
      
         | 691 |  |  |  
 | 
      
         | 692 |  |  |   /* The subreg offset cannot be outside the inner object.  */
 | 
      
         | 693 |  |  |   if (offset >= isize)
 | 
      
         | 694 |  |  |     return false;
 | 
      
         | 695 |  |  |  
 | 
      
         | 696 |  |  |   /* ??? This should not be here.  Temporarily continue to allow word_mode
 | 
      
         | 697 |  |  |      subregs of anything.  The most common offender is (subreg:SI (reg:DF)).
 | 
      
         | 698 |  |  |      Generally, backends are doing something sketchy but it'll take time to
 | 
      
         | 699 |  |  |      fix them all.  */
 | 
      
         | 700 |  |  |   if (omode == word_mode)
 | 
      
         | 701 |  |  |     ;
 | 
      
         | 702 |  |  |   /* ??? Similarly, e.g. with (subreg:DF (reg:TI)).  Though store_bit_field
 | 
      
         | 703 |  |  |      is the culprit here, and not the backends.  */
 | 
      
         | 704 |  |  |   else if (osize >= UNITS_PER_WORD && isize >= osize)
 | 
      
         | 705 |  |  |     ;
 | 
      
         | 706 |  |  |   /* Allow component subregs of complex and vector.  Though given the below
 | 
      
         | 707 |  |  |      extraction rules, it's not always clear what that means.  */
 | 
      
         | 708 |  |  |   else if ((COMPLEX_MODE_P (imode) || VECTOR_MODE_P (imode))
 | 
      
         | 709 |  |  |            && GET_MODE_INNER (imode) == omode)
 | 
      
         | 710 |  |  |     ;
 | 
      
         | 711 |  |  |   /* ??? x86 sse code makes heavy use of *paradoxical* vector subregs,
 | 
      
         | 712 |  |  |      i.e. (subreg:V4SF (reg:SF) 0).  This surely isn't the cleanest way to
 | 
      
         | 713 |  |  |      represent this.  It's questionable if this ought to be represented at
 | 
      
         | 714 |  |  |      all -- why can't this all be hidden in post-reload splitters that make
 | 
      
         | 715 |  |  |      arbitrarily mode changes to the registers themselves.  */
 | 
      
         | 716 |  |  |   else if (VECTOR_MODE_P (omode) && GET_MODE_INNER (omode) == imode)
 | 
      
         | 717 |  |  |     ;
 | 
      
         | 718 |  |  |   /* Subregs involving floating point modes are not allowed to
 | 
      
         | 719 |  |  |      change size.  Therefore (subreg:DI (reg:DF) 0) is fine, but
 | 
      
         | 720 |  |  |      (subreg:SI (reg:DF) 0) isn't.  */
 | 
      
         | 721 |  |  |   else if (FLOAT_MODE_P (imode) || FLOAT_MODE_P (omode))
 | 
      
         | 722 |  |  |     {
 | 
      
         | 723 |  |  |       if (isize != osize)
 | 
      
         | 724 |  |  |         return false;
 | 
      
         | 725 |  |  |     }
 | 
      
         | 726 |  |  |  
 | 
      
         | 727 |  |  |   /* Paradoxical subregs must have offset zero.  */
 | 
      
         | 728 |  |  |   if (osize > isize)
 | 
      
         | 729 |  |  |     return offset == 0;
 | 
      
         | 730 |  |  |  
 | 
      
         | 731 |  |  |   /* This is a normal subreg.  Verify that the offset is representable.  */
 | 
      
         | 732 |  |  |  
 | 
      
         | 733 |  |  |   /* For hard registers, we already have most of these rules collected in
 | 
      
         | 734 |  |  |      subreg_offset_representable_p.  */
 | 
      
         | 735 |  |  |   if (reg && REG_P (reg) && HARD_REGISTER_P (reg))
 | 
      
         | 736 |  |  |     {
 | 
      
         | 737 |  |  |       unsigned int regno = REGNO (reg);
 | 
      
         | 738 |  |  |  
 | 
      
         | 739 |  |  | #ifdef CANNOT_CHANGE_MODE_CLASS
 | 
      
         | 740 |  |  |       if ((COMPLEX_MODE_P (imode) || VECTOR_MODE_P (imode))
 | 
      
         | 741 |  |  |           && GET_MODE_INNER (imode) == omode)
 | 
      
         | 742 |  |  |         ;
 | 
      
         | 743 |  |  |       else if (REG_CANNOT_CHANGE_MODE_P (regno, imode, omode))
 | 
      
         | 744 |  |  |         return false;
 | 
      
         | 745 |  |  | #endif
 | 
      
         | 746 |  |  |  
 | 
      
         | 747 |  |  |       return subreg_offset_representable_p (regno, imode, offset, omode);
 | 
      
         | 748 |  |  |     }
 | 
      
         | 749 |  |  |  
 | 
      
         | 750 |  |  |   /* For pseudo registers, we want most of the same checks.  Namely:
 | 
      
         | 751 |  |  |      If the register no larger than a word, the subreg must be lowpart.
 | 
      
         | 752 |  |  |      If the register is larger than a word, the subreg must be the lowpart
 | 
      
         | 753 |  |  |      of a subword.  A subreg does *not* perform arbitrary bit extraction.
 | 
      
         | 754 |  |  |      Given that we've already checked mode/offset alignment, we only have
 | 
      
         | 755 |  |  |      to check subword subregs here.  */
 | 
      
         | 756 |  |  |   if (osize < UNITS_PER_WORD)
 | 
      
         | 757 |  |  |     {
 | 
      
         | 758 |  |  |       enum machine_mode wmode = isize > UNITS_PER_WORD ? word_mode : imode;
 | 
      
         | 759 |  |  |       unsigned int low_off = subreg_lowpart_offset (omode, wmode);
 | 
      
         | 760 |  |  |       if (offset % UNITS_PER_WORD != low_off)
 | 
      
         | 761 |  |  |         return false;
 | 
      
         | 762 |  |  |     }
 | 
      
         | 763 |  |  |   return true;
 | 
      
         | 764 |  |  | }
 | 
      
         | 765 |  |  |  
 | 
      
         | 766 |  |  | rtx
 | 
      
         | 767 |  |  | gen_rtx_SUBREG (enum machine_mode mode, rtx reg, int offset)
 | 
      
         | 768 |  |  | {
 | 
      
         | 769 |  |  |   gcc_assert (validate_subreg (mode, GET_MODE (reg), reg, offset));
 | 
      
         | 770 |  |  |   return gen_rtx_raw_SUBREG (mode, reg, offset);
 | 
      
         | 771 |  |  | }
 | 
      
         | 772 |  |  |  
 | 
      
         | 773 |  |  | /* Generate a SUBREG representing the least-significant part of REG if MODE
 | 
      
         | 774 |  |  |    is smaller than mode of REG, otherwise paradoxical SUBREG.  */
 | 
      
         | 775 |  |  |  
 | 
      
         | 776 |  |  | rtx
 | 
      
         | 777 |  |  | gen_lowpart_SUBREG (enum machine_mode mode, rtx reg)
 | 
      
         | 778 |  |  | {
 | 
      
         | 779 |  |  |   enum machine_mode inmode;
 | 
      
         | 780 |  |  |  
 | 
      
         | 781 |  |  |   inmode = GET_MODE (reg);
 | 
      
         | 782 |  |  |   if (inmode == VOIDmode)
 | 
      
         | 783 |  |  |     inmode = mode;
 | 
      
         | 784 |  |  |   return gen_rtx_SUBREG (mode, reg,
 | 
      
         | 785 |  |  |                          subreg_lowpart_offset (mode, inmode));
 | 
      
         | 786 |  |  | }
 | 
      
         | 787 |  |  |  
 | 
      
         | 788 |  |  |  
 | 
      
         | 789 |  |  | /* Create an rtvec and stores within it the RTXen passed in the arguments.  */
 | 
      
         | 790 |  |  |  
 | 
      
         | 791 |  |  | rtvec
 | 
      
         | 792 |  |  | gen_rtvec (int n, ...)
 | 
      
         | 793 |  |  | {
 | 
      
         | 794 |  |  |   int i;
 | 
      
         | 795 |  |  |   rtvec rt_val;
 | 
      
         | 796 |  |  |   va_list p;
 | 
      
         | 797 |  |  |  
 | 
      
         | 798 |  |  |   va_start (p, n);
 | 
      
         | 799 |  |  |  
 | 
      
         | 800 |  |  |   /* Don't allocate an empty rtvec...  */
 | 
      
         | 801 |  |  |   if (n == 0)
 | 
      
         | 802 |  |  |     {
 | 
      
         | 803 |  |  |       va_end (p);
 | 
      
         | 804 |  |  |       return NULL_RTVEC;
 | 
      
         | 805 |  |  |     }
 | 
      
         | 806 |  |  |  
 | 
      
         | 807 |  |  |   rt_val = rtvec_alloc (n);
 | 
      
         | 808 |  |  |  
 | 
      
         | 809 |  |  |   for (i = 0; i < n; i++)
 | 
      
         | 810 |  |  |     rt_val->elem[i] = va_arg (p, rtx);
 | 
      
         | 811 |  |  |  
 | 
      
         | 812 |  |  |   va_end (p);
 | 
      
         | 813 |  |  |   return rt_val;
 | 
      
         | 814 |  |  | }
 | 
      
         | 815 |  |  |  
 | 
      
         | 816 |  |  | rtvec
 | 
      
         | 817 |  |  | gen_rtvec_v (int n, rtx *argp)
 | 
      
         | 818 |  |  | {
 | 
      
         | 819 |  |  |   int i;
 | 
      
         | 820 |  |  |   rtvec rt_val;
 | 
      
         | 821 |  |  |  
 | 
      
         | 822 |  |  |   /* Don't allocate an empty rtvec...  */
 | 
      
         | 823 |  |  |   if (n == 0)
 | 
      
         | 824 |  |  |     return NULL_RTVEC;
 | 
      
         | 825 |  |  |  
 | 
      
         | 826 |  |  |   rt_val = rtvec_alloc (n);
 | 
      
         | 827 |  |  |  
 | 
      
         | 828 |  |  |   for (i = 0; i < n; i++)
 | 
      
         | 829 |  |  |     rt_val->elem[i] = *argp++;
 | 
      
         | 830 |  |  |  
 | 
      
         | 831 |  |  |   return rt_val;
 | 
      
         | 832 |  |  | }
 | 
      
         | 833 |  |  |  
 | 
      
         | 834 |  |  | /* Return the number of bytes between the start of an OUTER_MODE
 | 
      
         | 835 |  |  |    in-memory value and the start of an INNER_MODE in-memory value,
 | 
      
         | 836 |  |  |    given that the former is a lowpart of the latter.  It may be a
 | 
      
         | 837 |  |  |    paradoxical lowpart, in which case the offset will be negative
 | 
      
         | 838 |  |  |    on big-endian targets.  */
 | 
      
         | 839 |  |  |  
 | 
      
         | 840 |  |  | int
 | 
      
         | 841 |  |  | byte_lowpart_offset (enum machine_mode outer_mode,
 | 
      
         | 842 |  |  |                      enum machine_mode inner_mode)
 | 
      
         | 843 |  |  | {
 | 
      
         | 844 |  |  |   if (GET_MODE_SIZE (outer_mode) < GET_MODE_SIZE (inner_mode))
 | 
      
         | 845 |  |  |     return subreg_lowpart_offset (outer_mode, inner_mode);
 | 
      
         | 846 |  |  |   else
 | 
      
         | 847 |  |  |     return -subreg_lowpart_offset (inner_mode, outer_mode);
 | 
      
         | 848 |  |  | }
 | 
      
         | 849 |  |  |  
 | 
      
         | 850 |  |  | /* Generate a REG rtx for a new pseudo register of mode MODE.
 | 
      
         | 851 |  |  |    This pseudo is assigned the next sequential register number.  */
 | 
      
         | 852 |  |  |  
 | 
      
         | 853 |  |  | rtx
 | 
      
         | 854 |  |  | gen_reg_rtx (enum machine_mode mode)
 | 
      
         | 855 |  |  | {
 | 
      
         | 856 |  |  |   rtx val;
 | 
      
         | 857 |  |  |   unsigned int align = GET_MODE_ALIGNMENT (mode);
 | 
      
         | 858 |  |  |  
 | 
      
         | 859 |  |  |   gcc_assert (can_create_pseudo_p ());
 | 
      
         | 860 |  |  |  
 | 
      
         | 861 |  |  |   /* If a virtual register with bigger mode alignment is generated,
 | 
      
         | 862 |  |  |      increase stack alignment estimation because it might be spilled
 | 
      
         | 863 |  |  |      to stack later.  */
 | 
      
         | 864 |  |  |   if (SUPPORTS_STACK_ALIGNMENT
 | 
      
         | 865 |  |  |       && crtl->stack_alignment_estimated < align
 | 
      
         | 866 |  |  |       && !crtl->stack_realign_processed)
 | 
      
         | 867 |  |  |     {
 | 
      
         | 868 |  |  |       unsigned int min_align = MINIMUM_ALIGNMENT (NULL, mode, align);
 | 
      
         | 869 |  |  |       if (crtl->stack_alignment_estimated < min_align)
 | 
      
         | 870 |  |  |         crtl->stack_alignment_estimated = min_align;
 | 
      
         | 871 |  |  |     }
 | 
      
         | 872 |  |  |  
 | 
      
         | 873 |  |  |   if (generating_concat_p
 | 
      
         | 874 |  |  |       && (GET_MODE_CLASS (mode) == MODE_COMPLEX_FLOAT
 | 
      
         | 875 |  |  |           || GET_MODE_CLASS (mode) == MODE_COMPLEX_INT))
 | 
      
         | 876 |  |  |     {
 | 
      
         | 877 |  |  |       /* For complex modes, don't make a single pseudo.
 | 
      
         | 878 |  |  |          Instead, make a CONCAT of two pseudos.
 | 
      
         | 879 |  |  |          This allows noncontiguous allocation of the real and imaginary parts,
 | 
      
         | 880 |  |  |          which makes much better code.  Besides, allocating DCmode
 | 
      
         | 881 |  |  |          pseudos overstrains reload on some machines like the 386.  */
 | 
      
         | 882 |  |  |       rtx realpart, imagpart;
 | 
      
         | 883 |  |  |       enum machine_mode partmode = GET_MODE_INNER (mode);
 | 
      
         | 884 |  |  |  
 | 
      
         | 885 |  |  |       realpart = gen_reg_rtx (partmode);
 | 
      
         | 886 |  |  |       imagpart = gen_reg_rtx (partmode);
 | 
      
         | 887 |  |  |       return gen_rtx_CONCAT (mode, realpart, imagpart);
 | 
      
         | 888 |  |  |     }
 | 
      
         | 889 |  |  |  
 | 
      
         | 890 |  |  |   /* Make sure regno_pointer_align, and regno_reg_rtx are large
 | 
      
         | 891 |  |  |      enough to have an element for this pseudo reg number.  */
 | 
      
         | 892 |  |  |  
 | 
      
         | 893 |  |  |   if (reg_rtx_no == crtl->emit.regno_pointer_align_length)
 | 
      
         | 894 |  |  |     {
 | 
      
         | 895 |  |  |       int old_size = crtl->emit.regno_pointer_align_length;
 | 
      
         | 896 |  |  |       char *tmp;
 | 
      
         | 897 |  |  |       rtx *new1;
 | 
      
         | 898 |  |  |  
 | 
      
         | 899 |  |  |       tmp = XRESIZEVEC (char, crtl->emit.regno_pointer_align, old_size * 2);
 | 
      
         | 900 |  |  |       memset (tmp + old_size, 0, old_size);
 | 
      
         | 901 |  |  |       crtl->emit.regno_pointer_align = (unsigned char *) tmp;
 | 
      
         | 902 |  |  |  
 | 
      
         | 903 |  |  |       new1 = GGC_RESIZEVEC (rtx, regno_reg_rtx, old_size * 2);
 | 
      
         | 904 |  |  |       memset (new1 + old_size, 0, old_size * sizeof (rtx));
 | 
      
         | 905 |  |  |       regno_reg_rtx = new1;
 | 
      
         | 906 |  |  |  
 | 
      
         | 907 |  |  |       crtl->emit.regno_pointer_align_length = old_size * 2;
 | 
      
         | 908 |  |  |     }
 | 
      
         | 909 |  |  |  
 | 
      
         | 910 |  |  |   val = gen_raw_REG (mode, reg_rtx_no);
 | 
      
         | 911 |  |  |   regno_reg_rtx[reg_rtx_no++] = val;
 | 
      
         | 912 |  |  |   return val;
 | 
      
         | 913 |  |  | }
 | 
      
         | 914 |  |  |  
 | 
      
         | 915 |  |  | /* Update NEW with the same attributes as REG, but with OFFSET added
 | 
      
         | 916 |  |  |    to the REG_OFFSET.  */
 | 
      
         | 917 |  |  |  
 | 
      
         | 918 |  |  | static void
 | 
      
         | 919 |  |  | update_reg_offset (rtx new_rtx, rtx reg, int offset)
 | 
      
         | 920 |  |  | {
 | 
      
         | 921 |  |  |   REG_ATTRS (new_rtx) = get_reg_attrs (REG_EXPR (reg),
 | 
      
         | 922 |  |  |                                    REG_OFFSET (reg) + offset);
 | 
      
         | 923 |  |  | }
 | 
      
         | 924 |  |  |  
 | 
      
         | 925 |  |  | /* Generate a register with same attributes as REG, but with OFFSET
 | 
      
         | 926 |  |  |    added to the REG_OFFSET.  */
 | 
      
         | 927 |  |  |  
 | 
      
         | 928 |  |  | rtx
 | 
      
         | 929 |  |  | gen_rtx_REG_offset (rtx reg, enum machine_mode mode, unsigned int regno,
 | 
      
         | 930 |  |  |                     int offset)
 | 
      
         | 931 |  |  | {
 | 
      
         | 932 |  |  |   rtx new_rtx = gen_rtx_REG (mode, regno);
 | 
      
         | 933 |  |  |  
 | 
      
         | 934 |  |  |   update_reg_offset (new_rtx, reg, offset);
 | 
      
         | 935 |  |  |   return new_rtx;
 | 
      
         | 936 |  |  | }
 | 
      
         | 937 |  |  |  
 | 
      
         | 938 |  |  | /* Generate a new pseudo-register with the same attributes as REG, but
 | 
      
         | 939 |  |  |    with OFFSET added to the REG_OFFSET.  */
 | 
      
         | 940 |  |  |  
 | 
      
         | 941 |  |  | rtx
 | 
      
         | 942 |  |  | gen_reg_rtx_offset (rtx reg, enum machine_mode mode, int offset)
 | 
      
         | 943 |  |  | {
 | 
      
         | 944 |  |  |   rtx new_rtx = gen_reg_rtx (mode);
 | 
      
         | 945 |  |  |  
 | 
      
         | 946 |  |  |   update_reg_offset (new_rtx, reg, offset);
 | 
      
         | 947 |  |  |   return new_rtx;
 | 
      
         | 948 |  |  | }
 | 
      
         | 949 |  |  |  
 | 
      
         | 950 |  |  | /* Adjust REG in-place so that it has mode MODE.  It is assumed that the
 | 
      
         | 951 |  |  |    new register is a (possibly paradoxical) lowpart of the old one.  */
 | 
      
         | 952 |  |  |  
 | 
      
         | 953 |  |  | void
 | 
      
         | 954 |  |  | adjust_reg_mode (rtx reg, enum machine_mode mode)
 | 
      
         | 955 |  |  | {
 | 
      
         | 956 |  |  |   update_reg_offset (reg, reg, byte_lowpart_offset (mode, GET_MODE (reg)));
 | 
      
         | 957 |  |  |   PUT_MODE (reg, mode);
 | 
      
         | 958 |  |  | }
 | 
      
         | 959 |  |  |  
 | 
      
         | 960 |  |  | /* Copy REG's attributes from X, if X has any attributes.  If REG and X
 | 
      
         | 961 |  |  |    have different modes, REG is a (possibly paradoxical) lowpart of X.  */
 | 
      
         | 962 |  |  |  
 | 
      
         | 963 |  |  | void
 | 
      
         | 964 |  |  | set_reg_attrs_from_value (rtx reg, rtx x)
 | 
      
         | 965 |  |  | {
 | 
      
         | 966 |  |  |   int offset;
 | 
      
         | 967 |  |  |  
 | 
      
         | 968 |  |  |   /* Hard registers can be reused for multiple purposes within the same
 | 
      
         | 969 |  |  |      function, so setting REG_ATTRS, REG_POINTER and REG_POINTER_ALIGN
 | 
      
         | 970 |  |  |      on them is wrong.  */
 | 
      
         | 971 |  |  |   if (HARD_REGISTER_P (reg))
 | 
      
         | 972 |  |  |     return;
 | 
      
         | 973 |  |  |  
 | 
      
         | 974 |  |  |   offset = byte_lowpart_offset (GET_MODE (reg), GET_MODE (x));
 | 
      
         | 975 |  |  |   if (MEM_P (x))
 | 
      
         | 976 |  |  |     {
 | 
      
         | 977 |  |  |       if (MEM_OFFSET_KNOWN_P (x))
 | 
      
         | 978 |  |  |         REG_ATTRS (reg) = get_reg_attrs (MEM_EXPR (x),
 | 
      
         | 979 |  |  |                                          MEM_OFFSET (x) + offset);
 | 
      
         | 980 |  |  |       if (MEM_POINTER (x))
 | 
      
         | 981 |  |  |         mark_reg_pointer (reg, 0);
 | 
      
         | 982 |  |  |     }
 | 
      
         | 983 |  |  |   else if (REG_P (x))
 | 
      
         | 984 |  |  |     {
 | 
      
         | 985 |  |  |       if (REG_ATTRS (x))
 | 
      
         | 986 |  |  |         update_reg_offset (reg, x, offset);
 | 
      
         | 987 |  |  |       if (REG_POINTER (x))
 | 
      
         | 988 |  |  |         mark_reg_pointer (reg, REGNO_POINTER_ALIGN (REGNO (x)));
 | 
      
         | 989 |  |  |     }
 | 
      
         | 990 |  |  | }
 | 
      
         | 991 |  |  |  
 | 
      
         | 992 |  |  | /* Generate a REG rtx for a new pseudo register, copying the mode
 | 
      
         | 993 |  |  |    and attributes from X.  */
 | 
      
         | 994 |  |  |  
 | 
      
         | 995 |  |  | rtx
 | 
      
         | 996 |  |  | gen_reg_rtx_and_attrs (rtx x)
 | 
      
         | 997 |  |  | {
 | 
      
         | 998 |  |  |   rtx reg = gen_reg_rtx (GET_MODE (x));
 | 
      
         | 999 |  |  |   set_reg_attrs_from_value (reg, x);
 | 
      
         | 1000 |  |  |   return reg;
 | 
      
         | 1001 |  |  | }
 | 
      
         | 1002 |  |  |  
 | 
      
         | 1003 |  |  | /* Set the register attributes for registers contained in PARM_RTX.
 | 
      
         | 1004 |  |  |    Use needed values from memory attributes of MEM.  */
 | 
      
         | 1005 |  |  |  
 | 
      
         | 1006 |  |  | void
 | 
      
         | 1007 |  |  | set_reg_attrs_for_parm (rtx parm_rtx, rtx mem)
 | 
      
         | 1008 |  |  | {
 | 
      
         | 1009 |  |  |   if (REG_P (parm_rtx))
 | 
      
         | 1010 |  |  |     set_reg_attrs_from_value (parm_rtx, mem);
 | 
      
         | 1011 |  |  |   else if (GET_CODE (parm_rtx) == PARALLEL)
 | 
      
         | 1012 |  |  |     {
 | 
      
         | 1013 |  |  |       /* Check for a NULL entry in the first slot, used to indicate that the
 | 
      
         | 1014 |  |  |          parameter goes both on the stack and in registers.  */
 | 
      
         | 1015 |  |  |       int i = XEXP (XVECEXP (parm_rtx, 0, 0), 0) ? 0 : 1;
 | 
      
         | 1016 |  |  |       for (; i < XVECLEN (parm_rtx, 0); i++)
 | 
      
         | 1017 |  |  |         {
 | 
      
         | 1018 |  |  |           rtx x = XVECEXP (parm_rtx, 0, i);
 | 
      
         | 1019 |  |  |           if (REG_P (XEXP (x, 0)))
 | 
      
         | 1020 |  |  |             REG_ATTRS (XEXP (x, 0))
 | 
      
         | 1021 |  |  |               = get_reg_attrs (MEM_EXPR (mem),
 | 
      
         | 1022 |  |  |                                INTVAL (XEXP (x, 1)));
 | 
      
         | 1023 |  |  |         }
 | 
      
         | 1024 |  |  |     }
 | 
      
         | 1025 |  |  | }
 | 
      
         | 1026 |  |  |  
 | 
      
         | 1027 |  |  | /* Set the REG_ATTRS for registers in value X, given that X represents
 | 
      
         | 1028 |  |  |    decl T.  */
 | 
      
         | 1029 |  |  |  
 | 
      
         | 1030 |  |  | void
 | 
      
         | 1031 |  |  | set_reg_attrs_for_decl_rtl (tree t, rtx x)
 | 
      
         | 1032 |  |  | {
 | 
      
         | 1033 |  |  |   if (GET_CODE (x) == SUBREG)
 | 
      
         | 1034 |  |  |     {
 | 
      
         | 1035 |  |  |       gcc_assert (subreg_lowpart_p (x));
 | 
      
         | 1036 |  |  |       x = SUBREG_REG (x);
 | 
      
         | 1037 |  |  |     }
 | 
      
         | 1038 |  |  |   if (REG_P (x))
 | 
      
         | 1039 |  |  |     REG_ATTRS (x)
 | 
      
         | 1040 |  |  |       = get_reg_attrs (t, byte_lowpart_offset (GET_MODE (x),
 | 
      
         | 1041 |  |  |                                                DECL_MODE (t)));
 | 
      
         | 1042 |  |  |   if (GET_CODE (x) == CONCAT)
 | 
      
         | 1043 |  |  |     {
 | 
      
         | 1044 |  |  |       if (REG_P (XEXP (x, 0)))
 | 
      
         | 1045 |  |  |         REG_ATTRS (XEXP (x, 0)) = get_reg_attrs (t, 0);
 | 
      
         | 1046 |  |  |       if (REG_P (XEXP (x, 1)))
 | 
      
         | 1047 |  |  |         REG_ATTRS (XEXP (x, 1))
 | 
      
         | 1048 |  |  |           = get_reg_attrs (t, GET_MODE_UNIT_SIZE (GET_MODE (XEXP (x, 0))));
 | 
      
         | 1049 |  |  |     }
 | 
      
         | 1050 |  |  |   if (GET_CODE (x) == PARALLEL)
 | 
      
         | 1051 |  |  |     {
 | 
      
         | 1052 |  |  |       int i, start;
 | 
      
         | 1053 |  |  |  
 | 
      
         | 1054 |  |  |       /* Check for a NULL entry, used to indicate that the parameter goes
 | 
      
         | 1055 |  |  |          both on the stack and in registers.  */
 | 
      
         | 1056 |  |  |       if (XEXP (XVECEXP (x, 0, 0), 0))
 | 
      
         | 1057 |  |  |         start = 0;
 | 
      
         | 1058 |  |  |       else
 | 
      
         | 1059 |  |  |         start = 1;
 | 
      
         | 1060 |  |  |  
 | 
      
         | 1061 |  |  |       for (i = start; i < XVECLEN (x, 0); i++)
 | 
      
         | 1062 |  |  |         {
 | 
      
         | 1063 |  |  |           rtx y = XVECEXP (x, 0, i);
 | 
      
         | 1064 |  |  |           if (REG_P (XEXP (y, 0)))
 | 
      
         | 1065 |  |  |             REG_ATTRS (XEXP (y, 0)) = get_reg_attrs (t, INTVAL (XEXP (y, 1)));
 | 
      
         | 1066 |  |  |         }
 | 
      
         | 1067 |  |  |     }
 | 
      
         | 1068 |  |  | }
 | 
      
         | 1069 |  |  |  
 | 
      
         | 1070 |  |  | /* Assign the RTX X to declaration T.  */
 | 
      
         | 1071 |  |  |  
 | 
      
         | 1072 |  |  | void
 | 
      
         | 1073 |  |  | set_decl_rtl (tree t, rtx x)
 | 
      
         | 1074 |  |  | {
 | 
      
         | 1075 |  |  |   DECL_WRTL_CHECK (t)->decl_with_rtl.rtl = x;
 | 
      
         | 1076 |  |  |   if (x)
 | 
      
         | 1077 |  |  |     set_reg_attrs_for_decl_rtl (t, x);
 | 
      
         | 1078 |  |  | }
 | 
      
         | 1079 |  |  |  
 | 
      
         | 1080 |  |  | /* Assign the RTX X to parameter declaration T.  BY_REFERENCE_P is true
 | 
      
         | 1081 |  |  |    if the ABI requires the parameter to be passed by reference.  */
 | 
      
         | 1082 |  |  |  
 | 
      
         | 1083 |  |  | void
 | 
      
         | 1084 |  |  | set_decl_incoming_rtl (tree t, rtx x, bool by_reference_p)
 | 
      
         | 1085 |  |  | {
 | 
      
         | 1086 |  |  |   DECL_INCOMING_RTL (t) = x;
 | 
      
         | 1087 |  |  |   if (x && !by_reference_p)
 | 
      
         | 1088 |  |  |     set_reg_attrs_for_decl_rtl (t, x);
 | 
      
         | 1089 |  |  | }
 | 
      
         | 1090 |  |  |  
 | 
      
         | 1091 |  |  | /* Identify REG (which may be a CONCAT) as a user register.  */
 | 
      
         | 1092 |  |  |  
 | 
      
         | 1093 |  |  | void
 | 
      
         | 1094 |  |  | mark_user_reg (rtx reg)
 | 
      
         | 1095 |  |  | {
 | 
      
         | 1096 |  |  |   if (GET_CODE (reg) == CONCAT)
 | 
      
         | 1097 |  |  |     {
 | 
      
         | 1098 |  |  |       REG_USERVAR_P (XEXP (reg, 0)) = 1;
 | 
      
         | 1099 |  |  |       REG_USERVAR_P (XEXP (reg, 1)) = 1;
 | 
      
         | 1100 |  |  |     }
 | 
      
         | 1101 |  |  |   else
 | 
      
         | 1102 |  |  |     {
 | 
      
         | 1103 |  |  |       gcc_assert (REG_P (reg));
 | 
      
         | 1104 |  |  |       REG_USERVAR_P (reg) = 1;
 | 
      
         | 1105 |  |  |     }
 | 
      
         | 1106 |  |  | }
 | 
      
         | 1107 |  |  |  
 | 
      
         | 1108 |  |  | /* Identify REG as a probable pointer register and show its alignment
 | 
      
         | 1109 |  |  |    as ALIGN, if nonzero.  */
 | 
      
         | 1110 |  |  |  
 | 
      
         | 1111 |  |  | void
 | 
      
         | 1112 |  |  | mark_reg_pointer (rtx reg, int align)
 | 
      
         | 1113 |  |  | {
 | 
      
         | 1114 |  |  |   if (! REG_POINTER (reg))
 | 
      
         | 1115 |  |  |     {
 | 
      
         | 1116 |  |  |       REG_POINTER (reg) = 1;
 | 
      
         | 1117 |  |  |  
 | 
      
         | 1118 |  |  |       if (align)
 | 
      
         | 1119 |  |  |         REGNO_POINTER_ALIGN (REGNO (reg)) = align;
 | 
      
         | 1120 |  |  |     }
 | 
      
         | 1121 |  |  |   else if (align && align < REGNO_POINTER_ALIGN (REGNO (reg)))
 | 
      
         | 1122 |  |  |     /* We can no-longer be sure just how aligned this pointer is.  */
 | 
      
         | 1123 |  |  |     REGNO_POINTER_ALIGN (REGNO (reg)) = align;
 | 
      
         | 1124 |  |  | }
 | 
      
         | 1125 |  |  |  
 | 
      
         | 1126 |  |  | /* Return 1 plus largest pseudo reg number used in the current function.  */
 | 
      
         | 1127 |  |  |  
 | 
      
         | 1128 |  |  | int
 | 
      
         | 1129 |  |  | max_reg_num (void)
 | 
      
         | 1130 |  |  | {
 | 
      
         | 1131 |  |  |   return reg_rtx_no;
 | 
      
         | 1132 |  |  | }
 | 
      
         | 1133 |  |  |  
 | 
      
         | 1134 |  |  | /* Return 1 + the largest label number used so far in the current function.  */
 | 
      
         | 1135 |  |  |  
 | 
      
         | 1136 |  |  | int
 | 
      
         | 1137 |  |  | max_label_num (void)
 | 
      
         | 1138 |  |  | {
 | 
      
         | 1139 |  |  |   return label_num;
 | 
      
         | 1140 |  |  | }
 | 
      
         | 1141 |  |  |  
 | 
      
         | 1142 |  |  | /* Return first label number used in this function (if any were used).  */
 | 
      
         | 1143 |  |  |  
 | 
      
         | 1144 |  |  | int
 | 
      
         | 1145 |  |  | get_first_label_num (void)
 | 
      
         | 1146 |  |  | {
 | 
      
         | 1147 |  |  |   return first_label_num;
 | 
      
         | 1148 |  |  | }
 | 
      
         | 1149 |  |  |  
 | 
      
         | 1150 |  |  | /* If the rtx for label was created during the expansion of a nested
 | 
      
         | 1151 |  |  |    function, then first_label_num won't include this label number.
 | 
      
         | 1152 |  |  |    Fix this now so that array indices work later.  */
 | 
      
         | 1153 |  |  |  
 | 
      
         | 1154 |  |  | void
 | 
      
         | 1155 |  |  | maybe_set_first_label_num (rtx x)
 | 
      
         | 1156 |  |  | {
 | 
      
         | 1157 |  |  |   if (CODE_LABEL_NUMBER (x) < first_label_num)
 | 
      
         | 1158 |  |  |     first_label_num = CODE_LABEL_NUMBER (x);
 | 
      
         | 1159 |  |  | }
 | 
      
         | 1160 |  |  |  
 | 
      
         | 1161 |  |  | /* Return a value representing some low-order bits of X, where the number
 | 
      
         | 1162 |  |  |    of low-order bits is given by MODE.  Note that no conversion is done
 | 
      
         | 1163 |  |  |    between floating-point and fixed-point values, rather, the bit
 | 
      
         | 1164 |  |  |    representation is returned.
 | 
      
         | 1165 |  |  |  
 | 
      
         | 1166 |  |  |    This function handles the cases in common between gen_lowpart, below,
 | 
      
         | 1167 |  |  |    and two variants in cse.c and combine.c.  These are the cases that can
 | 
      
         | 1168 |  |  |    be safely handled at all points in the compilation.
 | 
      
         | 1169 |  |  |  
 | 
      
         | 1170 |  |  |    If this is not a case we can handle, return 0.  */
 | 
      
         | 1171 |  |  |  
 | 
      
         | 1172 |  |  | rtx
 | 
      
         | 1173 |  |  | gen_lowpart_common (enum machine_mode mode, rtx x)
 | 
      
         | 1174 |  |  | {
 | 
      
         | 1175 |  |  |   int msize = GET_MODE_SIZE (mode);
 | 
      
         | 1176 |  |  |   int xsize;
 | 
      
         | 1177 |  |  |   int offset = 0;
 | 
      
         | 1178 |  |  |   enum machine_mode innermode;
 | 
      
         | 1179 |  |  |  
 | 
      
         | 1180 |  |  |   /* Unfortunately, this routine doesn't take a parameter for the mode of X,
 | 
      
         | 1181 |  |  |      so we have to make one up.  Yuk.  */
 | 
      
         | 1182 |  |  |   innermode = GET_MODE (x);
 | 
      
         | 1183 |  |  |   if (CONST_INT_P (x)
 | 
      
         | 1184 |  |  |       && msize * BITS_PER_UNIT <= HOST_BITS_PER_WIDE_INT)
 | 
      
         | 1185 |  |  |     innermode = mode_for_size (HOST_BITS_PER_WIDE_INT, MODE_INT, 0);
 | 
      
         | 1186 |  |  |   else if (innermode == VOIDmode)
 | 
      
         | 1187 |  |  |     innermode = mode_for_size (HOST_BITS_PER_WIDE_INT * 2, MODE_INT, 0);
 | 
      
         | 1188 |  |  |  
 | 
      
         | 1189 |  |  |   xsize = GET_MODE_SIZE (innermode);
 | 
      
         | 1190 |  |  |  
 | 
      
         | 1191 |  |  |   gcc_assert (innermode != VOIDmode && innermode != BLKmode);
 | 
      
         | 1192 |  |  |  
 | 
      
         | 1193 |  |  |   if (innermode == mode)
 | 
      
         | 1194 |  |  |     return x;
 | 
      
         | 1195 |  |  |  
 | 
      
         | 1196 |  |  |   /* MODE must occupy no more words than the mode of X.  */
 | 
      
         | 1197 |  |  |   if ((msize + (UNITS_PER_WORD - 1)) / UNITS_PER_WORD
 | 
      
         | 1198 |  |  |       > ((xsize + (UNITS_PER_WORD - 1)) / UNITS_PER_WORD))
 | 
      
         | 1199 |  |  |     return 0;
 | 
      
         | 1200 |  |  |  
 | 
      
         | 1201 |  |  |   /* Don't allow generating paradoxical FLOAT_MODE subregs.  */
 | 
      
         | 1202 |  |  |   if (SCALAR_FLOAT_MODE_P (mode) && msize > xsize)
 | 
      
         | 1203 |  |  |     return 0;
 | 
      
         | 1204 |  |  |  
 | 
      
         | 1205 |  |  |   offset = subreg_lowpart_offset (mode, innermode);
 | 
      
         | 1206 |  |  |  
 | 
      
         | 1207 |  |  |   if ((GET_CODE (x) == ZERO_EXTEND || GET_CODE (x) == SIGN_EXTEND)
 | 
      
         | 1208 |  |  |       && (GET_MODE_CLASS (mode) == MODE_INT
 | 
      
         | 1209 |  |  |           || GET_MODE_CLASS (mode) == MODE_PARTIAL_INT))
 | 
      
         | 1210 |  |  |     {
 | 
      
         | 1211 |  |  |       /* If we are getting the low-order part of something that has been
 | 
      
         | 1212 |  |  |          sign- or zero-extended, we can either just use the object being
 | 
      
         | 1213 |  |  |          extended or make a narrower extension.  If we want an even smaller
 | 
      
         | 1214 |  |  |          piece than the size of the object being extended, call ourselves
 | 
      
         | 1215 |  |  |          recursively.
 | 
      
         | 1216 |  |  |  
 | 
      
         | 1217 |  |  |          This case is used mostly by combine and cse.  */
 | 
      
         | 1218 |  |  |  
 | 
      
         | 1219 |  |  |       if (GET_MODE (XEXP (x, 0)) == mode)
 | 
      
         | 1220 |  |  |         return XEXP (x, 0);
 | 
      
         | 1221 |  |  |       else if (msize < GET_MODE_SIZE (GET_MODE (XEXP (x, 0))))
 | 
      
         | 1222 |  |  |         return gen_lowpart_common (mode, XEXP (x, 0));
 | 
      
         | 1223 |  |  |       else if (msize < xsize)
 | 
      
         | 1224 |  |  |         return gen_rtx_fmt_e (GET_CODE (x), mode, XEXP (x, 0));
 | 
      
         | 1225 |  |  |     }
 | 
      
         | 1226 |  |  |   else if (GET_CODE (x) == SUBREG || REG_P (x)
 | 
      
         | 1227 |  |  |            || GET_CODE (x) == CONCAT || GET_CODE (x) == CONST_VECTOR
 | 
      
         | 1228 |  |  |            || GET_CODE (x) == CONST_DOUBLE || CONST_INT_P (x))
 | 
      
         | 1229 |  |  |     return simplify_gen_subreg (mode, x, innermode, offset);
 | 
      
         | 1230 |  |  |  
 | 
      
         | 1231 |  |  |   /* Otherwise, we can't do this.  */
 | 
      
         | 1232 |  |  |   return 0;
 | 
      
         | 1233 |  |  | }
 | 
      
         | 1234 |  |  |  
 | 
      
         | 1235 |  |  | rtx
 | 
      
         | 1236 |  |  | gen_highpart (enum machine_mode mode, rtx x)
 | 
      
         | 1237 |  |  | {
 | 
      
         | 1238 |  |  |   unsigned int msize = GET_MODE_SIZE (mode);
 | 
      
         | 1239 |  |  |   rtx result;
 | 
      
         | 1240 |  |  |  
 | 
      
         | 1241 |  |  |   /* This case loses if X is a subreg.  To catch bugs early,
 | 
      
         | 1242 |  |  |      complain if an invalid MODE is used even in other cases.  */
 | 
      
         | 1243 |  |  |   gcc_assert (msize <= UNITS_PER_WORD
 | 
      
         | 1244 |  |  |               || msize == (unsigned int) GET_MODE_UNIT_SIZE (GET_MODE (x)));
 | 
      
         | 1245 |  |  |  
 | 
      
         | 1246 |  |  |   result = simplify_gen_subreg (mode, x, GET_MODE (x),
 | 
      
         | 1247 |  |  |                                 subreg_highpart_offset (mode, GET_MODE (x)));
 | 
      
         | 1248 |  |  |   gcc_assert (result);
 | 
      
         | 1249 |  |  |  
 | 
      
         | 1250 |  |  |   /* simplify_gen_subreg is not guaranteed to return a valid operand for
 | 
      
         | 1251 |  |  |      the target if we have a MEM.  gen_highpart must return a valid operand,
 | 
      
         | 1252 |  |  |      emitting code if necessary to do so.  */
 | 
      
         | 1253 |  |  |   if (MEM_P (result))
 | 
      
         | 1254 |  |  |     {
 | 
      
         | 1255 |  |  |       result = validize_mem (result);
 | 
      
         | 1256 |  |  |       gcc_assert (result);
 | 
      
         | 1257 |  |  |     }
 | 
      
         | 1258 |  |  |  
 | 
      
         | 1259 |  |  |   return result;
 | 
      
         | 1260 |  |  | }
 | 
      
         | 1261 |  |  |  
 | 
      
         | 1262 |  |  | /* Like gen_highpart, but accept mode of EXP operand in case EXP can
 | 
      
         | 1263 |  |  |    be VOIDmode constant.  */
 | 
      
         | 1264 |  |  | rtx
 | 
      
         | 1265 |  |  | gen_highpart_mode (enum machine_mode outermode, enum machine_mode innermode, rtx exp)
 | 
      
         | 1266 |  |  | {
 | 
      
         | 1267 |  |  |   if (GET_MODE (exp) != VOIDmode)
 | 
      
         | 1268 |  |  |     {
 | 
      
         | 1269 |  |  |       gcc_assert (GET_MODE (exp) == innermode);
 | 
      
         | 1270 |  |  |       return gen_highpart (outermode, exp);
 | 
      
         | 1271 |  |  |     }
 | 
      
         | 1272 |  |  |   return simplify_gen_subreg (outermode, exp, innermode,
 | 
      
         | 1273 |  |  |                               subreg_highpart_offset (outermode, innermode));
 | 
      
         | 1274 |  |  | }
 | 
      
         | 1275 |  |  |  
 | 
      
         | 1276 |  |  | /* Return the SUBREG_BYTE for an OUTERMODE lowpart of an INNERMODE value.  */
 | 
      
         | 1277 |  |  |  
 | 
      
         | 1278 |  |  | unsigned int
 | 
      
         | 1279 |  |  | subreg_lowpart_offset (enum machine_mode outermode, enum machine_mode innermode)
 | 
      
         | 1280 |  |  | {
 | 
      
         | 1281 |  |  |   unsigned int offset = 0;
 | 
      
         | 1282 |  |  |   int difference = (GET_MODE_SIZE (innermode) - GET_MODE_SIZE (outermode));
 | 
      
         | 1283 |  |  |  
 | 
      
         | 1284 |  |  |   if (difference > 0)
 | 
      
         | 1285 |  |  |     {
 | 
      
         | 1286 |  |  |       if (WORDS_BIG_ENDIAN)
 | 
      
         | 1287 |  |  |         offset += (difference / UNITS_PER_WORD) * UNITS_PER_WORD;
 | 
      
         | 1288 |  |  |       if (BYTES_BIG_ENDIAN)
 | 
      
         | 1289 |  |  |         offset += difference % UNITS_PER_WORD;
 | 
      
         | 1290 |  |  |     }
 | 
      
         | 1291 |  |  |  
 | 
      
         | 1292 |  |  |   return offset;
 | 
      
         | 1293 |  |  | }
 | 
      
         | 1294 |  |  |  
 | 
      
         | 1295 |  |  | /* Return offset in bytes to get OUTERMODE high part
 | 
      
         | 1296 |  |  |    of the value in mode INNERMODE stored in memory in target format.  */
 | 
      
         | 1297 |  |  | unsigned int
 | 
      
         | 1298 |  |  | subreg_highpart_offset (enum machine_mode outermode, enum machine_mode innermode)
 | 
      
         | 1299 |  |  | {
 | 
      
         | 1300 |  |  |   unsigned int offset = 0;
 | 
      
         | 1301 |  |  |   int difference = (GET_MODE_SIZE (innermode) - GET_MODE_SIZE (outermode));
 | 
      
         | 1302 |  |  |  
 | 
      
         | 1303 |  |  |   gcc_assert (GET_MODE_SIZE (innermode) >= GET_MODE_SIZE (outermode));
 | 
      
         | 1304 |  |  |  
 | 
      
         | 1305 |  |  |   if (difference > 0)
 | 
      
         | 1306 |  |  |     {
 | 
      
         | 1307 |  |  |       if (! WORDS_BIG_ENDIAN)
 | 
      
         | 1308 |  |  |         offset += (difference / UNITS_PER_WORD) * UNITS_PER_WORD;
 | 
      
         | 1309 |  |  |       if (! BYTES_BIG_ENDIAN)
 | 
      
         | 1310 |  |  |         offset += difference % UNITS_PER_WORD;
 | 
      
         | 1311 |  |  |     }
 | 
      
         | 1312 |  |  |  
 | 
      
         | 1313 |  |  |   return offset;
 | 
      
         | 1314 |  |  | }
 | 
      
         | 1315 |  |  |  
 | 
      
         | 1316 |  |  | /* Return 1 iff X, assumed to be a SUBREG,
 | 
      
         | 1317 |  |  |    refers to the least significant part of its containing reg.
 | 
      
         | 1318 |  |  |    If X is not a SUBREG, always return 1 (it is its own low part!).  */
 | 
      
         | 1319 |  |  |  
 | 
      
         | 1320 |  |  | int
 | 
      
         | 1321 |  |  | subreg_lowpart_p (const_rtx x)
 | 
      
         | 1322 |  |  | {
 | 
      
         | 1323 |  |  |   if (GET_CODE (x) != SUBREG)
 | 
      
         | 1324 |  |  |     return 1;
 | 
      
         | 1325 |  |  |   else if (GET_MODE (SUBREG_REG (x)) == VOIDmode)
 | 
      
         | 1326 |  |  |     return 0;
 | 
      
         | 1327 |  |  |  
 | 
      
         | 1328 |  |  |   return (subreg_lowpart_offset (GET_MODE (x), GET_MODE (SUBREG_REG (x)))
 | 
      
         | 1329 |  |  |           == SUBREG_BYTE (x));
 | 
      
         | 1330 |  |  | }
 | 
      
         | 1331 |  |  |  
 | 
      
         | 1332 |  |  | /* Return true if X is a paradoxical subreg, false otherwise.  */
 | 
      
         | 1333 |  |  | bool
 | 
      
         | 1334 |  |  | paradoxical_subreg_p (const_rtx x)
 | 
      
         | 1335 |  |  | {
 | 
      
         | 1336 |  |  |   if (GET_CODE (x) != SUBREG)
 | 
      
         | 1337 |  |  |     return false;
 | 
      
         | 1338 |  |  |   return (GET_MODE_PRECISION (GET_MODE (x))
 | 
      
         | 1339 |  |  |           > GET_MODE_PRECISION (GET_MODE (SUBREG_REG (x))));
 | 
      
         | 1340 |  |  | }
 | 
      
         | 1341 |  |  |  
 | 
      
         | 1342 |  |  | /* Return subword OFFSET of operand OP.
 | 
      
         | 1343 |  |  |    The word number, OFFSET, is interpreted as the word number starting
 | 
      
         | 1344 |  |  |    at the low-order address.  OFFSET 0 is the low-order word if not
 | 
      
         | 1345 |  |  |    WORDS_BIG_ENDIAN, otherwise it is the high-order word.
 | 
      
         | 1346 |  |  |  
 | 
      
         | 1347 |  |  |    If we cannot extract the required word, we return zero.  Otherwise,
 | 
      
         | 1348 |  |  |    an rtx corresponding to the requested word will be returned.
 | 
      
         | 1349 |  |  |  
 | 
      
         | 1350 |  |  |    VALIDATE_ADDRESS is nonzero if the address should be validated.  Before
 | 
      
         | 1351 |  |  |    reload has completed, a valid address will always be returned.  After
 | 
      
         | 1352 |  |  |    reload, if a valid address cannot be returned, we return zero.
 | 
      
         | 1353 |  |  |  
 | 
      
         | 1354 |  |  |    If VALIDATE_ADDRESS is zero, we simply form the required address; validating
 | 
      
         | 1355 |  |  |    it is the responsibility of the caller.
 | 
      
         | 1356 |  |  |  
 | 
      
         | 1357 |  |  |    MODE is the mode of OP in case it is a CONST_INT.
 | 
      
         | 1358 |  |  |  
 | 
      
         | 1359 |  |  |    ??? This is still rather broken for some cases.  The problem for the
 | 
      
         | 1360 |  |  |    moment is that all callers of this thing provide no 'goal mode' to
 | 
      
         | 1361 |  |  |    tell us to work with.  This exists because all callers were written
 | 
      
         | 1362 |  |  |    in a word based SUBREG world.
 | 
      
         | 1363 |  |  |    Now use of this function can be deprecated by simplify_subreg in most
 | 
      
         | 1364 |  |  |    cases.
 | 
      
         | 1365 |  |  |  */
 | 
      
         | 1366 |  |  |  
 | 
      
         | 1367 |  |  | rtx
 | 
      
         | 1368 |  |  | operand_subword (rtx op, unsigned int offset, int validate_address, enum machine_mode mode)
 | 
      
         | 1369 |  |  | {
 | 
      
         | 1370 |  |  |   if (mode == VOIDmode)
 | 
      
         | 1371 |  |  |     mode = GET_MODE (op);
 | 
      
         | 1372 |  |  |  
 | 
      
         | 1373 |  |  |   gcc_assert (mode != VOIDmode);
 | 
      
         | 1374 |  |  |  
 | 
      
         | 1375 |  |  |   /* If OP is narrower than a word, fail.  */
 | 
      
         | 1376 |  |  |   if (mode != BLKmode
 | 
      
         | 1377 |  |  |       && (GET_MODE_SIZE (mode) < UNITS_PER_WORD))
 | 
      
         | 1378 |  |  |     return 0;
 | 
      
         | 1379 |  |  |  
 | 
      
         | 1380 |  |  |   /* If we want a word outside OP, return zero.  */
 | 
      
         | 1381 |  |  |   if (mode != BLKmode
 | 
      
         | 1382 |  |  |       && (offset + 1) * UNITS_PER_WORD > GET_MODE_SIZE (mode))
 | 
      
         | 1383 |  |  |     return const0_rtx;
 | 
      
         | 1384 |  |  |  
 | 
      
         | 1385 |  |  |   /* Form a new MEM at the requested address.  */
 | 
      
         | 1386 |  |  |   if (MEM_P (op))
 | 
      
         | 1387 |  |  |     {
 | 
      
         | 1388 |  |  |       rtx new_rtx = adjust_address_nv (op, word_mode, offset * UNITS_PER_WORD);
 | 
      
         | 1389 |  |  |  
 | 
      
         | 1390 |  |  |       if (! validate_address)
 | 
      
         | 1391 |  |  |         return new_rtx;
 | 
      
         | 1392 |  |  |  
 | 
      
         | 1393 |  |  |       else if (reload_completed)
 | 
      
         | 1394 |  |  |         {
 | 
      
         | 1395 |  |  |           if (! strict_memory_address_addr_space_p (word_mode,
 | 
      
         | 1396 |  |  |                                                     XEXP (new_rtx, 0),
 | 
      
         | 1397 |  |  |                                                     MEM_ADDR_SPACE (op)))
 | 
      
         | 1398 |  |  |             return 0;
 | 
      
         | 1399 |  |  |         }
 | 
      
         | 1400 |  |  |       else
 | 
      
         | 1401 |  |  |         return replace_equiv_address (new_rtx, XEXP (new_rtx, 0));
 | 
      
         | 1402 |  |  |     }
 | 
      
         | 1403 |  |  |  
 | 
      
         | 1404 |  |  |   /* Rest can be handled by simplify_subreg.  */
 | 
      
         | 1405 |  |  |   return simplify_gen_subreg (word_mode, op, mode, (offset * UNITS_PER_WORD));
 | 
      
         | 1406 |  |  | }
 | 
      
         | 1407 |  |  |  
 | 
      
         | 1408 |  |  | /* Similar to `operand_subword', but never return 0.  If we can't
 | 
      
         | 1409 |  |  |    extract the required subword, put OP into a register and try again.
 | 
      
         | 1410 |  |  |    The second attempt must succeed.  We always validate the address in
 | 
      
         | 1411 |  |  |    this case.
 | 
      
         | 1412 |  |  |  
 | 
      
         | 1413 |  |  |    MODE is the mode of OP, in case it is CONST_INT.  */
 | 
      
         | 1414 |  |  |  
 | 
      
         | 1415 |  |  | rtx
 | 
      
         | 1416 |  |  | operand_subword_force (rtx op, unsigned int offset, enum machine_mode mode)
 | 
      
         | 1417 |  |  | {
 | 
      
         | 1418 |  |  |   rtx result = operand_subword (op, offset, 1, mode);
 | 
      
         | 1419 |  |  |  
 | 
      
         | 1420 |  |  |   if (result)
 | 
      
         | 1421 |  |  |     return result;
 | 
      
         | 1422 |  |  |  
 | 
      
         | 1423 |  |  |   if (mode != BLKmode && mode != VOIDmode)
 | 
      
         | 1424 |  |  |     {
 | 
      
         | 1425 |  |  |       /* If this is a register which can not be accessed by words, copy it
 | 
      
         | 1426 |  |  |          to a pseudo register.  */
 | 
      
         | 1427 |  |  |       if (REG_P (op))
 | 
      
         | 1428 |  |  |         op = copy_to_reg (op);
 | 
      
         | 1429 |  |  |       else
 | 
      
         | 1430 |  |  |         op = force_reg (mode, op);
 | 
      
         | 1431 |  |  |     }
 | 
      
         | 1432 |  |  |  
 | 
      
         | 1433 |  |  |   result = operand_subword (op, offset, 1, mode);
 | 
      
         | 1434 |  |  |   gcc_assert (result);
 | 
      
         | 1435 |  |  |  
 | 
      
         | 1436 |  |  |   return result;
 | 
      
         | 1437 |  |  | }
 | 
      
         | 1438 |  |  |  
 | 
      
         | 1439 |  |  | /* Returns 1 if both MEM_EXPR can be considered equal
 | 
      
         | 1440 |  |  |    and 0 otherwise.  */
 | 
      
         | 1441 |  |  |  
 | 
      
         | 1442 |  |  | int
 | 
      
         | 1443 |  |  | mem_expr_equal_p (const_tree expr1, const_tree expr2)
 | 
      
         | 1444 |  |  | {
 | 
      
         | 1445 |  |  |   if (expr1 == expr2)
 | 
      
         | 1446 |  |  |     return 1;
 | 
      
         | 1447 |  |  |  
 | 
      
         | 1448 |  |  |   if (! expr1 || ! expr2)
 | 
      
         | 1449 |  |  |     return 0;
 | 
      
         | 1450 |  |  |  
 | 
      
         | 1451 |  |  |   if (TREE_CODE (expr1) != TREE_CODE (expr2))
 | 
      
         | 1452 |  |  |     return 0;
 | 
      
         | 1453 |  |  |  
 | 
      
         | 1454 |  |  |   return operand_equal_p (expr1, expr2, 0);
 | 
      
         | 1455 |  |  | }
 | 
      
         | 1456 |  |  |  
 | 
      
         | 1457 |  |  | /* Return OFFSET if XEXP (MEM, 0) - OFFSET is known to be ALIGN
 | 
      
         | 1458 |  |  |    bits aligned for 0 <= OFFSET < ALIGN / BITS_PER_UNIT, or
 | 
      
         | 1459 |  |  |    -1 if not known.  */
 | 
      
         | 1460 |  |  |  
 | 
      
         | 1461 |  |  | int
 | 
      
         | 1462 |  |  | get_mem_align_offset (rtx mem, unsigned int align)
 | 
      
         | 1463 |  |  | {
 | 
      
         | 1464 |  |  |   tree expr;
 | 
      
         | 1465 |  |  |   unsigned HOST_WIDE_INT offset;
 | 
      
         | 1466 |  |  |  
 | 
      
         | 1467 |  |  |   /* This function can't use
 | 
      
         | 1468 |  |  |      if (!MEM_EXPR (mem) || !MEM_OFFSET_KNOWN_P (mem)
 | 
      
         | 1469 |  |  |          || (MAX (MEM_ALIGN (mem),
 | 
      
         | 1470 |  |  |                   MAX (align, get_object_alignment (MEM_EXPR (mem))))
 | 
      
         | 1471 |  |  |              < align))
 | 
      
         | 1472 |  |  |        return -1;
 | 
      
         | 1473 |  |  |      else
 | 
      
         | 1474 |  |  |        return (- MEM_OFFSET (mem)) & (align / BITS_PER_UNIT - 1);
 | 
      
         | 1475 |  |  |      for two reasons:
 | 
      
         | 1476 |  |  |      - COMPONENT_REFs in MEM_EXPR can have NULL first operand,
 | 
      
         | 1477 |  |  |        for <variable>.  get_inner_reference doesn't handle it and
 | 
      
         | 1478 |  |  |        even if it did, the alignment in that case needs to be determined
 | 
      
         | 1479 |  |  |        from DECL_FIELD_CONTEXT's TYPE_ALIGN.
 | 
      
         | 1480 |  |  |      - it would do suboptimal job for COMPONENT_REFs, even if MEM_EXPR
 | 
      
         | 1481 |  |  |        isn't sufficiently aligned, the object it is in might be.  */
 | 
      
         | 1482 |  |  |   gcc_assert (MEM_P (mem));
 | 
      
         | 1483 |  |  |   expr = MEM_EXPR (mem);
 | 
      
         | 1484 |  |  |   if (expr == NULL_TREE || !MEM_OFFSET_KNOWN_P (mem))
 | 
      
         | 1485 |  |  |     return -1;
 | 
      
         | 1486 |  |  |  
 | 
      
         | 1487 |  |  |   offset = MEM_OFFSET (mem);
 | 
      
         | 1488 |  |  |   if (DECL_P (expr))
 | 
      
         | 1489 |  |  |     {
 | 
      
         | 1490 |  |  |       if (DECL_ALIGN (expr) < align)
 | 
      
         | 1491 |  |  |         return -1;
 | 
      
         | 1492 |  |  |     }
 | 
      
         | 1493 |  |  |   else if (INDIRECT_REF_P (expr))
 | 
      
         | 1494 |  |  |     {
 | 
      
         | 1495 |  |  |       if (TYPE_ALIGN (TREE_TYPE (expr)) < (unsigned int) align)
 | 
      
         | 1496 |  |  |         return -1;
 | 
      
         | 1497 |  |  |     }
 | 
      
         | 1498 |  |  |   else if (TREE_CODE (expr) == COMPONENT_REF)
 | 
      
         | 1499 |  |  |     {
 | 
      
         | 1500 |  |  |       while (1)
 | 
      
         | 1501 |  |  |         {
 | 
      
         | 1502 |  |  |           tree inner = TREE_OPERAND (expr, 0);
 | 
      
         | 1503 |  |  |           tree field = TREE_OPERAND (expr, 1);
 | 
      
         | 1504 |  |  |           tree byte_offset = component_ref_field_offset (expr);
 | 
      
         | 1505 |  |  |           tree bit_offset = DECL_FIELD_BIT_OFFSET (field);
 | 
      
         | 1506 |  |  |  
 | 
      
         | 1507 |  |  |           if (!byte_offset
 | 
      
         | 1508 |  |  |               || !host_integerp (byte_offset, 1)
 | 
      
         | 1509 |  |  |               || !host_integerp (bit_offset, 1))
 | 
      
         | 1510 |  |  |             return -1;
 | 
      
         | 1511 |  |  |  
 | 
      
         | 1512 |  |  |           offset += tree_low_cst (byte_offset, 1);
 | 
      
         | 1513 |  |  |           offset += tree_low_cst (bit_offset, 1) / BITS_PER_UNIT;
 | 
      
         | 1514 |  |  |  
 | 
      
         | 1515 |  |  |           if (inner == NULL_TREE)
 | 
      
         | 1516 |  |  |             {
 | 
      
         | 1517 |  |  |               if (TYPE_ALIGN (DECL_FIELD_CONTEXT (field))
 | 
      
         | 1518 |  |  |                   < (unsigned int) align)
 | 
      
         | 1519 |  |  |                 return -1;
 | 
      
         | 1520 |  |  |               break;
 | 
      
         | 1521 |  |  |             }
 | 
      
         | 1522 |  |  |           else if (DECL_P (inner))
 | 
      
         | 1523 |  |  |             {
 | 
      
         | 1524 |  |  |               if (DECL_ALIGN (inner) < align)
 | 
      
         | 1525 |  |  |                 return -1;
 | 
      
         | 1526 |  |  |               break;
 | 
      
         | 1527 |  |  |             }
 | 
      
         | 1528 |  |  |           else if (TREE_CODE (inner) != COMPONENT_REF)
 | 
      
         | 1529 |  |  |             return -1;
 | 
      
         | 1530 |  |  |           expr = inner;
 | 
      
         | 1531 |  |  |         }
 | 
      
         | 1532 |  |  |     }
 | 
      
         | 1533 |  |  |   else
 | 
      
         | 1534 |  |  |     return -1;
 | 
      
         | 1535 |  |  |  
 | 
      
         | 1536 |  |  |   return offset & ((align / BITS_PER_UNIT) - 1);
 | 
      
         | 1537 |  |  | }
 | 
      
         | 1538 |  |  |  
 | 
      
         | 1539 |  |  | /* Given REF (a MEM) and T, either the type of X or the expression
 | 
      
         | 1540 |  |  |    corresponding to REF, set the memory attributes.  OBJECTP is nonzero
 | 
      
         | 1541 |  |  |    if we are making a new object of this type.  BITPOS is nonzero if
 | 
      
         | 1542 |  |  |    there is an offset outstanding on T that will be applied later.  */
 | 
      
         | 1543 |  |  |  
 | 
      
         | 1544 |  |  | void
 | 
      
         | 1545 |  |  | set_mem_attributes_minus_bitpos (rtx ref, tree t, int objectp,
 | 
      
         | 1546 |  |  |                                  HOST_WIDE_INT bitpos)
 | 
      
         | 1547 |  |  | {
 | 
      
         | 1548 |  |  |   HOST_WIDE_INT apply_bitpos = 0;
 | 
      
         | 1549 |  |  |   tree type;
 | 
      
         | 1550 |  |  |   struct mem_attrs attrs, *defattrs, *refattrs;
 | 
      
         | 1551 |  |  |   addr_space_t as;
 | 
      
         | 1552 |  |  |  
 | 
      
         | 1553 |  |  |   /* It can happen that type_for_mode was given a mode for which there
 | 
      
         | 1554 |  |  |      is no language-level type.  In which case it returns NULL, which
 | 
      
         | 1555 |  |  |      we can see here.  */
 | 
      
         | 1556 |  |  |   if (t == NULL_TREE)
 | 
      
         | 1557 |  |  |     return;
 | 
      
         | 1558 |  |  |  
 | 
      
         | 1559 |  |  |   type = TYPE_P (t) ? t : TREE_TYPE (t);
 | 
      
         | 1560 |  |  |   if (type == error_mark_node)
 | 
      
         | 1561 |  |  |     return;
 | 
      
         | 1562 |  |  |  
 | 
      
         | 1563 |  |  |   /* If we have already set DECL_RTL = ref, get_alias_set will get the
 | 
      
         | 1564 |  |  |      wrong answer, as it assumes that DECL_RTL already has the right alias
 | 
      
         | 1565 |  |  |      info.  Callers should not set DECL_RTL until after the call to
 | 
      
         | 1566 |  |  |      set_mem_attributes.  */
 | 
      
         | 1567 |  |  |   gcc_assert (!DECL_P (t) || ref != DECL_RTL_IF_SET (t));
 | 
      
         | 1568 |  |  |  
 | 
      
         | 1569 |  |  |   memset (&attrs, 0, sizeof (attrs));
 | 
      
         | 1570 |  |  |  
 | 
      
         | 1571 |  |  |   /* Get the alias set from the expression or type (perhaps using a
 | 
      
         | 1572 |  |  |      front-end routine) and use it.  */
 | 
      
         | 1573 |  |  |   attrs.alias = get_alias_set (t);
 | 
      
         | 1574 |  |  |  
 | 
      
         | 1575 |  |  |   MEM_VOLATILE_P (ref) |= TYPE_VOLATILE (type);
 | 
      
         | 1576 |  |  |   MEM_POINTER (ref) = POINTER_TYPE_P (type);
 | 
      
         | 1577 |  |  |  
 | 
      
         | 1578 |  |  |   /* Default values from pre-existing memory attributes if present.  */
 | 
      
         | 1579 |  |  |   refattrs = MEM_ATTRS (ref);
 | 
      
         | 1580 |  |  |   if (refattrs)
 | 
      
         | 1581 |  |  |     {
 | 
      
         | 1582 |  |  |       /* ??? Can this ever happen?  Calling this routine on a MEM that
 | 
      
         | 1583 |  |  |          already carries memory attributes should probably be invalid.  */
 | 
      
         | 1584 |  |  |       attrs.expr = refattrs->expr;
 | 
      
         | 1585 |  |  |       attrs.offset_known_p = refattrs->offset_known_p;
 | 
      
         | 1586 |  |  |       attrs.offset = refattrs->offset;
 | 
      
         | 1587 |  |  |       attrs.size_known_p = refattrs->size_known_p;
 | 
      
         | 1588 |  |  |       attrs.size = refattrs->size;
 | 
      
         | 1589 |  |  |       attrs.align = refattrs->align;
 | 
      
         | 1590 |  |  |     }
 | 
      
         | 1591 |  |  |  
 | 
      
         | 1592 |  |  |   /* Otherwise, default values from the mode of the MEM reference.  */
 | 
      
         | 1593 |  |  |   else
 | 
      
         | 1594 |  |  |     {
 | 
      
         | 1595 |  |  |       defattrs = mode_mem_attrs[(int) GET_MODE (ref)];
 | 
      
         | 1596 |  |  |       gcc_assert (!defattrs->expr);
 | 
      
         | 1597 |  |  |       gcc_assert (!defattrs->offset_known_p);
 | 
      
         | 1598 |  |  |  
 | 
      
         | 1599 |  |  |       /* Respect mode size.  */
 | 
      
         | 1600 |  |  |       attrs.size_known_p = defattrs->size_known_p;
 | 
      
         | 1601 |  |  |       attrs.size = defattrs->size;
 | 
      
         | 1602 |  |  |       /* ??? Is this really necessary?  We probably should always get
 | 
      
         | 1603 |  |  |          the size from the type below.  */
 | 
      
         | 1604 |  |  |  
 | 
      
         | 1605 |  |  |       /* Respect mode alignment for STRICT_ALIGNMENT targets if T is a type;
 | 
      
         | 1606 |  |  |          if T is an object, always compute the object alignment below.  */
 | 
      
         | 1607 |  |  |       if (TYPE_P (t))
 | 
      
         | 1608 |  |  |         attrs.align = defattrs->align;
 | 
      
         | 1609 |  |  |       else
 | 
      
         | 1610 |  |  |         attrs.align = BITS_PER_UNIT;
 | 
      
         | 1611 |  |  |       /* ??? If T is a type, respecting mode alignment may *also* be wrong
 | 
      
         | 1612 |  |  |          e.g. if the type carries an alignment attribute.  Should we be
 | 
      
         | 1613 |  |  |          able to simply always use TYPE_ALIGN?  */
 | 
      
         | 1614 |  |  |     }
 | 
      
         | 1615 |  |  |  
 | 
      
         | 1616 |  |  |   /* We can set the alignment from the type if we are making an object,
 | 
      
         | 1617 |  |  |      this is an INDIRECT_REF, or if TYPE_ALIGN_OK.  */
 | 
      
         | 1618 |  |  |   if (objectp || TREE_CODE (t) == INDIRECT_REF || TYPE_ALIGN_OK (type))
 | 
      
         | 1619 |  |  |     attrs.align = MAX (attrs.align, TYPE_ALIGN (type));
 | 
      
         | 1620 |  |  |  
 | 
      
         | 1621 |  |  |   else if (TREE_CODE (t) == MEM_REF)
 | 
      
         | 1622 |  |  |     {
 | 
      
         | 1623 |  |  |       tree op0 = TREE_OPERAND (t, 0);
 | 
      
         | 1624 |  |  |       if (TREE_CODE (op0) == ADDR_EXPR
 | 
      
         | 1625 |  |  |           && (DECL_P (TREE_OPERAND (op0, 0))
 | 
      
         | 1626 |  |  |               || CONSTANT_CLASS_P (TREE_OPERAND (op0, 0))))
 | 
      
         | 1627 |  |  |         {
 | 
      
         | 1628 |  |  |           if (DECL_P (TREE_OPERAND (op0, 0)))
 | 
      
         | 1629 |  |  |             attrs.align = DECL_ALIGN (TREE_OPERAND (op0, 0));
 | 
      
         | 1630 |  |  |           else if (CONSTANT_CLASS_P (TREE_OPERAND (op0, 0)))
 | 
      
         | 1631 |  |  |             {
 | 
      
         | 1632 |  |  |               attrs.align = TYPE_ALIGN (TREE_TYPE (TREE_OPERAND (op0, 0)));
 | 
      
         | 1633 |  |  | #ifdef CONSTANT_ALIGNMENT
 | 
      
         | 1634 |  |  |               attrs.align = CONSTANT_ALIGNMENT (TREE_OPERAND (op0, 0),
 | 
      
         | 1635 |  |  |                                                 attrs.align);
 | 
      
         | 1636 |  |  | #endif
 | 
      
         | 1637 |  |  |             }
 | 
      
         | 1638 |  |  |           if (TREE_INT_CST_LOW (TREE_OPERAND (t, 1)) != 0)
 | 
      
         | 1639 |  |  |             {
 | 
      
         | 1640 |  |  |               unsigned HOST_WIDE_INT ioff
 | 
      
         | 1641 |  |  |                 = TREE_INT_CST_LOW (TREE_OPERAND (t, 1));
 | 
      
         | 1642 |  |  |               unsigned HOST_WIDE_INT aoff = (ioff & -ioff) * BITS_PER_UNIT;
 | 
      
         | 1643 |  |  |               attrs.align = MIN (aoff, attrs.align);
 | 
      
         | 1644 |  |  |             }
 | 
      
         | 1645 |  |  |         }
 | 
      
         | 1646 |  |  |       else
 | 
      
         | 1647 |  |  |         /* ??? This isn't fully correct, we can't set the alignment from the
 | 
      
         | 1648 |  |  |            type in all cases.  */
 | 
      
         | 1649 |  |  |         attrs.align = MAX (attrs.align, TYPE_ALIGN (type));
 | 
      
         | 1650 |  |  |     }
 | 
      
         | 1651 |  |  |  
 | 
      
         | 1652 |  |  |   else if (TREE_CODE (t) == TARGET_MEM_REF)
 | 
      
         | 1653 |  |  |     /* ??? This isn't fully correct, we can't set the alignment from the
 | 
      
         | 1654 |  |  |        type in all cases.  */
 | 
      
         | 1655 |  |  |     attrs.align = MAX (attrs.align, TYPE_ALIGN (type));
 | 
      
         | 1656 |  |  |  
 | 
      
         | 1657 |  |  |   /* If the size is known, we can set that.  */
 | 
      
         | 1658 |  |  |   if (TYPE_SIZE_UNIT (type) && host_integerp (TYPE_SIZE_UNIT (type), 1))
 | 
      
         | 1659 |  |  |     {
 | 
      
         | 1660 |  |  |       attrs.size_known_p = true;
 | 
      
         | 1661 |  |  |       attrs.size = tree_low_cst (TYPE_SIZE_UNIT (type), 1);
 | 
      
         | 1662 |  |  |     }
 | 
      
         | 1663 |  |  |  
 | 
      
         | 1664 |  |  |   /* If T is not a type, we may be able to deduce some more information about
 | 
      
         | 1665 |  |  |      the expression.  */
 | 
      
         | 1666 |  |  |   if (! TYPE_P (t))
 | 
      
         | 1667 |  |  |     {
 | 
      
         | 1668 |  |  |       tree base;
 | 
      
         | 1669 |  |  |       bool align_computed = false;
 | 
      
         | 1670 |  |  |  
 | 
      
         | 1671 |  |  |       if (TREE_THIS_VOLATILE (t))
 | 
      
         | 1672 |  |  |         MEM_VOLATILE_P (ref) = 1;
 | 
      
         | 1673 |  |  |  
 | 
      
         | 1674 |  |  |       /* Now remove any conversions: they don't change what the underlying
 | 
      
         | 1675 |  |  |          object is.  Likewise for SAVE_EXPR.  */
 | 
      
         | 1676 |  |  |       while (CONVERT_EXPR_P (t)
 | 
      
         | 1677 |  |  |              || TREE_CODE (t) == VIEW_CONVERT_EXPR
 | 
      
         | 1678 |  |  |              || TREE_CODE (t) == SAVE_EXPR)
 | 
      
         | 1679 |  |  |         t = TREE_OPERAND (t, 0);
 | 
      
         | 1680 |  |  |  
 | 
      
         | 1681 |  |  |       /* Note whether this expression can trap.  */
 | 
      
         | 1682 |  |  |       MEM_NOTRAP_P (ref) = !tree_could_trap_p (t);
 | 
      
         | 1683 |  |  |  
 | 
      
         | 1684 |  |  |       base = get_base_address (t);
 | 
      
         | 1685 |  |  |       if (base)
 | 
      
         | 1686 |  |  |         {
 | 
      
         | 1687 |  |  |           if (DECL_P (base)
 | 
      
         | 1688 |  |  |               && TREE_READONLY (base)
 | 
      
         | 1689 |  |  |               && (TREE_STATIC (base) || DECL_EXTERNAL (base))
 | 
      
         | 1690 |  |  |               && !TREE_THIS_VOLATILE (base))
 | 
      
         | 1691 |  |  |             MEM_READONLY_P (ref) = 1;
 | 
      
         | 1692 |  |  |  
 | 
      
         | 1693 |  |  |           /* Mark static const strings readonly as well.  */
 | 
      
         | 1694 |  |  |           if (TREE_CODE (base) == STRING_CST
 | 
      
         | 1695 |  |  |               && TREE_READONLY (base)
 | 
      
         | 1696 |  |  |               && TREE_STATIC (base))
 | 
      
         | 1697 |  |  |             MEM_READONLY_P (ref) = 1;
 | 
      
         | 1698 |  |  |  
 | 
      
         | 1699 |  |  |           if (TREE_CODE (base) == MEM_REF
 | 
      
         | 1700 |  |  |               || TREE_CODE (base) == TARGET_MEM_REF)
 | 
      
         | 1701 |  |  |             as = TYPE_ADDR_SPACE (TREE_TYPE (TREE_TYPE (TREE_OPERAND (base,
 | 
      
         | 1702 |  |  |                                                                       0))));
 | 
      
         | 1703 |  |  |           else
 | 
      
         | 1704 |  |  |             as = TYPE_ADDR_SPACE (TREE_TYPE (base));
 | 
      
         | 1705 |  |  |         }
 | 
      
         | 1706 |  |  |       else
 | 
      
         | 1707 |  |  |         as = TYPE_ADDR_SPACE (type);
 | 
      
         | 1708 |  |  |  
 | 
      
         | 1709 |  |  |       /* If this expression uses it's parent's alias set, mark it such
 | 
      
         | 1710 |  |  |          that we won't change it.  */
 | 
      
         | 1711 |  |  |       if (component_uses_parent_alias_set (t))
 | 
      
         | 1712 |  |  |         MEM_KEEP_ALIAS_SET_P (ref) = 1;
 | 
      
         | 1713 |  |  |  
 | 
      
         | 1714 |  |  |       /* If this is a decl, set the attributes of the MEM from it.  */
 | 
      
         | 1715 |  |  |       if (DECL_P (t))
 | 
      
         | 1716 |  |  |         {
 | 
      
         | 1717 |  |  |           attrs.expr = t;
 | 
      
         | 1718 |  |  |           attrs.offset_known_p = true;
 | 
      
         | 1719 |  |  |           attrs.offset = 0;
 | 
      
         | 1720 |  |  |           apply_bitpos = bitpos;
 | 
      
         | 1721 |  |  |           if (DECL_SIZE_UNIT (t) && host_integerp (DECL_SIZE_UNIT (t), 1))
 | 
      
         | 1722 |  |  |             {
 | 
      
         | 1723 |  |  |               attrs.size_known_p = true;
 | 
      
         | 1724 |  |  |               attrs.size = tree_low_cst (DECL_SIZE_UNIT (t), 1);
 | 
      
         | 1725 |  |  |             }
 | 
      
         | 1726 |  |  |           else
 | 
      
         | 1727 |  |  |             attrs.size_known_p = false;
 | 
      
         | 1728 |  |  |           attrs.align = DECL_ALIGN (t);
 | 
      
         | 1729 |  |  |           align_computed = true;
 | 
      
         | 1730 |  |  |         }
 | 
      
         | 1731 |  |  |  
 | 
      
         | 1732 |  |  |       /* If this is a constant, we know the alignment.  */
 | 
      
         | 1733 |  |  |       else if (CONSTANT_CLASS_P (t))
 | 
      
         | 1734 |  |  |         {
 | 
      
         | 1735 |  |  |           attrs.align = TYPE_ALIGN (type);
 | 
      
         | 1736 |  |  | #ifdef CONSTANT_ALIGNMENT
 | 
      
         | 1737 |  |  |           attrs.align = CONSTANT_ALIGNMENT (t, attrs.align);
 | 
      
         | 1738 |  |  | #endif
 | 
      
         | 1739 |  |  |           align_computed = true;
 | 
      
         | 1740 |  |  |         }
 | 
      
         | 1741 |  |  |  
 | 
      
         | 1742 |  |  |       /* If this is a field reference and not a bit-field, record it.  */
 | 
      
         | 1743 |  |  |       /* ??? There is some information that can be gleaned from bit-fields,
 | 
      
         | 1744 |  |  |          such as the word offset in the structure that might be modified.
 | 
      
         | 1745 |  |  |          But skip it for now.  */
 | 
      
         | 1746 |  |  |       else if (TREE_CODE (t) == COMPONENT_REF
 | 
      
         | 1747 |  |  |                && ! DECL_BIT_FIELD (TREE_OPERAND (t, 1)))
 | 
      
         | 1748 |  |  |         {
 | 
      
         | 1749 |  |  |           attrs.expr = t;
 | 
      
         | 1750 |  |  |           attrs.offset_known_p = true;
 | 
      
         | 1751 |  |  |           attrs.offset = 0;
 | 
      
         | 1752 |  |  |           apply_bitpos = bitpos;
 | 
      
         | 1753 |  |  |           /* ??? Any reason the field size would be different than
 | 
      
         | 1754 |  |  |              the size we got from the type?  */
 | 
      
         | 1755 |  |  |         }
 | 
      
         | 1756 |  |  |  
 | 
      
         | 1757 |  |  |       /* If this is an array reference, look for an outer field reference.  */
 | 
      
         | 1758 |  |  |       else if (TREE_CODE (t) == ARRAY_REF)
 | 
      
         | 1759 |  |  |         {
 | 
      
         | 1760 |  |  |           tree off_tree = size_zero_node;
 | 
      
         | 1761 |  |  |           /* We can't modify t, because we use it at the end of the
 | 
      
         | 1762 |  |  |              function.  */
 | 
      
         | 1763 |  |  |           tree t2 = t;
 | 
      
         | 1764 |  |  |  
 | 
      
         | 1765 |  |  |           do
 | 
      
         | 1766 |  |  |             {
 | 
      
         | 1767 |  |  |               tree index = TREE_OPERAND (t2, 1);
 | 
      
         | 1768 |  |  |               tree low_bound = array_ref_low_bound (t2);
 | 
      
         | 1769 |  |  |               tree unit_size = array_ref_element_size (t2);
 | 
      
         | 1770 |  |  |  
 | 
      
         | 1771 |  |  |               /* We assume all arrays have sizes that are a multiple of a byte.
 | 
      
         | 1772 |  |  |                  First subtract the lower bound, if any, in the type of the
 | 
      
         | 1773 |  |  |                  index, then convert to sizetype and multiply by the size of
 | 
      
         | 1774 |  |  |                  the array element.  */
 | 
      
         | 1775 |  |  |               if (! integer_zerop (low_bound))
 | 
      
         | 1776 |  |  |                 index = fold_build2 (MINUS_EXPR, TREE_TYPE (index),
 | 
      
         | 1777 |  |  |                                      index, low_bound);
 | 
      
         | 1778 |  |  |  
 | 
      
         | 1779 |  |  |               off_tree = size_binop (PLUS_EXPR,
 | 
      
         | 1780 |  |  |                                      size_binop (MULT_EXPR,
 | 
      
         | 1781 |  |  |                                                  fold_convert (sizetype,
 | 
      
         | 1782 |  |  |                                                                index),
 | 
      
         | 1783 |  |  |                                                  unit_size),
 | 
      
         | 1784 |  |  |                                      off_tree);
 | 
      
         | 1785 |  |  |               t2 = TREE_OPERAND (t2, 0);
 | 
      
         | 1786 |  |  |             }
 | 
      
         | 1787 |  |  |           while (TREE_CODE (t2) == ARRAY_REF);
 | 
      
         | 1788 |  |  |  
 | 
      
         | 1789 |  |  |           if (DECL_P (t2))
 | 
      
         | 1790 |  |  |             {
 | 
      
         | 1791 |  |  |               attrs.expr = t2;
 | 
      
         | 1792 |  |  |               attrs.offset_known_p = false;
 | 
      
         | 1793 |  |  |               if (host_integerp (off_tree, 1))
 | 
      
         | 1794 |  |  |                 {
 | 
      
         | 1795 |  |  |                   HOST_WIDE_INT ioff = tree_low_cst (off_tree, 1);
 | 
      
         | 1796 |  |  |                   HOST_WIDE_INT aoff = (ioff & -ioff) * BITS_PER_UNIT;
 | 
      
         | 1797 |  |  |                   attrs.align = DECL_ALIGN (t2);
 | 
      
         | 1798 |  |  |                   if (aoff && (unsigned HOST_WIDE_INT) aoff < attrs.align)
 | 
      
         | 1799 |  |  |                     attrs.align = aoff;
 | 
      
         | 1800 |  |  |                   align_computed = true;
 | 
      
         | 1801 |  |  |                   attrs.offset_known_p = true;
 | 
      
         | 1802 |  |  |                   attrs.offset = ioff;
 | 
      
         | 1803 |  |  |                   apply_bitpos = bitpos;
 | 
      
         | 1804 |  |  |                 }
 | 
      
         | 1805 |  |  |             }
 | 
      
         | 1806 |  |  |           else if (TREE_CODE (t2) == COMPONENT_REF)
 | 
      
         | 1807 |  |  |             {
 | 
      
         | 1808 |  |  |               attrs.expr = t2;
 | 
      
         | 1809 |  |  |               attrs.offset_known_p = false;
 | 
      
         | 1810 |  |  |               if (host_integerp (off_tree, 1))
 | 
      
         | 1811 |  |  |                 {
 | 
      
         | 1812 |  |  |                   attrs.offset_known_p = true;
 | 
      
         | 1813 |  |  |                   attrs.offset = tree_low_cst (off_tree, 1);
 | 
      
         | 1814 |  |  |                   apply_bitpos = bitpos;
 | 
      
         | 1815 |  |  |                 }
 | 
      
         | 1816 |  |  |               /* ??? Any reason the field size would be different than
 | 
      
         | 1817 |  |  |                  the size we got from the type?  */
 | 
      
         | 1818 |  |  |             }
 | 
      
         | 1819 |  |  |  
 | 
      
         | 1820 |  |  |           /* If this is an indirect reference, record it.  */
 | 
      
         | 1821 |  |  |           else if (TREE_CODE (t) == MEM_REF)
 | 
      
         | 1822 |  |  |             {
 | 
      
         | 1823 |  |  |               attrs.expr = t;
 | 
      
         | 1824 |  |  |               attrs.offset_known_p = true;
 | 
      
         | 1825 |  |  |               attrs.offset = 0;
 | 
      
         | 1826 |  |  |               apply_bitpos = bitpos;
 | 
      
         | 1827 |  |  |             }
 | 
      
         | 1828 |  |  |         }
 | 
      
         | 1829 |  |  |  
 | 
      
         | 1830 |  |  |       /* If this is an indirect reference, record it.  */
 | 
      
         | 1831 |  |  |       else if (TREE_CODE (t) == MEM_REF
 | 
      
         | 1832 |  |  |                || TREE_CODE (t) == TARGET_MEM_REF)
 | 
      
         | 1833 |  |  |         {
 | 
      
         | 1834 |  |  |           attrs.expr = t;
 | 
      
         | 1835 |  |  |           attrs.offset_known_p = true;
 | 
      
         | 1836 |  |  |           attrs.offset = 0;
 | 
      
         | 1837 |  |  |           apply_bitpos = bitpos;
 | 
      
         | 1838 |  |  |         }
 | 
      
         | 1839 |  |  |  
 | 
      
         | 1840 |  |  |       if (!align_computed)
 | 
      
         | 1841 |  |  |         {
 | 
      
         | 1842 |  |  |           unsigned int obj_align = get_object_alignment (t);
 | 
      
         | 1843 |  |  |           attrs.align = MAX (attrs.align, obj_align);
 | 
      
         | 1844 |  |  |         }
 | 
      
         | 1845 |  |  |     }
 | 
      
         | 1846 |  |  |   else
 | 
      
         | 1847 |  |  |     as = TYPE_ADDR_SPACE (type);
 | 
      
         | 1848 |  |  |  
 | 
      
         | 1849 |  |  |   /* If we modified OFFSET based on T, then subtract the outstanding
 | 
      
         | 1850 |  |  |      bit position offset.  Similarly, increase the size of the accessed
 | 
      
         | 1851 |  |  |      object to contain the negative offset.  */
 | 
      
         | 1852 |  |  |   if (apply_bitpos)
 | 
      
         | 1853 |  |  |     {
 | 
      
         | 1854 |  |  |       gcc_assert (attrs.offset_known_p);
 | 
      
         | 1855 |  |  |       attrs.offset -= apply_bitpos / BITS_PER_UNIT;
 | 
      
         | 1856 |  |  |       if (attrs.size_known_p)
 | 
      
         | 1857 |  |  |         attrs.size += apply_bitpos / BITS_PER_UNIT;
 | 
      
         | 1858 |  |  |     }
 | 
      
         | 1859 |  |  |  
 | 
      
         | 1860 |  |  |   /* Now set the attributes we computed above.  */
 | 
      
         | 1861 |  |  |   attrs.addrspace = as;
 | 
      
         | 1862 |  |  |   set_mem_attrs (ref, &attrs);
 | 
      
         | 1863 |  |  | }
 | 
      
         | 1864 |  |  |  
 | 
      
         | 1865 |  |  | void
 | 
      
         | 1866 |  |  | set_mem_attributes (rtx ref, tree t, int objectp)
 | 
      
         | 1867 |  |  | {
 | 
      
         | 1868 |  |  |   set_mem_attributes_minus_bitpos (ref, t, objectp, 0);
 | 
      
         | 1869 |  |  | }
 | 
      
         | 1870 |  |  |  
 | 
      
         | 1871 |  |  | /* Set the alias set of MEM to SET.  */
 | 
      
         | 1872 |  |  |  
 | 
      
         | 1873 |  |  | void
 | 
      
         | 1874 |  |  | set_mem_alias_set (rtx mem, alias_set_type set)
 | 
      
         | 1875 |  |  | {
 | 
      
         | 1876 |  |  |   struct mem_attrs attrs;
 | 
      
         | 1877 |  |  |  
 | 
      
         | 1878 |  |  |   /* If the new and old alias sets don't conflict, something is wrong.  */
 | 
      
         | 1879 |  |  |   gcc_checking_assert (alias_sets_conflict_p (set, MEM_ALIAS_SET (mem)));
 | 
      
         | 1880 |  |  |   attrs = *get_mem_attrs (mem);
 | 
      
         | 1881 |  |  |   attrs.alias = set;
 | 
      
         | 1882 |  |  |   set_mem_attrs (mem, &attrs);
 | 
      
         | 1883 |  |  | }
 | 
      
         | 1884 |  |  |  
 | 
      
         | 1885 |  |  | /* Set the address space of MEM to ADDRSPACE (target-defined).  */
 | 
      
         | 1886 |  |  |  
 | 
      
         | 1887 |  |  | void
 | 
      
         | 1888 |  |  | set_mem_addr_space (rtx mem, addr_space_t addrspace)
 | 
      
         | 1889 |  |  | {
 | 
      
         | 1890 |  |  |   struct mem_attrs attrs;
 | 
      
         | 1891 |  |  |  
 | 
      
         | 1892 |  |  |   attrs = *get_mem_attrs (mem);
 | 
      
         | 1893 |  |  |   attrs.addrspace = addrspace;
 | 
      
         | 1894 |  |  |   set_mem_attrs (mem, &attrs);
 | 
      
         | 1895 |  |  | }
 | 
      
         | 1896 |  |  |  
 | 
      
         | 1897 |  |  | /* Set the alignment of MEM to ALIGN bits.  */
 | 
      
         | 1898 |  |  |  
 | 
      
         | 1899 |  |  | void
 | 
      
         | 1900 |  |  | set_mem_align (rtx mem, unsigned int align)
 | 
      
         | 1901 |  |  | {
 | 
      
         | 1902 |  |  |   struct mem_attrs attrs;
 | 
      
         | 1903 |  |  |  
 | 
      
         | 1904 |  |  |   attrs = *get_mem_attrs (mem);
 | 
      
         | 1905 |  |  |   attrs.align = align;
 | 
      
         | 1906 |  |  |   set_mem_attrs (mem, &attrs);
 | 
      
         | 1907 |  |  | }
 | 
      
         | 1908 |  |  |  
 | 
      
         | 1909 |  |  | /* Set the expr for MEM to EXPR.  */
 | 
      
         | 1910 |  |  |  
 | 
      
         | 1911 |  |  | void
 | 
      
         | 1912 |  |  | set_mem_expr (rtx mem, tree expr)
 | 
      
         | 1913 |  |  | {
 | 
      
         | 1914 |  |  |   struct mem_attrs attrs;
 | 
      
         | 1915 |  |  |  
 | 
      
         | 1916 |  |  |   attrs = *get_mem_attrs (mem);
 | 
      
         | 1917 |  |  |   attrs.expr = expr;
 | 
      
         | 1918 |  |  |   set_mem_attrs (mem, &attrs);
 | 
      
         | 1919 |  |  | }
 | 
      
         | 1920 |  |  |  
 | 
      
         | 1921 |  |  | /* Set the offset of MEM to OFFSET.  */
 | 
      
         | 1922 |  |  |  
 | 
      
         | 1923 |  |  | void
 | 
      
         | 1924 |  |  | set_mem_offset (rtx mem, HOST_WIDE_INT offset)
 | 
      
         | 1925 |  |  | {
 | 
      
         | 1926 |  |  |   struct mem_attrs attrs;
 | 
      
         | 1927 |  |  |  
 | 
      
         | 1928 |  |  |   attrs = *get_mem_attrs (mem);
 | 
      
         | 1929 |  |  |   attrs.offset_known_p = true;
 | 
      
         | 1930 |  |  |   attrs.offset = offset;
 | 
      
         | 1931 |  |  |   set_mem_attrs (mem, &attrs);
 | 
      
         | 1932 |  |  | }
 | 
      
         | 1933 |  |  |  
 | 
      
         | 1934 |  |  | /* Clear the offset of MEM.  */
 | 
      
         | 1935 |  |  |  
 | 
      
         | 1936 |  |  | void
 | 
      
         | 1937 |  |  | clear_mem_offset (rtx mem)
 | 
      
         | 1938 |  |  | {
 | 
      
         | 1939 |  |  |   struct mem_attrs attrs;
 | 
      
         | 1940 |  |  |  
 | 
      
         | 1941 |  |  |   attrs = *get_mem_attrs (mem);
 | 
      
         | 1942 |  |  |   attrs.offset_known_p = false;
 | 
      
         | 1943 |  |  |   set_mem_attrs (mem, &attrs);
 | 
      
         | 1944 |  |  | }
 | 
      
         | 1945 |  |  |  
 | 
      
         | 1946 |  |  | /* Set the size of MEM to SIZE.  */
 | 
      
         | 1947 |  |  |  
 | 
      
         | 1948 |  |  | void
 | 
      
         | 1949 |  |  | set_mem_size (rtx mem, HOST_WIDE_INT size)
 | 
      
         | 1950 |  |  | {
 | 
      
         | 1951 |  |  |   struct mem_attrs attrs;
 | 
      
         | 1952 |  |  |  
 | 
      
         | 1953 |  |  |   attrs = *get_mem_attrs (mem);
 | 
      
         | 1954 |  |  |   attrs.size_known_p = true;
 | 
      
         | 1955 |  |  |   attrs.size = size;
 | 
      
         | 1956 |  |  |   set_mem_attrs (mem, &attrs);
 | 
      
         | 1957 |  |  | }
 | 
      
         | 1958 |  |  |  
 | 
      
         | 1959 |  |  | /* Clear the size of MEM.  */
 | 
      
         | 1960 |  |  |  
 | 
      
         | 1961 |  |  | void
 | 
      
         | 1962 |  |  | clear_mem_size (rtx mem)
 | 
      
         | 1963 |  |  | {
 | 
      
         | 1964 |  |  |   struct mem_attrs attrs;
 | 
      
         | 1965 |  |  |  
 | 
      
         | 1966 |  |  |   attrs = *get_mem_attrs (mem);
 | 
      
         | 1967 |  |  |   attrs.size_known_p = false;
 | 
      
         | 1968 |  |  |   set_mem_attrs (mem, &attrs);
 | 
      
         | 1969 |  |  | }
 | 
      
         | 1970 |  |  |  
 | 
      
         | 1971 |  |  | /* Return a memory reference like MEMREF, but with its mode changed to MODE
 | 
      
         | 1972 |  |  |    and its address changed to ADDR.  (VOIDmode means don't change the mode.
 | 
      
         | 1973 |  |  |    NULL for ADDR means don't change the address.)  VALIDATE is nonzero if the
 | 
      
         | 1974 |  |  |    returned memory location is required to be valid.  The memory
 | 
      
         | 1975 |  |  |    attributes are not changed.  */
 | 
      
         | 1976 |  |  |  
 | 
      
         | 1977 |  |  | static rtx
 | 
      
         | 1978 |  |  | change_address_1 (rtx memref, enum machine_mode mode, rtx addr, int validate)
 | 
      
         | 1979 |  |  | {
 | 
      
         | 1980 |  |  |   addr_space_t as;
 | 
      
         | 1981 |  |  |   rtx new_rtx;
 | 
      
         | 1982 |  |  |  
 | 
      
         | 1983 |  |  |   gcc_assert (MEM_P (memref));
 | 
      
         | 1984 |  |  |   as = MEM_ADDR_SPACE (memref);
 | 
      
         | 1985 |  |  |   if (mode == VOIDmode)
 | 
      
         | 1986 |  |  |     mode = GET_MODE (memref);
 | 
      
         | 1987 |  |  |   if (addr == 0)
 | 
      
         | 1988 |  |  |     addr = XEXP (memref, 0);
 | 
      
         | 1989 |  |  |   if (mode == GET_MODE (memref) && addr == XEXP (memref, 0)
 | 
      
         | 1990 |  |  |       && (!validate || memory_address_addr_space_p (mode, addr, as)))
 | 
      
         | 1991 |  |  |     return memref;
 | 
      
         | 1992 |  |  |  
 | 
      
         | 1993 |  |  |   if (validate)
 | 
      
         | 1994 |  |  |     {
 | 
      
         | 1995 |  |  |       if (reload_in_progress || reload_completed)
 | 
      
         | 1996 |  |  |         gcc_assert (memory_address_addr_space_p (mode, addr, as));
 | 
      
         | 1997 |  |  |       else
 | 
      
         | 1998 |  |  |         addr = memory_address_addr_space (mode, addr, as);
 | 
      
         | 1999 |  |  |     }
 | 
      
         | 2000 |  |  |  
 | 
      
         | 2001 |  |  |   if (rtx_equal_p (addr, XEXP (memref, 0)) && mode == GET_MODE (memref))
 | 
      
         | 2002 |  |  |     return memref;
 | 
      
         | 2003 |  |  |  
 | 
      
         | 2004 |  |  |   new_rtx = gen_rtx_MEM (mode, addr);
 | 
      
         | 2005 |  |  |   MEM_COPY_ATTRIBUTES (new_rtx, memref);
 | 
      
         | 2006 |  |  |   return new_rtx;
 | 
      
         | 2007 |  |  | }
 | 
      
         | 2008 |  |  |  
 | 
      
         | 2009 |  |  | /* Like change_address_1 with VALIDATE nonzero, but we are not saying in what
 | 
      
         | 2010 |  |  |    way we are changing MEMREF, so we only preserve the alias set.  */
 | 
      
         | 2011 |  |  |  
 | 
      
         | 2012 |  |  | rtx
 | 
      
         | 2013 |  |  | change_address (rtx memref, enum machine_mode mode, rtx addr)
 | 
      
         | 2014 |  |  | {
 | 
      
         | 2015 |  |  |   rtx new_rtx = change_address_1 (memref, mode, addr, 1);
 | 
      
         | 2016 |  |  |   enum machine_mode mmode = GET_MODE (new_rtx);
 | 
      
         | 2017 |  |  |   struct mem_attrs attrs, *defattrs;
 | 
      
         | 2018 |  |  |  
 | 
      
         | 2019 |  |  |   attrs = *get_mem_attrs (memref);
 | 
      
         | 2020 |  |  |   defattrs = mode_mem_attrs[(int) mmode];
 | 
      
         | 2021 |  |  |   attrs.expr = NULL_TREE;
 | 
      
         | 2022 |  |  |   attrs.offset_known_p = false;
 | 
      
         | 2023 |  |  |   attrs.size_known_p = defattrs->size_known_p;
 | 
      
         | 2024 |  |  |   attrs.size = defattrs->size;
 | 
      
         | 2025 |  |  |   attrs.align = defattrs->align;
 | 
      
         | 2026 |  |  |  
 | 
      
         | 2027 |  |  |   /* If there are no changes, just return the original memory reference.  */
 | 
      
         | 2028 |  |  |   if (new_rtx == memref)
 | 
      
         | 2029 |  |  |     {
 | 
      
         | 2030 |  |  |       if (mem_attrs_eq_p (get_mem_attrs (memref), &attrs))
 | 
      
         | 2031 |  |  |         return new_rtx;
 | 
      
         | 2032 |  |  |  
 | 
      
         | 2033 |  |  |       new_rtx = gen_rtx_MEM (mmode, XEXP (memref, 0));
 | 
      
         | 2034 |  |  |       MEM_COPY_ATTRIBUTES (new_rtx, memref);
 | 
      
         | 2035 |  |  |     }
 | 
      
         | 2036 |  |  |  
 | 
      
         | 2037 |  |  |   set_mem_attrs (new_rtx, &attrs);
 | 
      
         | 2038 |  |  |   return new_rtx;
 | 
      
         | 2039 |  |  | }
 | 
      
         | 2040 |  |  |  
 | 
      
         | 2041 |  |  | /* Return a memory reference like MEMREF, but with its mode changed
 | 
      
         | 2042 |  |  |    to MODE and its address offset by OFFSET bytes.  If VALIDATE is
 | 
      
         | 2043 |  |  |    nonzero, the memory address is forced to be valid.
 | 
      
         | 2044 |  |  |    If ADJUST is zero, OFFSET is only used to update MEM_ATTRS
 | 
      
         | 2045 |  |  |    and caller is responsible for adjusting MEMREF base register.  */
 | 
      
         | 2046 |  |  |  
 | 
      
         | 2047 |  |  | rtx
 | 
      
         | 2048 |  |  | adjust_address_1 (rtx memref, enum machine_mode mode, HOST_WIDE_INT offset,
 | 
      
         | 2049 |  |  |                   int validate, int adjust)
 | 
      
         | 2050 |  |  | {
 | 
      
         | 2051 |  |  |   rtx addr = XEXP (memref, 0);
 | 
      
         | 2052 |  |  |   rtx new_rtx;
 | 
      
         | 2053 |  |  |   enum machine_mode address_mode;
 | 
      
         | 2054 |  |  |   int pbits;
 | 
      
         | 2055 |  |  |   struct mem_attrs attrs, *defattrs;
 | 
      
         | 2056 |  |  |   unsigned HOST_WIDE_INT max_align;
 | 
      
         | 2057 |  |  |  
 | 
      
         | 2058 |  |  |   attrs = *get_mem_attrs (memref);
 | 
      
         | 2059 |  |  |  
 | 
      
         | 2060 |  |  |   /* If there are no changes, just return the original memory reference.  */
 | 
      
         | 2061 |  |  |   if (mode == GET_MODE (memref) && !offset
 | 
      
         | 2062 |  |  |       && (!validate || memory_address_addr_space_p (mode, addr,
 | 
      
         | 2063 |  |  |                                                     attrs.addrspace)))
 | 
      
         | 2064 |  |  |     return memref;
 | 
      
         | 2065 |  |  |  
 | 
      
         | 2066 |  |  |   /* ??? Prefer to create garbage instead of creating shared rtl.
 | 
      
         | 2067 |  |  |      This may happen even if offset is nonzero -- consider
 | 
      
         | 2068 |  |  |      (plus (plus reg reg) const_int) -- so do this always.  */
 | 
      
         | 2069 |  |  |   addr = copy_rtx (addr);
 | 
      
         | 2070 |  |  |  
 | 
      
         | 2071 |  |  |   /* Convert a possibly large offset to a signed value within the
 | 
      
         | 2072 |  |  |      range of the target address space.  */
 | 
      
         | 2073 |  |  |   address_mode = targetm.addr_space.address_mode (attrs.addrspace);
 | 
      
         | 2074 |  |  |   pbits = GET_MODE_BITSIZE (address_mode);
 | 
      
         | 2075 |  |  |   if (HOST_BITS_PER_WIDE_INT > pbits)
 | 
      
         | 2076 |  |  |     {
 | 
      
         | 2077 |  |  |       int shift = HOST_BITS_PER_WIDE_INT - pbits;
 | 
      
         | 2078 |  |  |       offset = (((HOST_WIDE_INT) ((unsigned HOST_WIDE_INT) offset << shift))
 | 
      
         | 2079 |  |  |                 >> shift);
 | 
      
         | 2080 |  |  |     }
 | 
      
         | 2081 |  |  |  
 | 
      
         | 2082 |  |  |   if (adjust)
 | 
      
         | 2083 |  |  |     {
 | 
      
         | 2084 |  |  |       /* If MEMREF is a LO_SUM and the offset is within the alignment of the
 | 
      
         | 2085 |  |  |          object, we can merge it into the LO_SUM.  */
 | 
      
         | 2086 |  |  |       if (GET_MODE (memref) != BLKmode && GET_CODE (addr) == LO_SUM
 | 
      
         | 2087 |  |  |           && offset >= 0
 | 
      
         | 2088 |  |  |           && (unsigned HOST_WIDE_INT) offset
 | 
      
         | 2089 |  |  |               < GET_MODE_ALIGNMENT (GET_MODE (memref)) / BITS_PER_UNIT)
 | 
      
         | 2090 |  |  |         addr = gen_rtx_LO_SUM (address_mode, XEXP (addr, 0),
 | 
      
         | 2091 |  |  |                                plus_constant (XEXP (addr, 1), offset));
 | 
      
         | 2092 |  |  |       else
 | 
      
         | 2093 |  |  |         addr = plus_constant (addr, offset);
 | 
      
         | 2094 |  |  |     }
 | 
      
         | 2095 |  |  |  
 | 
      
         | 2096 |  |  |   new_rtx = change_address_1 (memref, mode, addr, validate);
 | 
      
         | 2097 |  |  |  
 | 
      
         | 2098 |  |  |   /* If the address is a REG, change_address_1 rightfully returns memref,
 | 
      
         | 2099 |  |  |      but this would destroy memref's MEM_ATTRS.  */
 | 
      
         | 2100 |  |  |   if (new_rtx == memref && offset != 0)
 | 
      
         | 2101 |  |  |     new_rtx = copy_rtx (new_rtx);
 | 
      
         | 2102 |  |  |  
 | 
      
         | 2103 |  |  |   /* Compute the new values of the memory attributes due to this adjustment.
 | 
      
         | 2104 |  |  |      We add the offsets and update the alignment.  */
 | 
      
         | 2105 |  |  |   if (attrs.offset_known_p)
 | 
      
         | 2106 |  |  |     attrs.offset += offset;
 | 
      
         | 2107 |  |  |  
 | 
      
         | 2108 |  |  |   /* Compute the new alignment by taking the MIN of the alignment and the
 | 
      
         | 2109 |  |  |      lowest-order set bit in OFFSET, but don't change the alignment if OFFSET
 | 
      
         | 2110 |  |  |      if zero.  */
 | 
      
         | 2111 |  |  |   if (offset != 0)
 | 
      
         | 2112 |  |  |     {
 | 
      
         | 2113 |  |  |       max_align = (offset & -offset) * BITS_PER_UNIT;
 | 
      
         | 2114 |  |  |       attrs.align = MIN (attrs.align, max_align);
 | 
      
         | 2115 |  |  |     }
 | 
      
         | 2116 |  |  |  
 | 
      
         | 2117 |  |  |   /* We can compute the size in a number of ways.  */
 | 
      
         | 2118 |  |  |   defattrs = mode_mem_attrs[(int) GET_MODE (new_rtx)];
 | 
      
         | 2119 |  |  |   if (defattrs->size_known_p)
 | 
      
         | 2120 |  |  |     {
 | 
      
         | 2121 |  |  |       attrs.size_known_p = true;
 | 
      
         | 2122 |  |  |       attrs.size = defattrs->size;
 | 
      
         | 2123 |  |  |     }
 | 
      
         | 2124 |  |  |   else if (attrs.size_known_p)
 | 
      
         | 2125 |  |  |     attrs.size -= offset;
 | 
      
         | 2126 |  |  |  
 | 
      
         | 2127 |  |  |   set_mem_attrs (new_rtx, &attrs);
 | 
      
         | 2128 |  |  |  
 | 
      
         | 2129 |  |  |   /* At some point, we should validate that this offset is within the object,
 | 
      
         | 2130 |  |  |      if all the appropriate values are known.  */
 | 
      
         | 2131 |  |  |   return new_rtx;
 | 
      
         | 2132 |  |  | }
 | 
      
         | 2133 |  |  |  
 | 
      
         | 2134 |  |  | /* Return a memory reference like MEMREF, but with its mode changed
 | 
      
         | 2135 |  |  |    to MODE and its address changed to ADDR, which is assumed to be
 | 
      
         | 2136 |  |  |    MEMREF offset by OFFSET bytes.  If VALIDATE is
 | 
      
         | 2137 |  |  |    nonzero, the memory address is forced to be valid.  */
 | 
      
         | 2138 |  |  |  
 | 
      
         | 2139 |  |  | rtx
 | 
      
         | 2140 |  |  | adjust_automodify_address_1 (rtx memref, enum machine_mode mode, rtx addr,
 | 
      
         | 2141 |  |  |                              HOST_WIDE_INT offset, int validate)
 | 
      
         | 2142 |  |  | {
 | 
      
         | 2143 |  |  |   memref = change_address_1 (memref, VOIDmode, addr, validate);
 | 
      
         | 2144 |  |  |   return adjust_address_1 (memref, mode, offset, validate, 0);
 | 
      
         | 2145 |  |  | }
 | 
      
         | 2146 |  |  |  
 | 
      
         | 2147 |  |  | /* Return a memory reference like MEMREF, but whose address is changed by
 | 
      
         | 2148 |  |  |    adding OFFSET, an RTX, to it.  POW2 is the highest power of two factor
 | 
      
         | 2149 |  |  |    known to be in OFFSET (possibly 1).  */
 | 
      
         | 2150 |  |  |  
 | 
      
         | 2151 |  |  | rtx
 | 
      
         | 2152 |  |  | offset_address (rtx memref, rtx offset, unsigned HOST_WIDE_INT pow2)
 | 
      
         | 2153 |  |  | {
 | 
      
         | 2154 |  |  |   rtx new_rtx, addr = XEXP (memref, 0);
 | 
      
         | 2155 |  |  |   enum machine_mode address_mode;
 | 
      
         | 2156 |  |  |   struct mem_attrs attrs, *defattrs;
 | 
      
         | 2157 |  |  |  
 | 
      
         | 2158 |  |  |   attrs = *get_mem_attrs (memref);
 | 
      
         | 2159 |  |  |   address_mode = targetm.addr_space.address_mode (attrs.addrspace);
 | 
      
         | 2160 |  |  |   new_rtx = simplify_gen_binary (PLUS, address_mode, addr, offset);
 | 
      
         | 2161 |  |  |  
 | 
      
         | 2162 |  |  |   /* At this point we don't know _why_ the address is invalid.  It
 | 
      
         | 2163 |  |  |      could have secondary memory references, multiplies or anything.
 | 
      
         | 2164 |  |  |  
 | 
      
         | 2165 |  |  |      However, if we did go and rearrange things, we can wind up not
 | 
      
         | 2166 |  |  |      being able to recognize the magic around pic_offset_table_rtx.
 | 
      
         | 2167 |  |  |      This stuff is fragile, and is yet another example of why it is
 | 
      
         | 2168 |  |  |      bad to expose PIC machinery too early.  */
 | 
      
         | 2169 |  |  |   if (! memory_address_addr_space_p (GET_MODE (memref), new_rtx,
 | 
      
         | 2170 |  |  |                                      attrs.addrspace)
 | 
      
         | 2171 |  |  |       && GET_CODE (addr) == PLUS
 | 
      
         | 2172 |  |  |       && XEXP (addr, 0) == pic_offset_table_rtx)
 | 
      
         | 2173 |  |  |     {
 | 
      
         | 2174 |  |  |       addr = force_reg (GET_MODE (addr), addr);
 | 
      
         | 2175 |  |  |       new_rtx = simplify_gen_binary (PLUS, address_mode, addr, offset);
 | 
      
         | 2176 |  |  |     }
 | 
      
         | 2177 |  |  |  
 | 
      
         | 2178 |  |  |   update_temp_slot_address (XEXP (memref, 0), new_rtx);
 | 
      
         | 2179 |  |  |   new_rtx = change_address_1 (memref, VOIDmode, new_rtx, 1);
 | 
      
         | 2180 |  |  |  
 | 
      
         | 2181 |  |  |   /* If there are no changes, just return the original memory reference.  */
 | 
      
         | 2182 |  |  |   if (new_rtx == memref)
 | 
      
         | 2183 |  |  |     return new_rtx;
 | 
      
         | 2184 |  |  |  
 | 
      
         | 2185 |  |  |   /* Update the alignment to reflect the offset.  Reset the offset, which
 | 
      
         | 2186 |  |  |      we don't know.  */
 | 
      
         | 2187 |  |  |   defattrs = mode_mem_attrs[(int) GET_MODE (new_rtx)];
 | 
      
         | 2188 |  |  |   attrs.offset_known_p = false;
 | 
      
         | 2189 |  |  |   attrs.size_known_p = defattrs->size_known_p;
 | 
      
         | 2190 |  |  |   attrs.size = defattrs->size;
 | 
      
         | 2191 |  |  |   attrs.align = MIN (attrs.align, pow2 * BITS_PER_UNIT);
 | 
      
         | 2192 |  |  |   set_mem_attrs (new_rtx, &attrs);
 | 
      
         | 2193 |  |  |   return new_rtx;
 | 
      
         | 2194 |  |  | }
 | 
      
         | 2195 |  |  |  
 | 
      
         | 2196 |  |  | /* Return a memory reference like MEMREF, but with its address changed to
 | 
      
         | 2197 |  |  |    ADDR.  The caller is asserting that the actual piece of memory pointed
 | 
      
         | 2198 |  |  |    to is the same, just the form of the address is being changed, such as
 | 
      
         | 2199 |  |  |    by putting something into a register.  */
 | 
      
         | 2200 |  |  |  
 | 
      
         | 2201 |  |  | rtx
 | 
      
         | 2202 |  |  | replace_equiv_address (rtx memref, rtx addr)
 | 
      
         | 2203 |  |  | {
 | 
      
         | 2204 |  |  |   /* change_address_1 copies the memory attribute structure without change
 | 
      
         | 2205 |  |  |      and that's exactly what we want here.  */
 | 
      
         | 2206 |  |  |   update_temp_slot_address (XEXP (memref, 0), addr);
 | 
      
         | 2207 |  |  |   return change_address_1 (memref, VOIDmode, addr, 1);
 | 
      
         | 2208 |  |  | }
 | 
      
         | 2209 |  |  |  
 | 
      
         | 2210 |  |  | /* Likewise, but the reference is not required to be valid.  */
 | 
      
         | 2211 |  |  |  
 | 
      
         | 2212 |  |  | rtx
 | 
      
         | 2213 |  |  | replace_equiv_address_nv (rtx memref, rtx addr)
 | 
      
         | 2214 |  |  | {
 | 
      
         | 2215 |  |  |   return change_address_1 (memref, VOIDmode, addr, 0);
 | 
      
         | 2216 |  |  | }
 | 
      
         | 2217 |  |  |  
 | 
      
         | 2218 |  |  | /* Return a memory reference like MEMREF, but with its mode widened to
 | 
      
         | 2219 |  |  |    MODE and offset by OFFSET.  This would be used by targets that e.g.
 | 
      
         | 2220 |  |  |    cannot issue QImode memory operations and have to use SImode memory
 | 
      
         | 2221 |  |  |    operations plus masking logic.  */
 | 
      
         | 2222 |  |  |  
 | 
      
         | 2223 |  |  | rtx
 | 
      
         | 2224 |  |  | widen_memory_access (rtx memref, enum machine_mode mode, HOST_WIDE_INT offset)
 | 
      
         | 2225 |  |  | {
 | 
      
         | 2226 |  |  |   rtx new_rtx = adjust_address_1 (memref, mode, offset, 1, 1);
 | 
      
         | 2227 |  |  |   struct mem_attrs attrs;
 | 
      
         | 2228 |  |  |   unsigned int size = GET_MODE_SIZE (mode);
 | 
      
         | 2229 |  |  |  
 | 
      
         | 2230 |  |  |   /* If there are no changes, just return the original memory reference.  */
 | 
      
         | 2231 |  |  |   if (new_rtx == memref)
 | 
      
         | 2232 |  |  |     return new_rtx;
 | 
      
         | 2233 |  |  |  
 | 
      
         | 2234 |  |  |   attrs = *get_mem_attrs (new_rtx);
 | 
      
         | 2235 |  |  |  
 | 
      
         | 2236 |  |  |   /* If we don't know what offset we were at within the expression, then
 | 
      
         | 2237 |  |  |      we can't know if we've overstepped the bounds.  */
 | 
      
         | 2238 |  |  |   if (! attrs.offset_known_p)
 | 
      
         | 2239 |  |  |     attrs.expr = NULL_TREE;
 | 
      
         | 2240 |  |  |  
 | 
      
         | 2241 |  |  |   while (attrs.expr)
 | 
      
         | 2242 |  |  |     {
 | 
      
         | 2243 |  |  |       if (TREE_CODE (attrs.expr) == COMPONENT_REF)
 | 
      
         | 2244 |  |  |         {
 | 
      
         | 2245 |  |  |           tree field = TREE_OPERAND (attrs.expr, 1);
 | 
      
         | 2246 |  |  |           tree offset = component_ref_field_offset (attrs.expr);
 | 
      
         | 2247 |  |  |  
 | 
      
         | 2248 |  |  |           if (! DECL_SIZE_UNIT (field))
 | 
      
         | 2249 |  |  |             {
 | 
      
         | 2250 |  |  |               attrs.expr = NULL_TREE;
 | 
      
         | 2251 |  |  |               break;
 | 
      
         | 2252 |  |  |             }
 | 
      
         | 2253 |  |  |  
 | 
      
         | 2254 |  |  |           /* Is the field at least as large as the access?  If so, ok,
 | 
      
         | 2255 |  |  |              otherwise strip back to the containing structure.  */
 | 
      
         | 2256 |  |  |           if (TREE_CODE (DECL_SIZE_UNIT (field)) == INTEGER_CST
 | 
      
         | 2257 |  |  |               && compare_tree_int (DECL_SIZE_UNIT (field), size) >= 0
 | 
      
         | 2258 |  |  |               && attrs.offset >= 0)
 | 
      
         | 2259 |  |  |             break;
 | 
      
         | 2260 |  |  |  
 | 
      
         | 2261 |  |  |           if (! host_integerp (offset, 1))
 | 
      
         | 2262 |  |  |             {
 | 
      
         | 2263 |  |  |               attrs.expr = NULL_TREE;
 | 
      
         | 2264 |  |  |               break;
 | 
      
         | 2265 |  |  |             }
 | 
      
         | 2266 |  |  |  
 | 
      
         | 2267 |  |  |           attrs.expr = TREE_OPERAND (attrs.expr, 0);
 | 
      
         | 2268 |  |  |           attrs.offset += tree_low_cst (offset, 1);
 | 
      
         | 2269 |  |  |           attrs.offset += (tree_low_cst (DECL_FIELD_BIT_OFFSET (field), 1)
 | 
      
         | 2270 |  |  |                            / BITS_PER_UNIT);
 | 
      
         | 2271 |  |  |         }
 | 
      
         | 2272 |  |  |       /* Similarly for the decl.  */
 | 
      
         | 2273 |  |  |       else if (DECL_P (attrs.expr)
 | 
      
         | 2274 |  |  |                && DECL_SIZE_UNIT (attrs.expr)
 | 
      
         | 2275 |  |  |                && TREE_CODE (DECL_SIZE_UNIT (attrs.expr)) == INTEGER_CST
 | 
      
         | 2276 |  |  |                && compare_tree_int (DECL_SIZE_UNIT (attrs.expr), size) >= 0
 | 
      
         | 2277 |  |  |                && (! attrs.offset_known_p || attrs.offset >= 0))
 | 
      
         | 2278 |  |  |         break;
 | 
      
         | 2279 |  |  |       else
 | 
      
         | 2280 |  |  |         {
 | 
      
         | 2281 |  |  |           /* The widened memory access overflows the expression, which means
 | 
      
         | 2282 |  |  |              that it could alias another expression.  Zap it.  */
 | 
      
         | 2283 |  |  |           attrs.expr = NULL_TREE;
 | 
      
         | 2284 |  |  |           break;
 | 
      
         | 2285 |  |  |         }
 | 
      
         | 2286 |  |  |     }
 | 
      
         | 2287 |  |  |  
 | 
      
         | 2288 |  |  |   if (! attrs.expr)
 | 
      
         | 2289 |  |  |     attrs.offset_known_p = false;
 | 
      
         | 2290 |  |  |  
 | 
      
         | 2291 |  |  |   /* The widened memory may alias other stuff, so zap the alias set.  */
 | 
      
         | 2292 |  |  |   /* ??? Maybe use get_alias_set on any remaining expression.  */
 | 
      
         | 2293 |  |  |   attrs.alias = 0;
 | 
      
         | 2294 |  |  |   attrs.size_known_p = true;
 | 
      
         | 2295 |  |  |   attrs.size = size;
 | 
      
         | 2296 |  |  |   set_mem_attrs (new_rtx, &attrs);
 | 
      
         | 2297 |  |  |   return new_rtx;
 | 
      
         | 2298 |  |  | }
 | 
      
         | 2299 |  |  |  
 | 
      
         | 2300 |  |  | /* A fake decl that is used as the MEM_EXPR of spill slots.  */
 | 
      
         | 2301 |  |  | static GTY(()) tree spill_slot_decl;
 | 
      
         | 2302 |  |  |  
 | 
      
         | 2303 |  |  | tree
 | 
      
         | 2304 |  |  | get_spill_slot_decl (bool force_build_p)
 | 
      
         | 2305 |  |  | {
 | 
      
         | 2306 |  |  |   tree d = spill_slot_decl;
 | 
      
         | 2307 |  |  |   rtx rd;
 | 
      
         | 2308 |  |  |   struct mem_attrs attrs;
 | 
      
         | 2309 |  |  |  
 | 
      
         | 2310 |  |  |   if (d || !force_build_p)
 | 
      
         | 2311 |  |  |     return d;
 | 
      
         | 2312 |  |  |  
 | 
      
         | 2313 |  |  |   d = build_decl (DECL_SOURCE_LOCATION (current_function_decl),
 | 
      
         | 2314 |  |  |                   VAR_DECL, get_identifier ("%sfp"), void_type_node);
 | 
      
         | 2315 |  |  |   DECL_ARTIFICIAL (d) = 1;
 | 
      
         | 2316 |  |  |   DECL_IGNORED_P (d) = 1;
 | 
      
         | 2317 |  |  |   TREE_USED (d) = 1;
 | 
      
         | 2318 |  |  |   spill_slot_decl = d;
 | 
      
         | 2319 |  |  |  
 | 
      
         | 2320 |  |  |   rd = gen_rtx_MEM (BLKmode, frame_pointer_rtx);
 | 
      
         | 2321 |  |  |   MEM_NOTRAP_P (rd) = 1;
 | 
      
         | 2322 |  |  |   attrs = *mode_mem_attrs[(int) BLKmode];
 | 
      
         | 2323 |  |  |   attrs.alias = new_alias_set ();
 | 
      
         | 2324 |  |  |   attrs.expr = d;
 | 
      
         | 2325 |  |  |   set_mem_attrs (rd, &attrs);
 | 
      
         | 2326 |  |  |   SET_DECL_RTL (d, rd);
 | 
      
         | 2327 |  |  |  
 | 
      
         | 2328 |  |  |   return d;
 | 
      
         | 2329 |  |  | }
 | 
      
         | 2330 |  |  |  
 | 
      
         | 2331 |  |  | /* Given MEM, a result from assign_stack_local, fill in the memory
 | 
      
         | 2332 |  |  |    attributes as appropriate for a register allocator spill slot.
 | 
      
         | 2333 |  |  |    These slots are not aliasable by other memory.  We arrange for
 | 
      
         | 2334 |  |  |    them all to use a single MEM_EXPR, so that the aliasing code can
 | 
      
         | 2335 |  |  |    work properly in the case of shared spill slots.  */
 | 
      
         | 2336 |  |  |  
 | 
      
         | 2337 |  |  | void
 | 
      
         | 2338 |  |  | set_mem_attrs_for_spill (rtx mem)
 | 
      
         | 2339 |  |  | {
 | 
      
         | 2340 |  |  |   struct mem_attrs attrs;
 | 
      
         | 2341 |  |  |   rtx addr;
 | 
      
         | 2342 |  |  |  
 | 
      
         | 2343 |  |  |   attrs = *get_mem_attrs (mem);
 | 
      
         | 2344 |  |  |   attrs.expr = get_spill_slot_decl (true);
 | 
      
         | 2345 |  |  |   attrs.alias = MEM_ALIAS_SET (DECL_RTL (attrs.expr));
 | 
      
         | 2346 |  |  |   attrs.addrspace = ADDR_SPACE_GENERIC;
 | 
      
         | 2347 |  |  |  
 | 
      
         | 2348 |  |  |   /* We expect the incoming memory to be of the form:
 | 
      
         | 2349 |  |  |         (mem:MODE (plus (reg sfp) (const_int offset)))
 | 
      
         | 2350 |  |  |      with perhaps the plus missing for offset = 0.  */
 | 
      
         | 2351 |  |  |   addr = XEXP (mem, 0);
 | 
      
         | 2352 |  |  |   attrs.offset_known_p = true;
 | 
      
         | 2353 |  |  |   attrs.offset = 0;
 | 
      
         | 2354 |  |  |   if (GET_CODE (addr) == PLUS
 | 
      
         | 2355 |  |  |       && CONST_INT_P (XEXP (addr, 1)))
 | 
      
         | 2356 |  |  |     attrs.offset = INTVAL (XEXP (addr, 1));
 | 
      
         | 2357 |  |  |  
 | 
      
         | 2358 |  |  |   set_mem_attrs (mem, &attrs);
 | 
      
         | 2359 |  |  |   MEM_NOTRAP_P (mem) = 1;
 | 
      
         | 2360 |  |  | }
 | 
      
         | 2361 |  |  |  
 | 
      
         | 2362 |  |  | /* Return a newly created CODE_LABEL rtx with a unique label number.  */
 | 
      
         | 2363 |  |  |  
 | 
      
         | 2364 |  |  | rtx
 | 
      
         | 2365 |  |  | gen_label_rtx (void)
 | 
      
         | 2366 |  |  | {
 | 
      
         | 2367 |  |  |   return gen_rtx_CODE_LABEL (VOIDmode, 0, NULL_RTX, NULL_RTX,
 | 
      
         | 2368 |  |  |                              NULL, label_num++, NULL);
 | 
      
         | 2369 |  |  | }
 | 
      
         | 2370 |  |  |  
 | 
      
         | 2371 |  |  | /* For procedure integration.  */
 | 
      
         | 2372 |  |  |  
 | 
      
         | 2373 |  |  | /* Install new pointers to the first and last insns in the chain.
 | 
      
         | 2374 |  |  |    Also, set cur_insn_uid to one higher than the last in use.
 | 
      
         | 2375 |  |  |    Used for an inline-procedure after copying the insn chain.  */
 | 
      
         | 2376 |  |  |  
 | 
      
         | 2377 |  |  | void
 | 
      
         | 2378 |  |  | set_new_first_and_last_insn (rtx first, rtx last)
 | 
      
         | 2379 |  |  | {
 | 
      
         | 2380 |  |  |   rtx insn;
 | 
      
         | 2381 |  |  |  
 | 
      
         | 2382 |  |  |   set_first_insn (first);
 | 
      
         | 2383 |  |  |   set_last_insn (last);
 | 
      
         | 2384 |  |  |   cur_insn_uid = 0;
 | 
      
         | 2385 |  |  |  
 | 
      
         | 2386 |  |  |   if (MIN_NONDEBUG_INSN_UID || MAY_HAVE_DEBUG_INSNS)
 | 
      
         | 2387 |  |  |     {
 | 
      
         | 2388 |  |  |       int debug_count = 0;
 | 
      
         | 2389 |  |  |  
 | 
      
         | 2390 |  |  |       cur_insn_uid = MIN_NONDEBUG_INSN_UID - 1;
 | 
      
         | 2391 |  |  |       cur_debug_insn_uid = 0;
 | 
      
         | 2392 |  |  |  
 | 
      
         | 2393 |  |  |       for (insn = first; insn; insn = NEXT_INSN (insn))
 | 
      
         | 2394 |  |  |         if (INSN_UID (insn) < MIN_NONDEBUG_INSN_UID)
 | 
      
         | 2395 |  |  |           cur_debug_insn_uid = MAX (cur_debug_insn_uid, INSN_UID (insn));
 | 
      
         | 2396 |  |  |         else
 | 
      
         | 2397 |  |  |           {
 | 
      
         | 2398 |  |  |             cur_insn_uid = MAX (cur_insn_uid, INSN_UID (insn));
 | 
      
         | 2399 |  |  |             if (DEBUG_INSN_P (insn))
 | 
      
         | 2400 |  |  |               debug_count++;
 | 
      
         | 2401 |  |  |           }
 | 
      
         | 2402 |  |  |  
 | 
      
         | 2403 |  |  |       if (debug_count)
 | 
      
         | 2404 |  |  |         cur_debug_insn_uid = MIN_NONDEBUG_INSN_UID + debug_count;
 | 
      
         | 2405 |  |  |       else
 | 
      
         | 2406 |  |  |         cur_debug_insn_uid++;
 | 
      
         | 2407 |  |  |     }
 | 
      
         | 2408 |  |  |   else
 | 
      
         | 2409 |  |  |     for (insn = first; insn; insn = NEXT_INSN (insn))
 | 
      
         | 2410 |  |  |       cur_insn_uid = MAX (cur_insn_uid, INSN_UID (insn));
 | 
      
         | 2411 |  |  |  
 | 
      
         | 2412 |  |  |   cur_insn_uid++;
 | 
      
         | 2413 |  |  | }
 | 
      
         | 2414 |  |  |  
 | 
      
         | 2415 |  |  | /* Go through all the RTL insn bodies and copy any invalid shared
 | 
      
         | 2416 |  |  |    structure.  This routine should only be called once.  */
 | 
      
         | 2417 |  |  |  
 | 
      
         | 2418 |  |  | static void
 | 
      
         | 2419 |  |  | unshare_all_rtl_1 (rtx insn)
 | 
      
         | 2420 |  |  | {
 | 
      
         | 2421 |  |  |   /* Unshare just about everything else.  */
 | 
      
         | 2422 |  |  |   unshare_all_rtl_in_chain (insn);
 | 
      
         | 2423 |  |  |  
 | 
      
         | 2424 |  |  |   /* Make sure the addresses of stack slots found outside the insn chain
 | 
      
         | 2425 |  |  |      (such as, in DECL_RTL of a variable) are not shared
 | 
      
         | 2426 |  |  |      with the insn chain.
 | 
      
         | 2427 |  |  |  
 | 
      
         | 2428 |  |  |      This special care is necessary when the stack slot MEM does not
 | 
      
         | 2429 |  |  |      actually appear in the insn chain.  If it does appear, its address
 | 
      
         | 2430 |  |  |      is unshared from all else at that point.  */
 | 
      
         | 2431 |  |  |   stack_slot_list = copy_rtx_if_shared (stack_slot_list);
 | 
      
         | 2432 |  |  | }
 | 
      
         | 2433 |  |  |  
 | 
      
         | 2434 |  |  | /* Go through all the RTL insn bodies and copy any invalid shared
 | 
      
         | 2435 |  |  |    structure, again.  This is a fairly expensive thing to do so it
 | 
      
         | 2436 |  |  |    should be done sparingly.  */
 | 
      
         | 2437 |  |  |  
 | 
      
         | 2438 |  |  | void
 | 
      
         | 2439 |  |  | unshare_all_rtl_again (rtx insn)
 | 
      
         | 2440 |  |  | {
 | 
      
         | 2441 |  |  |   rtx p;
 | 
      
         | 2442 |  |  |   tree decl;
 | 
      
         | 2443 |  |  |  
 | 
      
         | 2444 |  |  |   for (p = insn; p; p = NEXT_INSN (p))
 | 
      
         | 2445 |  |  |     if (INSN_P (p))
 | 
      
         | 2446 |  |  |       {
 | 
      
         | 2447 |  |  |         reset_used_flags (PATTERN (p));
 | 
      
         | 2448 |  |  |         reset_used_flags (REG_NOTES (p));
 | 
      
         | 2449 |  |  |         if (CALL_P (p))
 | 
      
         | 2450 |  |  |           reset_used_flags (CALL_INSN_FUNCTION_USAGE (p));
 | 
      
         | 2451 |  |  |       }
 | 
      
         | 2452 |  |  |  
 | 
      
         | 2453 |  |  |   /* Make sure that virtual stack slots are not shared.  */
 | 
      
         | 2454 |  |  |   set_used_decls (DECL_INITIAL (cfun->decl));
 | 
      
         | 2455 |  |  |  
 | 
      
         | 2456 |  |  |   /* Make sure that virtual parameters are not shared.  */
 | 
      
         | 2457 |  |  |   for (decl = DECL_ARGUMENTS (cfun->decl); decl; decl = DECL_CHAIN (decl))
 | 
      
         | 2458 |  |  |     set_used_flags (DECL_RTL (decl));
 | 
      
         | 2459 |  |  |  
 | 
      
         | 2460 |  |  |   reset_used_flags (stack_slot_list);
 | 
      
         | 2461 |  |  |  
 | 
      
         | 2462 |  |  |   unshare_all_rtl_1 (insn);
 | 
      
         | 2463 |  |  | }
 | 
      
         | 2464 |  |  |  
 | 
      
         | 2465 |  |  | unsigned int
 | 
      
         | 2466 |  |  | unshare_all_rtl (void)
 | 
      
         | 2467 |  |  | {
 | 
      
         | 2468 |  |  |   unshare_all_rtl_1 (get_insns ());
 | 
      
         | 2469 |  |  |   return 0;
 | 
      
         | 2470 |  |  | }
 | 
      
         | 2471 |  |  |  
 | 
      
         | 2472 |  |  | struct rtl_opt_pass pass_unshare_all_rtl =
 | 
      
         | 2473 |  |  | {
 | 
      
         | 2474 |  |  |  {
 | 
      
         | 2475 |  |  |   RTL_PASS,
 | 
      
         | 2476 |  |  |   "unshare",                            /* name */
 | 
      
         | 2477 |  |  |   NULL,                                 /* gate */
 | 
      
         | 2478 |  |  |   unshare_all_rtl,                      /* execute */
 | 
      
         | 2479 |  |  |   NULL,                                 /* sub */
 | 
      
         | 2480 |  |  |   NULL,                                 /* next */
 | 
      
         | 2481 |  |  |   0,                                    /* static_pass_number */
 | 
      
         | 2482 |  |  |   TV_NONE,                              /* tv_id */
 | 
      
         | 2483 |  |  |   0,                                    /* properties_required */
 | 
      
         | 2484 |  |  |   0,                                    /* properties_provided */
 | 
      
         | 2485 |  |  |   0,                                    /* properties_destroyed */
 | 
      
         | 2486 |  |  |   0,                                    /* todo_flags_start */
 | 
      
         | 2487 |  |  |   TODO_verify_rtl_sharing               /* todo_flags_finish */
 | 
      
         | 2488 |  |  |  }
 | 
      
         | 2489 |  |  | };
 | 
      
         | 2490 |  |  |  
 | 
      
         | 2491 |  |  |  
 | 
      
         | 2492 |  |  | /* Check that ORIG is not marked when it should not be and mark ORIG as in use,
 | 
      
         | 2493 |  |  |    Recursively does the same for subexpressions.  */
 | 
      
         | 2494 |  |  |  
 | 
      
         | 2495 |  |  | static void
 | 
      
         | 2496 |  |  | verify_rtx_sharing (rtx orig, rtx insn)
 | 
      
         | 2497 |  |  | {
 | 
      
         | 2498 |  |  |   rtx x = orig;
 | 
      
         | 2499 |  |  |   int i;
 | 
      
         | 2500 |  |  |   enum rtx_code code;
 | 
      
         | 2501 |  |  |   const char *format_ptr;
 | 
      
         | 2502 |  |  |  
 | 
      
         | 2503 |  |  |   if (x == 0)
 | 
      
         | 2504 |  |  |     return;
 | 
      
         | 2505 |  |  |  
 | 
      
         | 2506 |  |  |   code = GET_CODE (x);
 | 
      
         | 2507 |  |  |  
 | 
      
         | 2508 |  |  |   /* These types may be freely shared.  */
 | 
      
         | 2509 |  |  |  
 | 
      
         | 2510 |  |  |   switch (code)
 | 
      
         | 2511 |  |  |     {
 | 
      
         | 2512 |  |  |     case REG:
 | 
      
         | 2513 |  |  |     case DEBUG_EXPR:
 | 
      
         | 2514 |  |  |     case VALUE:
 | 
      
         | 2515 |  |  |     case CONST_INT:
 | 
      
         | 2516 |  |  |     case CONST_DOUBLE:
 | 
      
         | 2517 |  |  |     case CONST_FIXED:
 | 
      
         | 2518 |  |  |     case CONST_VECTOR:
 | 
      
         | 2519 |  |  |     case SYMBOL_REF:
 | 
      
         | 2520 |  |  |     case LABEL_REF:
 | 
      
         | 2521 |  |  |     case CODE_LABEL:
 | 
      
         | 2522 |  |  |     case PC:
 | 
      
         | 2523 |  |  |     case CC0:
 | 
      
         | 2524 |  |  |     case RETURN:
 | 
      
         | 2525 |  |  |     case SIMPLE_RETURN:
 | 
      
         | 2526 |  |  |     case SCRATCH:
 | 
      
         | 2527 |  |  |       return;
 | 
      
         | 2528 |  |  |       /* SCRATCH must be shared because they represent distinct values.  */
 | 
      
         | 2529 |  |  |     case CLOBBER:
 | 
      
         | 2530 |  |  |       if (REG_P (XEXP (x, 0)) && REGNO (XEXP (x, 0)) < FIRST_PSEUDO_REGISTER)
 | 
      
         | 2531 |  |  |         return;
 | 
      
         | 2532 |  |  |       break;
 | 
      
         | 2533 |  |  |  
 | 
      
         | 2534 |  |  |     case CONST:
 | 
      
         | 2535 |  |  |       if (shared_const_p (orig))
 | 
      
         | 2536 |  |  |         return;
 | 
      
         | 2537 |  |  |       break;
 | 
      
         | 2538 |  |  |  
 | 
      
         | 2539 |  |  |     case MEM:
 | 
      
         | 2540 |  |  |       /* A MEM is allowed to be shared if its address is constant.  */
 | 
      
         | 2541 |  |  |       if (CONSTANT_ADDRESS_P (XEXP (x, 0))
 | 
      
         | 2542 |  |  |           || reload_completed || reload_in_progress)
 | 
      
         | 2543 |  |  |         return;
 | 
      
         | 2544 |  |  |  
 | 
      
         | 2545 |  |  |       break;
 | 
      
         | 2546 |  |  |  
 | 
      
         | 2547 |  |  |     default:
 | 
      
         | 2548 |  |  |       break;
 | 
      
         | 2549 |  |  |     }
 | 
      
         | 2550 |  |  |  
 | 
      
         | 2551 |  |  |   /* This rtx may not be shared.  If it has already been seen,
 | 
      
         | 2552 |  |  |      replace it with a copy of itself.  */
 | 
      
         | 2553 |  |  | #ifdef ENABLE_CHECKING
 | 
      
         | 2554 |  |  |   if (RTX_FLAG (x, used))
 | 
      
         | 2555 |  |  |     {
 | 
      
         | 2556 |  |  |       error ("invalid rtl sharing found in the insn");
 | 
      
         | 2557 |  |  |       debug_rtx (insn);
 | 
      
         | 2558 |  |  |       error ("shared rtx");
 | 
      
         | 2559 |  |  |       debug_rtx (x);
 | 
      
         | 2560 |  |  |       internal_error ("internal consistency failure");
 | 
      
         | 2561 |  |  |     }
 | 
      
         | 2562 |  |  | #endif
 | 
      
         | 2563 |  |  |   gcc_assert (!RTX_FLAG (x, used));
 | 
      
         | 2564 |  |  |  
 | 
      
         | 2565 |  |  |   RTX_FLAG (x, used) = 1;
 | 
      
         | 2566 |  |  |  
 | 
      
         | 2567 |  |  |   /* Now scan the subexpressions recursively.  */
 | 
      
         | 2568 |  |  |  
 | 
      
         | 2569 |  |  |   format_ptr = GET_RTX_FORMAT (code);
 | 
      
         | 2570 |  |  |  
 | 
      
         | 2571 |  |  |   for (i = 0; i < GET_RTX_LENGTH (code); i++)
 | 
      
         | 2572 |  |  |     {
 | 
      
         | 2573 |  |  |       switch (*format_ptr++)
 | 
      
         | 2574 |  |  |         {
 | 
      
         | 2575 |  |  |         case 'e':
 | 
      
         | 2576 |  |  |           verify_rtx_sharing (XEXP (x, i), insn);
 | 
      
         | 2577 |  |  |           break;
 | 
      
         | 2578 |  |  |  
 | 
      
         | 2579 |  |  |         case 'E':
 | 
      
         | 2580 |  |  |           if (XVEC (x, i) != NULL)
 | 
      
         | 2581 |  |  |             {
 | 
      
         | 2582 |  |  |               int j;
 | 
      
         | 2583 |  |  |               int len = XVECLEN (x, i);
 | 
      
         | 2584 |  |  |  
 | 
      
         | 2585 |  |  |               for (j = 0; j < len; j++)
 | 
      
         | 2586 |  |  |                 {
 | 
      
         | 2587 |  |  |                   /* We allow sharing of ASM_OPERANDS inside single
 | 
      
         | 2588 |  |  |                      instruction.  */
 | 
      
         | 2589 |  |  |                   if (j && GET_CODE (XVECEXP (x, i, j)) == SET
 | 
      
         | 2590 |  |  |                       && (GET_CODE (SET_SRC (XVECEXP (x, i, j)))
 | 
      
         | 2591 |  |  |                           == ASM_OPERANDS))
 | 
      
         | 2592 |  |  |                     verify_rtx_sharing (SET_DEST (XVECEXP (x, i, j)), insn);
 | 
      
         | 2593 |  |  |                   else
 | 
      
         | 2594 |  |  |                     verify_rtx_sharing (XVECEXP (x, i, j), insn);
 | 
      
         | 2595 |  |  |                 }
 | 
      
         | 2596 |  |  |             }
 | 
      
         | 2597 |  |  |           break;
 | 
      
         | 2598 |  |  |         }
 | 
      
         | 2599 |  |  |     }
 | 
      
         | 2600 |  |  |   return;
 | 
      
         | 2601 |  |  | }
 | 
      
         | 2602 |  |  |  
 | 
      
         | 2603 |  |  | /* Go through all the RTL insn bodies and check that there is no unexpected
 | 
      
         | 2604 |  |  |    sharing in between the subexpressions.  */
 | 
      
         | 2605 |  |  |  
 | 
      
         | 2606 |  |  | DEBUG_FUNCTION void
 | 
      
         | 2607 |  |  | verify_rtl_sharing (void)
 | 
      
         | 2608 |  |  | {
 | 
      
         | 2609 |  |  |   rtx p;
 | 
      
         | 2610 |  |  |  
 | 
      
         | 2611 |  |  |   timevar_push (TV_VERIFY_RTL_SHARING);
 | 
      
         | 2612 |  |  |  
 | 
      
         | 2613 |  |  |   for (p = get_insns (); p; p = NEXT_INSN (p))
 | 
      
         | 2614 |  |  |     if (INSN_P (p))
 | 
      
         | 2615 |  |  |       {
 | 
      
         | 2616 |  |  |         reset_used_flags (PATTERN (p));
 | 
      
         | 2617 |  |  |         reset_used_flags (REG_NOTES (p));
 | 
      
         | 2618 |  |  |         if (CALL_P (p))
 | 
      
         | 2619 |  |  |           reset_used_flags (CALL_INSN_FUNCTION_USAGE (p));
 | 
      
         | 2620 |  |  |         if (GET_CODE (PATTERN (p)) == SEQUENCE)
 | 
      
         | 2621 |  |  |           {
 | 
      
         | 2622 |  |  |             int i;
 | 
      
         | 2623 |  |  |             rtx q, sequence = PATTERN (p);
 | 
      
         | 2624 |  |  |  
 | 
      
         | 2625 |  |  |             for (i = 0; i < XVECLEN (sequence, 0); i++)
 | 
      
         | 2626 |  |  |               {
 | 
      
         | 2627 |  |  |                 q = XVECEXP (sequence, 0, i);
 | 
      
         | 2628 |  |  |                 gcc_assert (INSN_P (q));
 | 
      
         | 2629 |  |  |                 reset_used_flags (PATTERN (q));
 | 
      
         | 2630 |  |  |                 reset_used_flags (REG_NOTES (q));
 | 
      
         | 2631 |  |  |                 if (CALL_P (q))
 | 
      
         | 2632 |  |  |                   reset_used_flags (CALL_INSN_FUNCTION_USAGE (q));
 | 
      
         | 2633 |  |  |               }
 | 
      
         | 2634 |  |  |           }
 | 
      
         | 2635 |  |  |       }
 | 
      
         | 2636 |  |  |  
 | 
      
         | 2637 |  |  |   for (p = get_insns (); p; p = NEXT_INSN (p))
 | 
      
         | 2638 |  |  |     if (INSN_P (p))
 | 
      
         | 2639 |  |  |       {
 | 
      
         | 2640 |  |  |         verify_rtx_sharing (PATTERN (p), p);
 | 
      
         | 2641 |  |  |         verify_rtx_sharing (REG_NOTES (p), p);
 | 
      
         | 2642 |  |  |         if (CALL_P (p))
 | 
      
         | 2643 |  |  |           verify_rtx_sharing (CALL_INSN_FUNCTION_USAGE (p), p);
 | 
      
         | 2644 |  |  |       }
 | 
      
         | 2645 |  |  |  
 | 
      
         | 2646 |  |  |   timevar_pop (TV_VERIFY_RTL_SHARING);
 | 
      
         | 2647 |  |  | }
 | 
      
         | 2648 |  |  |  
 | 
      
         | 2649 |  |  | /* Go through all the RTL insn bodies and copy any invalid shared structure.
 | 
      
         | 2650 |  |  |    Assumes the mark bits are cleared at entry.  */
 | 
      
         | 2651 |  |  |  
 | 
      
         | 2652 |  |  | void
 | 
      
         | 2653 |  |  | unshare_all_rtl_in_chain (rtx insn)
 | 
      
         | 2654 |  |  | {
 | 
      
         | 2655 |  |  |   for (; insn; insn = NEXT_INSN (insn))
 | 
      
         | 2656 |  |  |     if (INSN_P (insn))
 | 
      
         | 2657 |  |  |       {
 | 
      
         | 2658 |  |  |         PATTERN (insn) = copy_rtx_if_shared (PATTERN (insn));
 | 
      
         | 2659 |  |  |         REG_NOTES (insn) = copy_rtx_if_shared (REG_NOTES (insn));
 | 
      
         | 2660 |  |  |         if (CALL_P (insn))
 | 
      
         | 2661 |  |  |           CALL_INSN_FUNCTION_USAGE (insn)
 | 
      
         | 2662 |  |  |             = copy_rtx_if_shared (CALL_INSN_FUNCTION_USAGE (insn));
 | 
      
         | 2663 |  |  |       }
 | 
      
         | 2664 |  |  | }
 | 
      
         | 2665 |  |  |  
 | 
      
         | 2666 |  |  | /* Go through all virtual stack slots of a function and mark them as
 | 
      
         | 2667 |  |  |    shared.  We never replace the DECL_RTLs themselves with a copy,
 | 
      
         | 2668 |  |  |    but expressions mentioned into a DECL_RTL cannot be shared with
 | 
      
         | 2669 |  |  |    expressions in the instruction stream.
 | 
      
         | 2670 |  |  |  
 | 
      
         | 2671 |  |  |    Note that reload may convert pseudo registers into memories in-place.
 | 
      
         | 2672 |  |  |    Pseudo registers are always shared, but MEMs never are.  Thus if we
 | 
      
         | 2673 |  |  |    reset the used flags on MEMs in the instruction stream, we must set
 | 
      
         | 2674 |  |  |    them again on MEMs that appear in DECL_RTLs.  */
 | 
      
         | 2675 |  |  |  
 | 
      
         | 2676 |  |  | static void
 | 
      
         | 2677 |  |  | set_used_decls (tree blk)
 | 
      
         | 2678 |  |  | {
 | 
      
         | 2679 |  |  |   tree t;
 | 
      
         | 2680 |  |  |  
 | 
      
         | 2681 |  |  |   /* Mark decls.  */
 | 
      
         | 2682 |  |  |   for (t = BLOCK_VARS (blk); t; t = DECL_CHAIN (t))
 | 
      
         | 2683 |  |  |     if (DECL_RTL_SET_P (t))
 | 
      
         | 2684 |  |  |       set_used_flags (DECL_RTL (t));
 | 
      
         | 2685 |  |  |  
 | 
      
         | 2686 |  |  |   /* Now process sub-blocks.  */
 | 
      
         | 2687 |  |  |   for (t = BLOCK_SUBBLOCKS (blk); t; t = BLOCK_CHAIN (t))
 | 
      
         | 2688 |  |  |     set_used_decls (t);
 | 
      
         | 2689 |  |  | }
 | 
      
         | 2690 |  |  |  
 | 
      
         | 2691 |  |  | /* Mark ORIG as in use, and return a copy of it if it was already in use.
 | 
      
         | 2692 |  |  |    Recursively does the same for subexpressions.  Uses
 | 
      
         | 2693 |  |  |    copy_rtx_if_shared_1 to reduce stack space.  */
 | 
      
         | 2694 |  |  |  
 | 
      
         | 2695 |  |  | rtx
 | 
      
         | 2696 |  |  | copy_rtx_if_shared (rtx orig)
 | 
      
         | 2697 |  |  | {
 | 
      
         | 2698 |  |  |   copy_rtx_if_shared_1 (&orig);
 | 
      
         | 2699 |  |  |   return orig;
 | 
      
         | 2700 |  |  | }
 | 
      
         | 2701 |  |  |  
 | 
      
         | 2702 |  |  | /* Mark *ORIG1 as in use, and set it to a copy of it if it was already in
 | 
      
         | 2703 |  |  |    use.  Recursively does the same for subexpressions.  */
 | 
      
         | 2704 |  |  |  
 | 
      
         | 2705 |  |  | static void
 | 
      
         | 2706 |  |  | copy_rtx_if_shared_1 (rtx *orig1)
 | 
      
         | 2707 |  |  | {
 | 
      
         | 2708 |  |  |   rtx x;
 | 
      
         | 2709 |  |  |   int i;
 | 
      
         | 2710 |  |  |   enum rtx_code code;
 | 
      
         | 2711 |  |  |   rtx *last_ptr;
 | 
      
         | 2712 |  |  |   const char *format_ptr;
 | 
      
         | 2713 |  |  |   int copied = 0;
 | 
      
         | 2714 |  |  |   int length;
 | 
      
         | 2715 |  |  |  
 | 
      
         | 2716 |  |  |   /* Repeat is used to turn tail-recursion into iteration.  */
 | 
      
         | 2717 |  |  | repeat:
 | 
      
         | 2718 |  |  |   x = *orig1;
 | 
      
         | 2719 |  |  |  
 | 
      
         | 2720 |  |  |   if (x == 0)
 | 
      
         | 2721 |  |  |     return;
 | 
      
         | 2722 |  |  |  
 | 
      
         | 2723 |  |  |   code = GET_CODE (x);
 | 
      
         | 2724 |  |  |  
 | 
      
         | 2725 |  |  |   /* These types may be freely shared.  */
 | 
      
         | 2726 |  |  |  
 | 
      
         | 2727 |  |  |   switch (code)
 | 
      
         | 2728 |  |  |     {
 | 
      
         | 2729 |  |  |     case REG:
 | 
      
         | 2730 |  |  |     case DEBUG_EXPR:
 | 
      
         | 2731 |  |  |     case VALUE:
 | 
      
         | 2732 |  |  |     case CONST_INT:
 | 
      
         | 2733 |  |  |     case CONST_DOUBLE:
 | 
      
         | 2734 |  |  |     case CONST_FIXED:
 | 
      
         | 2735 |  |  |     case CONST_VECTOR:
 | 
      
         | 2736 |  |  |     case SYMBOL_REF:
 | 
      
         | 2737 |  |  |     case LABEL_REF:
 | 
      
         | 2738 |  |  |     case CODE_LABEL:
 | 
      
         | 2739 |  |  |     case PC:
 | 
      
         | 2740 |  |  |     case CC0:
 | 
      
         | 2741 |  |  |     case RETURN:
 | 
      
         | 2742 |  |  |     case SIMPLE_RETURN:
 | 
      
         | 2743 |  |  |     case SCRATCH:
 | 
      
         | 2744 |  |  |       /* SCRATCH must be shared because they represent distinct values.  */
 | 
      
         | 2745 |  |  |       return;
 | 
      
         | 2746 |  |  |     case CLOBBER:
 | 
      
         | 2747 |  |  |       if (REG_P (XEXP (x, 0)) && REGNO (XEXP (x, 0)) < FIRST_PSEUDO_REGISTER)
 | 
      
         | 2748 |  |  |         return;
 | 
      
         | 2749 |  |  |       break;
 | 
      
         | 2750 |  |  |  
 | 
      
         | 2751 |  |  |     case CONST:
 | 
      
         | 2752 |  |  |       if (shared_const_p (x))
 | 
      
         | 2753 |  |  |         return;
 | 
      
         | 2754 |  |  |       break;
 | 
      
         | 2755 |  |  |  
 | 
      
         | 2756 |  |  |     case DEBUG_INSN:
 | 
      
         | 2757 |  |  |     case INSN:
 | 
      
         | 2758 |  |  |     case JUMP_INSN:
 | 
      
         | 2759 |  |  |     case CALL_INSN:
 | 
      
         | 2760 |  |  |     case NOTE:
 | 
      
         | 2761 |  |  |     case BARRIER:
 | 
      
         | 2762 |  |  |       /* The chain of insns is not being copied.  */
 | 
      
         | 2763 |  |  |       return;
 | 
      
         | 2764 |  |  |  
 | 
      
         | 2765 |  |  |     default:
 | 
      
         | 2766 |  |  |       break;
 | 
      
         | 2767 |  |  |     }
 | 
      
         | 2768 |  |  |  
 | 
      
         | 2769 |  |  |   /* This rtx may not be shared.  If it has already been seen,
 | 
      
         | 2770 |  |  |      replace it with a copy of itself.  */
 | 
      
         | 2771 |  |  |  
 | 
      
         | 2772 |  |  |   if (RTX_FLAG (x, used))
 | 
      
         | 2773 |  |  |     {
 | 
      
         | 2774 |  |  |       x = shallow_copy_rtx (x);
 | 
      
         | 2775 |  |  |       copied = 1;
 | 
      
         | 2776 |  |  |     }
 | 
      
         | 2777 |  |  |   RTX_FLAG (x, used) = 1;
 | 
      
         | 2778 |  |  |  
 | 
      
         | 2779 |  |  |   /* Now scan the subexpressions recursively.
 | 
      
         | 2780 |  |  |      We can store any replaced subexpressions directly into X
 | 
      
         | 2781 |  |  |      since we know X is not shared!  Any vectors in X
 | 
      
         | 2782 |  |  |      must be copied if X was copied.  */
 | 
      
         | 2783 |  |  |  
 | 
      
         | 2784 |  |  |   format_ptr = GET_RTX_FORMAT (code);
 | 
      
         | 2785 |  |  |   length = GET_RTX_LENGTH (code);
 | 
      
         | 2786 |  |  |   last_ptr = NULL;
 | 
      
         | 2787 |  |  |  
 | 
      
         | 2788 |  |  |   for (i = 0; i < length; i++)
 | 
      
         | 2789 |  |  |     {
 | 
      
         | 2790 |  |  |       switch (*format_ptr++)
 | 
      
         | 2791 |  |  |         {
 | 
      
         | 2792 |  |  |         case 'e':
 | 
      
         | 2793 |  |  |           if (last_ptr)
 | 
      
         | 2794 |  |  |             copy_rtx_if_shared_1 (last_ptr);
 | 
      
         | 2795 |  |  |           last_ptr = &XEXP (x, i);
 | 
      
         | 2796 |  |  |           break;
 | 
      
         | 2797 |  |  |  
 | 
      
         | 2798 |  |  |         case 'E':
 | 
      
         | 2799 |  |  |           if (XVEC (x, i) != NULL)
 | 
      
         | 2800 |  |  |             {
 | 
      
         | 2801 |  |  |               int j;
 | 
      
         | 2802 |  |  |               int len = XVECLEN (x, i);
 | 
      
         | 2803 |  |  |  
 | 
      
         | 2804 |  |  |               /* Copy the vector iff I copied the rtx and the length
 | 
      
         | 2805 |  |  |                  is nonzero.  */
 | 
      
         | 2806 |  |  |               if (copied && len > 0)
 | 
      
         | 2807 |  |  |                 XVEC (x, i) = gen_rtvec_v (len, XVEC (x, i)->elem);
 | 
      
         | 2808 |  |  |  
 | 
      
         | 2809 |  |  |               /* Call recursively on all inside the vector.  */
 | 
      
         | 2810 |  |  |               for (j = 0; j < len; j++)
 | 
      
         | 2811 |  |  |                 {
 | 
      
         | 2812 |  |  |                   if (last_ptr)
 | 
      
         | 2813 |  |  |                     copy_rtx_if_shared_1 (last_ptr);
 | 
      
         | 2814 |  |  |                   last_ptr = &XVECEXP (x, i, j);
 | 
      
         | 2815 |  |  |                 }
 | 
      
         | 2816 |  |  |             }
 | 
      
         | 2817 |  |  |           break;
 | 
      
         | 2818 |  |  |         }
 | 
      
         | 2819 |  |  |     }
 | 
      
         | 2820 |  |  |   *orig1 = x;
 | 
      
         | 2821 |  |  |   if (last_ptr)
 | 
      
         | 2822 |  |  |     {
 | 
      
         | 2823 |  |  |       orig1 = last_ptr;
 | 
      
         | 2824 |  |  |       goto repeat;
 | 
      
         | 2825 |  |  |     }
 | 
      
         | 2826 |  |  |   return;
 | 
      
         | 2827 |  |  | }
 | 
      
         | 2828 |  |  |  
 | 
      
         | 2829 |  |  | /* Set the USED bit in X and its non-shareable subparts to FLAG.  */
 | 
      
         | 2830 |  |  |  
 | 
      
         | 2831 |  |  | static void
 | 
      
         | 2832 |  |  | mark_used_flags (rtx x, int flag)
 | 
      
         | 2833 |  |  | {
 | 
      
         | 2834 |  |  |   int i, j;
 | 
      
         | 2835 |  |  |   enum rtx_code code;
 | 
      
         | 2836 |  |  |   const char *format_ptr;
 | 
      
         | 2837 |  |  |   int length;
 | 
      
         | 2838 |  |  |  
 | 
      
         | 2839 |  |  |   /* Repeat is used to turn tail-recursion into iteration.  */
 | 
      
         | 2840 |  |  | repeat:
 | 
      
         | 2841 |  |  |   if (x == 0)
 | 
      
         | 2842 |  |  |     return;
 | 
      
         | 2843 |  |  |  
 | 
      
         | 2844 |  |  |   code = GET_CODE (x);
 | 
      
         | 2845 |  |  |  
 | 
      
         | 2846 |  |  |   /* These types may be freely shared so we needn't do any resetting
 | 
      
         | 2847 |  |  |      for them.  */
 | 
      
         | 2848 |  |  |  
 | 
      
         | 2849 |  |  |   switch (code)
 | 
      
         | 2850 |  |  |     {
 | 
      
         | 2851 |  |  |     case REG:
 | 
      
         | 2852 |  |  |     case DEBUG_EXPR:
 | 
      
         | 2853 |  |  |     case VALUE:
 | 
      
         | 2854 |  |  |     case CONST_INT:
 | 
      
         | 2855 |  |  |     case CONST_DOUBLE:
 | 
      
         | 2856 |  |  |     case CONST_FIXED:
 | 
      
         | 2857 |  |  |     case CONST_VECTOR:
 | 
      
         | 2858 |  |  |     case SYMBOL_REF:
 | 
      
         | 2859 |  |  |     case CODE_LABEL:
 | 
      
         | 2860 |  |  |     case PC:
 | 
      
         | 2861 |  |  |     case CC0:
 | 
      
         | 2862 |  |  |     case RETURN:
 | 
      
         | 2863 |  |  |     case SIMPLE_RETURN:
 | 
      
         | 2864 |  |  |       return;
 | 
      
         | 2865 |  |  |  
 | 
      
         | 2866 |  |  |     case DEBUG_INSN:
 | 
      
         | 2867 |  |  |     case INSN:
 | 
      
         | 2868 |  |  |     case JUMP_INSN:
 | 
      
         | 2869 |  |  |     case CALL_INSN:
 | 
      
         | 2870 |  |  |     case NOTE:
 | 
      
         | 2871 |  |  |     case LABEL_REF:
 | 
      
         | 2872 |  |  |     case BARRIER:
 | 
      
         | 2873 |  |  |       /* The chain of insns is not being copied.  */
 | 
      
         | 2874 |  |  |       return;
 | 
      
         | 2875 |  |  |  
 | 
      
         | 2876 |  |  |     default:
 | 
      
         | 2877 |  |  |       break;
 | 
      
         | 2878 |  |  |     }
 | 
      
         | 2879 |  |  |  
 | 
      
         | 2880 |  |  |   RTX_FLAG (x, used) = flag;
 | 
      
         | 2881 |  |  |  
 | 
      
         | 2882 |  |  |   format_ptr = GET_RTX_FORMAT (code);
 | 
      
         | 2883 |  |  |   length = GET_RTX_LENGTH (code);
 | 
      
         | 2884 |  |  |  
 | 
      
         | 2885 |  |  |   for (i = 0; i < length; i++)
 | 
      
         | 2886 |  |  |     {
 | 
      
         | 2887 |  |  |       switch (*format_ptr++)
 | 
      
         | 2888 |  |  |         {
 | 
      
         | 2889 |  |  |         case 'e':
 | 
      
         | 2890 |  |  |           if (i == length-1)
 | 
      
         | 2891 |  |  |             {
 | 
      
         | 2892 |  |  |               x = XEXP (x, i);
 | 
      
         | 2893 |  |  |               goto repeat;
 | 
      
         | 2894 |  |  |             }
 | 
      
         | 2895 |  |  |           mark_used_flags (XEXP (x, i), flag);
 | 
      
         | 2896 |  |  |           break;
 | 
      
         | 2897 |  |  |  
 | 
      
         | 2898 |  |  |         case 'E':
 | 
      
         | 2899 |  |  |           for (j = 0; j < XVECLEN (x, i); j++)
 | 
      
         | 2900 |  |  |             mark_used_flags (XVECEXP (x, i, j), flag);
 | 
      
         | 2901 |  |  |           break;
 | 
      
         | 2902 |  |  |         }
 | 
      
         | 2903 |  |  |     }
 | 
      
         | 2904 |  |  | }
 | 
      
         | 2905 |  |  |  
 | 
      
         | 2906 |  |  | /* Clear all the USED bits in X to allow copy_rtx_if_shared to be used
 | 
      
         | 2907 |  |  |    to look for shared sub-parts.  */
 | 
      
         | 2908 |  |  |  
 | 
      
         | 2909 |  |  | void
 | 
      
         | 2910 |  |  | reset_used_flags (rtx x)
 | 
      
         | 2911 |  |  | {
 | 
      
         | 2912 |  |  |   mark_used_flags (x, 0);
 | 
      
         | 2913 |  |  | }
 | 
      
         | 2914 |  |  |  
 | 
      
         | 2915 |  |  | /* Set all the USED bits in X to allow copy_rtx_if_shared to be used
 | 
      
         | 2916 |  |  |    to look for shared sub-parts.  */
 | 
      
         | 2917 |  |  |  
 | 
      
         | 2918 |  |  | void
 | 
      
         | 2919 |  |  | set_used_flags (rtx x)
 | 
      
         | 2920 |  |  | {
 | 
      
         | 2921 |  |  |   mark_used_flags (x, 1);
 | 
      
         | 2922 |  |  | }
 | 
      
         | 2923 |  |  |  
 | 
      
         | 2924 |  |  | /* Copy X if necessary so that it won't be altered by changes in OTHER.
 | 
      
         | 2925 |  |  |    Return X or the rtx for the pseudo reg the value of X was copied into.
 | 
      
         | 2926 |  |  |    OTHER must be valid as a SET_DEST.  */
 | 
      
         | 2927 |  |  |  
 | 
      
         | 2928 |  |  | rtx
 | 
      
         | 2929 |  |  | make_safe_from (rtx x, rtx other)
 | 
      
         | 2930 |  |  | {
 | 
      
         | 2931 |  |  |   while (1)
 | 
      
         | 2932 |  |  |     switch (GET_CODE (other))
 | 
      
         | 2933 |  |  |       {
 | 
      
         | 2934 |  |  |       case SUBREG:
 | 
      
         | 2935 |  |  |         other = SUBREG_REG (other);
 | 
      
         | 2936 |  |  |         break;
 | 
      
         | 2937 |  |  |       case STRICT_LOW_PART:
 | 
      
         | 2938 |  |  |       case SIGN_EXTEND:
 | 
      
         | 2939 |  |  |       case ZERO_EXTEND:
 | 
      
         | 2940 |  |  |         other = XEXP (other, 0);
 | 
      
         | 2941 |  |  |         break;
 | 
      
         | 2942 |  |  |       default:
 | 
      
         | 2943 |  |  |         goto done;
 | 
      
         | 2944 |  |  |       }
 | 
      
         | 2945 |  |  |  done:
 | 
      
         | 2946 |  |  |   if ((MEM_P (other)
 | 
      
         | 2947 |  |  |        && ! CONSTANT_P (x)
 | 
      
         | 2948 |  |  |        && !REG_P (x)
 | 
      
         | 2949 |  |  |        && GET_CODE (x) != SUBREG)
 | 
      
         | 2950 |  |  |       || (REG_P (other)
 | 
      
         | 2951 |  |  |           && (REGNO (other) < FIRST_PSEUDO_REGISTER
 | 
      
         | 2952 |  |  |               || reg_mentioned_p (other, x))))
 | 
      
         | 2953 |  |  |     {
 | 
      
         | 2954 |  |  |       rtx temp = gen_reg_rtx (GET_MODE (x));
 | 
      
         | 2955 |  |  |       emit_move_insn (temp, x);
 | 
      
         | 2956 |  |  |       return temp;
 | 
      
         | 2957 |  |  |     }
 | 
      
         | 2958 |  |  |   return x;
 | 
      
         | 2959 |  |  | }
 | 
      
         | 2960 |  |  |  
 | 
      
         | 2961 |  |  | /* Emission of insns (adding them to the doubly-linked list).  */
 | 
      
         | 2962 |  |  |  
 | 
      
         | 2963 |  |  | /* Return the last insn emitted, even if it is in a sequence now pushed.  */
 | 
      
         | 2964 |  |  |  
 | 
      
         | 2965 |  |  | rtx
 | 
      
         | 2966 |  |  | get_last_insn_anywhere (void)
 | 
      
         | 2967 |  |  | {
 | 
      
         | 2968 |  |  |   struct sequence_stack *stack;
 | 
      
         | 2969 |  |  |   if (get_last_insn ())
 | 
      
         | 2970 |  |  |     return get_last_insn ();
 | 
      
         | 2971 |  |  |   for (stack = seq_stack; stack; stack = stack->next)
 | 
      
         | 2972 |  |  |     if (stack->last != 0)
 | 
      
         | 2973 |  |  |       return stack->last;
 | 
      
         | 2974 |  |  |   return 0;
 | 
      
         | 2975 |  |  | }
 | 
      
         | 2976 |  |  |  
 | 
      
         | 2977 |  |  | /* Return the first nonnote insn emitted in current sequence or current
 | 
      
         | 2978 |  |  |    function.  This routine looks inside SEQUENCEs.  */
 | 
      
         | 2979 |  |  |  
 | 
      
         | 2980 |  |  | rtx
 | 
      
         | 2981 |  |  | get_first_nonnote_insn (void)
 | 
      
         | 2982 |  |  | {
 | 
      
         | 2983 |  |  |   rtx insn = get_insns ();
 | 
      
         | 2984 |  |  |  
 | 
      
         | 2985 |  |  |   if (insn)
 | 
      
         | 2986 |  |  |     {
 | 
      
         | 2987 |  |  |       if (NOTE_P (insn))
 | 
      
         | 2988 |  |  |         for (insn = next_insn (insn);
 | 
      
         | 2989 |  |  |              insn && NOTE_P (insn);
 | 
      
         | 2990 |  |  |              insn = next_insn (insn))
 | 
      
         | 2991 |  |  |           continue;
 | 
      
         | 2992 |  |  |       else
 | 
      
         | 2993 |  |  |         {
 | 
      
         | 2994 |  |  |           if (NONJUMP_INSN_P (insn)
 | 
      
         | 2995 |  |  |               && GET_CODE (PATTERN (insn)) == SEQUENCE)
 | 
      
         | 2996 |  |  |             insn = XVECEXP (PATTERN (insn), 0, 0);
 | 
      
         | 2997 |  |  |         }
 | 
      
         | 2998 |  |  |     }
 | 
      
         | 2999 |  |  |  
 | 
      
         | 3000 |  |  |   return insn;
 | 
      
         | 3001 |  |  | }
 | 
      
         | 3002 |  |  |  
 | 
      
         | 3003 |  |  | /* Return the last nonnote insn emitted in current sequence or current
 | 
      
         | 3004 |  |  |    function.  This routine looks inside SEQUENCEs.  */
 | 
      
         | 3005 |  |  |  
 | 
      
         | 3006 |  |  | rtx
 | 
      
         | 3007 |  |  | get_last_nonnote_insn (void)
 | 
      
         | 3008 |  |  | {
 | 
      
         | 3009 |  |  |   rtx insn = get_last_insn ();
 | 
      
         | 3010 |  |  |  
 | 
      
         | 3011 |  |  |   if (insn)
 | 
      
         | 3012 |  |  |     {
 | 
      
         | 3013 |  |  |       if (NOTE_P (insn))
 | 
      
         | 3014 |  |  |         for (insn = previous_insn (insn);
 | 
      
         | 3015 |  |  |              insn && NOTE_P (insn);
 | 
      
         | 3016 |  |  |              insn = previous_insn (insn))
 | 
      
         | 3017 |  |  |           continue;
 | 
      
         | 3018 |  |  |       else
 | 
      
         | 3019 |  |  |         {
 | 
      
         | 3020 |  |  |           if (NONJUMP_INSN_P (insn)
 | 
      
         | 3021 |  |  |               && GET_CODE (PATTERN (insn)) == SEQUENCE)
 | 
      
         | 3022 |  |  |             insn = XVECEXP (PATTERN (insn), 0,
 | 
      
         | 3023 |  |  |                             XVECLEN (PATTERN (insn), 0) - 1);
 | 
      
         | 3024 |  |  |         }
 | 
      
         | 3025 |  |  |     }
 | 
      
         | 3026 |  |  |  
 | 
      
         | 3027 |  |  |   return insn;
 | 
      
         | 3028 |  |  | }
 | 
      
         | 3029 |  |  |  
 | 
      
         | 3030 |  |  | /* Return the number of actual (non-debug) insns emitted in this
 | 
      
         | 3031 |  |  |    function.  */
 | 
      
         | 3032 |  |  |  
 | 
      
         | 3033 |  |  | int
 | 
      
         | 3034 |  |  | get_max_insn_count (void)
 | 
      
         | 3035 |  |  | {
 | 
      
         | 3036 |  |  |   int n = cur_insn_uid;
 | 
      
         | 3037 |  |  |  
 | 
      
         | 3038 |  |  |   /* The table size must be stable across -g, to avoid codegen
 | 
      
         | 3039 |  |  |      differences due to debug insns, and not be affected by
 | 
      
         | 3040 |  |  |      -fmin-insn-uid, to avoid excessive table size and to simplify
 | 
      
         | 3041 |  |  |      debugging of -fcompare-debug failures.  */
 | 
      
         | 3042 |  |  |   if (cur_debug_insn_uid > MIN_NONDEBUG_INSN_UID)
 | 
      
         | 3043 |  |  |     n -= cur_debug_insn_uid;
 | 
      
         | 3044 |  |  |   else
 | 
      
         | 3045 |  |  |     n -= MIN_NONDEBUG_INSN_UID;
 | 
      
         | 3046 |  |  |  
 | 
      
         | 3047 |  |  |   return n;
 | 
      
         | 3048 |  |  | }
 | 
      
         | 3049 |  |  |  
 | 
      
         | 3050 |  |  |  
 | 
      
         | 3051 |  |  | /* Return the next insn.  If it is a SEQUENCE, return the first insn
 | 
      
         | 3052 |  |  |    of the sequence.  */
 | 
      
         | 3053 |  |  |  
 | 
      
         | 3054 |  |  | rtx
 | 
      
         | 3055 |  |  | next_insn (rtx insn)
 | 
      
         | 3056 |  |  | {
 | 
      
         | 3057 |  |  |   if (insn)
 | 
      
         | 3058 |  |  |     {
 | 
      
         | 3059 |  |  |       insn = NEXT_INSN (insn);
 | 
      
         | 3060 |  |  |       if (insn && NONJUMP_INSN_P (insn)
 | 
      
         | 3061 |  |  |           && GET_CODE (PATTERN (insn)) == SEQUENCE)
 | 
      
         | 3062 |  |  |         insn = XVECEXP (PATTERN (insn), 0, 0);
 | 
      
         | 3063 |  |  |     }
 | 
      
         | 3064 |  |  |  
 | 
      
         | 3065 |  |  |   return insn;
 | 
      
         | 3066 |  |  | }
 | 
      
         | 3067 |  |  |  
 | 
      
         | 3068 |  |  | /* Return the previous insn.  If it is a SEQUENCE, return the last insn
 | 
      
         | 3069 |  |  |    of the sequence.  */
 | 
      
         | 3070 |  |  |  
 | 
      
         | 3071 |  |  | rtx
 | 
      
         | 3072 |  |  | previous_insn (rtx insn)
 | 
      
         | 3073 |  |  | {
 | 
      
         | 3074 |  |  |   if (insn)
 | 
      
         | 3075 |  |  |     {
 | 
      
         | 3076 |  |  |       insn = PREV_INSN (insn);
 | 
      
         | 3077 |  |  |       if (insn && NONJUMP_INSN_P (insn)
 | 
      
         | 3078 |  |  |           && GET_CODE (PATTERN (insn)) == SEQUENCE)
 | 
      
         | 3079 |  |  |         insn = XVECEXP (PATTERN (insn), 0, XVECLEN (PATTERN (insn), 0) - 1);
 | 
      
         | 3080 |  |  |     }
 | 
      
         | 3081 |  |  |  
 | 
      
         | 3082 |  |  |   return insn;
 | 
      
         | 3083 |  |  | }
 | 
      
         | 3084 |  |  |  
 | 
      
         | 3085 |  |  | /* Return the next insn after INSN that is not a NOTE.  This routine does not
 | 
      
         | 3086 |  |  |    look inside SEQUENCEs.  */
 | 
      
         | 3087 |  |  |  
 | 
      
         | 3088 |  |  | rtx
 | 
      
         | 3089 |  |  | next_nonnote_insn (rtx insn)
 | 
      
         | 3090 |  |  | {
 | 
      
         | 3091 |  |  |   while (insn)
 | 
      
         | 3092 |  |  |     {
 | 
      
         | 3093 |  |  |       insn = NEXT_INSN (insn);
 | 
      
         | 3094 |  |  |       if (insn == 0 || !NOTE_P (insn))
 | 
      
         | 3095 |  |  |         break;
 | 
      
         | 3096 |  |  |     }
 | 
      
         | 3097 |  |  |  
 | 
      
         | 3098 |  |  |   return insn;
 | 
      
         | 3099 |  |  | }
 | 
      
         | 3100 |  |  |  
 | 
      
         | 3101 |  |  | /* Return the next insn after INSN that is not a NOTE, but stop the
 | 
      
         | 3102 |  |  |    search before we enter another basic block.  This routine does not
 | 
      
         | 3103 |  |  |    look inside SEQUENCEs.  */
 | 
      
         | 3104 |  |  |  
 | 
      
         | 3105 |  |  | rtx
 | 
      
         | 3106 |  |  | next_nonnote_insn_bb (rtx insn)
 | 
      
         | 3107 |  |  | {
 | 
      
         | 3108 |  |  |   while (insn)
 | 
      
         | 3109 |  |  |     {
 | 
      
         | 3110 |  |  |       insn = NEXT_INSN (insn);
 | 
      
         | 3111 |  |  |       if (insn == 0 || !NOTE_P (insn))
 | 
      
         | 3112 |  |  |         break;
 | 
      
         | 3113 |  |  |       if (NOTE_INSN_BASIC_BLOCK_P (insn))
 | 
      
         | 3114 |  |  |         return NULL_RTX;
 | 
      
         | 3115 |  |  |     }
 | 
      
         | 3116 |  |  |  
 | 
      
         | 3117 |  |  |   return insn;
 | 
      
         | 3118 |  |  | }
 | 
      
         | 3119 |  |  |  
 | 
      
         | 3120 |  |  | /* Return the previous insn before INSN that is not a NOTE.  This routine does
 | 
      
         | 3121 |  |  |    not look inside SEQUENCEs.  */
 | 
      
         | 3122 |  |  |  
 | 
      
         | 3123 |  |  | rtx
 | 
      
         | 3124 |  |  | prev_nonnote_insn (rtx insn)
 | 
      
         | 3125 |  |  | {
 | 
      
         | 3126 |  |  |   while (insn)
 | 
      
         | 3127 |  |  |     {
 | 
      
         | 3128 |  |  |       insn = PREV_INSN (insn);
 | 
      
         | 3129 |  |  |       if (insn == 0 || !NOTE_P (insn))
 | 
      
         | 3130 |  |  |         break;
 | 
      
         | 3131 |  |  |     }
 | 
      
         | 3132 |  |  |  
 | 
      
         | 3133 |  |  |   return insn;
 | 
      
         | 3134 |  |  | }
 | 
      
         | 3135 |  |  |  
 | 
      
         | 3136 |  |  | /* Return the previous insn before INSN that is not a NOTE, but stop
 | 
      
         | 3137 |  |  |    the search before we enter another basic block.  This routine does
 | 
      
         | 3138 |  |  |    not look inside SEQUENCEs.  */
 | 
      
         | 3139 |  |  |  
 | 
      
         | 3140 |  |  | rtx
 | 
      
         | 3141 |  |  | prev_nonnote_insn_bb (rtx insn)
 | 
      
         | 3142 |  |  | {
 | 
      
         | 3143 |  |  |   while (insn)
 | 
      
         | 3144 |  |  |     {
 | 
      
         | 3145 |  |  |       insn = PREV_INSN (insn);
 | 
      
         | 3146 |  |  |       if (insn == 0 || !NOTE_P (insn))
 | 
      
         | 3147 |  |  |         break;
 | 
      
         | 3148 |  |  |       if (NOTE_INSN_BASIC_BLOCK_P (insn))
 | 
      
         | 3149 |  |  |         return NULL_RTX;
 | 
      
         | 3150 |  |  |     }
 | 
      
         | 3151 |  |  |  
 | 
      
         | 3152 |  |  |   return insn;
 | 
      
         | 3153 |  |  | }
 | 
      
         | 3154 |  |  |  
 | 
      
         | 3155 |  |  | /* Return the next insn after INSN that is not a DEBUG_INSN.  This
 | 
      
         | 3156 |  |  |    routine does not look inside SEQUENCEs.  */
 | 
      
         | 3157 |  |  |  
 | 
      
         | 3158 |  |  | rtx
 | 
      
         | 3159 |  |  | next_nondebug_insn (rtx insn)
 | 
      
         | 3160 |  |  | {
 | 
      
         | 3161 |  |  |   while (insn)
 | 
      
         | 3162 |  |  |     {
 | 
      
         | 3163 |  |  |       insn = NEXT_INSN (insn);
 | 
      
         | 3164 |  |  |       if (insn == 0 || !DEBUG_INSN_P (insn))
 | 
      
         | 3165 |  |  |         break;
 | 
      
         | 3166 |  |  |     }
 | 
      
         | 3167 |  |  |  
 | 
      
         | 3168 |  |  |   return insn;
 | 
      
         | 3169 |  |  | }
 | 
      
         | 3170 |  |  |  
 | 
      
         | 3171 |  |  | /* Return the previous insn before INSN that is not a DEBUG_INSN.
 | 
      
         | 3172 |  |  |    This routine does not look inside SEQUENCEs.  */
 | 
      
         | 3173 |  |  |  
 | 
      
         | 3174 |  |  | rtx
 | 
      
         | 3175 |  |  | prev_nondebug_insn (rtx insn)
 | 
      
         | 3176 |  |  | {
 | 
      
         | 3177 |  |  |   while (insn)
 | 
      
         | 3178 |  |  |     {
 | 
      
         | 3179 |  |  |       insn = PREV_INSN (insn);
 | 
      
         | 3180 |  |  |       if (insn == 0 || !DEBUG_INSN_P (insn))
 | 
      
         | 3181 |  |  |         break;
 | 
      
         | 3182 |  |  |     }
 | 
      
         | 3183 |  |  |  
 | 
      
         | 3184 |  |  |   return insn;
 | 
      
         | 3185 |  |  | }
 | 
      
         | 3186 |  |  |  
 | 
      
         | 3187 |  |  | /* Return the next insn after INSN that is not a NOTE nor DEBUG_INSN.
 | 
      
         | 3188 |  |  |    This routine does not look inside SEQUENCEs.  */
 | 
      
         | 3189 |  |  |  
 | 
      
         | 3190 |  |  | rtx
 | 
      
         | 3191 |  |  | next_nonnote_nondebug_insn (rtx insn)
 | 
      
         | 3192 |  |  | {
 | 
      
         | 3193 |  |  |   while (insn)
 | 
      
         | 3194 |  |  |     {
 | 
      
         | 3195 |  |  |       insn = NEXT_INSN (insn);
 | 
      
         | 3196 |  |  |       if (insn == 0 || (!NOTE_P (insn) && !DEBUG_INSN_P (insn)))
 | 
      
         | 3197 |  |  |         break;
 | 
      
         | 3198 |  |  |     }
 | 
      
         | 3199 |  |  |  
 | 
      
         | 3200 |  |  |   return insn;
 | 
      
         | 3201 |  |  | }
 | 
      
         | 3202 |  |  |  
 | 
      
         | 3203 |  |  | /* Return the previous insn before INSN that is not a NOTE nor DEBUG_INSN.
 | 
      
         | 3204 |  |  |    This routine does not look inside SEQUENCEs.  */
 | 
      
         | 3205 |  |  |  
 | 
      
         | 3206 |  |  | rtx
 | 
      
         | 3207 |  |  | prev_nonnote_nondebug_insn (rtx insn)
 | 
      
         | 3208 |  |  | {
 | 
      
         | 3209 |  |  |   while (insn)
 | 
      
         | 3210 |  |  |     {
 | 
      
         | 3211 |  |  |       insn = PREV_INSN (insn);
 | 
      
         | 3212 |  |  |       if (insn == 0 || (!NOTE_P (insn) && !DEBUG_INSN_P (insn)))
 | 
      
         | 3213 |  |  |         break;
 | 
      
         | 3214 |  |  |     }
 | 
      
         | 3215 |  |  |  
 | 
      
         | 3216 |  |  |   return insn;
 | 
      
         | 3217 |  |  | }
 | 
      
         | 3218 |  |  |  
 | 
      
         | 3219 |  |  | /* Return the next INSN, CALL_INSN or JUMP_INSN after INSN;
 | 
      
         | 3220 |  |  |    or 0, if there is none.  This routine does not look inside
 | 
      
         | 3221 |  |  |    SEQUENCEs.  */
 | 
      
         | 3222 |  |  |  
 | 
      
         | 3223 |  |  | rtx
 | 
      
         | 3224 |  |  | next_real_insn (rtx insn)
 | 
      
         | 3225 |  |  | {
 | 
      
         | 3226 |  |  |   while (insn)
 | 
      
         | 3227 |  |  |     {
 | 
      
         | 3228 |  |  |       insn = NEXT_INSN (insn);
 | 
      
         | 3229 |  |  |       if (insn == 0 || INSN_P (insn))
 | 
      
         | 3230 |  |  |         break;
 | 
      
         | 3231 |  |  |     }
 | 
      
         | 3232 |  |  |  
 | 
      
         | 3233 |  |  |   return insn;
 | 
      
         | 3234 |  |  | }
 | 
      
         | 3235 |  |  |  
 | 
      
         | 3236 |  |  | /* Return the last INSN, CALL_INSN or JUMP_INSN before INSN;
 | 
      
         | 3237 |  |  |    or 0, if there is none.  This routine does not look inside
 | 
      
         | 3238 |  |  |    SEQUENCEs.  */
 | 
      
         | 3239 |  |  |  
 | 
      
         | 3240 |  |  | rtx
 | 
      
         | 3241 |  |  | prev_real_insn (rtx insn)
 | 
      
         | 3242 |  |  | {
 | 
      
         | 3243 |  |  |   while (insn)
 | 
      
         | 3244 |  |  |     {
 | 
      
         | 3245 |  |  |       insn = PREV_INSN (insn);
 | 
      
         | 3246 |  |  |       if (insn == 0 || INSN_P (insn))
 | 
      
         | 3247 |  |  |         break;
 | 
      
         | 3248 |  |  |     }
 | 
      
         | 3249 |  |  |  
 | 
      
         | 3250 |  |  |   return insn;
 | 
      
         | 3251 |  |  | }
 | 
      
         | 3252 |  |  |  
 | 
      
         | 3253 |  |  | /* Return the last CALL_INSN in the current list, or 0 if there is none.
 | 
      
         | 3254 |  |  |    This routine does not look inside SEQUENCEs.  */
 | 
      
         | 3255 |  |  |  
 | 
      
         | 3256 |  |  | rtx
 | 
      
         | 3257 |  |  | last_call_insn (void)
 | 
      
         | 3258 |  |  | {
 | 
      
         | 3259 |  |  |   rtx insn;
 | 
      
         | 3260 |  |  |  
 | 
      
         | 3261 |  |  |   for (insn = get_last_insn ();
 | 
      
         | 3262 |  |  |        insn && !CALL_P (insn);
 | 
      
         | 3263 |  |  |        insn = PREV_INSN (insn))
 | 
      
         | 3264 |  |  |     ;
 | 
      
         | 3265 |  |  |  
 | 
      
         | 3266 |  |  |   return insn;
 | 
      
         | 3267 |  |  | }
 | 
      
         | 3268 |  |  |  
 | 
      
         | 3269 |  |  | /* Find the next insn after INSN that really does something.  This routine
 | 
      
         | 3270 |  |  |    does not look inside SEQUENCEs.  After reload this also skips over
 | 
      
         | 3271 |  |  |    standalone USE and CLOBBER insn.  */
 | 
      
         | 3272 |  |  |  
 | 
      
         | 3273 |  |  | int
 | 
      
         | 3274 |  |  | active_insn_p (const_rtx insn)
 | 
      
         | 3275 |  |  | {
 | 
      
         | 3276 |  |  |   return (CALL_P (insn) || JUMP_P (insn)
 | 
      
         | 3277 |  |  |           || (NONJUMP_INSN_P (insn)
 | 
      
         | 3278 |  |  |               && (! reload_completed
 | 
      
         | 3279 |  |  |                   || (GET_CODE (PATTERN (insn)) != USE
 | 
      
         | 3280 |  |  |                       && GET_CODE (PATTERN (insn)) != CLOBBER))));
 | 
      
         | 3281 |  |  | }
 | 
      
         | 3282 |  |  |  
 | 
      
         | 3283 |  |  | rtx
 | 
      
         | 3284 |  |  | next_active_insn (rtx insn)
 | 
      
         | 3285 |  |  | {
 | 
      
         | 3286 |  |  |   while (insn)
 | 
      
         | 3287 |  |  |     {
 | 
      
         | 3288 |  |  |       insn = NEXT_INSN (insn);
 | 
      
         | 3289 |  |  |       if (insn == 0 || active_insn_p (insn))
 | 
      
         | 3290 |  |  |         break;
 | 
      
         | 3291 |  |  |     }
 | 
      
         | 3292 |  |  |  
 | 
      
         | 3293 |  |  |   return insn;
 | 
      
         | 3294 |  |  | }
 | 
      
         | 3295 |  |  |  
 | 
      
         | 3296 |  |  | /* Find the last insn before INSN that really does something.  This routine
 | 
      
         | 3297 |  |  |    does not look inside SEQUENCEs.  After reload this also skips over
 | 
      
         | 3298 |  |  |    standalone USE and CLOBBER insn.  */
 | 
      
         | 3299 |  |  |  
 | 
      
         | 3300 |  |  | rtx
 | 
      
         | 3301 |  |  | prev_active_insn (rtx insn)
 | 
      
         | 3302 |  |  | {
 | 
      
         | 3303 |  |  |   while (insn)
 | 
      
         | 3304 |  |  |     {
 | 
      
         | 3305 |  |  |       insn = PREV_INSN (insn);
 | 
      
         | 3306 |  |  |       if (insn == 0 || active_insn_p (insn))
 | 
      
         | 3307 |  |  |         break;
 | 
      
         | 3308 |  |  |     }
 | 
      
         | 3309 |  |  |  
 | 
      
         | 3310 |  |  |   return insn;
 | 
      
         | 3311 |  |  | }
 | 
      
         | 3312 |  |  |  
 | 
      
         | 3313 |  |  | /* Return the next CODE_LABEL after the insn INSN, or 0 if there is none.  */
 | 
      
         | 3314 |  |  |  
 | 
      
         | 3315 |  |  | rtx
 | 
      
         | 3316 |  |  | next_label (rtx insn)
 | 
      
         | 3317 |  |  | {
 | 
      
         | 3318 |  |  |   while (insn)
 | 
      
         | 3319 |  |  |     {
 | 
      
         | 3320 |  |  |       insn = NEXT_INSN (insn);
 | 
      
         | 3321 |  |  |       if (insn == 0 || LABEL_P (insn))
 | 
      
         | 3322 |  |  |         break;
 | 
      
         | 3323 |  |  |     }
 | 
      
         | 3324 |  |  |  
 | 
      
         | 3325 |  |  |   return insn;
 | 
      
         | 3326 |  |  | }
 | 
      
         | 3327 |  |  |  
 | 
      
         | 3328 |  |  | /* Return the last label to mark the same position as LABEL.  Return LABEL
 | 
      
         | 3329 |  |  |    itself if it is null or any return rtx.  */
 | 
      
         | 3330 |  |  |  
 | 
      
         | 3331 |  |  | rtx
 | 
      
         | 3332 |  |  | skip_consecutive_labels (rtx label)
 | 
      
         | 3333 |  |  | {
 | 
      
         | 3334 |  |  |   rtx insn;
 | 
      
         | 3335 |  |  |  
 | 
      
         | 3336 |  |  |   if (label && ANY_RETURN_P (label))
 | 
      
         | 3337 |  |  |     return label;
 | 
      
         | 3338 |  |  |  
 | 
      
         | 3339 |  |  |   for (insn = label; insn != 0 && !INSN_P (insn); insn = NEXT_INSN (insn))
 | 
      
         | 3340 |  |  |     if (LABEL_P (insn))
 | 
      
         | 3341 |  |  |       label = insn;
 | 
      
         | 3342 |  |  |  
 | 
      
         | 3343 |  |  |   return label;
 | 
      
         | 3344 |  |  | }
 | 
      
         | 3345 |  |  |  
 | 
      
         | 3346 |  |  | #ifdef HAVE_cc0
 | 
      
         | 3347 |  |  | /* INSN uses CC0 and is being moved into a delay slot.  Set up REG_CC_SETTER
 | 
      
         | 3348 |  |  |    and REG_CC_USER notes so we can find it.  */
 | 
      
         | 3349 |  |  |  
 | 
      
         | 3350 |  |  | void
 | 
      
         | 3351 |  |  | link_cc0_insns (rtx insn)
 | 
      
         | 3352 |  |  | {
 | 
      
         | 3353 |  |  |   rtx user = next_nonnote_insn (insn);
 | 
      
         | 3354 |  |  |  
 | 
      
         | 3355 |  |  |   if (NONJUMP_INSN_P (user) && GET_CODE (PATTERN (user)) == SEQUENCE)
 | 
      
         | 3356 |  |  |     user = XVECEXP (PATTERN (user), 0, 0);
 | 
      
         | 3357 |  |  |  
 | 
      
         | 3358 |  |  |   add_reg_note (user, REG_CC_SETTER, insn);
 | 
      
         | 3359 |  |  |   add_reg_note (insn, REG_CC_USER, user);
 | 
      
         | 3360 |  |  | }
 | 
      
         | 3361 |  |  |  
 | 
      
         | 3362 |  |  | /* Return the next insn that uses CC0 after INSN, which is assumed to
 | 
      
         | 3363 |  |  |    set it.  This is the inverse of prev_cc0_setter (i.e., prev_cc0_setter
 | 
      
         | 3364 |  |  |    applied to the result of this function should yield INSN).
 | 
      
         | 3365 |  |  |  
 | 
      
         | 3366 |  |  |    Normally, this is simply the next insn.  However, if a REG_CC_USER note
 | 
      
         | 3367 |  |  |    is present, it contains the insn that uses CC0.
 | 
      
         | 3368 |  |  |  
 | 
      
         | 3369 |  |  |    Return 0 if we can't find the insn.  */
 | 
      
         | 3370 |  |  |  
 | 
      
         | 3371 |  |  | rtx
 | 
      
         | 3372 |  |  | next_cc0_user (rtx insn)
 | 
      
         | 3373 |  |  | {
 | 
      
         | 3374 |  |  |   rtx note = find_reg_note (insn, REG_CC_USER, NULL_RTX);
 | 
      
         | 3375 |  |  |  
 | 
      
         | 3376 |  |  |   if (note)
 | 
      
         | 3377 |  |  |     return XEXP (note, 0);
 | 
      
         | 3378 |  |  |  
 | 
      
         | 3379 |  |  |   insn = next_nonnote_insn (insn);
 | 
      
         | 3380 |  |  |   if (insn && NONJUMP_INSN_P (insn) && GET_CODE (PATTERN (insn)) == SEQUENCE)
 | 
      
         | 3381 |  |  |     insn = XVECEXP (PATTERN (insn), 0, 0);
 | 
      
         | 3382 |  |  |  
 | 
      
         | 3383 |  |  |   if (insn && INSN_P (insn) && reg_mentioned_p (cc0_rtx, PATTERN (insn)))
 | 
      
         | 3384 |  |  |     return insn;
 | 
      
         | 3385 |  |  |  
 | 
      
         | 3386 |  |  |   return 0;
 | 
      
         | 3387 |  |  | }
 | 
      
         | 3388 |  |  |  
 | 
      
         | 3389 |  |  | /* Find the insn that set CC0 for INSN.  Unless INSN has a REG_CC_SETTER
 | 
      
         | 3390 |  |  |    note, it is the previous insn.  */
 | 
      
         | 3391 |  |  |  
 | 
      
         | 3392 |  |  | rtx
 | 
      
         | 3393 |  |  | prev_cc0_setter (rtx insn)
 | 
      
         | 3394 |  |  | {
 | 
      
         | 3395 |  |  |   rtx note = find_reg_note (insn, REG_CC_SETTER, NULL_RTX);
 | 
      
         | 3396 |  |  |  
 | 
      
         | 3397 |  |  |   if (note)
 | 
      
         | 3398 |  |  |     return XEXP (note, 0);
 | 
      
         | 3399 |  |  |  
 | 
      
         | 3400 |  |  |   insn = prev_nonnote_insn (insn);
 | 
      
         | 3401 |  |  |   gcc_assert (sets_cc0_p (PATTERN (insn)));
 | 
      
         | 3402 |  |  |  
 | 
      
         | 3403 |  |  |   return insn;
 | 
      
         | 3404 |  |  | }
 | 
      
         | 3405 |  |  | #endif
 | 
      
         | 3406 |  |  |  
 | 
      
         | 3407 |  |  | #ifdef AUTO_INC_DEC
 | 
      
         | 3408 |  |  | /* Find a RTX_AUTOINC class rtx which matches DATA.  */
 | 
      
         | 3409 |  |  |  
 | 
      
         | 3410 |  |  | static int
 | 
      
         | 3411 |  |  | find_auto_inc (rtx *xp, void *data)
 | 
      
         | 3412 |  |  | {
 | 
      
         | 3413 |  |  |   rtx x = *xp;
 | 
      
         | 3414 |  |  |   rtx reg = (rtx) data;
 | 
      
         | 3415 |  |  |  
 | 
      
         | 3416 |  |  |   if (GET_RTX_CLASS (GET_CODE (x)) != RTX_AUTOINC)
 | 
      
         | 3417 |  |  |     return 0;
 | 
      
         | 3418 |  |  |  
 | 
      
         | 3419 |  |  |   switch (GET_CODE (x))
 | 
      
         | 3420 |  |  |     {
 | 
      
         | 3421 |  |  |       case PRE_DEC:
 | 
      
         | 3422 |  |  |       case PRE_INC:
 | 
      
         | 3423 |  |  |       case POST_DEC:
 | 
      
         | 3424 |  |  |       case POST_INC:
 | 
      
         | 3425 |  |  |       case PRE_MODIFY:
 | 
      
         | 3426 |  |  |       case POST_MODIFY:
 | 
      
         | 3427 |  |  |         if (rtx_equal_p (reg, XEXP (x, 0)))
 | 
      
         | 3428 |  |  |           return 1;
 | 
      
         | 3429 |  |  |         break;
 | 
      
         | 3430 |  |  |  
 | 
      
         | 3431 |  |  |       default:
 | 
      
         | 3432 |  |  |         gcc_unreachable ();
 | 
      
         | 3433 |  |  |     }
 | 
      
         | 3434 |  |  |   return -1;
 | 
      
         | 3435 |  |  | }
 | 
      
         | 3436 |  |  | #endif
 | 
      
         | 3437 |  |  |  
 | 
      
         | 3438 |  |  | /* Increment the label uses for all labels present in rtx.  */
 | 
      
         | 3439 |  |  |  
 | 
      
         | 3440 |  |  | static void
 | 
      
         | 3441 |  |  | mark_label_nuses (rtx x)
 | 
      
         | 3442 |  |  | {
 | 
      
         | 3443 |  |  |   enum rtx_code code;
 | 
      
         | 3444 |  |  |   int i, j;
 | 
      
         | 3445 |  |  |   const char *fmt;
 | 
      
         | 3446 |  |  |  
 | 
      
         | 3447 |  |  |   code = GET_CODE (x);
 | 
      
         | 3448 |  |  |   if (code == LABEL_REF && LABEL_P (XEXP (x, 0)))
 | 
      
         | 3449 |  |  |     LABEL_NUSES (XEXP (x, 0))++;
 | 
      
         | 3450 |  |  |  
 | 
      
         | 3451 |  |  |   fmt = GET_RTX_FORMAT (code);
 | 
      
         | 3452 |  |  |   for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
 | 
      
         | 3453 |  |  |     {
 | 
      
         | 3454 |  |  |       if (fmt[i] == 'e')
 | 
      
         | 3455 |  |  |         mark_label_nuses (XEXP (x, i));
 | 
      
         | 3456 |  |  |       else if (fmt[i] == 'E')
 | 
      
         | 3457 |  |  |         for (j = XVECLEN (x, i) - 1; j >= 0; j--)
 | 
      
         | 3458 |  |  |           mark_label_nuses (XVECEXP (x, i, j));
 | 
      
         | 3459 |  |  |     }
 | 
      
         | 3460 |  |  | }
 | 
      
         | 3461 |  |  |  
 | 
      
         | 3462 |  |  |  
 | 
      
         | 3463 |  |  | /* Try splitting insns that can be split for better scheduling.
 | 
      
         | 3464 |  |  |    PAT is the pattern which might split.
 | 
      
         | 3465 |  |  |    TRIAL is the insn providing PAT.
 | 
      
         | 3466 |  |  |    LAST is nonzero if we should return the last insn of the sequence produced.
 | 
      
         | 3467 |  |  |  
 | 
      
         | 3468 |  |  |    If this routine succeeds in splitting, it returns the first or last
 | 
      
         | 3469 |  |  |    replacement insn depending on the value of LAST.  Otherwise, it
 | 
      
         | 3470 |  |  |    returns TRIAL.  If the insn to be returned can be split, it will be.  */
 | 
      
         | 3471 |  |  |  
 | 
      
         | 3472 |  |  | rtx
 | 
      
         | 3473 |  |  | try_split (rtx pat, rtx trial, int last)
 | 
      
         | 3474 |  |  | {
 | 
      
         | 3475 |  |  |   rtx before = PREV_INSN (trial);
 | 
      
         | 3476 |  |  |   rtx after = NEXT_INSN (trial);
 | 
      
         | 3477 |  |  |   int has_barrier = 0;
 | 
      
         | 3478 |  |  |   rtx note, seq, tem;
 | 
      
         | 3479 |  |  |   int probability;
 | 
      
         | 3480 |  |  |   rtx insn_last, insn;
 | 
      
         | 3481 |  |  |   int njumps = 0;
 | 
      
         | 3482 |  |  |  
 | 
      
         | 3483 |  |  |   /* We're not good at redistributing frame information.  */
 | 
      
         | 3484 |  |  |   if (RTX_FRAME_RELATED_P (trial))
 | 
      
         | 3485 |  |  |     return trial;
 | 
      
         | 3486 |  |  |  
 | 
      
         | 3487 |  |  |   if (any_condjump_p (trial)
 | 
      
         | 3488 |  |  |       && (note = find_reg_note (trial, REG_BR_PROB, 0)))
 | 
      
         | 3489 |  |  |     split_branch_probability = INTVAL (XEXP (note, 0));
 | 
      
         | 3490 |  |  |   probability = split_branch_probability;
 | 
      
         | 3491 |  |  |  
 | 
      
         | 3492 |  |  |   seq = split_insns (pat, trial);
 | 
      
         | 3493 |  |  |  
 | 
      
         | 3494 |  |  |   split_branch_probability = -1;
 | 
      
         | 3495 |  |  |  
 | 
      
         | 3496 |  |  |   /* If we are splitting a JUMP_INSN, it might be followed by a BARRIER.
 | 
      
         | 3497 |  |  |      We may need to handle this specially.  */
 | 
      
         | 3498 |  |  |   if (after && BARRIER_P (after))
 | 
      
         | 3499 |  |  |     {
 | 
      
         | 3500 |  |  |       has_barrier = 1;
 | 
      
         | 3501 |  |  |       after = NEXT_INSN (after);
 | 
      
         | 3502 |  |  |     }
 | 
      
         | 3503 |  |  |  
 | 
      
         | 3504 |  |  |   if (!seq)
 | 
      
         | 3505 |  |  |     return trial;
 | 
      
         | 3506 |  |  |  
 | 
      
         | 3507 |  |  |   /* Avoid infinite loop if any insn of the result matches
 | 
      
         | 3508 |  |  |      the original pattern.  */
 | 
      
         | 3509 |  |  |   insn_last = seq;
 | 
      
         | 3510 |  |  |   while (1)
 | 
      
         | 3511 |  |  |     {
 | 
      
         | 3512 |  |  |       if (INSN_P (insn_last)
 | 
      
         | 3513 |  |  |           && rtx_equal_p (PATTERN (insn_last), pat))
 | 
      
         | 3514 |  |  |         return trial;
 | 
      
         | 3515 |  |  |       if (!NEXT_INSN (insn_last))
 | 
      
         | 3516 |  |  |         break;
 | 
      
         | 3517 |  |  |       insn_last = NEXT_INSN (insn_last);
 | 
      
         | 3518 |  |  |     }
 | 
      
         | 3519 |  |  |  
 | 
      
         | 3520 |  |  |   /* We will be adding the new sequence to the function.  The splitters
 | 
      
         | 3521 |  |  |      may have introduced invalid RTL sharing, so unshare the sequence now.  */
 | 
      
         | 3522 |  |  |   unshare_all_rtl_in_chain (seq);
 | 
      
         | 3523 |  |  |  
 | 
      
         | 3524 |  |  |   /* Mark labels.  */
 | 
      
         | 3525 |  |  |   for (insn = insn_last; insn ; insn = PREV_INSN (insn))
 | 
      
         | 3526 |  |  |     {
 | 
      
         | 3527 |  |  |       if (JUMP_P (insn))
 | 
      
         | 3528 |  |  |         {
 | 
      
         | 3529 |  |  |           mark_jump_label (PATTERN (insn), insn, 0);
 | 
      
         | 3530 |  |  |           njumps++;
 | 
      
         | 3531 |  |  |           if (probability != -1
 | 
      
         | 3532 |  |  |               && any_condjump_p (insn)
 | 
      
         | 3533 |  |  |               && !find_reg_note (insn, REG_BR_PROB, 0))
 | 
      
         | 3534 |  |  |             {
 | 
      
         | 3535 |  |  |               /* We can preserve the REG_BR_PROB notes only if exactly
 | 
      
         | 3536 |  |  |                  one jump is created, otherwise the machine description
 | 
      
         | 3537 |  |  |                  is responsible for this step using
 | 
      
         | 3538 |  |  |                  split_branch_probability variable.  */
 | 
      
         | 3539 |  |  |               gcc_assert (njumps == 1);
 | 
      
         | 3540 |  |  |               add_reg_note (insn, REG_BR_PROB, GEN_INT (probability));
 | 
      
         | 3541 |  |  |             }
 | 
      
         | 3542 |  |  |         }
 | 
      
         | 3543 |  |  |     }
 | 
      
         | 3544 |  |  |  
 | 
      
         | 3545 |  |  |   /* If we are splitting a CALL_INSN, look for the CALL_INSN
 | 
      
         | 3546 |  |  |      in SEQ and copy any additional information across.  */
 | 
      
         | 3547 |  |  |   if (CALL_P (trial))
 | 
      
         | 3548 |  |  |     {
 | 
      
         | 3549 |  |  |       for (insn = insn_last; insn ; insn = PREV_INSN (insn))
 | 
      
         | 3550 |  |  |         if (CALL_P (insn))
 | 
      
         | 3551 |  |  |           {
 | 
      
         | 3552 |  |  |             rtx next, *p;
 | 
      
         | 3553 |  |  |  
 | 
      
         | 3554 |  |  |             /* Add the old CALL_INSN_FUNCTION_USAGE to whatever the
 | 
      
         | 3555 |  |  |                target may have explicitly specified.  */
 | 
      
         | 3556 |  |  |             p = &CALL_INSN_FUNCTION_USAGE (insn);
 | 
      
         | 3557 |  |  |             while (*p)
 | 
      
         | 3558 |  |  |               p = &XEXP (*p, 1);
 | 
      
         | 3559 |  |  |             *p = CALL_INSN_FUNCTION_USAGE (trial);
 | 
      
         | 3560 |  |  |  
 | 
      
         | 3561 |  |  |             /* If the old call was a sibling call, the new one must
 | 
      
         | 3562 |  |  |                be too.  */
 | 
      
         | 3563 |  |  |             SIBLING_CALL_P (insn) = SIBLING_CALL_P (trial);
 | 
      
         | 3564 |  |  |  
 | 
      
         | 3565 |  |  |             /* If the new call is the last instruction in the sequence,
 | 
      
         | 3566 |  |  |                it will effectively replace the old call in-situ.  Otherwise
 | 
      
         | 3567 |  |  |                we must move any following NOTE_INSN_CALL_ARG_LOCATION note
 | 
      
         | 3568 |  |  |                so that it comes immediately after the new call.  */
 | 
      
         | 3569 |  |  |             if (NEXT_INSN (insn))
 | 
      
         | 3570 |  |  |               for (next = NEXT_INSN (trial);
 | 
      
         | 3571 |  |  |                    next && NOTE_P (next);
 | 
      
         | 3572 |  |  |                    next = NEXT_INSN (next))
 | 
      
         | 3573 |  |  |                 if (NOTE_KIND (next) == NOTE_INSN_CALL_ARG_LOCATION)
 | 
      
         | 3574 |  |  |                   {
 | 
      
         | 3575 |  |  |                     remove_insn (next);
 | 
      
         | 3576 |  |  |                     add_insn_after (next, insn, NULL);
 | 
      
         | 3577 |  |  |                     break;
 | 
      
         | 3578 |  |  |                   }
 | 
      
         | 3579 |  |  |           }
 | 
      
         | 3580 |  |  |     }
 | 
      
         | 3581 |  |  |  
 | 
      
         | 3582 |  |  |   /* Copy notes, particularly those related to the CFG.  */
 | 
      
         | 3583 |  |  |   for (note = REG_NOTES (trial); note; note = XEXP (note, 1))
 | 
      
         | 3584 |  |  |     {
 | 
      
         | 3585 |  |  |       switch (REG_NOTE_KIND (note))
 | 
      
         | 3586 |  |  |         {
 | 
      
         | 3587 |  |  |         case REG_EH_REGION:
 | 
      
         | 3588 |  |  |           copy_reg_eh_region_note_backward (note, insn_last, NULL);
 | 
      
         | 3589 |  |  |           break;
 | 
      
         | 3590 |  |  |  
 | 
      
         | 3591 |  |  |         case REG_NORETURN:
 | 
      
         | 3592 |  |  |         case REG_SETJMP:
 | 
      
         | 3593 |  |  |         case REG_TM:
 | 
      
         | 3594 |  |  |           for (insn = insn_last; insn != NULL_RTX; insn = PREV_INSN (insn))
 | 
      
         | 3595 |  |  |             {
 | 
      
         | 3596 |  |  |               if (CALL_P (insn))
 | 
      
         | 3597 |  |  |                 add_reg_note (insn, REG_NOTE_KIND (note), XEXP (note, 0));
 | 
      
         | 3598 |  |  |             }
 | 
      
         | 3599 |  |  |           break;
 | 
      
         | 3600 |  |  |  
 | 
      
         | 3601 |  |  |         case REG_NON_LOCAL_GOTO:
 | 
      
         | 3602 |  |  |           for (insn = insn_last; insn != NULL_RTX; insn = PREV_INSN (insn))
 | 
      
         | 3603 |  |  |             {
 | 
      
         | 3604 |  |  |               if (JUMP_P (insn))
 | 
      
         | 3605 |  |  |                 add_reg_note (insn, REG_NOTE_KIND (note), XEXP (note, 0));
 | 
      
         | 3606 |  |  |             }
 | 
      
         | 3607 |  |  |           break;
 | 
      
         | 3608 |  |  |  
 | 
      
         | 3609 |  |  | #ifdef AUTO_INC_DEC
 | 
      
         | 3610 |  |  |         case REG_INC:
 | 
      
         | 3611 |  |  |           for (insn = insn_last; insn != NULL_RTX; insn = PREV_INSN (insn))
 | 
      
         | 3612 |  |  |             {
 | 
      
         | 3613 |  |  |               rtx reg = XEXP (note, 0);
 | 
      
         | 3614 |  |  |               if (!FIND_REG_INC_NOTE (insn, reg)
 | 
      
         | 3615 |  |  |                   && for_each_rtx (&PATTERN (insn), find_auto_inc, reg) > 0)
 | 
      
         | 3616 |  |  |                 add_reg_note (insn, REG_INC, reg);
 | 
      
         | 3617 |  |  |             }
 | 
      
         | 3618 |  |  |           break;
 | 
      
         | 3619 |  |  | #endif
 | 
      
         | 3620 |  |  |  
 | 
      
         | 3621 |  |  |         case REG_ARGS_SIZE:
 | 
      
         | 3622 |  |  |           fixup_args_size_notes (NULL_RTX, insn_last, INTVAL (XEXP (note, 0)));
 | 
      
         | 3623 |  |  |           break;
 | 
      
         | 3624 |  |  |  
 | 
      
         | 3625 |  |  |         default:
 | 
      
         | 3626 |  |  |           break;
 | 
      
         | 3627 |  |  |         }
 | 
      
         | 3628 |  |  |     }
 | 
      
         | 3629 |  |  |  
 | 
      
         | 3630 |  |  |   /* If there are LABELS inside the split insns increment the
 | 
      
         | 3631 |  |  |      usage count so we don't delete the label.  */
 | 
      
         | 3632 |  |  |   if (INSN_P (trial))
 | 
      
         | 3633 |  |  |     {
 | 
      
         | 3634 |  |  |       insn = insn_last;
 | 
      
         | 3635 |  |  |       while (insn != NULL_RTX)
 | 
      
         | 3636 |  |  |         {
 | 
      
         | 3637 |  |  |           /* JUMP_P insns have already been "marked" above.  */
 | 
      
         | 3638 |  |  |           if (NONJUMP_INSN_P (insn))
 | 
      
         | 3639 |  |  |             mark_label_nuses (PATTERN (insn));
 | 
      
         | 3640 |  |  |  
 | 
      
         | 3641 |  |  |           insn = PREV_INSN (insn);
 | 
      
         | 3642 |  |  |         }
 | 
      
         | 3643 |  |  |     }
 | 
      
         | 3644 |  |  |  
 | 
      
         | 3645 |  |  |   tem = emit_insn_after_setloc (seq, trial, INSN_LOCATOR (trial));
 | 
      
         | 3646 |  |  |  
 | 
      
         | 3647 |  |  |   delete_insn (trial);
 | 
      
         | 3648 |  |  |   if (has_barrier)
 | 
      
         | 3649 |  |  |     emit_barrier_after (tem);
 | 
      
         | 3650 |  |  |  
 | 
      
         | 3651 |  |  |   /* Recursively call try_split for each new insn created; by the
 | 
      
         | 3652 |  |  |      time control returns here that insn will be fully split, so
 | 
      
         | 3653 |  |  |      set LAST and continue from the insn after the one returned.
 | 
      
         | 3654 |  |  |      We can't use next_active_insn here since AFTER may be a note.
 | 
      
         | 3655 |  |  |      Ignore deleted insns, which can be occur if not optimizing.  */
 | 
      
         | 3656 |  |  |   for (tem = NEXT_INSN (before); tem != after; tem = NEXT_INSN (tem))
 | 
      
         | 3657 |  |  |     if (! INSN_DELETED_P (tem) && INSN_P (tem))
 | 
      
         | 3658 |  |  |       tem = try_split (PATTERN (tem), tem, 1);
 | 
      
         | 3659 |  |  |  
 | 
      
         | 3660 |  |  |   /* Return either the first or the last insn, depending on which was
 | 
      
         | 3661 |  |  |      requested.  */
 | 
      
         | 3662 |  |  |   return last
 | 
      
         | 3663 |  |  |     ? (after ? PREV_INSN (after) : get_last_insn ())
 | 
      
         | 3664 |  |  |     : NEXT_INSN (before);
 | 
      
         | 3665 |  |  | }
 | 
      
         | 3666 |  |  |  
 | 
      
         | 3667 |  |  | /* Make and return an INSN rtx, initializing all its slots.
 | 
      
         | 3668 |  |  |    Store PATTERN in the pattern slots.  */
 | 
      
         | 3669 |  |  |  
 | 
      
         | 3670 |  |  | rtx
 | 
      
         | 3671 |  |  | make_insn_raw (rtx pattern)
 | 
      
         | 3672 |  |  | {
 | 
      
         | 3673 |  |  |   rtx insn;
 | 
      
         | 3674 |  |  |  
 | 
      
         | 3675 |  |  |   insn = rtx_alloc (INSN);
 | 
      
         | 3676 |  |  |  
 | 
      
         | 3677 |  |  |   INSN_UID (insn) = cur_insn_uid++;
 | 
      
         | 3678 |  |  |   PATTERN (insn) = pattern;
 | 
      
         | 3679 |  |  |   INSN_CODE (insn) = -1;
 | 
      
         | 3680 |  |  |   REG_NOTES (insn) = NULL;
 | 
      
         | 3681 |  |  |   INSN_LOCATOR (insn) = curr_insn_locator ();
 | 
      
         | 3682 |  |  |   BLOCK_FOR_INSN (insn) = NULL;
 | 
      
         | 3683 |  |  |  
 | 
      
         | 3684 |  |  | #ifdef ENABLE_RTL_CHECKING
 | 
      
         | 3685 |  |  |   if (insn
 | 
      
         | 3686 |  |  |       && INSN_P (insn)
 | 
      
         | 3687 |  |  |       && (returnjump_p (insn)
 | 
      
         | 3688 |  |  |           || (GET_CODE (insn) == SET
 | 
      
         | 3689 |  |  |               && SET_DEST (insn) == pc_rtx)))
 | 
      
         | 3690 |  |  |     {
 | 
      
         | 3691 |  |  |       warning (0, "ICE: emit_insn used where emit_jump_insn needed:\n");
 | 
      
         | 3692 |  |  |       debug_rtx (insn);
 | 
      
         | 3693 |  |  |     }
 | 
      
         | 3694 |  |  | #endif
 | 
      
         | 3695 |  |  |  
 | 
      
         | 3696 |  |  |   return insn;
 | 
      
         | 3697 |  |  | }
 | 
      
         | 3698 |  |  |  
 | 
      
         | 3699 |  |  | /* Like `make_insn_raw' but make a DEBUG_INSN instead of an insn.  */
 | 
      
         | 3700 |  |  |  
 | 
      
         | 3701 |  |  | rtx
 | 
      
         | 3702 |  |  | make_debug_insn_raw (rtx pattern)
 | 
      
         | 3703 |  |  | {
 | 
      
         | 3704 |  |  |   rtx insn;
 | 
      
         | 3705 |  |  |  
 | 
      
         | 3706 |  |  |   insn = rtx_alloc (DEBUG_INSN);
 | 
      
         | 3707 |  |  |   INSN_UID (insn) = cur_debug_insn_uid++;
 | 
      
         | 3708 |  |  |   if (cur_debug_insn_uid > MIN_NONDEBUG_INSN_UID)
 | 
      
         | 3709 |  |  |     INSN_UID (insn) = cur_insn_uid++;
 | 
      
         | 3710 |  |  |  
 | 
      
         | 3711 |  |  |   PATTERN (insn) = pattern;
 | 
      
         | 3712 |  |  |   INSN_CODE (insn) = -1;
 | 
      
         | 3713 |  |  |   REG_NOTES (insn) = NULL;
 | 
      
         | 3714 |  |  |   INSN_LOCATOR (insn) = curr_insn_locator ();
 | 
      
         | 3715 |  |  |   BLOCK_FOR_INSN (insn) = NULL;
 | 
      
         | 3716 |  |  |  
 | 
      
         | 3717 |  |  |   return insn;
 | 
      
         | 3718 |  |  | }
 | 
      
         | 3719 |  |  |  
 | 
      
         | 3720 |  |  | /* Like `make_insn_raw' but make a JUMP_INSN instead of an insn.  */
 | 
      
         | 3721 |  |  |  
 | 
      
         | 3722 |  |  | rtx
 | 
      
         | 3723 |  |  | make_jump_insn_raw (rtx pattern)
 | 
      
         | 3724 |  |  | {
 | 
      
         | 3725 |  |  |   rtx insn;
 | 
      
         | 3726 |  |  |  
 | 
      
         | 3727 |  |  |   insn = rtx_alloc (JUMP_INSN);
 | 
      
         | 3728 |  |  |   INSN_UID (insn) = cur_insn_uid++;
 | 
      
         | 3729 |  |  |  
 | 
      
         | 3730 |  |  |   PATTERN (insn) = pattern;
 | 
      
         | 3731 |  |  |   INSN_CODE (insn) = -1;
 | 
      
         | 3732 |  |  |   REG_NOTES (insn) = NULL;
 | 
      
         | 3733 |  |  |   JUMP_LABEL (insn) = NULL;
 | 
      
         | 3734 |  |  |   INSN_LOCATOR (insn) = curr_insn_locator ();
 | 
      
         | 3735 |  |  |   BLOCK_FOR_INSN (insn) = NULL;
 | 
      
         | 3736 |  |  |  
 | 
      
         | 3737 |  |  |   return insn;
 | 
      
         | 3738 |  |  | }
 | 
      
         | 3739 |  |  |  
 | 
      
         | 3740 |  |  | /* Like `make_insn_raw' but make a CALL_INSN instead of an insn.  */
 | 
      
         | 3741 |  |  |  
 | 
      
         | 3742 |  |  | static rtx
 | 
      
         | 3743 |  |  | make_call_insn_raw (rtx pattern)
 | 
      
         | 3744 |  |  | {
 | 
      
         | 3745 |  |  |   rtx insn;
 | 
      
         | 3746 |  |  |  
 | 
      
         | 3747 |  |  |   insn = rtx_alloc (CALL_INSN);
 | 
      
         | 3748 |  |  |   INSN_UID (insn) = cur_insn_uid++;
 | 
      
         | 3749 |  |  |  
 | 
      
         | 3750 |  |  |   PATTERN (insn) = pattern;
 | 
      
         | 3751 |  |  |   INSN_CODE (insn) = -1;
 | 
      
         | 3752 |  |  |   REG_NOTES (insn) = NULL;
 | 
      
         | 3753 |  |  |   CALL_INSN_FUNCTION_USAGE (insn) = NULL;
 | 
      
         | 3754 |  |  |   INSN_LOCATOR (insn) = curr_insn_locator ();
 | 
      
         | 3755 |  |  |   BLOCK_FOR_INSN (insn) = NULL;
 | 
      
         | 3756 |  |  |  
 | 
      
         | 3757 |  |  |   return insn;
 | 
      
         | 3758 |  |  | }
 | 
      
         | 3759 |  |  |  
 | 
      
         | 3760 |  |  | /* Add INSN to the end of the doubly-linked list.
 | 
      
         | 3761 |  |  |    INSN may be an INSN, JUMP_INSN, CALL_INSN, CODE_LABEL, BARRIER or NOTE.  */
 | 
      
         | 3762 |  |  |  
 | 
      
         | 3763 |  |  | void
 | 
      
         | 3764 |  |  | add_insn (rtx insn)
 | 
      
         | 3765 |  |  | {
 | 
      
         | 3766 |  |  |   PREV_INSN (insn) = get_last_insn();
 | 
      
         | 3767 |  |  |   NEXT_INSN (insn) = 0;
 | 
      
         | 3768 |  |  |  
 | 
      
         | 3769 |  |  |   if (NULL != get_last_insn())
 | 
      
         | 3770 |  |  |     NEXT_INSN (get_last_insn ()) = insn;
 | 
      
         | 3771 |  |  |  
 | 
      
         | 3772 |  |  |   if (NULL == get_insns ())
 | 
      
         | 3773 |  |  |     set_first_insn (insn);
 | 
      
         | 3774 |  |  |  
 | 
      
         | 3775 |  |  |   set_last_insn (insn);
 | 
      
         | 3776 |  |  | }
 | 
      
         | 3777 |  |  |  
 | 
      
         | 3778 |  |  | /* Add INSN into the doubly-linked list after insn AFTER.  This and
 | 
      
         | 3779 |  |  |    the next should be the only functions called to insert an insn once
 | 
      
         | 3780 |  |  |    delay slots have been filled since only they know how to update a
 | 
      
         | 3781 |  |  |    SEQUENCE.  */
 | 
      
         | 3782 |  |  |  
 | 
      
         | 3783 |  |  | void
 | 
      
         | 3784 |  |  | add_insn_after (rtx insn, rtx after, basic_block bb)
 | 
      
         | 3785 |  |  | {
 | 
      
         | 3786 |  |  |   rtx next = NEXT_INSN (after);
 | 
      
         | 3787 |  |  |  
 | 
      
         | 3788 |  |  |   gcc_assert (!optimize || !INSN_DELETED_P (after));
 | 
      
         | 3789 |  |  |  
 | 
      
         | 3790 |  |  |   NEXT_INSN (insn) = next;
 | 
      
         | 3791 |  |  |   PREV_INSN (insn) = after;
 | 
      
         | 3792 |  |  |  
 | 
      
         | 3793 |  |  |   if (next)
 | 
      
         | 3794 |  |  |     {
 | 
      
         | 3795 |  |  |       PREV_INSN (next) = insn;
 | 
      
         | 3796 |  |  |       if (NONJUMP_INSN_P (next) && GET_CODE (PATTERN (next)) == SEQUENCE)
 | 
      
         | 3797 |  |  |         PREV_INSN (XVECEXP (PATTERN (next), 0, 0)) = insn;
 | 
      
         | 3798 |  |  |     }
 | 
      
         | 3799 |  |  |   else if (get_last_insn () == after)
 | 
      
         | 3800 |  |  |     set_last_insn (insn);
 | 
      
         | 3801 |  |  |   else
 | 
      
         | 3802 |  |  |     {
 | 
      
         | 3803 |  |  |       struct sequence_stack *stack = seq_stack;
 | 
      
         | 3804 |  |  |       /* Scan all pending sequences too.  */
 | 
      
         | 3805 |  |  |       for (; stack; stack = stack->next)
 | 
      
         | 3806 |  |  |         if (after == stack->last)
 | 
      
         | 3807 |  |  |           {
 | 
      
         | 3808 |  |  |             stack->last = insn;
 | 
      
         | 3809 |  |  |             break;
 | 
      
         | 3810 |  |  |           }
 | 
      
         | 3811 |  |  |  
 | 
      
         | 3812 |  |  |       gcc_assert (stack);
 | 
      
         | 3813 |  |  |     }
 | 
      
         | 3814 |  |  |  
 | 
      
         | 3815 |  |  |   if (!BARRIER_P (after)
 | 
      
         | 3816 |  |  |       && !BARRIER_P (insn)
 | 
      
         | 3817 |  |  |       && (bb = BLOCK_FOR_INSN (after)))
 | 
      
         | 3818 |  |  |     {
 | 
      
         | 3819 |  |  |       set_block_for_insn (insn, bb);
 | 
      
         | 3820 |  |  |       if (INSN_P (insn))
 | 
      
         | 3821 |  |  |         df_insn_rescan (insn);
 | 
      
         | 3822 |  |  |       /* Should not happen as first in the BB is always
 | 
      
         | 3823 |  |  |          either NOTE or LABEL.  */
 | 
      
         | 3824 |  |  |       if (BB_END (bb) == after
 | 
      
         | 3825 |  |  |           /* Avoid clobbering of structure when creating new BB.  */
 | 
      
         | 3826 |  |  |           && !BARRIER_P (insn)
 | 
      
         | 3827 |  |  |           && !NOTE_INSN_BASIC_BLOCK_P (insn))
 | 
      
         | 3828 |  |  |         BB_END (bb) = insn;
 | 
      
         | 3829 |  |  |     }
 | 
      
         | 3830 |  |  |  
 | 
      
         | 3831 |  |  |   NEXT_INSN (after) = insn;
 | 
      
         | 3832 |  |  |   if (NONJUMP_INSN_P (after) && GET_CODE (PATTERN (after)) == SEQUENCE)
 | 
      
         | 3833 |  |  |     {
 | 
      
         | 3834 |  |  |       rtx sequence = PATTERN (after);
 | 
      
         | 3835 |  |  |       NEXT_INSN (XVECEXP (sequence, 0, XVECLEN (sequence, 0) - 1)) = insn;
 | 
      
         | 3836 |  |  |     }
 | 
      
         | 3837 |  |  | }
 | 
      
         | 3838 |  |  |  
 | 
      
         | 3839 |  |  | /* Add INSN into the doubly-linked list before insn BEFORE.  This and
 | 
      
         | 3840 |  |  |    the previous should be the only functions called to insert an insn
 | 
      
         | 3841 |  |  |    once delay slots have been filled since only they know how to
 | 
      
         | 3842 |  |  |    update a SEQUENCE.  If BB is NULL, an attempt is made to infer the
 | 
      
         | 3843 |  |  |    bb from before.  */
 | 
      
         | 3844 |  |  |  
 | 
      
         | 3845 |  |  | void
 | 
      
         | 3846 |  |  | add_insn_before (rtx insn, rtx before, basic_block bb)
 | 
      
         | 3847 |  |  | {
 | 
      
         | 3848 |  |  |   rtx prev = PREV_INSN (before);
 | 
      
         | 3849 |  |  |  
 | 
      
         | 3850 |  |  |   gcc_assert (!optimize || !INSN_DELETED_P (before));
 | 
      
         | 3851 |  |  |  
 | 
      
         | 3852 |  |  |   PREV_INSN (insn) = prev;
 | 
      
         | 3853 |  |  |   NEXT_INSN (insn) = before;
 | 
      
         | 3854 |  |  |  
 | 
      
         | 3855 |  |  |   if (prev)
 | 
      
         | 3856 |  |  |     {
 | 
      
         | 3857 |  |  |       NEXT_INSN (prev) = insn;
 | 
      
         | 3858 |  |  |       if (NONJUMP_INSN_P (prev) && GET_CODE (PATTERN (prev)) == SEQUENCE)
 | 
      
         | 3859 |  |  |         {
 | 
      
         | 3860 |  |  |           rtx sequence = PATTERN (prev);
 | 
      
         | 3861 |  |  |           NEXT_INSN (XVECEXP (sequence, 0, XVECLEN (sequence, 0) - 1)) = insn;
 | 
      
         | 3862 |  |  |         }
 | 
      
         | 3863 |  |  |     }
 | 
      
         | 3864 |  |  |   else if (get_insns () == before)
 | 
      
         | 3865 |  |  |     set_first_insn (insn);
 | 
      
         | 3866 |  |  |   else
 | 
      
         | 3867 |  |  |     {
 | 
      
         | 3868 |  |  |       struct sequence_stack *stack = seq_stack;
 | 
      
         | 3869 |  |  |       /* Scan all pending sequences too.  */
 | 
      
         | 3870 |  |  |       for (; stack; stack = stack->next)
 | 
      
         | 3871 |  |  |         if (before == stack->first)
 | 
      
         | 3872 |  |  |           {
 | 
      
         | 3873 |  |  |             stack->first = insn;
 | 
      
         | 3874 |  |  |             break;
 | 
      
         | 3875 |  |  |           }
 | 
      
         | 3876 |  |  |  
 | 
      
         | 3877 |  |  |       gcc_assert (stack);
 | 
      
         | 3878 |  |  |     }
 | 
      
         | 3879 |  |  |  
 | 
      
         | 3880 |  |  |   if (!bb
 | 
      
         | 3881 |  |  |       && !BARRIER_P (before)
 | 
      
         | 3882 |  |  |       && !BARRIER_P (insn))
 | 
      
         | 3883 |  |  |     bb = BLOCK_FOR_INSN (before);
 | 
      
         | 3884 |  |  |  
 | 
      
         | 3885 |  |  |   if (bb)
 | 
      
         | 3886 |  |  |     {
 | 
      
         | 3887 |  |  |       set_block_for_insn (insn, bb);
 | 
      
         | 3888 |  |  |       if (INSN_P (insn))
 | 
      
         | 3889 |  |  |         df_insn_rescan (insn);
 | 
      
         | 3890 |  |  |       /* Should not happen as first in the BB is always either NOTE or
 | 
      
         | 3891 |  |  |          LABEL.  */
 | 
      
         | 3892 |  |  |       gcc_assert (BB_HEAD (bb) != insn
 | 
      
         | 3893 |  |  |                   /* Avoid clobbering of structure when creating new BB.  */
 | 
      
         | 3894 |  |  |                   || BARRIER_P (insn)
 | 
      
         | 3895 |  |  |                   || NOTE_INSN_BASIC_BLOCK_P (insn));
 | 
      
         | 3896 |  |  |     }
 | 
      
         | 3897 |  |  |  
 | 
      
         | 3898 |  |  |   PREV_INSN (before) = insn;
 | 
      
         | 3899 |  |  |   if (NONJUMP_INSN_P (before) && GET_CODE (PATTERN (before)) == SEQUENCE)
 | 
      
         | 3900 |  |  |     PREV_INSN (XVECEXP (PATTERN (before), 0, 0)) = insn;
 | 
      
         | 3901 |  |  | }
 | 
      
         | 3902 |  |  |  
 | 
      
         | 3903 |  |  |  
 | 
      
         | 3904 |  |  | /* Replace insn with an deleted instruction note.  */
 | 
      
         | 3905 |  |  |  
 | 
      
         | 3906 |  |  | void
 | 
      
         | 3907 |  |  | set_insn_deleted (rtx insn)
 | 
      
         | 3908 |  |  | {
 | 
      
         | 3909 |  |  |   df_insn_delete (BLOCK_FOR_INSN (insn), INSN_UID (insn));
 | 
      
         | 3910 |  |  |   PUT_CODE (insn, NOTE);
 | 
      
         | 3911 |  |  |   NOTE_KIND (insn) = NOTE_INSN_DELETED;
 | 
      
         | 3912 |  |  | }
 | 
      
         | 3913 |  |  |  
 | 
      
         | 3914 |  |  |  
 | 
      
         | 3915 |  |  | /* Remove an insn from its doubly-linked list.  This function knows how
 | 
      
         | 3916 |  |  |    to handle sequences.  */
 | 
      
         | 3917 |  |  | void
 | 
      
         | 3918 |  |  | remove_insn (rtx insn)
 | 
      
         | 3919 |  |  | {
 | 
      
         | 3920 |  |  |   rtx next = NEXT_INSN (insn);
 | 
      
         | 3921 |  |  |   rtx prev = PREV_INSN (insn);
 | 
      
         | 3922 |  |  |   basic_block bb;
 | 
      
         | 3923 |  |  |  
 | 
      
         | 3924 |  |  |   /* Later in the code, the block will be marked dirty.  */
 | 
      
         | 3925 |  |  |   df_insn_delete (NULL, INSN_UID (insn));
 | 
      
         | 3926 |  |  |  
 | 
      
         | 3927 |  |  |   if (prev)
 | 
      
         | 3928 |  |  |     {
 | 
      
         | 3929 |  |  |       NEXT_INSN (prev) = next;
 | 
      
         | 3930 |  |  |       if (NONJUMP_INSN_P (prev) && GET_CODE (PATTERN (prev)) == SEQUENCE)
 | 
      
         | 3931 |  |  |         {
 | 
      
         | 3932 |  |  |           rtx sequence = PATTERN (prev);
 | 
      
         | 3933 |  |  |           NEXT_INSN (XVECEXP (sequence, 0, XVECLEN (sequence, 0) - 1)) = next;
 | 
      
         | 3934 |  |  |         }
 | 
      
         | 3935 |  |  |     }
 | 
      
         | 3936 |  |  |   else if (get_insns () == insn)
 | 
      
         | 3937 |  |  |     {
 | 
      
         | 3938 |  |  |       if (next)
 | 
      
         | 3939 |  |  |         PREV_INSN (next) = NULL;
 | 
      
         | 3940 |  |  |       set_first_insn (next);
 | 
      
         | 3941 |  |  |     }
 | 
      
         | 3942 |  |  |   else
 | 
      
         | 3943 |  |  |     {
 | 
      
         | 3944 |  |  |       struct sequence_stack *stack = seq_stack;
 | 
      
         | 3945 |  |  |       /* Scan all pending sequences too.  */
 | 
      
         | 3946 |  |  |       for (; stack; stack = stack->next)
 | 
      
         | 3947 |  |  |         if (insn == stack->first)
 | 
      
         | 3948 |  |  |           {
 | 
      
         | 3949 |  |  |             stack->first = next;
 | 
      
         | 3950 |  |  |             break;
 | 
      
         | 3951 |  |  |           }
 | 
      
         | 3952 |  |  |  
 | 
      
         | 3953 |  |  |       gcc_assert (stack);
 | 
      
         | 3954 |  |  |     }
 | 
      
         | 3955 |  |  |  
 | 
      
         | 3956 |  |  |   if (next)
 | 
      
         | 3957 |  |  |     {
 | 
      
         | 3958 |  |  |       PREV_INSN (next) = prev;
 | 
      
         | 3959 |  |  |       if (NONJUMP_INSN_P (next) && GET_CODE (PATTERN (next)) == SEQUENCE)
 | 
      
         | 3960 |  |  |         PREV_INSN (XVECEXP (PATTERN (next), 0, 0)) = prev;
 | 
      
         | 3961 |  |  |     }
 | 
      
         | 3962 |  |  |   else if (get_last_insn () == insn)
 | 
      
         | 3963 |  |  |     set_last_insn (prev);
 | 
      
         | 3964 |  |  |   else
 | 
      
         | 3965 |  |  |     {
 | 
      
         | 3966 |  |  |       struct sequence_stack *stack = seq_stack;
 | 
      
         | 3967 |  |  |       /* Scan all pending sequences too.  */
 | 
      
         | 3968 |  |  |       for (; stack; stack = stack->next)
 | 
      
         | 3969 |  |  |         if (insn == stack->last)
 | 
      
         | 3970 |  |  |           {
 | 
      
         | 3971 |  |  |             stack->last = prev;
 | 
      
         | 3972 |  |  |             break;
 | 
      
         | 3973 |  |  |           }
 | 
      
         | 3974 |  |  |  
 | 
      
         | 3975 |  |  |       gcc_assert (stack);
 | 
      
         | 3976 |  |  |     }
 | 
      
         | 3977 |  |  |   if (!BARRIER_P (insn)
 | 
      
         | 3978 |  |  |       && (bb = BLOCK_FOR_INSN (insn)))
 | 
      
         | 3979 |  |  |     {
 | 
      
         | 3980 |  |  |       if (NONDEBUG_INSN_P (insn))
 | 
      
         | 3981 |  |  |         df_set_bb_dirty (bb);
 | 
      
         | 3982 |  |  |       if (BB_HEAD (bb) == insn)
 | 
      
         | 3983 |  |  |         {
 | 
      
         | 3984 |  |  |           /* Never ever delete the basic block note without deleting whole
 | 
      
         | 3985 |  |  |              basic block.  */
 | 
      
         | 3986 |  |  |           gcc_assert (!NOTE_P (insn));
 | 
      
         | 3987 |  |  |           BB_HEAD (bb) = next;
 | 
      
         | 3988 |  |  |         }
 | 
      
         | 3989 |  |  |       if (BB_END (bb) == insn)
 | 
      
         | 3990 |  |  |         BB_END (bb) = prev;
 | 
      
         | 3991 |  |  |     }
 | 
      
         | 3992 |  |  | }
 | 
      
         | 3993 |  |  |  
 | 
      
         | 3994 |  |  | /* Append CALL_FUSAGE to the CALL_INSN_FUNCTION_USAGE for CALL_INSN.  */
 | 
      
         | 3995 |  |  |  
 | 
      
         | 3996 |  |  | void
 | 
      
         | 3997 |  |  | add_function_usage_to (rtx call_insn, rtx call_fusage)
 | 
      
         | 3998 |  |  | {
 | 
      
         | 3999 |  |  |   gcc_assert (call_insn && CALL_P (call_insn));
 | 
      
         | 4000 |  |  |  
 | 
      
         | 4001 |  |  |   /* Put the register usage information on the CALL.  If there is already
 | 
      
         | 4002 |  |  |      some usage information, put ours at the end.  */
 | 
      
         | 4003 |  |  |   if (CALL_INSN_FUNCTION_USAGE (call_insn))
 | 
      
         | 4004 |  |  |     {
 | 
      
         | 4005 |  |  |       rtx link;
 | 
      
         | 4006 |  |  |  
 | 
      
         | 4007 |  |  |       for (link = CALL_INSN_FUNCTION_USAGE (call_insn); XEXP (link, 1) != 0;
 | 
      
         | 4008 |  |  |            link = XEXP (link, 1))
 | 
      
         | 4009 |  |  |         ;
 | 
      
         | 4010 |  |  |  
 | 
      
         | 4011 |  |  |       XEXP (link, 1) = call_fusage;
 | 
      
         | 4012 |  |  |     }
 | 
      
         | 4013 |  |  |   else
 | 
      
         | 4014 |  |  |     CALL_INSN_FUNCTION_USAGE (call_insn) = call_fusage;
 | 
      
         | 4015 |  |  | }
 | 
      
         | 4016 |  |  |  
 | 
      
         | 4017 |  |  | /* Delete all insns made since FROM.
 | 
      
         | 4018 |  |  |    FROM becomes the new last instruction.  */
 | 
      
         | 4019 |  |  |  
 | 
      
         | 4020 |  |  | void
 | 
      
         | 4021 |  |  | delete_insns_since (rtx from)
 | 
      
         | 4022 |  |  | {
 | 
      
         | 4023 |  |  |   if (from == 0)
 | 
      
         | 4024 |  |  |     set_first_insn (0);
 | 
      
         | 4025 |  |  |   else
 | 
      
         | 4026 |  |  |     NEXT_INSN (from) = 0;
 | 
      
         | 4027 |  |  |   set_last_insn (from);
 | 
      
         | 4028 |  |  | }
 | 
      
         | 4029 |  |  |  
 | 
      
         | 4030 |  |  | /* This function is deprecated, please use sequences instead.
 | 
      
         | 4031 |  |  |  
 | 
      
         | 4032 |  |  |    Move a consecutive bunch of insns to a different place in the chain.
 | 
      
         | 4033 |  |  |    The insns to be moved are those between FROM and TO.
 | 
      
         | 4034 |  |  |    They are moved to a new position after the insn AFTER.
 | 
      
         | 4035 |  |  |    AFTER must not be FROM or TO or any insn in between.
 | 
      
         | 4036 |  |  |  
 | 
      
         | 4037 |  |  |    This function does not know about SEQUENCEs and hence should not be
 | 
      
         | 4038 |  |  |    called after delay-slot filling has been done.  */
 | 
      
         | 4039 |  |  |  
 | 
      
         | 4040 |  |  | void
 | 
      
         | 4041 |  |  | reorder_insns_nobb (rtx from, rtx to, rtx after)
 | 
      
         | 4042 |  |  | {
 | 
      
         | 4043 |  |  | #ifdef ENABLE_CHECKING
 | 
      
         | 4044 |  |  |   rtx x;
 | 
      
         | 4045 |  |  |   for (x = from; x != to; x = NEXT_INSN (x))
 | 
      
         | 4046 |  |  |     gcc_assert (after != x);
 | 
      
         | 4047 |  |  |   gcc_assert (after != to);
 | 
      
         | 4048 |  |  | #endif
 | 
      
         | 4049 |  |  |  
 | 
      
         | 4050 |  |  |   /* Splice this bunch out of where it is now.  */
 | 
      
         | 4051 |  |  |   if (PREV_INSN (from))
 | 
      
         | 4052 |  |  |     NEXT_INSN (PREV_INSN (from)) = NEXT_INSN (to);
 | 
      
         | 4053 |  |  |   if (NEXT_INSN (to))
 | 
      
         | 4054 |  |  |     PREV_INSN (NEXT_INSN (to)) = PREV_INSN (from);
 | 
      
         | 4055 |  |  |   if (get_last_insn () == to)
 | 
      
         | 4056 |  |  |     set_last_insn (PREV_INSN (from));
 | 
      
         | 4057 |  |  |   if (get_insns () == from)
 | 
      
         | 4058 |  |  |     set_first_insn (NEXT_INSN (to));
 | 
      
         | 4059 |  |  |  
 | 
      
         | 4060 |  |  |   /* Make the new neighbors point to it and it to them.  */
 | 
      
         | 4061 |  |  |   if (NEXT_INSN (after))
 | 
      
         | 4062 |  |  |     PREV_INSN (NEXT_INSN (after)) = to;
 | 
      
         | 4063 |  |  |  
 | 
      
         | 4064 |  |  |   NEXT_INSN (to) = NEXT_INSN (after);
 | 
      
         | 4065 |  |  |   PREV_INSN (from) = after;
 | 
      
         | 4066 |  |  |   NEXT_INSN (after) = from;
 | 
      
         | 4067 |  |  |   if (after == get_last_insn())
 | 
      
         | 4068 |  |  |     set_last_insn (to);
 | 
      
         | 4069 |  |  | }
 | 
      
         | 4070 |  |  |  
 | 
      
         | 4071 |  |  | /* Same as function above, but take care to update BB boundaries.  */
 | 
      
         | 4072 |  |  | void
 | 
      
         | 4073 |  |  | reorder_insns (rtx from, rtx to, rtx after)
 | 
      
         | 4074 |  |  | {
 | 
      
         | 4075 |  |  |   rtx prev = PREV_INSN (from);
 | 
      
         | 4076 |  |  |   basic_block bb, bb2;
 | 
      
         | 4077 |  |  |  
 | 
      
         | 4078 |  |  |   reorder_insns_nobb (from, to, after);
 | 
      
         | 4079 |  |  |  
 | 
      
         | 4080 |  |  |   if (!BARRIER_P (after)
 | 
      
         | 4081 |  |  |       && (bb = BLOCK_FOR_INSN (after)))
 | 
      
         | 4082 |  |  |     {
 | 
      
         | 4083 |  |  |       rtx x;
 | 
      
         | 4084 |  |  |       df_set_bb_dirty (bb);
 | 
      
         | 4085 |  |  |  
 | 
      
         | 4086 |  |  |       if (!BARRIER_P (from)
 | 
      
         | 4087 |  |  |           && (bb2 = BLOCK_FOR_INSN (from)))
 | 
      
         | 4088 |  |  |         {
 | 
      
         | 4089 |  |  |           if (BB_END (bb2) == to)
 | 
      
         | 4090 |  |  |             BB_END (bb2) = prev;
 | 
      
         | 4091 |  |  |           df_set_bb_dirty (bb2);
 | 
      
         | 4092 |  |  |         }
 | 
      
         | 4093 |  |  |  
 | 
      
         | 4094 |  |  |       if (BB_END (bb) == after)
 | 
      
         | 4095 |  |  |         BB_END (bb) = to;
 | 
      
         | 4096 |  |  |  
 | 
      
         | 4097 |  |  |       for (x = from; x != NEXT_INSN (to); x = NEXT_INSN (x))
 | 
      
         | 4098 |  |  |         if (!BARRIER_P (x))
 | 
      
         | 4099 |  |  |           df_insn_change_bb (x, bb);
 | 
      
         | 4100 |  |  |     }
 | 
      
         | 4101 |  |  | }
 | 
      
         | 4102 |  |  |  
 | 
      
         | 4103 |  |  |  
 | 
      
         | 4104 |  |  | /* Emit insn(s) of given code and pattern
 | 
      
         | 4105 |  |  |    at a specified place within the doubly-linked list.
 | 
      
         | 4106 |  |  |  
 | 
      
         | 4107 |  |  |    All of the emit_foo global entry points accept an object
 | 
      
         | 4108 |  |  |    X which is either an insn list or a PATTERN of a single
 | 
      
         | 4109 |  |  |    instruction.
 | 
      
         | 4110 |  |  |  
 | 
      
         | 4111 |  |  |    There are thus a few canonical ways to generate code and
 | 
      
         | 4112 |  |  |    emit it at a specific place in the instruction stream.  For
 | 
      
         | 4113 |  |  |    example, consider the instruction named SPOT and the fact that
 | 
      
         | 4114 |  |  |    we would like to emit some instructions before SPOT.  We might
 | 
      
         | 4115 |  |  |    do it like this:
 | 
      
         | 4116 |  |  |  
 | 
      
         | 4117 |  |  |         start_sequence ();
 | 
      
         | 4118 |  |  |         ... emit the new instructions ...
 | 
      
         | 4119 |  |  |         insns_head = get_insns ();
 | 
      
         | 4120 |  |  |         end_sequence ();
 | 
      
         | 4121 |  |  |  
 | 
      
         | 4122 |  |  |         emit_insn_before (insns_head, SPOT);
 | 
      
         | 4123 |  |  |  
 | 
      
         | 4124 |  |  |    It used to be common to generate SEQUENCE rtl instead, but that
 | 
      
         | 4125 |  |  |    is a relic of the past which no longer occurs.  The reason is that
 | 
      
         | 4126 |  |  |    SEQUENCE rtl results in much fragmented RTL memory since the SEQUENCE
 | 
      
         | 4127 |  |  |    generated would almost certainly die right after it was created.  */
 | 
      
         | 4128 |  |  |  
 | 
      
         | 4129 |  |  | static rtx
 | 
      
         | 4130 |  |  | emit_pattern_before_noloc (rtx x, rtx before, rtx last, basic_block bb,
 | 
      
         | 4131 |  |  |                            rtx (*make_raw) (rtx))
 | 
      
         | 4132 |  |  | {
 | 
      
         | 4133 |  |  |   rtx insn;
 | 
      
         | 4134 |  |  |  
 | 
      
         | 4135 |  |  |   gcc_assert (before);
 | 
      
         | 4136 |  |  |  
 | 
      
         | 4137 |  |  |   if (x == NULL_RTX)
 | 
      
         | 4138 |  |  |     return last;
 | 
      
         | 4139 |  |  |  
 | 
      
         | 4140 |  |  |   switch (GET_CODE (x))
 | 
      
         | 4141 |  |  |     {
 | 
      
         | 4142 |  |  |     case DEBUG_INSN:
 | 
      
         | 4143 |  |  |     case INSN:
 | 
      
         | 4144 |  |  |     case JUMP_INSN:
 | 
      
         | 4145 |  |  |     case CALL_INSN:
 | 
      
         | 4146 |  |  |     case CODE_LABEL:
 | 
      
         | 4147 |  |  |     case BARRIER:
 | 
      
         | 4148 |  |  |     case NOTE:
 | 
      
         | 4149 |  |  |       insn = x;
 | 
      
         | 4150 |  |  |       while (insn)
 | 
      
         | 4151 |  |  |         {
 | 
      
         | 4152 |  |  |           rtx next = NEXT_INSN (insn);
 | 
      
         | 4153 |  |  |           add_insn_before (insn, before, bb);
 | 
      
         | 4154 |  |  |           last = insn;
 | 
      
         | 4155 |  |  |           insn = next;
 | 
      
         | 4156 |  |  |         }
 | 
      
         | 4157 |  |  |       break;
 | 
      
         | 4158 |  |  |  
 | 
      
         | 4159 |  |  | #ifdef ENABLE_RTL_CHECKING
 | 
      
         | 4160 |  |  |     case SEQUENCE:
 | 
      
         | 4161 |  |  |       gcc_unreachable ();
 | 
      
         | 4162 |  |  |       break;
 | 
      
         | 4163 |  |  | #endif
 | 
      
         | 4164 |  |  |  
 | 
      
         | 4165 |  |  |     default:
 | 
      
         | 4166 |  |  |       last = (*make_raw) (x);
 | 
      
         | 4167 |  |  |       add_insn_before (last, before, bb);
 | 
      
         | 4168 |  |  |       break;
 | 
      
         | 4169 |  |  |     }
 | 
      
         | 4170 |  |  |  
 | 
      
         | 4171 |  |  |   return last;
 | 
      
         | 4172 |  |  | }
 | 
      
         | 4173 |  |  |  
 | 
      
         | 4174 |  |  | /* Make X be output before the instruction BEFORE.  */
 | 
      
         | 4175 |  |  |  
 | 
      
         | 4176 |  |  | rtx
 | 
      
         | 4177 |  |  | emit_insn_before_noloc (rtx x, rtx before, basic_block bb)
 | 
      
         | 4178 |  |  | {
 | 
      
         | 4179 |  |  |   return emit_pattern_before_noloc (x, before, before, bb, make_insn_raw);
 | 
      
         | 4180 |  |  | }
 | 
      
         | 4181 |  |  |  
 | 
      
         | 4182 |  |  | /* Make an instruction with body X and code JUMP_INSN
 | 
      
         | 4183 |  |  |    and output it before the instruction BEFORE.  */
 | 
      
         | 4184 |  |  |  
 | 
      
         | 4185 |  |  | rtx
 | 
      
         | 4186 |  |  | emit_jump_insn_before_noloc (rtx x, rtx before)
 | 
      
         | 4187 |  |  | {
 | 
      
         | 4188 |  |  |   return emit_pattern_before_noloc (x, before, NULL_RTX, NULL,
 | 
      
         | 4189 |  |  |                                     make_jump_insn_raw);
 | 
      
         | 4190 |  |  | }
 | 
      
         | 4191 |  |  |  
 | 
      
         | 4192 |  |  | /* Make an instruction with body X and code CALL_INSN
 | 
      
         | 4193 |  |  |    and output it before the instruction BEFORE.  */
 | 
      
         | 4194 |  |  |  
 | 
      
         | 4195 |  |  | rtx
 | 
      
         | 4196 |  |  | emit_call_insn_before_noloc (rtx x, rtx before)
 | 
      
         | 4197 |  |  | {
 | 
      
         | 4198 |  |  |   return emit_pattern_before_noloc (x, before, NULL_RTX, NULL,
 | 
      
         | 4199 |  |  |                                     make_call_insn_raw);
 | 
      
         | 4200 |  |  | }
 | 
      
         | 4201 |  |  |  
 | 
      
         | 4202 |  |  | /* Make an instruction with body X and code DEBUG_INSN
 | 
      
         | 4203 |  |  |    and output it before the instruction BEFORE.  */
 | 
      
         | 4204 |  |  |  
 | 
      
         | 4205 |  |  | rtx
 | 
      
         | 4206 |  |  | emit_debug_insn_before_noloc (rtx x, rtx before)
 | 
      
         | 4207 |  |  | {
 | 
      
         | 4208 |  |  |   return emit_pattern_before_noloc (x, before, NULL_RTX, NULL,
 | 
      
         | 4209 |  |  |                                     make_debug_insn_raw);
 | 
      
         | 4210 |  |  | }
 | 
      
         | 4211 |  |  |  
 | 
      
         | 4212 |  |  | /* Make an insn of code BARRIER
 | 
      
         | 4213 |  |  |    and output it before the insn BEFORE.  */
 | 
      
         | 4214 |  |  |  
 | 
      
         | 4215 |  |  | rtx
 | 
      
         | 4216 |  |  | emit_barrier_before (rtx before)
 | 
      
         | 4217 |  |  | {
 | 
      
         | 4218 |  |  |   rtx insn = rtx_alloc (BARRIER);
 | 
      
         | 4219 |  |  |  
 | 
      
         | 4220 |  |  |   INSN_UID (insn) = cur_insn_uid++;
 | 
      
         | 4221 |  |  |  
 | 
      
         | 4222 |  |  |   add_insn_before (insn, before, NULL);
 | 
      
         | 4223 |  |  |   return insn;
 | 
      
         | 4224 |  |  | }
 | 
      
         | 4225 |  |  |  
 | 
      
         | 4226 |  |  | /* Emit the label LABEL before the insn BEFORE.  */
 | 
      
         | 4227 |  |  |  
 | 
      
         | 4228 |  |  | rtx
 | 
      
         | 4229 |  |  | emit_label_before (rtx label, rtx before)
 | 
      
         | 4230 |  |  | {
 | 
      
         | 4231 |  |  |   /* This can be called twice for the same label as a result of the
 | 
      
         | 4232 |  |  |      confusion that follows a syntax error!  So make it harmless.  */
 | 
      
         | 4233 |  |  |   if (INSN_UID (label) == 0)
 | 
      
         | 4234 |  |  |     {
 | 
      
         | 4235 |  |  |       INSN_UID (label) = cur_insn_uid++;
 | 
      
         | 4236 |  |  |       add_insn_before (label, before, NULL);
 | 
      
         | 4237 |  |  |     }
 | 
      
         | 4238 |  |  |  
 | 
      
         | 4239 |  |  |   return label;
 | 
      
         | 4240 |  |  | }
 | 
      
         | 4241 |  |  |  
 | 
      
         | 4242 |  |  | /* Emit a note of subtype SUBTYPE before the insn BEFORE.  */
 | 
      
         | 4243 |  |  |  
 | 
      
         | 4244 |  |  | rtx
 | 
      
         | 4245 |  |  | emit_note_before (enum insn_note subtype, rtx before)
 | 
      
         | 4246 |  |  | {
 | 
      
         | 4247 |  |  |   rtx note = rtx_alloc (NOTE);
 | 
      
         | 4248 |  |  |   INSN_UID (note) = cur_insn_uid++;
 | 
      
         | 4249 |  |  |   NOTE_KIND (note) = subtype;
 | 
      
         | 4250 |  |  |   BLOCK_FOR_INSN (note) = NULL;
 | 
      
         | 4251 |  |  |   memset (&NOTE_DATA (note), 0, sizeof (NOTE_DATA (note)));
 | 
      
         | 4252 |  |  |  
 | 
      
         | 4253 |  |  |   add_insn_before (note, before, NULL);
 | 
      
         | 4254 |  |  |   return note;
 | 
      
         | 4255 |  |  | }
 | 
      
         | 4256 |  |  |  
 | 
      
         | 4257 |  |  | /* Helper for emit_insn_after, handles lists of instructions
 | 
      
         | 4258 |  |  |    efficiently.  */
 | 
      
         | 4259 |  |  |  
 | 
      
         | 4260 |  |  | static rtx
 | 
      
         | 4261 |  |  | emit_insn_after_1 (rtx first, rtx after, basic_block bb)
 | 
      
         | 4262 |  |  | {
 | 
      
         | 4263 |  |  |   rtx last;
 | 
      
         | 4264 |  |  |   rtx after_after;
 | 
      
         | 4265 |  |  |   if (!bb && !BARRIER_P (after))
 | 
      
         | 4266 |  |  |     bb = BLOCK_FOR_INSN (after);
 | 
      
         | 4267 |  |  |  
 | 
      
         | 4268 |  |  |   if (bb)
 | 
      
         | 4269 |  |  |     {
 | 
      
         | 4270 |  |  |       df_set_bb_dirty (bb);
 | 
      
         | 4271 |  |  |       for (last = first; NEXT_INSN (last); last = NEXT_INSN (last))
 | 
      
         | 4272 |  |  |         if (!BARRIER_P (last))
 | 
      
         | 4273 |  |  |           {
 | 
      
         | 4274 |  |  |             set_block_for_insn (last, bb);
 | 
      
         | 4275 |  |  |             df_insn_rescan (last);
 | 
      
         | 4276 |  |  |           }
 | 
      
         | 4277 |  |  |       if (!BARRIER_P (last))
 | 
      
         | 4278 |  |  |         {
 | 
      
         | 4279 |  |  |           set_block_for_insn (last, bb);
 | 
      
         | 4280 |  |  |           df_insn_rescan (last);
 | 
      
         | 4281 |  |  |         }
 | 
      
         | 4282 |  |  |       if (BB_END (bb) == after)
 | 
      
         | 4283 |  |  |         BB_END (bb) = last;
 | 
      
         | 4284 |  |  |     }
 | 
      
         | 4285 |  |  |   else
 | 
      
         | 4286 |  |  |     for (last = first; NEXT_INSN (last); last = NEXT_INSN (last))
 | 
      
         | 4287 |  |  |       continue;
 | 
      
         | 4288 |  |  |  
 | 
      
         | 4289 |  |  |   after_after = NEXT_INSN (after);
 | 
      
         | 4290 |  |  |  
 | 
      
         | 4291 |  |  |   NEXT_INSN (after) = first;
 | 
      
         | 4292 |  |  |   PREV_INSN (first) = after;
 | 
      
         | 4293 |  |  |   NEXT_INSN (last) = after_after;
 | 
      
         | 4294 |  |  |   if (after_after)
 | 
      
         | 4295 |  |  |     PREV_INSN (after_after) = last;
 | 
      
         | 4296 |  |  |  
 | 
      
         | 4297 |  |  |   if (after == get_last_insn())
 | 
      
         | 4298 |  |  |     set_last_insn (last);
 | 
      
         | 4299 |  |  |  
 | 
      
         | 4300 |  |  |   return last;
 | 
      
         | 4301 |  |  | }
 | 
      
         | 4302 |  |  |  
 | 
      
         | 4303 |  |  | static rtx
 | 
      
         | 4304 |  |  | emit_pattern_after_noloc (rtx x, rtx after, basic_block bb,
 | 
      
         | 4305 |  |  |                           rtx (*make_raw)(rtx))
 | 
      
         | 4306 |  |  | {
 | 
      
         | 4307 |  |  |   rtx last = after;
 | 
      
         | 4308 |  |  |  
 | 
      
         | 4309 |  |  |   gcc_assert (after);
 | 
      
         | 4310 |  |  |  
 | 
      
         | 4311 |  |  |   if (x == NULL_RTX)
 | 
      
         | 4312 |  |  |     return last;
 | 
      
         | 4313 |  |  |  
 | 
      
         | 4314 |  |  |   switch (GET_CODE (x))
 | 
      
         | 4315 |  |  |     {
 | 
      
         | 4316 |  |  |     case DEBUG_INSN:
 | 
      
         | 4317 |  |  |     case INSN:
 | 
      
         | 4318 |  |  |     case JUMP_INSN:
 | 
      
         | 4319 |  |  |     case CALL_INSN:
 | 
      
         | 4320 |  |  |     case CODE_LABEL:
 | 
      
         | 4321 |  |  |     case BARRIER:
 | 
      
         | 4322 |  |  |     case NOTE:
 | 
      
         | 4323 |  |  |       last = emit_insn_after_1 (x, after, bb);
 | 
      
         | 4324 |  |  |       break;
 | 
      
         | 4325 |  |  |  
 | 
      
         | 4326 |  |  | #ifdef ENABLE_RTL_CHECKING
 | 
      
         | 4327 |  |  |     case SEQUENCE:
 | 
      
         | 4328 |  |  |       gcc_unreachable ();
 | 
      
         | 4329 |  |  |       break;
 | 
      
         | 4330 |  |  | #endif
 | 
      
         | 4331 |  |  |  
 | 
      
         | 4332 |  |  |     default:
 | 
      
         | 4333 |  |  |       last = (*make_raw) (x);
 | 
      
         | 4334 |  |  |       add_insn_after (last, after, bb);
 | 
      
         | 4335 |  |  |       break;
 | 
      
         | 4336 |  |  |     }
 | 
      
         | 4337 |  |  |  
 | 
      
         | 4338 |  |  |   return last;
 | 
      
         | 4339 |  |  | }
 | 
      
         | 4340 |  |  |  
 | 
      
         | 4341 |  |  | /* Make X be output after the insn AFTER and set the BB of insn.  If
 | 
      
         | 4342 |  |  |    BB is NULL, an attempt is made to infer the BB from AFTER.  */
 | 
      
         | 4343 |  |  |  
 | 
      
         | 4344 |  |  | rtx
 | 
      
         | 4345 |  |  | emit_insn_after_noloc (rtx x, rtx after, basic_block bb)
 | 
      
         | 4346 |  |  | {
 | 
      
         | 4347 |  |  |   return emit_pattern_after_noloc (x, after, bb, make_insn_raw);
 | 
      
         | 4348 |  |  | }
 | 
      
         | 4349 |  |  |  
 | 
      
         | 4350 |  |  |  
 | 
      
         | 4351 |  |  | /* Make an insn of code JUMP_INSN with body X
 | 
      
         | 4352 |  |  |    and output it after the insn AFTER.  */
 | 
      
         | 4353 |  |  |  
 | 
      
         | 4354 |  |  | rtx
 | 
      
         | 4355 |  |  | emit_jump_insn_after_noloc (rtx x, rtx after)
 | 
      
         | 4356 |  |  | {
 | 
      
         | 4357 |  |  |   return emit_pattern_after_noloc (x, after, NULL, make_jump_insn_raw);
 | 
      
         | 4358 |  |  | }
 | 
      
         | 4359 |  |  |  
 | 
      
         | 4360 |  |  | /* Make an instruction with body X and code CALL_INSN
 | 
      
         | 4361 |  |  |    and output it after the instruction AFTER.  */
 | 
      
         | 4362 |  |  |  
 | 
      
         | 4363 |  |  | rtx
 | 
      
         | 4364 |  |  | emit_call_insn_after_noloc (rtx x, rtx after)
 | 
      
         | 4365 |  |  | {
 | 
      
         | 4366 |  |  |   return emit_pattern_after_noloc (x, after, NULL, make_call_insn_raw);
 | 
      
         | 4367 |  |  | }
 | 
      
         | 4368 |  |  |  
 | 
      
         | 4369 |  |  | /* Make an instruction with body X and code CALL_INSN
 | 
      
         | 4370 |  |  |    and output it after the instruction AFTER.  */
 | 
      
         | 4371 |  |  |  
 | 
      
         | 4372 |  |  | rtx
 | 
      
         | 4373 |  |  | emit_debug_insn_after_noloc (rtx x, rtx after)
 | 
      
         | 4374 |  |  | {
 | 
      
         | 4375 |  |  |   return emit_pattern_after_noloc (x, after, NULL, make_debug_insn_raw);
 | 
      
         | 4376 |  |  | }
 | 
      
         | 4377 |  |  |  
 | 
      
         | 4378 |  |  | /* Make an insn of code BARRIER
 | 
      
         | 4379 |  |  |    and output it after the insn AFTER.  */
 | 
      
         | 4380 |  |  |  
 | 
      
         | 4381 |  |  | rtx
 | 
      
         | 4382 |  |  | emit_barrier_after (rtx after)
 | 
      
         | 4383 |  |  | {
 | 
      
         | 4384 |  |  |   rtx insn = rtx_alloc (BARRIER);
 | 
      
         | 4385 |  |  |  
 | 
      
         | 4386 |  |  |   INSN_UID (insn) = cur_insn_uid++;
 | 
      
         | 4387 |  |  |  
 | 
      
         | 4388 |  |  |   add_insn_after (insn, after, NULL);
 | 
      
         | 4389 |  |  |   return insn;
 | 
      
         | 4390 |  |  | }
 | 
      
         | 4391 |  |  |  
 | 
      
         | 4392 |  |  | /* Emit the label LABEL after the insn AFTER.  */
 | 
      
         | 4393 |  |  |  
 | 
      
         | 4394 |  |  | rtx
 | 
      
         | 4395 |  |  | emit_label_after (rtx label, rtx after)
 | 
      
         | 4396 |  |  | {
 | 
      
         | 4397 |  |  |   /* This can be called twice for the same label
 | 
      
         | 4398 |  |  |      as a result of the confusion that follows a syntax error!
 | 
      
         | 4399 |  |  |      So make it harmless.  */
 | 
      
         | 4400 |  |  |   if (INSN_UID (label) == 0)
 | 
      
         | 4401 |  |  |     {
 | 
      
         | 4402 |  |  |       INSN_UID (label) = cur_insn_uid++;
 | 
      
         | 4403 |  |  |       add_insn_after (label, after, NULL);
 | 
      
         | 4404 |  |  |     }
 | 
      
         | 4405 |  |  |  
 | 
      
         | 4406 |  |  |   return label;
 | 
      
         | 4407 |  |  | }
 | 
      
         | 4408 |  |  |  
 | 
      
         | 4409 |  |  | /* Emit a note of subtype SUBTYPE after the insn AFTER.  */
 | 
      
         | 4410 |  |  |  
 | 
      
         | 4411 |  |  | rtx
 | 
      
         | 4412 |  |  | emit_note_after (enum insn_note subtype, rtx after)
 | 
      
         | 4413 |  |  | {
 | 
      
         | 4414 |  |  |   rtx note = rtx_alloc (NOTE);
 | 
      
         | 4415 |  |  |   INSN_UID (note) = cur_insn_uid++;
 | 
      
         | 4416 |  |  |   NOTE_KIND (note) = subtype;
 | 
      
         | 4417 |  |  |   BLOCK_FOR_INSN (note) = NULL;
 | 
      
         | 4418 |  |  |   memset (&NOTE_DATA (note), 0, sizeof (NOTE_DATA (note)));
 | 
      
         | 4419 |  |  |   add_insn_after (note, after, NULL);
 | 
      
         | 4420 |  |  |   return note;
 | 
      
         | 4421 |  |  | }
 | 
      
         | 4422 |  |  |  
 | 
      
         | 4423 |  |  | /* Insert PATTERN after AFTER, setting its INSN_LOCATION to LOC.
 | 
      
         | 4424 |  |  |    MAKE_RAW indicates how to turn PATTERN into a real insn.  */
 | 
      
         | 4425 |  |  |  
 | 
      
         | 4426 |  |  | static rtx
 | 
      
         | 4427 |  |  | emit_pattern_after_setloc (rtx pattern, rtx after, int loc,
 | 
      
         | 4428 |  |  |                            rtx (*make_raw) (rtx))
 | 
      
         | 4429 |  |  | {
 | 
      
         | 4430 |  |  |   rtx last = emit_pattern_after_noloc (pattern, after, NULL, make_raw);
 | 
      
         | 4431 |  |  |  
 | 
      
         | 4432 |  |  |   if (pattern == NULL_RTX || !loc)
 | 
      
         | 4433 |  |  |     return last;
 | 
      
         | 4434 |  |  |  
 | 
      
         | 4435 |  |  |   after = NEXT_INSN (after);
 | 
      
         | 4436 |  |  |   while (1)
 | 
      
         | 4437 |  |  |     {
 | 
      
         | 4438 |  |  |       if (active_insn_p (after) && !INSN_LOCATOR (after))
 | 
      
         | 4439 |  |  |         INSN_LOCATOR (after) = loc;
 | 
      
         | 4440 |  |  |       if (after == last)
 | 
      
         | 4441 |  |  |         break;
 | 
      
         | 4442 |  |  |       after = NEXT_INSN (after);
 | 
      
         | 4443 |  |  |     }
 | 
      
         | 4444 |  |  |   return last;
 | 
      
         | 4445 |  |  | }
 | 
      
         | 4446 |  |  |  
 | 
      
         | 4447 |  |  | /* Insert PATTERN after AFTER.  MAKE_RAW indicates how to turn PATTERN
 | 
      
         | 4448 |  |  |    into a real insn.  SKIP_DEBUG_INSNS indicates whether to insert after
 | 
      
         | 4449 |  |  |    any DEBUG_INSNs.  */
 | 
      
         | 4450 |  |  |  
 | 
      
         | 4451 |  |  | static rtx
 | 
      
         | 4452 |  |  | emit_pattern_after (rtx pattern, rtx after, bool skip_debug_insns,
 | 
      
         | 4453 |  |  |                     rtx (*make_raw) (rtx))
 | 
      
         | 4454 |  |  | {
 | 
      
         | 4455 |  |  |   rtx prev = after;
 | 
      
         | 4456 |  |  |  
 | 
      
         | 4457 |  |  |   if (skip_debug_insns)
 | 
      
         | 4458 |  |  |     while (DEBUG_INSN_P (prev))
 | 
      
         | 4459 |  |  |       prev = PREV_INSN (prev);
 | 
      
         | 4460 |  |  |  
 | 
      
         | 4461 |  |  |   if (INSN_P (prev))
 | 
      
         | 4462 |  |  |     return emit_pattern_after_setloc (pattern, after, INSN_LOCATOR (prev),
 | 
      
         | 4463 |  |  |                                       make_raw);
 | 
      
         | 4464 |  |  |   else
 | 
      
         | 4465 |  |  |     return emit_pattern_after_noloc (pattern, after, NULL, make_raw);
 | 
      
         | 4466 |  |  | }
 | 
      
         | 4467 |  |  |  
 | 
      
         | 4468 |  |  | /* Like emit_insn_after_noloc, but set INSN_LOCATOR according to LOC.  */
 | 
      
         | 4469 |  |  | rtx
 | 
      
         | 4470 |  |  | emit_insn_after_setloc (rtx pattern, rtx after, int loc)
 | 
      
         | 4471 |  |  | {
 | 
      
         | 4472 |  |  |   return emit_pattern_after_setloc (pattern, after, loc, make_insn_raw);
 | 
      
         | 4473 |  |  | }
 | 
      
         | 4474 |  |  |  
 | 
      
         | 4475 |  |  | /* Like emit_insn_after_noloc, but set INSN_LOCATOR according to AFTER.  */
 | 
      
         | 4476 |  |  | rtx
 | 
      
         | 4477 |  |  | emit_insn_after (rtx pattern, rtx after)
 | 
      
         | 4478 |  |  | {
 | 
      
         | 4479 |  |  |   return emit_pattern_after (pattern, after, true, make_insn_raw);
 | 
      
         | 4480 |  |  | }
 | 
      
         | 4481 |  |  |  
 | 
      
         | 4482 |  |  | /* Like emit_jump_insn_after_noloc, but set INSN_LOCATOR according to LOC.  */
 | 
      
         | 4483 |  |  | rtx
 | 
      
         | 4484 |  |  | emit_jump_insn_after_setloc (rtx pattern, rtx after, int loc)
 | 
      
         | 4485 |  |  | {
 | 
      
         | 4486 |  |  |   return emit_pattern_after_setloc (pattern, after, loc, make_jump_insn_raw);
 | 
      
         | 4487 |  |  | }
 | 
      
         | 4488 |  |  |  
 | 
      
         | 4489 |  |  | /* Like emit_jump_insn_after_noloc, but set INSN_LOCATOR according to AFTER.  */
 | 
      
         | 4490 |  |  | rtx
 | 
      
         | 4491 |  |  | emit_jump_insn_after (rtx pattern, rtx after)
 | 
      
         | 4492 |  |  | {
 | 
      
         | 4493 |  |  |   return emit_pattern_after (pattern, after, true, make_jump_insn_raw);
 | 
      
         | 4494 |  |  | }
 | 
      
         | 4495 |  |  |  
 | 
      
         | 4496 |  |  | /* Like emit_call_insn_after_noloc, but set INSN_LOCATOR according to LOC.  */
 | 
      
         | 4497 |  |  | rtx
 | 
      
         | 4498 |  |  | emit_call_insn_after_setloc (rtx pattern, rtx after, int loc)
 | 
      
         | 4499 |  |  | {
 | 
      
         | 4500 |  |  |   return emit_pattern_after_setloc (pattern, after, loc, make_call_insn_raw);
 | 
      
         | 4501 |  |  | }
 | 
      
         | 4502 |  |  |  
 | 
      
         | 4503 |  |  | /* Like emit_call_insn_after_noloc, but set INSN_LOCATOR according to AFTER.  */
 | 
      
         | 4504 |  |  | rtx
 | 
      
         | 4505 |  |  | emit_call_insn_after (rtx pattern, rtx after)
 | 
      
         | 4506 |  |  | {
 | 
      
         | 4507 |  |  |   return emit_pattern_after (pattern, after, true, make_call_insn_raw);
 | 
      
         | 4508 |  |  | }
 | 
      
         | 4509 |  |  |  
 | 
      
         | 4510 |  |  | /* Like emit_debug_insn_after_noloc, but set INSN_LOCATOR according to LOC.  */
 | 
      
         | 4511 |  |  | rtx
 | 
      
         | 4512 |  |  | emit_debug_insn_after_setloc (rtx pattern, rtx after, int loc)
 | 
      
         | 4513 |  |  | {
 | 
      
         | 4514 |  |  |   return emit_pattern_after_setloc (pattern, after, loc, make_debug_insn_raw);
 | 
      
         | 4515 |  |  | }
 | 
      
         | 4516 |  |  |  
 | 
      
         | 4517 |  |  | /* Like emit_debug_insn_after_noloc, but set INSN_LOCATOR according to AFTER.  */
 | 
      
         | 4518 |  |  | rtx
 | 
      
         | 4519 |  |  | emit_debug_insn_after (rtx pattern, rtx after)
 | 
      
         | 4520 |  |  | {
 | 
      
         | 4521 |  |  |   return emit_pattern_after (pattern, after, false, make_debug_insn_raw);
 | 
      
         | 4522 |  |  | }
 | 
      
         | 4523 |  |  |  
 | 
      
         | 4524 |  |  | /* Insert PATTERN before BEFORE, setting its INSN_LOCATION to LOC.
 | 
      
         | 4525 |  |  |    MAKE_RAW indicates how to turn PATTERN into a real insn.  INSNP
 | 
      
         | 4526 |  |  |    indicates if PATTERN is meant for an INSN as opposed to a JUMP_INSN,
 | 
      
         | 4527 |  |  |    CALL_INSN, etc.  */
 | 
      
         | 4528 |  |  |  
 | 
      
         | 4529 |  |  | static rtx
 | 
      
         | 4530 |  |  | emit_pattern_before_setloc (rtx pattern, rtx before, int loc, bool insnp,
 | 
      
         | 4531 |  |  |                             rtx (*make_raw) (rtx))
 | 
      
         | 4532 |  |  | {
 | 
      
         | 4533 |  |  |   rtx first = PREV_INSN (before);
 | 
      
         | 4534 |  |  |   rtx last = emit_pattern_before_noloc (pattern, before,
 | 
      
         | 4535 |  |  |                                         insnp ? before : NULL_RTX,
 | 
      
         | 4536 |  |  |                                         NULL, make_raw);
 | 
      
         | 4537 |  |  |  
 | 
      
         | 4538 |  |  |   if (pattern == NULL_RTX || !loc)
 | 
      
         | 4539 |  |  |     return last;
 | 
      
         | 4540 |  |  |  
 | 
      
         | 4541 |  |  |   if (!first)
 | 
      
         | 4542 |  |  |     first = get_insns ();
 | 
      
         | 4543 |  |  |   else
 | 
      
         | 4544 |  |  |     first = NEXT_INSN (first);
 | 
      
         | 4545 |  |  |   while (1)
 | 
      
         | 4546 |  |  |     {
 | 
      
         | 4547 |  |  |       if (active_insn_p (first) && !INSN_LOCATOR (first))
 | 
      
         | 4548 |  |  |         INSN_LOCATOR (first) = loc;
 | 
      
         | 4549 |  |  |       if (first == last)
 | 
      
         | 4550 |  |  |         break;
 | 
      
         | 4551 |  |  |       first = NEXT_INSN (first);
 | 
      
         | 4552 |  |  |     }
 | 
      
         | 4553 |  |  |   return last;
 | 
      
         | 4554 |  |  | }
 | 
      
         | 4555 |  |  |  
 | 
      
         | 4556 |  |  | /* Insert PATTERN before BEFORE.  MAKE_RAW indicates how to turn PATTERN
 | 
      
         | 4557 |  |  |    into a real insn.  SKIP_DEBUG_INSNS indicates whether to insert
 | 
      
         | 4558 |  |  |    before any DEBUG_INSNs.  INSNP indicates if PATTERN is meant for an
 | 
      
         | 4559 |  |  |    INSN as opposed to a JUMP_INSN, CALL_INSN, etc.  */
 | 
      
         | 4560 |  |  |  
 | 
      
         | 4561 |  |  | static rtx
 | 
      
         | 4562 |  |  | emit_pattern_before (rtx pattern, rtx before, bool skip_debug_insns,
 | 
      
         | 4563 |  |  |                      bool insnp, rtx (*make_raw) (rtx))
 | 
      
         | 4564 |  |  | {
 | 
      
         | 4565 |  |  |   rtx next = before;
 | 
      
         | 4566 |  |  |  
 | 
      
         | 4567 |  |  |   if (skip_debug_insns)
 | 
      
         | 4568 |  |  |     while (DEBUG_INSN_P (next))
 | 
      
         | 4569 |  |  |       next = PREV_INSN (next);
 | 
      
         | 4570 |  |  |  
 | 
      
         | 4571 |  |  |   if (INSN_P (next))
 | 
      
         | 4572 |  |  |     return emit_pattern_before_setloc (pattern, before, INSN_LOCATOR (next),
 | 
      
         | 4573 |  |  |                                        insnp, make_raw);
 | 
      
         | 4574 |  |  |   else
 | 
      
         | 4575 |  |  |     return emit_pattern_before_noloc (pattern, before,
 | 
      
         | 4576 |  |  |                                       insnp ? before : NULL_RTX,
 | 
      
         | 4577 |  |  |                                       NULL, make_raw);
 | 
      
         | 4578 |  |  | }
 | 
      
         | 4579 |  |  |  
 | 
      
         | 4580 |  |  | /* Like emit_insn_before_noloc, but set INSN_LOCATOR according to LOC.  */
 | 
      
         | 4581 |  |  | rtx
 | 
      
         | 4582 |  |  | emit_insn_before_setloc (rtx pattern, rtx before, int loc)
 | 
      
         | 4583 |  |  | {
 | 
      
         | 4584 |  |  |   return emit_pattern_before_setloc (pattern, before, loc, true,
 | 
      
         | 4585 |  |  |                                      make_insn_raw);
 | 
      
         | 4586 |  |  | }
 | 
      
         | 4587 |  |  |  
 | 
      
         | 4588 |  |  | /* Like emit_insn_before_noloc, but set INSN_LOCATOR according to BEFORE.  */
 | 
      
         | 4589 |  |  | rtx
 | 
      
         | 4590 |  |  | emit_insn_before (rtx pattern, rtx before)
 | 
      
         | 4591 |  |  | {
 | 
      
         | 4592 |  |  |   return emit_pattern_before (pattern, before, true, true, make_insn_raw);
 | 
      
         | 4593 |  |  | }
 | 
      
         | 4594 |  |  |  
 | 
      
         | 4595 |  |  | /* like emit_insn_before_noloc, but set INSN_LOCATOR according to LOC.  */
 | 
      
         | 4596 |  |  | rtx
 | 
      
         | 4597 |  |  | emit_jump_insn_before_setloc (rtx pattern, rtx before, int loc)
 | 
      
         | 4598 |  |  | {
 | 
      
         | 4599 |  |  |   return emit_pattern_before_setloc (pattern, before, loc, false,
 | 
      
         | 4600 |  |  |                                      make_jump_insn_raw);
 | 
      
         | 4601 |  |  | }
 | 
      
         | 4602 |  |  |  
 | 
      
         | 4603 |  |  | /* Like emit_jump_insn_before_noloc, but set INSN_LOCATOR according to BEFORE.  */
 | 
      
         | 4604 |  |  | rtx
 | 
      
         | 4605 |  |  | emit_jump_insn_before (rtx pattern, rtx before)
 | 
      
         | 4606 |  |  | {
 | 
      
         | 4607 |  |  |   return emit_pattern_before (pattern, before, true, false,
 | 
      
         | 4608 |  |  |                               make_jump_insn_raw);
 | 
      
         | 4609 |  |  | }
 | 
      
         | 4610 |  |  |  
 | 
      
         | 4611 |  |  | /* Like emit_insn_before_noloc, but set INSN_LOCATOR according to LOC.  */
 | 
      
         | 4612 |  |  | rtx
 | 
      
         | 4613 |  |  | emit_call_insn_before_setloc (rtx pattern, rtx before, int loc)
 | 
      
         | 4614 |  |  | {
 | 
      
         | 4615 |  |  |   return emit_pattern_before_setloc (pattern, before, loc, false,
 | 
      
         | 4616 |  |  |                                      make_call_insn_raw);
 | 
      
         | 4617 |  |  | }
 | 
      
         | 4618 |  |  |  
 | 
      
         | 4619 |  |  | /* Like emit_call_insn_before_noloc,
 | 
      
         | 4620 |  |  |    but set insn_locator according to BEFORE.  */
 | 
      
         | 4621 |  |  | rtx
 | 
      
         | 4622 |  |  | emit_call_insn_before (rtx pattern, rtx before)
 | 
      
         | 4623 |  |  | {
 | 
      
         | 4624 |  |  |   return emit_pattern_before (pattern, before, true, false,
 | 
      
         | 4625 |  |  |                               make_call_insn_raw);
 | 
      
         | 4626 |  |  | }
 | 
      
         | 4627 |  |  |  
 | 
      
         | 4628 |  |  | /* Like emit_insn_before_noloc, but set INSN_LOCATOR according to LOC.  */
 | 
      
         | 4629 |  |  | rtx
 | 
      
         | 4630 |  |  | emit_debug_insn_before_setloc (rtx pattern, rtx before, int loc)
 | 
      
         | 4631 |  |  | {
 | 
      
         | 4632 |  |  |   return emit_pattern_before_setloc (pattern, before, loc, false,
 | 
      
         | 4633 |  |  |                                      make_debug_insn_raw);
 | 
      
         | 4634 |  |  | }
 | 
      
         | 4635 |  |  |  
 | 
      
         | 4636 |  |  | /* Like emit_debug_insn_before_noloc,
 | 
      
         | 4637 |  |  |    but set insn_locator according to BEFORE.  */
 | 
      
         | 4638 |  |  | rtx
 | 
      
         | 4639 |  |  | emit_debug_insn_before (rtx pattern, rtx before)
 | 
      
         | 4640 |  |  | {
 | 
      
         | 4641 |  |  |   return emit_pattern_before (pattern, before, false, false,
 | 
      
         | 4642 |  |  |                               make_debug_insn_raw);
 | 
      
         | 4643 |  |  | }
 | 
      
         | 4644 |  |  |  
 | 
      
         | 4645 |  |  | /* Take X and emit it at the end of the doubly-linked
 | 
      
         | 4646 |  |  |    INSN list.
 | 
      
         | 4647 |  |  |  
 | 
      
         | 4648 |  |  |    Returns the last insn emitted.  */
 | 
      
         | 4649 |  |  |  
 | 
      
         | 4650 |  |  | rtx
 | 
      
         | 4651 |  |  | emit_insn (rtx x)
 | 
      
         | 4652 |  |  | {
 | 
      
         | 4653 |  |  |   rtx last = get_last_insn();
 | 
      
         | 4654 |  |  |   rtx insn;
 | 
      
         | 4655 |  |  |  
 | 
      
         | 4656 |  |  |   if (x == NULL_RTX)
 | 
      
         | 4657 |  |  |     return last;
 | 
      
         | 4658 |  |  |  
 | 
      
         | 4659 |  |  |   switch (GET_CODE (x))
 | 
      
         | 4660 |  |  |     {
 | 
      
         | 4661 |  |  |     case DEBUG_INSN:
 | 
      
         | 4662 |  |  |     case INSN:
 | 
      
         | 4663 |  |  |     case JUMP_INSN:
 | 
      
         | 4664 |  |  |     case CALL_INSN:
 | 
      
         | 4665 |  |  |     case CODE_LABEL:
 | 
      
         | 4666 |  |  |     case BARRIER:
 | 
      
         | 4667 |  |  |     case NOTE:
 | 
      
         | 4668 |  |  |       insn = x;
 | 
      
         | 4669 |  |  |       while (insn)
 | 
      
         | 4670 |  |  |         {
 | 
      
         | 4671 |  |  |           rtx next = NEXT_INSN (insn);
 | 
      
         | 4672 |  |  |           add_insn (insn);
 | 
      
         | 4673 |  |  |           last = insn;
 | 
      
         | 4674 |  |  |           insn = next;
 | 
      
         | 4675 |  |  |         }
 | 
      
         | 4676 |  |  |       break;
 | 
      
         | 4677 |  |  |  
 | 
      
         | 4678 |  |  | #ifdef ENABLE_RTL_CHECKING
 | 
      
         | 4679 |  |  |     case SEQUENCE:
 | 
      
         | 4680 |  |  |       gcc_unreachable ();
 | 
      
         | 4681 |  |  |       break;
 | 
      
         | 4682 |  |  | #endif
 | 
      
         | 4683 |  |  |  
 | 
      
         | 4684 |  |  |     default:
 | 
      
         | 4685 |  |  |       last = make_insn_raw (x);
 | 
      
         | 4686 |  |  |       add_insn (last);
 | 
      
         | 4687 |  |  |       break;
 | 
      
         | 4688 |  |  |     }
 | 
      
         | 4689 |  |  |  
 | 
      
         | 4690 |  |  |   return last;
 | 
      
         | 4691 |  |  | }
 | 
      
         | 4692 |  |  |  
 | 
      
         | 4693 |  |  | /* Make an insn of code DEBUG_INSN with pattern X
 | 
      
         | 4694 |  |  |    and add it to the end of the doubly-linked list.  */
 | 
      
         | 4695 |  |  |  
 | 
      
         | 4696 |  |  | rtx
 | 
      
         | 4697 |  |  | emit_debug_insn (rtx x)
 | 
      
         | 4698 |  |  | {
 | 
      
         | 4699 |  |  |   rtx last = get_last_insn();
 | 
      
         | 4700 |  |  |   rtx insn;
 | 
      
         | 4701 |  |  |  
 | 
      
         | 4702 |  |  |   if (x == NULL_RTX)
 | 
      
         | 4703 |  |  |     return last;
 | 
      
         | 4704 |  |  |  
 | 
      
         | 4705 |  |  |   switch (GET_CODE (x))
 | 
      
         | 4706 |  |  |     {
 | 
      
         | 4707 |  |  |     case DEBUG_INSN:
 | 
      
         | 4708 |  |  |     case INSN:
 | 
      
         | 4709 |  |  |     case JUMP_INSN:
 | 
      
         | 4710 |  |  |     case CALL_INSN:
 | 
      
         | 4711 |  |  |     case CODE_LABEL:
 | 
      
         | 4712 |  |  |     case BARRIER:
 | 
      
         | 4713 |  |  |     case NOTE:
 | 
      
         | 4714 |  |  |       insn = x;
 | 
      
         | 4715 |  |  |       while (insn)
 | 
      
         | 4716 |  |  |         {
 | 
      
         | 4717 |  |  |           rtx next = NEXT_INSN (insn);
 | 
      
         | 4718 |  |  |           add_insn (insn);
 | 
      
         | 4719 |  |  |           last = insn;
 | 
      
         | 4720 |  |  |           insn = next;
 | 
      
         | 4721 |  |  |         }
 | 
      
         | 4722 |  |  |       break;
 | 
      
         | 4723 |  |  |  
 | 
      
         | 4724 |  |  | #ifdef ENABLE_RTL_CHECKING
 | 
      
         | 4725 |  |  |     case SEQUENCE:
 | 
      
         | 4726 |  |  |       gcc_unreachable ();
 | 
      
         | 4727 |  |  |       break;
 | 
      
         | 4728 |  |  | #endif
 | 
      
         | 4729 |  |  |  
 | 
      
         | 4730 |  |  |     default:
 | 
      
         | 4731 |  |  |       last = make_debug_insn_raw (x);
 | 
      
         | 4732 |  |  |       add_insn (last);
 | 
      
         | 4733 |  |  |       break;
 | 
      
         | 4734 |  |  |     }
 | 
      
         | 4735 |  |  |  
 | 
      
         | 4736 |  |  |   return last;
 | 
      
         | 4737 |  |  | }
 | 
      
         | 4738 |  |  |  
 | 
      
         | 4739 |  |  | /* Make an insn of code JUMP_INSN with pattern X
 | 
      
         | 4740 |  |  |    and add it to the end of the doubly-linked list.  */
 | 
      
         | 4741 |  |  |  
 | 
      
         | 4742 |  |  | rtx
 | 
      
         | 4743 |  |  | emit_jump_insn (rtx x)
 | 
      
         | 4744 |  |  | {
 | 
      
         | 4745 |  |  |   rtx last = NULL_RTX, insn;
 | 
      
         | 4746 |  |  |  
 | 
      
         | 4747 |  |  |   switch (GET_CODE (x))
 | 
      
         | 4748 |  |  |     {
 | 
      
         | 4749 |  |  |     case DEBUG_INSN:
 | 
      
         | 4750 |  |  |     case INSN:
 | 
      
         | 4751 |  |  |     case JUMP_INSN:
 | 
      
         | 4752 |  |  |     case CALL_INSN:
 | 
      
         | 4753 |  |  |     case CODE_LABEL:
 | 
      
         | 4754 |  |  |     case BARRIER:
 | 
      
         | 4755 |  |  |     case NOTE:
 | 
      
         | 4756 |  |  |       insn = x;
 | 
      
         | 4757 |  |  |       while (insn)
 | 
      
         | 4758 |  |  |         {
 | 
      
         | 4759 |  |  |           rtx next = NEXT_INSN (insn);
 | 
      
         | 4760 |  |  |           add_insn (insn);
 | 
      
         | 4761 |  |  |           last = insn;
 | 
      
         | 4762 |  |  |           insn = next;
 | 
      
         | 4763 |  |  |         }
 | 
      
         | 4764 |  |  |       break;
 | 
      
         | 4765 |  |  |  
 | 
      
         | 4766 |  |  | #ifdef ENABLE_RTL_CHECKING
 | 
      
         | 4767 |  |  |     case SEQUENCE:
 | 
      
         | 4768 |  |  |       gcc_unreachable ();
 | 
      
         | 4769 |  |  |       break;
 | 
      
         | 4770 |  |  | #endif
 | 
      
         | 4771 |  |  |  
 | 
      
         | 4772 |  |  |     default:
 | 
      
         | 4773 |  |  |       last = make_jump_insn_raw (x);
 | 
      
         | 4774 |  |  |       add_insn (last);
 | 
      
         | 4775 |  |  |       break;
 | 
      
         | 4776 |  |  |     }
 | 
      
         | 4777 |  |  |  
 | 
      
         | 4778 |  |  |   return last;
 | 
      
         | 4779 |  |  | }
 | 
      
         | 4780 |  |  |  
 | 
      
         | 4781 |  |  | /* Make an insn of code CALL_INSN with pattern X
 | 
      
         | 4782 |  |  |    and add it to the end of the doubly-linked list.  */
 | 
      
         | 4783 |  |  |  
 | 
      
         | 4784 |  |  | rtx
 | 
      
         | 4785 |  |  | emit_call_insn (rtx x)
 | 
      
         | 4786 |  |  | {
 | 
      
         | 4787 |  |  |   rtx insn;
 | 
      
         | 4788 |  |  |  
 | 
      
         | 4789 |  |  |   switch (GET_CODE (x))
 | 
      
         | 4790 |  |  |     {
 | 
      
         | 4791 |  |  |     case DEBUG_INSN:
 | 
      
         | 4792 |  |  |     case INSN:
 | 
      
         | 4793 |  |  |     case JUMP_INSN:
 | 
      
         | 4794 |  |  |     case CALL_INSN:
 | 
      
         | 4795 |  |  |     case CODE_LABEL:
 | 
      
         | 4796 |  |  |     case BARRIER:
 | 
      
         | 4797 |  |  |     case NOTE:
 | 
      
         | 4798 |  |  |       insn = emit_insn (x);
 | 
      
         | 4799 |  |  |       break;
 | 
      
         | 4800 |  |  |  
 | 
      
         | 4801 |  |  | #ifdef ENABLE_RTL_CHECKING
 | 
      
         | 4802 |  |  |     case SEQUENCE:
 | 
      
         | 4803 |  |  |       gcc_unreachable ();
 | 
      
         | 4804 |  |  |       break;
 | 
      
         | 4805 |  |  | #endif
 | 
      
         | 4806 |  |  |  
 | 
      
         | 4807 |  |  |     default:
 | 
      
         | 4808 |  |  |       insn = make_call_insn_raw (x);
 | 
      
         | 4809 |  |  |       add_insn (insn);
 | 
      
         | 4810 |  |  |       break;
 | 
      
         | 4811 |  |  |     }
 | 
      
         | 4812 |  |  |  
 | 
      
         | 4813 |  |  |   return insn;
 | 
      
         | 4814 |  |  | }
 | 
      
         | 4815 |  |  |  
 | 
      
         | 4816 |  |  | /* Add the label LABEL to the end of the doubly-linked list.  */
 | 
      
         | 4817 |  |  |  
 | 
      
         | 4818 |  |  | rtx
 | 
      
         | 4819 |  |  | emit_label (rtx label)
 | 
      
         | 4820 |  |  | {
 | 
      
         | 4821 |  |  |   /* This can be called twice for the same label
 | 
      
         | 4822 |  |  |      as a result of the confusion that follows a syntax error!
 | 
      
         | 4823 |  |  |      So make it harmless.  */
 | 
      
         | 4824 |  |  |   if (INSN_UID (label) == 0)
 | 
      
         | 4825 |  |  |     {
 | 
      
         | 4826 |  |  |       INSN_UID (label) = cur_insn_uid++;
 | 
      
         | 4827 |  |  |       add_insn (label);
 | 
      
         | 4828 |  |  |     }
 | 
      
         | 4829 |  |  |   return label;
 | 
      
         | 4830 |  |  | }
 | 
      
         | 4831 |  |  |  
 | 
      
         | 4832 |  |  | /* Make an insn of code BARRIER
 | 
      
         | 4833 |  |  |    and add it to the end of the doubly-linked list.  */
 | 
      
         | 4834 |  |  |  
 | 
      
         | 4835 |  |  | rtx
 | 
      
         | 4836 |  |  | emit_barrier (void)
 | 
      
         | 4837 |  |  | {
 | 
      
         | 4838 |  |  |   rtx barrier = rtx_alloc (BARRIER);
 | 
      
         | 4839 |  |  |   INSN_UID (barrier) = cur_insn_uid++;
 | 
      
         | 4840 |  |  |   add_insn (barrier);
 | 
      
         | 4841 |  |  |   return barrier;
 | 
      
         | 4842 |  |  | }
 | 
      
         | 4843 |  |  |  
 | 
      
         | 4844 |  |  | /* Emit a copy of note ORIG.  */
 | 
      
         | 4845 |  |  |  
 | 
      
         | 4846 |  |  | rtx
 | 
      
         | 4847 |  |  | emit_note_copy (rtx orig)
 | 
      
         | 4848 |  |  | {
 | 
      
         | 4849 |  |  |   rtx note;
 | 
      
         | 4850 |  |  |  
 | 
      
         | 4851 |  |  |   note = rtx_alloc (NOTE);
 | 
      
         | 4852 |  |  |  
 | 
      
         | 4853 |  |  |   INSN_UID (note) = cur_insn_uid++;
 | 
      
         | 4854 |  |  |   NOTE_DATA (note) = NOTE_DATA (orig);
 | 
      
         | 4855 |  |  |   NOTE_KIND (note) = NOTE_KIND (orig);
 | 
      
         | 4856 |  |  |   BLOCK_FOR_INSN (note) = NULL;
 | 
      
         | 4857 |  |  |   add_insn (note);
 | 
      
         | 4858 |  |  |  
 | 
      
         | 4859 |  |  |   return note;
 | 
      
         | 4860 |  |  | }
 | 
      
         | 4861 |  |  |  
 | 
      
         | 4862 |  |  | /* Make an insn of code NOTE or type NOTE_NO
 | 
      
         | 4863 |  |  |    and add it to the end of the doubly-linked list.  */
 | 
      
         | 4864 |  |  |  
 | 
      
         | 4865 |  |  | rtx
 | 
      
         | 4866 |  |  | emit_note (enum insn_note kind)
 | 
      
         | 4867 |  |  | {
 | 
      
         | 4868 |  |  |   rtx note;
 | 
      
         | 4869 |  |  |  
 | 
      
         | 4870 |  |  |   note = rtx_alloc (NOTE);
 | 
      
         | 4871 |  |  |   INSN_UID (note) = cur_insn_uid++;
 | 
      
         | 4872 |  |  |   NOTE_KIND (note) = kind;
 | 
      
         | 4873 |  |  |   memset (&NOTE_DATA (note), 0, sizeof (NOTE_DATA (note)));
 | 
      
         | 4874 |  |  |   BLOCK_FOR_INSN (note) = NULL;
 | 
      
         | 4875 |  |  |   add_insn (note);
 | 
      
         | 4876 |  |  |   return note;
 | 
      
         | 4877 |  |  | }
 | 
      
         | 4878 |  |  |  
 | 
      
         | 4879 |  |  | /* Emit a clobber of lvalue X.  */
 | 
      
         | 4880 |  |  |  
 | 
      
         | 4881 |  |  | rtx
 | 
      
         | 4882 |  |  | emit_clobber (rtx x)
 | 
      
         | 4883 |  |  | {
 | 
      
         | 4884 |  |  |   /* CONCATs should not appear in the insn stream.  */
 | 
      
         | 4885 |  |  |   if (GET_CODE (x) == CONCAT)
 | 
      
         | 4886 |  |  |     {
 | 
      
         | 4887 |  |  |       emit_clobber (XEXP (x, 0));
 | 
      
         | 4888 |  |  |       return emit_clobber (XEXP (x, 1));
 | 
      
         | 4889 |  |  |     }
 | 
      
         | 4890 |  |  |   return emit_insn (gen_rtx_CLOBBER (VOIDmode, x));
 | 
      
         | 4891 |  |  | }
 | 
      
         | 4892 |  |  |  
 | 
      
         | 4893 |  |  | /* Return a sequence of insns to clobber lvalue X.  */
 | 
      
         | 4894 |  |  |  
 | 
      
         | 4895 |  |  | rtx
 | 
      
         | 4896 |  |  | gen_clobber (rtx x)
 | 
      
         | 4897 |  |  | {
 | 
      
         | 4898 |  |  |   rtx seq;
 | 
      
         | 4899 |  |  |  
 | 
      
         | 4900 |  |  |   start_sequence ();
 | 
      
         | 4901 |  |  |   emit_clobber (x);
 | 
      
         | 4902 |  |  |   seq = get_insns ();
 | 
      
         | 4903 |  |  |   end_sequence ();
 | 
      
         | 4904 |  |  |   return seq;
 | 
      
         | 4905 |  |  | }
 | 
      
         | 4906 |  |  |  
 | 
      
         | 4907 |  |  | /* Emit a use of rvalue X.  */
 | 
      
         | 4908 |  |  |  
 | 
      
         | 4909 |  |  | rtx
 | 
      
         | 4910 |  |  | emit_use (rtx x)
 | 
      
         | 4911 |  |  | {
 | 
      
         | 4912 |  |  |   /* CONCATs should not appear in the insn stream.  */
 | 
      
         | 4913 |  |  |   if (GET_CODE (x) == CONCAT)
 | 
      
         | 4914 |  |  |     {
 | 
      
         | 4915 |  |  |       emit_use (XEXP (x, 0));
 | 
      
         | 4916 |  |  |       return emit_use (XEXP (x, 1));
 | 
      
         | 4917 |  |  |     }
 | 
      
         | 4918 |  |  |   return emit_insn (gen_rtx_USE (VOIDmode, x));
 | 
      
         | 4919 |  |  | }
 | 
      
         | 4920 |  |  |  
 | 
      
         | 4921 |  |  | /* Return a sequence of insns to use rvalue X.  */
 | 
      
         | 4922 |  |  |  
 | 
      
         | 4923 |  |  | rtx
 | 
      
         | 4924 |  |  | gen_use (rtx x)
 | 
      
         | 4925 |  |  | {
 | 
      
         | 4926 |  |  |   rtx seq;
 | 
      
         | 4927 |  |  |  
 | 
      
         | 4928 |  |  |   start_sequence ();
 | 
      
         | 4929 |  |  |   emit_use (x);
 | 
      
         | 4930 |  |  |   seq = get_insns ();
 | 
      
         | 4931 |  |  |   end_sequence ();
 | 
      
         | 4932 |  |  |   return seq;
 | 
      
         | 4933 |  |  | }
 | 
      
         | 4934 |  |  |  
 | 
      
         | 4935 |  |  | /* Cause next statement to emit a line note even if the line number
 | 
      
         | 4936 |  |  |    has not changed.  */
 | 
      
         | 4937 |  |  |  
 | 
      
         | 4938 |  |  | void
 | 
      
         | 4939 |  |  | force_next_line_note (void)
 | 
      
         | 4940 |  |  | {
 | 
      
         | 4941 |  |  |   last_location = -1;
 | 
      
         | 4942 |  |  | }
 | 
      
         | 4943 |  |  |  
 | 
      
         | 4944 |  |  | /* Place a note of KIND on insn INSN with DATUM as the datum. If a
 | 
      
         | 4945 |  |  |    note of this type already exists, remove it first.  */
 | 
      
         | 4946 |  |  |  
 | 
      
         | 4947 |  |  | rtx
 | 
      
         | 4948 |  |  | set_unique_reg_note (rtx insn, enum reg_note kind, rtx datum)
 | 
      
         | 4949 |  |  | {
 | 
      
         | 4950 |  |  |   rtx note = find_reg_note (insn, kind, NULL_RTX);
 | 
      
         | 4951 |  |  |  
 | 
      
         | 4952 |  |  |   switch (kind)
 | 
      
         | 4953 |  |  |     {
 | 
      
         | 4954 |  |  |     case REG_EQUAL:
 | 
      
         | 4955 |  |  |     case REG_EQUIV:
 | 
      
         | 4956 |  |  |       /* Don't add REG_EQUAL/REG_EQUIV notes if the insn
 | 
      
         | 4957 |  |  |          has multiple sets (some callers assume single_set
 | 
      
         | 4958 |  |  |          means the insn only has one set, when in fact it
 | 
      
         | 4959 |  |  |          means the insn only has one * useful * set).  */
 | 
      
         | 4960 |  |  |       if (GET_CODE (PATTERN (insn)) == PARALLEL && multiple_sets (insn))
 | 
      
         | 4961 |  |  |         {
 | 
      
         | 4962 |  |  |           gcc_assert (!note);
 | 
      
         | 4963 |  |  |           return NULL_RTX;
 | 
      
         | 4964 |  |  |         }
 | 
      
         | 4965 |  |  |  
 | 
      
         | 4966 |  |  |       /* Don't add ASM_OPERAND REG_EQUAL/REG_EQUIV notes.
 | 
      
         | 4967 |  |  |          It serves no useful purpose and breaks eliminate_regs.  */
 | 
      
         | 4968 |  |  |       if (GET_CODE (datum) == ASM_OPERANDS)
 | 
      
         | 4969 |  |  |         return NULL_RTX;
 | 
      
         | 4970 |  |  |  
 | 
      
         | 4971 |  |  |       if (note)
 | 
      
         | 4972 |  |  |         {
 | 
      
         | 4973 |  |  |           XEXP (note, 0) = datum;
 | 
      
         | 4974 |  |  |           df_notes_rescan (insn);
 | 
      
         | 4975 |  |  |           return note;
 | 
      
         | 4976 |  |  |         }
 | 
      
         | 4977 |  |  |       break;
 | 
      
         | 4978 |  |  |  
 | 
      
         | 4979 |  |  |     default:
 | 
      
         | 4980 |  |  |       if (note)
 | 
      
         | 4981 |  |  |         {
 | 
      
         | 4982 |  |  |           XEXP (note, 0) = datum;
 | 
      
         | 4983 |  |  |           return note;
 | 
      
         | 4984 |  |  |         }
 | 
      
         | 4985 |  |  |       break;
 | 
      
         | 4986 |  |  |     }
 | 
      
         | 4987 |  |  |  
 | 
      
         | 4988 |  |  |   add_reg_note (insn, kind, datum);
 | 
      
         | 4989 |  |  |  
 | 
      
         | 4990 |  |  |   switch (kind)
 | 
      
         | 4991 |  |  |     {
 | 
      
         | 4992 |  |  |     case REG_EQUAL:
 | 
      
         | 4993 |  |  |     case REG_EQUIV:
 | 
      
         | 4994 |  |  |       df_notes_rescan (insn);
 | 
      
         | 4995 |  |  |       break;
 | 
      
         | 4996 |  |  |     default:
 | 
      
         | 4997 |  |  |       break;
 | 
      
         | 4998 |  |  |     }
 | 
      
         | 4999 |  |  |  
 | 
      
         | 5000 |  |  |   return REG_NOTES (insn);
 | 
      
         | 5001 |  |  | }
 | 
      
         | 5002 |  |  |  
 | 
      
         | 5003 |  |  | /* Like set_unique_reg_note, but don't do anything unless INSN sets DST.  */
 | 
      
         | 5004 |  |  | rtx
 | 
      
         | 5005 |  |  | set_dst_reg_note (rtx insn, enum reg_note kind, rtx datum, rtx dst)
 | 
      
         | 5006 |  |  | {
 | 
      
         | 5007 |  |  |   rtx set = single_set (insn);
 | 
      
         | 5008 |  |  |  
 | 
      
         | 5009 |  |  |   if (set && SET_DEST (set) == dst)
 | 
      
         | 5010 |  |  |     return set_unique_reg_note (insn, kind, datum);
 | 
      
         | 5011 |  |  |   return NULL_RTX;
 | 
      
         | 5012 |  |  | }
 | 
      
         | 5013 |  |  |  
 | 
      
         | 5014 |  |  | /* Return an indication of which type of insn should have X as a body.
 | 
      
         | 5015 |  |  |    The value is CODE_LABEL, INSN, CALL_INSN or JUMP_INSN.  */
 | 
      
         | 5016 |  |  |  
 | 
      
         | 5017 |  |  | static enum rtx_code
 | 
      
         | 5018 |  |  | classify_insn (rtx x)
 | 
      
         | 5019 |  |  | {
 | 
      
         | 5020 |  |  |   if (LABEL_P (x))
 | 
      
         | 5021 |  |  |     return CODE_LABEL;
 | 
      
         | 5022 |  |  |   if (GET_CODE (x) == CALL)
 | 
      
         | 5023 |  |  |     return CALL_INSN;
 | 
      
         | 5024 |  |  |   if (ANY_RETURN_P (x))
 | 
      
         | 5025 |  |  |     return JUMP_INSN;
 | 
      
         | 5026 |  |  |   if (GET_CODE (x) == SET)
 | 
      
         | 5027 |  |  |     {
 | 
      
         | 5028 |  |  |       if (SET_DEST (x) == pc_rtx)
 | 
      
         | 5029 |  |  |         return JUMP_INSN;
 | 
      
         | 5030 |  |  |       else if (GET_CODE (SET_SRC (x)) == CALL)
 | 
      
         | 5031 |  |  |         return CALL_INSN;
 | 
      
         | 5032 |  |  |       else
 | 
      
         | 5033 |  |  |         return INSN;
 | 
      
         | 5034 |  |  |     }
 | 
      
         | 5035 |  |  |   if (GET_CODE (x) == PARALLEL)
 | 
      
         | 5036 |  |  |     {
 | 
      
         | 5037 |  |  |       int j;
 | 
      
         | 5038 |  |  |       for (j = XVECLEN (x, 0) - 1; j >= 0; j--)
 | 
      
         | 5039 |  |  |         if (GET_CODE (XVECEXP (x, 0, j)) == CALL)
 | 
      
         | 5040 |  |  |           return CALL_INSN;
 | 
      
         | 5041 |  |  |         else if (GET_CODE (XVECEXP (x, 0, j)) == SET
 | 
      
         | 5042 |  |  |                  && SET_DEST (XVECEXP (x, 0, j)) == pc_rtx)
 | 
      
         | 5043 |  |  |           return JUMP_INSN;
 | 
      
         | 5044 |  |  |         else if (GET_CODE (XVECEXP (x, 0, j)) == SET
 | 
      
         | 5045 |  |  |                  && GET_CODE (SET_SRC (XVECEXP (x, 0, j))) == CALL)
 | 
      
         | 5046 |  |  |           return CALL_INSN;
 | 
      
         | 5047 |  |  |     }
 | 
      
         | 5048 |  |  |   return INSN;
 | 
      
         | 5049 |  |  | }
 | 
      
         | 5050 |  |  |  
 | 
      
         | 5051 |  |  | /* Emit the rtl pattern X as an appropriate kind of insn.
 | 
      
         | 5052 |  |  |    If X is a label, it is simply added into the insn chain.  */
 | 
      
         | 5053 |  |  |  
 | 
      
         | 5054 |  |  | rtx
 | 
      
         | 5055 |  |  | emit (rtx x)
 | 
      
         | 5056 |  |  | {
 | 
      
         | 5057 |  |  |   enum rtx_code code = classify_insn (x);
 | 
      
         | 5058 |  |  |  
 | 
      
         | 5059 |  |  |   switch (code)
 | 
      
         | 5060 |  |  |     {
 | 
      
         | 5061 |  |  |     case CODE_LABEL:
 | 
      
         | 5062 |  |  |       return emit_label (x);
 | 
      
         | 5063 |  |  |     case INSN:
 | 
      
         | 5064 |  |  |       return emit_insn (x);
 | 
      
         | 5065 |  |  |     case  JUMP_INSN:
 | 
      
         | 5066 |  |  |       {
 | 
      
         | 5067 |  |  |         rtx insn = emit_jump_insn (x);
 | 
      
         | 5068 |  |  |         if (any_uncondjump_p (insn) || GET_CODE (x) == RETURN)
 | 
      
         | 5069 |  |  |           return emit_barrier ();
 | 
      
         | 5070 |  |  |         return insn;
 | 
      
         | 5071 |  |  |       }
 | 
      
         | 5072 |  |  |     case CALL_INSN:
 | 
      
         | 5073 |  |  |       return emit_call_insn (x);
 | 
      
         | 5074 |  |  |     case DEBUG_INSN:
 | 
      
         | 5075 |  |  |       return emit_debug_insn (x);
 | 
      
         | 5076 |  |  |     default:
 | 
      
         | 5077 |  |  |       gcc_unreachable ();
 | 
      
         | 5078 |  |  |     }
 | 
      
         | 5079 |  |  | }
 | 
      
         | 5080 |  |  |  
 | 
      
         | 5081 |  |  | /* Space for free sequence stack entries.  */
 | 
      
         | 5082 |  |  | static GTY ((deletable)) struct sequence_stack *free_sequence_stack;
 | 
      
         | 5083 |  |  |  
 | 
      
         | 5084 |  |  | /* Begin emitting insns to a sequence.  If this sequence will contain
 | 
      
         | 5085 |  |  |    something that might cause the compiler to pop arguments to function
 | 
      
         | 5086 |  |  |    calls (because those pops have previously been deferred; see
 | 
      
         | 5087 |  |  |    INHIBIT_DEFER_POP for more details), use do_pending_stack_adjust
 | 
      
         | 5088 |  |  |    before calling this function.  That will ensure that the deferred
 | 
      
         | 5089 |  |  |    pops are not accidentally emitted in the middle of this sequence.  */
 | 
      
         | 5090 |  |  |  
 | 
      
         | 5091 |  |  | void
 | 
      
         | 5092 |  |  | start_sequence (void)
 | 
      
         | 5093 |  |  | {
 | 
      
         | 5094 |  |  |   struct sequence_stack *tem;
 | 
      
         | 5095 |  |  |  
 | 
      
         | 5096 |  |  |   if (free_sequence_stack != NULL)
 | 
      
         | 5097 |  |  |     {
 | 
      
         | 5098 |  |  |       tem = free_sequence_stack;
 | 
      
         | 5099 |  |  |       free_sequence_stack = tem->next;
 | 
      
         | 5100 |  |  |     }
 | 
      
         | 5101 |  |  |   else
 | 
      
         | 5102 |  |  |     tem = ggc_alloc_sequence_stack ();
 | 
      
         | 5103 |  |  |  
 | 
      
         | 5104 |  |  |   tem->next = seq_stack;
 | 
      
         | 5105 |  |  |   tem->first = get_insns ();
 | 
      
         | 5106 |  |  |   tem->last = get_last_insn ();
 | 
      
         | 5107 |  |  |  
 | 
      
         | 5108 |  |  |   seq_stack = tem;
 | 
      
         | 5109 |  |  |  
 | 
      
         | 5110 |  |  |   set_first_insn (0);
 | 
      
         | 5111 |  |  |   set_last_insn (0);
 | 
      
         | 5112 |  |  | }
 | 
      
         | 5113 |  |  |  
 | 
      
         | 5114 |  |  | /* Set up the insn chain starting with FIRST as the current sequence,
 | 
      
         | 5115 |  |  |    saving the previously current one.  See the documentation for
 | 
      
         | 5116 |  |  |    start_sequence for more information about how to use this function.  */
 | 
      
         | 5117 |  |  |  
 | 
      
         | 5118 |  |  | void
 | 
      
         | 5119 |  |  | push_to_sequence (rtx first)
 | 
      
         | 5120 |  |  | {
 | 
      
         | 5121 |  |  |   rtx last;
 | 
      
         | 5122 |  |  |  
 | 
      
         | 5123 |  |  |   start_sequence ();
 | 
      
         | 5124 |  |  |  
 | 
      
         | 5125 |  |  |   for (last = first; last && NEXT_INSN (last); last = NEXT_INSN (last))
 | 
      
         | 5126 |  |  |     ;
 | 
      
         | 5127 |  |  |  
 | 
      
         | 5128 |  |  |   set_first_insn (first);
 | 
      
         | 5129 |  |  |   set_last_insn (last);
 | 
      
         | 5130 |  |  | }
 | 
      
         | 5131 |  |  |  
 | 
      
         | 5132 |  |  | /* Like push_to_sequence, but take the last insn as an argument to avoid
 | 
      
         | 5133 |  |  |    looping through the list.  */
 | 
      
         | 5134 |  |  |  
 | 
      
         | 5135 |  |  | void
 | 
      
         | 5136 |  |  | push_to_sequence2 (rtx first, rtx last)
 | 
      
         | 5137 |  |  | {
 | 
      
         | 5138 |  |  |   start_sequence ();
 | 
      
         | 5139 |  |  |  
 | 
      
         | 5140 |  |  |   set_first_insn (first);
 | 
      
         | 5141 |  |  |   set_last_insn (last);
 | 
      
         | 5142 |  |  | }
 | 
      
         | 5143 |  |  |  
 | 
      
         | 5144 |  |  | /* Set up the outer-level insn chain
 | 
      
         | 5145 |  |  |    as the current sequence, saving the previously current one.  */
 | 
      
         | 5146 |  |  |  
 | 
      
         | 5147 |  |  | void
 | 
      
         | 5148 |  |  | push_topmost_sequence (void)
 | 
      
         | 5149 |  |  | {
 | 
      
         | 5150 |  |  |   struct sequence_stack *stack, *top = NULL;
 | 
      
         | 5151 |  |  |  
 | 
      
         | 5152 |  |  |   start_sequence ();
 | 
      
         | 5153 |  |  |  
 | 
      
         | 5154 |  |  |   for (stack = seq_stack; stack; stack = stack->next)
 | 
      
         | 5155 |  |  |     top = stack;
 | 
      
         | 5156 |  |  |  
 | 
      
         | 5157 |  |  |   set_first_insn (top->first);
 | 
      
         | 5158 |  |  |   set_last_insn (top->last);
 | 
      
         | 5159 |  |  | }
 | 
      
         | 5160 |  |  |  
 | 
      
         | 5161 |  |  | /* After emitting to the outer-level insn chain, update the outer-level
 | 
      
         | 5162 |  |  |    insn chain, and restore the previous saved state.  */
 | 
      
         | 5163 |  |  |  
 | 
      
         | 5164 |  |  | void
 | 
      
         | 5165 |  |  | pop_topmost_sequence (void)
 | 
      
         | 5166 |  |  | {
 | 
      
         | 5167 |  |  |   struct sequence_stack *stack, *top = NULL;
 | 
      
         | 5168 |  |  |  
 | 
      
         | 5169 |  |  |   for (stack = seq_stack; stack; stack = stack->next)
 | 
      
         | 5170 |  |  |     top = stack;
 | 
      
         | 5171 |  |  |  
 | 
      
         | 5172 |  |  |   top->first = get_insns ();
 | 
      
         | 5173 |  |  |   top->last = get_last_insn ();
 | 
      
         | 5174 |  |  |  
 | 
      
         | 5175 |  |  |   end_sequence ();
 | 
      
         | 5176 |  |  | }
 | 
      
         | 5177 |  |  |  
 | 
      
         | 5178 |  |  | /* After emitting to a sequence, restore previous saved state.
 | 
      
         | 5179 |  |  |  
 | 
      
         | 5180 |  |  |    To get the contents of the sequence just made, you must call
 | 
      
         | 5181 |  |  |    `get_insns' *before* calling here.
 | 
      
         | 5182 |  |  |  
 | 
      
         | 5183 |  |  |    If the compiler might have deferred popping arguments while
 | 
      
         | 5184 |  |  |    generating this sequence, and this sequence will not be immediately
 | 
      
         | 5185 |  |  |    inserted into the instruction stream, use do_pending_stack_adjust
 | 
      
         | 5186 |  |  |    before calling get_insns.  That will ensure that the deferred
 | 
      
         | 5187 |  |  |    pops are inserted into this sequence, and not into some random
 | 
      
         | 5188 |  |  |    location in the instruction stream.  See INHIBIT_DEFER_POP for more
 | 
      
         | 5189 |  |  |    information about deferred popping of arguments.  */
 | 
      
         | 5190 |  |  |  
 | 
      
         | 5191 |  |  | void
 | 
      
         | 5192 |  |  | end_sequence (void)
 | 
      
         | 5193 |  |  | {
 | 
      
         | 5194 |  |  |   struct sequence_stack *tem = seq_stack;
 | 
      
         | 5195 |  |  |  
 | 
      
         | 5196 |  |  |   set_first_insn (tem->first);
 | 
      
         | 5197 |  |  |   set_last_insn (tem->last);
 | 
      
         | 5198 |  |  |   seq_stack = tem->next;
 | 
      
         | 5199 |  |  |  
 | 
      
         | 5200 |  |  |   memset (tem, 0, sizeof (*tem));
 | 
      
         | 5201 |  |  |   tem->next = free_sequence_stack;
 | 
      
         | 5202 |  |  |   free_sequence_stack = tem;
 | 
      
         | 5203 |  |  | }
 | 
      
         | 5204 |  |  |  
 | 
      
         | 5205 |  |  | /* Return 1 if currently emitting into a sequence.  */
 | 
      
         | 5206 |  |  |  
 | 
      
         | 5207 |  |  | int
 | 
      
         | 5208 |  |  | in_sequence_p (void)
 | 
      
         | 5209 |  |  | {
 | 
      
         | 5210 |  |  |   return seq_stack != 0;
 | 
      
         | 5211 |  |  | }
 | 
      
         | 5212 |  |  |  
 | 
      
         | 5213 |  |  | /* Put the various virtual registers into REGNO_REG_RTX.  */
 | 
      
         | 5214 |  |  |  
 | 
      
         | 5215 |  |  | static void
 | 
      
         | 5216 |  |  | init_virtual_regs (void)
 | 
      
         | 5217 |  |  | {
 | 
      
         | 5218 |  |  |   regno_reg_rtx[VIRTUAL_INCOMING_ARGS_REGNUM] = virtual_incoming_args_rtx;
 | 
      
         | 5219 |  |  |   regno_reg_rtx[VIRTUAL_STACK_VARS_REGNUM] = virtual_stack_vars_rtx;
 | 
      
         | 5220 |  |  |   regno_reg_rtx[VIRTUAL_STACK_DYNAMIC_REGNUM] = virtual_stack_dynamic_rtx;
 | 
      
         | 5221 |  |  |   regno_reg_rtx[VIRTUAL_OUTGOING_ARGS_REGNUM] = virtual_outgoing_args_rtx;
 | 
      
         | 5222 |  |  |   regno_reg_rtx[VIRTUAL_CFA_REGNUM] = virtual_cfa_rtx;
 | 
      
         | 5223 |  |  |   regno_reg_rtx[VIRTUAL_PREFERRED_STACK_BOUNDARY_REGNUM]
 | 
      
         | 5224 |  |  |     = virtual_preferred_stack_boundary_rtx;
 | 
      
         | 5225 |  |  | }
 | 
      
         | 5226 |  |  |  
 | 
      
         | 5227 |  |  |  
 | 
      
         | 5228 |  |  | /* Used by copy_insn_1 to avoid copying SCRATCHes more than once.  */
 | 
      
         | 5229 |  |  | static rtx copy_insn_scratch_in[MAX_RECOG_OPERANDS];
 | 
      
         | 5230 |  |  | static rtx copy_insn_scratch_out[MAX_RECOG_OPERANDS];
 | 
      
         | 5231 |  |  | static int copy_insn_n_scratches;
 | 
      
         | 5232 |  |  |  
 | 
      
         | 5233 |  |  | /* When an insn is being copied by copy_insn_1, this is nonzero if we have
 | 
      
         | 5234 |  |  |    copied an ASM_OPERANDS.
 | 
      
         | 5235 |  |  |    In that case, it is the original input-operand vector.  */
 | 
      
         | 5236 |  |  | static rtvec orig_asm_operands_vector;
 | 
      
         | 5237 |  |  |  
 | 
      
         | 5238 |  |  | /* When an insn is being copied by copy_insn_1, this is nonzero if we have
 | 
      
         | 5239 |  |  |    copied an ASM_OPERANDS.
 | 
      
         | 5240 |  |  |    In that case, it is the copied input-operand vector.  */
 | 
      
         | 5241 |  |  | static rtvec copy_asm_operands_vector;
 | 
      
         | 5242 |  |  |  
 | 
      
         | 5243 |  |  | /* Likewise for the constraints vector.  */
 | 
      
         | 5244 |  |  | static rtvec orig_asm_constraints_vector;
 | 
      
         | 5245 |  |  | static rtvec copy_asm_constraints_vector;
 | 
      
         | 5246 |  |  |  
 | 
      
         | 5247 |  |  | /* Recursively create a new copy of an rtx for copy_insn.
 | 
      
         | 5248 |  |  |    This function differs from copy_rtx in that it handles SCRATCHes and
 | 
      
         | 5249 |  |  |    ASM_OPERANDs properly.
 | 
      
         | 5250 |  |  |    Normally, this function is not used directly; use copy_insn as front end.
 | 
      
         | 5251 |  |  |    However, you could first copy an insn pattern with copy_insn and then use
 | 
      
         | 5252 |  |  |    this function afterwards to properly copy any REG_NOTEs containing
 | 
      
         | 5253 |  |  |    SCRATCHes.  */
 | 
      
         | 5254 |  |  |  
 | 
      
         | 5255 |  |  | rtx
 | 
      
         | 5256 |  |  | copy_insn_1 (rtx orig)
 | 
      
         | 5257 |  |  | {
 | 
      
         | 5258 |  |  |   rtx copy;
 | 
      
         | 5259 |  |  |   int i, j;
 | 
      
         | 5260 |  |  |   RTX_CODE code;
 | 
      
         | 5261 |  |  |   const char *format_ptr;
 | 
      
         | 5262 |  |  |  
 | 
      
         | 5263 |  |  |   if (orig == NULL)
 | 
      
         | 5264 |  |  |     return NULL;
 | 
      
         | 5265 |  |  |  
 | 
      
         | 5266 |  |  |   code = GET_CODE (orig);
 | 
      
         | 5267 |  |  |  
 | 
      
         | 5268 |  |  |   switch (code)
 | 
      
         | 5269 |  |  |     {
 | 
      
         | 5270 |  |  |     case REG:
 | 
      
         | 5271 |  |  |     case DEBUG_EXPR:
 | 
      
         | 5272 |  |  |     case CONST_INT:
 | 
      
         | 5273 |  |  |     case CONST_DOUBLE:
 | 
      
         | 5274 |  |  |     case CONST_FIXED:
 | 
      
         | 5275 |  |  |     case CONST_VECTOR:
 | 
      
         | 5276 |  |  |     case SYMBOL_REF:
 | 
      
         | 5277 |  |  |     case CODE_LABEL:
 | 
      
         | 5278 |  |  |     case PC:
 | 
      
         | 5279 |  |  |     case CC0:
 | 
      
         | 5280 |  |  |     case RETURN:
 | 
      
         | 5281 |  |  |     case SIMPLE_RETURN:
 | 
      
         | 5282 |  |  |       return orig;
 | 
      
         | 5283 |  |  |     case CLOBBER:
 | 
      
         | 5284 |  |  |       if (REG_P (XEXP (orig, 0)) && REGNO (XEXP (orig, 0)) < FIRST_PSEUDO_REGISTER)
 | 
      
         | 5285 |  |  |         return orig;
 | 
      
         | 5286 |  |  |       break;
 | 
      
         | 5287 |  |  |  
 | 
      
         | 5288 |  |  |     case SCRATCH:
 | 
      
         | 5289 |  |  |       for (i = 0; i < copy_insn_n_scratches; i++)
 | 
      
         | 5290 |  |  |         if (copy_insn_scratch_in[i] == orig)
 | 
      
         | 5291 |  |  |           return copy_insn_scratch_out[i];
 | 
      
         | 5292 |  |  |       break;
 | 
      
         | 5293 |  |  |  
 | 
      
         | 5294 |  |  |     case CONST:
 | 
      
         | 5295 |  |  |       if (shared_const_p (orig))
 | 
      
         | 5296 |  |  |         return orig;
 | 
      
         | 5297 |  |  |       break;
 | 
      
         | 5298 |  |  |  
 | 
      
         | 5299 |  |  |       /* A MEM with a constant address is not sharable.  The problem is that
 | 
      
         | 5300 |  |  |          the constant address may need to be reloaded.  If the mem is shared,
 | 
      
         | 5301 |  |  |          then reloading one copy of this mem will cause all copies to appear
 | 
      
         | 5302 |  |  |          to have been reloaded.  */
 | 
      
         | 5303 |  |  |  
 | 
      
         | 5304 |  |  |     default:
 | 
      
         | 5305 |  |  |       break;
 | 
      
         | 5306 |  |  |     }
 | 
      
         | 5307 |  |  |  
 | 
      
         | 5308 |  |  |   /* Copy the various flags, fields, and other information.  We assume
 | 
      
         | 5309 |  |  |      that all fields need copying, and then clear the fields that should
 | 
      
         | 5310 |  |  |      not be copied.  That is the sensible default behavior, and forces
 | 
      
         | 5311 |  |  |      us to explicitly document why we are *not* copying a flag.  */
 | 
      
         | 5312 |  |  |   copy = shallow_copy_rtx (orig);
 | 
      
         | 5313 |  |  |  
 | 
      
         | 5314 |  |  |   /* We do not copy the USED flag, which is used as a mark bit during
 | 
      
         | 5315 |  |  |      walks over the RTL.  */
 | 
      
         | 5316 |  |  |   RTX_FLAG (copy, used) = 0;
 | 
      
         | 5317 |  |  |  
 | 
      
         | 5318 |  |  |   /* We do not copy JUMP, CALL, or FRAME_RELATED for INSNs.  */
 | 
      
         | 5319 |  |  |   if (INSN_P (orig))
 | 
      
         | 5320 |  |  |     {
 | 
      
         | 5321 |  |  |       RTX_FLAG (copy, jump) = 0;
 | 
      
         | 5322 |  |  |       RTX_FLAG (copy, call) = 0;
 | 
      
         | 5323 |  |  |       RTX_FLAG (copy, frame_related) = 0;
 | 
      
         | 5324 |  |  |     }
 | 
      
         | 5325 |  |  |  
 | 
      
         | 5326 |  |  |   format_ptr = GET_RTX_FORMAT (GET_CODE (copy));
 | 
      
         | 5327 |  |  |  
 | 
      
         | 5328 |  |  |   for (i = 0; i < GET_RTX_LENGTH (GET_CODE (copy)); i++)
 | 
      
         | 5329 |  |  |     switch (*format_ptr++)
 | 
      
         | 5330 |  |  |       {
 | 
      
         | 5331 |  |  |       case 'e':
 | 
      
         | 5332 |  |  |         if (XEXP (orig, i) != NULL)
 | 
      
         | 5333 |  |  |           XEXP (copy, i) = copy_insn_1 (XEXP (orig, i));
 | 
      
         | 5334 |  |  |         break;
 | 
      
         | 5335 |  |  |  
 | 
      
         | 5336 |  |  |       case 'E':
 | 
      
         | 5337 |  |  |       case 'V':
 | 
      
         | 5338 |  |  |         if (XVEC (orig, i) == orig_asm_constraints_vector)
 | 
      
         | 5339 |  |  |           XVEC (copy, i) = copy_asm_constraints_vector;
 | 
      
         | 5340 |  |  |         else if (XVEC (orig, i) == orig_asm_operands_vector)
 | 
      
         | 5341 |  |  |           XVEC (copy, i) = copy_asm_operands_vector;
 | 
      
         | 5342 |  |  |         else if (XVEC (orig, i) != NULL)
 | 
      
         | 5343 |  |  |           {
 | 
      
         | 5344 |  |  |             XVEC (copy, i) = rtvec_alloc (XVECLEN (orig, i));
 | 
      
         | 5345 |  |  |             for (j = 0; j < XVECLEN (copy, i); j++)
 | 
      
         | 5346 |  |  |               XVECEXP (copy, i, j) = copy_insn_1 (XVECEXP (orig, i, j));
 | 
      
         | 5347 |  |  |           }
 | 
      
         | 5348 |  |  |         break;
 | 
      
         | 5349 |  |  |  
 | 
      
         | 5350 |  |  |       case 't':
 | 
      
         | 5351 |  |  |       case 'w':
 | 
      
         | 5352 |  |  |       case 'i':
 | 
      
         | 5353 |  |  |       case 's':
 | 
      
         | 5354 |  |  |       case 'S':
 | 
      
         | 5355 |  |  |       case 'u':
 | 
      
         | 5356 |  |  |       case '0':
 | 
      
         | 5357 |  |  |         /* These are left unchanged.  */
 | 
      
         | 5358 |  |  |         break;
 | 
      
         | 5359 |  |  |  
 | 
      
         | 5360 |  |  |       default:
 | 
      
         | 5361 |  |  |         gcc_unreachable ();
 | 
      
         | 5362 |  |  |       }
 | 
      
         | 5363 |  |  |  
 | 
      
         | 5364 |  |  |   if (code == SCRATCH)
 | 
      
         | 5365 |  |  |     {
 | 
      
         | 5366 |  |  |       i = copy_insn_n_scratches++;
 | 
      
         | 5367 |  |  |       gcc_assert (i < MAX_RECOG_OPERANDS);
 | 
      
         | 5368 |  |  |       copy_insn_scratch_in[i] = orig;
 | 
      
         | 5369 |  |  |       copy_insn_scratch_out[i] = copy;
 | 
      
         | 5370 |  |  |     }
 | 
      
         | 5371 |  |  |   else if (code == ASM_OPERANDS)
 | 
      
         | 5372 |  |  |     {
 | 
      
         | 5373 |  |  |       orig_asm_operands_vector = ASM_OPERANDS_INPUT_VEC (orig);
 | 
      
         | 5374 |  |  |       copy_asm_operands_vector = ASM_OPERANDS_INPUT_VEC (copy);
 | 
      
         | 5375 |  |  |       orig_asm_constraints_vector = ASM_OPERANDS_INPUT_CONSTRAINT_VEC (orig);
 | 
      
         | 5376 |  |  |       copy_asm_constraints_vector = ASM_OPERANDS_INPUT_CONSTRAINT_VEC (copy);
 | 
      
         | 5377 |  |  |     }
 | 
      
         | 5378 |  |  |  
 | 
      
         | 5379 |  |  |   return copy;
 | 
      
         | 5380 |  |  | }
 | 
      
         | 5381 |  |  |  
 | 
      
         | 5382 |  |  | /* Create a new copy of an rtx.
 | 
      
         | 5383 |  |  |    This function differs from copy_rtx in that it handles SCRATCHes and
 | 
      
         | 5384 |  |  |    ASM_OPERANDs properly.
 | 
      
         | 5385 |  |  |    INSN doesn't really have to be a full INSN; it could be just the
 | 
      
         | 5386 |  |  |    pattern.  */
 | 
      
         | 5387 |  |  | rtx
 | 
      
         | 5388 |  |  | copy_insn (rtx insn)
 | 
      
         | 5389 |  |  | {
 | 
      
         | 5390 |  |  |   copy_insn_n_scratches = 0;
 | 
      
         | 5391 |  |  |   orig_asm_operands_vector = 0;
 | 
      
         | 5392 |  |  |   orig_asm_constraints_vector = 0;
 | 
      
         | 5393 |  |  |   copy_asm_operands_vector = 0;
 | 
      
         | 5394 |  |  |   copy_asm_constraints_vector = 0;
 | 
      
         | 5395 |  |  |   return copy_insn_1 (insn);
 | 
      
         | 5396 |  |  | }
 | 
      
         | 5397 |  |  |  
 | 
      
         | 5398 |  |  | /* Initialize data structures and variables in this file
 | 
      
         | 5399 |  |  |    before generating rtl for each function.  */
 | 
      
         | 5400 |  |  |  
 | 
      
         | 5401 |  |  | void
 | 
      
         | 5402 |  |  | init_emit (void)
 | 
      
         | 5403 |  |  | {
 | 
      
         | 5404 |  |  |   set_first_insn (NULL);
 | 
      
         | 5405 |  |  |   set_last_insn (NULL);
 | 
      
         | 5406 |  |  |   if (MIN_NONDEBUG_INSN_UID)
 | 
      
         | 5407 |  |  |     cur_insn_uid = MIN_NONDEBUG_INSN_UID;
 | 
      
         | 5408 |  |  |   else
 | 
      
         | 5409 |  |  |     cur_insn_uid = 1;
 | 
      
         | 5410 |  |  |   cur_debug_insn_uid = 1;
 | 
      
         | 5411 |  |  |   reg_rtx_no = LAST_VIRTUAL_REGISTER + 1;
 | 
      
         | 5412 |  |  |   last_location = UNKNOWN_LOCATION;
 | 
      
         | 5413 |  |  |   first_label_num = label_num;
 | 
      
         | 5414 |  |  |   seq_stack = NULL;
 | 
      
         | 5415 |  |  |  
 | 
      
         | 5416 |  |  |   /* Init the tables that describe all the pseudo regs.  */
 | 
      
         | 5417 |  |  |  
 | 
      
         | 5418 |  |  |   crtl->emit.regno_pointer_align_length = LAST_VIRTUAL_REGISTER + 101;
 | 
      
         | 5419 |  |  |  
 | 
      
         | 5420 |  |  |   crtl->emit.regno_pointer_align
 | 
      
         | 5421 |  |  |     = XCNEWVEC (unsigned char, crtl->emit.regno_pointer_align_length);
 | 
      
         | 5422 |  |  |  
 | 
      
         | 5423 |  |  |   regno_reg_rtx = ggc_alloc_vec_rtx (crtl->emit.regno_pointer_align_length);
 | 
      
         | 5424 |  |  |  
 | 
      
         | 5425 |  |  |   /* Put copies of all the hard registers into regno_reg_rtx.  */
 | 
      
         | 5426 |  |  |   memcpy (regno_reg_rtx,
 | 
      
         | 5427 |  |  |           initial_regno_reg_rtx,
 | 
      
         | 5428 |  |  |           FIRST_PSEUDO_REGISTER * sizeof (rtx));
 | 
      
         | 5429 |  |  |  
 | 
      
         | 5430 |  |  |   /* Put copies of all the virtual register rtx into regno_reg_rtx.  */
 | 
      
         | 5431 |  |  |   init_virtual_regs ();
 | 
      
         | 5432 |  |  |  
 | 
      
         | 5433 |  |  |   /* Indicate that the virtual registers and stack locations are
 | 
      
         | 5434 |  |  |      all pointers.  */
 | 
      
         | 5435 |  |  |   REG_POINTER (stack_pointer_rtx) = 1;
 | 
      
         | 5436 |  |  |   REG_POINTER (frame_pointer_rtx) = 1;
 | 
      
         | 5437 |  |  |   REG_POINTER (hard_frame_pointer_rtx) = 1;
 | 
      
         | 5438 |  |  |   REG_POINTER (arg_pointer_rtx) = 1;
 | 
      
         | 5439 |  |  |  
 | 
      
         | 5440 |  |  |   REG_POINTER (virtual_incoming_args_rtx) = 1;
 | 
      
         | 5441 |  |  |   REG_POINTER (virtual_stack_vars_rtx) = 1;
 | 
      
         | 5442 |  |  |   REG_POINTER (virtual_stack_dynamic_rtx) = 1;
 | 
      
         | 5443 |  |  |   REG_POINTER (virtual_outgoing_args_rtx) = 1;
 | 
      
         | 5444 |  |  |   REG_POINTER (virtual_cfa_rtx) = 1;
 | 
      
         | 5445 |  |  |  
 | 
      
         | 5446 |  |  | #ifdef STACK_BOUNDARY
 | 
      
         | 5447 |  |  |   REGNO_POINTER_ALIGN (STACK_POINTER_REGNUM) = STACK_BOUNDARY;
 | 
      
         | 5448 |  |  |   REGNO_POINTER_ALIGN (FRAME_POINTER_REGNUM) = STACK_BOUNDARY;
 | 
      
         | 5449 |  |  |   REGNO_POINTER_ALIGN (HARD_FRAME_POINTER_REGNUM) = STACK_BOUNDARY;
 | 
      
         | 5450 |  |  |   REGNO_POINTER_ALIGN (ARG_POINTER_REGNUM) = STACK_BOUNDARY;
 | 
      
         | 5451 |  |  |  
 | 
      
         | 5452 |  |  |   REGNO_POINTER_ALIGN (VIRTUAL_INCOMING_ARGS_REGNUM) = STACK_BOUNDARY;
 | 
      
         | 5453 |  |  |   REGNO_POINTER_ALIGN (VIRTUAL_STACK_VARS_REGNUM) = STACK_BOUNDARY;
 | 
      
         | 5454 |  |  |   REGNO_POINTER_ALIGN (VIRTUAL_STACK_DYNAMIC_REGNUM) = STACK_BOUNDARY;
 | 
      
         | 5455 |  |  |   REGNO_POINTER_ALIGN (VIRTUAL_OUTGOING_ARGS_REGNUM) = STACK_BOUNDARY;
 | 
      
         | 5456 |  |  |   REGNO_POINTER_ALIGN (VIRTUAL_CFA_REGNUM) = BITS_PER_WORD;
 | 
      
         | 5457 |  |  | #endif
 | 
      
         | 5458 |  |  |  
 | 
      
         | 5459 |  |  | #ifdef INIT_EXPANDERS
 | 
      
         | 5460 |  |  |   INIT_EXPANDERS;
 | 
      
         | 5461 |  |  | #endif
 | 
      
         | 5462 |  |  | }
 | 
      
         | 5463 |  |  |  
 | 
      
         | 5464 |  |  | /* Generate a vector constant for mode MODE and constant value CONSTANT.  */
 | 
      
         | 5465 |  |  |  
 | 
      
         | 5466 |  |  | static rtx
 | 
      
         | 5467 |  |  | gen_const_vector (enum machine_mode mode, int constant)
 | 
      
         | 5468 |  |  | {
 | 
      
         | 5469 |  |  |   rtx tem;
 | 
      
         | 5470 |  |  |   rtvec v;
 | 
      
         | 5471 |  |  |   int units, i;
 | 
      
         | 5472 |  |  |   enum machine_mode inner;
 | 
      
         | 5473 |  |  |  
 | 
      
         | 5474 |  |  |   units = GET_MODE_NUNITS (mode);
 | 
      
         | 5475 |  |  |   inner = GET_MODE_INNER (mode);
 | 
      
         | 5476 |  |  |  
 | 
      
         | 5477 |  |  |   gcc_assert (!DECIMAL_FLOAT_MODE_P (inner));
 | 
      
         | 5478 |  |  |  
 | 
      
         | 5479 |  |  |   v = rtvec_alloc (units);
 | 
      
         | 5480 |  |  |  
 | 
      
         | 5481 |  |  |   /* We need to call this function after we set the scalar const_tiny_rtx
 | 
      
         | 5482 |  |  |      entries.  */
 | 
      
         | 5483 |  |  |   gcc_assert (const_tiny_rtx[constant][(int) inner]);
 | 
      
         | 5484 |  |  |  
 | 
      
         | 5485 |  |  |   for (i = 0; i < units; ++i)
 | 
      
         | 5486 |  |  |     RTVEC_ELT (v, i) = const_tiny_rtx[constant][(int) inner];
 | 
      
         | 5487 |  |  |  
 | 
      
         | 5488 |  |  |   tem = gen_rtx_raw_CONST_VECTOR (mode, v);
 | 
      
         | 5489 |  |  |   return tem;
 | 
      
         | 5490 |  |  | }
 | 
      
         | 5491 |  |  |  
 | 
      
         | 5492 |  |  | /* Generate a vector like gen_rtx_raw_CONST_VEC, but use the zero vector when
 | 
      
         | 5493 |  |  |    all elements are zero, and the one vector when all elements are one.  */
 | 
      
         | 5494 |  |  | rtx
 | 
      
         | 5495 |  |  | gen_rtx_CONST_VECTOR (enum machine_mode mode, rtvec v)
 | 
      
         | 5496 |  |  | {
 | 
      
         | 5497 |  |  |   enum machine_mode inner = GET_MODE_INNER (mode);
 | 
      
         | 5498 |  |  |   int nunits = GET_MODE_NUNITS (mode);
 | 
      
         | 5499 |  |  |   rtx x;
 | 
      
         | 5500 |  |  |   int i;
 | 
      
         | 5501 |  |  |  
 | 
      
         | 5502 |  |  |   /* Check to see if all of the elements have the same value.  */
 | 
      
         | 5503 |  |  |   x = RTVEC_ELT (v, nunits - 1);
 | 
      
         | 5504 |  |  |   for (i = nunits - 2; i >= 0; i--)
 | 
      
         | 5505 |  |  |     if (RTVEC_ELT (v, i) != x)
 | 
      
         | 5506 |  |  |       break;
 | 
      
         | 5507 |  |  |  
 | 
      
         | 5508 |  |  |   /* If the values are all the same, check to see if we can use one of the
 | 
      
         | 5509 |  |  |      standard constant vectors.  */
 | 
      
         | 5510 |  |  |   if (i == -1)
 | 
      
         | 5511 |  |  |     {
 | 
      
         | 5512 |  |  |       if (x == CONST0_RTX (inner))
 | 
      
         | 5513 |  |  |         return CONST0_RTX (mode);
 | 
      
         | 5514 |  |  |       else if (x == CONST1_RTX (inner))
 | 
      
         | 5515 |  |  |         return CONST1_RTX (mode);
 | 
      
         | 5516 |  |  |       else if (x == CONSTM1_RTX (inner))
 | 
      
         | 5517 |  |  |         return CONSTM1_RTX (mode);
 | 
      
         | 5518 |  |  |     }
 | 
      
         | 5519 |  |  |  
 | 
      
         | 5520 |  |  |   return gen_rtx_raw_CONST_VECTOR (mode, v);
 | 
      
         | 5521 |  |  | }
 | 
      
         | 5522 |  |  |  
 | 
      
         | 5523 |  |  | /* Initialise global register information required by all functions.  */
 | 
      
         | 5524 |  |  |  
 | 
      
         | 5525 |  |  | void
 | 
      
         | 5526 |  |  | init_emit_regs (void)
 | 
      
         | 5527 |  |  | {
 | 
      
         | 5528 |  |  |   int i;
 | 
      
         | 5529 |  |  |   enum machine_mode mode;
 | 
      
         | 5530 |  |  |   mem_attrs *attrs;
 | 
      
         | 5531 |  |  |  
 | 
      
         | 5532 |  |  |   /* Reset register attributes */
 | 
      
         | 5533 |  |  |   htab_empty (reg_attrs_htab);
 | 
      
         | 5534 |  |  |  
 | 
      
         | 5535 |  |  |   /* We need reg_raw_mode, so initialize the modes now.  */
 | 
      
         | 5536 |  |  |   init_reg_modes_target ();
 | 
      
         | 5537 |  |  |  
 | 
      
         | 5538 |  |  |   /* Assign register numbers to the globally defined register rtx.  */
 | 
      
         | 5539 |  |  |   pc_rtx = gen_rtx_fmt_ (PC, VOIDmode);
 | 
      
         | 5540 |  |  |   ret_rtx = gen_rtx_fmt_ (RETURN, VOIDmode);
 | 
      
         | 5541 |  |  |   simple_return_rtx = gen_rtx_fmt_ (SIMPLE_RETURN, VOIDmode);
 | 
      
         | 5542 |  |  |   cc0_rtx = gen_rtx_fmt_ (CC0, VOIDmode);
 | 
      
         | 5543 |  |  |   stack_pointer_rtx = gen_raw_REG (Pmode, STACK_POINTER_REGNUM);
 | 
      
         | 5544 |  |  |   frame_pointer_rtx = gen_raw_REG (Pmode, FRAME_POINTER_REGNUM);
 | 
      
         | 5545 |  |  |   hard_frame_pointer_rtx = gen_raw_REG (Pmode, HARD_FRAME_POINTER_REGNUM);
 | 
      
         | 5546 |  |  |   arg_pointer_rtx = gen_raw_REG (Pmode, ARG_POINTER_REGNUM);
 | 
      
         | 5547 |  |  |   virtual_incoming_args_rtx =
 | 
      
         | 5548 |  |  |     gen_raw_REG (Pmode, VIRTUAL_INCOMING_ARGS_REGNUM);
 | 
      
         | 5549 |  |  |   virtual_stack_vars_rtx =
 | 
      
         | 5550 |  |  |     gen_raw_REG (Pmode, VIRTUAL_STACK_VARS_REGNUM);
 | 
      
         | 5551 |  |  |   virtual_stack_dynamic_rtx =
 | 
      
         | 5552 |  |  |     gen_raw_REG (Pmode, VIRTUAL_STACK_DYNAMIC_REGNUM);
 | 
      
         | 5553 |  |  |   virtual_outgoing_args_rtx =
 | 
      
         | 5554 |  |  |     gen_raw_REG (Pmode, VIRTUAL_OUTGOING_ARGS_REGNUM);
 | 
      
         | 5555 |  |  |   virtual_cfa_rtx = gen_raw_REG (Pmode, VIRTUAL_CFA_REGNUM);
 | 
      
         | 5556 |  |  |   virtual_preferred_stack_boundary_rtx =
 | 
      
         | 5557 |  |  |     gen_raw_REG (Pmode, VIRTUAL_PREFERRED_STACK_BOUNDARY_REGNUM);
 | 
      
         | 5558 |  |  |  
 | 
      
         | 5559 |  |  |   /* Initialize RTL for commonly used hard registers.  These are
 | 
      
         | 5560 |  |  |      copied into regno_reg_rtx as we begin to compile each function.  */
 | 
      
         | 5561 |  |  |   for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
 | 
      
         | 5562 |  |  |     initial_regno_reg_rtx[i] = gen_raw_REG (reg_raw_mode[i], i);
 | 
      
         | 5563 |  |  |  
 | 
      
         | 5564 |  |  | #ifdef RETURN_ADDRESS_POINTER_REGNUM
 | 
      
         | 5565 |  |  |   return_address_pointer_rtx
 | 
      
         | 5566 |  |  |     = gen_raw_REG (Pmode, RETURN_ADDRESS_POINTER_REGNUM);
 | 
      
         | 5567 |  |  | #endif
 | 
      
         | 5568 |  |  |  
 | 
      
         | 5569 |  |  |   if ((unsigned) PIC_OFFSET_TABLE_REGNUM != INVALID_REGNUM)
 | 
      
         | 5570 |  |  |     pic_offset_table_rtx = gen_raw_REG (Pmode, PIC_OFFSET_TABLE_REGNUM);
 | 
      
         | 5571 |  |  |   else
 | 
      
         | 5572 |  |  |     pic_offset_table_rtx = NULL_RTX;
 | 
      
         | 5573 |  |  |  
 | 
      
         | 5574 |  |  |   for (i = 0; i < (int) MAX_MACHINE_MODE; i++)
 | 
      
         | 5575 |  |  |     {
 | 
      
         | 5576 |  |  |       mode = (enum machine_mode) i;
 | 
      
         | 5577 |  |  |       attrs = ggc_alloc_cleared_mem_attrs ();
 | 
      
         | 5578 |  |  |       attrs->align = BITS_PER_UNIT;
 | 
      
         | 5579 |  |  |       attrs->addrspace = ADDR_SPACE_GENERIC;
 | 
      
         | 5580 |  |  |       if (mode != BLKmode)
 | 
      
         | 5581 |  |  |         {
 | 
      
         | 5582 |  |  |           attrs->size_known_p = true;
 | 
      
         | 5583 |  |  |           attrs->size = GET_MODE_SIZE (mode);
 | 
      
         | 5584 |  |  |           if (STRICT_ALIGNMENT)
 | 
      
         | 5585 |  |  |             attrs->align = GET_MODE_ALIGNMENT (mode);
 | 
      
         | 5586 |  |  |         }
 | 
      
         | 5587 |  |  |       mode_mem_attrs[i] = attrs;
 | 
      
         | 5588 |  |  |     }
 | 
      
         | 5589 |  |  | }
 | 
      
         | 5590 |  |  |  
 | 
      
         | 5591 |  |  | /* Create some permanent unique rtl objects shared between all functions.  */
 | 
      
         | 5592 |  |  |  
 | 
      
         | 5593 |  |  | void
 | 
      
         | 5594 |  |  | init_emit_once (void)
 | 
      
         | 5595 |  |  | {
 | 
      
         | 5596 |  |  |   int i;
 | 
      
         | 5597 |  |  |   enum machine_mode mode;
 | 
      
         | 5598 |  |  |   enum machine_mode double_mode;
 | 
      
         | 5599 |  |  |  
 | 
      
         | 5600 |  |  |   /* Initialize the CONST_INT, CONST_DOUBLE, CONST_FIXED, and memory attribute
 | 
      
         | 5601 |  |  |      hash tables.  */
 | 
      
         | 5602 |  |  |   const_int_htab = htab_create_ggc (37, const_int_htab_hash,
 | 
      
         | 5603 |  |  |                                     const_int_htab_eq, NULL);
 | 
      
         | 5604 |  |  |  
 | 
      
         | 5605 |  |  |   const_double_htab = htab_create_ggc (37, const_double_htab_hash,
 | 
      
         | 5606 |  |  |                                        const_double_htab_eq, NULL);
 | 
      
         | 5607 |  |  |  
 | 
      
         | 5608 |  |  |   const_fixed_htab = htab_create_ggc (37, const_fixed_htab_hash,
 | 
      
         | 5609 |  |  |                                       const_fixed_htab_eq, NULL);
 | 
      
         | 5610 |  |  |  
 | 
      
         | 5611 |  |  |   mem_attrs_htab = htab_create_ggc (37, mem_attrs_htab_hash,
 | 
      
         | 5612 |  |  |                                     mem_attrs_htab_eq, NULL);
 | 
      
         | 5613 |  |  |   reg_attrs_htab = htab_create_ggc (37, reg_attrs_htab_hash,
 | 
      
         | 5614 |  |  |                                     reg_attrs_htab_eq, NULL);
 | 
      
         | 5615 |  |  |  
 | 
      
         | 5616 |  |  |   /* Compute the word and byte modes.  */
 | 
      
         | 5617 |  |  |  
 | 
      
         | 5618 |  |  |   byte_mode = VOIDmode;
 | 
      
         | 5619 |  |  |   word_mode = VOIDmode;
 | 
      
         | 5620 |  |  |   double_mode = VOIDmode;
 | 
      
         | 5621 |  |  |  
 | 
      
         | 5622 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_INT);
 | 
      
         | 5623 |  |  |        mode != VOIDmode;
 | 
      
         | 5624 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5625 |  |  |     {
 | 
      
         | 5626 |  |  |       if (GET_MODE_BITSIZE (mode) == BITS_PER_UNIT
 | 
      
         | 5627 |  |  |           && byte_mode == VOIDmode)
 | 
      
         | 5628 |  |  |         byte_mode = mode;
 | 
      
         | 5629 |  |  |  
 | 
      
         | 5630 |  |  |       if (GET_MODE_BITSIZE (mode) == BITS_PER_WORD
 | 
      
         | 5631 |  |  |           && word_mode == VOIDmode)
 | 
      
         | 5632 |  |  |         word_mode = mode;
 | 
      
         | 5633 |  |  |     }
 | 
      
         | 5634 |  |  |  
 | 
      
         | 5635 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_FLOAT);
 | 
      
         | 5636 |  |  |        mode != VOIDmode;
 | 
      
         | 5637 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5638 |  |  |     {
 | 
      
         | 5639 |  |  |       if (GET_MODE_BITSIZE (mode) == DOUBLE_TYPE_SIZE
 | 
      
         | 5640 |  |  |           && double_mode == VOIDmode)
 | 
      
         | 5641 |  |  |         double_mode = mode;
 | 
      
         | 5642 |  |  |     }
 | 
      
         | 5643 |  |  |  
 | 
      
         | 5644 |  |  |   ptr_mode = mode_for_size (POINTER_SIZE, GET_MODE_CLASS (Pmode), 0);
 | 
      
         | 5645 |  |  |  
 | 
      
         | 5646 |  |  | #ifdef INIT_EXPANDERS
 | 
      
         | 5647 |  |  |   /* This is to initialize {init|mark|free}_machine_status before the first
 | 
      
         | 5648 |  |  |      call to push_function_context_to.  This is needed by the Chill front
 | 
      
         | 5649 |  |  |      end which calls push_function_context_to before the first call to
 | 
      
         | 5650 |  |  |      init_function_start.  */
 | 
      
         | 5651 |  |  |   INIT_EXPANDERS;
 | 
      
         | 5652 |  |  | #endif
 | 
      
         | 5653 |  |  |  
 | 
      
         | 5654 |  |  |   /* Create the unique rtx's for certain rtx codes and operand values.  */
 | 
      
         | 5655 |  |  |  
 | 
      
         | 5656 |  |  |   /* Don't use gen_rtx_CONST_INT here since gen_rtx_CONST_INT in this case
 | 
      
         | 5657 |  |  |      tries to use these variables.  */
 | 
      
         | 5658 |  |  |   for (i = - MAX_SAVED_CONST_INT; i <= MAX_SAVED_CONST_INT; i++)
 | 
      
         | 5659 |  |  |     const_int_rtx[i + MAX_SAVED_CONST_INT] =
 | 
      
         | 5660 |  |  |       gen_rtx_raw_CONST_INT (VOIDmode, (HOST_WIDE_INT) i);
 | 
      
         | 5661 |  |  |  
 | 
      
         | 5662 |  |  |   if (STORE_FLAG_VALUE >= - MAX_SAVED_CONST_INT
 | 
      
         | 5663 |  |  |       && STORE_FLAG_VALUE <= MAX_SAVED_CONST_INT)
 | 
      
         | 5664 |  |  |     const_true_rtx = const_int_rtx[STORE_FLAG_VALUE + MAX_SAVED_CONST_INT];
 | 
      
         | 5665 |  |  |   else
 | 
      
         | 5666 |  |  |     const_true_rtx = gen_rtx_CONST_INT (VOIDmode, STORE_FLAG_VALUE);
 | 
      
         | 5667 |  |  |  
 | 
      
         | 5668 |  |  |   REAL_VALUE_FROM_INT (dconst0,   0,  0, double_mode);
 | 
      
         | 5669 |  |  |   REAL_VALUE_FROM_INT (dconst1,   1,  0, double_mode);
 | 
      
         | 5670 |  |  |   REAL_VALUE_FROM_INT (dconst2,   2,  0, double_mode);
 | 
      
         | 5671 |  |  |  
 | 
      
         | 5672 |  |  |   dconstm1 = dconst1;
 | 
      
         | 5673 |  |  |   dconstm1.sign = 1;
 | 
      
         | 5674 |  |  |  
 | 
      
         | 5675 |  |  |   dconsthalf = dconst1;
 | 
      
         | 5676 |  |  |   SET_REAL_EXP (&dconsthalf, REAL_EXP (&dconsthalf) - 1);
 | 
      
         | 5677 |  |  |  
 | 
      
         | 5678 |  |  |   for (i = 0; i < 3; i++)
 | 
      
         | 5679 |  |  |     {
 | 
      
         | 5680 |  |  |       const REAL_VALUE_TYPE *const r =
 | 
      
         | 5681 |  |  |         (i == 0 ? &dconst0 : i == 1 ? &dconst1 : &dconst2);
 | 
      
         | 5682 |  |  |  
 | 
      
         | 5683 |  |  |       for (mode = GET_CLASS_NARROWEST_MODE (MODE_FLOAT);
 | 
      
         | 5684 |  |  |            mode != VOIDmode;
 | 
      
         | 5685 |  |  |            mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5686 |  |  |         const_tiny_rtx[i][(int) mode] =
 | 
      
         | 5687 |  |  |           CONST_DOUBLE_FROM_REAL_VALUE (*r, mode);
 | 
      
         | 5688 |  |  |  
 | 
      
         | 5689 |  |  |       for (mode = GET_CLASS_NARROWEST_MODE (MODE_DECIMAL_FLOAT);
 | 
      
         | 5690 |  |  |            mode != VOIDmode;
 | 
      
         | 5691 |  |  |            mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5692 |  |  |         const_tiny_rtx[i][(int) mode] =
 | 
      
         | 5693 |  |  |           CONST_DOUBLE_FROM_REAL_VALUE (*r, mode);
 | 
      
         | 5694 |  |  |  
 | 
      
         | 5695 |  |  |       const_tiny_rtx[i][(int) VOIDmode] = GEN_INT (i);
 | 
      
         | 5696 |  |  |  
 | 
      
         | 5697 |  |  |       for (mode = GET_CLASS_NARROWEST_MODE (MODE_INT);
 | 
      
         | 5698 |  |  |            mode != VOIDmode;
 | 
      
         | 5699 |  |  |            mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5700 |  |  |         const_tiny_rtx[i][(int) mode] = GEN_INT (i);
 | 
      
         | 5701 |  |  |  
 | 
      
         | 5702 |  |  |       for (mode = GET_CLASS_NARROWEST_MODE (MODE_PARTIAL_INT);
 | 
      
         | 5703 |  |  |            mode != VOIDmode;
 | 
      
         | 5704 |  |  |            mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5705 |  |  |         const_tiny_rtx[i][(int) mode] = GEN_INT (i);
 | 
      
         | 5706 |  |  |     }
 | 
      
         | 5707 |  |  |  
 | 
      
         | 5708 |  |  |   const_tiny_rtx[3][(int) VOIDmode] = constm1_rtx;
 | 
      
         | 5709 |  |  |  
 | 
      
         | 5710 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_INT);
 | 
      
         | 5711 |  |  |        mode != VOIDmode;
 | 
      
         | 5712 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5713 |  |  |     const_tiny_rtx[3][(int) mode] = constm1_rtx;
 | 
      
         | 5714 |  |  |  
 | 
      
         | 5715 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_PARTIAL_INT);
 | 
      
         | 5716 |  |  |        mode != VOIDmode;
 | 
      
         | 5717 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5718 |  |  |     const_tiny_rtx[3][(int) mode] = constm1_rtx;
 | 
      
         | 5719 |  |  |  
 | 
      
         | 5720 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_COMPLEX_INT);
 | 
      
         | 5721 |  |  |        mode != VOIDmode;
 | 
      
         | 5722 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5723 |  |  |     {
 | 
      
         | 5724 |  |  |       rtx inner = const_tiny_rtx[0][(int)GET_MODE_INNER (mode)];
 | 
      
         | 5725 |  |  |       const_tiny_rtx[0][(int) mode] = gen_rtx_CONCAT (mode, inner, inner);
 | 
      
         | 5726 |  |  |     }
 | 
      
         | 5727 |  |  |  
 | 
      
         | 5728 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_COMPLEX_FLOAT);
 | 
      
         | 5729 |  |  |        mode != VOIDmode;
 | 
      
         | 5730 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5731 |  |  |     {
 | 
      
         | 5732 |  |  |       rtx inner = const_tiny_rtx[0][(int)GET_MODE_INNER (mode)];
 | 
      
         | 5733 |  |  |       const_tiny_rtx[0][(int) mode] = gen_rtx_CONCAT (mode, inner, inner);
 | 
      
         | 5734 |  |  |     }
 | 
      
         | 5735 |  |  |  
 | 
      
         | 5736 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_VECTOR_INT);
 | 
      
         | 5737 |  |  |        mode != VOIDmode;
 | 
      
         | 5738 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5739 |  |  |     {
 | 
      
         | 5740 |  |  |       const_tiny_rtx[0][(int) mode] = gen_const_vector (mode, 0);
 | 
      
         | 5741 |  |  |       const_tiny_rtx[1][(int) mode] = gen_const_vector (mode, 1);
 | 
      
         | 5742 |  |  |       const_tiny_rtx[3][(int) mode] = gen_const_vector (mode, 3);
 | 
      
         | 5743 |  |  |     }
 | 
      
         | 5744 |  |  |  
 | 
      
         | 5745 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_VECTOR_FLOAT);
 | 
      
         | 5746 |  |  |        mode != VOIDmode;
 | 
      
         | 5747 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5748 |  |  |     {
 | 
      
         | 5749 |  |  |       const_tiny_rtx[0][(int) mode] = gen_const_vector (mode, 0);
 | 
      
         | 5750 |  |  |       const_tiny_rtx[1][(int) mode] = gen_const_vector (mode, 1);
 | 
      
         | 5751 |  |  |     }
 | 
      
         | 5752 |  |  |  
 | 
      
         | 5753 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_FRACT);
 | 
      
         | 5754 |  |  |        mode != VOIDmode;
 | 
      
         | 5755 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5756 |  |  |     {
 | 
      
         | 5757 |  |  |       FCONST0(mode).data.high = 0;
 | 
      
         | 5758 |  |  |       FCONST0(mode).data.low = 0;
 | 
      
         | 5759 |  |  |       FCONST0(mode).mode = mode;
 | 
      
         | 5760 |  |  |       const_tiny_rtx[0][(int) mode] = CONST_FIXED_FROM_FIXED_VALUE (
 | 
      
         | 5761 |  |  |                                       FCONST0 (mode), mode);
 | 
      
         | 5762 |  |  |     }
 | 
      
         | 5763 |  |  |  
 | 
      
         | 5764 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_UFRACT);
 | 
      
         | 5765 |  |  |        mode != VOIDmode;
 | 
      
         | 5766 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5767 |  |  |     {
 | 
      
         | 5768 |  |  |       FCONST0(mode).data.high = 0;
 | 
      
         | 5769 |  |  |       FCONST0(mode).data.low = 0;
 | 
      
         | 5770 |  |  |       FCONST0(mode).mode = mode;
 | 
      
         | 5771 |  |  |       const_tiny_rtx[0][(int) mode] = CONST_FIXED_FROM_FIXED_VALUE (
 | 
      
         | 5772 |  |  |                                       FCONST0 (mode), mode);
 | 
      
         | 5773 |  |  |     }
 | 
      
         | 5774 |  |  |  
 | 
      
         | 5775 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_ACCUM);
 | 
      
         | 5776 |  |  |        mode != VOIDmode;
 | 
      
         | 5777 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5778 |  |  |     {
 | 
      
         | 5779 |  |  |       FCONST0(mode).data.high = 0;
 | 
      
         | 5780 |  |  |       FCONST0(mode).data.low = 0;
 | 
      
         | 5781 |  |  |       FCONST0(mode).mode = mode;
 | 
      
         | 5782 |  |  |       const_tiny_rtx[0][(int) mode] = CONST_FIXED_FROM_FIXED_VALUE (
 | 
      
         | 5783 |  |  |                                       FCONST0 (mode), mode);
 | 
      
         | 5784 |  |  |  
 | 
      
         | 5785 |  |  |       /* We store the value 1.  */
 | 
      
         | 5786 |  |  |       FCONST1(mode).data.high = 0;
 | 
      
         | 5787 |  |  |       FCONST1(mode).data.low = 0;
 | 
      
         | 5788 |  |  |       FCONST1(mode).mode = mode;
 | 
      
         | 5789 |  |  |       lshift_double (1, 0, GET_MODE_FBIT (mode),
 | 
      
         | 5790 |  |  |                      2 * HOST_BITS_PER_WIDE_INT,
 | 
      
         | 5791 |  |  |                      &FCONST1(mode).data.low,
 | 
      
         | 5792 |  |  |                      &FCONST1(mode).data.high,
 | 
      
         | 5793 |  |  |                      SIGNED_FIXED_POINT_MODE_P (mode));
 | 
      
         | 5794 |  |  |       const_tiny_rtx[1][(int) mode] = CONST_FIXED_FROM_FIXED_VALUE (
 | 
      
         | 5795 |  |  |                                       FCONST1 (mode), mode);
 | 
      
         | 5796 |  |  |     }
 | 
      
         | 5797 |  |  |  
 | 
      
         | 5798 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_UACCUM);
 | 
      
         | 5799 |  |  |        mode != VOIDmode;
 | 
      
         | 5800 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5801 |  |  |     {
 | 
      
         | 5802 |  |  |       FCONST0(mode).data.high = 0;
 | 
      
         | 5803 |  |  |       FCONST0(mode).data.low = 0;
 | 
      
         | 5804 |  |  |       FCONST0(mode).mode = mode;
 | 
      
         | 5805 |  |  |       const_tiny_rtx[0][(int) mode] = CONST_FIXED_FROM_FIXED_VALUE (
 | 
      
         | 5806 |  |  |                                       FCONST0 (mode), mode);
 | 
      
         | 5807 |  |  |  
 | 
      
         | 5808 |  |  |       /* We store the value 1.  */
 | 
      
         | 5809 |  |  |       FCONST1(mode).data.high = 0;
 | 
      
         | 5810 |  |  |       FCONST1(mode).data.low = 0;
 | 
      
         | 5811 |  |  |       FCONST1(mode).mode = mode;
 | 
      
         | 5812 |  |  |       lshift_double (1, 0, GET_MODE_FBIT (mode),
 | 
      
         | 5813 |  |  |                      2 * HOST_BITS_PER_WIDE_INT,
 | 
      
         | 5814 |  |  |                      &FCONST1(mode).data.low,
 | 
      
         | 5815 |  |  |                      &FCONST1(mode).data.high,
 | 
      
         | 5816 |  |  |                      SIGNED_FIXED_POINT_MODE_P (mode));
 | 
      
         | 5817 |  |  |       const_tiny_rtx[1][(int) mode] = CONST_FIXED_FROM_FIXED_VALUE (
 | 
      
         | 5818 |  |  |                                       FCONST1 (mode), mode);
 | 
      
         | 5819 |  |  |     }
 | 
      
         | 5820 |  |  |  
 | 
      
         | 5821 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_VECTOR_FRACT);
 | 
      
         | 5822 |  |  |        mode != VOIDmode;
 | 
      
         | 5823 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5824 |  |  |     {
 | 
      
         | 5825 |  |  |       const_tiny_rtx[0][(int) mode] = gen_const_vector (mode, 0);
 | 
      
         | 5826 |  |  |     }
 | 
      
         | 5827 |  |  |  
 | 
      
         | 5828 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_VECTOR_UFRACT);
 | 
      
         | 5829 |  |  |        mode != VOIDmode;
 | 
      
         | 5830 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5831 |  |  |     {
 | 
      
         | 5832 |  |  |       const_tiny_rtx[0][(int) mode] = gen_const_vector (mode, 0);
 | 
      
         | 5833 |  |  |     }
 | 
      
         | 5834 |  |  |  
 | 
      
         | 5835 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_VECTOR_ACCUM);
 | 
      
         | 5836 |  |  |        mode != VOIDmode;
 | 
      
         | 5837 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5838 |  |  |     {
 | 
      
         | 5839 |  |  |       const_tiny_rtx[0][(int) mode] = gen_const_vector (mode, 0);
 | 
      
         | 5840 |  |  |       const_tiny_rtx[1][(int) mode] = gen_const_vector (mode, 1);
 | 
      
         | 5841 |  |  |     }
 | 
      
         | 5842 |  |  |  
 | 
      
         | 5843 |  |  |   for (mode = GET_CLASS_NARROWEST_MODE (MODE_VECTOR_UACCUM);
 | 
      
         | 5844 |  |  |        mode != VOIDmode;
 | 
      
         | 5845 |  |  |        mode = GET_MODE_WIDER_MODE (mode))
 | 
      
         | 5846 |  |  |     {
 | 
      
         | 5847 |  |  |       const_tiny_rtx[0][(int) mode] = gen_const_vector (mode, 0);
 | 
      
         | 5848 |  |  |       const_tiny_rtx[1][(int) mode] = gen_const_vector (mode, 1);
 | 
      
         | 5849 |  |  |     }
 | 
      
         | 5850 |  |  |  
 | 
      
         | 5851 |  |  |   for (i = (int) CCmode; i < (int) MAX_MACHINE_MODE; ++i)
 | 
      
         | 5852 |  |  |     if (GET_MODE_CLASS ((enum machine_mode) i) == MODE_CC)
 | 
      
         | 5853 |  |  |       const_tiny_rtx[0][i] = const0_rtx;
 | 
      
         | 5854 |  |  |  
 | 
      
         | 5855 |  |  |   const_tiny_rtx[0][(int) BImode] = const0_rtx;
 | 
      
         | 5856 |  |  |   if (STORE_FLAG_VALUE == 1)
 | 
      
         | 5857 |  |  |     const_tiny_rtx[1][(int) BImode] = const1_rtx;
 | 
      
         | 5858 |  |  | }
 | 
      
         | 5859 |  |  |  
 | 
      
         | 5860 |  |  | /* Produce exact duplicate of insn INSN after AFTER.
 | 
      
         | 5861 |  |  |    Care updating of libcall regions if present.  */
 | 
      
         | 5862 |  |  |  
 | 
      
         | 5863 |  |  | rtx
 | 
      
         | 5864 |  |  | emit_copy_of_insn_after (rtx insn, rtx after)
 | 
      
         | 5865 |  |  | {
 | 
      
         | 5866 |  |  |   rtx new_rtx, link;
 | 
      
         | 5867 |  |  |  
 | 
      
         | 5868 |  |  |   switch (GET_CODE (insn))
 | 
      
         | 5869 |  |  |     {
 | 
      
         | 5870 |  |  |     case INSN:
 | 
      
         | 5871 |  |  |       new_rtx = emit_insn_after (copy_insn (PATTERN (insn)), after);
 | 
      
         | 5872 |  |  |       break;
 | 
      
         | 5873 |  |  |  
 | 
      
         | 5874 |  |  |     case JUMP_INSN:
 | 
      
         | 5875 |  |  |       new_rtx = emit_jump_insn_after (copy_insn (PATTERN (insn)), after);
 | 
      
         | 5876 |  |  |       break;
 | 
      
         | 5877 |  |  |  
 | 
      
         | 5878 |  |  |     case DEBUG_INSN:
 | 
      
         | 5879 |  |  |       new_rtx = emit_debug_insn_after (copy_insn (PATTERN (insn)), after);
 | 
      
         | 5880 |  |  |       break;
 | 
      
         | 5881 |  |  |  
 | 
      
         | 5882 |  |  |     case CALL_INSN:
 | 
      
         | 5883 |  |  |       new_rtx = emit_call_insn_after (copy_insn (PATTERN (insn)), after);
 | 
      
         | 5884 |  |  |       if (CALL_INSN_FUNCTION_USAGE (insn))
 | 
      
         | 5885 |  |  |         CALL_INSN_FUNCTION_USAGE (new_rtx)
 | 
      
         | 5886 |  |  |           = copy_insn (CALL_INSN_FUNCTION_USAGE (insn));
 | 
      
         | 5887 |  |  |       SIBLING_CALL_P (new_rtx) = SIBLING_CALL_P (insn);
 | 
      
         | 5888 |  |  |       RTL_CONST_CALL_P (new_rtx) = RTL_CONST_CALL_P (insn);
 | 
      
         | 5889 |  |  |       RTL_PURE_CALL_P (new_rtx) = RTL_PURE_CALL_P (insn);
 | 
      
         | 5890 |  |  |       RTL_LOOPING_CONST_OR_PURE_CALL_P (new_rtx)
 | 
      
         | 5891 |  |  |         = RTL_LOOPING_CONST_OR_PURE_CALL_P (insn);
 | 
      
         | 5892 |  |  |       break;
 | 
      
         | 5893 |  |  |  
 | 
      
         | 5894 |  |  |     default:
 | 
      
         | 5895 |  |  |       gcc_unreachable ();
 | 
      
         | 5896 |  |  |     }
 | 
      
         | 5897 |  |  |  
 | 
      
         | 5898 |  |  |   /* Update LABEL_NUSES.  */
 | 
      
         | 5899 |  |  |   mark_jump_label (PATTERN (new_rtx), new_rtx, 0);
 | 
      
         | 5900 |  |  |  
 | 
      
         | 5901 |  |  |   INSN_LOCATOR (new_rtx) = INSN_LOCATOR (insn);
 | 
      
         | 5902 |  |  |  
 | 
      
         | 5903 |  |  |   /* If the old insn is frame related, then so is the new one.  This is
 | 
      
         | 5904 |  |  |      primarily needed for IA-64 unwind info which marks epilogue insns,
 | 
      
         | 5905 |  |  |      which may be duplicated by the basic block reordering code.  */
 | 
      
         | 5906 |  |  |   RTX_FRAME_RELATED_P (new_rtx) = RTX_FRAME_RELATED_P (insn);
 | 
      
         | 5907 |  |  |  
 | 
      
         | 5908 |  |  |   /* Copy all REG_NOTES except REG_LABEL_OPERAND since mark_jump_label
 | 
      
         | 5909 |  |  |      will make them.  REG_LABEL_TARGETs are created there too, but are
 | 
      
         | 5910 |  |  |      supposed to be sticky, so we copy them.  */
 | 
      
         | 5911 |  |  |   for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
 | 
      
         | 5912 |  |  |     if (REG_NOTE_KIND (link) != REG_LABEL_OPERAND)
 | 
      
         | 5913 |  |  |       {
 | 
      
         | 5914 |  |  |         if (GET_CODE (link) == EXPR_LIST)
 | 
      
         | 5915 |  |  |           add_reg_note (new_rtx, REG_NOTE_KIND (link),
 | 
      
         | 5916 |  |  |                         copy_insn_1 (XEXP (link, 0)));
 | 
      
         | 5917 |  |  |         else
 | 
      
         | 5918 |  |  |           add_reg_note (new_rtx, REG_NOTE_KIND (link), XEXP (link, 0));
 | 
      
         | 5919 |  |  |       }
 | 
      
         | 5920 |  |  |  
 | 
      
         | 5921 |  |  |   INSN_CODE (new_rtx) = INSN_CODE (insn);
 | 
      
         | 5922 |  |  |   return new_rtx;
 | 
      
         | 5923 |  |  | }
 | 
      
         | 5924 |  |  |  
 | 
      
         | 5925 |  |  | static GTY((deletable)) rtx hard_reg_clobbers [NUM_MACHINE_MODES][FIRST_PSEUDO_REGISTER];
 | 
      
         | 5926 |  |  | rtx
 | 
      
         | 5927 |  |  | gen_hard_reg_clobber (enum machine_mode mode, unsigned int regno)
 | 
      
         | 5928 |  |  | {
 | 
      
         | 5929 |  |  |   if (hard_reg_clobbers[mode][regno])
 | 
      
         | 5930 |  |  |     return hard_reg_clobbers[mode][regno];
 | 
      
         | 5931 |  |  |   else
 | 
      
         | 5932 |  |  |     return (hard_reg_clobbers[mode][regno] =
 | 
      
         | 5933 |  |  |             gen_rtx_CLOBBER (VOIDmode, gen_rtx_REG (mode, regno)));
 | 
      
         | 5934 |  |  | }
 | 
      
         | 5935 |  |  |  
 | 
      
         | 5936 |  |  | #include "gt-emit-rtl.h"
 |