1 |
282 |
jeremybenn |
/* Output routines for GCC for Renesas / SuperH SH.
|
2 |
|
|
Copyright (C) 1993, 1994, 1995, 1996, 1997, 1998, 1999, 2000, 2001, 2002,
|
3 |
|
|
2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010
|
4 |
|
|
Free Software Foundation, Inc.
|
5 |
|
|
Contributed by Steve Chamberlain (sac@cygnus.com).
|
6 |
|
|
Improved by Jim Wilson (wilson@cygnus.com).
|
7 |
|
|
|
8 |
|
|
This file is part of GCC.
|
9 |
|
|
|
10 |
|
|
GCC is free software; you can redistribute it and/or modify
|
11 |
|
|
it under the terms of the GNU General Public License as published by
|
12 |
|
|
the Free Software Foundation; either version 3, or (at your option)
|
13 |
|
|
any later version.
|
14 |
|
|
|
15 |
|
|
GCC is distributed in the hope that it will be useful,
|
16 |
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
17 |
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
18 |
|
|
GNU General Public License for more details.
|
19 |
|
|
|
20 |
|
|
You should have received a copy of the GNU General Public License
|
21 |
|
|
along with GCC; see the file COPYING3. If not see
|
22 |
|
|
<http://www.gnu.org/licenses/>. */
|
23 |
|
|
|
24 |
|
|
#include "config.h"
|
25 |
|
|
#include "system.h"
|
26 |
|
|
#include "coretypes.h"
|
27 |
|
|
#include "tm.h"
|
28 |
|
|
#include "insn-config.h"
|
29 |
|
|
#include "rtl.h"
|
30 |
|
|
#include "tree.h"
|
31 |
|
|
#include "flags.h"
|
32 |
|
|
#include "expr.h"
|
33 |
|
|
#include "optabs.h"
|
34 |
|
|
#include "function.h"
|
35 |
|
|
#include "regs.h"
|
36 |
|
|
#include "hard-reg-set.h"
|
37 |
|
|
#include "output.h"
|
38 |
|
|
#include "insn-attr.h"
|
39 |
|
|
#include "toplev.h"
|
40 |
|
|
#include "recog.h"
|
41 |
|
|
#include "integrate.h"
|
42 |
|
|
#include "dwarf2.h"
|
43 |
|
|
#include "tm_p.h"
|
44 |
|
|
#include "target.h"
|
45 |
|
|
#include "target-def.h"
|
46 |
|
|
#include "real.h"
|
47 |
|
|
#include "langhooks.h"
|
48 |
|
|
#include "basic-block.h"
|
49 |
|
|
#include "df.h"
|
50 |
|
|
#include "cfglayout.h"
|
51 |
|
|
#include "intl.h"
|
52 |
|
|
#include "sched-int.h"
|
53 |
|
|
#include "params.h"
|
54 |
|
|
#include "ggc.h"
|
55 |
|
|
#include "gimple.h"
|
56 |
|
|
#include "cfgloop.h"
|
57 |
|
|
#include "alloc-pool.h"
|
58 |
|
|
#include "tm-constrs.h"
|
59 |
|
|
|
60 |
|
|
|
61 |
|
|
int code_for_indirect_jump_scratch = CODE_FOR_indirect_jump_scratch;
|
62 |
|
|
|
63 |
|
|
#define MSW (TARGET_LITTLE_ENDIAN ? 1 : 0)
|
64 |
|
|
#define LSW (TARGET_LITTLE_ENDIAN ? 0 : 1)
|
65 |
|
|
|
66 |
|
|
/* These are some macros to abstract register modes. */
|
67 |
|
|
#define CONST_OK_FOR_ADD(size) \
|
68 |
|
|
(TARGET_SHMEDIA ? CONST_OK_FOR_I10 (size) : CONST_OK_FOR_I08 (size))
|
69 |
|
|
#define GEN_MOV (*(TARGET_SHMEDIA64 ? gen_movdi : gen_movsi))
|
70 |
|
|
#define GEN_ADD3 (*(TARGET_SHMEDIA64 ? gen_adddi3 : gen_addsi3))
|
71 |
|
|
#define GEN_SUB3 (*(TARGET_SHMEDIA64 ? gen_subdi3 : gen_subsi3))
|
72 |
|
|
|
73 |
|
|
/* Used to simplify the logic below. Find the attributes wherever
|
74 |
|
|
they may be. */
|
75 |
|
|
#define SH_ATTRIBUTES(decl) \
|
76 |
|
|
(TYPE_P (decl)) ? TYPE_ATTRIBUTES (decl) \
|
77 |
|
|
: DECL_ATTRIBUTES (decl) \
|
78 |
|
|
? (DECL_ATTRIBUTES (decl)) \
|
79 |
|
|
: TYPE_ATTRIBUTES (TREE_TYPE (decl))
|
80 |
|
|
|
81 |
|
|
/* Set to 1 by expand_prologue() when the function is an interrupt handler. */
|
82 |
|
|
int current_function_interrupt;
|
83 |
|
|
|
84 |
|
|
tree sh_deferred_function_attributes;
|
85 |
|
|
tree *sh_deferred_function_attributes_tail = &sh_deferred_function_attributes;
|
86 |
|
|
|
87 |
|
|
/* Global variables for machine-dependent things. */
|
88 |
|
|
|
89 |
|
|
/* Which cpu are we scheduling for. */
|
90 |
|
|
enum processor_type sh_cpu;
|
91 |
|
|
|
92 |
|
|
/* Definitions used in ready queue reordering for first scheduling pass. */
|
93 |
|
|
|
94 |
|
|
/* Reg weights arrays for modes SFmode and SImode, indexed by insn LUID. */
|
95 |
|
|
static short *regmode_weight[2];
|
96 |
|
|
|
97 |
|
|
/* Total SFmode and SImode weights of scheduled insns. */
|
98 |
|
|
static int curr_regmode_pressure[2];
|
99 |
|
|
|
100 |
|
|
/* Number of r0 life regions. */
|
101 |
|
|
static int r0_life_regions;
|
102 |
|
|
|
103 |
|
|
/* If true, skip cycles for Q -> R movement. */
|
104 |
|
|
static int skip_cycles = 0;
|
105 |
|
|
|
106 |
|
|
/* Cached value of can_issue_more. This is cached in sh_variable_issue hook
|
107 |
|
|
and returned from sh_reorder2. */
|
108 |
|
|
static short cached_can_issue_more;
|
109 |
|
|
|
110 |
|
|
/* Unique number for UNSPEC_BBR pattern. */
|
111 |
|
|
static unsigned int unspec_bbr_uid = 1;
|
112 |
|
|
|
113 |
|
|
/* Provides the class number of the smallest class containing
|
114 |
|
|
reg number. */
|
115 |
|
|
|
116 |
|
|
enum reg_class regno_reg_class[FIRST_PSEUDO_REGISTER] =
|
117 |
|
|
{
|
118 |
|
|
R0_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
119 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
120 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
121 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
122 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
123 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
124 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
125 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
126 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
127 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
128 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
129 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
130 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
131 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
132 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
133 |
|
|
GENERAL_REGS, GENERAL_REGS, GENERAL_REGS, GENERAL_REGS,
|
134 |
|
|
FP0_REGS,FP_REGS, FP_REGS, FP_REGS,
|
135 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
136 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
137 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
138 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
139 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
140 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
141 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
142 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
143 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
144 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
145 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
146 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
147 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
148 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
149 |
|
|
FP_REGS, FP_REGS, FP_REGS, FP_REGS,
|
150 |
|
|
TARGET_REGS, TARGET_REGS, TARGET_REGS, TARGET_REGS,
|
151 |
|
|
TARGET_REGS, TARGET_REGS, TARGET_REGS, TARGET_REGS,
|
152 |
|
|
DF_REGS, DF_REGS, DF_REGS, DF_REGS,
|
153 |
|
|
DF_REGS, DF_REGS, DF_REGS, DF_REGS,
|
154 |
|
|
NO_REGS, GENERAL_REGS, PR_REGS, T_REGS,
|
155 |
|
|
MAC_REGS, MAC_REGS, FPUL_REGS, FPSCR_REGS,
|
156 |
|
|
GENERAL_REGS, GENERAL_REGS,
|
157 |
|
|
};
|
158 |
|
|
|
159 |
|
|
char sh_register_names[FIRST_PSEUDO_REGISTER] \
|
160 |
|
|
[MAX_REGISTER_NAME_LENGTH + 1] = SH_REGISTER_NAMES_INITIALIZER;
|
161 |
|
|
|
162 |
|
|
char sh_additional_register_names[ADDREGNAMES_SIZE] \
|
163 |
|
|
[MAX_ADDITIONAL_REGISTER_NAME_LENGTH + 1]
|
164 |
|
|
= SH_ADDITIONAL_REGISTER_NAMES_INITIALIZER;
|
165 |
|
|
|
166 |
|
|
int assembler_dialect;
|
167 |
|
|
|
168 |
|
|
static bool shmedia_space_reserved_for_target_registers;
|
169 |
|
|
|
170 |
|
|
static bool sh_handle_option (size_t, const char *, int);
|
171 |
|
|
static void split_branches (rtx);
|
172 |
|
|
static int branch_dest (rtx);
|
173 |
|
|
static void force_into (rtx, rtx);
|
174 |
|
|
static void print_slot (rtx);
|
175 |
|
|
static rtx add_constant (rtx, enum machine_mode, rtx);
|
176 |
|
|
static void dump_table (rtx, rtx);
|
177 |
|
|
static int hi_const (rtx);
|
178 |
|
|
static int broken_move (rtx);
|
179 |
|
|
static int mova_p (rtx);
|
180 |
|
|
static rtx find_barrier (int, rtx, rtx);
|
181 |
|
|
static int noncall_uses_reg (rtx, rtx, rtx *);
|
182 |
|
|
static rtx gen_block_redirect (rtx, int, int);
|
183 |
|
|
static void sh_reorg (void);
|
184 |
|
|
static void output_stack_adjust (int, rtx, int, HARD_REG_SET *, bool);
|
185 |
|
|
static rtx frame_insn (rtx);
|
186 |
|
|
static rtx push (int);
|
187 |
|
|
static void pop (int);
|
188 |
|
|
static void push_regs (HARD_REG_SET *, int);
|
189 |
|
|
static int calc_live_regs (HARD_REG_SET *);
|
190 |
|
|
static HOST_WIDE_INT rounded_frame_size (int);
|
191 |
|
|
static rtx mark_constant_pool_use (rtx);
|
192 |
|
|
static tree sh_handle_interrupt_handler_attribute (tree *, tree, tree, int, bool *);
|
193 |
|
|
static tree sh_handle_resbank_handler_attribute (tree *, tree,
|
194 |
|
|
tree, int, bool *);
|
195 |
|
|
static tree sh2a_handle_function_vector_handler_attribute (tree *, tree,
|
196 |
|
|
tree, int, bool *);
|
197 |
|
|
static tree sh_handle_sp_switch_attribute (tree *, tree, tree, int, bool *);
|
198 |
|
|
static tree sh_handle_trap_exit_attribute (tree *, tree, tree, int, bool *);
|
199 |
|
|
static tree sh_handle_renesas_attribute (tree *, tree, tree, int, bool *);
|
200 |
|
|
static void sh_output_function_epilogue (FILE *, HOST_WIDE_INT);
|
201 |
|
|
static void sh_insert_attributes (tree, tree *);
|
202 |
|
|
static const char *sh_check_pch_target_flags (int);
|
203 |
|
|
static int sh_adjust_cost (rtx, rtx, rtx, int);
|
204 |
|
|
static int sh_issue_rate (void);
|
205 |
|
|
static int sh_dfa_new_cycle (FILE *, int, rtx, int, int, int *sort_p);
|
206 |
|
|
static short find_set_regmode_weight (rtx, enum machine_mode);
|
207 |
|
|
static short find_insn_regmode_weight (rtx, enum machine_mode);
|
208 |
|
|
static void find_regmode_weight (basic_block, enum machine_mode);
|
209 |
|
|
static int find_r0_life_regions (basic_block);
|
210 |
|
|
static void sh_md_init_global (FILE *, int, int);
|
211 |
|
|
static void sh_md_finish_global (FILE *, int);
|
212 |
|
|
static int rank_for_reorder (const void *, const void *);
|
213 |
|
|
static void swap_reorder (rtx *, int);
|
214 |
|
|
static void ready_reorder (rtx *, int);
|
215 |
|
|
static short high_pressure (enum machine_mode);
|
216 |
|
|
static int sh_reorder (FILE *, int, rtx *, int *, int);
|
217 |
|
|
static int sh_reorder2 (FILE *, int, rtx *, int *, int);
|
218 |
|
|
static void sh_md_init (FILE *, int, int);
|
219 |
|
|
static int sh_variable_issue (FILE *, int, rtx, int);
|
220 |
|
|
|
221 |
|
|
static bool sh_function_ok_for_sibcall (tree, tree);
|
222 |
|
|
|
223 |
|
|
static bool sh_cannot_modify_jumps_p (void);
|
224 |
|
|
static enum reg_class sh_target_reg_class (void);
|
225 |
|
|
static bool sh_optimize_target_register_callee_saved (bool);
|
226 |
|
|
static bool sh_ms_bitfield_layout_p (const_tree);
|
227 |
|
|
|
228 |
|
|
static void sh_init_builtins (void);
|
229 |
|
|
static tree sh_builtin_decl (unsigned, bool);
|
230 |
|
|
static void sh_media_init_builtins (void);
|
231 |
|
|
static tree sh_media_builtin_decl (unsigned, bool);
|
232 |
|
|
static rtx sh_expand_builtin (tree, rtx, rtx, enum machine_mode, int);
|
233 |
|
|
static void sh_output_mi_thunk (FILE *, tree, HOST_WIDE_INT, HOST_WIDE_INT, tree);
|
234 |
|
|
static void sh_file_start (void);
|
235 |
|
|
static int flow_dependent_p (rtx, rtx);
|
236 |
|
|
static void flow_dependent_p_1 (rtx, const_rtx, void *);
|
237 |
|
|
static int shiftcosts (rtx);
|
238 |
|
|
static int andcosts (rtx);
|
239 |
|
|
static int addsubcosts (rtx);
|
240 |
|
|
static int multcosts (rtx);
|
241 |
|
|
static bool unspec_caller_rtx_p (rtx);
|
242 |
|
|
static bool sh_cannot_copy_insn_p (rtx);
|
243 |
|
|
static bool sh_rtx_costs (rtx, int, int, int *, bool);
|
244 |
|
|
static int sh_address_cost (rtx, bool);
|
245 |
|
|
static int sh_pr_n_sets (void);
|
246 |
|
|
static rtx sh_allocate_initial_value (rtx);
|
247 |
|
|
static bool sh_legitimate_address_p (enum machine_mode, rtx, bool);
|
248 |
|
|
static rtx sh_legitimize_address (rtx, rtx, enum machine_mode);
|
249 |
|
|
static int shmedia_target_regs_stack_space (HARD_REG_SET *);
|
250 |
|
|
static int shmedia_reserve_space_for_target_registers_p (int, HARD_REG_SET *);
|
251 |
|
|
static int shmedia_target_regs_stack_adjust (HARD_REG_SET *);
|
252 |
|
|
static int scavenge_reg (HARD_REG_SET *s);
|
253 |
|
|
struct save_schedule_s;
|
254 |
|
|
static struct save_entry_s *sh5_schedule_saves (HARD_REG_SET *,
|
255 |
|
|
struct save_schedule_s *, int);
|
256 |
|
|
|
257 |
|
|
static rtx sh_struct_value_rtx (tree, int);
|
258 |
|
|
static rtx sh_function_value (const_tree, const_tree, bool);
|
259 |
|
|
static rtx sh_libcall_value (enum machine_mode, const_rtx);
|
260 |
|
|
static bool sh_return_in_memory (const_tree, const_tree);
|
261 |
|
|
static rtx sh_builtin_saveregs (void);
|
262 |
|
|
static void sh_setup_incoming_varargs (CUMULATIVE_ARGS *, enum machine_mode, tree, int *, int);
|
263 |
|
|
static bool sh_strict_argument_naming (CUMULATIVE_ARGS *);
|
264 |
|
|
static bool sh_pretend_outgoing_varargs_named (CUMULATIVE_ARGS *);
|
265 |
|
|
static tree sh_build_builtin_va_list (void);
|
266 |
|
|
static void sh_va_start (tree, rtx);
|
267 |
|
|
static tree sh_gimplify_va_arg_expr (tree, tree, gimple_seq *, gimple_seq *);
|
268 |
|
|
static bool sh_promote_prototypes (const_tree);
|
269 |
|
|
static enum machine_mode sh_promote_function_mode (const_tree type,
|
270 |
|
|
enum machine_mode,
|
271 |
|
|
int *punsignedp,
|
272 |
|
|
const_tree funtype,
|
273 |
|
|
int for_return);
|
274 |
|
|
static bool sh_pass_by_reference (CUMULATIVE_ARGS *, enum machine_mode,
|
275 |
|
|
const_tree, bool);
|
276 |
|
|
static bool sh_callee_copies (CUMULATIVE_ARGS *, enum machine_mode,
|
277 |
|
|
const_tree, bool);
|
278 |
|
|
static int sh_arg_partial_bytes (CUMULATIVE_ARGS *, enum machine_mode,
|
279 |
|
|
tree, bool);
|
280 |
|
|
static bool sh_scalar_mode_supported_p (enum machine_mode);
|
281 |
|
|
static int sh_dwarf_calling_convention (const_tree);
|
282 |
|
|
static void sh_encode_section_info (tree, rtx, int);
|
283 |
|
|
static int sh2a_function_vector_p (tree);
|
284 |
|
|
static void sh_trampoline_init (rtx, tree, rtx);
|
285 |
|
|
static rtx sh_trampoline_adjust_address (rtx);
|
286 |
|
|
|
287 |
|
|
static const struct attribute_spec sh_attribute_table[] =
|
288 |
|
|
{
|
289 |
|
|
/* { name, min_len, max_len, decl_req, type_req, fn_type_req, handler } */
|
290 |
|
|
{ "interrupt_handler", 0, 0, true, false, false, sh_handle_interrupt_handler_attribute },
|
291 |
|
|
{ "sp_switch", 1, 1, true, false, false, sh_handle_sp_switch_attribute },
|
292 |
|
|
{ "trap_exit", 1, 1, true, false, false, sh_handle_trap_exit_attribute },
|
293 |
|
|
{ "renesas", 0, 0, false, true, false, sh_handle_renesas_attribute },
|
294 |
|
|
{ "trapa_handler", 0, 0, true, false, false, sh_handle_interrupt_handler_attribute },
|
295 |
|
|
{ "nosave_low_regs", 0, 0, true, false, false, sh_handle_interrupt_handler_attribute },
|
296 |
|
|
{ "resbank", 0, 0, true, false, false, sh_handle_resbank_handler_attribute },
|
297 |
|
|
{ "function_vector", 1, 1, true, false, false, sh2a_handle_function_vector_handler_attribute },
|
298 |
|
|
#ifdef SYMBIAN
|
299 |
|
|
/* Symbian support adds three new attributes:
|
300 |
|
|
dllexport - for exporting a function/variable that will live in a dll
|
301 |
|
|
dllimport - for importing a function/variable from a dll
|
302 |
|
|
|
303 |
|
|
Microsoft allows multiple declspecs in one __declspec, separating
|
304 |
|
|
them with spaces. We do NOT support this. Instead, use __declspec
|
305 |
|
|
multiple times. */
|
306 |
|
|
{ "dllimport", 0, 0, true, false, false, sh_symbian_handle_dll_attribute },
|
307 |
|
|
{ "dllexport", 0, 0, true, false, false, sh_symbian_handle_dll_attribute },
|
308 |
|
|
#endif
|
309 |
|
|
{ NULL, 0, 0, false, false, false, NULL }
|
310 |
|
|
};
|
311 |
|
|
|
312 |
|
|
/* Initialize the GCC target structure. */
|
313 |
|
|
#undef TARGET_ATTRIBUTE_TABLE
|
314 |
|
|
#define TARGET_ATTRIBUTE_TABLE sh_attribute_table
|
315 |
|
|
|
316 |
|
|
/* The next two are used for debug info when compiling with -gdwarf. */
|
317 |
|
|
#undef TARGET_ASM_UNALIGNED_HI_OP
|
318 |
|
|
#define TARGET_ASM_UNALIGNED_HI_OP "\t.uaword\t"
|
319 |
|
|
#undef TARGET_ASM_UNALIGNED_SI_OP
|
320 |
|
|
#define TARGET_ASM_UNALIGNED_SI_OP "\t.ualong\t"
|
321 |
|
|
|
322 |
|
|
/* These are NULLed out on non-SH5 in OVERRIDE_OPTIONS. */
|
323 |
|
|
#undef TARGET_ASM_UNALIGNED_DI_OP
|
324 |
|
|
#define TARGET_ASM_UNALIGNED_DI_OP "\t.uaquad\t"
|
325 |
|
|
#undef TARGET_ASM_ALIGNED_DI_OP
|
326 |
|
|
#define TARGET_ASM_ALIGNED_DI_OP "\t.quad\t"
|
327 |
|
|
|
328 |
|
|
#undef TARGET_ASM_FUNCTION_EPILOGUE
|
329 |
|
|
#define TARGET_ASM_FUNCTION_EPILOGUE sh_output_function_epilogue
|
330 |
|
|
|
331 |
|
|
#undef TARGET_ASM_OUTPUT_MI_THUNK
|
332 |
|
|
#define TARGET_ASM_OUTPUT_MI_THUNK sh_output_mi_thunk
|
333 |
|
|
|
334 |
|
|
#undef TARGET_ASM_CAN_OUTPUT_MI_THUNK
|
335 |
|
|
#define TARGET_ASM_CAN_OUTPUT_MI_THUNK hook_bool_const_tree_hwi_hwi_const_tree_true
|
336 |
|
|
|
337 |
|
|
#undef TARGET_ASM_FILE_START
|
338 |
|
|
#define TARGET_ASM_FILE_START sh_file_start
|
339 |
|
|
#undef TARGET_ASM_FILE_START_FILE_DIRECTIVE
|
340 |
|
|
#define TARGET_ASM_FILE_START_FILE_DIRECTIVE true
|
341 |
|
|
|
342 |
|
|
#undef TARGET_DEFAULT_TARGET_FLAGS
|
343 |
|
|
#define TARGET_DEFAULT_TARGET_FLAGS TARGET_DEFAULT
|
344 |
|
|
#undef TARGET_HANDLE_OPTION
|
345 |
|
|
#define TARGET_HANDLE_OPTION sh_handle_option
|
346 |
|
|
|
347 |
|
|
#undef TARGET_INSERT_ATTRIBUTES
|
348 |
|
|
#define TARGET_INSERT_ATTRIBUTES sh_insert_attributes
|
349 |
|
|
|
350 |
|
|
#undef TARGET_SCHED_ADJUST_COST
|
351 |
|
|
#define TARGET_SCHED_ADJUST_COST sh_adjust_cost
|
352 |
|
|
|
353 |
|
|
#undef TARGET_SCHED_ISSUE_RATE
|
354 |
|
|
#define TARGET_SCHED_ISSUE_RATE sh_issue_rate
|
355 |
|
|
|
356 |
|
|
/* The next 5 hooks have been implemented for reenabling sched1. With the
|
357 |
|
|
help of these macros we are limiting the movement of insns in sched1 to
|
358 |
|
|
reduce the register pressure. The overall idea is to keep count of SImode
|
359 |
|
|
and SFmode regs required by already scheduled insns. When these counts
|
360 |
|
|
cross some threshold values; give priority to insns that free registers.
|
361 |
|
|
The insn that frees registers is most likely to be the insn with lowest
|
362 |
|
|
LUID (original insn order); but such an insn might be there in the stalled
|
363 |
|
|
queue (Q) instead of the ready queue (R). To solve this, we skip cycles
|
364 |
|
|
upto a max of 8 cycles so that such insns may move from Q -> R.
|
365 |
|
|
|
366 |
|
|
The description of the hooks are as below:
|
367 |
|
|
|
368 |
|
|
TARGET_SCHED_INIT_GLOBAL: Added a new target hook in the generic
|
369 |
|
|
scheduler; it is called inside the sched_init function just after
|
370 |
|
|
find_insn_reg_weights function call. It is used to calculate the SImode
|
371 |
|
|
and SFmode weights of insns of basic blocks; much similar to what
|
372 |
|
|
find_insn_reg_weights does.
|
373 |
|
|
TARGET_SCHED_FINISH_GLOBAL: Corresponding cleanup hook.
|
374 |
|
|
|
375 |
|
|
TARGET_SCHED_DFA_NEW_CYCLE: Skip cycles if high register pressure is
|
376 |
|
|
indicated by TARGET_SCHED_REORDER2; doing this may move insns from
|
377 |
|
|
(Q)->(R).
|
378 |
|
|
|
379 |
|
|
TARGET_SCHED_REORDER: If the register pressure for SImode or SFmode is
|
380 |
|
|
high; reorder the ready queue so that the insn with lowest LUID will be
|
381 |
|
|
issued next.
|
382 |
|
|
|
383 |
|
|
TARGET_SCHED_REORDER2: If the register pressure is high, indicate to
|
384 |
|
|
TARGET_SCHED_DFA_NEW_CYCLE to skip cycles.
|
385 |
|
|
|
386 |
|
|
TARGET_SCHED_VARIABLE_ISSUE: Cache the value of can_issue_more so that it
|
387 |
|
|
can be returned from TARGET_SCHED_REORDER2.
|
388 |
|
|
|
389 |
|
|
TARGET_SCHED_INIT: Reset the register pressure counting variables. */
|
390 |
|
|
|
391 |
|
|
#undef TARGET_SCHED_DFA_NEW_CYCLE
|
392 |
|
|
#define TARGET_SCHED_DFA_NEW_CYCLE sh_dfa_new_cycle
|
393 |
|
|
|
394 |
|
|
#undef TARGET_SCHED_INIT_GLOBAL
|
395 |
|
|
#define TARGET_SCHED_INIT_GLOBAL sh_md_init_global
|
396 |
|
|
|
397 |
|
|
#undef TARGET_SCHED_FINISH_GLOBAL
|
398 |
|
|
#define TARGET_SCHED_FINISH_GLOBAL sh_md_finish_global
|
399 |
|
|
|
400 |
|
|
#undef TARGET_SCHED_VARIABLE_ISSUE
|
401 |
|
|
#define TARGET_SCHED_VARIABLE_ISSUE sh_variable_issue
|
402 |
|
|
|
403 |
|
|
#undef TARGET_SCHED_REORDER
|
404 |
|
|
#define TARGET_SCHED_REORDER sh_reorder
|
405 |
|
|
|
406 |
|
|
#undef TARGET_SCHED_REORDER2
|
407 |
|
|
#define TARGET_SCHED_REORDER2 sh_reorder2
|
408 |
|
|
|
409 |
|
|
#undef TARGET_SCHED_INIT
|
410 |
|
|
#define TARGET_SCHED_INIT sh_md_init
|
411 |
|
|
|
412 |
|
|
#undef TARGET_LEGITIMIZE_ADDRESS
|
413 |
|
|
#define TARGET_LEGITIMIZE_ADDRESS sh_legitimize_address
|
414 |
|
|
|
415 |
|
|
#undef TARGET_CANNOT_MODIFY_JUMPS_P
|
416 |
|
|
#define TARGET_CANNOT_MODIFY_JUMPS_P sh_cannot_modify_jumps_p
|
417 |
|
|
#undef TARGET_BRANCH_TARGET_REGISTER_CLASS
|
418 |
|
|
#define TARGET_BRANCH_TARGET_REGISTER_CLASS sh_target_reg_class
|
419 |
|
|
#undef TARGET_BRANCH_TARGET_REGISTER_CALLEE_SAVED
|
420 |
|
|
#define TARGET_BRANCH_TARGET_REGISTER_CALLEE_SAVED \
|
421 |
|
|
sh_optimize_target_register_callee_saved
|
422 |
|
|
|
423 |
|
|
#undef TARGET_MS_BITFIELD_LAYOUT_P
|
424 |
|
|
#define TARGET_MS_BITFIELD_LAYOUT_P sh_ms_bitfield_layout_p
|
425 |
|
|
|
426 |
|
|
#undef TARGET_INIT_BUILTINS
|
427 |
|
|
#define TARGET_INIT_BUILTINS sh_init_builtins
|
428 |
|
|
#undef TARGET_BUILTIN_DECL
|
429 |
|
|
#define TARGET_BUILTIN_DECL sh_builtin_decl
|
430 |
|
|
#undef TARGET_EXPAND_BUILTIN
|
431 |
|
|
#define TARGET_EXPAND_BUILTIN sh_expand_builtin
|
432 |
|
|
|
433 |
|
|
#undef TARGET_FUNCTION_OK_FOR_SIBCALL
|
434 |
|
|
#define TARGET_FUNCTION_OK_FOR_SIBCALL sh_function_ok_for_sibcall
|
435 |
|
|
|
436 |
|
|
#undef TARGET_CANNOT_COPY_INSN_P
|
437 |
|
|
#define TARGET_CANNOT_COPY_INSN_P sh_cannot_copy_insn_p
|
438 |
|
|
#undef TARGET_RTX_COSTS
|
439 |
|
|
#define TARGET_RTX_COSTS sh_rtx_costs
|
440 |
|
|
#undef TARGET_ADDRESS_COST
|
441 |
|
|
#define TARGET_ADDRESS_COST sh_address_cost
|
442 |
|
|
#undef TARGET_ALLOCATE_INITIAL_VALUE
|
443 |
|
|
#define TARGET_ALLOCATE_INITIAL_VALUE sh_allocate_initial_value
|
444 |
|
|
|
445 |
|
|
#undef TARGET_MACHINE_DEPENDENT_REORG
|
446 |
|
|
#define TARGET_MACHINE_DEPENDENT_REORG sh_reorg
|
447 |
|
|
|
448 |
|
|
#undef TARGET_DWARF_REGISTER_SPAN
|
449 |
|
|
#define TARGET_DWARF_REGISTER_SPAN sh_dwarf_register_span
|
450 |
|
|
|
451 |
|
|
#ifdef HAVE_AS_TLS
|
452 |
|
|
#undef TARGET_HAVE_TLS
|
453 |
|
|
#define TARGET_HAVE_TLS true
|
454 |
|
|
#endif
|
455 |
|
|
|
456 |
|
|
#undef TARGET_PROMOTE_PROTOTYPES
|
457 |
|
|
#define TARGET_PROMOTE_PROTOTYPES sh_promote_prototypes
|
458 |
|
|
#undef TARGET_PROMOTE_FUNCTION_MODE
|
459 |
|
|
#define TARGET_PROMOTE_FUNCTION_MODE sh_promote_function_mode
|
460 |
|
|
|
461 |
|
|
#undef TARGET_FUNCTION_VALUE
|
462 |
|
|
#define TARGET_FUNCTION_VALUE sh_function_value
|
463 |
|
|
#undef TARGET_LIBCALL_VALUE
|
464 |
|
|
#define TARGET_LIBCALL_VALUE sh_libcall_value
|
465 |
|
|
#undef TARGET_STRUCT_VALUE_RTX
|
466 |
|
|
#define TARGET_STRUCT_VALUE_RTX sh_struct_value_rtx
|
467 |
|
|
#undef TARGET_RETURN_IN_MEMORY
|
468 |
|
|
#define TARGET_RETURN_IN_MEMORY sh_return_in_memory
|
469 |
|
|
|
470 |
|
|
#undef TARGET_EXPAND_BUILTIN_SAVEREGS
|
471 |
|
|
#define TARGET_EXPAND_BUILTIN_SAVEREGS sh_builtin_saveregs
|
472 |
|
|
#undef TARGET_SETUP_INCOMING_VARARGS
|
473 |
|
|
#define TARGET_SETUP_INCOMING_VARARGS sh_setup_incoming_varargs
|
474 |
|
|
#undef TARGET_STRICT_ARGUMENT_NAMING
|
475 |
|
|
#define TARGET_STRICT_ARGUMENT_NAMING sh_strict_argument_naming
|
476 |
|
|
#undef TARGET_PRETEND_OUTGOING_VARARGS_NAMED
|
477 |
|
|
#define TARGET_PRETEND_OUTGOING_VARARGS_NAMED sh_pretend_outgoing_varargs_named
|
478 |
|
|
#undef TARGET_MUST_PASS_IN_STACK
|
479 |
|
|
#define TARGET_MUST_PASS_IN_STACK must_pass_in_stack_var_size
|
480 |
|
|
#undef TARGET_PASS_BY_REFERENCE
|
481 |
|
|
#define TARGET_PASS_BY_REFERENCE sh_pass_by_reference
|
482 |
|
|
#undef TARGET_CALLEE_COPIES
|
483 |
|
|
#define TARGET_CALLEE_COPIES sh_callee_copies
|
484 |
|
|
#undef TARGET_ARG_PARTIAL_BYTES
|
485 |
|
|
#define TARGET_ARG_PARTIAL_BYTES sh_arg_partial_bytes
|
486 |
|
|
|
487 |
|
|
#undef TARGET_BUILD_BUILTIN_VA_LIST
|
488 |
|
|
#define TARGET_BUILD_BUILTIN_VA_LIST sh_build_builtin_va_list
|
489 |
|
|
#undef TARGET_EXPAND_BUILTIN_VA_START
|
490 |
|
|
#define TARGET_EXPAND_BUILTIN_VA_START sh_va_start
|
491 |
|
|
#undef TARGET_GIMPLIFY_VA_ARG_EXPR
|
492 |
|
|
#define TARGET_GIMPLIFY_VA_ARG_EXPR sh_gimplify_va_arg_expr
|
493 |
|
|
|
494 |
|
|
#undef TARGET_SCALAR_MODE_SUPPORTED_P
|
495 |
|
|
#define TARGET_SCALAR_MODE_SUPPORTED_P sh_scalar_mode_supported_p
|
496 |
|
|
#undef TARGET_VECTOR_MODE_SUPPORTED_P
|
497 |
|
|
#define TARGET_VECTOR_MODE_SUPPORTED_P sh_vector_mode_supported_p
|
498 |
|
|
|
499 |
|
|
#undef TARGET_CHECK_PCH_TARGET_FLAGS
|
500 |
|
|
#define TARGET_CHECK_PCH_TARGET_FLAGS sh_check_pch_target_flags
|
501 |
|
|
|
502 |
|
|
#undef TARGET_DWARF_CALLING_CONVENTION
|
503 |
|
|
#define TARGET_DWARF_CALLING_CONVENTION sh_dwarf_calling_convention
|
504 |
|
|
|
505 |
|
|
/* Return regmode weight for insn. */
|
506 |
|
|
#define INSN_REGMODE_WEIGHT(INSN, MODE) regmode_weight[((MODE) == SImode) ? 0 : 1][INSN_UID (INSN)]
|
507 |
|
|
|
508 |
|
|
/* Return current register pressure for regmode. */
|
509 |
|
|
#define CURR_REGMODE_PRESSURE(MODE) curr_regmode_pressure[((MODE) == SImode) ? 0 : 1]
|
510 |
|
|
|
511 |
|
|
#undef TARGET_ENCODE_SECTION_INFO
|
512 |
|
|
#define TARGET_ENCODE_SECTION_INFO sh_encode_section_info
|
513 |
|
|
|
514 |
|
|
#ifdef SYMBIAN
|
515 |
|
|
|
516 |
|
|
#undef TARGET_ENCODE_SECTION_INFO
|
517 |
|
|
#define TARGET_ENCODE_SECTION_INFO sh_symbian_encode_section_info
|
518 |
|
|
#undef TARGET_STRIP_NAME_ENCODING
|
519 |
|
|
#define TARGET_STRIP_NAME_ENCODING sh_symbian_strip_name_encoding
|
520 |
|
|
#undef TARGET_CXX_IMPORT_EXPORT_CLASS
|
521 |
|
|
#define TARGET_CXX_IMPORT_EXPORT_CLASS sh_symbian_import_export_class
|
522 |
|
|
|
523 |
|
|
#endif /* SYMBIAN */
|
524 |
|
|
|
525 |
|
|
#undef TARGET_SECONDARY_RELOAD
|
526 |
|
|
#define TARGET_SECONDARY_RELOAD sh_secondary_reload
|
527 |
|
|
|
528 |
|
|
#undef TARGET_LEGITIMATE_ADDRESS_P
|
529 |
|
|
#define TARGET_LEGITIMATE_ADDRESS_P sh_legitimate_address_p
|
530 |
|
|
|
531 |
|
|
#undef TARGET_TRAMPOLINE_INIT
|
532 |
|
|
#define TARGET_TRAMPOLINE_INIT sh_trampoline_init
|
533 |
|
|
#undef TARGET_TRAMPOLINE_ADJUST_ADDRESS
|
534 |
|
|
#define TARGET_TRAMPOLINE_ADJUST_ADDRESS sh_trampoline_adjust_address
|
535 |
|
|
|
536 |
|
|
/* Machine-specific symbol_ref flags. */
|
537 |
|
|
#define SYMBOL_FLAG_FUNCVEC_FUNCTION (SYMBOL_FLAG_MACH_DEP << 0)
|
538 |
|
|
|
539 |
|
|
struct gcc_target targetm = TARGET_INITIALIZER;
|
540 |
|
|
|
541 |
|
|
/* Implement TARGET_HANDLE_OPTION. */
|
542 |
|
|
|
543 |
|
|
static bool
|
544 |
|
|
sh_handle_option (size_t code, const char *arg ATTRIBUTE_UNUSED,
|
545 |
|
|
int value ATTRIBUTE_UNUSED)
|
546 |
|
|
{
|
547 |
|
|
switch (code)
|
548 |
|
|
{
|
549 |
|
|
case OPT_m1:
|
550 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH1;
|
551 |
|
|
return true;
|
552 |
|
|
|
553 |
|
|
case OPT_m2:
|
554 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH2;
|
555 |
|
|
return true;
|
556 |
|
|
|
557 |
|
|
case OPT_m2a:
|
558 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH2A;
|
559 |
|
|
return true;
|
560 |
|
|
|
561 |
|
|
case OPT_m2a_nofpu:
|
562 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH2A_NOFPU;
|
563 |
|
|
return true;
|
564 |
|
|
|
565 |
|
|
case OPT_m2a_single:
|
566 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH2A_SINGLE;
|
567 |
|
|
return true;
|
568 |
|
|
|
569 |
|
|
case OPT_m2a_single_only:
|
570 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH2A_SINGLE_ONLY;
|
571 |
|
|
return true;
|
572 |
|
|
|
573 |
|
|
case OPT_m2e:
|
574 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH2E;
|
575 |
|
|
return true;
|
576 |
|
|
|
577 |
|
|
case OPT_m3:
|
578 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH3;
|
579 |
|
|
return true;
|
580 |
|
|
|
581 |
|
|
case OPT_m3e:
|
582 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH3E;
|
583 |
|
|
return true;
|
584 |
|
|
|
585 |
|
|
case OPT_m4:
|
586 |
|
|
case OPT_m4_100:
|
587 |
|
|
case OPT_m4_200:
|
588 |
|
|
case OPT_m4_300:
|
589 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH4;
|
590 |
|
|
return true;
|
591 |
|
|
|
592 |
|
|
case OPT_m4_nofpu:
|
593 |
|
|
case OPT_m4_100_nofpu:
|
594 |
|
|
case OPT_m4_200_nofpu:
|
595 |
|
|
case OPT_m4_300_nofpu:
|
596 |
|
|
case OPT_m4_340:
|
597 |
|
|
case OPT_m4_400:
|
598 |
|
|
case OPT_m4_500:
|
599 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH4_NOFPU;
|
600 |
|
|
return true;
|
601 |
|
|
|
602 |
|
|
case OPT_m4_single:
|
603 |
|
|
case OPT_m4_100_single:
|
604 |
|
|
case OPT_m4_200_single:
|
605 |
|
|
case OPT_m4_300_single:
|
606 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH4_SINGLE;
|
607 |
|
|
return true;
|
608 |
|
|
|
609 |
|
|
case OPT_m4_single_only:
|
610 |
|
|
case OPT_m4_100_single_only:
|
611 |
|
|
case OPT_m4_200_single_only:
|
612 |
|
|
case OPT_m4_300_single_only:
|
613 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH4_SINGLE_ONLY;
|
614 |
|
|
return true;
|
615 |
|
|
|
616 |
|
|
case OPT_m4a:
|
617 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH4A;
|
618 |
|
|
return true;
|
619 |
|
|
|
620 |
|
|
case OPT_m4a_nofpu:
|
621 |
|
|
case OPT_m4al:
|
622 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH4A_NOFPU;
|
623 |
|
|
return true;
|
624 |
|
|
|
625 |
|
|
case OPT_m4a_single:
|
626 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH4A_SINGLE;
|
627 |
|
|
return true;
|
628 |
|
|
|
629 |
|
|
case OPT_m4a_single_only:
|
630 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH4A_SINGLE_ONLY;
|
631 |
|
|
return true;
|
632 |
|
|
|
633 |
|
|
case OPT_m5_32media:
|
634 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH5_32MEDIA;
|
635 |
|
|
return true;
|
636 |
|
|
|
637 |
|
|
case OPT_m5_32media_nofpu:
|
638 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH5_32MEDIA_NOFPU;
|
639 |
|
|
return true;
|
640 |
|
|
|
641 |
|
|
case OPT_m5_64media:
|
642 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH5_64MEDIA;
|
643 |
|
|
return true;
|
644 |
|
|
|
645 |
|
|
case OPT_m5_64media_nofpu:
|
646 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH5_64MEDIA_NOFPU;
|
647 |
|
|
return true;
|
648 |
|
|
|
649 |
|
|
case OPT_m5_compact:
|
650 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH5_COMPACT;
|
651 |
|
|
return true;
|
652 |
|
|
|
653 |
|
|
case OPT_m5_compact_nofpu:
|
654 |
|
|
target_flags = (target_flags & ~MASK_ARCH) | SELECT_SH5_COMPACT_NOFPU;
|
655 |
|
|
return true;
|
656 |
|
|
|
657 |
|
|
default:
|
658 |
|
|
return true;
|
659 |
|
|
}
|
660 |
|
|
}
|
661 |
|
|
|
662 |
|
|
/* Set default optimization options. */
|
663 |
|
|
void
|
664 |
|
|
sh_optimization_options (int level ATTRIBUTE_UNUSED, int size ATTRIBUTE_UNUSED)
|
665 |
|
|
{
|
666 |
|
|
if (level)
|
667 |
|
|
{
|
668 |
|
|
flag_omit_frame_pointer = 2;
|
669 |
|
|
if (!size)
|
670 |
|
|
sh_div_str = "inv:minlat";
|
671 |
|
|
}
|
672 |
|
|
if (size)
|
673 |
|
|
{
|
674 |
|
|
target_flags |= MASK_SMALLCODE;
|
675 |
|
|
sh_div_str = SH_DIV_STR_FOR_SIZE ;
|
676 |
|
|
}
|
677 |
|
|
else
|
678 |
|
|
TARGET_CBRANCHDI4 = 1;
|
679 |
|
|
/* We can't meaningfully test TARGET_SHMEDIA here, because -m options
|
680 |
|
|
haven't been parsed yet, hence we'd read only the default.
|
681 |
|
|
sh_target_reg_class will return NO_REGS if this is not SHMEDIA, so
|
682 |
|
|
it's OK to always set flag_branch_target_load_optimize. */
|
683 |
|
|
if (level > 1)
|
684 |
|
|
{
|
685 |
|
|
flag_branch_target_load_optimize = 1;
|
686 |
|
|
if (!size)
|
687 |
|
|
target_flags |= MASK_SAVE_ALL_TARGET_REGS;
|
688 |
|
|
}
|
689 |
|
|
/* Likewise, we can't meaningfully test TARGET_SH2E / TARGET_IEEE
|
690 |
|
|
here, so leave it to OVERRIDE_OPTIONS to set
|
691 |
|
|
flag_finite_math_only. We set it to 2 here so we know if the user
|
692 |
|
|
explicitly requested this to be on or off. */
|
693 |
|
|
flag_finite_math_only = 2;
|
694 |
|
|
/* If flag_schedule_insns is 1, we set it to 2 here so we know if
|
695 |
|
|
the user explicitly requested this to be on or off. */
|
696 |
|
|
if (flag_schedule_insns > 0)
|
697 |
|
|
flag_schedule_insns = 2;
|
698 |
|
|
|
699 |
|
|
set_param_value ("simultaneous-prefetches", 2);
|
700 |
|
|
}
|
701 |
|
|
|
702 |
|
|
/* Implement OVERRIDE_OPTIONS macro. Validate and override various
|
703 |
|
|
options, and do some machine dependent initialization. */
|
704 |
|
|
void
|
705 |
|
|
sh_override_options (void)
|
706 |
|
|
{
|
707 |
|
|
int regno;
|
708 |
|
|
|
709 |
|
|
SUBTARGET_OVERRIDE_OPTIONS;
|
710 |
|
|
if (flag_finite_math_only == 2)
|
711 |
|
|
flag_finite_math_only
|
712 |
|
|
= !flag_signaling_nans && TARGET_SH2E && ! TARGET_IEEE;
|
713 |
|
|
if (TARGET_SH2E && !flag_finite_math_only)
|
714 |
|
|
target_flags |= MASK_IEEE;
|
715 |
|
|
sh_cpu = PROCESSOR_SH1;
|
716 |
|
|
assembler_dialect = 0;
|
717 |
|
|
if (TARGET_SH2)
|
718 |
|
|
sh_cpu = PROCESSOR_SH2;
|
719 |
|
|
if (TARGET_SH2E)
|
720 |
|
|
sh_cpu = PROCESSOR_SH2E;
|
721 |
|
|
if (TARGET_SH2A)
|
722 |
|
|
sh_cpu = PROCESSOR_SH2A;
|
723 |
|
|
if (TARGET_SH3)
|
724 |
|
|
sh_cpu = PROCESSOR_SH3;
|
725 |
|
|
if (TARGET_SH3E)
|
726 |
|
|
sh_cpu = PROCESSOR_SH3E;
|
727 |
|
|
if (TARGET_SH4)
|
728 |
|
|
{
|
729 |
|
|
assembler_dialect = 1;
|
730 |
|
|
sh_cpu = PROCESSOR_SH4;
|
731 |
|
|
}
|
732 |
|
|
if (TARGET_SH4A_ARCH)
|
733 |
|
|
{
|
734 |
|
|
assembler_dialect = 1;
|
735 |
|
|
sh_cpu = PROCESSOR_SH4A;
|
736 |
|
|
}
|
737 |
|
|
if (TARGET_SH5)
|
738 |
|
|
{
|
739 |
|
|
sh_cpu = PROCESSOR_SH5;
|
740 |
|
|
target_flags |= MASK_ALIGN_DOUBLE;
|
741 |
|
|
if (TARGET_SHMEDIA_FPU)
|
742 |
|
|
target_flags |= MASK_FMOVD;
|
743 |
|
|
if (TARGET_SHMEDIA)
|
744 |
|
|
{
|
745 |
|
|
/* There are no delay slots on SHmedia. */
|
746 |
|
|
flag_delayed_branch = 0;
|
747 |
|
|
/* Relaxation isn't yet supported for SHmedia */
|
748 |
|
|
target_flags &= ~MASK_RELAX;
|
749 |
|
|
/* After reload, if conversion does little good but can cause
|
750 |
|
|
ICEs:
|
751 |
|
|
- find_if_block doesn't do anything for SH because we don't
|
752 |
|
|
have conditional execution patterns. (We use conditional
|
753 |
|
|
move patterns, which are handled differently, and only
|
754 |
|
|
before reload).
|
755 |
|
|
- find_cond_trap doesn't do anything for the SH because we
|
756 |
|
|
don't have conditional traps.
|
757 |
|
|
- find_if_case_1 uses redirect_edge_and_branch_force in
|
758 |
|
|
the only path that does an optimization, and this causes
|
759 |
|
|
an ICE when branch targets are in registers.
|
760 |
|
|
- find_if_case_2 doesn't do anything for the SHmedia after
|
761 |
|
|
reload except when it can redirect a tablejump - and
|
762 |
|
|
that's rather rare. */
|
763 |
|
|
flag_if_conversion2 = 0;
|
764 |
|
|
if (! strcmp (sh_div_str, "call"))
|
765 |
|
|
sh_div_strategy = SH_DIV_CALL;
|
766 |
|
|
else if (! strcmp (sh_div_str, "call2"))
|
767 |
|
|
sh_div_strategy = SH_DIV_CALL2;
|
768 |
|
|
if (! strcmp (sh_div_str, "fp") && TARGET_FPU_ANY)
|
769 |
|
|
sh_div_strategy = SH_DIV_FP;
|
770 |
|
|
else if (! strcmp (sh_div_str, "inv"))
|
771 |
|
|
sh_div_strategy = SH_DIV_INV;
|
772 |
|
|
else if (! strcmp (sh_div_str, "inv:minlat"))
|
773 |
|
|
sh_div_strategy = SH_DIV_INV_MINLAT;
|
774 |
|
|
else if (! strcmp (sh_div_str, "inv20u"))
|
775 |
|
|
sh_div_strategy = SH_DIV_INV20U;
|
776 |
|
|
else if (! strcmp (sh_div_str, "inv20l"))
|
777 |
|
|
sh_div_strategy = SH_DIV_INV20L;
|
778 |
|
|
else if (! strcmp (sh_div_str, "inv:call2"))
|
779 |
|
|
sh_div_strategy = SH_DIV_INV_CALL2;
|
780 |
|
|
else if (! strcmp (sh_div_str, "inv:call"))
|
781 |
|
|
sh_div_strategy = SH_DIV_INV_CALL;
|
782 |
|
|
else if (! strcmp (sh_div_str, "inv:fp"))
|
783 |
|
|
{
|
784 |
|
|
if (TARGET_FPU_ANY)
|
785 |
|
|
sh_div_strategy = SH_DIV_INV_FP;
|
786 |
|
|
else
|
787 |
|
|
sh_div_strategy = SH_DIV_INV;
|
788 |
|
|
}
|
789 |
|
|
TARGET_CBRANCHDI4 = 0;
|
790 |
|
|
/* Assembler CFI isn't yet fully supported for SHmedia. */
|
791 |
|
|
flag_dwarf2_cfi_asm = 0;
|
792 |
|
|
}
|
793 |
|
|
}
|
794 |
|
|
else
|
795 |
|
|
{
|
796 |
|
|
/* Only the sh64-elf assembler fully supports .quad properly. */
|
797 |
|
|
targetm.asm_out.aligned_op.di = NULL;
|
798 |
|
|
targetm.asm_out.unaligned_op.di = NULL;
|
799 |
|
|
}
|
800 |
|
|
if (TARGET_SH1)
|
801 |
|
|
{
|
802 |
|
|
if (! strcmp (sh_div_str, "call-div1"))
|
803 |
|
|
sh_div_strategy = SH_DIV_CALL_DIV1;
|
804 |
|
|
else if (! strcmp (sh_div_str, "call-fp")
|
805 |
|
|
&& (TARGET_FPU_DOUBLE
|
806 |
|
|
|| (TARGET_HARD_SH4 && TARGET_SH2E)
|
807 |
|
|
|| (TARGET_SHCOMPACT && TARGET_FPU_ANY)))
|
808 |
|
|
sh_div_strategy = SH_DIV_CALL_FP;
|
809 |
|
|
else if (! strcmp (sh_div_str, "call-table") && TARGET_SH2)
|
810 |
|
|
sh_div_strategy = SH_DIV_CALL_TABLE;
|
811 |
|
|
else
|
812 |
|
|
/* Pick one that makes most sense for the target in general.
|
813 |
|
|
It is not much good to use different functions depending
|
814 |
|
|
on -Os, since then we'll end up with two different functions
|
815 |
|
|
when some of the code is compiled for size, and some for
|
816 |
|
|
speed. */
|
817 |
|
|
|
818 |
|
|
/* SH4 tends to emphasize speed. */
|
819 |
|
|
if (TARGET_HARD_SH4)
|
820 |
|
|
sh_div_strategy = SH_DIV_CALL_TABLE;
|
821 |
|
|
/* These have their own way of doing things. */
|
822 |
|
|
else if (TARGET_SH2A)
|
823 |
|
|
sh_div_strategy = SH_DIV_INTRINSIC;
|
824 |
|
|
/* ??? Should we use the integer SHmedia function instead? */
|
825 |
|
|
else if (TARGET_SHCOMPACT && TARGET_FPU_ANY)
|
826 |
|
|
sh_div_strategy = SH_DIV_CALL_FP;
|
827 |
|
|
/* SH1 .. SH3 cores often go into small-footprint systems, so
|
828 |
|
|
default to the smallest implementation available. */
|
829 |
|
|
else if (TARGET_SH2) /* ??? EXPERIMENTAL */
|
830 |
|
|
sh_div_strategy = SH_DIV_CALL_TABLE;
|
831 |
|
|
else
|
832 |
|
|
sh_div_strategy = SH_DIV_CALL_DIV1;
|
833 |
|
|
}
|
834 |
|
|
if (!TARGET_SH1)
|
835 |
|
|
TARGET_PRETEND_CMOVE = 0;
|
836 |
|
|
if (sh_divsi3_libfunc[0])
|
837 |
|
|
; /* User supplied - leave it alone. */
|
838 |
|
|
else if (TARGET_DIVIDE_CALL_FP)
|
839 |
|
|
sh_divsi3_libfunc = "__sdivsi3_i4";
|
840 |
|
|
else if (TARGET_DIVIDE_CALL_TABLE)
|
841 |
|
|
sh_divsi3_libfunc = "__sdivsi3_i4i";
|
842 |
|
|
else if (TARGET_SH5)
|
843 |
|
|
sh_divsi3_libfunc = "__sdivsi3_1";
|
844 |
|
|
else
|
845 |
|
|
sh_divsi3_libfunc = "__sdivsi3";
|
846 |
|
|
if (sh_branch_cost == -1)
|
847 |
|
|
sh_branch_cost
|
848 |
|
|
= TARGET_SH5 ? 1 : ! TARGET_SH2 || TARGET_HARD_SH4 ? 2 : 1;
|
849 |
|
|
|
850 |
|
|
for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
|
851 |
|
|
if (! VALID_REGISTER_P (regno))
|
852 |
|
|
sh_register_names[regno][0] = '\0';
|
853 |
|
|
|
854 |
|
|
for (regno = 0; regno < ADDREGNAMES_SIZE; regno++)
|
855 |
|
|
if (! VALID_REGISTER_P (ADDREGNAMES_REGNO (regno)))
|
856 |
|
|
sh_additional_register_names[regno][0] = '\0';
|
857 |
|
|
|
858 |
|
|
if (flag_omit_frame_pointer == 2)
|
859 |
|
|
{
|
860 |
|
|
/* The debugging information is sufficient,
|
861 |
|
|
but gdb doesn't implement this yet */
|
862 |
|
|
if (0)
|
863 |
|
|
flag_omit_frame_pointer
|
864 |
|
|
= (PREFERRED_DEBUGGING_TYPE == DWARF2_DEBUG);
|
865 |
|
|
else
|
866 |
|
|
flag_omit_frame_pointer = 0;
|
867 |
|
|
}
|
868 |
|
|
|
869 |
|
|
if ((flag_pic && ! TARGET_PREFERGOT)
|
870 |
|
|
|| (TARGET_SHMEDIA && !TARGET_PT_FIXED))
|
871 |
|
|
flag_no_function_cse = 1;
|
872 |
|
|
|
873 |
|
|
if (SMALL_REGISTER_CLASSES)
|
874 |
|
|
{
|
875 |
|
|
/* Never run scheduling before reload, since that can
|
876 |
|
|
break global alloc, and generates slower code anyway due
|
877 |
|
|
to the pressure on R0. */
|
878 |
|
|
/* Enable sched1 for SH4 if the user explicitly requests.
|
879 |
|
|
When sched1 is enabled, the ready queue will be reordered by
|
880 |
|
|
the target hooks if pressure is high. We can not do this for
|
881 |
|
|
PIC, SH3 and lower as they give spill failures for R0. */
|
882 |
|
|
if (!TARGET_HARD_SH4 || flag_pic)
|
883 |
|
|
flag_schedule_insns = 0;
|
884 |
|
|
/* ??? Current exception handling places basic block boundaries
|
885 |
|
|
after call_insns. It causes the high pressure on R0 and gives
|
886 |
|
|
spill failures for R0 in reload. See PR 22553 and the thread
|
887 |
|
|
on gcc-patches
|
888 |
|
|
<http://gcc.gnu.org/ml/gcc-patches/2005-10/msg00816.html>. */
|
889 |
|
|
else if (flag_exceptions)
|
890 |
|
|
{
|
891 |
|
|
if (flag_schedule_insns == 1)
|
892 |
|
|
warning (0, "ignoring -fschedule-insns because of exception handling bug");
|
893 |
|
|
flag_schedule_insns = 0;
|
894 |
|
|
}
|
895 |
|
|
else if (flag_schedule_insns == 2)
|
896 |
|
|
flag_schedule_insns = 0;
|
897 |
|
|
}
|
898 |
|
|
|
899 |
|
|
/* Unwinding with -freorder-blocks-and-partition does not work on this
|
900 |
|
|
architecture, because it requires far jumps to label crossing between
|
901 |
|
|
hot/cold sections which are rejected on this architecture. */
|
902 |
|
|
if (flag_reorder_blocks_and_partition)
|
903 |
|
|
{
|
904 |
|
|
if (flag_exceptions)
|
905 |
|
|
{
|
906 |
|
|
inform (input_location,
|
907 |
|
|
"-freorder-blocks-and-partition does not work with "
|
908 |
|
|
"exceptions on this architecture");
|
909 |
|
|
flag_reorder_blocks_and_partition = 0;
|
910 |
|
|
flag_reorder_blocks = 1;
|
911 |
|
|
}
|
912 |
|
|
else if (flag_unwind_tables)
|
913 |
|
|
{
|
914 |
|
|
inform (input_location,
|
915 |
|
|
"-freorder-blocks-and-partition does not support unwind "
|
916 |
|
|
"info on this architecture");
|
917 |
|
|
flag_reorder_blocks_and_partition = 0;
|
918 |
|
|
flag_reorder_blocks = 1;
|
919 |
|
|
}
|
920 |
|
|
}
|
921 |
|
|
|
922 |
|
|
if (align_loops == 0)
|
923 |
|
|
align_loops = 1 << (TARGET_SH5 ? 3 : 2);
|
924 |
|
|
if (align_jumps == 0)
|
925 |
|
|
align_jumps = 1 << CACHE_LOG;
|
926 |
|
|
else if (align_jumps < (TARGET_SHMEDIA ? 4 : 2))
|
927 |
|
|
align_jumps = TARGET_SHMEDIA ? 4 : 2;
|
928 |
|
|
|
929 |
|
|
/* Allocation boundary (in *bytes*) for the code of a function.
|
930 |
|
|
SH1: 32 bit alignment is faster, because instructions are always
|
931 |
|
|
fetched as a pair from a longword boundary.
|
932 |
|
|
SH2 .. SH5 : align to cache line start. */
|
933 |
|
|
if (align_functions == 0)
|
934 |
|
|
align_functions
|
935 |
|
|
= TARGET_SMALLCODE ? FUNCTION_BOUNDARY/8 : (1 << CACHE_LOG);
|
936 |
|
|
/* The linker relaxation code breaks when a function contains
|
937 |
|
|
alignments that are larger than that at the start of a
|
938 |
|
|
compilation unit. */
|
939 |
|
|
if (TARGET_RELAX)
|
940 |
|
|
{
|
941 |
|
|
int min_align
|
942 |
|
|
= align_loops > align_jumps ? align_loops : align_jumps;
|
943 |
|
|
|
944 |
|
|
/* Also take possible .long constants / mova tables int account. */
|
945 |
|
|
if (min_align < 4)
|
946 |
|
|
min_align = 4;
|
947 |
|
|
if (align_functions < min_align)
|
948 |
|
|
align_functions = min_align;
|
949 |
|
|
}
|
950 |
|
|
|
951 |
|
|
if (sh_fixed_range_str)
|
952 |
|
|
sh_fix_range (sh_fixed_range_str);
|
953 |
|
|
}
|
954 |
|
|
|
955 |
|
|
/* Print the operand address in x to the stream. */
|
956 |
|
|
|
957 |
|
|
void
|
958 |
|
|
print_operand_address (FILE *stream, rtx x)
|
959 |
|
|
{
|
960 |
|
|
switch (GET_CODE (x))
|
961 |
|
|
{
|
962 |
|
|
case REG:
|
963 |
|
|
case SUBREG:
|
964 |
|
|
fprintf (stream, "@%s", reg_names[true_regnum (x)]);
|
965 |
|
|
break;
|
966 |
|
|
|
967 |
|
|
case PLUS:
|
968 |
|
|
{
|
969 |
|
|
rtx base = XEXP (x, 0);
|
970 |
|
|
rtx index = XEXP (x, 1);
|
971 |
|
|
|
972 |
|
|
switch (GET_CODE (index))
|
973 |
|
|
{
|
974 |
|
|
case CONST_INT:
|
975 |
|
|
fprintf (stream, "@(%d,%s)", (int) INTVAL (index),
|
976 |
|
|
reg_names[true_regnum (base)]);
|
977 |
|
|
break;
|
978 |
|
|
|
979 |
|
|
case REG:
|
980 |
|
|
case SUBREG:
|
981 |
|
|
{
|
982 |
|
|
int base_num = true_regnum (base);
|
983 |
|
|
int index_num = true_regnum (index);
|
984 |
|
|
|
985 |
|
|
fprintf (stream, "@(r0,%s)",
|
986 |
|
|
reg_names[MAX (base_num, index_num)]);
|
987 |
|
|
break;
|
988 |
|
|
}
|
989 |
|
|
|
990 |
|
|
default:
|
991 |
|
|
gcc_unreachable ();
|
992 |
|
|
}
|
993 |
|
|
}
|
994 |
|
|
break;
|
995 |
|
|
|
996 |
|
|
case PRE_DEC:
|
997 |
|
|
fprintf (stream, "@-%s", reg_names[true_regnum (XEXP (x, 0))]);
|
998 |
|
|
break;
|
999 |
|
|
|
1000 |
|
|
case POST_INC:
|
1001 |
|
|
fprintf (stream, "@%s+", reg_names[true_regnum (XEXP (x, 0))]);
|
1002 |
|
|
break;
|
1003 |
|
|
|
1004 |
|
|
default:
|
1005 |
|
|
x = mark_constant_pool_use (x);
|
1006 |
|
|
output_addr_const (stream, x);
|
1007 |
|
|
break;
|
1008 |
|
|
}
|
1009 |
|
|
}
|
1010 |
|
|
|
1011 |
|
|
/* Print operand x (an rtx) in assembler syntax to file stream
|
1012 |
|
|
according to modifier code.
|
1013 |
|
|
|
1014 |
|
|
'.' print a .s if insn needs delay slot
|
1015 |
|
|
',' print LOCAL_LABEL_PREFIX
|
1016 |
|
|
'@' print trap, rte or rts depending upon pragma interruptness
|
1017 |
|
|
'#' output a nop if there is nothing to put in the delay slot
|
1018 |
|
|
''' print likelihood suffix (/u for unlikely).
|
1019 |
|
|
'>' print branch target if -fverbose-asm
|
1020 |
|
|
'O' print a constant without the #
|
1021 |
|
|
'R' print the LSW of a dp value - changes if in little endian
|
1022 |
|
|
'S' print the MSW of a dp value - changes if in little endian
|
1023 |
|
|
'T' print the next word of a dp value - same as 'R' in big endian mode.
|
1024 |
|
|
'M' SHMEDIA: print an `x' if `m' will print `base,index'.
|
1025 |
|
|
otherwise: print .b / .w / .l / .s / .d suffix if operand is a MEM.
|
1026 |
|
|
'N' print 'r63' if the operand is (const_int 0).
|
1027 |
|
|
'd' print a V2SF reg as dN instead of fpN.
|
1028 |
|
|
'm' print a pair `base,offset' or `base,index', for LD and ST.
|
1029 |
|
|
'U' Likewise for {LD,ST}{HI,LO}.
|
1030 |
|
|
'V' print the position of a single bit set.
|
1031 |
|
|
'W' print the position of a single bit cleared.
|
1032 |
|
|
't' print a memory address which is a register.
|
1033 |
|
|
'u' prints the lowest 16 bits of CONST_INT, as an unsigned value.
|
1034 |
|
|
'o' output an operator. */
|
1035 |
|
|
|
1036 |
|
|
void
|
1037 |
|
|
print_operand (FILE *stream, rtx x, int code)
|
1038 |
|
|
{
|
1039 |
|
|
int regno;
|
1040 |
|
|
enum machine_mode mode;
|
1041 |
|
|
|
1042 |
|
|
switch (code)
|
1043 |
|
|
{
|
1044 |
|
|
tree trapa_attr;
|
1045 |
|
|
|
1046 |
|
|
case '.':
|
1047 |
|
|
if (final_sequence
|
1048 |
|
|
&& ! INSN_ANNULLED_BRANCH_P (XVECEXP (final_sequence, 0, 0))
|
1049 |
|
|
&& get_attr_length (XVECEXP (final_sequence, 0, 1)))
|
1050 |
|
|
fprintf (stream, ASSEMBLER_DIALECT ? "/s" : ".s");
|
1051 |
|
|
break;
|
1052 |
|
|
case ',':
|
1053 |
|
|
fprintf (stream, "%s", LOCAL_LABEL_PREFIX);
|
1054 |
|
|
break;
|
1055 |
|
|
case '@':
|
1056 |
|
|
trapa_attr = lookup_attribute ("trap_exit",
|
1057 |
|
|
DECL_ATTRIBUTES (current_function_decl));
|
1058 |
|
|
if (trapa_attr)
|
1059 |
|
|
fprintf (stream, "trapa #%ld",
|
1060 |
|
|
(long) TREE_INT_CST_LOW (TREE_VALUE (TREE_VALUE (trapa_attr))));
|
1061 |
|
|
else if (sh_cfun_interrupt_handler_p ())
|
1062 |
|
|
{
|
1063 |
|
|
if (sh_cfun_resbank_handler_p ())
|
1064 |
|
|
fprintf (stream, "resbank\n");
|
1065 |
|
|
fprintf (stream, "rte");
|
1066 |
|
|
}
|
1067 |
|
|
else
|
1068 |
|
|
fprintf (stream, "rts");
|
1069 |
|
|
break;
|
1070 |
|
|
case '#':
|
1071 |
|
|
/* Output a nop if there's nothing in the delay slot. */
|
1072 |
|
|
if (dbr_sequence_length () == 0)
|
1073 |
|
|
fprintf (stream, "\n\tnop");
|
1074 |
|
|
break;
|
1075 |
|
|
case '\'':
|
1076 |
|
|
{
|
1077 |
|
|
rtx note = find_reg_note (current_output_insn, REG_BR_PROB, 0);
|
1078 |
|
|
|
1079 |
|
|
if (note && INTVAL (XEXP (note, 0)) * 2 < REG_BR_PROB_BASE)
|
1080 |
|
|
fputs ("/u", stream);
|
1081 |
|
|
break;
|
1082 |
|
|
}
|
1083 |
|
|
case '>':
|
1084 |
|
|
if (flag_verbose_asm && JUMP_LABEL (current_output_insn))
|
1085 |
|
|
{
|
1086 |
|
|
fputs ("\t! target: ", stream);
|
1087 |
|
|
output_addr_const (stream, JUMP_LABEL (current_output_insn));
|
1088 |
|
|
}
|
1089 |
|
|
break;
|
1090 |
|
|
case 'O':
|
1091 |
|
|
x = mark_constant_pool_use (x);
|
1092 |
|
|
output_addr_const (stream, x);
|
1093 |
|
|
break;
|
1094 |
|
|
/* N.B.: %R / %S / %T adjust memory addresses by four.
|
1095 |
|
|
For SHMEDIA, that means they can be used to access the first and
|
1096 |
|
|
second 32 bit part of a 64 bit (or larger) value that
|
1097 |
|
|
might be held in floating point registers or memory.
|
1098 |
|
|
While they can be used to access 64 bit parts of a larger value
|
1099 |
|
|
held in general purpose registers, that won't work with memory -
|
1100 |
|
|
neither for fp registers, since the frxx names are used. */
|
1101 |
|
|
case 'R':
|
1102 |
|
|
if (REG_P (x) || GET_CODE (x) == SUBREG)
|
1103 |
|
|
{
|
1104 |
|
|
regno = true_regnum (x);
|
1105 |
|
|
regno += FP_REGISTER_P (regno) ? 1 : LSW;
|
1106 |
|
|
fputs (reg_names[regno], (stream));
|
1107 |
|
|
}
|
1108 |
|
|
else if (MEM_P (x))
|
1109 |
|
|
{
|
1110 |
|
|
x = adjust_address (x, SImode, 4 * LSW);
|
1111 |
|
|
print_operand_address (stream, XEXP (x, 0));
|
1112 |
|
|
}
|
1113 |
|
|
else
|
1114 |
|
|
{
|
1115 |
|
|
rtx sub = NULL_RTX;
|
1116 |
|
|
|
1117 |
|
|
mode = GET_MODE (x);
|
1118 |
|
|
if (mode == VOIDmode)
|
1119 |
|
|
mode = DImode;
|
1120 |
|
|
if (GET_MODE_SIZE (mode) >= 8)
|
1121 |
|
|
sub = simplify_subreg (SImode, x, mode, 4 * LSW);
|
1122 |
|
|
if (sub)
|
1123 |
|
|
print_operand (stream, sub, 0);
|
1124 |
|
|
else
|
1125 |
|
|
output_operand_lossage ("invalid operand to %%R");
|
1126 |
|
|
}
|
1127 |
|
|
break;
|
1128 |
|
|
case 'S':
|
1129 |
|
|
if (REG_P (x) || GET_CODE (x) == SUBREG)
|
1130 |
|
|
{
|
1131 |
|
|
regno = true_regnum (x);
|
1132 |
|
|
regno += FP_REGISTER_P (regno) ? 0 : MSW;
|
1133 |
|
|
fputs (reg_names[regno], (stream));
|
1134 |
|
|
}
|
1135 |
|
|
else if (MEM_P (x))
|
1136 |
|
|
{
|
1137 |
|
|
x = adjust_address (x, SImode, 4 * MSW);
|
1138 |
|
|
print_operand_address (stream, XEXP (x, 0));
|
1139 |
|
|
}
|
1140 |
|
|
else
|
1141 |
|
|
{
|
1142 |
|
|
rtx sub = NULL_RTX;
|
1143 |
|
|
|
1144 |
|
|
mode = GET_MODE (x);
|
1145 |
|
|
if (mode == VOIDmode)
|
1146 |
|
|
mode = DImode;
|
1147 |
|
|
if (GET_MODE_SIZE (mode) >= 8)
|
1148 |
|
|
sub = simplify_subreg (SImode, x, mode, 4 * MSW);
|
1149 |
|
|
if (sub)
|
1150 |
|
|
print_operand (stream, sub, 0);
|
1151 |
|
|
else
|
1152 |
|
|
output_operand_lossage ("invalid operand to %%S");
|
1153 |
|
|
}
|
1154 |
|
|
break;
|
1155 |
|
|
case 'T':
|
1156 |
|
|
/* Next word of a double. */
|
1157 |
|
|
switch (GET_CODE (x))
|
1158 |
|
|
{
|
1159 |
|
|
case REG:
|
1160 |
|
|
fputs (reg_names[REGNO (x) + 1], (stream));
|
1161 |
|
|
break;
|
1162 |
|
|
case MEM:
|
1163 |
|
|
if (GET_CODE (XEXP (x, 0)) != PRE_DEC
|
1164 |
|
|
&& GET_CODE (XEXP (x, 0)) != POST_INC)
|
1165 |
|
|
x = adjust_address (x, SImode, 4);
|
1166 |
|
|
print_operand_address (stream, XEXP (x, 0));
|
1167 |
|
|
break;
|
1168 |
|
|
default:
|
1169 |
|
|
break;
|
1170 |
|
|
}
|
1171 |
|
|
break;
|
1172 |
|
|
|
1173 |
|
|
case 't':
|
1174 |
|
|
gcc_assert (MEM_P (x));
|
1175 |
|
|
x = XEXP (x, 0);
|
1176 |
|
|
switch (GET_CODE (x))
|
1177 |
|
|
{
|
1178 |
|
|
case REG:
|
1179 |
|
|
case SUBREG:
|
1180 |
|
|
print_operand (stream, x, 0);
|
1181 |
|
|
break;
|
1182 |
|
|
default:
|
1183 |
|
|
break;
|
1184 |
|
|
}
|
1185 |
|
|
break;
|
1186 |
|
|
|
1187 |
|
|
case 'o':
|
1188 |
|
|
switch (GET_CODE (x))
|
1189 |
|
|
{
|
1190 |
|
|
case PLUS: fputs ("add", stream); break;
|
1191 |
|
|
case MINUS: fputs ("sub", stream); break;
|
1192 |
|
|
case MULT: fputs ("mul", stream); break;
|
1193 |
|
|
case DIV: fputs ("div", stream); break;
|
1194 |
|
|
case EQ: fputs ("eq", stream); break;
|
1195 |
|
|
case NE: fputs ("ne", stream); break;
|
1196 |
|
|
case GT: case LT: fputs ("gt", stream); break;
|
1197 |
|
|
case GE: case LE: fputs ("ge", stream); break;
|
1198 |
|
|
case GTU: case LTU: fputs ("gtu", stream); break;
|
1199 |
|
|
case GEU: case LEU: fputs ("geu", stream); break;
|
1200 |
|
|
default:
|
1201 |
|
|
break;
|
1202 |
|
|
}
|
1203 |
|
|
break;
|
1204 |
|
|
case 'M':
|
1205 |
|
|
if (TARGET_SHMEDIA)
|
1206 |
|
|
{
|
1207 |
|
|
if (MEM_P (x)
|
1208 |
|
|
&& GET_CODE (XEXP (x, 0)) == PLUS
|
1209 |
|
|
&& (REG_P (XEXP (XEXP (x, 0), 1))
|
1210 |
|
|
|| GET_CODE (XEXP (XEXP (x, 0), 1)) == SUBREG))
|
1211 |
|
|
fputc ('x', stream);
|
1212 |
|
|
}
|
1213 |
|
|
else
|
1214 |
|
|
{
|
1215 |
|
|
if (MEM_P (x))
|
1216 |
|
|
{
|
1217 |
|
|
switch (GET_MODE (x))
|
1218 |
|
|
{
|
1219 |
|
|
case QImode: fputs (".b", stream); break;
|
1220 |
|
|
case HImode: fputs (".w", stream); break;
|
1221 |
|
|
case SImode: fputs (".l", stream); break;
|
1222 |
|
|
case SFmode: fputs (".s", stream); break;
|
1223 |
|
|
case DFmode: fputs (".d", stream); break;
|
1224 |
|
|
default: gcc_unreachable ();
|
1225 |
|
|
}
|
1226 |
|
|
}
|
1227 |
|
|
}
|
1228 |
|
|
break;
|
1229 |
|
|
|
1230 |
|
|
case 'm':
|
1231 |
|
|
gcc_assert (MEM_P (x));
|
1232 |
|
|
x = XEXP (x, 0);
|
1233 |
|
|
/* Fall through. */
|
1234 |
|
|
case 'U':
|
1235 |
|
|
switch (GET_CODE (x))
|
1236 |
|
|
{
|
1237 |
|
|
case REG:
|
1238 |
|
|
case SUBREG:
|
1239 |
|
|
print_operand (stream, x, 0);
|
1240 |
|
|
fputs (", 0", stream);
|
1241 |
|
|
break;
|
1242 |
|
|
|
1243 |
|
|
case PLUS:
|
1244 |
|
|
print_operand (stream, XEXP (x, 0), 0);
|
1245 |
|
|
fputs (", ", stream);
|
1246 |
|
|
print_operand (stream, XEXP (x, 1), 0);
|
1247 |
|
|
break;
|
1248 |
|
|
|
1249 |
|
|
default:
|
1250 |
|
|
gcc_unreachable ();
|
1251 |
|
|
}
|
1252 |
|
|
break;
|
1253 |
|
|
|
1254 |
|
|
case 'V':
|
1255 |
|
|
{
|
1256 |
|
|
int num = exact_log2 (INTVAL (x));
|
1257 |
|
|
gcc_assert (num >= 0);
|
1258 |
|
|
fprintf (stream, "#%d", num);
|
1259 |
|
|
}
|
1260 |
|
|
break;
|
1261 |
|
|
|
1262 |
|
|
case 'W':
|
1263 |
|
|
{
|
1264 |
|
|
int num = exact_log2 (~INTVAL (x));
|
1265 |
|
|
gcc_assert (num >= 0);
|
1266 |
|
|
fprintf (stream, "#%d", num);
|
1267 |
|
|
}
|
1268 |
|
|
break;
|
1269 |
|
|
|
1270 |
|
|
case 'd':
|
1271 |
|
|
gcc_assert (REG_P (x) && GET_MODE (x) == V2SFmode);
|
1272 |
|
|
|
1273 |
|
|
fprintf ((stream), "d%s", reg_names[REGNO (x)] + 1);
|
1274 |
|
|
break;
|
1275 |
|
|
|
1276 |
|
|
case 'N':
|
1277 |
|
|
if (x == CONST0_RTX (GET_MODE (x)))
|
1278 |
|
|
{
|
1279 |
|
|
fprintf ((stream), "r63");
|
1280 |
|
|
break;
|
1281 |
|
|
}
|
1282 |
|
|
goto default_output;
|
1283 |
|
|
case 'u':
|
1284 |
|
|
if (CONST_INT_P (x))
|
1285 |
|
|
{
|
1286 |
|
|
fprintf ((stream), "%u", (unsigned) INTVAL (x) & (0x10000 - 1));
|
1287 |
|
|
break;
|
1288 |
|
|
}
|
1289 |
|
|
/* Fall through. */
|
1290 |
|
|
|
1291 |
|
|
default_output:
|
1292 |
|
|
default:
|
1293 |
|
|
regno = 0;
|
1294 |
|
|
mode = GET_MODE (x);
|
1295 |
|
|
|
1296 |
|
|
switch (GET_CODE (x))
|
1297 |
|
|
{
|
1298 |
|
|
case TRUNCATE:
|
1299 |
|
|
{
|
1300 |
|
|
rtx inner = XEXP (x, 0);
|
1301 |
|
|
int offset = 0;
|
1302 |
|
|
enum machine_mode inner_mode;
|
1303 |
|
|
|
1304 |
|
|
/* We might see SUBREGs with vector mode registers inside. */
|
1305 |
|
|
if (GET_CODE (inner) == SUBREG
|
1306 |
|
|
&& (GET_MODE_SIZE (GET_MODE (inner))
|
1307 |
|
|
== GET_MODE_SIZE (GET_MODE (SUBREG_REG (inner))))
|
1308 |
|
|
&& subreg_lowpart_p (inner))
|
1309 |
|
|
inner = SUBREG_REG (inner);
|
1310 |
|
|
if (CONST_INT_P (inner))
|
1311 |
|
|
{
|
1312 |
|
|
x = GEN_INT (trunc_int_for_mode (INTVAL (inner), GET_MODE (x)));
|
1313 |
|
|
goto default_output;
|
1314 |
|
|
}
|
1315 |
|
|
inner_mode = GET_MODE (inner);
|
1316 |
|
|
if (GET_CODE (inner) == SUBREG
|
1317 |
|
|
&& (GET_MODE_SIZE (GET_MODE (inner))
|
1318 |
|
|
< GET_MODE_SIZE (GET_MODE (SUBREG_REG (inner))))
|
1319 |
|
|
&& REG_P (SUBREG_REG (inner)))
|
1320 |
|
|
{
|
1321 |
|
|
offset = subreg_regno_offset (REGNO (SUBREG_REG (inner)),
|
1322 |
|
|
GET_MODE (SUBREG_REG (inner)),
|
1323 |
|
|
SUBREG_BYTE (inner),
|
1324 |
|
|
GET_MODE (inner));
|
1325 |
|
|
inner = SUBREG_REG (inner);
|
1326 |
|
|
}
|
1327 |
|
|
if (!REG_P (inner) || GET_MODE_SIZE (inner_mode) > 8)
|
1328 |
|
|
abort ();
|
1329 |
|
|
/* Floating point register pairs are always big endian;
|
1330 |
|
|
general purpose registers are 64 bit wide. */
|
1331 |
|
|
regno = REGNO (inner);
|
1332 |
|
|
regno = (HARD_REGNO_NREGS (regno, inner_mode)
|
1333 |
|
|
- HARD_REGNO_NREGS (regno, mode))
|
1334 |
|
|
+ offset;
|
1335 |
|
|
x = inner;
|
1336 |
|
|
goto reg;
|
1337 |
|
|
}
|
1338 |
|
|
case SIGN_EXTEND:
|
1339 |
|
|
x = XEXP (x, 0);
|
1340 |
|
|
goto reg;
|
1341 |
|
|
/* FIXME: We need this on SHmedia32 because reload generates
|
1342 |
|
|
some sign-extended HI or QI loads into DImode registers
|
1343 |
|
|
but, because Pmode is SImode, the address ends up with a
|
1344 |
|
|
subreg:SI of the DImode register. Maybe reload should be
|
1345 |
|
|
fixed so as to apply alter_subreg to such loads? */
|
1346 |
|
|
case IF_THEN_ELSE:
|
1347 |
|
|
gcc_assert (trapping_target_operand (x, VOIDmode));
|
1348 |
|
|
x = XEXP (XEXP (x, 2), 0);
|
1349 |
|
|
goto default_output;
|
1350 |
|
|
case SUBREG:
|
1351 |
|
|
gcc_assert (SUBREG_BYTE (x) == 0
|
1352 |
|
|
&& REG_P (SUBREG_REG (x)));
|
1353 |
|
|
|
1354 |
|
|
x = SUBREG_REG (x);
|
1355 |
|
|
/* Fall through. */
|
1356 |
|
|
|
1357 |
|
|
reg:
|
1358 |
|
|
case REG:
|
1359 |
|
|
regno += REGNO (x);
|
1360 |
|
|
if (FP_REGISTER_P (regno)
|
1361 |
|
|
&& mode == V16SFmode)
|
1362 |
|
|
fprintf ((stream), "mtrx%s", reg_names[regno] + 2);
|
1363 |
|
|
else if (FP_REGISTER_P (REGNO (x))
|
1364 |
|
|
&& mode == V4SFmode)
|
1365 |
|
|
fprintf ((stream), "fv%s", reg_names[regno] + 2);
|
1366 |
|
|
else if (REG_P (x)
|
1367 |
|
|
&& mode == V2SFmode)
|
1368 |
|
|
fprintf ((stream), "fp%s", reg_names[regno] + 2);
|
1369 |
|
|
else if (FP_REGISTER_P (REGNO (x))
|
1370 |
|
|
&& GET_MODE_SIZE (mode) > 4)
|
1371 |
|
|
fprintf ((stream), "d%s", reg_names[regno] + 1);
|
1372 |
|
|
else
|
1373 |
|
|
fputs (reg_names[regno], (stream));
|
1374 |
|
|
break;
|
1375 |
|
|
|
1376 |
|
|
case MEM:
|
1377 |
|
|
output_address (XEXP (x, 0));
|
1378 |
|
|
break;
|
1379 |
|
|
|
1380 |
|
|
default:
|
1381 |
|
|
if (TARGET_SH1)
|
1382 |
|
|
fputc ('#', stream);
|
1383 |
|
|
output_addr_const (stream, x);
|
1384 |
|
|
break;
|
1385 |
|
|
}
|
1386 |
|
|
break;
|
1387 |
|
|
}
|
1388 |
|
|
}
|
1389 |
|
|
|
1390 |
|
|
|
1391 |
|
|
/* Encode symbol attributes of a SYMBOL_REF into its
|
1392 |
|
|
SYMBOL_REF_FLAGS. */
|
1393 |
|
|
static void
|
1394 |
|
|
sh_encode_section_info (tree decl, rtx rtl, int first)
|
1395 |
|
|
{
|
1396 |
|
|
default_encode_section_info (decl, rtl, first);
|
1397 |
|
|
|
1398 |
|
|
if (TREE_CODE (decl) == FUNCTION_DECL
|
1399 |
|
|
&& sh2a_function_vector_p (decl) && TARGET_SH2A)
|
1400 |
|
|
SYMBOL_REF_FLAGS (XEXP (rtl, 0)) |= SYMBOL_FLAG_FUNCVEC_FUNCTION;
|
1401 |
|
|
}
|
1402 |
|
|
|
1403 |
|
|
/* Like force_operand, but guarantees that VALUE ends up in TARGET. */
|
1404 |
|
|
static void
|
1405 |
|
|
force_into (rtx value, rtx target)
|
1406 |
|
|
{
|
1407 |
|
|
value = force_operand (value, target);
|
1408 |
|
|
if (! rtx_equal_p (value, target))
|
1409 |
|
|
emit_insn (gen_move_insn (target, value));
|
1410 |
|
|
}
|
1411 |
|
|
|
1412 |
|
|
/* Emit code to perform a block move. Choose the best method.
|
1413 |
|
|
|
1414 |
|
|
OPERANDS[0] is the destination.
|
1415 |
|
|
OPERANDS[1] is the source.
|
1416 |
|
|
OPERANDS[2] is the size.
|
1417 |
|
|
OPERANDS[3] is the alignment safe to use. */
|
1418 |
|
|
|
1419 |
|
|
int
|
1420 |
|
|
expand_block_move (rtx *operands)
|
1421 |
|
|
{
|
1422 |
|
|
int align = INTVAL (operands[3]);
|
1423 |
|
|
int constp = (CONST_INT_P (operands[2]));
|
1424 |
|
|
int bytes = (constp ? INTVAL (operands[2]) : 0);
|
1425 |
|
|
|
1426 |
|
|
if (! constp)
|
1427 |
|
|
return 0;
|
1428 |
|
|
|
1429 |
|
|
/* If we could use mov.l to move words and dest is word-aligned, we
|
1430 |
|
|
can use movua.l for loads and still generate a relatively short
|
1431 |
|
|
and efficient sequence. */
|
1432 |
|
|
if (TARGET_SH4A_ARCH && align < 4
|
1433 |
|
|
&& MEM_ALIGN (operands[0]) >= 32
|
1434 |
|
|
&& can_move_by_pieces (bytes, 32))
|
1435 |
|
|
{
|
1436 |
|
|
rtx dest = copy_rtx (operands[0]);
|
1437 |
|
|
rtx src = copy_rtx (operands[1]);
|
1438 |
|
|
/* We could use different pseudos for each copied word, but
|
1439 |
|
|
since movua can only load into r0, it's kind of
|
1440 |
|
|
pointless. */
|
1441 |
|
|
rtx temp = gen_reg_rtx (SImode);
|
1442 |
|
|
rtx src_addr = copy_addr_to_reg (XEXP (src, 0));
|
1443 |
|
|
int copied = 0;
|
1444 |
|
|
|
1445 |
|
|
while (copied + 4 <= bytes)
|
1446 |
|
|
{
|
1447 |
|
|
rtx to = adjust_address (dest, SImode, copied);
|
1448 |
|
|
rtx from = adjust_automodify_address (src, BLKmode,
|
1449 |
|
|
src_addr, copied);
|
1450 |
|
|
|
1451 |
|
|
set_mem_size (from, GEN_INT (4));
|
1452 |
|
|
emit_insn (gen_movua (temp, from));
|
1453 |
|
|
emit_move_insn (src_addr, plus_constant (src_addr, 4));
|
1454 |
|
|
emit_move_insn (to, temp);
|
1455 |
|
|
copied += 4;
|
1456 |
|
|
}
|
1457 |
|
|
|
1458 |
|
|
if (copied < bytes)
|
1459 |
|
|
move_by_pieces (adjust_address (dest, BLKmode, copied),
|
1460 |
|
|
adjust_automodify_address (src, BLKmode,
|
1461 |
|
|
src_addr, copied),
|
1462 |
|
|
bytes - copied, align, 0);
|
1463 |
|
|
|
1464 |
|
|
return 1;
|
1465 |
|
|
}
|
1466 |
|
|
|
1467 |
|
|
/* If it isn't a constant number of bytes, or if it doesn't have 4 byte
|
1468 |
|
|
alignment, or if it isn't a multiple of 4 bytes, then fail. */
|
1469 |
|
|
if (align < 4 || (bytes % 4 != 0))
|
1470 |
|
|
return 0;
|
1471 |
|
|
|
1472 |
|
|
if (TARGET_HARD_SH4)
|
1473 |
|
|
{
|
1474 |
|
|
if (bytes < 12)
|
1475 |
|
|
return 0;
|
1476 |
|
|
else if (bytes == 12)
|
1477 |
|
|
{
|
1478 |
|
|
rtx func_addr_rtx = gen_reg_rtx (Pmode);
|
1479 |
|
|
rtx r4 = gen_rtx_REG (SImode, 4);
|
1480 |
|
|
rtx r5 = gen_rtx_REG (SImode, 5);
|
1481 |
|
|
|
1482 |
|
|
function_symbol (func_addr_rtx, "__movmemSI12_i4", SFUNC_STATIC);
|
1483 |
|
|
force_into (XEXP (operands[0], 0), r4);
|
1484 |
|
|
force_into (XEXP (operands[1], 0), r5);
|
1485 |
|
|
emit_insn (gen_block_move_real_i4 (func_addr_rtx));
|
1486 |
|
|
return 1;
|
1487 |
|
|
}
|
1488 |
|
|
else if (! TARGET_SMALLCODE)
|
1489 |
|
|
{
|
1490 |
|
|
const char *entry_name;
|
1491 |
|
|
rtx func_addr_rtx = gen_reg_rtx (Pmode);
|
1492 |
|
|
int dwords;
|
1493 |
|
|
rtx r4 = gen_rtx_REG (SImode, 4);
|
1494 |
|
|
rtx r5 = gen_rtx_REG (SImode, 5);
|
1495 |
|
|
rtx r6 = gen_rtx_REG (SImode, 6);
|
1496 |
|
|
|
1497 |
|
|
entry_name = (bytes & 4 ? "__movmem_i4_odd" : "__movmem_i4_even");
|
1498 |
|
|
function_symbol (func_addr_rtx, entry_name, SFUNC_STATIC);
|
1499 |
|
|
force_into (XEXP (operands[0], 0), r4);
|
1500 |
|
|
force_into (XEXP (operands[1], 0), r5);
|
1501 |
|
|
|
1502 |
|
|
dwords = bytes >> 3;
|
1503 |
|
|
emit_insn (gen_move_insn (r6, GEN_INT (dwords - 1)));
|
1504 |
|
|
emit_insn (gen_block_lump_real_i4 (func_addr_rtx));
|
1505 |
|
|
return 1;
|
1506 |
|
|
}
|
1507 |
|
|
else
|
1508 |
|
|
return 0;
|
1509 |
|
|
}
|
1510 |
|
|
if (bytes < 64)
|
1511 |
|
|
{
|
1512 |
|
|
char entry[30];
|
1513 |
|
|
rtx func_addr_rtx = gen_reg_rtx (Pmode);
|
1514 |
|
|
rtx r4 = gen_rtx_REG (SImode, 4);
|
1515 |
|
|
rtx r5 = gen_rtx_REG (SImode, 5);
|
1516 |
|
|
|
1517 |
|
|
sprintf (entry, "__movmemSI%d", bytes);
|
1518 |
|
|
function_symbol (func_addr_rtx, entry, SFUNC_STATIC);
|
1519 |
|
|
force_into (XEXP (operands[0], 0), r4);
|
1520 |
|
|
force_into (XEXP (operands[1], 0), r5);
|
1521 |
|
|
emit_insn (gen_block_move_real (func_addr_rtx));
|
1522 |
|
|
return 1;
|
1523 |
|
|
}
|
1524 |
|
|
|
1525 |
|
|
/* This is the same number of bytes as a memcpy call, but to a different
|
1526 |
|
|
less common function name, so this will occasionally use more space. */
|
1527 |
|
|
if (! TARGET_SMALLCODE)
|
1528 |
|
|
{
|
1529 |
|
|
rtx func_addr_rtx = gen_reg_rtx (Pmode);
|
1530 |
|
|
int final_switch, while_loop;
|
1531 |
|
|
rtx r4 = gen_rtx_REG (SImode, 4);
|
1532 |
|
|
rtx r5 = gen_rtx_REG (SImode, 5);
|
1533 |
|
|
rtx r6 = gen_rtx_REG (SImode, 6);
|
1534 |
|
|
|
1535 |
|
|
function_symbol (func_addr_rtx, "__movmem", SFUNC_STATIC);
|
1536 |
|
|
force_into (XEXP (operands[0], 0), r4);
|
1537 |
|
|
force_into (XEXP (operands[1], 0), r5);
|
1538 |
|
|
|
1539 |
|
|
/* r6 controls the size of the move. 16 is decremented from it
|
1540 |
|
|
for each 64 bytes moved. Then the negative bit left over is used
|
1541 |
|
|
as an index into a list of move instructions. e.g., a 72 byte move
|
1542 |
|
|
would be set up with size(r6) = 14, for one iteration through the
|
1543 |
|
|
big while loop, and a switch of -2 for the last part. */
|
1544 |
|
|
|
1545 |
|
|
final_switch = 16 - ((bytes / 4) % 16);
|
1546 |
|
|
while_loop = ((bytes / 4) / 16 - 1) * 16;
|
1547 |
|
|
emit_insn (gen_move_insn (r6, GEN_INT (while_loop + final_switch)));
|
1548 |
|
|
emit_insn (gen_block_lump_real (func_addr_rtx));
|
1549 |
|
|
return 1;
|
1550 |
|
|
}
|
1551 |
|
|
|
1552 |
|
|
return 0;
|
1553 |
|
|
}
|
1554 |
|
|
|
1555 |
|
|
/* Prepare operands for a move define_expand; specifically, one of the
|
1556 |
|
|
operands must be in a register. */
|
1557 |
|
|
|
1558 |
|
|
int
|
1559 |
|
|
prepare_move_operands (rtx operands[], enum machine_mode mode)
|
1560 |
|
|
{
|
1561 |
|
|
if ((mode == SImode || mode == DImode)
|
1562 |
|
|
&& flag_pic
|
1563 |
|
|
&& ! ((mode == Pmode || mode == ptr_mode)
|
1564 |
|
|
&& tls_symbolic_operand (operands[1], Pmode) != TLS_MODEL_NONE))
|
1565 |
|
|
{
|
1566 |
|
|
rtx temp;
|
1567 |
|
|
if (SYMBOLIC_CONST_P (operands[1]))
|
1568 |
|
|
{
|
1569 |
|
|
if (MEM_P (operands[0]))
|
1570 |
|
|
operands[1] = force_reg (Pmode, operands[1]);
|
1571 |
|
|
else if (TARGET_SHMEDIA
|
1572 |
|
|
&& GET_CODE (operands[1]) == LABEL_REF
|
1573 |
|
|
&& target_reg_operand (operands[0], mode))
|
1574 |
|
|
/* It's ok. */;
|
1575 |
|
|
else
|
1576 |
|
|
{
|
1577 |
|
|
temp = (!can_create_pseudo_p ()
|
1578 |
|
|
? operands[0]
|
1579 |
|
|
: gen_reg_rtx (Pmode));
|
1580 |
|
|
operands[1] = legitimize_pic_address (operands[1], mode, temp);
|
1581 |
|
|
}
|
1582 |
|
|
}
|
1583 |
|
|
else if (GET_CODE (operands[1]) == CONST
|
1584 |
|
|
&& GET_CODE (XEXP (operands[1], 0)) == PLUS
|
1585 |
|
|
&& SYMBOLIC_CONST_P (XEXP (XEXP (operands[1], 0), 0)))
|
1586 |
|
|
{
|
1587 |
|
|
temp = !can_create_pseudo_p () ? operands[0] : gen_reg_rtx (Pmode);
|
1588 |
|
|
temp = legitimize_pic_address (XEXP (XEXP (operands[1], 0), 0),
|
1589 |
|
|
mode, temp);
|
1590 |
|
|
operands[1] = expand_binop (mode, add_optab, temp,
|
1591 |
|
|
XEXP (XEXP (operands[1], 0), 1),
|
1592 |
|
|
(!can_create_pseudo_p ()
|
1593 |
|
|
? temp
|
1594 |
|
|
: gen_reg_rtx (Pmode)),
|
1595 |
|
|
0, OPTAB_LIB_WIDEN);
|
1596 |
|
|
}
|
1597 |
|
|
}
|
1598 |
|
|
|
1599 |
|
|
if (! reload_in_progress && ! reload_completed)
|
1600 |
|
|
{
|
1601 |
|
|
/* Copy the source to a register if both operands aren't registers. */
|
1602 |
|
|
if (! register_operand (operands[0], mode)
|
1603 |
|
|
&& ! sh_register_operand (operands[1], mode))
|
1604 |
|
|
operands[1] = copy_to_mode_reg (mode, operands[1]);
|
1605 |
|
|
|
1606 |
|
|
if (MEM_P (operands[0]) && ! memory_operand (operands[0], mode))
|
1607 |
|
|
{
|
1608 |
|
|
/* This is like change_address_1 (operands[0], mode, 0, 1) ,
|
1609 |
|
|
except that we can't use that function because it is static. */
|
1610 |
|
|
rtx new_rtx = change_address (operands[0], mode, 0);
|
1611 |
|
|
MEM_COPY_ATTRIBUTES (new_rtx, operands[0]);
|
1612 |
|
|
operands[0] = new_rtx;
|
1613 |
|
|
}
|
1614 |
|
|
|
1615 |
|
|
/* This case can happen while generating code to move the result
|
1616 |
|
|
of a library call to the target. Reject `st r0,@(rX,rY)' because
|
1617 |
|
|
reload will fail to find a spill register for rX, since r0 is already
|
1618 |
|
|
being used for the source. */
|
1619 |
|
|
else if (TARGET_SH1
|
1620 |
|
|
&& refers_to_regno_p (R0_REG, R0_REG + 1, operands[1], (rtx *)0)
|
1621 |
|
|
&& MEM_P (operands[0])
|
1622 |
|
|
&& GET_CODE (XEXP (operands[0], 0)) == PLUS
|
1623 |
|
|
&& REG_P (XEXP (XEXP (operands[0], 0), 1)))
|
1624 |
|
|
operands[1] = copy_to_mode_reg (mode, operands[1]);
|
1625 |
|
|
}
|
1626 |
|
|
|
1627 |
|
|
if (mode == Pmode || mode == ptr_mode)
|
1628 |
|
|
{
|
1629 |
|
|
rtx op0, op1, opc;
|
1630 |
|
|
enum tls_model tls_kind;
|
1631 |
|
|
|
1632 |
|
|
op0 = operands[0];
|
1633 |
|
|
op1 = operands[1];
|
1634 |
|
|
if (GET_CODE (op1) == CONST
|
1635 |
|
|
&& GET_CODE (XEXP (op1, 0)) == PLUS
|
1636 |
|
|
&& (tls_symbolic_operand (XEXP (XEXP (op1, 0), 0), Pmode)
|
1637 |
|
|
!= TLS_MODEL_NONE))
|
1638 |
|
|
{
|
1639 |
|
|
opc = XEXP (XEXP (op1, 0), 1);
|
1640 |
|
|
op1 = XEXP (XEXP (op1, 0), 0);
|
1641 |
|
|
}
|
1642 |
|
|
else
|
1643 |
|
|
opc = NULL_RTX;
|
1644 |
|
|
|
1645 |
|
|
if ((tls_kind = tls_symbolic_operand (op1, Pmode)) != TLS_MODEL_NONE)
|
1646 |
|
|
{
|
1647 |
|
|
rtx tga_op1, tga_ret, tmp, tmp2;
|
1648 |
|
|
|
1649 |
|
|
switch (tls_kind)
|
1650 |
|
|
{
|
1651 |
|
|
case TLS_MODEL_GLOBAL_DYNAMIC:
|
1652 |
|
|
tga_ret = gen_rtx_REG (Pmode, R0_REG);
|
1653 |
|
|
emit_call_insn (gen_tls_global_dynamic (tga_ret, op1));
|
1654 |
|
|
op1 = tga_ret;
|
1655 |
|
|
break;
|
1656 |
|
|
|
1657 |
|
|
case TLS_MODEL_LOCAL_DYNAMIC:
|
1658 |
|
|
tga_ret = gen_rtx_REG (Pmode, R0_REG);
|
1659 |
|
|
emit_call_insn (gen_tls_local_dynamic (tga_ret, op1));
|
1660 |
|
|
|
1661 |
|
|
tmp = gen_reg_rtx (Pmode);
|
1662 |
|
|
emit_move_insn (tmp, tga_ret);
|
1663 |
|
|
|
1664 |
|
|
if (register_operand (op0, Pmode))
|
1665 |
|
|
tmp2 = op0;
|
1666 |
|
|
else
|
1667 |
|
|
tmp2 = gen_reg_rtx (Pmode);
|
1668 |
|
|
|
1669 |
|
|
emit_insn (gen_symDTPOFF2reg (tmp2, op1, tmp));
|
1670 |
|
|
op1 = tmp2;
|
1671 |
|
|
break;
|
1672 |
|
|
|
1673 |
|
|
case TLS_MODEL_INITIAL_EXEC:
|
1674 |
|
|
if (! flag_pic)
|
1675 |
|
|
{
|
1676 |
|
|
/* Don't schedule insns for getting GOT address when
|
1677 |
|
|
the first scheduling is enabled, to avoid spill
|
1678 |
|
|
failures for R0. */
|
1679 |
|
|
if (flag_schedule_insns)
|
1680 |
|
|
emit_insn (gen_blockage ());
|
1681 |
|
|
emit_insn (gen_GOTaddr2picreg ());
|
1682 |
|
|
emit_use (gen_rtx_REG (SImode, PIC_REG));
|
1683 |
|
|
if (flag_schedule_insns)
|
1684 |
|
|
emit_insn (gen_blockage ());
|
1685 |
|
|
}
|
1686 |
|
|
tga_op1 = !can_create_pseudo_p () ? op0 : gen_reg_rtx (Pmode);
|
1687 |
|
|
tmp = gen_sym2GOTTPOFF (op1);
|
1688 |
|
|
emit_insn (gen_tls_initial_exec (tga_op1, tmp));
|
1689 |
|
|
op1 = tga_op1;
|
1690 |
|
|
break;
|
1691 |
|
|
|
1692 |
|
|
case TLS_MODEL_LOCAL_EXEC:
|
1693 |
|
|
tmp2 = gen_reg_rtx (Pmode);
|
1694 |
|
|
emit_insn (gen_load_gbr (tmp2));
|
1695 |
|
|
tmp = gen_reg_rtx (Pmode);
|
1696 |
|
|
emit_insn (gen_symTPOFF2reg (tmp, op1));
|
1697 |
|
|
|
1698 |
|
|
if (register_operand (op0, Pmode))
|
1699 |
|
|
op1 = op0;
|
1700 |
|
|
else
|
1701 |
|
|
op1 = gen_reg_rtx (Pmode);
|
1702 |
|
|
|
1703 |
|
|
emit_insn (gen_addsi3 (op1, tmp, tmp2));
|
1704 |
|
|
break;
|
1705 |
|
|
|
1706 |
|
|
default:
|
1707 |
|
|
gcc_unreachable ();
|
1708 |
|
|
}
|
1709 |
|
|
if (opc)
|
1710 |
|
|
emit_insn (gen_addsi3 (op1, op1, force_reg (SImode, opc)));
|
1711 |
|
|
operands[1] = op1;
|
1712 |
|
|
}
|
1713 |
|
|
}
|
1714 |
|
|
|
1715 |
|
|
return 0;
|
1716 |
|
|
}
|
1717 |
|
|
|
1718 |
|
|
enum rtx_code
|
1719 |
|
|
prepare_cbranch_operands (rtx *operands, enum machine_mode mode,
|
1720 |
|
|
enum rtx_code comparison)
|
1721 |
|
|
{
|
1722 |
|
|
rtx op1;
|
1723 |
|
|
rtx scratch = NULL_RTX;
|
1724 |
|
|
|
1725 |
|
|
if (comparison == LAST_AND_UNUSED_RTX_CODE)
|
1726 |
|
|
comparison = GET_CODE (operands[0]);
|
1727 |
|
|
else
|
1728 |
|
|
scratch = operands[4];
|
1729 |
|
|
if (CONST_INT_P (operands[1])
|
1730 |
|
|
&& !CONST_INT_P (operands[2]))
|
1731 |
|
|
{
|
1732 |
|
|
rtx tmp = operands[1];
|
1733 |
|
|
|
1734 |
|
|
operands[1] = operands[2];
|
1735 |
|
|
operands[2] = tmp;
|
1736 |
|
|
comparison = swap_condition (comparison);
|
1737 |
|
|
}
|
1738 |
|
|
if (CONST_INT_P (operands[2]))
|
1739 |
|
|
{
|
1740 |
|
|
HOST_WIDE_INT val = INTVAL (operands[2]);
|
1741 |
|
|
if ((val == -1 || val == -0x81)
|
1742 |
|
|
&& (comparison == GT || comparison == LE))
|
1743 |
|
|
{
|
1744 |
|
|
comparison = (comparison == GT) ? GE : LT;
|
1745 |
|
|
operands[2] = gen_int_mode (val + 1, mode);
|
1746 |
|
|
}
|
1747 |
|
|
else if ((val == 1 || val == 0x80)
|
1748 |
|
|
&& (comparison == GE || comparison == LT))
|
1749 |
|
|
{
|
1750 |
|
|
comparison = (comparison == GE) ? GT : LE;
|
1751 |
|
|
operands[2] = gen_int_mode (val - 1, mode);
|
1752 |
|
|
}
|
1753 |
|
|
else if (val == 1 && (comparison == GEU || comparison == LTU))
|
1754 |
|
|
{
|
1755 |
|
|
comparison = (comparison == GEU) ? NE : EQ;
|
1756 |
|
|
operands[2] = CONST0_RTX (mode);
|
1757 |
|
|
}
|
1758 |
|
|
else if (val == 0x80 && (comparison == GEU || comparison == LTU))
|
1759 |
|
|
{
|
1760 |
|
|
comparison = (comparison == GEU) ? GTU : LEU;
|
1761 |
|
|
operands[2] = gen_int_mode (val - 1, mode);
|
1762 |
|
|
}
|
1763 |
|
|
else if (val == 0 && (comparison == GTU || comparison == LEU))
|
1764 |
|
|
comparison = (comparison == GTU) ? NE : EQ;
|
1765 |
|
|
else if (mode == SImode
|
1766 |
|
|
&& ((val == 0x7fffffff
|
1767 |
|
|
&& (comparison == GTU || comparison == LEU))
|
1768 |
|
|
|| ((unsigned HOST_WIDE_INT) val
|
1769 |
|
|
== (unsigned HOST_WIDE_INT) 0x7fffffff + 1
|
1770 |
|
|
&& (comparison == GEU || comparison == LTU))))
|
1771 |
|
|
{
|
1772 |
|
|
comparison = (comparison == GTU || comparison == GEU) ? LT : GE;
|
1773 |
|
|
operands[2] = CONST0_RTX (mode);
|
1774 |
|
|
}
|
1775 |
|
|
}
|
1776 |
|
|
op1 = operands[1];
|
1777 |
|
|
if (can_create_pseudo_p ())
|
1778 |
|
|
operands[1] = force_reg (mode, op1);
|
1779 |
|
|
/* When we are handling DImode comparisons, we want to keep constants so
|
1780 |
|
|
that we can optimize the component comparisons; however, memory loads
|
1781 |
|
|
are better issued as a whole so that they can be scheduled well.
|
1782 |
|
|
SImode equality comparisons allow I08 constants, but only when they
|
1783 |
|
|
compare r0. Hence, if operands[1] has to be loaded from somewhere else
|
1784 |
|
|
into a register, that register might as well be r0, and we allow the
|
1785 |
|
|
constant. If it is already in a register, this is likely to be
|
1786 |
|
|
allocated to a different hard register, thus we load the constant into
|
1787 |
|
|
a register unless it is zero. */
|
1788 |
|
|
if (!REG_P (operands[2])
|
1789 |
|
|
&& (!CONST_INT_P (operands[2])
|
1790 |
|
|
|| (mode == SImode && operands[2] != CONST0_RTX (SImode)
|
1791 |
|
|
&& ((comparison != EQ && comparison != NE)
|
1792 |
|
|
|| (REG_P (op1) && REGNO (op1) != R0_REG)
|
1793 |
|
|
|| !satisfies_constraint_I08 (operands[2])))))
|
1794 |
|
|
{
|
1795 |
|
|
if (scratch && GET_MODE (scratch) == mode)
|
1796 |
|
|
{
|
1797 |
|
|
emit_move_insn (scratch, operands[2]);
|
1798 |
|
|
operands[2] = scratch;
|
1799 |
|
|
}
|
1800 |
|
|
else if (can_create_pseudo_p ())
|
1801 |
|
|
operands[2] = force_reg (mode, operands[2]);
|
1802 |
|
|
}
|
1803 |
|
|
return comparison;
|
1804 |
|
|
}
|
1805 |
|
|
|
1806 |
|
|
void
|
1807 |
|
|
expand_cbranchsi4 (rtx *operands, enum rtx_code comparison, int probability)
|
1808 |
|
|
{
|
1809 |
|
|
rtx (*branch_expander) (rtx) = gen_branch_true;
|
1810 |
|
|
rtx jump;
|
1811 |
|
|
|
1812 |
|
|
comparison = prepare_cbranch_operands (operands, SImode, comparison);
|
1813 |
|
|
switch (comparison)
|
1814 |
|
|
{
|
1815 |
|
|
case NE: case LT: case LE: case LTU: case LEU:
|
1816 |
|
|
comparison = reverse_condition (comparison);
|
1817 |
|
|
branch_expander = gen_branch_false;
|
1818 |
|
|
default: ;
|
1819 |
|
|
}
|
1820 |
|
|
emit_insn (gen_rtx_SET (VOIDmode, gen_rtx_REG (SImode, T_REG),
|
1821 |
|
|
gen_rtx_fmt_ee (comparison, SImode,
|
1822 |
|
|
operands[1], operands[2])));
|
1823 |
|
|
jump = emit_jump_insn (branch_expander (operands[3]));
|
1824 |
|
|
if (probability >= 0)
|
1825 |
|
|
add_reg_note (jump, REG_BR_PROB, GEN_INT (probability));
|
1826 |
|
|
|
1827 |
|
|
}
|
1828 |
|
|
|
1829 |
|
|
/* ??? How should we distribute probabilities when more than one branch
|
1830 |
|
|
is generated. So far we only have soem ad-hoc observations:
|
1831 |
|
|
- If the operands are random, they are likely to differ in both parts.
|
1832 |
|
|
- If comparing items in a hash chain, the operands are random or equal;
|
1833 |
|
|
operation should be EQ or NE.
|
1834 |
|
|
- If items are searched in an ordered tree from the root, we can expect
|
1835 |
|
|
the highpart to be unequal about half of the time; operation should be
|
1836 |
|
|
an inequality comparison, operands non-constant, and overall probability
|
1837 |
|
|
about 50%. Likewise for quicksort.
|
1838 |
|
|
- Range checks will be often made against constants. Even if we assume for
|
1839 |
|
|
simplicity an even distribution of the non-constant operand over a
|
1840 |
|
|
sub-range here, the same probability could be generated with differently
|
1841 |
|
|
wide sub-ranges - as long as the ratio of the part of the subrange that
|
1842 |
|
|
is before the threshold to the part that comes after the threshold stays
|
1843 |
|
|
the same. Thus, we can't really tell anything here;
|
1844 |
|
|
assuming random distribution is at least simple.
|
1845 |
|
|
*/
|
1846 |
|
|
|
1847 |
|
|
bool
|
1848 |
|
|
expand_cbranchdi4 (rtx *operands, enum rtx_code comparison)
|
1849 |
|
|
{
|
1850 |
|
|
enum rtx_code msw_taken, msw_skip, lsw_taken;
|
1851 |
|
|
rtx skip_label = NULL_RTX;
|
1852 |
|
|
rtx op1h, op1l, op2h, op2l;
|
1853 |
|
|
int num_branches;
|
1854 |
|
|
int prob, rev_prob;
|
1855 |
|
|
int msw_taken_prob = -1, msw_skip_prob = -1, lsw_taken_prob = -1;
|
1856 |
|
|
rtx scratch = operands[4];
|
1857 |
|
|
|
1858 |
|
|
comparison = prepare_cbranch_operands (operands, DImode, comparison);
|
1859 |
|
|
op1h = gen_highpart_mode (SImode, DImode, operands[1]);
|
1860 |
|
|
op2h = gen_highpart_mode (SImode, DImode, operands[2]);
|
1861 |
|
|
op1l = gen_lowpart (SImode, operands[1]);
|
1862 |
|
|
op2l = gen_lowpart (SImode, operands[2]);
|
1863 |
|
|
msw_taken = msw_skip = lsw_taken = LAST_AND_UNUSED_RTX_CODE;
|
1864 |
|
|
prob = split_branch_probability;
|
1865 |
|
|
rev_prob = REG_BR_PROB_BASE - prob;
|
1866 |
|
|
switch (comparison)
|
1867 |
|
|
{
|
1868 |
|
|
/* ??? Should we use the cmpeqdi_t pattern for equality comparisons?
|
1869 |
|
|
That costs 1 cycle more when the first branch can be predicted taken,
|
1870 |
|
|
but saves us mispredicts because only one branch needs prediction.
|
1871 |
|
|
It also enables generating the cmpeqdi_t-1 pattern. */
|
1872 |
|
|
case EQ:
|
1873 |
|
|
if (TARGET_CMPEQDI_T)
|
1874 |
|
|
{
|
1875 |
|
|
emit_insn (gen_cmpeqdi_t (operands[1], operands[2]));
|
1876 |
|
|
emit_jump_insn (gen_branch_true (operands[3]));
|
1877 |
|
|
return true;
|
1878 |
|
|
}
|
1879 |
|
|
msw_skip = NE;
|
1880 |
|
|
lsw_taken = EQ;
|
1881 |
|
|
if (prob >= 0)
|
1882 |
|
|
{
|
1883 |
|
|
/* If we had more precision, we'd use rev_prob - (rev_prob >> 32) .
|
1884 |
|
|
*/
|
1885 |
|
|
msw_skip_prob = rev_prob;
|
1886 |
|
|
if (REG_BR_PROB_BASE <= 65535)
|
1887 |
|
|
lsw_taken_prob = prob ? REG_BR_PROB_BASE : 0;
|
1888 |
|
|
else
|
1889 |
|
|
{
|
1890 |
|
|
gcc_assert (HOST_BITS_PER_WIDEST_INT >= 64);
|
1891 |
|
|
lsw_taken_prob
|
1892 |
|
|
= (prob
|
1893 |
|
|
? (REG_BR_PROB_BASE
|
1894 |
|
|
- ((HOST_WIDEST_INT) REG_BR_PROB_BASE * rev_prob
|
1895 |
|
|
/ ((HOST_WIDEST_INT) prob << 32)))
|
1896 |
|
|
: 0);
|
1897 |
|
|
}
|
1898 |
|
|
}
|
1899 |
|
|
break;
|
1900 |
|
|
case NE:
|
1901 |
|
|
if (TARGET_CMPEQDI_T)
|
1902 |
|
|
{
|
1903 |
|
|
emit_insn (gen_cmpeqdi_t (operands[1], operands[2]));
|
1904 |
|
|
emit_jump_insn (gen_branch_false (operands[3]));
|
1905 |
|
|
return true;
|
1906 |
|
|
}
|
1907 |
|
|
msw_taken = NE;
|
1908 |
|
|
msw_taken_prob = prob;
|
1909 |
|
|
lsw_taken = NE;
|
1910 |
|
|
lsw_taken_prob = 0;
|
1911 |
|
|
break;
|
1912 |
|
|
case GTU: case GT:
|
1913 |
|
|
msw_taken = comparison;
|
1914 |
|
|
if (CONST_INT_P (op2l) && INTVAL (op2l) == -1)
|
1915 |
|
|
break;
|
1916 |
|
|
if (comparison != GTU || op2h != CONST0_RTX (SImode))
|
1917 |
|
|
msw_skip = swap_condition (msw_taken);
|
1918 |
|
|
lsw_taken = GTU;
|
1919 |
|
|
break;
|
1920 |
|
|
case GEU: case GE:
|
1921 |
|
|
if (op2l == CONST0_RTX (SImode))
|
1922 |
|
|
msw_taken = comparison;
|
1923 |
|
|
else
|
1924 |
|
|
{
|
1925 |
|
|
msw_taken = comparison == GE ? GT : GTU;
|
1926 |
|
|
msw_skip = swap_condition (msw_taken);
|
1927 |
|
|
lsw_taken = GEU;
|
1928 |
|
|
}
|
1929 |
|
|
break;
|
1930 |
|
|
case LTU: case LT:
|
1931 |
|
|
msw_taken = comparison;
|
1932 |
|
|
if (op2l == CONST0_RTX (SImode))
|
1933 |
|
|
break;
|
1934 |
|
|
msw_skip = swap_condition (msw_taken);
|
1935 |
|
|
lsw_taken = LTU;
|
1936 |
|
|
break;
|
1937 |
|
|
case LEU: case LE:
|
1938 |
|
|
if (CONST_INT_P (op2l) && INTVAL (op2l) == -1)
|
1939 |
|
|
msw_taken = comparison;
|
1940 |
|
|
else
|
1941 |
|
|
{
|
1942 |
|
|
lsw_taken = LEU;
|
1943 |
|
|
if (comparison == LE)
|
1944 |
|
|
msw_taken = LT;
|
1945 |
|
|
else if (op2h != CONST0_RTX (SImode))
|
1946 |
|
|
msw_taken = LTU;
|
1947 |
|
|
else
|
1948 |
|
|
break;
|
1949 |
|
|
msw_skip = swap_condition (msw_taken);
|
1950 |
|
|
}
|
1951 |
|
|
break;
|
1952 |
|
|
default: return false;
|
1953 |
|
|
}
|
1954 |
|
|
num_branches = ((msw_taken != LAST_AND_UNUSED_RTX_CODE)
|
1955 |
|
|
+ (msw_skip != LAST_AND_UNUSED_RTX_CODE)
|
1956 |
|
|
+ (lsw_taken != LAST_AND_UNUSED_RTX_CODE));
|
1957 |
|
|
if (comparison != EQ && comparison != NE && num_branches > 1)
|
1958 |
|
|
{
|
1959 |
|
|
if (!CONSTANT_P (operands[2])
|
1960 |
|
|
&& prob >= (int) (REG_BR_PROB_BASE * 3 / 8U)
|
1961 |
|
|
&& prob <= (int) (REG_BR_PROB_BASE * 5 / 8U))
|
1962 |
|
|
{
|
1963 |
|
|
msw_taken_prob = prob / 2U;
|
1964 |
|
|
msw_skip_prob
|
1965 |
|
|
= REG_BR_PROB_BASE * rev_prob / (REG_BR_PROB_BASE + rev_prob);
|
1966 |
|
|
lsw_taken_prob = prob;
|
1967 |
|
|
}
|
1968 |
|
|
else
|
1969 |
|
|
{
|
1970 |
|
|
msw_taken_prob = prob;
|
1971 |
|
|
msw_skip_prob = REG_BR_PROB_BASE;
|
1972 |
|
|
/* ??? If we have a constant op2h, should we use that when
|
1973 |
|
|
calculating lsw_taken_prob? */
|
1974 |
|
|
lsw_taken_prob = prob;
|
1975 |
|
|
}
|
1976 |
|
|
}
|
1977 |
|
|
operands[1] = op1h;
|
1978 |
|
|
operands[2] = op2h;
|
1979 |
|
|
operands[4] = NULL_RTX;
|
1980 |
|
|
if (reload_completed
|
1981 |
|
|
&& ! arith_reg_or_0_operand (op2h, SImode)
|
1982 |
|
|
&& (true_regnum (op1h) || (comparison != EQ && comparison != NE))
|
1983 |
|
|
&& (msw_taken != LAST_AND_UNUSED_RTX_CODE
|
1984 |
|
|
|| msw_skip != LAST_AND_UNUSED_RTX_CODE))
|
1985 |
|
|
{
|
1986 |
|
|
emit_move_insn (scratch, operands[2]);
|
1987 |
|
|
operands[2] = scratch;
|
1988 |
|
|
}
|
1989 |
|
|
if (msw_taken != LAST_AND_UNUSED_RTX_CODE)
|
1990 |
|
|
expand_cbranchsi4 (operands, msw_taken, msw_taken_prob);
|
1991 |
|
|
if (msw_skip != LAST_AND_UNUSED_RTX_CODE)
|
1992 |
|
|
{
|
1993 |
|
|
rtx taken_label = operands[3];
|
1994 |
|
|
|
1995 |
|
|
/* Operands were possibly modified, but msw_skip doesn't expect this.
|
1996 |
|
|
Always use the original ones. */
|
1997 |
|
|
if (msw_taken != LAST_AND_UNUSED_RTX_CODE)
|
1998 |
|
|
{
|
1999 |
|
|
operands[1] = op1h;
|
2000 |
|
|
operands[2] = op2h;
|
2001 |
|
|
}
|
2002 |
|
|
|
2003 |
|
|
operands[3] = skip_label = gen_label_rtx ();
|
2004 |
|
|
expand_cbranchsi4 (operands, msw_skip, msw_skip_prob);
|
2005 |
|
|
operands[3] = taken_label;
|
2006 |
|
|
}
|
2007 |
|
|
operands[1] = op1l;
|
2008 |
|
|
operands[2] = op2l;
|
2009 |
|
|
if (lsw_taken != LAST_AND_UNUSED_RTX_CODE)
|
2010 |
|
|
{
|
2011 |
|
|
if (reload_completed
|
2012 |
|
|
&& ! arith_reg_or_0_operand (op2l, SImode)
|
2013 |
|
|
&& (true_regnum (op1l) || (lsw_taken != EQ && lsw_taken != NE)))
|
2014 |
|
|
{
|
2015 |
|
|
emit_move_insn (scratch, operands[2]);
|
2016 |
|
|
operands[2] = scratch;
|
2017 |
|
|
}
|
2018 |
|
|
expand_cbranchsi4 (operands, lsw_taken, lsw_taken_prob);
|
2019 |
|
|
}
|
2020 |
|
|
if (msw_skip != LAST_AND_UNUSED_RTX_CODE)
|
2021 |
|
|
emit_label (skip_label);
|
2022 |
|
|
return true;
|
2023 |
|
|
}
|
2024 |
|
|
|
2025 |
|
|
/* Emit INSN, possibly in a PARALLEL with an USE of fpscr for SH4. */
|
2026 |
|
|
|
2027 |
|
|
static void
|
2028 |
|
|
sh_emit_set_t_insn (rtx insn, enum machine_mode mode)
|
2029 |
|
|
{
|
2030 |
|
|
if ((TARGET_SH4 || TARGET_SH2A) && GET_MODE_CLASS (mode) == MODE_FLOAT)
|
2031 |
|
|
{
|
2032 |
|
|
insn = gen_rtx_PARALLEL (VOIDmode,
|
2033 |
|
|
gen_rtvec (2, insn,
|
2034 |
|
|
gen_rtx_USE (VOIDmode, get_fpscr_rtx ())));
|
2035 |
|
|
(mode == SFmode ? emit_sf_insn : emit_df_insn) (insn);
|
2036 |
|
|
}
|
2037 |
|
|
else
|
2038 |
|
|
emit_insn (insn);
|
2039 |
|
|
}
|
2040 |
|
|
|
2041 |
|
|
/* Prepare the operands for an scc instruction; make sure that the
|
2042 |
|
|
compare has been done and the result is in T_REG. */
|
2043 |
|
|
void
|
2044 |
|
|
sh_emit_scc_to_t (enum rtx_code code, rtx op0, rtx op1)
|
2045 |
|
|
{
|
2046 |
|
|
rtx t_reg = gen_rtx_REG (SImode, T_REG);
|
2047 |
|
|
enum rtx_code oldcode = code;
|
2048 |
|
|
enum machine_mode mode;
|
2049 |
|
|
|
2050 |
|
|
/* First need a compare insn. */
|
2051 |
|
|
switch (code)
|
2052 |
|
|
{
|
2053 |
|
|
case NE:
|
2054 |
|
|
/* It isn't possible to handle this case. */
|
2055 |
|
|
gcc_unreachable ();
|
2056 |
|
|
case LT:
|
2057 |
|
|
code = GT;
|
2058 |
|
|
break;
|
2059 |
|
|
case LE:
|
2060 |
|
|
code = GE;
|
2061 |
|
|
break;
|
2062 |
|
|
case LTU:
|
2063 |
|
|
code = GTU;
|
2064 |
|
|
break;
|
2065 |
|
|
case LEU:
|
2066 |
|
|
code = GEU;
|
2067 |
|
|
break;
|
2068 |
|
|
default:
|
2069 |
|
|
break;
|
2070 |
|
|
}
|
2071 |
|
|
if (code != oldcode)
|
2072 |
|
|
{
|
2073 |
|
|
rtx tmp = op0;
|
2074 |
|
|
op0 = op1;
|
2075 |
|
|
op1 = tmp;
|
2076 |
|
|
}
|
2077 |
|
|
|
2078 |
|
|
mode = GET_MODE (op0);
|
2079 |
|
|
if (mode == VOIDmode)
|
2080 |
|
|
mode = GET_MODE (op1);
|
2081 |
|
|
|
2082 |
|
|
op0 = force_reg (mode, op0);
|
2083 |
|
|
if ((code != EQ && code != NE
|
2084 |
|
|
&& (op1 != const0_rtx
|
2085 |
|
|
|| code == GTU || code == GEU || code == LTU || code == LEU))
|
2086 |
|
|
|| (mode == DImode && op1 != const0_rtx)
|
2087 |
|
|
|| (TARGET_SH2E && GET_MODE_CLASS (mode) == MODE_FLOAT))
|
2088 |
|
|
op1 = force_reg (mode, op1);
|
2089 |
|
|
|
2090 |
|
|
sh_emit_set_t_insn (gen_rtx_SET (VOIDmode, t_reg,
|
2091 |
|
|
gen_rtx_fmt_ee (code, SImode, op0, op1)),
|
2092 |
|
|
mode);
|
2093 |
|
|
}
|
2094 |
|
|
|
2095 |
|
|
rtx
|
2096 |
|
|
sh_emit_cheap_store_flag (enum machine_mode mode, enum rtx_code code,
|
2097 |
|
|
rtx op0, rtx op1)
|
2098 |
|
|
{
|
2099 |
|
|
rtx target = gen_reg_rtx (SImode);
|
2100 |
|
|
rtx tmp;
|
2101 |
|
|
|
2102 |
|
|
gcc_assert (TARGET_SHMEDIA);
|
2103 |
|
|
switch (code)
|
2104 |
|
|
{
|
2105 |
|
|
case EQ:
|
2106 |
|
|
case GT:
|
2107 |
|
|
case LT:
|
2108 |
|
|
case UNORDERED:
|
2109 |
|
|
case GTU:
|
2110 |
|
|
case LTU:
|
2111 |
|
|
tmp = gen_rtx_fmt_ee (code, SImode, op0, op1);
|
2112 |
|
|
emit_insn (gen_cstore4_media (target, tmp, op0, op1));
|
2113 |
|
|
code = NE;
|
2114 |
|
|
break;
|
2115 |
|
|
|
2116 |
|
|
case NE:
|
2117 |
|
|
case GE:
|
2118 |
|
|
case LE:
|
2119 |
|
|
case ORDERED:
|
2120 |
|
|
case GEU:
|
2121 |
|
|
case LEU:
|
2122 |
|
|
tmp = gen_rtx_fmt_ee (reverse_condition (code), mode, op0, op1);
|
2123 |
|
|
emit_insn (gen_cstore4_media (target, tmp, op0, op1));
|
2124 |
|
|
code = EQ;
|
2125 |
|
|
break;
|
2126 |
|
|
|
2127 |
|
|
case UNEQ:
|
2128 |
|
|
case UNGE:
|
2129 |
|
|
case UNGT:
|
2130 |
|
|
case UNLE:
|
2131 |
|
|
case UNLT:
|
2132 |
|
|
case LTGT:
|
2133 |
|
|
return NULL_RTX;
|
2134 |
|
|
|
2135 |
|
|
default:
|
2136 |
|
|
gcc_unreachable ();
|
2137 |
|
|
}
|
2138 |
|
|
|
2139 |
|
|
if (mode == DImode)
|
2140 |
|
|
{
|
2141 |
|
|
rtx t2 = gen_reg_rtx (DImode);
|
2142 |
|
|
emit_insn (gen_extendsidi2 (t2, target));
|
2143 |
|
|
target = t2;
|
2144 |
|
|
}
|
2145 |
|
|
|
2146 |
|
|
return gen_rtx_fmt_ee (code, VOIDmode, target, const0_rtx);
|
2147 |
|
|
}
|
2148 |
|
|
|
2149 |
|
|
/* Called from the md file, set up the operands of a compare instruction. */
|
2150 |
|
|
|
2151 |
|
|
void
|
2152 |
|
|
sh_emit_compare_and_branch (rtx *operands, enum machine_mode mode)
|
2153 |
|
|
{
|
2154 |
|
|
enum rtx_code code = GET_CODE (operands[0]);
|
2155 |
|
|
enum rtx_code branch_code;
|
2156 |
|
|
rtx op0 = operands[1];
|
2157 |
|
|
rtx op1 = operands[2];
|
2158 |
|
|
rtx insn, tem;
|
2159 |
|
|
bool need_ccmpeq = false;
|
2160 |
|
|
|
2161 |
|
|
if (TARGET_SH2E && GET_MODE_CLASS (mode) == MODE_FLOAT)
|
2162 |
|
|
{
|
2163 |
|
|
op0 = force_reg (mode, op0);
|
2164 |
|
|
op1 = force_reg (mode, op1);
|
2165 |
|
|
}
|
2166 |
|
|
else
|
2167 |
|
|
{
|
2168 |
|
|
if (code != EQ || mode == DImode)
|
2169 |
|
|
{
|
2170 |
|
|
/* Force args into regs, since we can't use constants here. */
|
2171 |
|
|
op0 = force_reg (mode, op0);
|
2172 |
|
|
if (op1 != const0_rtx || code == GTU || code == GEU)
|
2173 |
|
|
op1 = force_reg (mode, op1);
|
2174 |
|
|
}
|
2175 |
|
|
}
|
2176 |
|
|
|
2177 |
|
|
if (GET_MODE_CLASS (mode) == MODE_FLOAT)
|
2178 |
|
|
{
|
2179 |
|
|
if (code == LT
|
2180 |
|
|
|| (code == LE && TARGET_IEEE && TARGET_SH2E)
|
2181 |
|
|
|| (code == GE && !(TARGET_IEEE && TARGET_SH2E)))
|
2182 |
|
|
{
|
2183 |
|
|
tem = op0, op0 = op1, op1 = tem;
|
2184 |
|
|
code = swap_condition (code);
|
2185 |
|
|
}
|
2186 |
|
|
|
2187 |
|
|
/* GE becomes fcmp/gt+fcmp/eq, for SH2E and TARGET_IEEE only. */
|
2188 |
|
|
if (code == GE)
|
2189 |
|
|
{
|
2190 |
|
|
gcc_assert (TARGET_IEEE && TARGET_SH2E);
|
2191 |
|
|
need_ccmpeq = true;
|
2192 |
|
|
code = GT;
|
2193 |
|
|
}
|
2194 |
|
|
|
2195 |
|
|
/* Now we can have EQ, NE, GT, LE. NE and LE are then transformed
|
2196 |
|
|
to EQ/GT respectively. */
|
2197 |
|
|
gcc_assert (code == EQ || code == GT || code == NE || code == LE);
|
2198 |
|
|
}
|
2199 |
|
|
|
2200 |
|
|
switch (code)
|
2201 |
|
|
{
|
2202 |
|
|
case EQ:
|
2203 |
|
|
case GT:
|
2204 |
|
|
case GE:
|
2205 |
|
|
case GTU:
|
2206 |
|
|
case GEU:
|
2207 |
|
|
branch_code = code;
|
2208 |
|
|
break;
|
2209 |
|
|
case NE:
|
2210 |
|
|
case LT:
|
2211 |
|
|
case LE:
|
2212 |
|
|
case LTU:
|
2213 |
|
|
case LEU:
|
2214 |
|
|
branch_code = reverse_condition (code);
|
2215 |
|
|
break;
|
2216 |
|
|
default:
|
2217 |
|
|
gcc_unreachable ();
|
2218 |
|
|
}
|
2219 |
|
|
|
2220 |
|
|
insn = gen_rtx_SET (VOIDmode,
|
2221 |
|
|
gen_rtx_REG (SImode, T_REG),
|
2222 |
|
|
gen_rtx_fmt_ee (branch_code, SImode, op0, op1));
|
2223 |
|
|
|
2224 |
|
|
sh_emit_set_t_insn (insn, mode);
|
2225 |
|
|
if (need_ccmpeq)
|
2226 |
|
|
sh_emit_set_t_insn (gen_ieee_ccmpeqsf_t (op0, op1), mode);
|
2227 |
|
|
|
2228 |
|
|
if (branch_code == code)
|
2229 |
|
|
emit_jump_insn (gen_branch_true (operands[3]));
|
2230 |
|
|
else
|
2231 |
|
|
emit_jump_insn (gen_branch_false (operands[3]));
|
2232 |
|
|
}
|
2233 |
|
|
|
2234 |
|
|
void
|
2235 |
|
|
sh_emit_compare_and_set (rtx *operands, enum machine_mode mode)
|
2236 |
|
|
{
|
2237 |
|
|
enum rtx_code code = GET_CODE (operands[1]);
|
2238 |
|
|
rtx op0 = operands[2];
|
2239 |
|
|
rtx op1 = operands[3];
|
2240 |
|
|
rtx lab = NULL_RTX;
|
2241 |
|
|
bool invert = false;
|
2242 |
|
|
rtx tem;
|
2243 |
|
|
|
2244 |
|
|
op0 = force_reg (mode, op0);
|
2245 |
|
|
if ((code != EQ && code != NE
|
2246 |
|
|
&& (op1 != const0_rtx
|
2247 |
|
|
|| code == GTU || code == GEU || code == LTU || code == LEU))
|
2248 |
|
|
|| (mode == DImode && op1 != const0_rtx)
|
2249 |
|
|
|| (TARGET_SH2E && GET_MODE_CLASS (mode) == MODE_FLOAT))
|
2250 |
|
|
op1 = force_reg (mode, op1);
|
2251 |
|
|
|
2252 |
|
|
if (GET_MODE_CLASS (mode) == MODE_FLOAT)
|
2253 |
|
|
{
|
2254 |
|
|
if (code == LT || code == LE)
|
2255 |
|
|
{
|
2256 |
|
|
code = swap_condition (code);
|
2257 |
|
|
tem = op0, op0 = op1, op1 = tem;
|
2258 |
|
|
}
|
2259 |
|
|
if (code == GE)
|
2260 |
|
|
{
|
2261 |
|
|
if (TARGET_IEEE)
|
2262 |
|
|
{
|
2263 |
|
|
lab = gen_label_rtx ();
|
2264 |
|
|
sh_emit_scc_to_t (EQ, op0, op1);
|
2265 |
|
|
emit_jump_insn (gen_branch_true (lab));
|
2266 |
|
|
code = GT;
|
2267 |
|
|
}
|
2268 |
|
|
else
|
2269 |
|
|
{
|
2270 |
|
|
code = LT;
|
2271 |
|
|
invert = true;
|
2272 |
|
|
}
|
2273 |
|
|
}
|
2274 |
|
|
}
|
2275 |
|
|
|
2276 |
|
|
if (code == NE)
|
2277 |
|
|
{
|
2278 |
|
|
code = EQ;
|
2279 |
|
|
invert = true;
|
2280 |
|
|
}
|
2281 |
|
|
|
2282 |
|
|
sh_emit_scc_to_t (code, op0, op1);
|
2283 |
|
|
if (lab)
|
2284 |
|
|
emit_label (lab);
|
2285 |
|
|
if (invert)
|
2286 |
|
|
emit_insn (gen_movnegt (operands[0]));
|
2287 |
|
|
else
|
2288 |
|
|
emit_move_insn (operands[0], gen_rtx_REG (SImode, T_REG));
|
2289 |
|
|
}
|
2290 |
|
|
|
2291 |
|
|
/* Functions to output assembly code. */
|
2292 |
|
|
|
2293 |
|
|
/* Return a sequence of instructions to perform DI or DF move.
|
2294 |
|
|
|
2295 |
|
|
Since the SH cannot move a DI or DF in one instruction, we have
|
2296 |
|
|
to take care when we see overlapping source and dest registers. */
|
2297 |
|
|
|
2298 |
|
|
const char *
|
2299 |
|
|
output_movedouble (rtx insn ATTRIBUTE_UNUSED, rtx operands[],
|
2300 |
|
|
enum machine_mode mode)
|
2301 |
|
|
{
|
2302 |
|
|
rtx dst = operands[0];
|
2303 |
|
|
rtx src = operands[1];
|
2304 |
|
|
|
2305 |
|
|
if (MEM_P (dst)
|
2306 |
|
|
&& GET_CODE (XEXP (dst, 0)) == PRE_DEC)
|
2307 |
|
|
return "mov.l %T1,%0\n\tmov.l %1,%0";
|
2308 |
|
|
|
2309 |
|
|
if (register_operand (dst, mode)
|
2310 |
|
|
&& register_operand (src, mode))
|
2311 |
|
|
{
|
2312 |
|
|
if (REGNO (src) == MACH_REG)
|
2313 |
|
|
return "sts mach,%S0\n\tsts macl,%R0";
|
2314 |
|
|
|
2315 |
|
|
/* When mov.d r1,r2 do r2->r3 then r1->r2;
|
2316 |
|
|
when mov.d r1,r0 do r1->r0 then r2->r1. */
|
2317 |
|
|
|
2318 |
|
|
if (REGNO (src) + 1 == REGNO (dst))
|
2319 |
|
|
return "mov %T1,%T0\n\tmov %1,%0";
|
2320 |
|
|
else
|
2321 |
|
|
return "mov %1,%0\n\tmov %T1,%T0";
|
2322 |
|
|
}
|
2323 |
|
|
else if (CONST_INT_P (src))
|
2324 |
|
|
{
|
2325 |
|
|
if (INTVAL (src) < 0)
|
2326 |
|
|
output_asm_insn ("mov #-1,%S0", operands);
|
2327 |
|
|
else
|
2328 |
|
|
output_asm_insn ("mov #0,%S0", operands);
|
2329 |
|
|
|
2330 |
|
|
return "mov %1,%R0";
|
2331 |
|
|
}
|
2332 |
|
|
else if (MEM_P (src))
|
2333 |
|
|
{
|
2334 |
|
|
int ptrreg = -1;
|
2335 |
|
|
int dreg = REGNO (dst);
|
2336 |
|
|
rtx inside = XEXP (src, 0);
|
2337 |
|
|
|
2338 |
|
|
switch (GET_CODE (inside))
|
2339 |
|
|
{
|
2340 |
|
|
case REG:
|
2341 |
|
|
ptrreg = REGNO (inside);
|
2342 |
|
|
break;
|
2343 |
|
|
|
2344 |
|
|
case SUBREG:
|
2345 |
|
|
ptrreg = subreg_regno (inside);
|
2346 |
|
|
break;
|
2347 |
|
|
|
2348 |
|
|
case PLUS:
|
2349 |
|
|
ptrreg = REGNO (XEXP (inside, 0));
|
2350 |
|
|
/* ??? A r0+REG address shouldn't be possible here, because it isn't
|
2351 |
|
|
an offsettable address. Unfortunately, offsettable addresses use
|
2352 |
|
|
QImode to check the offset, and a QImode offsettable address
|
2353 |
|
|
requires r0 for the other operand, which is not currently
|
2354 |
|
|
supported, so we can't use the 'o' constraint.
|
2355 |
|
|
Thus we must check for and handle r0+REG addresses here.
|
2356 |
|
|
We punt for now, since this is likely very rare. */
|
2357 |
|
|
gcc_assert (!REG_P (XEXP (inside, 1)));
|
2358 |
|
|
break;
|
2359 |
|
|
|
2360 |
|
|
case LABEL_REF:
|
2361 |
|
|
return "mov.l %1,%0\n\tmov.l %1+4,%T0";
|
2362 |
|
|
case POST_INC:
|
2363 |
|
|
return "mov.l %1,%0\n\tmov.l %1,%T0";
|
2364 |
|
|
default:
|
2365 |
|
|
gcc_unreachable ();
|
2366 |
|
|
}
|
2367 |
|
|
|
2368 |
|
|
/* Work out the safe way to copy. Copy into the second half first. */
|
2369 |
|
|
if (dreg == ptrreg)
|
2370 |
|
|
return "mov.l %T1,%T0\n\tmov.l %1,%0";
|
2371 |
|
|
}
|
2372 |
|
|
|
2373 |
|
|
return "mov.l %1,%0\n\tmov.l %T1,%T0";
|
2374 |
|
|
}
|
2375 |
|
|
|
2376 |
|
|
/* Print an instruction which would have gone into a delay slot after
|
2377 |
|
|
another instruction, but couldn't because the other instruction expanded
|
2378 |
|
|
into a sequence where putting the slot insn at the end wouldn't work. */
|
2379 |
|
|
|
2380 |
|
|
static void
|
2381 |
|
|
print_slot (rtx insn)
|
2382 |
|
|
{
|
2383 |
|
|
final_scan_insn (XVECEXP (insn, 0, 1), asm_out_file, optimize, 1, NULL);
|
2384 |
|
|
|
2385 |
|
|
INSN_DELETED_P (XVECEXP (insn, 0, 1)) = 1;
|
2386 |
|
|
}
|
2387 |
|
|
|
2388 |
|
|
const char *
|
2389 |
|
|
output_far_jump (rtx insn, rtx op)
|
2390 |
|
|
{
|
2391 |
|
|
struct { rtx lab, reg, op; } this_jmp;
|
2392 |
|
|
rtx braf_base_lab = NULL_RTX;
|
2393 |
|
|
const char *jump;
|
2394 |
|
|
int far;
|
2395 |
|
|
int offset = branch_dest (insn) - INSN_ADDRESSES (INSN_UID (insn));
|
2396 |
|
|
rtx prev;
|
2397 |
|
|
|
2398 |
|
|
this_jmp.lab = gen_label_rtx ();
|
2399 |
|
|
|
2400 |
|
|
if (TARGET_SH2
|
2401 |
|
|
&& offset >= -32764
|
2402 |
|
|
&& offset - get_attr_length (insn) <= 32766)
|
2403 |
|
|
{
|
2404 |
|
|
far = 0;
|
2405 |
|
|
jump = "mov.w %O0,%1; braf %1";
|
2406 |
|
|
}
|
2407 |
|
|
else
|
2408 |
|
|
{
|
2409 |
|
|
far = 1;
|
2410 |
|
|
if (flag_pic)
|
2411 |
|
|
{
|
2412 |
|
|
if (TARGET_SH2)
|
2413 |
|
|
jump = "mov.l %O0,%1; braf %1";
|
2414 |
|
|
else
|
2415 |
|
|
jump = "mov.l r0,@-r15; mova %O0,r0; mov.l @r0,%1; add r0,%1; mov.l @r15+,r0; jmp @%1";
|
2416 |
|
|
}
|
2417 |
|
|
else
|
2418 |
|
|
jump = "mov.l %O0,%1; jmp @%1";
|
2419 |
|
|
}
|
2420 |
|
|
/* If we have a scratch register available, use it. */
|
2421 |
|
|
if (NONJUMP_INSN_P ((prev = prev_nonnote_insn (insn)))
|
2422 |
|
|
&& INSN_CODE (prev) == CODE_FOR_indirect_jump_scratch)
|
2423 |
|
|
{
|
2424 |
|
|
this_jmp.reg = SET_DEST (XVECEXP (PATTERN (prev), 0, 0));
|
2425 |
|
|
if (REGNO (this_jmp.reg) == R0_REG && flag_pic && ! TARGET_SH2)
|
2426 |
|
|
jump = "mov.l r1,@-r15; mova %O0,r0; mov.l @r0,r1; add r1,r0; mov.l @r15+,r1; jmp @%1";
|
2427 |
|
|
output_asm_insn (jump, &this_jmp.lab);
|
2428 |
|
|
if (dbr_sequence_length ())
|
2429 |
|
|
print_slot (final_sequence);
|
2430 |
|
|
else
|
2431 |
|
|
output_asm_insn ("nop", 0);
|
2432 |
|
|
}
|
2433 |
|
|
else
|
2434 |
|
|
{
|
2435 |
|
|
/* Output the delay slot insn first if any. */
|
2436 |
|
|
if (dbr_sequence_length ())
|
2437 |
|
|
print_slot (final_sequence);
|
2438 |
|
|
|
2439 |
|
|
this_jmp.reg = gen_rtx_REG (SImode, 13);
|
2440 |
|
|
/* We must keep the stack aligned to 8-byte boundaries on SH5.
|
2441 |
|
|
Fortunately, MACL is fixed and call-clobbered, and we never
|
2442 |
|
|
need its value across jumps, so save r13 in it instead of in
|
2443 |
|
|
the stack. */
|
2444 |
|
|
if (TARGET_SH5)
|
2445 |
|
|
output_asm_insn ("lds r13, macl", 0);
|
2446 |
|
|
else
|
2447 |
|
|
output_asm_insn ("mov.l r13,@-r15", 0);
|
2448 |
|
|
output_asm_insn (jump, &this_jmp.lab);
|
2449 |
|
|
if (TARGET_SH5)
|
2450 |
|
|
output_asm_insn ("sts macl, r13", 0);
|
2451 |
|
|
else
|
2452 |
|
|
output_asm_insn ("mov.l @r15+,r13", 0);
|
2453 |
|
|
}
|
2454 |
|
|
if (far && flag_pic && TARGET_SH2)
|
2455 |
|
|
{
|
2456 |
|
|
braf_base_lab = gen_label_rtx ();
|
2457 |
|
|
(*targetm.asm_out.internal_label) (asm_out_file, "L",
|
2458 |
|
|
CODE_LABEL_NUMBER (braf_base_lab));
|
2459 |
|
|
}
|
2460 |
|
|
if (far)
|
2461 |
|
|
output_asm_insn (".align 2", 0);
|
2462 |
|
|
(*targetm.asm_out.internal_label) (asm_out_file, "L", CODE_LABEL_NUMBER (this_jmp.lab));
|
2463 |
|
|
this_jmp.op = op;
|
2464 |
|
|
if (far && flag_pic)
|
2465 |
|
|
{
|
2466 |
|
|
if (TARGET_SH2)
|
2467 |
|
|
this_jmp.lab = braf_base_lab;
|
2468 |
|
|
output_asm_insn (".long %O2-%O0", &this_jmp.lab);
|
2469 |
|
|
}
|
2470 |
|
|
else
|
2471 |
|
|
output_asm_insn (far ? ".long %O2" : ".word %O2-%O0", &this_jmp.lab);
|
2472 |
|
|
return "";
|
2473 |
|
|
}
|
2474 |
|
|
|
2475 |
|
|
/* Local label counter, used for constants in the pool and inside
|
2476 |
|
|
pattern branches. */
|
2477 |
|
|
|
2478 |
|
|
static int lf = 100;
|
2479 |
|
|
|
2480 |
|
|
/* Output code for ordinary branches. */
|
2481 |
|
|
|
2482 |
|
|
const char *
|
2483 |
|
|
output_branch (int logic, rtx insn, rtx *operands)
|
2484 |
|
|
{
|
2485 |
|
|
switch (get_attr_length (insn))
|
2486 |
|
|
{
|
2487 |
|
|
case 6:
|
2488 |
|
|
/* This can happen if filling the delay slot has caused a forward
|
2489 |
|
|
branch to exceed its range (we could reverse it, but only
|
2490 |
|
|
when we know we won't overextend other branches; this should
|
2491 |
|
|
best be handled by relaxation).
|
2492 |
|
|
It can also happen when other condbranches hoist delay slot insn
|
2493 |
|
|
from their destination, thus leading to code size increase.
|
2494 |
|
|
But the branch will still be in the range -4092..+4098 bytes. */
|
2495 |
|
|
|
2496 |
|
|
if (! TARGET_RELAX)
|
2497 |
|
|
{
|
2498 |
|
|
int label = lf++;
|
2499 |
|
|
/* The call to print_slot will clobber the operands. */
|
2500 |
|
|
rtx op0 = operands[0];
|
2501 |
|
|
|
2502 |
|
|
/* If the instruction in the delay slot is annulled (true), then
|
2503 |
|
|
there is no delay slot where we can put it now. The only safe
|
2504 |
|
|
place for it is after the label. final will do that by default. */
|
2505 |
|
|
|
2506 |
|
|
if (final_sequence
|
2507 |
|
|
&& ! INSN_ANNULLED_BRANCH_P (XVECEXP (final_sequence, 0, 0))
|
2508 |
|
|
&& get_attr_length (XVECEXP (final_sequence, 0, 1)))
|
2509 |
|
|
{
|
2510 |
|
|
asm_fprintf (asm_out_file, "\tb%s%ss\t%LLF%d\n", logic ? "f" : "t",
|
2511 |
|
|
ASSEMBLER_DIALECT ? "/" : ".", label);
|
2512 |
|
|
print_slot (final_sequence);
|
2513 |
|
|
}
|
2514 |
|
|
else
|
2515 |
|
|
asm_fprintf (asm_out_file, "\tb%s\t%LLF%d\n", logic ? "f" : "t", label);
|
2516 |
|
|
|
2517 |
|
|
output_asm_insn ("bra\t%l0", &op0);
|
2518 |
|
|
fprintf (asm_out_file, "\tnop\n");
|
2519 |
|
|
(*targetm.asm_out.internal_label) (asm_out_file, "LF", label);
|
2520 |
|
|
|
2521 |
|
|
return "";
|
2522 |
|
|
}
|
2523 |
|
|
/* When relaxing, handle this like a short branch. The linker
|
2524 |
|
|
will fix it up if it still doesn't fit after relaxation. */
|
2525 |
|
|
case 2:
|
2526 |
|
|
return logic ? "bt%.\t%l0" : "bf%.\t%l0";
|
2527 |
|
|
|
2528 |
|
|
/* These are for SH2e, in which we have to account for the
|
2529 |
|
|
extra nop because of the hardware bug in annulled branches. */
|
2530 |
|
|
case 8:
|
2531 |
|
|
if (! TARGET_RELAX)
|
2532 |
|
|
{
|
2533 |
|
|
int label = lf++;
|
2534 |
|
|
|
2535 |
|
|
gcc_assert (!final_sequence
|
2536 |
|
|
|| !(INSN_ANNULLED_BRANCH_P
|
2537 |
|
|
(XVECEXP (final_sequence, 0, 0))));
|
2538 |
|
|
asm_fprintf (asm_out_file, "b%s%ss\t%LLF%d\n",
|
2539 |
|
|
logic ? "f" : "t",
|
2540 |
|
|
ASSEMBLER_DIALECT ? "/" : ".", label);
|
2541 |
|
|
fprintf (asm_out_file, "\tnop\n");
|
2542 |
|
|
output_asm_insn ("bra\t%l0", operands);
|
2543 |
|
|
fprintf (asm_out_file, "\tnop\n");
|
2544 |
|
|
(*targetm.asm_out.internal_label) (asm_out_file, "LF", label);
|
2545 |
|
|
|
2546 |
|
|
return "";
|
2547 |
|
|
}
|
2548 |
|
|
/* When relaxing, fall through. */
|
2549 |
|
|
case 4:
|
2550 |
|
|
{
|
2551 |
|
|
char buffer[10];
|
2552 |
|
|
|
2553 |
|
|
sprintf (buffer, "b%s%ss\t%%l0",
|
2554 |
|
|
logic ? "t" : "f",
|
2555 |
|
|
ASSEMBLER_DIALECT ? "/" : ".");
|
2556 |
|
|
output_asm_insn (buffer, &operands[0]);
|
2557 |
|
|
return "nop";
|
2558 |
|
|
}
|
2559 |
|
|
|
2560 |
|
|
default:
|
2561 |
|
|
/* There should be no longer branches now - that would
|
2562 |
|
|
indicate that something has destroyed the branches set
|
2563 |
|
|
up in machine_dependent_reorg. */
|
2564 |
|
|
gcc_unreachable ();
|
2565 |
|
|
}
|
2566 |
|
|
}
|
2567 |
|
|
|
2568 |
|
|
/* Output a code sequence for INSN using TEMPL with OPERANDS; but before,
|
2569 |
|
|
fill in operands 9 as a label to the successor insn.
|
2570 |
|
|
We try to use jump threading where possible.
|
2571 |
|
|
IF CODE matches the comparison in the IF_THEN_ELSE of a following jump,
|
2572 |
|
|
we assume the jump is taken. I.e. EQ means follow jmp and bf, NE means
|
2573 |
|
|
follow jmp and bt, if the address is in range. */
|
2574 |
|
|
const char *
|
2575 |
|
|
output_branchy_insn (enum rtx_code code, const char *templ,
|
2576 |
|
|
rtx insn, rtx *operands)
|
2577 |
|
|
{
|
2578 |
|
|
rtx next_insn = NEXT_INSN (insn);
|
2579 |
|
|
|
2580 |
|
|
if (next_insn && JUMP_P (next_insn) && condjump_p (next_insn))
|
2581 |
|
|
{
|
2582 |
|
|
rtx src = SET_SRC (PATTERN (next_insn));
|
2583 |
|
|
if (GET_CODE (src) == IF_THEN_ELSE && GET_CODE (XEXP (src, 0)) != code)
|
2584 |
|
|
{
|
2585 |
|
|
/* Following branch not taken */
|
2586 |
|
|
operands[9] = gen_label_rtx ();
|
2587 |
|
|
emit_label_after (operands[9], next_insn);
|
2588 |
|
|
INSN_ADDRESSES_NEW (operands[9],
|
2589 |
|
|
INSN_ADDRESSES (INSN_UID (next_insn))
|
2590 |
|
|
+ get_attr_length (next_insn));
|
2591 |
|
|
return templ;
|
2592 |
|
|
}
|
2593 |
|
|
else
|
2594 |
|
|
{
|
2595 |
|
|
int offset = (branch_dest (next_insn)
|
2596 |
|
|
- INSN_ADDRESSES (INSN_UID (next_insn)) + 4);
|
2597 |
|
|
if (offset >= -252 && offset <= 258)
|
2598 |
|
|
{
|
2599 |
|
|
if (GET_CODE (src) == IF_THEN_ELSE)
|
2600 |
|
|
/* branch_true */
|
2601 |
|
|
src = XEXP (src, 1);
|
2602 |
|
|
operands[9] = src;
|
2603 |
|
|
return templ;
|
2604 |
|
|
}
|
2605 |
|
|
}
|
2606 |
|
|
}
|
2607 |
|
|
operands[9] = gen_label_rtx ();
|
2608 |
|
|
emit_label_after (operands[9], insn);
|
2609 |
|
|
INSN_ADDRESSES_NEW (operands[9],
|
2610 |
|
|
INSN_ADDRESSES (INSN_UID (insn))
|
2611 |
|
|
+ get_attr_length (insn));
|
2612 |
|
|
return templ;
|
2613 |
|
|
}
|
2614 |
|
|
|
2615 |
|
|
const char *
|
2616 |
|
|
output_ieee_ccmpeq (rtx insn, rtx *operands)
|
2617 |
|
|
{
|
2618 |
|
|
return output_branchy_insn (NE, "bt\t%l9\n\tfcmp/eq\t%1,%0",
|
2619 |
|
|
insn, operands);
|
2620 |
|
|
}
|
2621 |
|
|
|
2622 |
|
|
/* Output the start of the assembler file. */
|
2623 |
|
|
|
2624 |
|
|
static void
|
2625 |
|
|
sh_file_start (void)
|
2626 |
|
|
{
|
2627 |
|
|
default_file_start ();
|
2628 |
|
|
|
2629 |
|
|
#ifdef SYMBIAN
|
2630 |
|
|
/* Declare the .directive section before it is used. */
|
2631 |
|
|
fputs ("\t.section .directive, \"SM\", @progbits, 1\n", asm_out_file);
|
2632 |
|
|
fputs ("\t.asciz \"#<SYMEDIT>#\\n\"\n", asm_out_file);
|
2633 |
|
|
#endif
|
2634 |
|
|
|
2635 |
|
|
if (TARGET_ELF)
|
2636 |
|
|
/* We need to show the text section with the proper
|
2637 |
|
|
attributes as in TEXT_SECTION_ASM_OP, before dwarf2out
|
2638 |
|
|
emits it without attributes in TEXT_SECTION_ASM_OP, else GAS
|
2639 |
|
|
will complain. We can teach GAS specifically about the
|
2640 |
|
|
default attributes for our choice of text section, but
|
2641 |
|
|
then we would have to change GAS again if/when we change
|
2642 |
|
|
the text section name. */
|
2643 |
|
|
fprintf (asm_out_file, "%s\n", TEXT_SECTION_ASM_OP);
|
2644 |
|
|
else
|
2645 |
|
|
/* Switch to the data section so that the coffsem symbol
|
2646 |
|
|
isn't in the text section. */
|
2647 |
|
|
switch_to_section (data_section);
|
2648 |
|
|
|
2649 |
|
|
if (TARGET_LITTLE_ENDIAN)
|
2650 |
|
|
fputs ("\t.little\n", asm_out_file);
|
2651 |
|
|
|
2652 |
|
|
if (!TARGET_ELF)
|
2653 |
|
|
{
|
2654 |
|
|
if (TARGET_SHCOMPACT)
|
2655 |
|
|
fputs ("\t.mode\tSHcompact\n", asm_out_file);
|
2656 |
|
|
else if (TARGET_SHMEDIA)
|
2657 |
|
|
fprintf (asm_out_file, "\t.mode\tSHmedia\n\t.abi\t%i\n",
|
2658 |
|
|
TARGET_SHMEDIA64 ? 64 : 32);
|
2659 |
|
|
}
|
2660 |
|
|
}
|
2661 |
|
|
|
2662 |
|
|
/* Check if PAT includes UNSPEC_CALLER unspec pattern. */
|
2663 |
|
|
|
2664 |
|
|
static bool
|
2665 |
|
|
unspec_caller_rtx_p (rtx pat)
|
2666 |
|
|
{
|
2667 |
|
|
rtx base, offset;
|
2668 |
|
|
int i;
|
2669 |
|
|
|
2670 |
|
|
split_const (pat, &base, &offset);
|
2671 |
|
|
if (GET_CODE (base) == UNSPEC)
|
2672 |
|
|
{
|
2673 |
|
|
if (XINT (base, 1) == UNSPEC_CALLER)
|
2674 |
|
|
return true;
|
2675 |
|
|
for (i = 0; i < XVECLEN (base, 0); i++)
|
2676 |
|
|
if (unspec_caller_rtx_p (XVECEXP (base, 0, i)))
|
2677 |
|
|
return true;
|
2678 |
|
|
}
|
2679 |
|
|
return false;
|
2680 |
|
|
}
|
2681 |
|
|
|
2682 |
|
|
/* Indicate that INSN cannot be duplicated. This is true for insn
|
2683 |
|
|
that generates a unique label. */
|
2684 |
|
|
|
2685 |
|
|
static bool
|
2686 |
|
|
sh_cannot_copy_insn_p (rtx insn)
|
2687 |
|
|
{
|
2688 |
|
|
rtx pat;
|
2689 |
|
|
|
2690 |
|
|
if (!reload_completed || !flag_pic)
|
2691 |
|
|
return false;
|
2692 |
|
|
|
2693 |
|
|
if (!NONJUMP_INSN_P (insn))
|
2694 |
|
|
return false;
|
2695 |
|
|
if (asm_noperands (insn) >= 0)
|
2696 |
|
|
return false;
|
2697 |
|
|
|
2698 |
|
|
pat = PATTERN (insn);
|
2699 |
|
|
if (GET_CODE (pat) != SET)
|
2700 |
|
|
return false;
|
2701 |
|
|
pat = SET_SRC (pat);
|
2702 |
|
|
|
2703 |
|
|
if (unspec_caller_rtx_p (pat))
|
2704 |
|
|
return true;
|
2705 |
|
|
|
2706 |
|
|
return false;
|
2707 |
|
|
}
|
2708 |
|
|
|
2709 |
|
|
/* Actual number of instructions used to make a shift by N. */
|
2710 |
|
|
static const char ashiftrt_insns[] =
|
2711 |
|
|
{ 0,1,2,3,4,5,8,8,8,8,8,8,8,8,8,8,2,3,4,5,8,8,8,8,8,8,8,8,8,8,8,2};
|
2712 |
|
|
|
2713 |
|
|
/* Left shift and logical right shift are the same. */
|
2714 |
|
|
static const char shift_insns[] =
|
2715 |
|
|
{ 0,1,1,2,2,3,3,4,1,2,2,3,3,4,3,3,1,2,2,3,3,4,3,3,2,3,3,4,4,4,3,3};
|
2716 |
|
|
|
2717 |
|
|
/* Individual shift amounts needed to get the above length sequences.
|
2718 |
|
|
One bit right shifts clobber the T bit, so when possible, put one bit
|
2719 |
|
|
shifts in the middle of the sequence, so the ends are eligible for
|
2720 |
|
|
branch delay slots. */
|
2721 |
|
|
static const short shift_amounts[32][5] = {
|
2722 |
|
|
{0}, {1}, {2}, {2, 1},
|
2723 |
|
|
{2, 2}, {2, 1, 2}, {2, 2, 2}, {2, 2, 1, 2},
|
2724 |
|
|
{8}, {8, 1}, {8, 2}, {8, 1, 2},
|
2725 |
|
|
{8, 2, 2}, {8, 2, 1, 2}, {8, -2, 8}, {8, -1, 8},
|
2726 |
|
|
{16}, {16, 1}, {16, 2}, {16, 1, 2},
|
2727 |
|
|
{16, 2, 2}, {16, 2, 1, 2}, {16, -2, 8}, {16, -1, 8},
|
2728 |
|
|
{16, 8}, {16, 1, 8}, {16, 8, 2}, {16, 8, 1, 2},
|
2729 |
|
|
{16, 8, 2, 2}, {16, -1, -2, 16}, {16, -2, 16}, {16, -1, 16}};
|
2730 |
|
|
|
2731 |
|
|
/* Likewise, but for shift amounts < 16, up to three highmost bits
|
2732 |
|
|
might be clobbered. This is typically used when combined with some
|
2733 |
|
|
kind of sign or zero extension. */
|
2734 |
|
|
|
2735 |
|
|
static const char ext_shift_insns[] =
|
2736 |
|
|
{ 0,1,1,2,2,3,2,2,1,2,2,3,3,3,2,2,1,2,2,3,3,4,3,3,2,3,3,4,4,4,3,3};
|
2737 |
|
|
|
2738 |
|
|
static const short ext_shift_amounts[32][4] = {
|
2739 |
|
|
{0}, {1}, {2}, {2, 1},
|
2740 |
|
|
{2, 2}, {2, 1, 2}, {8, -2}, {8, -1},
|
2741 |
|
|
{8}, {8, 1}, {8, 2}, {8, 1, 2},
|
2742 |
|
|
{8, 2, 2}, {16, -2, -1}, {16, -2}, {16, -1},
|
2743 |
|
|
{16}, {16, 1}, {16, 2}, {16, 1, 2},
|
2744 |
|
|
{16, 2, 2}, {16, 2, 1, 2}, {16, -2, 8}, {16, -1, 8},
|
2745 |
|
|
{16, 8}, {16, 1, 8}, {16, 8, 2}, {16, 8, 1, 2},
|
2746 |
|
|
{16, 8, 2, 2}, {16, -1, -2, 16}, {16, -2, 16}, {16, -1, 16}};
|
2747 |
|
|
|
2748 |
|
|
/* Assuming we have a value that has been sign-extended by at least one bit,
|
2749 |
|
|
can we use the ext_shift_amounts with the last shift turned to an arithmetic shift
|
2750 |
|
|
to shift it by N without data loss, and quicker than by other means? */
|
2751 |
|
|
#define EXT_SHIFT_SIGNED(n) (((n) | 8) == 15)
|
2752 |
|
|
|
2753 |
|
|
/* This is used in length attributes in sh.md to help compute the length
|
2754 |
|
|
of arbitrary constant shift instructions. */
|
2755 |
|
|
|
2756 |
|
|
int
|
2757 |
|
|
shift_insns_rtx (rtx insn)
|
2758 |
|
|
{
|
2759 |
|
|
rtx set_src = SET_SRC (XVECEXP (PATTERN (insn), 0, 0));
|
2760 |
|
|
int shift_count = INTVAL (XEXP (set_src, 1)) & 31;
|
2761 |
|
|
enum rtx_code shift_code = GET_CODE (set_src);
|
2762 |
|
|
|
2763 |
|
|
switch (shift_code)
|
2764 |
|
|
{
|
2765 |
|
|
case ASHIFTRT:
|
2766 |
|
|
return ashiftrt_insns[shift_count];
|
2767 |
|
|
case LSHIFTRT:
|
2768 |
|
|
case ASHIFT:
|
2769 |
|
|
return shift_insns[shift_count];
|
2770 |
|
|
default:
|
2771 |
|
|
gcc_unreachable ();
|
2772 |
|
|
}
|
2773 |
|
|
}
|
2774 |
|
|
|
2775 |
|
|
/* Return the cost of a shift. */
|
2776 |
|
|
|
2777 |
|
|
static inline int
|
2778 |
|
|
shiftcosts (rtx x)
|
2779 |
|
|
{
|
2780 |
|
|
int value;
|
2781 |
|
|
|
2782 |
|
|
if (TARGET_SHMEDIA)
|
2783 |
|
|
return 1;
|
2784 |
|
|
|
2785 |
|
|
if (GET_MODE_SIZE (GET_MODE (x)) > UNITS_PER_WORD)
|
2786 |
|
|
{
|
2787 |
|
|
if (GET_MODE (x) == DImode
|
2788 |
|
|
&& CONST_INT_P (XEXP (x, 1))
|
2789 |
|
|
&& INTVAL (XEXP (x, 1)) == 1)
|
2790 |
|
|
return 2;
|
2791 |
|
|
|
2792 |
|
|
/* Everything else is invalid, because there is no pattern for it. */
|
2793 |
|
|
return MAX_COST;
|
2794 |
|
|
}
|
2795 |
|
|
/* If shift by a non constant, then this will be expensive. */
|
2796 |
|
|
if (!CONST_INT_P (XEXP (x, 1)))
|
2797 |
|
|
return SH_DYNAMIC_SHIFT_COST;
|
2798 |
|
|
|
2799 |
|
|
/* Otherwise, return the true cost in instructions. Cope with out of range
|
2800 |
|
|
shift counts more or less arbitrarily. */
|
2801 |
|
|
value = INTVAL (XEXP (x, 1)) & 31;
|
2802 |
|
|
|
2803 |
|
|
if (GET_CODE (x) == ASHIFTRT)
|
2804 |
|
|
{
|
2805 |
|
|
int cost = ashiftrt_insns[value];
|
2806 |
|
|
/* If SH3, then we put the constant in a reg and use shad. */
|
2807 |
|
|
if (cost > 1 + SH_DYNAMIC_SHIFT_COST)
|
2808 |
|
|
cost = 1 + SH_DYNAMIC_SHIFT_COST;
|
2809 |
|
|
return cost;
|
2810 |
|
|
}
|
2811 |
|
|
else
|
2812 |
|
|
return shift_insns[value];
|
2813 |
|
|
}
|
2814 |
|
|
|
2815 |
|
|
/* Return the cost of an AND operation. */
|
2816 |
|
|
|
2817 |
|
|
static inline int
|
2818 |
|
|
andcosts (rtx x)
|
2819 |
|
|
{
|
2820 |
|
|
int i;
|
2821 |
|
|
|
2822 |
|
|
/* Anding with a register is a single cycle and instruction. */
|
2823 |
|
|
if (!CONST_INT_P (XEXP (x, 1)))
|
2824 |
|
|
return 1;
|
2825 |
|
|
|
2826 |
|
|
i = INTVAL (XEXP (x, 1));
|
2827 |
|
|
|
2828 |
|
|
if (TARGET_SHMEDIA)
|
2829 |
|
|
{
|
2830 |
|
|
if (satisfies_constraint_I10 (XEXP (x, 1))
|
2831 |
|
|
|| satisfies_constraint_J16 (XEXP (x, 1)))
|
2832 |
|
|
return 1;
|
2833 |
|
|
else
|
2834 |
|
|
return 1 + rtx_cost (XEXP (x, 1), AND, !optimize_size);
|
2835 |
|
|
}
|
2836 |
|
|
|
2837 |
|
|
/* These constants are single cycle extu.[bw] instructions. */
|
2838 |
|
|
if (i == 0xff || i == 0xffff)
|
2839 |
|
|
return 1;
|
2840 |
|
|
/* Constants that can be used in an and immediate instruction in a single
|
2841 |
|
|
cycle, but this requires r0, so make it a little more expensive. */
|
2842 |
|
|
if (CONST_OK_FOR_K08 (i))
|
2843 |
|
|
return 2;
|
2844 |
|
|
/* Constants that can be loaded with a mov immediate and an and.
|
2845 |
|
|
This case is probably unnecessary. */
|
2846 |
|
|
if (CONST_OK_FOR_I08 (i))
|
2847 |
|
|
return 2;
|
2848 |
|
|
/* Any other constants requires a 2 cycle pc-relative load plus an and.
|
2849 |
|
|
This case is probably unnecessary. */
|
2850 |
|
|
return 3;
|
2851 |
|
|
}
|
2852 |
|
|
|
2853 |
|
|
/* Return the cost of an addition or a subtraction. */
|
2854 |
|
|
|
2855 |
|
|
static inline int
|
2856 |
|
|
addsubcosts (rtx x)
|
2857 |
|
|
{
|
2858 |
|
|
/* Adding a register is a single cycle insn. */
|
2859 |
|
|
if (REG_P (XEXP (x, 1))
|
2860 |
|
|
|| GET_CODE (XEXP (x, 1)) == SUBREG)
|
2861 |
|
|
return 1;
|
2862 |
|
|
|
2863 |
|
|
/* Likewise for small constants. */
|
2864 |
|
|
if (CONST_INT_P (XEXP (x, 1))
|
2865 |
|
|
&& CONST_OK_FOR_ADD (INTVAL (XEXP (x, 1))))
|
2866 |
|
|
return 1;
|
2867 |
|
|
|
2868 |
|
|
if (TARGET_SHMEDIA)
|
2869 |
|
|
switch (GET_CODE (XEXP (x, 1)))
|
2870 |
|
|
{
|
2871 |
|
|
case CONST:
|
2872 |
|
|
case LABEL_REF:
|
2873 |
|
|
case SYMBOL_REF:
|
2874 |
|
|
return TARGET_SHMEDIA64 ? 5 : 3;
|
2875 |
|
|
|
2876 |
|
|
case CONST_INT:
|
2877 |
|
|
if (CONST_OK_FOR_I16 (INTVAL (XEXP (x, 1))))
|
2878 |
|
|
return 2;
|
2879 |
|
|
else if (CONST_OK_FOR_I16 (INTVAL (XEXP (x, 1)) >> 16))
|
2880 |
|
|
return 3;
|
2881 |
|
|
else if (CONST_OK_FOR_I16 ((INTVAL (XEXP (x, 1)) >> 16) >> 16))
|
2882 |
|
|
return 4;
|
2883 |
|
|
|
2884 |
|
|
/* Fall through. */
|
2885 |
|
|
default:
|
2886 |
|
|
return 5;
|
2887 |
|
|
}
|
2888 |
|
|
|
2889 |
|
|
/* Any other constant requires a 2 cycle pc-relative load plus an
|
2890 |
|
|
addition. */
|
2891 |
|
|
return 3;
|
2892 |
|
|
}
|
2893 |
|
|
|
2894 |
|
|
/* Return the cost of a multiply. */
|
2895 |
|
|
static inline int
|
2896 |
|
|
multcosts (rtx x ATTRIBUTE_UNUSED)
|
2897 |
|
|
{
|
2898 |
|
|
if (sh_multcost >= 0)
|
2899 |
|
|
return sh_multcost;
|
2900 |
|
|
if (TARGET_SHMEDIA)
|
2901 |
|
|
/* ??? We have a mul insn, but it has a latency of three, and doesn't
|
2902 |
|
|
accept constants. Ideally, we would use a cost of one or two and
|
2903 |
|
|
add the cost of the operand, but disregard the latter when inside loops
|
2904 |
|
|
and loop invariant code motion is still to follow.
|
2905 |
|
|
Using a multiply first and splitting it later if it's a loss
|
2906 |
|
|
doesn't work because of different sign / zero extension semantics
|
2907 |
|
|
of multiplies vs. shifts. */
|
2908 |
|
|
return TARGET_SMALLCODE ? 2 : 3;
|
2909 |
|
|
|
2910 |
|
|
if (TARGET_SH2)
|
2911 |
|
|
{
|
2912 |
|
|
/* We have a mul insn, so we can never take more than the mul and the
|
2913 |
|
|
read of the mac reg, but count more because of the latency and extra
|
2914 |
|
|
reg usage. */
|
2915 |
|
|
if (TARGET_SMALLCODE)
|
2916 |
|
|
return 2;
|
2917 |
|
|
return 3;
|
2918 |
|
|
}
|
2919 |
|
|
|
2920 |
|
|
/* If we're aiming at small code, then just count the number of
|
2921 |
|
|
insns in a multiply call sequence. */
|
2922 |
|
|
if (TARGET_SMALLCODE)
|
2923 |
|
|
return 5;
|
2924 |
|
|
|
2925 |
|
|
/* Otherwise count all the insns in the routine we'd be calling too. */
|
2926 |
|
|
return 20;
|
2927 |
|
|
}
|
2928 |
|
|
|
2929 |
|
|
/* Compute a (partial) cost for rtx X. Return true if the complete
|
2930 |
|
|
cost has been computed, and false if subexpressions should be
|
2931 |
|
|
scanned. In either case, *TOTAL contains the cost result. */
|
2932 |
|
|
|
2933 |
|
|
static bool
|
2934 |
|
|
sh_rtx_costs (rtx x, int code, int outer_code, int *total,
|
2935 |
|
|
bool speed ATTRIBUTE_UNUSED)
|
2936 |
|
|
{
|
2937 |
|
|
switch (code)
|
2938 |
|
|
{
|
2939 |
|
|
case CONST_INT:
|
2940 |
|
|
if (TARGET_SHMEDIA)
|
2941 |
|
|
{
|
2942 |
|
|
if (INTVAL (x) == 0)
|
2943 |
|
|
*total = 0;
|
2944 |
|
|
else if (outer_code == AND && and_operand ((x), DImode))
|
2945 |
|
|
*total = 0;
|
2946 |
|
|
else if ((outer_code == IOR || outer_code == XOR
|
2947 |
|
|
|| outer_code == PLUS)
|
2948 |
|
|
&& CONST_OK_FOR_I10 (INTVAL (x)))
|
2949 |
|
|
*total = 0;
|
2950 |
|
|
else if (CONST_OK_FOR_I16 (INTVAL (x)))
|
2951 |
|
|
*total = COSTS_N_INSNS (outer_code != SET);
|
2952 |
|
|
else if (CONST_OK_FOR_I16 (INTVAL (x) >> 16))
|
2953 |
|
|
*total = COSTS_N_INSNS ((outer_code != SET) + 1);
|
2954 |
|
|
else if (CONST_OK_FOR_I16 ((INTVAL (x) >> 16) >> 16))
|
2955 |
|
|
*total = COSTS_N_INSNS ((outer_code != SET) + 2);
|
2956 |
|
|
else
|
2957 |
|
|
*total = COSTS_N_INSNS ((outer_code != SET) + 3);
|
2958 |
|
|
return true;
|
2959 |
|
|
}
|
2960 |
|
|
if (CONST_OK_FOR_I08 (INTVAL (x)))
|
2961 |
|
|
*total = 0;
|
2962 |
|
|
else if ((outer_code == AND || outer_code == IOR || outer_code == XOR)
|
2963 |
|
|
&& CONST_OK_FOR_K08 (INTVAL (x)))
|
2964 |
|
|
*total = 1;
|
2965 |
|
|
/* prepare_cmp_insn will force costly constants int registers before
|
2966 |
|
|
the cbranch[sd]i4 patterns can see them, so preserve potentially
|
2967 |
|
|
interesting ones not covered by I08 above. */
|
2968 |
|
|
else if (outer_code == COMPARE
|
2969 |
|
|
&& ((unsigned HOST_WIDE_INT) INTVAL (x)
|
2970 |
|
|
== (unsigned HOST_WIDE_INT) 0x7fffffff + 1
|
2971 |
|
|
|| INTVAL (x) == 0x7fffffff
|
2972 |
|
|
|| INTVAL (x) == 0x80 || INTVAL (x) == -0x81))
|
2973 |
|
|
*total = 1;
|
2974 |
|
|
else
|
2975 |
|
|
*total = 8;
|
2976 |
|
|
return true;
|
2977 |
|
|
|
2978 |
|
|
case CONST:
|
2979 |
|
|
case LABEL_REF:
|
2980 |
|
|
case SYMBOL_REF:
|
2981 |
|
|
if (TARGET_SHMEDIA64)
|
2982 |
|
|
*total = COSTS_N_INSNS (4);
|
2983 |
|
|
else if (TARGET_SHMEDIA32)
|
2984 |
|
|
*total = COSTS_N_INSNS (2);
|
2985 |
|
|
else
|
2986 |
|
|
*total = 5;
|
2987 |
|
|
return true;
|
2988 |
|
|
|
2989 |
|
|
case CONST_DOUBLE:
|
2990 |
|
|
if (TARGET_SHMEDIA)
|
2991 |
|
|
*total = COSTS_N_INSNS (4);
|
2992 |
|
|
/* prepare_cmp_insn will force costly constants int registers before
|
2993 |
|
|
the cbranchdi4 pattern can see them, so preserve potentially
|
2994 |
|
|
interesting ones. */
|
2995 |
|
|
else if (outer_code == COMPARE && GET_MODE (x) == DImode)
|
2996 |
|
|
*total = 1;
|
2997 |
|
|
else
|
2998 |
|
|
*total = 10;
|
2999 |
|
|
return true;
|
3000 |
|
|
case CONST_VECTOR:
|
3001 |
|
|
if (x == CONST0_RTX (GET_MODE (x)))
|
3002 |
|
|
*total = 0;
|
3003 |
|
|
else if (sh_1el_vec (x, VOIDmode))
|
3004 |
|
|
*total = outer_code != SET;
|
3005 |
|
|
if (sh_rep_vec (x, VOIDmode))
|
3006 |
|
|
*total = ((GET_MODE_UNIT_SIZE (GET_MODE (x)) + 3) / 4
|
3007 |
|
|
+ (outer_code != SET));
|
3008 |
|
|
*total = COSTS_N_INSNS (3) + (outer_code != SET);
|
3009 |
|
|
return true;
|
3010 |
|
|
|
3011 |
|
|
case PLUS:
|
3012 |
|
|
case MINUS:
|
3013 |
|
|
*total = COSTS_N_INSNS (addsubcosts (x));
|
3014 |
|
|
return true;
|
3015 |
|
|
|
3016 |
|
|
case AND:
|
3017 |
|
|
*total = COSTS_N_INSNS (andcosts (x));
|
3018 |
|
|
return true;
|
3019 |
|
|
|
3020 |
|
|
case MULT:
|
3021 |
|
|
*total = COSTS_N_INSNS (multcosts (x));
|
3022 |
|
|
return true;
|
3023 |
|
|
|
3024 |
|
|
case ASHIFT:
|
3025 |
|
|
case ASHIFTRT:
|
3026 |
|
|
case LSHIFTRT:
|
3027 |
|
|
*total = COSTS_N_INSNS (shiftcosts (x));
|
3028 |
|
|
return true;
|
3029 |
|
|
|
3030 |
|
|
case DIV:
|
3031 |
|
|
case UDIV:
|
3032 |
|
|
case MOD:
|
3033 |
|
|
case UMOD:
|
3034 |
|
|
*total = COSTS_N_INSNS (20);
|
3035 |
|
|
return true;
|
3036 |
|
|
|
3037 |
|
|
case PARALLEL:
|
3038 |
|
|
if (sh_1el_vec (x, VOIDmode))
|
3039 |
|
|
*total = outer_code != SET;
|
3040 |
|
|
if (sh_rep_vec (x, VOIDmode))
|
3041 |
|
|
*total = ((GET_MODE_UNIT_SIZE (GET_MODE (x)) + 3) / 4
|
3042 |
|
|
+ (outer_code != SET));
|
3043 |
|
|
*total = COSTS_N_INSNS (3) + (outer_code != SET);
|
3044 |
|
|
return true;
|
3045 |
|
|
|
3046 |
|
|
case FLOAT:
|
3047 |
|
|
case FIX:
|
3048 |
|
|
*total = 100;
|
3049 |
|
|
return true;
|
3050 |
|
|
|
3051 |
|
|
default:
|
3052 |
|
|
return false;
|
3053 |
|
|
}
|
3054 |
|
|
}
|
3055 |
|
|
|
3056 |
|
|
/* Compute the cost of an address. For the SH, all valid addresses are
|
3057 |
|
|
the same cost. Use a slightly higher cost for reg + reg addressing,
|
3058 |
|
|
since it increases pressure on r0. */
|
3059 |
|
|
|
3060 |
|
|
static int
|
3061 |
|
|
sh_address_cost (rtx X,
|
3062 |
|
|
bool speed ATTRIBUTE_UNUSED)
|
3063 |
|
|
{
|
3064 |
|
|
return (GET_CODE (X) == PLUS
|
3065 |
|
|
&& ! CONSTANT_P (XEXP (X, 1))
|
3066 |
|
|
&& ! TARGET_SHMEDIA ? 1 : 0);
|
3067 |
|
|
}
|
3068 |
|
|
|
3069 |
|
|
/* Code to expand a shift. */
|
3070 |
|
|
|
3071 |
|
|
void
|
3072 |
|
|
gen_ashift (int type, int n, rtx reg)
|
3073 |
|
|
{
|
3074 |
|
|
/* Negative values here come from the shift_amounts array. */
|
3075 |
|
|
if (n < 0)
|
3076 |
|
|
{
|
3077 |
|
|
if (type == ASHIFT)
|
3078 |
|
|
type = LSHIFTRT;
|
3079 |
|
|
else
|
3080 |
|
|
type = ASHIFT;
|
3081 |
|
|
n = -n;
|
3082 |
|
|
}
|
3083 |
|
|
|
3084 |
|
|
switch (type)
|
3085 |
|
|
{
|
3086 |
|
|
case ASHIFTRT:
|
3087 |
|
|
emit_insn (gen_ashrsi3_k (reg, reg, GEN_INT (n)));
|
3088 |
|
|
break;
|
3089 |
|
|
case LSHIFTRT:
|
3090 |
|
|
if (n == 1)
|
3091 |
|
|
emit_insn (gen_lshrsi3_m (reg, reg, GEN_INT (n)));
|
3092 |
|
|
else
|
3093 |
|
|
emit_insn (gen_lshrsi3_k (reg, reg, GEN_INT (n)));
|
3094 |
|
|
break;
|
3095 |
|
|
case ASHIFT:
|
3096 |
|
|
emit_insn (gen_ashlsi3_std (reg, reg, GEN_INT (n)));
|
3097 |
|
|
break;
|
3098 |
|
|
}
|
3099 |
|
|
}
|
3100 |
|
|
|
3101 |
|
|
/* Same for HImode */
|
3102 |
|
|
|
3103 |
|
|
void
|
3104 |
|
|
gen_ashift_hi (int type, int n, rtx reg)
|
3105 |
|
|
{
|
3106 |
|
|
/* Negative values here come from the shift_amounts array. */
|
3107 |
|
|
if (n < 0)
|
3108 |
|
|
{
|
3109 |
|
|
if (type == ASHIFT)
|
3110 |
|
|
type = LSHIFTRT;
|
3111 |
|
|
else
|
3112 |
|
|
type = ASHIFT;
|
3113 |
|
|
n = -n;
|
3114 |
|
|
}
|
3115 |
|
|
|
3116 |
|
|
switch (type)
|
3117 |
|
|
{
|
3118 |
|
|
case ASHIFTRT:
|
3119 |
|
|
case LSHIFTRT:
|
3120 |
|
|
/* We don't have HImode right shift operations because using the
|
3121 |
|
|
ordinary 32 bit shift instructions for that doesn't generate proper
|
3122 |
|
|
zero/sign extension.
|
3123 |
|
|
gen_ashift_hi is only called in contexts where we know that the
|
3124 |
|
|
sign extension works out correctly. */
|
3125 |
|
|
{
|
3126 |
|
|
int offset = 0;
|
3127 |
|
|
if (GET_CODE (reg) == SUBREG)
|
3128 |
|
|
{
|
3129 |
|
|
offset = SUBREG_BYTE (reg);
|
3130 |
|
|
reg = SUBREG_REG (reg);
|
3131 |
|
|
}
|
3132 |
|
|
gen_ashift (type, n, gen_rtx_SUBREG (SImode, reg, offset));
|
3133 |
|
|
break;
|
3134 |
|
|
}
|
3135 |
|
|
case ASHIFT:
|
3136 |
|
|
emit_insn (gen_ashlhi3_k (reg, reg, GEN_INT (n)));
|
3137 |
|
|
break;
|
3138 |
|
|
}
|
3139 |
|
|
}
|
3140 |
|
|
|
3141 |
|
|
/* Output RTL to split a constant shift into its component SH constant
|
3142 |
|
|
shift instructions. */
|
3143 |
|
|
|
3144 |
|
|
void
|
3145 |
|
|
gen_shifty_op (int code, rtx *operands)
|
3146 |
|
|
{
|
3147 |
|
|
int value = INTVAL (operands[2]);
|
3148 |
|
|
int max, i;
|
3149 |
|
|
|
3150 |
|
|
/* Truncate the shift count in case it is out of bounds. */
|
3151 |
|
|
value = value & 31;
|
3152 |
|
|
|
3153 |
|
|
if (value == 31)
|
3154 |
|
|
{
|
3155 |
|
|
if (code == LSHIFTRT)
|
3156 |
|
|
{
|
3157 |
|
|
emit_insn (gen_rotlsi3_1 (operands[0], operands[0]));
|
3158 |
|
|
emit_insn (gen_movt (operands[0]));
|
3159 |
|
|
return;
|
3160 |
|
|
}
|
3161 |
|
|
else if (code == ASHIFT)
|
3162 |
|
|
{
|
3163 |
|
|
/* There is a two instruction sequence for 31 bit left shifts,
|
3164 |
|
|
but it requires r0. */
|
3165 |
|
|
if (REG_P (operands[0]) && REGNO (operands[0]) == 0)
|
3166 |
|
|
{
|
3167 |
|
|
emit_insn (gen_andsi3 (operands[0], operands[0], const1_rtx));
|
3168 |
|
|
emit_insn (gen_rotlsi3_31 (operands[0], operands[0]));
|
3169 |
|
|
return;
|
3170 |
|
|
}
|
3171 |
|
|
}
|
3172 |
|
|
}
|
3173 |
|
|
else if (value == 0)
|
3174 |
|
|
{
|
3175 |
|
|
/* This can happen even when optimizing, if there were subregs before
|
3176 |
|
|
reload. Don't output a nop here, as this is never optimized away;
|
3177 |
|
|
use a no-op move instead. */
|
3178 |
|
|
emit_insn (gen_rtx_SET (VOIDmode, operands[0], operands[0]));
|
3179 |
|
|
return;
|
3180 |
|
|
}
|
3181 |
|
|
|
3182 |
|
|
max = shift_insns[value];
|
3183 |
|
|
for (i = 0; i < max; i++)
|
3184 |
|
|
gen_ashift (code, shift_amounts[value][i], operands[0]);
|
3185 |
|
|
}
|
3186 |
|
|
|
3187 |
|
|
/* Same as above, but optimized for values where the topmost bits don't
|
3188 |
|
|
matter. */
|
3189 |
|
|
|
3190 |
|
|
void
|
3191 |
|
|
gen_shifty_hi_op (int code, rtx *operands)
|
3192 |
|
|
{
|
3193 |
|
|
int value = INTVAL (operands[2]);
|
3194 |
|
|
int max, i;
|
3195 |
|
|
void (*gen_fun) (int, int, rtx);
|
3196 |
|
|
|
3197 |
|
|
/* This operation is used by and_shl for SImode values with a few
|
3198 |
|
|
high bits known to be cleared. */
|
3199 |
|
|
value &= 31;
|
3200 |
|
|
if (value == 0)
|
3201 |
|
|
{
|
3202 |
|
|
emit_insn (gen_nop ());
|
3203 |
|
|
return;
|
3204 |
|
|
}
|
3205 |
|
|
|
3206 |
|
|
gen_fun = GET_MODE (operands[0]) == HImode ? gen_ashift_hi : gen_ashift;
|
3207 |
|
|
if (code == ASHIFT)
|
3208 |
|
|
{
|
3209 |
|
|
max = ext_shift_insns[value];
|
3210 |
|
|
for (i = 0; i < max; i++)
|
3211 |
|
|
gen_fun (code, ext_shift_amounts[value][i], operands[0]);
|
3212 |
|
|
}
|
3213 |
|
|
else
|
3214 |
|
|
/* When shifting right, emit the shifts in reverse order, so that
|
3215 |
|
|
solitary negative values come first. */
|
3216 |
|
|
for (i = ext_shift_insns[value] - 1; i >= 0; i--)
|
3217 |
|
|
gen_fun (code, ext_shift_amounts[value][i], operands[0]);
|
3218 |
|
|
}
|
3219 |
|
|
|
3220 |
|
|
/* Output RTL for an arithmetic right shift. */
|
3221 |
|
|
|
3222 |
|
|
/* ??? Rewrite to use super-optimizer sequences. */
|
3223 |
|
|
|
3224 |
|
|
int
|
3225 |
|
|
expand_ashiftrt (rtx *operands)
|
3226 |
|
|
{
|
3227 |
|
|
rtx wrk;
|
3228 |
|
|
char func[18];
|
3229 |
|
|
int value;
|
3230 |
|
|
|
3231 |
|
|
if (TARGET_SH3)
|
3232 |
|
|
{
|
3233 |
|
|
if (!CONST_INT_P (operands[2]))
|
3234 |
|
|
{
|
3235 |
|
|
rtx count = copy_to_mode_reg (SImode, operands[2]);
|
3236 |
|
|
emit_insn (gen_negsi2 (count, count));
|
3237 |
|
|
emit_insn (gen_ashrsi3_d (operands[0], operands[1], count));
|
3238 |
|
|
return 1;
|
3239 |
|
|
}
|
3240 |
|
|
else if (ashiftrt_insns[INTVAL (operands[2]) & 31]
|
3241 |
|
|
> 1 + SH_DYNAMIC_SHIFT_COST)
|
3242 |
|
|
{
|
3243 |
|
|
rtx count
|
3244 |
|
|
= force_reg (SImode, GEN_INT (- (INTVAL (operands[2]) & 31)));
|
3245 |
|
|
emit_insn (gen_ashrsi3_d (operands[0], operands[1], count));
|
3246 |
|
|
return 1;
|
3247 |
|
|
}
|
3248 |
|
|
}
|
3249 |
|
|
if (!CONST_INT_P (operands[2]))
|
3250 |
|
|
return 0;
|
3251 |
|
|
|
3252 |
|
|
value = INTVAL (operands[2]) & 31;
|
3253 |
|
|
|
3254 |
|
|
if (value == 31)
|
3255 |
|
|
{
|
3256 |
|
|
/* If we are called from abs expansion, arrange things so that we
|
3257 |
|
|
we can use a single MT instruction that doesn't clobber the source,
|
3258 |
|
|
if LICM can hoist out the load of the constant zero. */
|
3259 |
|
|
if (currently_expanding_to_rtl)
|
3260 |
|
|
{
|
3261 |
|
|
emit_insn (gen_cmpgtsi_t (force_reg (SImode, CONST0_RTX (SImode)),
|
3262 |
|
|
operands[1]));
|
3263 |
|
|
emit_insn (gen_mov_neg_si_t (operands[0]));
|
3264 |
|
|
return 1;
|
3265 |
|
|
}
|
3266 |
|
|
emit_insn (gen_ashrsi2_31 (operands[0], operands[1]));
|
3267 |
|
|
return 1;
|
3268 |
|
|
}
|
3269 |
|
|
else if (value >= 16 && value <= 19)
|
3270 |
|
|
{
|
3271 |
|
|
wrk = gen_reg_rtx (SImode);
|
3272 |
|
|
emit_insn (gen_ashrsi2_16 (wrk, operands[1]));
|
3273 |
|
|
value -= 16;
|
3274 |
|
|
while (value--)
|
3275 |
|
|
gen_ashift (ASHIFTRT, 1, wrk);
|
3276 |
|
|
emit_move_insn (operands[0], wrk);
|
3277 |
|
|
return 1;
|
3278 |
|
|
}
|
3279 |
|
|
/* Expand a short sequence inline, longer call a magic routine. */
|
3280 |
|
|
else if (value <= 5)
|
3281 |
|
|
{
|
3282 |
|
|
wrk = gen_reg_rtx (SImode);
|
3283 |
|
|
emit_move_insn (wrk, operands[1]);
|
3284 |
|
|
while (value--)
|
3285 |
|
|
gen_ashift (ASHIFTRT, 1, wrk);
|
3286 |
|
|
emit_move_insn (operands[0], wrk);
|
3287 |
|
|
return 1;
|
3288 |
|
|
}
|
3289 |
|
|
|
3290 |
|
|
wrk = gen_reg_rtx (Pmode);
|
3291 |
|
|
|
3292 |
|
|
/* Load the value into an arg reg and call a helper. */
|
3293 |
|
|
emit_move_insn (gen_rtx_REG (SImode, 4), operands[1]);
|
3294 |
|
|
sprintf (func, "__ashiftrt_r4_%d", value);
|
3295 |
|
|
function_symbol (wrk, func, SFUNC_STATIC);
|
3296 |
|
|
emit_insn (gen_ashrsi3_n (GEN_INT (value), wrk));
|
3297 |
|
|
emit_move_insn (operands[0], gen_rtx_REG (SImode, 4));
|
3298 |
|
|
return 1;
|
3299 |
|
|
}
|
3300 |
|
|
|
3301 |
|
|
int
|
3302 |
|
|
sh_dynamicalize_shift_p (rtx count)
|
3303 |
|
|
{
|
3304 |
|
|
return shift_insns[INTVAL (count) & 31] > 1 + SH_DYNAMIC_SHIFT_COST;
|
3305 |
|
|
}
|
3306 |
|
|
|
3307 |
|
|
/* Try to find a good way to implement the combiner pattern
|
3308 |
|
|
[(set (match_operand:SI 0 "register_operand" "r")
|
3309 |
|
|
(and:SI (ashift:SI (match_operand:SI 1 "register_operand" "r")
|
3310 |
|
|
(match_operand:SI 2 "const_int_operand" "n"))
|
3311 |
|
|
(match_operand:SI 3 "const_int_operand" "n"))) .
|
3312 |
|
|
LEFT_RTX is operand 2 in the above pattern, and MASK_RTX is operand 3.
|
3313 |
|
|
return 0 for simple right / left or left/right shift combination.
|
3314 |
|
|
return 1 for a combination of shifts with zero_extend.
|
3315 |
|
|
return 2 for a combination of shifts with an AND that needs r0.
|
3316 |
|
|
return 3 for a combination of shifts with an AND that needs an extra
|
3317 |
|
|
scratch register, when the three highmost bits of the AND mask are clear.
|
3318 |
|
|
return 4 for a combination of shifts with an AND that needs an extra
|
3319 |
|
|
scratch register, when any of the three highmost bits of the AND mask
|
3320 |
|
|
is set.
|
3321 |
|
|
If ATTRP is set, store an initial right shift width in ATTRP[0],
|
3322 |
|
|
and the instruction length in ATTRP[1] . These values are not valid
|
3323 |
|
|
when returning 0.
|
3324 |
|
|
When ATTRP is set and returning 1, ATTRP[2] gets set to the index into
|
3325 |
|
|
shift_amounts for the last shift value that is to be used before the
|
3326 |
|
|
sign extend. */
|
3327 |
|
|
int
|
3328 |
|
|
shl_and_kind (rtx left_rtx, rtx mask_rtx, int *attrp)
|
3329 |
|
|
{
|
3330 |
|
|
unsigned HOST_WIDE_INT mask, lsb, mask2, lsb2;
|
3331 |
|
|
int left = INTVAL (left_rtx), right;
|
3332 |
|
|
int best = 0;
|
3333 |
|
|
int cost, best_cost = 10000;
|
3334 |
|
|
int best_right = 0, best_len = 0;
|
3335 |
|
|
int i;
|
3336 |
|
|
int can_ext;
|
3337 |
|
|
|
3338 |
|
|
if (left < 0 || left > 31)
|
3339 |
|
|
return 0;
|
3340 |
|
|
if (CONST_INT_P (mask_rtx))
|
3341 |
|
|
mask = (unsigned HOST_WIDE_INT) INTVAL (mask_rtx) >> left;
|
3342 |
|
|
else
|
3343 |
|
|
mask = (unsigned HOST_WIDE_INT) GET_MODE_MASK (SImode) >> left;
|
3344 |
|
|
/* Can this be expressed as a right shift / left shift pair? */
|
3345 |
|
|
lsb = ((mask ^ (mask - 1)) >> 1) + 1;
|
3346 |
|
|
right = exact_log2 (lsb);
|
3347 |
|
|
mask2 = ~(mask + lsb - 1);
|
3348 |
|
|
lsb2 = ((mask2 ^ (mask2 - 1)) >> 1) + 1;
|
3349 |
|
|
/* mask has no zeroes but trailing zeroes <==> ! mask2 */
|
3350 |
|
|
if (! mask2)
|
3351 |
|
|
best_cost = shift_insns[right] + shift_insns[right + left];
|
3352 |
|
|
/* mask has no trailing zeroes <==> ! right */
|
3353 |
|
|
else if (! right && mask2 == ~(lsb2 - 1))
|
3354 |
|
|
{
|
3355 |
|
|
int late_right = exact_log2 (lsb2);
|
3356 |
|
|
best_cost = shift_insns[left + late_right] + shift_insns[late_right];
|
3357 |
|
|
}
|
3358 |
|
|
/* Try to use zero extend. */
|
3359 |
|
|
if (mask2 == ~(lsb2 - 1))
|
3360 |
|
|
{
|
3361 |
|
|
int width, first;
|
3362 |
|
|
|
3363 |
|
|
for (width = 8; width <= 16; width += 8)
|
3364 |
|
|
{
|
3365 |
|
|
/* Can we zero-extend right away? */
|
3366 |
|
|
if (lsb2 == (unsigned HOST_WIDE_INT) 1 << width)
|
3367 |
|
|
{
|
3368 |
|
|
cost
|
3369 |
|
|
= 1 + ext_shift_insns[right] + ext_shift_insns[left + right];
|
3370 |
|
|
if (cost < best_cost)
|
3371 |
|
|
{
|
3372 |
|
|
best = 1;
|
3373 |
|
|
best_cost = cost;
|
3374 |
|
|
best_right = right;
|
3375 |
|
|
best_len = cost;
|
3376 |
|
|
if (attrp)
|
3377 |
|
|
attrp[2] = -1;
|
3378 |
|
|
}
|
3379 |
|
|
continue;
|
3380 |
|
|
}
|
3381 |
|
|
/* ??? Could try to put zero extend into initial right shift,
|
3382 |
|
|
or even shift a bit left before the right shift. */
|
3383 |
|
|
/* Determine value of first part of left shift, to get to the
|
3384 |
|
|
zero extend cut-off point. */
|
3385 |
|
|
first = width - exact_log2 (lsb2) + right;
|
3386 |
|
|
if (first >= 0 && right + left - first >= 0)
|
3387 |
|
|
{
|
3388 |
|
|
cost = ext_shift_insns[right] + ext_shift_insns[first] + 1
|
3389 |
|
|
+ ext_shift_insns[right + left - first];
|
3390 |
|
|
if (cost < best_cost)
|
3391 |
|
|
{
|
3392 |
|
|
best = 1;
|
3393 |
|
|
best_cost = cost;
|
3394 |
|
|
best_right = right;
|
3395 |
|
|
best_len = cost;
|
3396 |
|
|
if (attrp)
|
3397 |
|
|
attrp[2] = first;
|
3398 |
|
|
}
|
3399 |
|
|
}
|
3400 |
|
|
}
|
3401 |
|
|
}
|
3402 |
|
|
/* Try to use r0 AND pattern */
|
3403 |
|
|
for (i = 0; i <= 2; i++)
|
3404 |
|
|
{
|
3405 |
|
|
if (i > right)
|
3406 |
|
|
break;
|
3407 |
|
|
if (! CONST_OK_FOR_K08 (mask >> i))
|
3408 |
|
|
continue;
|
3409 |
|
|
cost = (i != 0) + 2 + ext_shift_insns[left + i];
|
3410 |
|
|
if (cost < best_cost)
|
3411 |
|
|
{
|
3412 |
|
|
best = 2;
|
3413 |
|
|
best_cost = cost;
|
3414 |
|
|
best_right = i;
|
3415 |
|
|
best_len = cost - 1;
|
3416 |
|
|
}
|
3417 |
|
|
}
|
3418 |
|
|
/* Try to use a scratch register to hold the AND operand. */
|
3419 |
|
|
can_ext = ((mask << left) & ((unsigned HOST_WIDE_INT) 3 << 30)) == 0;
|
3420 |
|
|
for (i = 0; i <= 2; i++)
|
3421 |
|
|
{
|
3422 |
|
|
if (i > right)
|
3423 |
|
|
break;
|
3424 |
|
|
cost = (i != 0) + (CONST_OK_FOR_I08 (mask >> i) ? 2 : 3)
|
3425 |
|
|
+ (can_ext ? ext_shift_insns : shift_insns)[left + i];
|
3426 |
|
|
if (cost < best_cost)
|
3427 |
|
|
{
|
3428 |
|
|
best = 4 - can_ext;
|
3429 |
|
|
best_cost = cost;
|
3430 |
|
|
best_right = i;
|
3431 |
|
|
best_len = cost - 1 - ! CONST_OK_FOR_I08 (mask >> i);
|
3432 |
|
|
}
|
3433 |
|
|
}
|
3434 |
|
|
|
3435 |
|
|
if (attrp)
|
3436 |
|
|
{
|
3437 |
|
|
attrp[0] = best_right;
|
3438 |
|
|
attrp[1] = best_len;
|
3439 |
|
|
}
|
3440 |
|
|
return best;
|
3441 |
|
|
}
|
3442 |
|
|
|
3443 |
|
|
/* This is used in length attributes of the unnamed instructions
|
3444 |
|
|
corresponding to shl_and_kind return values of 1 and 2. */
|
3445 |
|
|
int
|
3446 |
|
|
shl_and_length (rtx insn)
|
3447 |
|
|
{
|
3448 |
|
|
rtx set_src, left_rtx, mask_rtx;
|
3449 |
|
|
int attributes[3];
|
3450 |
|
|
|
3451 |
|
|
set_src = SET_SRC (XVECEXP (PATTERN (insn), 0, 0));
|
3452 |
|
|
left_rtx = XEXP (XEXP (set_src, 0), 1);
|
3453 |
|
|
mask_rtx = XEXP (set_src, 1);
|
3454 |
|
|
shl_and_kind (left_rtx, mask_rtx, attributes);
|
3455 |
|
|
return attributes[1];
|
3456 |
|
|
}
|
3457 |
|
|
|
3458 |
|
|
/* This is used in length attribute of the and_shl_scratch instruction. */
|
3459 |
|
|
|
3460 |
|
|
int
|
3461 |
|
|
shl_and_scr_length (rtx insn)
|
3462 |
|
|
{
|
3463 |
|
|
rtx set_src = SET_SRC (XVECEXP (PATTERN (insn), 0, 0));
|
3464 |
|
|
int len = shift_insns[INTVAL (XEXP (set_src, 1)) & 31];
|
3465 |
|
|
rtx op = XEXP (set_src, 0);
|
3466 |
|
|
len += shift_insns[INTVAL (XEXP (op, 1)) & 31] + 1;
|
3467 |
|
|
op = XEXP (XEXP (op, 0), 0);
|
3468 |
|
|
return len + shift_insns[INTVAL (XEXP (op, 1)) & 31];
|
3469 |
|
|
}
|
3470 |
|
|
|
3471 |
|
|
/* Generate rtl for instructions for which shl_and_kind advised a particular
|
3472 |
|
|
method of generating them, i.e. returned zero. */
|
3473 |
|
|
|
3474 |
|
|
int
|
3475 |
|
|
gen_shl_and (rtx dest, rtx left_rtx, rtx mask_rtx, rtx source)
|
3476 |
|
|
{
|
3477 |
|
|
int attributes[3];
|
3478 |
|
|
unsigned HOST_WIDE_INT mask;
|
3479 |
|
|
int kind = shl_and_kind (left_rtx, mask_rtx, attributes);
|
3480 |
|
|
int right, total_shift;
|
3481 |
|
|
void (*shift_gen_fun) (int, rtx *) = gen_shifty_hi_op;
|
3482 |
|
|
|
3483 |
|
|
right = attributes[0];
|
3484 |
|
|
total_shift = INTVAL (left_rtx) + right;
|
3485 |
|
|
mask = (unsigned HOST_WIDE_INT) INTVAL (mask_rtx) >> total_shift;
|
3486 |
|
|
switch (kind)
|
3487 |
|
|
{
|
3488 |
|
|
default:
|
3489 |
|
|
return -1;
|
3490 |
|
|
case 1:
|
3491 |
|
|
{
|
3492 |
|
|
int first = attributes[2];
|
3493 |
|
|
rtx operands[3];
|
3494 |
|
|
|
3495 |
|
|
if (first < 0)
|
3496 |
|
|
{
|
3497 |
|
|
emit_insn ((mask << right) <= 0xff
|
3498 |
|
|
? gen_zero_extendqisi2 (dest,
|
3499 |
|
|
gen_lowpart (QImode, source))
|
3500 |
|
|
: gen_zero_extendhisi2 (dest,
|
3501 |
|
|
gen_lowpart (HImode, source)));
|
3502 |
|
|
source = dest;
|
3503 |
|
|
}
|
3504 |
|
|
if (source != dest)
|
3505 |
|
|
emit_insn (gen_movsi (dest, source));
|
3506 |
|
|
operands[0] = dest;
|
3507 |
|
|
if (right)
|
3508 |
|
|
{
|
3509 |
|
|
operands[2] = GEN_INT (right);
|
3510 |
|
|
gen_shifty_hi_op (LSHIFTRT, operands);
|
3511 |
|
|
}
|
3512 |
|
|
if (first > 0)
|
3513 |
|
|
{
|
3514 |
|
|
operands[2] = GEN_INT (first);
|
3515 |
|
|
gen_shifty_hi_op (ASHIFT, operands);
|
3516 |
|
|
total_shift -= first;
|
3517 |
|
|
mask <<= first;
|
3518 |
|
|
}
|
3519 |
|
|
if (first >= 0)
|
3520 |
|
|
emit_insn (mask <= 0xff
|
3521 |
|
|
? gen_zero_extendqisi2 (dest, gen_lowpart (QImode, dest))
|
3522 |
|
|
: gen_zero_extendhisi2 (dest, gen_lowpart (HImode, dest)));
|
3523 |
|
|
if (total_shift > 0)
|
3524 |
|
|
{
|
3525 |
|
|
operands[2] = GEN_INT (total_shift);
|
3526 |
|
|
gen_shifty_hi_op (ASHIFT, operands);
|
3527 |
|
|
}
|
3528 |
|
|
break;
|
3529 |
|
|
}
|
3530 |
|
|
case 4:
|
3531 |
|
|
shift_gen_fun = gen_shifty_op;
|
3532 |
|
|
case 3:
|
3533 |
|
|
/* If the topmost bit that matters is set, set the topmost bits
|
3534 |
|
|
that don't matter. This way, we might be able to get a shorter
|
3535 |
|
|
signed constant. */
|
3536 |
|
|
if (mask & ((HOST_WIDE_INT) 1 << (31 - total_shift)))
|
3537 |
|
|
mask |= (HOST_WIDE_INT) ~0 << (31 - total_shift);
|
3538 |
|
|
case 2:
|
3539 |
|
|
/* Don't expand fine-grained when combining, because that will
|
3540 |
|
|
make the pattern fail. */
|
3541 |
|
|
if (currently_expanding_to_rtl
|
3542 |
|
|
|| reload_in_progress || reload_completed)
|
3543 |
|
|
{
|
3544 |
|
|
rtx operands[3];
|
3545 |
|
|
|
3546 |
|
|
/* Cases 3 and 4 should be handled by this split
|
3547 |
|
|
only while combining */
|
3548 |
|
|
gcc_assert (kind <= 2);
|
3549 |
|
|
if (right)
|
3550 |
|
|
{
|
3551 |
|
|
emit_insn (gen_lshrsi3 (dest, source, GEN_INT (right)));
|
3552 |
|
|
source = dest;
|
3553 |
|
|
}
|
3554 |
|
|
emit_insn (gen_andsi3 (dest, source, GEN_INT (mask)));
|
3555 |
|
|
if (total_shift)
|
3556 |
|
|
{
|
3557 |
|
|
operands[0] = dest;
|
3558 |
|
|
operands[1] = dest;
|
3559 |
|
|
operands[2] = GEN_INT (total_shift);
|
3560 |
|
|
shift_gen_fun (ASHIFT, operands);
|
3561 |
|
|
}
|
3562 |
|
|
break;
|
3563 |
|
|
}
|
3564 |
|
|
else
|
3565 |
|
|
{
|
3566 |
|
|
int neg = 0;
|
3567 |
|
|
if (kind != 4 && total_shift < 16)
|
3568 |
|
|
{
|
3569 |
|
|
neg = -ext_shift_amounts[total_shift][1];
|
3570 |
|
|
if (neg > 0)
|
3571 |
|
|
neg -= ext_shift_amounts[total_shift][2];
|
3572 |
|
|
else
|
3573 |
|
|
neg = 0;
|
3574 |
|
|
}
|
3575 |
|
|
emit_insn (gen_and_shl_scratch (dest, source,
|
3576 |
|
|
GEN_INT (right),
|
3577 |
|
|
GEN_INT (mask),
|
3578 |
|
|
GEN_INT (total_shift + neg),
|
3579 |
|
|
GEN_INT (neg)));
|
3580 |
|
|
emit_insn (gen_movsi (dest, dest));
|
3581 |
|
|
break;
|
3582 |
|
|
}
|
3583 |
|
|
}
|
3584 |
|
|
return 0;
|
3585 |
|
|
}
|
3586 |
|
|
|
3587 |
|
|
/* Try to find a good way to implement the combiner pattern
|
3588 |
|
|
[(set (match_operand:SI 0 "register_operand" "=r")
|
3589 |
|
|
(sign_extract:SI (ashift:SI (match_operand:SI 1 "register_operand" "r")
|
3590 |
|
|
(match_operand:SI 2 "const_int_operand" "n")
|
3591 |
|
|
(match_operand:SI 3 "const_int_operand" "n")
|
3592 |
|
|
(const_int 0)))
|
3593 |
|
|
(clobber (reg:SI T_REG))]
|
3594 |
|
|
LEFT_RTX is operand 2 in the above pattern, and SIZE_RTX is operand 3.
|
3595 |
|
|
return 0 for simple left / right shift combination.
|
3596 |
|
|
return 1 for left shift / 8 bit sign extend / left shift.
|
3597 |
|
|
return 2 for left shift / 16 bit sign extend / left shift.
|
3598 |
|
|
return 3 for left shift / 8 bit sign extend / shift / sign extend.
|
3599 |
|
|
return 4 for left shift / 16 bit sign extend / shift / sign extend.
|
3600 |
|
|
return 5 for left shift / 16 bit sign extend / right shift
|
3601 |
|
|
return 6 for < 8 bit sign extend / left shift.
|
3602 |
|
|
return 7 for < 8 bit sign extend / left shift / single right shift.
|
3603 |
|
|
If COSTP is nonzero, assign the calculated cost to *COSTP. */
|
3604 |
|
|
|
3605 |
|
|
int
|
3606 |
|
|
shl_sext_kind (rtx left_rtx, rtx size_rtx, int *costp)
|
3607 |
|
|
{
|
3608 |
|
|
int left, size, insize, ext;
|
3609 |
|
|
int cost = 0, best_cost;
|
3610 |
|
|
int kind;
|
3611 |
|
|
|
3612 |
|
|
left = INTVAL (left_rtx);
|
3613 |
|
|
size = INTVAL (size_rtx);
|
3614 |
|
|
insize = size - left;
|
3615 |
|
|
gcc_assert (insize > 0);
|
3616 |
|
|
/* Default to left / right shift. */
|
3617 |
|
|
kind = 0;
|
3618 |
|
|
best_cost = shift_insns[32 - insize] + ashiftrt_insns[32 - size];
|
3619 |
|
|
if (size <= 16)
|
3620 |
|
|
{
|
3621 |
|
|
/* 16 bit shift / sign extend / 16 bit shift */
|
3622 |
|
|
cost = shift_insns[16 - insize] + 1 + ashiftrt_insns[16 - size];
|
3623 |
|
|
/* If ashiftrt_insns[16 - size] is 8, this choice will be overridden
|
3624 |
|
|
below, by alternative 3 or something even better. */
|
3625 |
|
|
if (cost < best_cost)
|
3626 |
|
|
{
|
3627 |
|
|
kind = 5;
|
3628 |
|
|
best_cost = cost;
|
3629 |
|
|
}
|
3630 |
|
|
}
|
3631 |
|
|
/* Try a plain sign extend between two shifts. */
|
3632 |
|
|
for (ext = 16; ext >= insize; ext -= 8)
|
3633 |
|
|
{
|
3634 |
|
|
if (ext <= size)
|
3635 |
|
|
{
|
3636 |
|
|
cost = ext_shift_insns[ext - insize] + 1 + shift_insns[size - ext];
|
3637 |
|
|
if (cost < best_cost)
|
3638 |
|
|
{
|
3639 |
|
|
kind = ext / (unsigned) 8;
|
3640 |
|
|
best_cost = cost;
|
3641 |
|
|
}
|
3642 |
|
|
}
|
3643 |
|
|
/* Check if we can do a sloppy shift with a final signed shift
|
3644 |
|
|
restoring the sign. */
|
3645 |
|
|
if (EXT_SHIFT_SIGNED (size - ext))
|
3646 |
|
|
cost = ext_shift_insns[ext - insize] + ext_shift_insns[size - ext] + 1;
|
3647 |
|
|
/* If not, maybe it's still cheaper to do the second shift sloppy,
|
3648 |
|
|
and do a final sign extend? */
|
3649 |
|
|
else if (size <= 16)
|
3650 |
|
|
cost = ext_shift_insns[ext - insize] + 1
|
3651 |
|
|
+ ext_shift_insns[size > ext ? size - ext : ext - size] + 1;
|
3652 |
|
|
else
|
3653 |
|
|
continue;
|
3654 |
|
|
if (cost < best_cost)
|
3655 |
|
|
{
|
3656 |
|
|
kind = ext / (unsigned) 8 + 2;
|
3657 |
|
|
best_cost = cost;
|
3658 |
|
|
}
|
3659 |
|
|
}
|
3660 |
|
|
/* Check if we can sign extend in r0 */
|
3661 |
|
|
if (insize < 8)
|
3662 |
|
|
{
|
3663 |
|
|
cost = 3 + shift_insns[left];
|
3664 |
|
|
if (cost < best_cost)
|
3665 |
|
|
{
|
3666 |
|
|
kind = 6;
|
3667 |
|
|
best_cost = cost;
|
3668 |
|
|
}
|
3669 |
|
|
/* Try the same with a final signed shift. */
|
3670 |
|
|
if (left < 31)
|
3671 |
|
|
{
|
3672 |
|
|
cost = 3 + ext_shift_insns[left + 1] + 1;
|
3673 |
|
|
if (cost < best_cost)
|
3674 |
|
|
{
|
3675 |
|
|
kind = 7;
|
3676 |
|
|
best_cost = cost;
|
3677 |
|
|
}
|
3678 |
|
|
}
|
3679 |
|
|
}
|
3680 |
|
|
if (TARGET_SH3)
|
3681 |
|
|
{
|
3682 |
|
|
/* Try to use a dynamic shift. */
|
3683 |
|
|
cost = shift_insns[32 - insize] + 1 + SH_DYNAMIC_SHIFT_COST;
|
3684 |
|
|
if (cost < best_cost)
|
3685 |
|
|
{
|
3686 |
|
|
kind = 0;
|
3687 |
|
|
best_cost = cost;
|
3688 |
|
|
}
|
3689 |
|
|
}
|
3690 |
|
|
if (costp)
|
3691 |
|
|
*costp = cost;
|
3692 |
|
|
return kind;
|
3693 |
|
|
}
|
3694 |
|
|
|
3695 |
|
|
/* Function to be used in the length attribute of the instructions
|
3696 |
|
|
implementing this pattern. */
|
3697 |
|
|
|
3698 |
|
|
int
|
3699 |
|
|
shl_sext_length (rtx insn)
|
3700 |
|
|
{
|
3701 |
|
|
rtx set_src, left_rtx, size_rtx;
|
3702 |
|
|
int cost;
|
3703 |
|
|
|
3704 |
|
|
set_src = SET_SRC (XVECEXP (PATTERN (insn), 0, 0));
|
3705 |
|
|
left_rtx = XEXP (XEXP (set_src, 0), 1);
|
3706 |
|
|
size_rtx = XEXP (set_src, 1);
|
3707 |
|
|
shl_sext_kind (left_rtx, size_rtx, &cost);
|
3708 |
|
|
return cost;
|
3709 |
|
|
}
|
3710 |
|
|
|
3711 |
|
|
/* Generate rtl for this pattern */
|
3712 |
|
|
|
3713 |
|
|
int
|
3714 |
|
|
gen_shl_sext (rtx dest, rtx left_rtx, rtx size_rtx, rtx source)
|
3715 |
|
|
{
|
3716 |
|
|
int kind;
|
3717 |
|
|
int left, size, insize, cost;
|
3718 |
|
|
rtx operands[3];
|
3719 |
|
|
|
3720 |
|
|
kind = shl_sext_kind (left_rtx, size_rtx, &cost);
|
3721 |
|
|
left = INTVAL (left_rtx);
|
3722 |
|
|
size = INTVAL (size_rtx);
|
3723 |
|
|
insize = size - left;
|
3724 |
|
|
switch (kind)
|
3725 |
|
|
{
|
3726 |
|
|
case 1:
|
3727 |
|
|
case 2:
|
3728 |
|
|
case 3:
|
3729 |
|
|
case 4:
|
3730 |
|
|
{
|
3731 |
|
|
int ext = kind & 1 ? 8 : 16;
|
3732 |
|
|
int shift2 = size - ext;
|
3733 |
|
|
|
3734 |
|
|
/* Don't expand fine-grained when combining, because that will
|
3735 |
|
|
make the pattern fail. */
|
3736 |
|
|
if (! currently_expanding_to_rtl
|
3737 |
|
|
&& ! reload_in_progress && ! reload_completed)
|
3738 |
|
|
{
|
3739 |
|
|
emit_insn (gen_shl_sext_ext (dest, source, left_rtx, size_rtx));
|
3740 |
|
|
emit_insn (gen_movsi (dest, source));
|
3741 |
|
|
break;
|
3742 |
|
|
}
|
3743 |
|
|
if (dest != source)
|
3744 |
|
|
emit_insn (gen_movsi (dest, source));
|
3745 |
|
|
operands[0] = dest;
|
3746 |
|
|
if (ext - insize)
|
3747 |
|
|
{
|
3748 |
|
|
operands[2] = GEN_INT (ext - insize);
|
3749 |
|
|
gen_shifty_hi_op (ASHIFT, operands);
|
3750 |
|
|
}
|
3751 |
|
|
emit_insn (kind & 1
|
3752 |
|
|
? gen_extendqisi2 (dest, gen_lowpart (QImode, dest))
|
3753 |
|
|
: gen_extendhisi2 (dest, gen_lowpart (HImode, dest)));
|
3754 |
|
|
if (kind <= 2)
|
3755 |
|
|
{
|
3756 |
|
|
if (shift2)
|
3757 |
|
|
{
|
3758 |
|
|
operands[2] = GEN_INT (shift2);
|
3759 |
|
|
gen_shifty_op (ASHIFT, operands);
|
3760 |
|
|
}
|
3761 |
|
|
}
|
3762 |
|
|
else
|
3763 |
|
|
{
|
3764 |
|
|
if (shift2 > 0)
|
3765 |
|
|
{
|
3766 |
|
|
if (EXT_SHIFT_SIGNED (shift2))
|
3767 |
|
|
{
|
3768 |
|
|
operands[2] = GEN_INT (shift2 + 1);
|
3769 |
|
|
gen_shifty_op (ASHIFT, operands);
|
3770 |
|
|
operands[2] = const1_rtx;
|
3771 |
|
|
gen_shifty_op (ASHIFTRT, operands);
|
3772 |
|
|
break;
|
3773 |
|
|
}
|
3774 |
|
|
operands[2] = GEN_INT (shift2);
|
3775 |
|
|
gen_shifty_hi_op (ASHIFT, operands);
|
3776 |
|
|
}
|
3777 |
|
|
else if (shift2)
|
3778 |
|
|
{
|
3779 |
|
|
operands[2] = GEN_INT (-shift2);
|
3780 |
|
|
gen_shifty_hi_op (LSHIFTRT, operands);
|
3781 |
|
|
}
|
3782 |
|
|
emit_insn (size <= 8
|
3783 |
|
|
? gen_extendqisi2 (dest, gen_lowpart (QImode, dest))
|
3784 |
|
|
: gen_extendhisi2 (dest, gen_lowpart (HImode, dest)));
|
3785 |
|
|
}
|
3786 |
|
|
break;
|
3787 |
|
|
}
|
3788 |
|
|
case 5:
|
3789 |
|
|
{
|
3790 |
|
|
int i = 16 - size;
|
3791 |
|
|
if (! currently_expanding_to_rtl
|
3792 |
|
|
&& ! reload_in_progress && ! reload_completed)
|
3793 |
|
|
emit_insn (gen_shl_sext_ext (dest, source, left_rtx, size_rtx));
|
3794 |
|
|
else
|
3795 |
|
|
{
|
3796 |
|
|
operands[0] = dest;
|
3797 |
|
|
operands[2] = GEN_INT (16 - insize);
|
3798 |
|
|
gen_shifty_hi_op (ASHIFT, operands);
|
3799 |
|
|
emit_insn (gen_extendhisi2 (dest, gen_lowpart (HImode, dest)));
|
3800 |
|
|
}
|
3801 |
|
|
/* Don't use gen_ashrsi3 because it generates new pseudos. */
|
3802 |
|
|
while (--i >= 0)
|
3803 |
|
|
gen_ashift (ASHIFTRT, 1, dest);
|
3804 |
|
|
break;
|
3805 |
|
|
}
|
3806 |
|
|
case 6:
|
3807 |
|
|
case 7:
|
3808 |
|
|
/* Don't expand fine-grained when combining, because that will
|
3809 |
|
|
make the pattern fail. */
|
3810 |
|
|
if (! currently_expanding_to_rtl
|
3811 |
|
|
&& ! reload_in_progress && ! reload_completed)
|
3812 |
|
|
{
|
3813 |
|
|
emit_insn (gen_shl_sext_ext (dest, source, left_rtx, size_rtx));
|
3814 |
|
|
emit_insn (gen_movsi (dest, source));
|
3815 |
|
|
break;
|
3816 |
|
|
}
|
3817 |
|
|
emit_insn (gen_andsi3 (dest, source, GEN_INT ((1 << insize) - 1)));
|
3818 |
|
|
emit_insn (gen_xorsi3 (dest, dest, GEN_INT (1 << (insize - 1))));
|
3819 |
|
|
emit_insn (gen_addsi3 (dest, dest, GEN_INT (-1 << (insize - 1))));
|
3820 |
|
|
operands[0] = dest;
|
3821 |
|
|
operands[2] = kind == 7 ? GEN_INT (left + 1) : left_rtx;
|
3822 |
|
|
gen_shifty_op (ASHIFT, operands);
|
3823 |
|
|
if (kind == 7)
|
3824 |
|
|
emit_insn (gen_ashrsi3_k (dest, dest, const1_rtx));
|
3825 |
|
|
break;
|
3826 |
|
|
default:
|
3827 |
|
|
return -1;
|
3828 |
|
|
}
|
3829 |
|
|
return 0;
|
3830 |
|
|
}
|
3831 |
|
|
|
3832 |
|
|
/* Prefix a symbol_ref name with "datalabel". */
|
3833 |
|
|
|
3834 |
|
|
rtx
|
3835 |
|
|
gen_datalabel_ref (rtx sym)
|
3836 |
|
|
{
|
3837 |
|
|
const char *str;
|
3838 |
|
|
|
3839 |
|
|
if (GET_CODE (sym) == LABEL_REF)
|
3840 |
|
|
return gen_rtx_CONST (GET_MODE (sym),
|
3841 |
|
|
gen_rtx_UNSPEC (GET_MODE (sym),
|
3842 |
|
|
gen_rtvec (1, sym),
|
3843 |
|
|
UNSPEC_DATALABEL));
|
3844 |
|
|
|
3845 |
|
|
gcc_assert (GET_CODE (sym) == SYMBOL_REF);
|
3846 |
|
|
|
3847 |
|
|
str = XSTR (sym, 0);
|
3848 |
|
|
/* Share all SYMBOL_REF strings with the same value - that is important
|
3849 |
|
|
for cse. */
|
3850 |
|
|
str = IDENTIFIER_POINTER (get_identifier (str));
|
3851 |
|
|
XSTR (sym, 0) = str;
|
3852 |
|
|
|
3853 |
|
|
return sym;
|
3854 |
|
|
}
|
3855 |
|
|
|
3856 |
|
|
|
3857 |
|
|
static alloc_pool label_ref_list_pool;
|
3858 |
|
|
|
3859 |
|
|
typedef struct label_ref_list_d
|
3860 |
|
|
{
|
3861 |
|
|
rtx label;
|
3862 |
|
|
struct label_ref_list_d *next;
|
3863 |
|
|
} *label_ref_list_t;
|
3864 |
|
|
|
3865 |
|
|
/* The SH cannot load a large constant into a register, constants have to
|
3866 |
|
|
come from a pc relative load. The reference of a pc relative load
|
3867 |
|
|
instruction must be less than 1k in front of the instruction. This
|
3868 |
|
|
means that we often have to dump a constant inside a function, and
|
3869 |
|
|
generate code to branch around it.
|
3870 |
|
|
|
3871 |
|
|
It is important to minimize this, since the branches will slow things
|
3872 |
|
|
down and make things bigger.
|
3873 |
|
|
|
3874 |
|
|
Worst case code looks like:
|
3875 |
|
|
|
3876 |
|
|
mov.l L1,rn
|
3877 |
|
|
bra L2
|
3878 |
|
|
nop
|
3879 |
|
|
align
|
3880 |
|
|
L1: .long value
|
3881 |
|
|
L2:
|
3882 |
|
|
..
|
3883 |
|
|
|
3884 |
|
|
mov.l L3,rn
|
3885 |
|
|
bra L4
|
3886 |
|
|
nop
|
3887 |
|
|
align
|
3888 |
|
|
L3: .long value
|
3889 |
|
|
L4:
|
3890 |
|
|
..
|
3891 |
|
|
|
3892 |
|
|
We fix this by performing a scan before scheduling, which notices which
|
3893 |
|
|
instructions need to have their operands fetched from the constant table
|
3894 |
|
|
and builds the table.
|
3895 |
|
|
|
3896 |
|
|
The algorithm is:
|
3897 |
|
|
|
3898 |
|
|
scan, find an instruction which needs a pcrel move. Look forward, find the
|
3899 |
|
|
last barrier which is within MAX_COUNT bytes of the requirement.
|
3900 |
|
|
If there isn't one, make one. Process all the instructions between
|
3901 |
|
|
the find and the barrier.
|
3902 |
|
|
|
3903 |
|
|
In the above example, we can tell that L3 is within 1k of L1, so
|
3904 |
|
|
the first move can be shrunk from the 3 insn+constant sequence into
|
3905 |
|
|
just 1 insn, and the constant moved to L3 to make:
|
3906 |
|
|
|
3907 |
|
|
mov.l L1,rn
|
3908 |
|
|
..
|
3909 |
|
|
mov.l L3,rn
|
3910 |
|
|
bra L4
|
3911 |
|
|
nop
|
3912 |
|
|
align
|
3913 |
|
|
L3:.long value
|
3914 |
|
|
L4:.long value
|
3915 |
|
|
|
3916 |
|
|
Then the second move becomes the target for the shortening process. */
|
3917 |
|
|
|
3918 |
|
|
typedef struct
|
3919 |
|
|
{
|
3920 |
|
|
rtx value; /* Value in table. */
|
3921 |
|
|
rtx label; /* Label of value. */
|
3922 |
|
|
label_ref_list_t wend; /* End of window. */
|
3923 |
|
|
enum machine_mode mode; /* Mode of value. */
|
3924 |
|
|
|
3925 |
|
|
/* True if this constant is accessed as part of a post-increment
|
3926 |
|
|
sequence. Note that HImode constants are never accessed in this way. */
|
3927 |
|
|
bool part_of_sequence_p;
|
3928 |
|
|
} pool_node;
|
3929 |
|
|
|
3930 |
|
|
/* The maximum number of constants that can fit into one pool, since
|
3931 |
|
|
constants in the range 0..510 are at least 2 bytes long, and in the
|
3932 |
|
|
range from there to 1018 at least 4 bytes. */
|
3933 |
|
|
|
3934 |
|
|
#define MAX_POOL_SIZE 372
|
3935 |
|
|
static pool_node pool_vector[MAX_POOL_SIZE];
|
3936 |
|
|
static int pool_size;
|
3937 |
|
|
static rtx pool_window_label;
|
3938 |
|
|
static int pool_window_last;
|
3939 |
|
|
|
3940 |
|
|
static int max_labelno_before_reorg;
|
3941 |
|
|
|
3942 |
|
|
/* ??? If we need a constant in HImode which is the truncated value of a
|
3943 |
|
|
constant we need in SImode, we could combine the two entries thus saving
|
3944 |
|
|
two bytes. Is this common enough to be worth the effort of implementing
|
3945 |
|
|
it? */
|
3946 |
|
|
|
3947 |
|
|
/* ??? This stuff should be done at the same time that we shorten branches.
|
3948 |
|
|
As it is now, we must assume that all branches are the maximum size, and
|
3949 |
|
|
this causes us to almost always output constant pools sooner than
|
3950 |
|
|
necessary. */
|
3951 |
|
|
|
3952 |
|
|
/* Add a constant to the pool and return its label. */
|
3953 |
|
|
|
3954 |
|
|
static rtx
|
3955 |
|
|
add_constant (rtx x, enum machine_mode mode, rtx last_value)
|
3956 |
|
|
{
|
3957 |
|
|
int i;
|
3958 |
|
|
rtx lab, new_rtx;
|
3959 |
|
|
label_ref_list_t ref, newref;
|
3960 |
|
|
|
3961 |
|
|
/* First see if we've already got it. */
|
3962 |
|
|
for (i = 0; i < pool_size; i++)
|
3963 |
|
|
{
|
3964 |
|
|
if (x->code == pool_vector[i].value->code
|
3965 |
|
|
&& mode == pool_vector[i].mode)
|
3966 |
|
|
{
|
3967 |
|
|
if (x->code == CODE_LABEL)
|
3968 |
|
|
{
|
3969 |
|
|
if (XINT (x, 3) != XINT (pool_vector[i].value, 3))
|
3970 |
|
|
continue;
|
3971 |
|
|
}
|
3972 |
|
|
if (rtx_equal_p (x, pool_vector[i].value))
|
3973 |
|
|
{
|
3974 |
|
|
lab = new_rtx = 0;
|
3975 |
|
|
if (! last_value
|
3976 |
|
|
|| ! i
|
3977 |
|
|
|| ! rtx_equal_p (last_value, pool_vector[i-1].value))
|
3978 |
|
|
{
|
3979 |
|
|
new_rtx = gen_label_rtx ();
|
3980 |
|
|
LABEL_REFS (new_rtx) = pool_vector[i].label;
|
3981 |
|
|
pool_vector[i].label = lab = new_rtx;
|
3982 |
|
|
}
|
3983 |
|
|
if (lab && pool_window_label)
|
3984 |
|
|
{
|
3985 |
|
|
newref = (label_ref_list_t) pool_alloc (label_ref_list_pool);
|
3986 |
|
|
newref->label = pool_window_label;
|
3987 |
|
|
ref = pool_vector[pool_window_last].wend;
|
3988 |
|
|
newref->next = ref;
|
3989 |
|
|
pool_vector[pool_window_last].wend = newref;
|
3990 |
|
|
}
|
3991 |
|
|
if (new_rtx)
|
3992 |
|
|
pool_window_label = new_rtx;
|
3993 |
|
|
pool_window_last = i;
|
3994 |
|
|
return lab;
|
3995 |
|
|
}
|
3996 |
|
|
}
|
3997 |
|
|
}
|
3998 |
|
|
|
3999 |
|
|
/* Need a new one. */
|
4000 |
|
|
pool_vector[pool_size].value = x;
|
4001 |
|
|
if (last_value && rtx_equal_p (last_value, pool_vector[pool_size - 1].value))
|
4002 |
|
|
{
|
4003 |
|
|
lab = 0;
|
4004 |
|
|
pool_vector[pool_size - 1].part_of_sequence_p = true;
|
4005 |
|
|
}
|
4006 |
|
|
else
|
4007 |
|
|
lab = gen_label_rtx ();
|
4008 |
|
|
pool_vector[pool_size].mode = mode;
|
4009 |
|
|
pool_vector[pool_size].label = lab;
|
4010 |
|
|
pool_vector[pool_size].wend = NULL;
|
4011 |
|
|
pool_vector[pool_size].part_of_sequence_p = (lab == 0);
|
4012 |
|
|
if (lab && pool_window_label)
|
4013 |
|
|
{
|
4014 |
|
|
newref = (label_ref_list_t) pool_alloc (label_ref_list_pool);
|
4015 |
|
|
newref->label = pool_window_label;
|
4016 |
|
|
ref = pool_vector[pool_window_last].wend;
|
4017 |
|
|
newref->next = ref;
|
4018 |
|
|
pool_vector[pool_window_last].wend = newref;
|
4019 |
|
|
}
|
4020 |
|
|
if (lab)
|
4021 |
|
|
pool_window_label = lab;
|
4022 |
|
|
pool_window_last = pool_size;
|
4023 |
|
|
pool_size++;
|
4024 |
|
|
return lab;
|
4025 |
|
|
}
|
4026 |
|
|
|
4027 |
|
|
/* Output the literal table. START, if nonzero, is the first instruction
|
4028 |
|
|
this table is needed for, and also indicates that there is at least one
|
4029 |
|
|
casesi_worker_2 instruction; We have to emit the operand3 labels from
|
4030 |
|
|
these insns at a 4-byte aligned position. BARRIER is the barrier
|
4031 |
|
|
after which we are to place the table. */
|
4032 |
|
|
|
4033 |
|
|
static void
|
4034 |
|
|
dump_table (rtx start, rtx barrier)
|
4035 |
|
|
{
|
4036 |
|
|
rtx scan = barrier;
|
4037 |
|
|
int i;
|
4038 |
|
|
int need_align = 1;
|
4039 |
|
|
rtx lab;
|
4040 |
|
|
label_ref_list_t ref;
|
4041 |
|
|
int have_df = 0;
|
4042 |
|
|
|
4043 |
|
|
/* Do two passes, first time dump out the HI sized constants. */
|
4044 |
|
|
|
4045 |
|
|
for (i = 0; i < pool_size; i++)
|
4046 |
|
|
{
|
4047 |
|
|
pool_node *p = &pool_vector[i];
|
4048 |
|
|
|
4049 |
|
|
if (p->mode == HImode)
|
4050 |
|
|
{
|
4051 |
|
|
if (need_align)
|
4052 |
|
|
{
|
4053 |
|
|
scan = emit_insn_after (gen_align_2 (), scan);
|
4054 |
|
|
need_align = 0;
|
4055 |
|
|
}
|
4056 |
|
|
for (lab = p->label; lab; lab = LABEL_REFS (lab))
|
4057 |
|
|
scan = emit_label_after (lab, scan);
|
4058 |
|
|
scan = emit_insn_after (gen_consttable_2 (p->value, const0_rtx),
|
4059 |
|
|
scan);
|
4060 |
|
|
for (ref = p->wend; ref; ref = ref->next)
|
4061 |
|
|
{
|
4062 |
|
|
lab = ref->label;
|
4063 |
|
|
scan = emit_insn_after (gen_consttable_window_end (lab), scan);
|
4064 |
|
|
}
|
4065 |
|
|
}
|
4066 |
|
|
else if (p->mode == DFmode)
|
4067 |
|
|
have_df = 1;
|
4068 |
|
|
}
|
4069 |
|
|
|
4070 |
|
|
need_align = 1;
|
4071 |
|
|
|
4072 |
|
|
if (start)
|
4073 |
|
|
{
|
4074 |
|
|
scan = emit_insn_after (gen_align_4 (), scan);
|
4075 |
|
|
need_align = 0;
|
4076 |
|
|
for (; start != barrier; start = NEXT_INSN (start))
|
4077 |
|
|
if (NONJUMP_INSN_P (start)
|
4078 |
|
|
&& recog_memoized (start) == CODE_FOR_casesi_worker_2)
|
4079 |
|
|
{
|
4080 |
|
|
rtx src = SET_SRC (XVECEXP (PATTERN (start), 0, 0));
|
4081 |
|
|
rtx lab = XEXP (XVECEXP (src, 0, 3), 0);
|
4082 |
|
|
|
4083 |
|
|
scan = emit_label_after (lab, scan);
|
4084 |
|
|
}
|
4085 |
|
|
}
|
4086 |
|
|
if (TARGET_FMOVD && TARGET_ALIGN_DOUBLE && have_df)
|
4087 |
|
|
{
|
4088 |
|
|
rtx align_insn = NULL_RTX;
|
4089 |
|
|
|
4090 |
|
|
scan = emit_label_after (gen_label_rtx (), scan);
|
4091 |
|
|
scan = emit_insn_after (gen_align_log (GEN_INT (3)), scan);
|
4092 |
|
|
need_align = 0;
|
4093 |
|
|
|
4094 |
|
|
for (i = 0; i < pool_size; i++)
|
4095 |
|
|
{
|
4096 |
|
|
pool_node *p = &pool_vector[i];
|
4097 |
|
|
|
4098 |
|
|
switch (p->mode)
|
4099 |
|
|
{
|
4100 |
|
|
case HImode:
|
4101 |
|
|
break;
|
4102 |
|
|
case SImode:
|
4103 |
|
|
case SFmode:
|
4104 |
|
|
if (align_insn && !p->part_of_sequence_p)
|
4105 |
|
|
{
|
4106 |
|
|
for (lab = p->label; lab; lab = LABEL_REFS (lab))
|
4107 |
|
|
emit_label_before (lab, align_insn);
|
4108 |
|
|
emit_insn_before (gen_consttable_4 (p->value, const0_rtx),
|
4109 |
|
|
align_insn);
|
4110 |
|
|
for (ref = p->wend; ref; ref = ref->next)
|
4111 |
|
|
{
|
4112 |
|
|
lab = ref->label;
|
4113 |
|
|
emit_insn_before (gen_consttable_window_end (lab),
|
4114 |
|
|
align_insn);
|
4115 |
|
|
}
|
4116 |
|
|
delete_insn (align_insn);
|
4117 |
|
|
align_insn = NULL_RTX;
|
4118 |
|
|
continue;
|
4119 |
|
|
}
|
4120 |
|
|
else
|
4121 |
|
|
{
|
4122 |
|
|
for (lab = p->label; lab; lab = LABEL_REFS (lab))
|
4123 |
|
|
scan = emit_label_after (lab, scan);
|
4124 |
|
|
scan = emit_insn_after (gen_consttable_4 (p->value,
|
4125 |
|
|
const0_rtx), scan);
|
4126 |
|
|
need_align = ! need_align;
|
4127 |
|
|
}
|
4128 |
|
|
break;
|
4129 |
|
|
case DFmode:
|
4130 |
|
|
if (need_align)
|
4131 |
|
|
{
|
4132 |
|
|
scan = emit_insn_after (gen_align_log (GEN_INT (3)), scan);
|
4133 |
|
|
align_insn = scan;
|
4134 |
|
|
need_align = 0;
|
4135 |
|
|
}
|
4136 |
|
|
case DImode:
|
4137 |
|
|
for (lab = p->label; lab; lab = LABEL_REFS (lab))
|
4138 |
|
|
scan = emit_label_after (lab, scan);
|
4139 |
|
|
scan = emit_insn_after (gen_consttable_8 (p->value, const0_rtx),
|
4140 |
|
|
scan);
|
4141 |
|
|
break;
|
4142 |
|
|
default:
|
4143 |
|
|
gcc_unreachable ();
|
4144 |
|
|
}
|
4145 |
|
|
|
4146 |
|
|
if (p->mode != HImode)
|
4147 |
|
|
{
|
4148 |
|
|
for (ref = p->wend; ref; ref = ref->next)
|
4149 |
|
|
{
|
4150 |
|
|
lab = ref->label;
|
4151 |
|
|
scan = emit_insn_after (gen_consttable_window_end (lab),
|
4152 |
|
|
scan);
|
4153 |
|
|
}
|
4154 |
|
|
}
|
4155 |
|
|
}
|
4156 |
|
|
|
4157 |
|
|
pool_size = 0;
|
4158 |
|
|
}
|
4159 |
|
|
|
4160 |
|
|
for (i = 0; i < pool_size; i++)
|
4161 |
|
|
{
|
4162 |
|
|
pool_node *p = &pool_vector[i];
|
4163 |
|
|
|
4164 |
|
|
switch (p->mode)
|
4165 |
|
|
{
|
4166 |
|
|
case HImode:
|
4167 |
|
|
break;
|
4168 |
|
|
case SImode:
|
4169 |
|
|
case SFmode:
|
4170 |
|
|
if (need_align)
|
4171 |
|
|
{
|
4172 |
|
|
need_align = 0;
|
4173 |
|
|
scan = emit_label_after (gen_label_rtx (), scan);
|
4174 |
|
|
scan = emit_insn_after (gen_align_4 (), scan);
|
4175 |
|
|
}
|
4176 |
|
|
for (lab = p->label; lab; lab = LABEL_REFS (lab))
|
4177 |
|
|
scan = emit_label_after (lab, scan);
|
4178 |
|
|
scan = emit_insn_after (gen_consttable_4 (p->value, const0_rtx),
|
4179 |
|
|
scan);
|
4180 |
|
|
break;
|
4181 |
|
|
case DFmode:
|
4182 |
|
|
case DImode:
|
4183 |
|
|
if (need_align)
|
4184 |
|
|
{
|
4185 |
|
|
need_align = 0;
|
4186 |
|
|
scan = emit_label_after (gen_label_rtx (), scan);
|
4187 |
|
|
scan = emit_insn_after (gen_align_4 (), scan);
|
4188 |
|
|
}
|
4189 |
|
|
for (lab = p->label; lab; lab = LABEL_REFS (lab))
|
4190 |
|
|
scan = emit_label_after (lab, scan);
|
4191 |
|
|
scan = emit_insn_after (gen_consttable_8 (p->value, const0_rtx),
|
4192 |
|
|
scan);
|
4193 |
|
|
break;
|
4194 |
|
|
default:
|
4195 |
|
|
gcc_unreachable ();
|
4196 |
|
|
}
|
4197 |
|
|
|
4198 |
|
|
if (p->mode != HImode)
|
4199 |
|
|
{
|
4200 |
|
|
for (ref = p->wend; ref; ref = ref->next)
|
4201 |
|
|
{
|
4202 |
|
|
lab = ref->label;
|
4203 |
|
|
scan = emit_insn_after (gen_consttable_window_end (lab), scan);
|
4204 |
|
|
}
|
4205 |
|
|
}
|
4206 |
|
|
}
|
4207 |
|
|
|
4208 |
|
|
scan = emit_insn_after (gen_consttable_end (), scan);
|
4209 |
|
|
scan = emit_barrier_after (scan);
|
4210 |
|
|
pool_size = 0;
|
4211 |
|
|
pool_window_label = NULL_RTX;
|
4212 |
|
|
pool_window_last = 0;
|
4213 |
|
|
}
|
4214 |
|
|
|
4215 |
|
|
/* Return nonzero if constant would be an ok source for a
|
4216 |
|
|
mov.w instead of a mov.l. */
|
4217 |
|
|
|
4218 |
|
|
static int
|
4219 |
|
|
hi_const (rtx src)
|
4220 |
|
|
{
|
4221 |
|
|
return (CONST_INT_P (src)
|
4222 |
|
|
&& INTVAL (src) >= -32768
|
4223 |
|
|
&& INTVAL (src) <= 32767);
|
4224 |
|
|
}
|
4225 |
|
|
|
4226 |
|
|
#define MOVA_LABELREF(mova) XVECEXP (SET_SRC (PATTERN (mova)), 0, 0)
|
4227 |
|
|
|
4228 |
|
|
/* Nonzero if the insn is a move instruction which needs to be fixed. */
|
4229 |
|
|
|
4230 |
|
|
/* ??? For a DImode/DFmode moves, we don't need to fix it if each half of the
|
4231 |
|
|
CONST_DOUBLE input value is CONST_OK_FOR_I08. For a SFmode move, we don't
|
4232 |
|
|
need to fix it if the input value is CONST_OK_FOR_I08. */
|
4233 |
|
|
|
4234 |
|
|
static int
|
4235 |
|
|
broken_move (rtx insn)
|
4236 |
|
|
{
|
4237 |
|
|
if (NONJUMP_INSN_P (insn))
|
4238 |
|
|
{
|
4239 |
|
|
rtx pat = PATTERN (insn);
|
4240 |
|
|
if (GET_CODE (pat) == PARALLEL)
|
4241 |
|
|
pat = XVECEXP (pat, 0, 0);
|
4242 |
|
|
if (GET_CODE (pat) == SET
|
4243 |
|
|
/* We can load any 8-bit value if we don't care what the high
|
4244 |
|
|
order bits end up as. */
|
4245 |
|
|
&& GET_MODE (SET_DEST (pat)) != QImode
|
4246 |
|
|
&& (CONSTANT_P (SET_SRC (pat))
|
4247 |
|
|
/* Match mova_const. */
|
4248 |
|
|
|| (GET_CODE (SET_SRC (pat)) == UNSPEC
|
4249 |
|
|
&& XINT (SET_SRC (pat), 1) == UNSPEC_MOVA
|
4250 |
|
|
&& GET_CODE (XVECEXP (SET_SRC (pat), 0, 0)) == CONST))
|
4251 |
|
|
&& ! (TARGET_SH2E
|
4252 |
|
|
&& GET_CODE (SET_SRC (pat)) == CONST_DOUBLE
|
4253 |
|
|
&& (fp_zero_operand (SET_SRC (pat))
|
4254 |
|
|
|| fp_one_operand (SET_SRC (pat)))
|
4255 |
|
|
/* In general we don't know the current setting of fpscr, so disable fldi.
|
4256 |
|
|
There is an exception if this was a register-register move
|
4257 |
|
|
before reload - and hence it was ascertained that we have
|
4258 |
|
|
single precision setting - and in a post-reload optimization
|
4259 |
|
|
we changed this to do a constant load. In that case
|
4260 |
|
|
we don't have an r0 clobber, hence we must use fldi. */
|
4261 |
|
|
&& (TARGET_FMOVD
|
4262 |
|
|
|| (GET_CODE (XEXP (XVECEXP (PATTERN (insn), 0, 2), 0))
|
4263 |
|
|
== SCRATCH))
|
4264 |
|
|
&& REG_P (SET_DEST (pat))
|
4265 |
|
|
&& FP_REGISTER_P (REGNO (SET_DEST (pat))))
|
4266 |
|
|
&& ! (TARGET_SH2A
|
4267 |
|
|
&& GET_MODE (SET_DEST (pat)) == SImode
|
4268 |
|
|
&& (satisfies_constraint_I20 (SET_SRC (pat))
|
4269 |
|
|
|| satisfies_constraint_I28 (SET_SRC (pat))))
|
4270 |
|
|
&& ! satisfies_constraint_I08 (SET_SRC (pat)))
|
4271 |
|
|
return 1;
|
4272 |
|
|
}
|
4273 |
|
|
|
4274 |
|
|
return 0;
|
4275 |
|
|
}
|
4276 |
|
|
|
4277 |
|
|
static int
|
4278 |
|
|
mova_p (rtx insn)
|
4279 |
|
|
{
|
4280 |
|
|
return (NONJUMP_INSN_P (insn)
|
4281 |
|
|
&& GET_CODE (PATTERN (insn)) == SET
|
4282 |
|
|
&& GET_CODE (SET_SRC (PATTERN (insn))) == UNSPEC
|
4283 |
|
|
&& XINT (SET_SRC (PATTERN (insn)), 1) == UNSPEC_MOVA
|
4284 |
|
|
/* Don't match mova_const. */
|
4285 |
|
|
&& GET_CODE (MOVA_LABELREF (insn)) == LABEL_REF);
|
4286 |
|
|
}
|
4287 |
|
|
|
4288 |
|
|
/* Fix up a mova from a switch that went out of range. */
|
4289 |
|
|
static void
|
4290 |
|
|
fixup_mova (rtx mova)
|
4291 |
|
|
{
|
4292 |
|
|
PUT_MODE (XEXP (MOVA_LABELREF (mova), 0), QImode);
|
4293 |
|
|
if (! flag_pic)
|
4294 |
|
|
{
|
4295 |
|
|
SET_SRC (PATTERN (mova)) = MOVA_LABELREF (mova);
|
4296 |
|
|
INSN_CODE (mova) = -1;
|
4297 |
|
|
}
|
4298 |
|
|
else
|
4299 |
|
|
{
|
4300 |
|
|
rtx worker = mova;
|
4301 |
|
|
rtx lab = gen_label_rtx ();
|
4302 |
|
|
rtx wpat, wpat0, wpat1, wsrc, target, base, diff;
|
4303 |
|
|
|
4304 |
|
|
do
|
4305 |
|
|
{
|
4306 |
|
|
worker = NEXT_INSN (worker);
|
4307 |
|
|
gcc_assert (worker
|
4308 |
|
|
&& !LABEL_P (worker)
|
4309 |
|
|
&& !JUMP_P (worker));
|
4310 |
|
|
} while (NOTE_P (worker)
|
4311 |
|
|
|| recog_memoized (worker) != CODE_FOR_casesi_worker_1);
|
4312 |
|
|
wpat = PATTERN (worker);
|
4313 |
|
|
wpat0 = XVECEXP (wpat, 0, 0);
|
4314 |
|
|
wpat1 = XVECEXP (wpat, 0, 1);
|
4315 |
|
|
wsrc = SET_SRC (wpat0);
|
4316 |
|
|
PATTERN (worker) = (gen_casesi_worker_2
|
4317 |
|
|
(SET_DEST (wpat0), XVECEXP (wsrc, 0, 1),
|
4318 |
|
|
XEXP (XVECEXP (wsrc, 0, 2), 0), lab,
|
4319 |
|
|
XEXP (wpat1, 0)));
|
4320 |
|
|
INSN_CODE (worker) = -1;
|
4321 |
|
|
target = XVECEXP (SET_SRC (PATTERN (mova)), 0, 0);
|
4322 |
|
|
base = gen_rtx_LABEL_REF (Pmode, lab);
|
4323 |
|
|
diff = gen_rtx_UNSPEC (Pmode, gen_rtvec (2, target, base), UNSPEC_SYMOFF);
|
4324 |
|
|
SET_SRC (PATTERN (mova)) = gen_rtx_CONST (Pmode, diff);
|
4325 |
|
|
INSN_CODE (mova) = -1;
|
4326 |
|
|
}
|
4327 |
|
|
}
|
4328 |
|
|
|
4329 |
|
|
/* NEW_MOVA is a mova we've just encountered while scanning forward. Update
|
4330 |
|
|
*num_mova, and check if the new mova is not nested within the first one.
|
4331 |
|
|
return 0 if *first_mova was replaced, 1 if new_mova was replaced,
|
4332 |
|
|
2 if new_mova has been assigned to *first_mova, -1 otherwise.. */
|
4333 |
|
|
static int
|
4334 |
|
|
untangle_mova (int *num_mova, rtx *first_mova, rtx new_mova)
|
4335 |
|
|
{
|
4336 |
|
|
int n_addr = 0; /* Initialization to shut up spurious warning. */
|
4337 |
|
|
int f_target, n_target = 0; /* Likewise. */
|
4338 |
|
|
|
4339 |
|
|
if (optimize)
|
4340 |
|
|
{
|
4341 |
|
|
/* If NEW_MOVA has no address yet, it will be handled later. */
|
4342 |
|
|
if (INSN_ADDRESSES_SIZE() <= (unsigned) INSN_UID (new_mova))
|
4343 |
|
|
return -1;
|
4344 |
|
|
|
4345 |
|
|
n_addr = INSN_ADDRESSES (INSN_UID (new_mova));
|
4346 |
|
|
n_target = INSN_ADDRESSES (INSN_UID (XEXP (MOVA_LABELREF (new_mova), 0)));
|
4347 |
|
|
if (n_addr > n_target || n_addr + 1022 < n_target)
|
4348 |
|
|
{
|
4349 |
|
|
/* Change the mova into a load.
|
4350 |
|
|
broken_move will then return true for it. */
|
4351 |
|
|
fixup_mova (new_mova);
|
4352 |
|
|
return 1;
|
4353 |
|
|
}
|
4354 |
|
|
}
|
4355 |
|
|
if (!(*num_mova)++)
|
4356 |
|
|
{
|
4357 |
|
|
*first_mova = new_mova;
|
4358 |
|
|
return 2;
|
4359 |
|
|
}
|
4360 |
|
|
if (!optimize
|
4361 |
|
|
|| ((f_target
|
4362 |
|
|
= INSN_ADDRESSES (INSN_UID (XEXP (MOVA_LABELREF (*first_mova), 0))))
|
4363 |
|
|
>= n_target))
|
4364 |
|
|
return -1;
|
4365 |
|
|
|
4366 |
|
|
(*num_mova)--;
|
4367 |
|
|
if (f_target - INSN_ADDRESSES (INSN_UID (*first_mova))
|
4368 |
|
|
> n_target - n_addr)
|
4369 |
|
|
{
|
4370 |
|
|
fixup_mova (*first_mova);
|
4371 |
|
|
return 0;
|
4372 |
|
|
}
|
4373 |
|
|
else
|
4374 |
|
|
{
|
4375 |
|
|
fixup_mova (new_mova);
|
4376 |
|
|
return 1;
|
4377 |
|
|
}
|
4378 |
|
|
}
|
4379 |
|
|
|
4380 |
|
|
/* Find the last barrier from insn FROM which is close enough to hold the
|
4381 |
|
|
constant pool. If we can't find one, then create one near the end of
|
4382 |
|
|
the range. */
|
4383 |
|
|
|
4384 |
|
|
static rtx
|
4385 |
|
|
find_barrier (int num_mova, rtx mova, rtx from)
|
4386 |
|
|
{
|
4387 |
|
|
int count_si = 0;
|
4388 |
|
|
int count_hi = 0;
|
4389 |
|
|
int found_hi = 0;
|
4390 |
|
|
int found_si = 0;
|
4391 |
|
|
int found_di = 0;
|
4392 |
|
|
int hi_align = 2;
|
4393 |
|
|
int si_align = 2;
|
4394 |
|
|
int leading_mova = num_mova;
|
4395 |
|
|
rtx barrier_before_mova = 0, found_barrier = 0, good_barrier = 0;
|
4396 |
|
|
int si_limit;
|
4397 |
|
|
int hi_limit;
|
4398 |
|
|
rtx orig = from;
|
4399 |
|
|
rtx last_got = NULL_RTX;
|
4400 |
|
|
rtx last_symoff = NULL_RTX;
|
4401 |
|
|
|
4402 |
|
|
/* For HImode: range is 510, add 4 because pc counts from address of
|
4403 |
|
|
second instruction after this one, subtract 2 for the jump instruction
|
4404 |
|
|
that we may need to emit before the table, subtract 2 for the instruction
|
4405 |
|
|
that fills the jump delay slot (in very rare cases, reorg will take an
|
4406 |
|
|
instruction from after the constant pool or will leave the delay slot
|
4407 |
|
|
empty). This gives 510.
|
4408 |
|
|
For SImode: range is 1020, add 4 because pc counts from address of
|
4409 |
|
|
second instruction after this one, subtract 2 in case pc is 2 byte
|
4410 |
|
|
aligned, subtract 2 for the jump instruction that we may need to emit
|
4411 |
|
|
before the table, subtract 2 for the instruction that fills the jump
|
4412 |
|
|
delay slot. This gives 1018. */
|
4413 |
|
|
|
4414 |
|
|
/* The branch will always be shortened now that the reference address for
|
4415 |
|
|
forward branches is the successor address, thus we need no longer make
|
4416 |
|
|
adjustments to the [sh]i_limit for -O0. */
|
4417 |
|
|
|
4418 |
|
|
si_limit = 1018;
|
4419 |
|
|
hi_limit = 510;
|
4420 |
|
|
|
4421 |
|
|
while (from && count_si < si_limit && count_hi < hi_limit)
|
4422 |
|
|
{
|
4423 |
|
|
int inc = get_attr_length (from);
|
4424 |
|
|
int new_align = 1;
|
4425 |
|
|
|
4426 |
|
|
/* If this is a label that existed at the time of the compute_alignments
|
4427 |
|
|
call, determine the alignment. N.B. When find_barrier recurses for
|
4428 |
|
|
an out-of-reach mova, we might see labels at the start of previously
|
4429 |
|
|
inserted constant tables. */
|
4430 |
|
|
if (LABEL_P (from)
|
4431 |
|
|
&& CODE_LABEL_NUMBER (from) <= max_labelno_before_reorg)
|
4432 |
|
|
{
|
4433 |
|
|
if (optimize)
|
4434 |
|
|
new_align = 1 << label_to_alignment (from);
|
4435 |
|
|
else if (BARRIER_P (prev_nonnote_insn (from)))
|
4436 |
|
|
new_align = 1 << barrier_align (from);
|
4437 |
|
|
else
|
4438 |
|
|
new_align = 1;
|
4439 |
|
|
inc = 0;
|
4440 |
|
|
}
|
4441 |
|
|
/* In case we are scanning a constant table because of recursion, check
|
4442 |
|
|
for explicit alignments. If the table is long, we might be forced
|
4443 |
|
|
to emit the new table in front of it; the length of the alignment
|
4444 |
|
|
might be the last straw. */
|
4445 |
|
|
else if (NONJUMP_INSN_P (from)
|
4446 |
|
|
&& GET_CODE (PATTERN (from)) == UNSPEC_VOLATILE
|
4447 |
|
|
&& XINT (PATTERN (from), 1) == UNSPECV_ALIGN)
|
4448 |
|
|
new_align = INTVAL (XVECEXP (PATTERN (from), 0, 0));
|
4449 |
|
|
/* When we find the end of a constant table, paste the new constant
|
4450 |
|
|
at the end. That is better than putting it in front because
|
4451 |
|
|
this way, we don't need extra alignment for adding a 4-byte-aligned
|
4452 |
|
|
mov(a) label to a 2/4 or 8/4 byte aligned table. */
|
4453 |
|
|
else if (NONJUMP_INSN_P (from)
|
4454 |
|
|
&& GET_CODE (PATTERN (from)) == UNSPEC_VOLATILE
|
4455 |
|
|
&& XINT (PATTERN (from), 1) == UNSPECV_CONST_END)
|
4456 |
|
|
return from;
|
4457 |
|
|
|
4458 |
|
|
if (BARRIER_P (from))
|
4459 |
|
|
{
|
4460 |
|
|
rtx next;
|
4461 |
|
|
|
4462 |
|
|
found_barrier = from;
|
4463 |
|
|
|
4464 |
|
|
/* If we are at the end of the function, or in front of an alignment
|
4465 |
|
|
instruction, we need not insert an extra alignment. We prefer
|
4466 |
|
|
this kind of barrier. */
|
4467 |
|
|
if (barrier_align (from) > 2)
|
4468 |
|
|
good_barrier = from;
|
4469 |
|
|
|
4470 |
|
|
/* If we are at the end of a hot/cold block, dump the constants
|
4471 |
|
|
here. */
|
4472 |
|
|
next = NEXT_INSN (from);
|
4473 |
|
|
if (next
|
4474 |
|
|
&& NOTE_P (next)
|
4475 |
|
|
&& NOTE_KIND (next) == NOTE_INSN_SWITCH_TEXT_SECTIONS)
|
4476 |
|
|
break;
|
4477 |
|
|
}
|
4478 |
|
|
|
4479 |
|
|
if (broken_move (from))
|
4480 |
|
|
{
|
4481 |
|
|
rtx pat, src, dst;
|
4482 |
|
|
enum machine_mode mode;
|
4483 |
|
|
|
4484 |
|
|
pat = PATTERN (from);
|
4485 |
|
|
if (GET_CODE (pat) == PARALLEL)
|
4486 |
|
|
pat = XVECEXP (pat, 0, 0);
|
4487 |
|
|
src = SET_SRC (pat);
|
4488 |
|
|
dst = SET_DEST (pat);
|
4489 |
|
|
mode = GET_MODE (dst);
|
4490 |
|
|
|
4491 |
|
|
/* GOT pcrelat setting comes in pair of
|
4492 |
|
|
mova .L8,r0
|
4493 |
|
|
mov.l .L8,r12
|
4494 |
|
|
instructions. (plus add r0,r12).
|
4495 |
|
|
Remember if we see one without the other. */
|
4496 |
|
|
if (GET_CODE (src) == UNSPEC && PIC_ADDR_P (XVECEXP (src, 0, 0)))
|
4497 |
|
|
last_got = last_got ? NULL_RTX : from;
|
4498 |
|
|
else if (PIC_ADDR_P (src))
|
4499 |
|
|
last_got = last_got ? NULL_RTX : from;
|
4500 |
|
|
|
4501 |
|
|
/* We must explicitly check the mode, because sometimes the
|
4502 |
|
|
front end will generate code to load unsigned constants into
|
4503 |
|
|
HImode targets without properly sign extending them. */
|
4504 |
|
|
if (mode == HImode
|
4505 |
|
|
|| (mode == SImode && hi_const (src) && REGNO (dst) != FPUL_REG))
|
4506 |
|
|
{
|
4507 |
|
|
found_hi += 2;
|
4508 |
|
|
/* We put the short constants before the long constants, so
|
4509 |
|
|
we must count the length of short constants in the range
|
4510 |
|
|
for the long constants. */
|
4511 |
|
|
/* ??? This isn't optimal, but is easy to do. */
|
4512 |
|
|
si_limit -= 2;
|
4513 |
|
|
}
|
4514 |
|
|
else
|
4515 |
|
|
{
|
4516 |
|
|
/* We dump DF/DI constants before SF/SI ones, because
|
4517 |
|
|
the limit is the same, but the alignment requirements
|
4518 |
|
|
are higher. We may waste up to 4 additional bytes
|
4519 |
|
|
for alignment, and the DF/DI constant may have
|
4520 |
|
|
another SF/SI constant placed before it. */
|
4521 |
|
|
if (TARGET_SHCOMPACT
|
4522 |
|
|
&& ! found_di
|
4523 |
|
|
&& (mode == DFmode || mode == DImode))
|
4524 |
|
|
{
|
4525 |
|
|
found_di = 1;
|
4526 |
|
|
si_limit -= 8;
|
4527 |
|
|
}
|
4528 |
|
|
while (si_align > 2 && found_si + si_align - 2 > count_si)
|
4529 |
|
|
si_align >>= 1;
|
4530 |
|
|
if (found_si > count_si)
|
4531 |
|
|
count_si = found_si;
|
4532 |
|
|
found_si += GET_MODE_SIZE (mode);
|
4533 |
|
|
if (num_mova)
|
4534 |
|
|
si_limit -= GET_MODE_SIZE (mode);
|
4535 |
|
|
}
|
4536 |
|
|
}
|
4537 |
|
|
|
4538 |
|
|
if (mova_p (from))
|
4539 |
|
|
{
|
4540 |
|
|
switch (untangle_mova (&num_mova, &mova, from))
|
4541 |
|
|
{
|
4542 |
|
|
case 1:
|
4543 |
|
|
if (flag_pic)
|
4544 |
|
|
{
|
4545 |
|
|
rtx src = SET_SRC (PATTERN (from));
|
4546 |
|
|
if (GET_CODE (src) == CONST
|
4547 |
|
|
&& GET_CODE (XEXP (src, 0)) == UNSPEC
|
4548 |
|
|
&& XINT (XEXP (src, 0), 1) == UNSPEC_SYMOFF)
|
4549 |
|
|
last_symoff = from;
|
4550 |
|
|
}
|
4551 |
|
|
break;
|
4552 |
|
|
case 0: return find_barrier (0, 0, mova);
|
4553 |
|
|
case 2:
|
4554 |
|
|
{
|
4555 |
|
|
leading_mova = 0;
|
4556 |
|
|
barrier_before_mova
|
4557 |
|
|
= good_barrier ? good_barrier : found_barrier;
|
4558 |
|
|
}
|
4559 |
|
|
default: break;
|
4560 |
|
|
}
|
4561 |
|
|
if (found_si > count_si)
|
4562 |
|
|
count_si = found_si;
|
4563 |
|
|
}
|
4564 |
|
|
else if (JUMP_TABLE_DATA_P (from))
|
4565 |
|
|
{
|
4566 |
|
|
if ((num_mova > 1 && GET_MODE (prev_nonnote_insn (from)) == VOIDmode)
|
4567 |
|
|
|| (num_mova
|
4568 |
|
|
&& (prev_nonnote_insn (from)
|
4569 |
|
|
== XEXP (MOVA_LABELREF (mova), 0))))
|
4570 |
|
|
num_mova--;
|
4571 |
|
|
if (barrier_align (next_real_insn (from)) == align_jumps_log)
|
4572 |
|
|
{
|
4573 |
|
|
/* We have just passed the barrier in front of the
|
4574 |
|
|
ADDR_DIFF_VEC, which is stored in found_barrier. Since
|
4575 |
|
|
the ADDR_DIFF_VEC is accessed as data, just like our pool
|
4576 |
|
|
constants, this is a good opportunity to accommodate what
|
4577 |
|
|
we have gathered so far.
|
4578 |
|
|
If we waited any longer, we could end up at a barrier in
|
4579 |
|
|
front of code, which gives worse cache usage for separated
|
4580 |
|
|
instruction / data caches. */
|
4581 |
|
|
good_barrier = found_barrier;
|
4582 |
|
|
break;
|
4583 |
|
|
}
|
4584 |
|
|
else
|
4585 |
|
|
{
|
4586 |
|
|
rtx body = PATTERN (from);
|
4587 |
|
|
inc = XVECLEN (body, 1) * GET_MODE_SIZE (GET_MODE (body));
|
4588 |
|
|
}
|
4589 |
|
|
}
|
4590 |
|
|
/* For the SH1, we generate alignments even after jumps-around-jumps. */
|
4591 |
|
|
else if (JUMP_P (from)
|
4592 |
|
|
&& ! TARGET_SH2
|
4593 |
|
|
&& ! TARGET_SMALLCODE)
|
4594 |
|
|
new_align = 4;
|
4595 |
|
|
|
4596 |
|
|
/* There is a possibility that a bf is transformed into a bf/s by the
|
4597 |
|
|
delay slot scheduler. */
|
4598 |
|
|
if (JUMP_P (from) && !JUMP_TABLE_DATA_P (from)
|
4599 |
|
|
&& get_attr_type (from) == TYPE_CBRANCH
|
4600 |
|
|
&& GET_CODE (PATTERN (NEXT_INSN (PREV_INSN (from)))) != SEQUENCE)
|
4601 |
|
|
inc += 2;
|
4602 |
|
|
|
4603 |
|
|
if (found_si)
|
4604 |
|
|
{
|
4605 |
|
|
count_si += inc;
|
4606 |
|
|
if (new_align > si_align)
|
4607 |
|
|
{
|
4608 |
|
|
si_limit -= (count_si - 1) & (new_align - si_align);
|
4609 |
|
|
si_align = new_align;
|
4610 |
|
|
}
|
4611 |
|
|
count_si = (count_si + new_align - 1) & -new_align;
|
4612 |
|
|
}
|
4613 |
|
|
if (found_hi)
|
4614 |
|
|
{
|
4615 |
|
|
count_hi += inc;
|
4616 |
|
|
if (new_align > hi_align)
|
4617 |
|
|
{
|
4618 |
|
|
hi_limit -= (count_hi - 1) & (new_align - hi_align);
|
4619 |
|
|
hi_align = new_align;
|
4620 |
|
|
}
|
4621 |
|
|
count_hi = (count_hi + new_align - 1) & -new_align;
|
4622 |
|
|
}
|
4623 |
|
|
from = NEXT_INSN (from);
|
4624 |
|
|
}
|
4625 |
|
|
|
4626 |
|
|
if (num_mova)
|
4627 |
|
|
{
|
4628 |
|
|
if (leading_mova)
|
4629 |
|
|
{
|
4630 |
|
|
/* Try as we might, the leading mova is out of range. Change
|
4631 |
|
|
it into a load (which will become a pcload) and retry. */
|
4632 |
|
|
fixup_mova (mova);
|
4633 |
|
|
return find_barrier (0, 0, mova);
|
4634 |
|
|
}
|
4635 |
|
|
else
|
4636 |
|
|
{
|
4637 |
|
|
/* Insert the constant pool table before the mova instruction,
|
4638 |
|
|
to prevent the mova label reference from going out of range. */
|
4639 |
|
|
from = mova;
|
4640 |
|
|
good_barrier = found_barrier = barrier_before_mova;
|
4641 |
|
|
}
|
4642 |
|
|
}
|
4643 |
|
|
|
4644 |
|
|
if (found_barrier)
|
4645 |
|
|
{
|
4646 |
|
|
if (good_barrier && next_real_insn (found_barrier))
|
4647 |
|
|
found_barrier = good_barrier;
|
4648 |
|
|
}
|
4649 |
|
|
else
|
4650 |
|
|
{
|
4651 |
|
|
/* We didn't find a barrier in time to dump our stuff,
|
4652 |
|
|
so we'll make one. */
|
4653 |
|
|
rtx label = gen_label_rtx ();
|
4654 |
|
|
|
4655 |
|
|
/* Don't emit a constant table in the middle of insns for
|
4656 |
|
|
casesi_worker_2. This is a bit overkill but is enough
|
4657 |
|
|
because casesi_worker_2 wouldn't appear so frequently. */
|
4658 |
|
|
if (last_symoff)
|
4659 |
|
|
from = last_symoff;
|
4660 |
|
|
|
4661 |
|
|
/* If we exceeded the range, then we must back up over the last
|
4662 |
|
|
instruction we looked at. Otherwise, we just need to undo the
|
4663 |
|
|
NEXT_INSN at the end of the loop. */
|
4664 |
|
|
if (PREV_INSN (from) != orig
|
4665 |
|
|
&& (count_hi > hi_limit || count_si > si_limit))
|
4666 |
|
|
from = PREV_INSN (PREV_INSN (from));
|
4667 |
|
|
else
|
4668 |
|
|
from = PREV_INSN (from);
|
4669 |
|
|
|
4670 |
|
|
/* Don't emit a constant table int the middle of global pointer setting,
|
4671 |
|
|
since that that would move the addressing base GOT into another table.
|
4672 |
|
|
We need the first mov instruction before the _GLOBAL_OFFSET_TABLE_
|
4673 |
|
|
in the pool anyway, so just move up the whole constant pool. */
|
4674 |
|
|
if (last_got)
|
4675 |
|
|
from = PREV_INSN (last_got);
|
4676 |
|
|
|
4677 |
|
|
/* Don't insert the constant pool table at the position which
|
4678 |
|
|
may be the landing pad. */
|
4679 |
|
|
if (flag_exceptions
|
4680 |
|
|
&& CALL_P (from)
|
4681 |
|
|
&& find_reg_note (from, REG_EH_REGION, NULL_RTX))
|
4682 |
|
|
from = PREV_INSN (from);
|
4683 |
|
|
|
4684 |
|
|
/* Walk back to be just before any jump or label.
|
4685 |
|
|
Putting it before a label reduces the number of times the branch
|
4686 |
|
|
around the constant pool table will be hit. Putting it before
|
4687 |
|
|
a jump makes it more likely that the bra delay slot will be
|
4688 |
|
|
filled. */
|
4689 |
|
|
while (NOTE_P (from) || JUMP_P (from)
|
4690 |
|
|
|| LABEL_P (from))
|
4691 |
|
|
from = PREV_INSN (from);
|
4692 |
|
|
|
4693 |
|
|
from = emit_jump_insn_after (gen_jump (label), from);
|
4694 |
|
|
JUMP_LABEL (from) = label;
|
4695 |
|
|
LABEL_NUSES (label) = 1;
|
4696 |
|
|
found_barrier = emit_barrier_after (from);
|
4697 |
|
|
emit_label_after (label, found_barrier);
|
4698 |
|
|
}
|
4699 |
|
|
|
4700 |
|
|
return found_barrier;
|
4701 |
|
|
}
|
4702 |
|
|
|
4703 |
|
|
/* If the instruction INSN is implemented by a special function, and we can
|
4704 |
|
|
positively find the register that is used to call the sfunc, and this
|
4705 |
|
|
register is not used anywhere else in this instruction - except as the
|
4706 |
|
|
destination of a set, return this register; else, return 0. */
|
4707 |
|
|
rtx
|
4708 |
|
|
sfunc_uses_reg (rtx insn)
|
4709 |
|
|
{
|
4710 |
|
|
int i;
|
4711 |
|
|
rtx pattern, part, reg_part, reg;
|
4712 |
|
|
|
4713 |
|
|
if (!NONJUMP_INSN_P (insn))
|
4714 |
|
|
return 0;
|
4715 |
|
|
pattern = PATTERN (insn);
|
4716 |
|
|
if (GET_CODE (pattern) != PARALLEL || get_attr_type (insn) != TYPE_SFUNC)
|
4717 |
|
|
return 0;
|
4718 |
|
|
|
4719 |
|
|
for (reg_part = 0, i = XVECLEN (pattern, 0) - 1; i >= 1; i--)
|
4720 |
|
|
{
|
4721 |
|
|
part = XVECEXP (pattern, 0, i);
|
4722 |
|
|
if (GET_CODE (part) == USE && GET_MODE (XEXP (part, 0)) == SImode)
|
4723 |
|
|
reg_part = part;
|
4724 |
|
|
}
|
4725 |
|
|
if (! reg_part)
|
4726 |
|
|
return 0;
|
4727 |
|
|
reg = XEXP (reg_part, 0);
|
4728 |
|
|
for (i = XVECLEN (pattern, 0) - 1; i >= 0; i--)
|
4729 |
|
|
{
|
4730 |
|
|
part = XVECEXP (pattern, 0, i);
|
4731 |
|
|
if (part == reg_part || GET_CODE (part) == CLOBBER)
|
4732 |
|
|
continue;
|
4733 |
|
|
if (reg_mentioned_p (reg, ((GET_CODE (part) == SET
|
4734 |
|
|
&& REG_P (SET_DEST (part)))
|
4735 |
|
|
? SET_SRC (part) : part)))
|
4736 |
|
|
return 0;
|
4737 |
|
|
}
|
4738 |
|
|
return reg;
|
4739 |
|
|
}
|
4740 |
|
|
|
4741 |
|
|
/* See if the only way in which INSN uses REG is by calling it, or by
|
4742 |
|
|
setting it while calling it. Set *SET to a SET rtx if the register
|
4743 |
|
|
is set by INSN. */
|
4744 |
|
|
|
4745 |
|
|
static int
|
4746 |
|
|
noncall_uses_reg (rtx reg, rtx insn, rtx *set)
|
4747 |
|
|
{
|
4748 |
|
|
rtx pattern, reg2;
|
4749 |
|
|
|
4750 |
|
|
*set = NULL_RTX;
|
4751 |
|
|
|
4752 |
|
|
reg2 = sfunc_uses_reg (insn);
|
4753 |
|
|
if (reg2 && REGNO (reg2) == REGNO (reg))
|
4754 |
|
|
{
|
4755 |
|
|
pattern = single_set (insn);
|
4756 |
|
|
if (pattern
|
4757 |
|
|
&& REG_P (SET_DEST (pattern))
|
4758 |
|
|
&& REGNO (reg) == REGNO (SET_DEST (pattern)))
|
4759 |
|
|
*set = pattern;
|
4760 |
|
|
return 0;
|
4761 |
|
|
}
|
4762 |
|
|
if (!CALL_P (insn))
|
4763 |
|
|
{
|
4764 |
|
|
/* We don't use rtx_equal_p because we don't care if the mode is
|
4765 |
|
|
different. */
|
4766 |
|
|
pattern = single_set (insn);
|
4767 |
|
|
if (pattern
|
4768 |
|
|
&& REG_P (SET_DEST (pattern))
|
4769 |
|
|
&& REGNO (reg) == REGNO (SET_DEST (pattern)))
|
4770 |
|
|
{
|
4771 |
|
|
rtx par, part;
|
4772 |
|
|
int i;
|
4773 |
|
|
|
4774 |
|
|
*set = pattern;
|
4775 |
|
|
par = PATTERN (insn);
|
4776 |
|
|
if (GET_CODE (par) == PARALLEL)
|
4777 |
|
|
for (i = XVECLEN (par, 0) - 1; i >= 0; i--)
|
4778 |
|
|
{
|
4779 |
|
|
part = XVECEXP (par, 0, i);
|
4780 |
|
|
if (GET_CODE (part) != SET && reg_mentioned_p (reg, part))
|
4781 |
|
|
return 1;
|
4782 |
|
|
}
|
4783 |
|
|
return reg_mentioned_p (reg, SET_SRC (pattern));
|
4784 |
|
|
}
|
4785 |
|
|
|
4786 |
|
|
return 1;
|
4787 |
|
|
}
|
4788 |
|
|
|
4789 |
|
|
pattern = PATTERN (insn);
|
4790 |
|
|
|
4791 |
|
|
if (GET_CODE (pattern) == PARALLEL)
|
4792 |
|
|
{
|
4793 |
|
|
int i;
|
4794 |
|
|
|
4795 |
|
|
for (i = XVECLEN (pattern, 0) - 1; i >= 1; i--)
|
4796 |
|
|
if (reg_mentioned_p (reg, XVECEXP (pattern, 0, i)))
|
4797 |
|
|
return 1;
|
4798 |
|
|
pattern = XVECEXP (pattern, 0, 0);
|
4799 |
|
|
}
|
4800 |
|
|
|
4801 |
|
|
if (GET_CODE (pattern) == SET)
|
4802 |
|
|
{
|
4803 |
|
|
if (reg_mentioned_p (reg, SET_DEST (pattern)))
|
4804 |
|
|
{
|
4805 |
|
|
/* We don't use rtx_equal_p, because we don't care if the
|
4806 |
|
|
mode is different. */
|
4807 |
|
|
if (!REG_P (SET_DEST (pattern))
|
4808 |
|
|
|| REGNO (reg) != REGNO (SET_DEST (pattern)))
|
4809 |
|
|
return 1;
|
4810 |
|
|
|
4811 |
|
|
*set = pattern;
|
4812 |
|
|
}
|
4813 |
|
|
|
4814 |
|
|
pattern = SET_SRC (pattern);
|
4815 |
|
|
}
|
4816 |
|
|
|
4817 |
|
|
if (GET_CODE (pattern) != CALL
|
4818 |
|
|
|| !MEM_P (XEXP (pattern, 0))
|
4819 |
|
|
|| ! rtx_equal_p (reg, XEXP (XEXP (pattern, 0), 0)))
|
4820 |
|
|
return 1;
|
4821 |
|
|
|
4822 |
|
|
return 0;
|
4823 |
|
|
}
|
4824 |
|
|
|
4825 |
|
|
/* Given a X, a pattern of an insn or a part of it, return a mask of used
|
4826 |
|
|
general registers. Bits 0..15 mean that the respective registers
|
4827 |
|
|
are used as inputs in the instruction. Bits 16..31 mean that the
|
4828 |
|
|
registers 0..15, respectively, are used as outputs, or are clobbered.
|
4829 |
|
|
IS_DEST should be set to 16 if X is the destination of a SET, else to 0. */
|
4830 |
|
|
int
|
4831 |
|
|
regs_used (rtx x, int is_dest)
|
4832 |
|
|
{
|
4833 |
|
|
enum rtx_code code;
|
4834 |
|
|
const char *fmt;
|
4835 |
|
|
int i, used = 0;
|
4836 |
|
|
|
4837 |
|
|
if (! x)
|
4838 |
|
|
return used;
|
4839 |
|
|
code = GET_CODE (x);
|
4840 |
|
|
switch (code)
|
4841 |
|
|
{
|
4842 |
|
|
case REG:
|
4843 |
|
|
if (REGNO (x) < 16)
|
4844 |
|
|
return (((1 << HARD_REGNO_NREGS (0, GET_MODE (x))) - 1)
|
4845 |
|
|
<< (REGNO (x) + is_dest));
|
4846 |
|
|
return 0;
|
4847 |
|
|
case SUBREG:
|
4848 |
|
|
{
|
4849 |
|
|
rtx y = SUBREG_REG (x);
|
4850 |
|
|
|
4851 |
|
|
if (!REG_P (y))
|
4852 |
|
|
break;
|
4853 |
|
|
if (REGNO (y) < 16)
|
4854 |
|
|
return (((1 << HARD_REGNO_NREGS (0, GET_MODE (x))) - 1)
|
4855 |
|
|
<< (REGNO (y) +
|
4856 |
|
|
subreg_regno_offset (REGNO (y),
|
4857 |
|
|
GET_MODE (y),
|
4858 |
|
|
SUBREG_BYTE (x),
|
4859 |
|
|
GET_MODE (x)) + is_dest));
|
4860 |
|
|
return 0;
|
4861 |
|
|
}
|
4862 |
|
|
case SET:
|
4863 |
|
|
return regs_used (SET_SRC (x), 0) | regs_used (SET_DEST (x), 16);
|
4864 |
|
|
case RETURN:
|
4865 |
|
|
/* If there was a return value, it must have been indicated with USE. */
|
4866 |
|
|
return 0x00ffff00;
|
4867 |
|
|
case CLOBBER:
|
4868 |
|
|
is_dest = 1;
|
4869 |
|
|
break;
|
4870 |
|
|
case MEM:
|
4871 |
|
|
is_dest = 0;
|
4872 |
|
|
break;
|
4873 |
|
|
case CALL:
|
4874 |
|
|
used |= 0x00ff00f0;
|
4875 |
|
|
break;
|
4876 |
|
|
default:
|
4877 |
|
|
break;
|
4878 |
|
|
}
|
4879 |
|
|
|
4880 |
|
|
fmt = GET_RTX_FORMAT (code);
|
4881 |
|
|
|
4882 |
|
|
for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
|
4883 |
|
|
{
|
4884 |
|
|
if (fmt[i] == 'E')
|
4885 |
|
|
{
|
4886 |
|
|
register int j;
|
4887 |
|
|
for (j = XVECLEN (x, i) - 1; j >= 0; j--)
|
4888 |
|
|
used |= regs_used (XVECEXP (x, i, j), is_dest);
|
4889 |
|
|
}
|
4890 |
|
|
else if (fmt[i] == 'e')
|
4891 |
|
|
used |= regs_used (XEXP (x, i), is_dest);
|
4892 |
|
|
}
|
4893 |
|
|
return used;
|
4894 |
|
|
}
|
4895 |
|
|
|
4896 |
|
|
/* Create an instruction that prevents redirection of a conditional branch
|
4897 |
|
|
to the destination of the JUMP with address ADDR.
|
4898 |
|
|
If the branch needs to be implemented as an indirect jump, try to find
|
4899 |
|
|
a scratch register for it.
|
4900 |
|
|
If NEED_BLOCK is 0, don't do anything unless we need a scratch register.
|
4901 |
|
|
If any preceding insn that doesn't fit into a delay slot is good enough,
|
4902 |
|
|
pass 1. Pass 2 if a definite blocking insn is needed.
|
4903 |
|
|
-1 is used internally to avoid deep recursion.
|
4904 |
|
|
If a blocking instruction is made or recognized, return it. */
|
4905 |
|
|
|
4906 |
|
|
static rtx
|
4907 |
|
|
gen_block_redirect (rtx jump, int addr, int need_block)
|
4908 |
|
|
{
|
4909 |
|
|
int dead = 0;
|
4910 |
|
|
rtx prev = prev_nonnote_insn (jump);
|
4911 |
|
|
rtx dest;
|
4912 |
|
|
|
4913 |
|
|
/* First, check if we already have an instruction that satisfies our need. */
|
4914 |
|
|
if (prev && NONJUMP_INSN_P (prev) && ! INSN_DELETED_P (prev))
|
4915 |
|
|
{
|
4916 |
|
|
if (INSN_CODE (prev) == CODE_FOR_indirect_jump_scratch)
|
4917 |
|
|
return prev;
|
4918 |
|
|
if (GET_CODE (PATTERN (prev)) == USE
|
4919 |
|
|
|| GET_CODE (PATTERN (prev)) == CLOBBER
|
4920 |
|
|
|| get_attr_in_delay_slot (prev) == IN_DELAY_SLOT_YES)
|
4921 |
|
|
prev = jump;
|
4922 |
|
|
else if ((need_block &= ~1) < 0)
|
4923 |
|
|
return prev;
|
4924 |
|
|
else if (recog_memoized (prev) == CODE_FOR_block_branch_redirect)
|
4925 |
|
|
need_block = 0;
|
4926 |
|
|
}
|
4927 |
|
|
if (GET_CODE (PATTERN (jump)) == RETURN)
|
4928 |
|
|
{
|
4929 |
|
|
if (! need_block)
|
4930 |
|
|
return prev;
|
4931 |
|
|
/* Reorg even does nasty things with return insns that cause branches
|
4932 |
|
|
to go out of range - see find_end_label and callers. */
|
4933 |
|
|
return emit_insn_before (gen_block_branch_redirect (const0_rtx) , jump);
|
4934 |
|
|
}
|
4935 |
|
|
/* We can't use JUMP_LABEL here because it might be undefined
|
4936 |
|
|
when not optimizing. */
|
4937 |
|
|
dest = XEXP (SET_SRC (PATTERN (jump)), 0);
|
4938 |
|
|
/* If the branch is out of range, try to find a scratch register for it. */
|
4939 |
|
|
if (optimize
|
4940 |
|
|
&& (INSN_ADDRESSES (INSN_UID (dest)) - addr + (unsigned) 4092
|
4941 |
|
|
> 4092 + 4098))
|
4942 |
|
|
{
|
4943 |
|
|
rtx scan;
|
4944 |
|
|
/* Don't look for the stack pointer as a scratch register,
|
4945 |
|
|
it would cause trouble if an interrupt occurred. */
|
4946 |
|
|
unsigned attempt = 0x7fff, used;
|
4947 |
|
|
int jump_left = flag_expensive_optimizations + 1;
|
4948 |
|
|
|
4949 |
|
|
/* It is likely that the most recent eligible instruction is wanted for
|
4950 |
|
|
the delay slot. Therefore, find out which registers it uses, and
|
4951 |
|
|
try to avoid using them. */
|
4952 |
|
|
|
4953 |
|
|
for (scan = jump; (scan = PREV_INSN (scan)); )
|
4954 |
|
|
{
|
4955 |
|
|
enum rtx_code code;
|
4956 |
|
|
|
4957 |
|
|
if (INSN_DELETED_P (scan))
|
4958 |
|
|
continue;
|
4959 |
|
|
code = GET_CODE (scan);
|
4960 |
|
|
if (code == CODE_LABEL || code == JUMP_INSN)
|
4961 |
|
|
break;
|
4962 |
|
|
if (code == INSN
|
4963 |
|
|
&& GET_CODE (PATTERN (scan)) != USE
|
4964 |
|
|
&& GET_CODE (PATTERN (scan)) != CLOBBER
|
4965 |
|
|
&& get_attr_in_delay_slot (scan) == IN_DELAY_SLOT_YES)
|
4966 |
|
|
{
|
4967 |
|
|
attempt &= ~regs_used (PATTERN (scan), 0);
|
4968 |
|
|
break;
|
4969 |
|
|
}
|
4970 |
|
|
}
|
4971 |
|
|
for (used = dead = 0, scan = JUMP_LABEL (jump);
|
4972 |
|
|
(scan = NEXT_INSN (scan)); )
|
4973 |
|
|
{
|
4974 |
|
|
enum rtx_code code;
|
4975 |
|
|
|
4976 |
|
|
if (INSN_DELETED_P (scan))
|
4977 |
|
|
continue;
|
4978 |
|
|
code = GET_CODE (scan);
|
4979 |
|
|
if (INSN_P (scan))
|
4980 |
|
|
{
|
4981 |
|
|
used |= regs_used (PATTERN (scan), 0);
|
4982 |
|
|
if (code == CALL_INSN)
|
4983 |
|
|
used |= regs_used (CALL_INSN_FUNCTION_USAGE (scan), 0);
|
4984 |
|
|
dead |= (used >> 16) & ~used;
|
4985 |
|
|
if (dead & attempt)
|
4986 |
|
|
{
|
4987 |
|
|
dead &= attempt;
|
4988 |
|
|
break;
|
4989 |
|
|
}
|
4990 |
|
|
if (code == JUMP_INSN)
|
4991 |
|
|
{
|
4992 |
|
|
if (jump_left-- && simplejump_p (scan))
|
4993 |
|
|
scan = JUMP_LABEL (scan);
|
4994 |
|
|
else
|
4995 |
|
|
break;
|
4996 |
|
|
}
|
4997 |
|
|
}
|
4998 |
|
|
}
|
4999 |
|
|
/* Mask out the stack pointer again, in case it was
|
5000 |
|
|
the only 'free' register we have found. */
|
5001 |
|
|
dead &= 0x7fff;
|
5002 |
|
|
}
|
5003 |
|
|
/* If the immediate destination is still in range, check for possible
|
5004 |
|
|
threading with a jump beyond the delay slot insn.
|
5005 |
|
|
Don't check if we are called recursively; the jump has been or will be
|
5006 |
|
|
checked in a different invocation then. */
|
5007 |
|
|
|
5008 |
|
|
else if (optimize && need_block >= 0)
|
5009 |
|
|
{
|
5010 |
|
|
rtx next = next_active_insn (next_active_insn (dest));
|
5011 |
|
|
if (next && JUMP_P (next)
|
5012 |
|
|
&& GET_CODE (PATTERN (next)) == SET
|
5013 |
|
|
&& recog_memoized (next) == CODE_FOR_jump_compact)
|
5014 |
|
|
{
|
5015 |
|
|
dest = JUMP_LABEL (next);
|
5016 |
|
|
if (dest
|
5017 |
|
|
&& (INSN_ADDRESSES (INSN_UID (dest)) - addr + (unsigned) 4092
|
5018 |
|
|
> 4092 + 4098))
|
5019 |
|
|
gen_block_redirect (next, INSN_ADDRESSES (INSN_UID (next)), -1);
|
5020 |
|
|
}
|
5021 |
|
|
}
|
5022 |
|
|
|
5023 |
|
|
if (dead)
|
5024 |
|
|
{
|
5025 |
|
|
rtx reg = gen_rtx_REG (SImode, exact_log2 (dead & -dead));
|
5026 |
|
|
|
5027 |
|
|
/* It would be nice if we could convert the jump into an indirect
|
5028 |
|
|
jump / far branch right now, and thus exposing all constituent
|
5029 |
|
|
instructions to further optimization. However, reorg uses
|
5030 |
|
|
simplejump_p to determine if there is an unconditional jump where
|
5031 |
|
|
it should try to schedule instructions from the target of the
|
5032 |
|
|
branch; simplejump_p fails for indirect jumps even if they have
|
5033 |
|
|
a JUMP_LABEL. */
|
5034 |
|
|
rtx insn = emit_insn_before (gen_indirect_jump_scratch
|
5035 |
|
|
(reg, GEN_INT (unspec_bbr_uid++)),
|
5036 |
|
|
jump);
|
5037 |
|
|
/* ??? We would like this to have the scope of the jump, but that
|
5038 |
|
|
scope will change when a delay slot insn of an inner scope is added.
|
5039 |
|
|
Hence, after delay slot scheduling, we'll have to expect
|
5040 |
|
|
NOTE_INSN_BLOCK_END notes between the indirect_jump_scratch and
|
5041 |
|
|
the jump. */
|
5042 |
|
|
|
5043 |
|
|
INSN_LOCATOR (insn) = INSN_LOCATOR (jump);
|
5044 |
|
|
INSN_CODE (insn) = CODE_FOR_indirect_jump_scratch;
|
5045 |
|
|
return insn;
|
5046 |
|
|
}
|
5047 |
|
|
else if (need_block)
|
5048 |
|
|
/* We can't use JUMP_LABEL here because it might be undefined
|
5049 |
|
|
when not optimizing. */
|
5050 |
|
|
return emit_insn_before (gen_block_branch_redirect
|
5051 |
|
|
(GEN_INT (unspec_bbr_uid++)),
|
5052 |
|
|
jump);
|
5053 |
|
|
return prev;
|
5054 |
|
|
}
|
5055 |
|
|
|
5056 |
|
|
#define CONDJUMP_MIN -252
|
5057 |
|
|
#define CONDJUMP_MAX 262
|
5058 |
|
|
struct far_branch
|
5059 |
|
|
{
|
5060 |
|
|
/* A label (to be placed) in front of the jump
|
5061 |
|
|
that jumps to our ultimate destination. */
|
5062 |
|
|
rtx near_label;
|
5063 |
|
|
/* Where we are going to insert it if we cannot move the jump any farther,
|
5064 |
|
|
or the jump itself if we have picked up an existing jump. */
|
5065 |
|
|
rtx insert_place;
|
5066 |
|
|
/* The ultimate destination. */
|
5067 |
|
|
rtx far_label;
|
5068 |
|
|
struct far_branch *prev;
|
5069 |
|
|
/* If the branch has already been created, its address;
|
5070 |
|
|
else the address of its first prospective user. */
|
5071 |
|
|
int address;
|
5072 |
|
|
};
|
5073 |
|
|
|
5074 |
|
|
static void gen_far_branch (struct far_branch *);
|
5075 |
|
|
enum mdep_reorg_phase_e mdep_reorg_phase;
|
5076 |
|
|
static void
|
5077 |
|
|
gen_far_branch (struct far_branch *bp)
|
5078 |
|
|
{
|
5079 |
|
|
rtx insn = bp->insert_place;
|
5080 |
|
|
rtx jump;
|
5081 |
|
|
rtx label = gen_label_rtx ();
|
5082 |
|
|
int ok;
|
5083 |
|
|
|
5084 |
|
|
emit_label_after (label, insn);
|
5085 |
|
|
if (bp->far_label)
|
5086 |
|
|
{
|
5087 |
|
|
jump = emit_jump_insn_after (gen_jump (bp->far_label), insn);
|
5088 |
|
|
LABEL_NUSES (bp->far_label)++;
|
5089 |
|
|
}
|
5090 |
|
|
else
|
5091 |
|
|
jump = emit_jump_insn_after (gen_return (), insn);
|
5092 |
|
|
/* Emit a barrier so that reorg knows that any following instructions
|
5093 |
|
|
are not reachable via a fall-through path.
|
5094 |
|
|
But don't do this when not optimizing, since we wouldn't suppress the
|
5095 |
|
|
alignment for the barrier then, and could end up with out-of-range
|
5096 |
|
|
pc-relative loads. */
|
5097 |
|
|
if (optimize)
|
5098 |
|
|
emit_barrier_after (jump);
|
5099 |
|
|
emit_label_after (bp->near_label, insn);
|
5100 |
|
|
JUMP_LABEL (jump) = bp->far_label;
|
5101 |
|
|
ok = invert_jump (insn, label, 1);
|
5102 |
|
|
gcc_assert (ok);
|
5103 |
|
|
|
5104 |
|
|
/* If we are branching around a jump (rather than a return), prevent
|
5105 |
|
|
reorg from using an insn from the jump target as the delay slot insn -
|
5106 |
|
|
when reorg did this, it pessimized code (we rather hide the delay slot)
|
5107 |
|
|
and it could cause branches to go out of range. */
|
5108 |
|
|
if (bp->far_label)
|
5109 |
|
|
(emit_insn_after
|
5110 |
|
|
(gen_stuff_delay_slot
|
5111 |
|
|
(GEN_INT (unspec_bbr_uid++),
|
5112 |
|
|
GEN_INT (recog_memoized (insn) == CODE_FOR_branch_false)),
|
5113 |
|
|
insn));
|
5114 |
|
|
/* Prevent reorg from undoing our splits. */
|
5115 |
|
|
gen_block_redirect (jump, bp->address += 2, 2);
|
5116 |
|
|
}
|
5117 |
|
|
|
5118 |
|
|
/* Fix up ADDR_DIFF_VECs. */
|
5119 |
|
|
void
|
5120 |
|
|
fixup_addr_diff_vecs (rtx first)
|
5121 |
|
|
{
|
5122 |
|
|
rtx insn;
|
5123 |
|
|
|
5124 |
|
|
for (insn = first; insn; insn = NEXT_INSN (insn))
|
5125 |
|
|
{
|
5126 |
|
|
rtx vec_lab, pat, prev, prevpat, x, braf_label;
|
5127 |
|
|
|
5128 |
|
|
if (!JUMP_P (insn)
|
5129 |
|
|
|| GET_CODE (PATTERN (insn)) != ADDR_DIFF_VEC)
|
5130 |
|
|
continue;
|
5131 |
|
|
pat = PATTERN (insn);
|
5132 |
|
|
vec_lab = XEXP (XEXP (pat, 0), 0);
|
5133 |
|
|
|
5134 |
|
|
/* Search the matching casesi_jump_2. */
|
5135 |
|
|
for (prev = vec_lab; ; prev = PREV_INSN (prev))
|
5136 |
|
|
{
|
5137 |
|
|
if (!JUMP_P (prev))
|
5138 |
|
|
continue;
|
5139 |
|
|
prevpat = PATTERN (prev);
|
5140 |
|
|
if (GET_CODE (prevpat) != PARALLEL || XVECLEN (prevpat, 0) != 2)
|
5141 |
|
|
continue;
|
5142 |
|
|
x = XVECEXP (prevpat, 0, 1);
|
5143 |
|
|
if (GET_CODE (x) != USE)
|
5144 |
|
|
continue;
|
5145 |
|
|
x = XEXP (x, 0);
|
5146 |
|
|
if (GET_CODE (x) == LABEL_REF && XEXP (x, 0) == vec_lab)
|
5147 |
|
|
break;
|
5148 |
|
|
}
|
5149 |
|
|
/* FIXME: This is a bug in the optimizer, but it seems harmless
|
5150 |
|
|
to just avoid panicing. */
|
5151 |
|
|
if (!prev)
|
5152 |
|
|
continue;
|
5153 |
|
|
|
5154 |
|
|
/* Emit the reference label of the braf where it belongs, right after
|
5155 |
|
|
the casesi_jump_2 (i.e. braf). */
|
5156 |
|
|
braf_label = XEXP (XEXP (SET_SRC (XVECEXP (prevpat, 0, 0)), 1), 0);
|
5157 |
|
|
emit_label_after (braf_label, prev);
|
5158 |
|
|
|
5159 |
|
|
/* Fix up the ADDR_DIF_VEC to be relative
|
5160 |
|
|
to the reference address of the braf. */
|
5161 |
|
|
XEXP (XEXP (pat, 0), 0) = braf_label;
|
5162 |
|
|
}
|
5163 |
|
|
}
|
5164 |
|
|
|
5165 |
|
|
/* BARRIER_OR_LABEL is either a BARRIER or a CODE_LABEL immediately following
|
5166 |
|
|
a barrier. Return the base 2 logarithm of the desired alignment. */
|
5167 |
|
|
int
|
5168 |
|
|
barrier_align (rtx barrier_or_label)
|
5169 |
|
|
{
|
5170 |
|
|
rtx next = next_real_insn (barrier_or_label), pat, prev;
|
5171 |
|
|
int slot, credit, jump_to_next = 0;
|
5172 |
|
|
|
5173 |
|
|
if (! next)
|
5174 |
|
|
return 0;
|
5175 |
|
|
|
5176 |
|
|
pat = PATTERN (next);
|
5177 |
|
|
|
5178 |
|
|
if (GET_CODE (pat) == ADDR_DIFF_VEC)
|
5179 |
|
|
return 2;
|
5180 |
|
|
|
5181 |
|
|
if (GET_CODE (pat) == UNSPEC_VOLATILE && XINT (pat, 1) == UNSPECV_ALIGN)
|
5182 |
|
|
/* This is a barrier in front of a constant table. */
|
5183 |
|
|
return 0;
|
5184 |
|
|
|
5185 |
|
|
prev = prev_real_insn (barrier_or_label);
|
5186 |
|
|
if (GET_CODE (PATTERN (prev)) == ADDR_DIFF_VEC)
|
5187 |
|
|
{
|
5188 |
|
|
pat = PATTERN (prev);
|
5189 |
|
|
/* If this is a very small table, we want to keep the alignment after
|
5190 |
|
|
the table to the minimum for proper code alignment. */
|
5191 |
|
|
return ((TARGET_SMALLCODE
|
5192 |
|
|
|| ((unsigned) XVECLEN (pat, 1) * GET_MODE_SIZE (GET_MODE (pat))
|
5193 |
|
|
<= (unsigned) 1 << (CACHE_LOG - 2)))
|
5194 |
|
|
? 1 << TARGET_SHMEDIA : align_jumps_log);
|
5195 |
|
|
}
|
5196 |
|
|
|
5197 |
|
|
if (TARGET_SMALLCODE)
|
5198 |
|
|
return 0;
|
5199 |
|
|
|
5200 |
|
|
if (! TARGET_SH2 || ! optimize)
|
5201 |
|
|
return align_jumps_log;
|
5202 |
|
|
|
5203 |
|
|
/* When fixing up pcloads, a constant table might be inserted just before
|
5204 |
|
|
the basic block that ends with the barrier. Thus, we can't trust the
|
5205 |
|
|
instruction lengths before that. */
|
5206 |
|
|
if (mdep_reorg_phase > SH_FIXUP_PCLOAD)
|
5207 |
|
|
{
|
5208 |
|
|
/* Check if there is an immediately preceding branch to the insn beyond
|
5209 |
|
|
the barrier. We must weight the cost of discarding useful information
|
5210 |
|
|
from the current cache line when executing this branch and there is
|
5211 |
|
|
an alignment, against that of fetching unneeded insn in front of the
|
5212 |
|
|
branch target when there is no alignment. */
|
5213 |
|
|
|
5214 |
|
|
/* There are two delay_slot cases to consider. One is the simple case
|
5215 |
|
|
where the preceding branch is to the insn beyond the barrier (simple
|
5216 |
|
|
delay slot filling), and the other is where the preceding branch has
|
5217 |
|
|
a delay slot that is a duplicate of the insn after the barrier
|
5218 |
|
|
(fill_eager_delay_slots) and the branch is to the insn after the insn
|
5219 |
|
|
after the barrier. */
|
5220 |
|
|
|
5221 |
|
|
/* PREV is presumed to be the JUMP_INSN for the barrier under
|
5222 |
|
|
investigation. Skip to the insn before it. */
|
5223 |
|
|
prev = prev_real_insn (prev);
|
5224 |
|
|
|
5225 |
|
|
for (slot = 2, credit = (1 << (CACHE_LOG - 2)) + 2;
|
5226 |
|
|
credit >= 0 && prev && NONJUMP_INSN_P (prev);
|
5227 |
|
|
prev = prev_real_insn (prev))
|
5228 |
|
|
{
|
5229 |
|
|
jump_to_next = 0;
|
5230 |
|
|
if (GET_CODE (PATTERN (prev)) == USE
|
5231 |
|
|
|| GET_CODE (PATTERN (prev)) == CLOBBER)
|
5232 |
|
|
continue;
|
5233 |
|
|
if (GET_CODE (PATTERN (prev)) == SEQUENCE)
|
5234 |
|
|
{
|
5235 |
|
|
prev = XVECEXP (PATTERN (prev), 0, 1);
|
5236 |
|
|
if (INSN_UID (prev) == INSN_UID (next))
|
5237 |
|
|
{
|
5238 |
|
|
/* Delay slot was filled with insn at jump target. */
|
5239 |
|
|
jump_to_next = 1;
|
5240 |
|
|
continue;
|
5241 |
|
|
}
|
5242 |
|
|
}
|
5243 |
|
|
|
5244 |
|
|
if (slot &&
|
5245 |
|
|
get_attr_in_delay_slot (prev) == IN_DELAY_SLOT_YES)
|
5246 |
|
|
slot = 0;
|
5247 |
|
|
credit -= get_attr_length (prev);
|
5248 |
|
|
}
|
5249 |
|
|
if (prev
|
5250 |
|
|
&& JUMP_P (prev)
|
5251 |
|
|
&& JUMP_LABEL (prev))
|
5252 |
|
|
{
|
5253 |
|
|
rtx x;
|
5254 |
|
|
if (jump_to_next
|
5255 |
|
|
|| next_real_insn (JUMP_LABEL (prev)) == next
|
5256 |
|
|
/* If relax_delay_slots() decides NEXT was redundant
|
5257 |
|
|
with some previous instruction, it will have
|
5258 |
|
|
redirected PREV's jump to the following insn. */
|
5259 |
|
|
|| JUMP_LABEL (prev) == next_nonnote_insn (next)
|
5260 |
|
|
/* There is no upper bound on redundant instructions
|
5261 |
|
|
that might have been skipped, but we must not put an
|
5262 |
|
|
alignment where none had been before. */
|
5263 |
|
|
|| (x = (NEXT_INSN (NEXT_INSN (PREV_INSN (prev)))),
|
5264 |
|
|
(INSN_P (x)
|
5265 |
|
|
&& (INSN_CODE (x) == CODE_FOR_block_branch_redirect
|
5266 |
|
|
|| INSN_CODE (x) == CODE_FOR_indirect_jump_scratch
|
5267 |
|
|
|| INSN_CODE (x) == CODE_FOR_stuff_delay_slot))))
|
5268 |
|
|
{
|
5269 |
|
|
rtx pat = PATTERN (prev);
|
5270 |
|
|
if (GET_CODE (pat) == PARALLEL)
|
5271 |
|
|
pat = XVECEXP (pat, 0, 0);
|
5272 |
|
|
if (credit - slot >= (GET_CODE (SET_SRC (pat)) == PC ? 2 : 0))
|
5273 |
|
|
return 0;
|
5274 |
|
|
}
|
5275 |
|
|
}
|
5276 |
|
|
}
|
5277 |
|
|
|
5278 |
|
|
return align_jumps_log;
|
5279 |
|
|
}
|
5280 |
|
|
|
5281 |
|
|
/* If we are inside a phony loop, almost any kind of label can turn up as the
|
5282 |
|
|
first one in the loop. Aligning a braf label causes incorrect switch
|
5283 |
|
|
destination addresses; we can detect braf labels because they are
|
5284 |
|
|
followed by a BARRIER.
|
5285 |
|
|
Applying loop alignment to small constant or switch tables is a waste
|
5286 |
|
|
of space, so we suppress this too. */
|
5287 |
|
|
int
|
5288 |
|
|
sh_loop_align (rtx label)
|
5289 |
|
|
{
|
5290 |
|
|
rtx next = label;
|
5291 |
|
|
|
5292 |
|
|
do
|
5293 |
|
|
next = next_nonnote_insn (next);
|
5294 |
|
|
while (next && LABEL_P (next));
|
5295 |
|
|
|
5296 |
|
|
if (! next
|
5297 |
|
|
|| ! INSN_P (next)
|
5298 |
|
|
|| GET_CODE (PATTERN (next)) == ADDR_DIFF_VEC
|
5299 |
|
|
|| recog_memoized (next) == CODE_FOR_consttable_2)
|
5300 |
|
|
return 0;
|
5301 |
|
|
|
5302 |
|
|
return align_loops_log;
|
5303 |
|
|
}
|
5304 |
|
|
|
5305 |
|
|
/* Do a final pass over the function, just before delayed branch
|
5306 |
|
|
scheduling. */
|
5307 |
|
|
|
5308 |
|
|
static void
|
5309 |
|
|
sh_reorg (void)
|
5310 |
|
|
{
|
5311 |
|
|
rtx first, insn, mova = NULL_RTX;
|
5312 |
|
|
int num_mova;
|
5313 |
|
|
rtx r0_rtx = gen_rtx_REG (Pmode, 0);
|
5314 |
|
|
rtx r0_inc_rtx = gen_rtx_POST_INC (Pmode, r0_rtx);
|
5315 |
|
|
|
5316 |
|
|
first = get_insns ();
|
5317 |
|
|
max_labelno_before_reorg = max_label_num ();
|
5318 |
|
|
|
5319 |
|
|
/* We must split call insns before introducing `mova's. If we're
|
5320 |
|
|
optimizing, they'll have already been split. Otherwise, make
|
5321 |
|
|
sure we don't split them too late. */
|
5322 |
|
|
if (! optimize)
|
5323 |
|
|
split_all_insns_noflow ();
|
5324 |
|
|
|
5325 |
|
|
if (TARGET_SHMEDIA)
|
5326 |
|
|
return;
|
5327 |
|
|
|
5328 |
|
|
/* If relaxing, generate pseudo-ops to associate function calls with
|
5329 |
|
|
the symbols they call. It does no harm to not generate these
|
5330 |
|
|
pseudo-ops. However, when we can generate them, it enables to
|
5331 |
|
|
linker to potentially relax the jsr to a bsr, and eliminate the
|
5332 |
|
|
register load and, possibly, the constant pool entry. */
|
5333 |
|
|
|
5334 |
|
|
mdep_reorg_phase = SH_INSERT_USES_LABELS;
|
5335 |
|
|
if (TARGET_RELAX)
|
5336 |
|
|
{
|
5337 |
|
|
/* Remove all REG_LABEL_OPERAND notes. We want to use them for our
|
5338 |
|
|
own purposes. This works because none of the remaining passes
|
5339 |
|
|
need to look at them.
|
5340 |
|
|
|
5341 |
|
|
??? But it may break in the future. We should use a machine
|
5342 |
|
|
dependent REG_NOTE, or some other approach entirely. */
|
5343 |
|
|
for (insn = first; insn; insn = NEXT_INSN (insn))
|
5344 |
|
|
{
|
5345 |
|
|
if (INSN_P (insn))
|
5346 |
|
|
{
|
5347 |
|
|
rtx note;
|
5348 |
|
|
|
5349 |
|
|
while ((note = find_reg_note (insn, REG_LABEL_OPERAND,
|
5350 |
|
|
NULL_RTX)) != 0)
|
5351 |
|
|
remove_note (insn, note);
|
5352 |
|
|
}
|
5353 |
|
|
}
|
5354 |
|
|
|
5355 |
|
|
for (insn = first; insn; insn = NEXT_INSN (insn))
|
5356 |
|
|
{
|
5357 |
|
|
rtx pattern, reg, link, set, scan, dies, label;
|
5358 |
|
|
int rescan = 0, foundinsn = 0;
|
5359 |
|
|
|
5360 |
|
|
if (CALL_P (insn))
|
5361 |
|
|
{
|
5362 |
|
|
pattern = PATTERN (insn);
|
5363 |
|
|
|
5364 |
|
|
if (GET_CODE (pattern) == PARALLEL)
|
5365 |
|
|
pattern = XVECEXP (pattern, 0, 0);
|
5366 |
|
|
if (GET_CODE (pattern) == SET)
|
5367 |
|
|
pattern = SET_SRC (pattern);
|
5368 |
|
|
|
5369 |
|
|
if (GET_CODE (pattern) != CALL
|
5370 |
|
|
|| !MEM_P (XEXP (pattern, 0)))
|
5371 |
|
|
continue;
|
5372 |
|
|
|
5373 |
|
|
reg = XEXP (XEXP (pattern, 0), 0);
|
5374 |
|
|
}
|
5375 |
|
|
else
|
5376 |
|
|
{
|
5377 |
|
|
reg = sfunc_uses_reg (insn);
|
5378 |
|
|
if (! reg)
|
5379 |
|
|
continue;
|
5380 |
|
|
}
|
5381 |
|
|
|
5382 |
|
|
if (!REG_P (reg))
|
5383 |
|
|
continue;
|
5384 |
|
|
|
5385 |
|
|
/* Try scanning backward to find where the register is set. */
|
5386 |
|
|
link = NULL;
|
5387 |
|
|
for (scan = PREV_INSN (insn);
|
5388 |
|
|
scan && !LABEL_P (scan);
|
5389 |
|
|
scan = PREV_INSN (scan))
|
5390 |
|
|
{
|
5391 |
|
|
if (! INSN_P (scan))
|
5392 |
|
|
continue;
|
5393 |
|
|
|
5394 |
|
|
if (! reg_mentioned_p (reg, scan))
|
5395 |
|
|
continue;
|
5396 |
|
|
|
5397 |
|
|
if (noncall_uses_reg (reg, scan, &set))
|
5398 |
|
|
break;
|
5399 |
|
|
|
5400 |
|
|
if (set)
|
5401 |
|
|
{
|
5402 |
|
|
link = scan;
|
5403 |
|
|
break;
|
5404 |
|
|
}
|
5405 |
|
|
}
|
5406 |
|
|
|
5407 |
|
|
if (! link)
|
5408 |
|
|
continue;
|
5409 |
|
|
|
5410 |
|
|
/* The register is set at LINK. */
|
5411 |
|
|
|
5412 |
|
|
/* We can only optimize the function call if the register is
|
5413 |
|
|
being set to a symbol. In theory, we could sometimes
|
5414 |
|
|
optimize calls to a constant location, but the assembler
|
5415 |
|
|
and linker do not support that at present. */
|
5416 |
|
|
if (GET_CODE (SET_SRC (set)) != SYMBOL_REF
|
5417 |
|
|
&& GET_CODE (SET_SRC (set)) != LABEL_REF)
|
5418 |
|
|
continue;
|
5419 |
|
|
|
5420 |
|
|
/* Scan forward from LINK to the place where REG dies, and
|
5421 |
|
|
make sure that the only insns which use REG are
|
5422 |
|
|
themselves function calls. */
|
5423 |
|
|
|
5424 |
|
|
/* ??? This doesn't work for call targets that were allocated
|
5425 |
|
|
by reload, since there may not be a REG_DEAD note for the
|
5426 |
|
|
register. */
|
5427 |
|
|
|
5428 |
|
|
dies = NULL_RTX;
|
5429 |
|
|
for (scan = NEXT_INSN (link); scan; scan = NEXT_INSN (scan))
|
5430 |
|
|
{
|
5431 |
|
|
rtx scanset;
|
5432 |
|
|
|
5433 |
|
|
/* Don't try to trace forward past a CODE_LABEL if we haven't
|
5434 |
|
|
seen INSN yet. Ordinarily, we will only find the setting insn
|
5435 |
|
|
if it is in the same basic block. However,
|
5436 |
|
|
cross-jumping can insert code labels in between the load and
|
5437 |
|
|
the call, and can result in situations where a single call
|
5438 |
|
|
insn may have two targets depending on where we came from. */
|
5439 |
|
|
|
5440 |
|
|
if (LABEL_P (scan) && ! foundinsn)
|
5441 |
|
|
break;
|
5442 |
|
|
|
5443 |
|
|
if (! INSN_P (scan))
|
5444 |
|
|
continue;
|
5445 |
|
|
|
5446 |
|
|
/* Don't try to trace forward past a JUMP. To optimize
|
5447 |
|
|
safely, we would have to check that all the
|
5448 |
|
|
instructions at the jump destination did not use REG. */
|
5449 |
|
|
|
5450 |
|
|
if (JUMP_P (scan))
|
5451 |
|
|
break;
|
5452 |
|
|
|
5453 |
|
|
if (! reg_mentioned_p (reg, scan))
|
5454 |
|
|
continue;
|
5455 |
|
|
|
5456 |
|
|
if (noncall_uses_reg (reg, scan, &scanset))
|
5457 |
|
|
break;
|
5458 |
|
|
|
5459 |
|
|
if (scan == insn)
|
5460 |
|
|
foundinsn = 1;
|
5461 |
|
|
|
5462 |
|
|
if (scan != insn
|
5463 |
|
|
&& (CALL_P (scan) || sfunc_uses_reg (scan)))
|
5464 |
|
|
{
|
5465 |
|
|
/* There is a function call to this register other
|
5466 |
|
|
than the one we are checking. If we optimize
|
5467 |
|
|
this call, we need to rescan again below. */
|
5468 |
|
|
rescan = 1;
|
5469 |
|
|
}
|
5470 |
|
|
|
5471 |
|
|
/* ??? We shouldn't have to worry about SCANSET here.
|
5472 |
|
|
We should just be able to check for a REG_DEAD note
|
5473 |
|
|
on a function call. However, the REG_DEAD notes are
|
5474 |
|
|
apparently not dependable around libcalls; c-torture
|
5475 |
|
|
execute/920501-2 is a test case. If SCANSET is set,
|
5476 |
|
|
then this insn sets the register, so it must have
|
5477 |
|
|
died earlier. Unfortunately, this will only handle
|
5478 |
|
|
the cases in which the register is, in fact, set in a
|
5479 |
|
|
later insn. */
|
5480 |
|
|
|
5481 |
|
|
/* ??? We shouldn't have to use FOUNDINSN here.
|
5482 |
|
|
This dates back to when we used LOG_LINKS to find
|
5483 |
|
|
the most recent insn which sets the register. */
|
5484 |
|
|
|
5485 |
|
|
if (foundinsn
|
5486 |
|
|
&& (scanset
|
5487 |
|
|
|| find_reg_note (scan, REG_DEAD, reg)))
|
5488 |
|
|
{
|
5489 |
|
|
dies = scan;
|
5490 |
|
|
break;
|
5491 |
|
|
}
|
5492 |
|
|
}
|
5493 |
|
|
|
5494 |
|
|
if (! dies)
|
5495 |
|
|
{
|
5496 |
|
|
/* Either there was a branch, or some insn used REG
|
5497 |
|
|
other than as a function call address. */
|
5498 |
|
|
continue;
|
5499 |
|
|
}
|
5500 |
|
|
|
5501 |
|
|
/* Create a code label, and put it in a REG_LABEL_OPERAND note
|
5502 |
|
|
on the insn which sets the register, and on each call insn
|
5503 |
|
|
which uses the register. In final_prescan_insn we look for
|
5504 |
|
|
the REG_LABEL_OPERAND notes, and output the appropriate label
|
5505 |
|
|
or pseudo-op. */
|
5506 |
|
|
|
5507 |
|
|
label = gen_label_rtx ();
|
5508 |
|
|
add_reg_note (link, REG_LABEL_OPERAND, label);
|
5509 |
|
|
add_reg_note (insn, REG_LABEL_OPERAND, label);
|
5510 |
|
|
if (rescan)
|
5511 |
|
|
{
|
5512 |
|
|
scan = link;
|
5513 |
|
|
do
|
5514 |
|
|
{
|
5515 |
|
|
rtx reg2;
|
5516 |
|
|
|
5517 |
|
|
scan = NEXT_INSN (scan);
|
5518 |
|
|
if (scan != insn
|
5519 |
|
|
&& ((CALL_P (scan)
|
5520 |
|
|
&& reg_mentioned_p (reg, scan))
|
5521 |
|
|
|| ((reg2 = sfunc_uses_reg (scan))
|
5522 |
|
|
&& REGNO (reg2) == REGNO (reg))))
|
5523 |
|
|
add_reg_note (scan, REG_LABEL_OPERAND, label);
|
5524 |
|
|
}
|
5525 |
|
|
while (scan != dies);
|
5526 |
|
|
}
|
5527 |
|
|
}
|
5528 |
|
|
}
|
5529 |
|
|
|
5530 |
|
|
if (TARGET_SH2)
|
5531 |
|
|
fixup_addr_diff_vecs (first);
|
5532 |
|
|
|
5533 |
|
|
if (optimize)
|
5534 |
|
|
{
|
5535 |
|
|
mdep_reorg_phase = SH_SHORTEN_BRANCHES0;
|
5536 |
|
|
shorten_branches (first);
|
5537 |
|
|
}
|
5538 |
|
|
|
5539 |
|
|
/* Scan the function looking for move instructions which have to be
|
5540 |
|
|
changed to pc-relative loads and insert the literal tables. */
|
5541 |
|
|
label_ref_list_pool = create_alloc_pool ("label references list",
|
5542 |
|
|
sizeof (struct label_ref_list_d),
|
5543 |
|
|
30);
|
5544 |
|
|
mdep_reorg_phase = SH_FIXUP_PCLOAD;
|
5545 |
|
|
for (insn = first, num_mova = 0; insn; insn = NEXT_INSN (insn))
|
5546 |
|
|
{
|
5547 |
|
|
if (mova_p (insn))
|
5548 |
|
|
{
|
5549 |
|
|
/* ??? basic block reordering can move a switch table dispatch
|
5550 |
|
|
below the switch table. Check if that has happened.
|
5551 |
|
|
We only have the addresses available when optimizing; but then,
|
5552 |
|
|
this check shouldn't be needed when not optimizing. */
|
5553 |
|
|
if (!untangle_mova (&num_mova, &mova, insn))
|
5554 |
|
|
{
|
5555 |
|
|
insn = mova;
|
5556 |
|
|
num_mova = 0;
|
5557 |
|
|
}
|
5558 |
|
|
}
|
5559 |
|
|
else if (JUMP_P (insn)
|
5560 |
|
|
&& GET_CODE (PATTERN (insn)) == ADDR_DIFF_VEC
|
5561 |
|
|
&& num_mova
|
5562 |
|
|
/* ??? loop invariant motion can also move a mova out of a
|
5563 |
|
|
loop. Since loop does this code motion anyway, maybe we
|
5564 |
|
|
should wrap UNSPEC_MOVA into a CONST, so that reload can
|
5565 |
|
|
move it back. */
|
5566 |
|
|
&& ((num_mova > 1
|
5567 |
|
|
&& GET_MODE (prev_nonnote_insn (insn)) == VOIDmode)
|
5568 |
|
|
|| (prev_nonnote_insn (insn)
|
5569 |
|
|
== XEXP (MOVA_LABELREF (mova), 0))))
|
5570 |
|
|
{
|
5571 |
|
|
rtx scan;
|
5572 |
|
|
int total;
|
5573 |
|
|
|
5574 |
|
|
num_mova--;
|
5575 |
|
|
|
5576 |
|
|
/* Some code might have been inserted between the mova and
|
5577 |
|
|
its ADDR_DIFF_VEC. Check if the mova is still in range. */
|
5578 |
|
|
for (scan = mova, total = 0; scan != insn; scan = NEXT_INSN (scan))
|
5579 |
|
|
total += get_attr_length (scan);
|
5580 |
|
|
|
5581 |
|
|
/* range of mova is 1020, add 4 because pc counts from address of
|
5582 |
|
|
second instruction after this one, subtract 2 in case pc is 2
|
5583 |
|
|
byte aligned. Possible alignment needed for the ADDR_DIFF_VEC
|
5584 |
|
|
cancels out with alignment effects of the mova itself. */
|
5585 |
|
|
if (total > 1022)
|
5586 |
|
|
{
|
5587 |
|
|
/* Change the mova into a load, and restart scanning
|
5588 |
|
|
there. broken_move will then return true for mova. */
|
5589 |
|
|
fixup_mova (mova);
|
5590 |
|
|
insn = mova;
|
5591 |
|
|
}
|
5592 |
|
|
}
|
5593 |
|
|
if (broken_move (insn)
|
5594 |
|
|
|| (NONJUMP_INSN_P (insn)
|
5595 |
|
|
&& recog_memoized (insn) == CODE_FOR_casesi_worker_2))
|
5596 |
|
|
{
|
5597 |
|
|
rtx scan;
|
5598 |
|
|
/* Scan ahead looking for a barrier to stick the constant table
|
5599 |
|
|
behind. */
|
5600 |
|
|
rtx barrier = find_barrier (num_mova, mova, insn);
|
5601 |
|
|
rtx last_float_move = NULL_RTX, last_float = 0, *last_float_addr = NULL;
|
5602 |
|
|
int need_aligned_label = 0;
|
5603 |
|
|
|
5604 |
|
|
if (num_mova && ! mova_p (mova))
|
5605 |
|
|
{
|
5606 |
|
|
/* find_barrier had to change the first mova into a
|
5607 |
|
|
pcload; thus, we have to start with this new pcload. */
|
5608 |
|
|
insn = mova;
|
5609 |
|
|
num_mova = 0;
|
5610 |
|
|
}
|
5611 |
|
|
/* Now find all the moves between the points and modify them. */
|
5612 |
|
|
for (scan = insn; scan != barrier; scan = NEXT_INSN (scan))
|
5613 |
|
|
{
|
5614 |
|
|
if (LABEL_P (scan))
|
5615 |
|
|
last_float = 0;
|
5616 |
|
|
if (NONJUMP_INSN_P (scan)
|
5617 |
|
|
&& recog_memoized (scan) == CODE_FOR_casesi_worker_2)
|
5618 |
|
|
need_aligned_label = 1;
|
5619 |
|
|
if (broken_move (scan))
|
5620 |
|
|
{
|
5621 |
|
|
rtx *patp = &PATTERN (scan), pat = *patp;
|
5622 |
|
|
rtx src, dst;
|
5623 |
|
|
rtx lab;
|
5624 |
|
|
rtx newsrc;
|
5625 |
|
|
enum machine_mode mode;
|
5626 |
|
|
|
5627 |
|
|
if (GET_CODE (pat) == PARALLEL)
|
5628 |
|
|
patp = &XVECEXP (pat, 0, 0), pat = *patp;
|
5629 |
|
|
src = SET_SRC (pat);
|
5630 |
|
|
dst = SET_DEST (pat);
|
5631 |
|
|
mode = GET_MODE (dst);
|
5632 |
|
|
|
5633 |
|
|
if (mode == SImode && hi_const (src)
|
5634 |
|
|
&& REGNO (dst) != FPUL_REG)
|
5635 |
|
|
{
|
5636 |
|
|
int offset = 0;
|
5637 |
|
|
|
5638 |
|
|
mode = HImode;
|
5639 |
|
|
while (GET_CODE (dst) == SUBREG)
|
5640 |
|
|
{
|
5641 |
|
|
offset += subreg_regno_offset (REGNO (SUBREG_REG (dst)),
|
5642 |
|
|
GET_MODE (SUBREG_REG (dst)),
|
5643 |
|
|
SUBREG_BYTE (dst),
|
5644 |
|
|
GET_MODE (dst));
|
5645 |
|
|
dst = SUBREG_REG (dst);
|
5646 |
|
|
}
|
5647 |
|
|
dst = gen_rtx_REG (HImode, REGNO (dst) + offset);
|
5648 |
|
|
}
|
5649 |
|
|
if (REG_P (dst) && FP_ANY_REGISTER_P (REGNO (dst)))
|
5650 |
|
|
{
|
5651 |
|
|
/* This must be an insn that clobbers r0. */
|
5652 |
|
|
rtx *clobberp = &XVECEXP (PATTERN (scan), 0,
|
5653 |
|
|
XVECLEN (PATTERN (scan), 0)
|
5654 |
|
|
- 1);
|
5655 |
|
|
rtx clobber = *clobberp;
|
5656 |
|
|
|
5657 |
|
|
gcc_assert (GET_CODE (clobber) == CLOBBER
|
5658 |
|
|
&& rtx_equal_p (XEXP (clobber, 0), r0_rtx));
|
5659 |
|
|
|
5660 |
|
|
if (last_float
|
5661 |
|
|
&& reg_set_between_p (r0_rtx, last_float_move, scan))
|
5662 |
|
|
last_float = 0;
|
5663 |
|
|
if (last_float
|
5664 |
|
|
&& TARGET_SHCOMPACT
|
5665 |
|
|
&& GET_MODE_SIZE (mode) != 4
|
5666 |
|
|
&& GET_MODE_SIZE (GET_MODE (last_float)) == 4)
|
5667 |
|
|
last_float = 0;
|
5668 |
|
|
lab = add_constant (src, mode, last_float);
|
5669 |
|
|
if (lab)
|
5670 |
|
|
emit_insn_before (gen_mova (lab), scan);
|
5671 |
|
|
else
|
5672 |
|
|
{
|
5673 |
|
|
/* There will be a REG_UNUSED note for r0 on
|
5674 |
|
|
LAST_FLOAT_MOVE; we have to change it to REG_INC,
|
5675 |
|
|
lest reorg:mark_target_live_regs will not
|
5676 |
|
|
consider r0 to be used, and we end up with delay
|
5677 |
|
|
slot insn in front of SCAN that clobbers r0. */
|
5678 |
|
|
rtx note
|
5679 |
|
|
= find_regno_note (last_float_move, REG_UNUSED, 0);
|
5680 |
|
|
|
5681 |
|
|
/* If we are not optimizing, then there may not be
|
5682 |
|
|
a note. */
|
5683 |
|
|
if (note)
|
5684 |
|
|
PUT_REG_NOTE_KIND (note, REG_INC);
|
5685 |
|
|
|
5686 |
|
|
*last_float_addr = r0_inc_rtx;
|
5687 |
|
|
}
|
5688 |
|
|
last_float_move = scan;
|
5689 |
|
|
last_float = src;
|
5690 |
|
|
newsrc = gen_const_mem (mode,
|
5691 |
|
|
(((TARGET_SH4 && ! TARGET_FMOVD)
|
5692 |
|
|
|| REGNO (dst) == FPUL_REG)
|
5693 |
|
|
? r0_inc_rtx
|
5694 |
|
|
: r0_rtx));
|
5695 |
|
|
last_float_addr = &XEXP (newsrc, 0);
|
5696 |
|
|
|
5697 |
|
|
/* Remove the clobber of r0. */
|
5698 |
|
|
*clobberp = gen_rtx_CLOBBER (GET_MODE (clobber),
|
5699 |
|
|
gen_rtx_SCRATCH (Pmode));
|
5700 |
|
|
}
|
5701 |
|
|
/* This is a mova needing a label. Create it. */
|
5702 |
|
|
else if (GET_CODE (src) == UNSPEC
|
5703 |
|
|
&& XINT (src, 1) == UNSPEC_MOVA
|
5704 |
|
|
&& GET_CODE (XVECEXP (src, 0, 0)) == CONST)
|
5705 |
|
|
{
|
5706 |
|
|
lab = add_constant (XVECEXP (src, 0, 0), mode, 0);
|
5707 |
|
|
newsrc = gen_rtx_LABEL_REF (VOIDmode, lab);
|
5708 |
|
|
newsrc = gen_rtx_UNSPEC (SImode,
|
5709 |
|
|
gen_rtvec (1, newsrc),
|
5710 |
|
|
UNSPEC_MOVA);
|
5711 |
|
|
}
|
5712 |
|
|
else
|
5713 |
|
|
{
|
5714 |
|
|
lab = add_constant (src, mode, 0);
|
5715 |
|
|
newsrc = gen_rtx_LABEL_REF (VOIDmode, lab);
|
5716 |
|
|
newsrc = gen_const_mem (mode, newsrc);
|
5717 |
|
|
}
|
5718 |
|
|
*patp = gen_rtx_SET (VOIDmode, dst, newsrc);
|
5719 |
|
|
INSN_CODE (scan) = -1;
|
5720 |
|
|
}
|
5721 |
|
|
}
|
5722 |
|
|
dump_table (need_aligned_label ? insn : 0, barrier);
|
5723 |
|
|
insn = barrier;
|
5724 |
|
|
}
|
5725 |
|
|
}
|
5726 |
|
|
free_alloc_pool (label_ref_list_pool);
|
5727 |
|
|
for (insn = first; insn; insn = NEXT_INSN (insn))
|
5728 |
|
|
PUT_MODE (insn, VOIDmode);
|
5729 |
|
|
|
5730 |
|
|
mdep_reorg_phase = SH_SHORTEN_BRANCHES1;
|
5731 |
|
|
INSN_ADDRESSES_FREE ();
|
5732 |
|
|
split_branches (first);
|
5733 |
|
|
|
5734 |
|
|
/* The INSN_REFERENCES_ARE_DELAYED in sh.h is problematic because it
|
5735 |
|
|
also has an effect on the register that holds the address of the sfunc.
|
5736 |
|
|
Insert an extra dummy insn in front of each sfunc that pretends to
|
5737 |
|
|
use this register. */
|
5738 |
|
|
if (flag_delayed_branch)
|
5739 |
|
|
{
|
5740 |
|
|
for (insn = first; insn; insn = NEXT_INSN (insn))
|
5741 |
|
|
{
|
5742 |
|
|
rtx reg = sfunc_uses_reg (insn);
|
5743 |
|
|
|
5744 |
|
|
if (! reg)
|
5745 |
|
|
continue;
|
5746 |
|
|
emit_insn_before (gen_use_sfunc_addr (reg), insn);
|
5747 |
|
|
}
|
5748 |
|
|
}
|
5749 |
|
|
#if 0
|
5750 |
|
|
/* fpscr is not actually a user variable, but we pretend it is for the
|
5751 |
|
|
sake of the previous optimization passes, since we want it handled like
|
5752 |
|
|
one. However, we don't have any debugging information for it, so turn
|
5753 |
|
|
it into a non-user variable now. */
|
5754 |
|
|
if (TARGET_SH4)
|
5755 |
|
|
REG_USERVAR_P (get_fpscr_rtx ()) = 0;
|
5756 |
|
|
#endif
|
5757 |
|
|
mdep_reorg_phase = SH_AFTER_MDEP_REORG;
|
5758 |
|
|
}
|
5759 |
|
|
|
5760 |
|
|
int
|
5761 |
|
|
get_dest_uid (rtx label, int max_uid)
|
5762 |
|
|
{
|
5763 |
|
|
rtx dest = next_real_insn (label);
|
5764 |
|
|
int dest_uid;
|
5765 |
|
|
if (! dest)
|
5766 |
|
|
/* This can happen for an undefined label. */
|
5767 |
|
|
return 0;
|
5768 |
|
|
dest_uid = INSN_UID (dest);
|
5769 |
|
|
/* If this is a newly created branch redirection blocking instruction,
|
5770 |
|
|
we cannot index the branch_uid or insn_addresses arrays with its
|
5771 |
|
|
uid. But then, we won't need to, because the actual destination is
|
5772 |
|
|
the following branch. */
|
5773 |
|
|
while (dest_uid >= max_uid)
|
5774 |
|
|
{
|
5775 |
|
|
dest = NEXT_INSN (dest);
|
5776 |
|
|
dest_uid = INSN_UID (dest);
|
5777 |
|
|
}
|
5778 |
|
|
if (JUMP_P (dest) && GET_CODE (PATTERN (dest)) == RETURN)
|
5779 |
|
|
return 0;
|
5780 |
|
|
return dest_uid;
|
5781 |
|
|
}
|
5782 |
|
|
|
5783 |
|
|
/* Split condbranches that are out of range. Also add clobbers for
|
5784 |
|
|
scratch registers that are needed in far jumps.
|
5785 |
|
|
We do this before delay slot scheduling, so that it can take our
|
5786 |
|
|
newly created instructions into account. It also allows us to
|
5787 |
|
|
find branches with common targets more easily. */
|
5788 |
|
|
|
5789 |
|
|
static void
|
5790 |
|
|
split_branches (rtx first)
|
5791 |
|
|
{
|
5792 |
|
|
rtx insn;
|
5793 |
|
|
struct far_branch **uid_branch, *far_branch_list = 0;
|
5794 |
|
|
int max_uid = get_max_uid ();
|
5795 |
|
|
int ok;
|
5796 |
|
|
|
5797 |
|
|
/* Find out which branches are out of range. */
|
5798 |
|
|
shorten_branches (first);
|
5799 |
|
|
|
5800 |
|
|
uid_branch = (struct far_branch **) alloca (max_uid * sizeof *uid_branch);
|
5801 |
|
|
memset ((char *) uid_branch, 0, max_uid * sizeof *uid_branch);
|
5802 |
|
|
|
5803 |
|
|
for (insn = first; insn; insn = NEXT_INSN (insn))
|
5804 |
|
|
if (! INSN_P (insn))
|
5805 |
|
|
continue;
|
5806 |
|
|
else if (INSN_DELETED_P (insn))
|
5807 |
|
|
{
|
5808 |
|
|
/* Shorten_branches would split this instruction again,
|
5809 |
|
|
so transform it into a note. */
|
5810 |
|
|
SET_INSN_DELETED (insn);
|
5811 |
|
|
}
|
5812 |
|
|
else if (JUMP_P (insn)
|
5813 |
|
|
/* Don't mess with ADDR_DIFF_VEC */
|
5814 |
|
|
&& (GET_CODE (PATTERN (insn)) == SET
|
5815 |
|
|
|| GET_CODE (PATTERN (insn)) == RETURN))
|
5816 |
|
|
{
|
5817 |
|
|
enum attr_type type = get_attr_type (insn);
|
5818 |
|
|
if (type == TYPE_CBRANCH)
|
5819 |
|
|
{
|
5820 |
|
|
rtx next, beyond;
|
5821 |
|
|
|
5822 |
|
|
if (get_attr_length (insn) > 4)
|
5823 |
|
|
{
|
5824 |
|
|
rtx src = SET_SRC (PATTERN (insn));
|
5825 |
|
|
rtx olabel = XEXP (XEXP (src, 1), 0);
|
5826 |
|
|
int addr = INSN_ADDRESSES (INSN_UID (insn));
|
5827 |
|
|
rtx label = 0;
|
5828 |
|
|
int dest_uid = get_dest_uid (olabel, max_uid);
|
5829 |
|
|
struct far_branch *bp = uid_branch[dest_uid];
|
5830 |
|
|
|
5831 |
|
|
/* redirect_jump needs a valid JUMP_LABEL, and it might delete
|
5832 |
|
|
the label if the LABEL_NUSES count drops to zero. There is
|
5833 |
|
|
always a jump_optimize pass that sets these values, but it
|
5834 |
|
|
proceeds to delete unreferenced code, and then if not
|
5835 |
|
|
optimizing, to un-delete the deleted instructions, thus
|
5836 |
|
|
leaving labels with too low uses counts. */
|
5837 |
|
|
if (! optimize)
|
5838 |
|
|
{
|
5839 |
|
|
JUMP_LABEL (insn) = olabel;
|
5840 |
|
|
LABEL_NUSES (olabel)++;
|
5841 |
|
|
}
|
5842 |
|
|
if (! bp)
|
5843 |
|
|
{
|
5844 |
|
|
bp = (struct far_branch *) alloca (sizeof *bp);
|
5845 |
|
|
uid_branch[dest_uid] = bp;
|
5846 |
|
|
bp->prev = far_branch_list;
|
5847 |
|
|
far_branch_list = bp;
|
5848 |
|
|
bp->far_label
|
5849 |
|
|
= XEXP (XEXP (SET_SRC (PATTERN (insn)), 1), 0);
|
5850 |
|
|
LABEL_NUSES (bp->far_label)++;
|
5851 |
|
|
}
|
5852 |
|
|
else
|
5853 |
|
|
{
|
5854 |
|
|
label = bp->near_label;
|
5855 |
|
|
if (! label && bp->address - addr >= CONDJUMP_MIN)
|
5856 |
|
|
{
|
5857 |
|
|
rtx block = bp->insert_place;
|
5858 |
|
|
|
5859 |
|
|
if (GET_CODE (PATTERN (block)) == RETURN)
|
5860 |
|
|
block = PREV_INSN (block);
|
5861 |
|
|
else
|
5862 |
|
|
block = gen_block_redirect (block,
|
5863 |
|
|
bp->address, 2);
|
5864 |
|
|
label = emit_label_after (gen_label_rtx (),
|
5865 |
|
|
PREV_INSN (block));
|
5866 |
|
|
bp->near_label = label;
|
5867 |
|
|
}
|
5868 |
|
|
else if (label && ! NEXT_INSN (label))
|
5869 |
|
|
{
|
5870 |
|
|
if (addr + 2 - bp->address <= CONDJUMP_MAX)
|
5871 |
|
|
bp->insert_place = insn;
|
5872 |
|
|
else
|
5873 |
|
|
gen_far_branch (bp);
|
5874 |
|
|
}
|
5875 |
|
|
}
|
5876 |
|
|
if (! label
|
5877 |
|
|
|| (NEXT_INSN (label) && bp->address - addr < CONDJUMP_MIN))
|
5878 |
|
|
{
|
5879 |
|
|
bp->near_label = label = gen_label_rtx ();
|
5880 |
|
|
bp->insert_place = insn;
|
5881 |
|
|
bp->address = addr;
|
5882 |
|
|
}
|
5883 |
|
|
ok = redirect_jump (insn, label, 0);
|
5884 |
|
|
gcc_assert (ok);
|
5885 |
|
|
}
|
5886 |
|
|
else
|
5887 |
|
|
{
|
5888 |
|
|
/* get_attr_length (insn) == 2 */
|
5889 |
|
|
/* Check if we have a pattern where reorg wants to redirect
|
5890 |
|
|
the branch to a label from an unconditional branch that
|
5891 |
|
|
is too far away. */
|
5892 |
|
|
/* We can't use JUMP_LABEL here because it might be undefined
|
5893 |
|
|
when not optimizing. */
|
5894 |
|
|
/* A syntax error might cause beyond to be NULL_RTX. */
|
5895 |
|
|
beyond
|
5896 |
|
|
= next_active_insn (XEXP (XEXP (SET_SRC (PATTERN (insn)), 1),
|
5897 |
|
|
0));
|
5898 |
|
|
|
5899 |
|
|
if (beyond
|
5900 |
|
|
&& (JUMP_P (beyond)
|
5901 |
|
|
|| ((beyond = next_active_insn (beyond))
|
5902 |
|
|
&& JUMP_P (beyond)))
|
5903 |
|
|
&& GET_CODE (PATTERN (beyond)) == SET
|
5904 |
|
|
&& recog_memoized (beyond) == CODE_FOR_jump_compact
|
5905 |
|
|
&& ((INSN_ADDRESSES
|
5906 |
|
|
(INSN_UID (XEXP (SET_SRC (PATTERN (beyond)), 0)))
|
5907 |
|
|
- INSN_ADDRESSES (INSN_UID (insn)) + (unsigned) 252)
|
5908 |
|
|
> 252 + 258 + 2))
|
5909 |
|
|
gen_block_redirect (beyond,
|
5910 |
|
|
INSN_ADDRESSES (INSN_UID (beyond)), 1);
|
5911 |
|
|
}
|
5912 |
|
|
|
5913 |
|
|
next = next_active_insn (insn);
|
5914 |
|
|
|
5915 |
|
|
if (next
|
5916 |
|
|
&& (JUMP_P (next)
|
5917 |
|
|
|| ((next = next_active_insn (next))
|
5918 |
|
|
&& JUMP_P (next)))
|
5919 |
|
|
&& GET_CODE (PATTERN (next)) == SET
|
5920 |
|
|
&& recog_memoized (next) == CODE_FOR_jump_compact
|
5921 |
|
|
&& ((INSN_ADDRESSES
|
5922 |
|
|
(INSN_UID (XEXP (SET_SRC (PATTERN (next)), 0)))
|
5923 |
|
|
- INSN_ADDRESSES (INSN_UID (insn)) + (unsigned) 252)
|
5924 |
|
|
> 252 + 258 + 2))
|
5925 |
|
|
gen_block_redirect (next, INSN_ADDRESSES (INSN_UID (next)), 1);
|
5926 |
|
|
}
|
5927 |
|
|
else if (type == TYPE_JUMP || type == TYPE_RETURN)
|
5928 |
|
|
{
|
5929 |
|
|
int addr = INSN_ADDRESSES (INSN_UID (insn));
|
5930 |
|
|
rtx far_label = 0;
|
5931 |
|
|
int dest_uid = 0;
|
5932 |
|
|
struct far_branch *bp;
|
5933 |
|
|
|
5934 |
|
|
if (type == TYPE_JUMP)
|
5935 |
|
|
{
|
5936 |
|
|
far_label = XEXP (SET_SRC (PATTERN (insn)), 0);
|
5937 |
|
|
dest_uid = get_dest_uid (far_label, max_uid);
|
5938 |
|
|
if (! dest_uid)
|
5939 |
|
|
{
|
5940 |
|
|
/* Parse errors can lead to labels outside
|
5941 |
|
|
the insn stream. */
|
5942 |
|
|
if (! NEXT_INSN (far_label))
|
5943 |
|
|
continue;
|
5944 |
|
|
|
5945 |
|
|
if (! optimize)
|
5946 |
|
|
{
|
5947 |
|
|
JUMP_LABEL (insn) = far_label;
|
5948 |
|
|
LABEL_NUSES (far_label)++;
|
5949 |
|
|
}
|
5950 |
|
|
redirect_jump (insn, NULL_RTX, 1);
|
5951 |
|
|
far_label = 0;
|
5952 |
|
|
}
|
5953 |
|
|
}
|
5954 |
|
|
bp = uid_branch[dest_uid];
|
5955 |
|
|
if (! bp)
|
5956 |
|
|
{
|
5957 |
|
|
bp = (struct far_branch *) alloca (sizeof *bp);
|
5958 |
|
|
uid_branch[dest_uid] = bp;
|
5959 |
|
|
bp->prev = far_branch_list;
|
5960 |
|
|
far_branch_list = bp;
|
5961 |
|
|
bp->near_label = 0;
|
5962 |
|
|
bp->far_label = far_label;
|
5963 |
|
|
if (far_label)
|
5964 |
|
|
LABEL_NUSES (far_label)++;
|
5965 |
|
|
}
|
5966 |
|
|
else if (bp->near_label && ! NEXT_INSN (bp->near_label))
|
5967 |
|
|
if (addr - bp->address <= CONDJUMP_MAX)
|
5968 |
|
|
emit_label_after (bp->near_label, PREV_INSN (insn));
|
5969 |
|
|
else
|
5970 |
|
|
{
|
5971 |
|
|
gen_far_branch (bp);
|
5972 |
|
|
bp->near_label = 0;
|
5973 |
|
|
}
|
5974 |
|
|
else
|
5975 |
|
|
bp->near_label = 0;
|
5976 |
|
|
bp->address = addr;
|
5977 |
|
|
bp->insert_place = insn;
|
5978 |
|
|
if (! far_label)
|
5979 |
|
|
emit_insn_before (gen_block_branch_redirect (const0_rtx), insn);
|
5980 |
|
|
else
|
5981 |
|
|
gen_block_redirect (insn, addr, bp->near_label ? 2 : 0);
|
5982 |
|
|
}
|
5983 |
|
|
}
|
5984 |
|
|
/* Generate all pending far branches,
|
5985 |
|
|
and free our references to the far labels. */
|
5986 |
|
|
while (far_branch_list)
|
5987 |
|
|
{
|
5988 |
|
|
if (far_branch_list->near_label
|
5989 |
|
|
&& ! NEXT_INSN (far_branch_list->near_label))
|
5990 |
|
|
gen_far_branch (far_branch_list);
|
5991 |
|
|
if (optimize
|
5992 |
|
|
&& far_branch_list->far_label
|
5993 |
|
|
&& ! --LABEL_NUSES (far_branch_list->far_label))
|
5994 |
|
|
delete_insn (far_branch_list->far_label);
|
5995 |
|
|
far_branch_list = far_branch_list->prev;
|
5996 |
|
|
}
|
5997 |
|
|
|
5998 |
|
|
/* Instruction length information is no longer valid due to the new
|
5999 |
|
|
instructions that have been generated. */
|
6000 |
|
|
init_insn_lengths ();
|
6001 |
|
|
}
|
6002 |
|
|
|
6003 |
|
|
/* Dump out instruction addresses, which is useful for debugging the
|
6004 |
|
|
constant pool table stuff.
|
6005 |
|
|
|
6006 |
|
|
If relaxing, output the label and pseudo-ops used to link together
|
6007 |
|
|
calls and the instruction which set the registers. */
|
6008 |
|
|
|
6009 |
|
|
/* ??? The addresses printed by this routine for insns are nonsense for
|
6010 |
|
|
insns which are inside of a sequence where none of the inner insns have
|
6011 |
|
|
variable length. This is because the second pass of shorten_branches
|
6012 |
|
|
does not bother to update them. */
|
6013 |
|
|
|
6014 |
|
|
void
|
6015 |
|
|
final_prescan_insn (rtx insn, rtx *opvec ATTRIBUTE_UNUSED,
|
6016 |
|
|
int noperands ATTRIBUTE_UNUSED)
|
6017 |
|
|
{
|
6018 |
|
|
if (TARGET_DUMPISIZE)
|
6019 |
|
|
fprintf (asm_out_file, "\n! at %04x\n", INSN_ADDRESSES (INSN_UID (insn)));
|
6020 |
|
|
|
6021 |
|
|
if (TARGET_RELAX)
|
6022 |
|
|
{
|
6023 |
|
|
rtx note;
|
6024 |
|
|
|
6025 |
|
|
note = find_reg_note (insn, REG_LABEL_OPERAND, NULL_RTX);
|
6026 |
|
|
if (note)
|
6027 |
|
|
{
|
6028 |
|
|
rtx pattern;
|
6029 |
|
|
|
6030 |
|
|
pattern = PATTERN (insn);
|
6031 |
|
|
if (GET_CODE (pattern) == PARALLEL)
|
6032 |
|
|
pattern = XVECEXP (pattern, 0, 0);
|
6033 |
|
|
switch (GET_CODE (pattern))
|
6034 |
|
|
{
|
6035 |
|
|
case SET:
|
6036 |
|
|
if (GET_CODE (SET_SRC (pattern)) != CALL
|
6037 |
|
|
&& get_attr_type (insn) != TYPE_SFUNC)
|
6038 |
|
|
{
|
6039 |
|
|
targetm.asm_out.internal_label
|
6040 |
|
|
(asm_out_file, "L", CODE_LABEL_NUMBER (XEXP (note, 0)));
|
6041 |
|
|
break;
|
6042 |
|
|
}
|
6043 |
|
|
/* else FALLTHROUGH */
|
6044 |
|
|
case CALL:
|
6045 |
|
|
asm_fprintf (asm_out_file, "\t.uses %LL%d\n",
|
6046 |
|
|
CODE_LABEL_NUMBER (XEXP (note, 0)));
|
6047 |
|
|
break;
|
6048 |
|
|
|
6049 |
|
|
default:
|
6050 |
|
|
gcc_unreachable ();
|
6051 |
|
|
}
|
6052 |
|
|
}
|
6053 |
|
|
}
|
6054 |
|
|
}
|
6055 |
|
|
|
6056 |
|
|
/* Dump out any constants accumulated in the final pass. These will
|
6057 |
|
|
only be labels. */
|
6058 |
|
|
|
6059 |
|
|
const char *
|
6060 |
|
|
output_jump_label_table (void)
|
6061 |
|
|
{
|
6062 |
|
|
int i;
|
6063 |
|
|
|
6064 |
|
|
if (pool_size)
|
6065 |
|
|
{
|
6066 |
|
|
fprintf (asm_out_file, "\t.align 2\n");
|
6067 |
|
|
for (i = 0; i < pool_size; i++)
|
6068 |
|
|
{
|
6069 |
|
|
pool_node *p = &pool_vector[i];
|
6070 |
|
|
|
6071 |
|
|
(*targetm.asm_out.internal_label) (asm_out_file, "L",
|
6072 |
|
|
CODE_LABEL_NUMBER (p->label));
|
6073 |
|
|
output_asm_insn (".long %O0", &p->value);
|
6074 |
|
|
}
|
6075 |
|
|
pool_size = 0;
|
6076 |
|
|
}
|
6077 |
|
|
|
6078 |
|
|
return "";
|
6079 |
|
|
}
|
6080 |
|
|
|
6081 |
|
|
/* A full frame looks like:
|
6082 |
|
|
|
6083 |
|
|
arg-5
|
6084 |
|
|
arg-4
|
6085 |
|
|
[ if current_function_anonymous_args
|
6086 |
|
|
arg-3
|
6087 |
|
|
arg-2
|
6088 |
|
|
arg-1
|
6089 |
|
|
arg-0 ]
|
6090 |
|
|
saved-fp
|
6091 |
|
|
saved-r10
|
6092 |
|
|
saved-r11
|
6093 |
|
|
saved-r12
|
6094 |
|
|
saved-pr
|
6095 |
|
|
local-n
|
6096 |
|
|
..
|
6097 |
|
|
local-1
|
6098 |
|
|
local-0 <- fp points here. */
|
6099 |
|
|
|
6100 |
|
|
/* Number of bytes pushed for anonymous args, used to pass information
|
6101 |
|
|
between expand_prologue and expand_epilogue. */
|
6102 |
|
|
|
6103 |
|
|
/* Adjust the stack by SIZE bytes. REG holds the rtl of the register to be
|
6104 |
|
|
adjusted. If epilogue_p is zero, this is for a prologue; otherwise, it's
|
6105 |
|
|
for an epilogue and a negative value means that it's for a sibcall
|
6106 |
|
|
epilogue. If LIVE_REGS_MASK is nonzero, it points to a HARD_REG_SET of
|
6107 |
|
|
all the registers that are about to be restored, and hence dead. */
|
6108 |
|
|
|
6109 |
|
|
static void
|
6110 |
|
|
output_stack_adjust (int size, rtx reg, int epilogue_p,
|
6111 |
|
|
HARD_REG_SET *live_regs_mask, bool frame_p)
|
6112 |
|
|
{
|
6113 |
|
|
rtx (*emit_fn) (rtx) = frame_p ? &frame_insn : &emit_insn;
|
6114 |
|
|
if (size)
|
6115 |
|
|
{
|
6116 |
|
|
HOST_WIDE_INT align = STACK_BOUNDARY / BITS_PER_UNIT;
|
6117 |
|
|
|
6118 |
|
|
/* This test is bogus, as output_stack_adjust is used to re-align the
|
6119 |
|
|
stack. */
|
6120 |
|
|
#if 0
|
6121 |
|
|
gcc_assert (!(size % align));
|
6122 |
|
|
#endif
|
6123 |
|
|
|
6124 |
|
|
if (CONST_OK_FOR_ADD (size))
|
6125 |
|
|
emit_fn (GEN_ADD3 (reg, reg, GEN_INT (size)));
|
6126 |
|
|
/* Try to do it with two partial adjustments; however, we must make
|
6127 |
|
|
sure that the stack is properly aligned at all times, in case
|
6128 |
|
|
an interrupt occurs between the two partial adjustments. */
|
6129 |
|
|
else if (CONST_OK_FOR_ADD (size / 2 & -align)
|
6130 |
|
|
&& CONST_OK_FOR_ADD (size - (size / 2 & -align)))
|
6131 |
|
|
{
|
6132 |
|
|
emit_fn (GEN_ADD3 (reg, reg, GEN_INT (size / 2 & -align)));
|
6133 |
|
|
emit_fn (GEN_ADD3 (reg, reg, GEN_INT (size - (size / 2 & -align))));
|
6134 |
|
|
}
|
6135 |
|
|
else
|
6136 |
|
|
{
|
6137 |
|
|
rtx const_reg;
|
6138 |
|
|
rtx insn;
|
6139 |
|
|
int temp = epilogue_p ? 7 : (TARGET_SH5 ? 0 : 1);
|
6140 |
|
|
int i;
|
6141 |
|
|
|
6142 |
|
|
/* If TEMP is invalid, we could temporarily save a general
|
6143 |
|
|
register to MACL. However, there is currently no need
|
6144 |
|
|
to handle this case, so just die when we see it. */
|
6145 |
|
|
if (epilogue_p < 0
|
6146 |
|
|
|| current_function_interrupt
|
6147 |
|
|
|| ! call_really_used_regs[temp] || fixed_regs[temp])
|
6148 |
|
|
temp = -1;
|
6149 |
|
|
if (temp < 0 && ! current_function_interrupt
|
6150 |
|
|
&& (TARGET_SHMEDIA || epilogue_p >= 0))
|
6151 |
|
|
{
|
6152 |
|
|
HARD_REG_SET temps;
|
6153 |
|
|
COPY_HARD_REG_SET (temps, call_used_reg_set);
|
6154 |
|
|
AND_COMPL_HARD_REG_SET (temps, call_fixed_reg_set);
|
6155 |
|
|
if (epilogue_p > 0)
|
6156 |
|
|
{
|
6157 |
|
|
int nreg = 0;
|
6158 |
|
|
if (crtl->return_rtx)
|
6159 |
|
|
{
|
6160 |
|
|
enum machine_mode mode;
|
6161 |
|
|
mode = GET_MODE (crtl->return_rtx);
|
6162 |
|
|
if (BASE_RETURN_VALUE_REG (mode) == FIRST_RET_REG)
|
6163 |
|
|
nreg = HARD_REGNO_NREGS (FIRST_RET_REG, mode);
|
6164 |
|
|
}
|
6165 |
|
|
for (i = 0; i < nreg; i++)
|
6166 |
|
|
CLEAR_HARD_REG_BIT (temps, FIRST_RET_REG + i);
|
6167 |
|
|
if (crtl->calls_eh_return)
|
6168 |
|
|
{
|
6169 |
|
|
CLEAR_HARD_REG_BIT (temps, EH_RETURN_STACKADJ_REGNO);
|
6170 |
|
|
for (i = 0; i <= 3; i++)
|
6171 |
|
|
CLEAR_HARD_REG_BIT (temps, EH_RETURN_DATA_REGNO (i));
|
6172 |
|
|
}
|
6173 |
|
|
}
|
6174 |
|
|
if (TARGET_SHMEDIA && epilogue_p < 0)
|
6175 |
|
|
for (i = FIRST_TARGET_REG; i <= LAST_TARGET_REG; i++)
|
6176 |
|
|
CLEAR_HARD_REG_BIT (temps, i);
|
6177 |
|
|
if (epilogue_p <= 0)
|
6178 |
|
|
{
|
6179 |
|
|
for (i = FIRST_PARM_REG;
|
6180 |
|
|
i < FIRST_PARM_REG + NPARM_REGS (SImode); i++)
|
6181 |
|
|
CLEAR_HARD_REG_BIT (temps, i);
|
6182 |
|
|
if (cfun->static_chain_decl != NULL)
|
6183 |
|
|
CLEAR_HARD_REG_BIT (temps, STATIC_CHAIN_REGNUM);
|
6184 |
|
|
}
|
6185 |
|
|
temp = scavenge_reg (&temps);
|
6186 |
|
|
}
|
6187 |
|
|
if (temp < 0 && live_regs_mask)
|
6188 |
|
|
{
|
6189 |
|
|
HARD_REG_SET temps;
|
6190 |
|
|
|
6191 |
|
|
COPY_HARD_REG_SET (temps, *live_regs_mask);
|
6192 |
|
|
CLEAR_HARD_REG_BIT (temps, REGNO (reg));
|
6193 |
|
|
temp = scavenge_reg (&temps);
|
6194 |
|
|
}
|
6195 |
|
|
if (temp < 0)
|
6196 |
|
|
{
|
6197 |
|
|
rtx adj_reg, tmp_reg, mem;
|
6198 |
|
|
|
6199 |
|
|
/* If we reached here, the most likely case is the (sibcall)
|
6200 |
|
|
epilogue for non SHmedia. Put a special push/pop sequence
|
6201 |
|
|
for such case as the last resort. This looks lengthy but
|
6202 |
|
|
would not be problem because it seems to be very
|
6203 |
|
|
rare. */
|
6204 |
|
|
|
6205 |
|
|
gcc_assert (!TARGET_SHMEDIA && epilogue_p);
|
6206 |
|
|
|
6207 |
|
|
|
6208 |
|
|
/* ??? There is still the slight possibility that r4 or
|
6209 |
|
|
r5 have been reserved as fixed registers or assigned
|
6210 |
|
|
as global registers, and they change during an
|
6211 |
|
|
interrupt. There are possible ways to handle this:
|
6212 |
|
|
|
6213 |
|
|
- If we are adjusting the frame pointer (r14), we can do
|
6214 |
|
|
with a single temp register and an ordinary push / pop
|
6215 |
|
|
on the stack.
|
6216 |
|
|
- Grab any call-used or call-saved registers (i.e. not
|
6217 |
|
|
fixed or globals) for the temps we need. We might
|
6218 |
|
|
also grab r14 if we are adjusting the stack pointer.
|
6219 |
|
|
If we can't find enough available registers, issue
|
6220 |
|
|
a diagnostic and die - the user must have reserved
|
6221 |
|
|
way too many registers.
|
6222 |
|
|
But since all this is rather unlikely to happen and
|
6223 |
|
|
would require extra testing, we just die if r4 / r5
|
6224 |
|
|
are not available. */
|
6225 |
|
|
gcc_assert (!fixed_regs[4] && !fixed_regs[5]
|
6226 |
|
|
&& !global_regs[4] && !global_regs[5]);
|
6227 |
|
|
|
6228 |
|
|
adj_reg = gen_rtx_REG (GET_MODE (reg), 4);
|
6229 |
|
|
tmp_reg = gen_rtx_REG (GET_MODE (reg), 5);
|
6230 |
|
|
emit_move_insn (gen_tmp_stack_mem (Pmode, reg), adj_reg);
|
6231 |
|
|
emit_insn (GEN_MOV (adj_reg, GEN_INT (size)));
|
6232 |
|
|
emit_insn (GEN_ADD3 (adj_reg, adj_reg, reg));
|
6233 |
|
|
mem = gen_tmp_stack_mem (Pmode, gen_rtx_PRE_DEC (Pmode, adj_reg));
|
6234 |
|
|
emit_move_insn (mem, tmp_reg);
|
6235 |
|
|
emit_move_insn (tmp_reg, gen_tmp_stack_mem (Pmode, reg));
|
6236 |
|
|
mem = gen_tmp_stack_mem (Pmode, gen_rtx_PRE_DEC (Pmode, adj_reg));
|
6237 |
|
|
emit_move_insn (mem, tmp_reg);
|
6238 |
|
|
emit_move_insn (reg, adj_reg);
|
6239 |
|
|
mem = gen_tmp_stack_mem (Pmode, gen_rtx_POST_INC (Pmode, reg));
|
6240 |
|
|
emit_move_insn (adj_reg, mem);
|
6241 |
|
|
mem = gen_tmp_stack_mem (Pmode, gen_rtx_POST_INC (Pmode, reg));
|
6242 |
|
|
emit_move_insn (tmp_reg, mem);
|
6243 |
|
|
/* Tell flow the insns that pop r4/r5 aren't dead. */
|
6244 |
|
|
emit_use (tmp_reg);
|
6245 |
|
|
emit_use (adj_reg);
|
6246 |
|
|
return;
|
6247 |
|
|
}
|
6248 |
|
|
const_reg = gen_rtx_REG (GET_MODE (reg), temp);
|
6249 |
|
|
|
6250 |
|
|
/* If SIZE is negative, subtract the positive value.
|
6251 |
|
|
This sometimes allows a constant pool entry to be shared
|
6252 |
|
|
between prologue and epilogue code. */
|
6253 |
|
|
if (size < 0)
|
6254 |
|
|
{
|
6255 |
|
|
emit_insn (GEN_MOV (const_reg, GEN_INT (-size)));
|
6256 |
|
|
insn = emit_fn (GEN_SUB3 (reg, reg, const_reg));
|
6257 |
|
|
}
|
6258 |
|
|
else
|
6259 |
|
|
{
|
6260 |
|
|
emit_insn (GEN_MOV (const_reg, GEN_INT (size)));
|
6261 |
|
|
insn = emit_fn (GEN_ADD3 (reg, reg, const_reg));
|
6262 |
|
|
}
|
6263 |
|
|
if (! epilogue_p)
|
6264 |
|
|
add_reg_note (insn, REG_FRAME_RELATED_EXPR,
|
6265 |
|
|
gen_rtx_SET (VOIDmode, reg,
|
6266 |
|
|
gen_rtx_PLUS (SImode, reg,
|
6267 |
|
|
GEN_INT (size))));
|
6268 |
|
|
}
|
6269 |
|
|
}
|
6270 |
|
|
}
|
6271 |
|
|
|
6272 |
|
|
static rtx
|
6273 |
|
|
frame_insn (rtx x)
|
6274 |
|
|
{
|
6275 |
|
|
x = emit_insn (x);
|
6276 |
|
|
RTX_FRAME_RELATED_P (x) = 1;
|
6277 |
|
|
return x;
|
6278 |
|
|
}
|
6279 |
|
|
|
6280 |
|
|
/* Output RTL to push register RN onto the stack. */
|
6281 |
|
|
|
6282 |
|
|
static rtx
|
6283 |
|
|
push (int rn)
|
6284 |
|
|
{
|
6285 |
|
|
rtx x;
|
6286 |
|
|
if (rn == FPUL_REG)
|
6287 |
|
|
x = gen_push_fpul ();
|
6288 |
|
|
else if (rn == FPSCR_REG)
|
6289 |
|
|
x = gen_push_fpscr ();
|
6290 |
|
|
else if ((TARGET_SH4 || TARGET_SH2A_DOUBLE) && TARGET_FMOVD && ! TARGET_FPU_SINGLE
|
6291 |
|
|
&& FP_OR_XD_REGISTER_P (rn))
|
6292 |
|
|
{
|
6293 |
|
|
if (FP_REGISTER_P (rn) && (rn - FIRST_FP_REG) & 1)
|
6294 |
|
|
return NULL_RTX;
|
6295 |
|
|
x = gen_push_4 (gen_rtx_REG (DFmode, rn));
|
6296 |
|
|
}
|
6297 |
|
|
else if (TARGET_SH2E && FP_REGISTER_P (rn))
|
6298 |
|
|
x = gen_push_e (gen_rtx_REG (SFmode, rn));
|
6299 |
|
|
else
|
6300 |
|
|
x = gen_push (gen_rtx_REG (SImode, rn));
|
6301 |
|
|
|
6302 |
|
|
x = frame_insn (x);
|
6303 |
|
|
add_reg_note (x, REG_INC, gen_rtx_REG (SImode, STACK_POINTER_REGNUM));
|
6304 |
|
|
return x;
|
6305 |
|
|
}
|
6306 |
|
|
|
6307 |
|
|
/* Output RTL to pop register RN from the stack. */
|
6308 |
|
|
|
6309 |
|
|
static void
|
6310 |
|
|
pop (int rn)
|
6311 |
|
|
{
|
6312 |
|
|
rtx x;
|
6313 |
|
|
if (rn == FPUL_REG)
|
6314 |
|
|
x = gen_pop_fpul ();
|
6315 |
|
|
else if (rn == FPSCR_REG)
|
6316 |
|
|
x = gen_pop_fpscr ();
|
6317 |
|
|
else if ((TARGET_SH4 || TARGET_SH2A_DOUBLE) && TARGET_FMOVD && ! TARGET_FPU_SINGLE
|
6318 |
|
|
&& FP_OR_XD_REGISTER_P (rn))
|
6319 |
|
|
{
|
6320 |
|
|
if (FP_REGISTER_P (rn) && (rn - FIRST_FP_REG) & 1)
|
6321 |
|
|
return;
|
6322 |
|
|
x = gen_pop_4 (gen_rtx_REG (DFmode, rn));
|
6323 |
|
|
}
|
6324 |
|
|
else if (TARGET_SH2E && FP_REGISTER_P (rn))
|
6325 |
|
|
x = gen_pop_e (gen_rtx_REG (SFmode, rn));
|
6326 |
|
|
else
|
6327 |
|
|
x = gen_pop (gen_rtx_REG (SImode, rn));
|
6328 |
|
|
|
6329 |
|
|
x = emit_insn (x);
|
6330 |
|
|
add_reg_note (x, REG_INC, gen_rtx_REG (SImode, STACK_POINTER_REGNUM));
|
6331 |
|
|
}
|
6332 |
|
|
|
6333 |
|
|
/* Generate code to push the regs specified in the mask. */
|
6334 |
|
|
|
6335 |
|
|
static void
|
6336 |
|
|
push_regs (HARD_REG_SET *mask, int interrupt_handler)
|
6337 |
|
|
{
|
6338 |
|
|
int i = interrupt_handler ? LAST_BANKED_REG + 1 : 0;
|
6339 |
|
|
int skip_fpscr = 0;
|
6340 |
|
|
|
6341 |
|
|
/* Push PR last; this gives better latencies after the prologue, and
|
6342 |
|
|
candidates for the return delay slot when there are no general
|
6343 |
|
|
registers pushed. */
|
6344 |
|
|
for (; i < FIRST_PSEUDO_REGISTER; i++)
|
6345 |
|
|
{
|
6346 |
|
|
/* If this is an interrupt handler, and the SZ bit varies,
|
6347 |
|
|
and we have to push any floating point register, we need
|
6348 |
|
|
to switch to the correct precision first. */
|
6349 |
|
|
if (i == FIRST_FP_REG && interrupt_handler && TARGET_FMOVD
|
6350 |
|
|
&& hard_reg_set_intersect_p (*mask, reg_class_contents[DF_REGS]))
|
6351 |
|
|
{
|
6352 |
|
|
HARD_REG_SET unsaved;
|
6353 |
|
|
|
6354 |
|
|
push (FPSCR_REG);
|
6355 |
|
|
COMPL_HARD_REG_SET (unsaved, *mask);
|
6356 |
|
|
fpscr_set_from_mem (NORMAL_MODE (FP_MODE), unsaved);
|
6357 |
|
|
skip_fpscr = 1;
|
6358 |
|
|
}
|
6359 |
|
|
if (i != PR_REG
|
6360 |
|
|
&& (i != FPSCR_REG || ! skip_fpscr)
|
6361 |
|
|
&& TEST_HARD_REG_BIT (*mask, i))
|
6362 |
|
|
{
|
6363 |
|
|
/* If the ISR has RESBANK attribute assigned, don't push any of
|
6364 |
|
|
the following registers - R0-R14, MACH, MACL and GBR. */
|
6365 |
|
|
if (! (sh_cfun_resbank_handler_p ()
|
6366 |
|
|
&& ((i >= FIRST_GENERAL_REG && i < LAST_GENERAL_REG)
|
6367 |
|
|
|| i == MACH_REG
|
6368 |
|
|
|| i == MACL_REG
|
6369 |
|
|
|| i == GBR_REG)))
|
6370 |
|
|
push (i);
|
6371 |
|
|
}
|
6372 |
|
|
}
|
6373 |
|
|
|
6374 |
|
|
/* Push banked registers last to improve delay slot opportunities. */
|
6375 |
|
|
if (interrupt_handler)
|
6376 |
|
|
for (i = FIRST_BANKED_REG; i <= LAST_BANKED_REG; i++)
|
6377 |
|
|
if (TEST_HARD_REG_BIT (*mask, i))
|
6378 |
|
|
push (i);
|
6379 |
|
|
|
6380 |
|
|
/* Don't push PR register for an ISR with RESBANK attribute assigned. */
|
6381 |
|
|
if (TEST_HARD_REG_BIT (*mask, PR_REG) && !sh_cfun_resbank_handler_p ())
|
6382 |
|
|
push (PR_REG);
|
6383 |
|
|
}
|
6384 |
|
|
|
6385 |
|
|
/* Calculate how much extra space is needed to save all callee-saved
|
6386 |
|
|
target registers.
|
6387 |
|
|
LIVE_REGS_MASK is the register mask calculated by calc_live_regs. */
|
6388 |
|
|
|
6389 |
|
|
static int
|
6390 |
|
|
shmedia_target_regs_stack_space (HARD_REG_SET *live_regs_mask)
|
6391 |
|
|
{
|
6392 |
|
|
int reg;
|
6393 |
|
|
int stack_space = 0;
|
6394 |
|
|
int interrupt_handler = sh_cfun_interrupt_handler_p ();
|
6395 |
|
|
|
6396 |
|
|
for (reg = LAST_TARGET_REG; reg >= FIRST_TARGET_REG; reg--)
|
6397 |
|
|
if ((! call_really_used_regs[reg] || interrupt_handler)
|
6398 |
|
|
&& ! TEST_HARD_REG_BIT (*live_regs_mask, reg))
|
6399 |
|
|
/* Leave space to save this target register on the stack,
|
6400 |
|
|
in case target register allocation wants to use it. */
|
6401 |
|
|
stack_space += GET_MODE_SIZE (REGISTER_NATURAL_MODE (reg));
|
6402 |
|
|
return stack_space;
|
6403 |
|
|
}
|
6404 |
|
|
|
6405 |
|
|
/* Decide whether we should reserve space for callee-save target registers,
|
6406 |
|
|
in case target register allocation wants to use them. REGS_SAVED is
|
6407 |
|
|
the space, in bytes, that is already required for register saves.
|
6408 |
|
|
LIVE_REGS_MASK is the register mask calculated by calc_live_regs. */
|
6409 |
|
|
|
6410 |
|
|
static int
|
6411 |
|
|
shmedia_reserve_space_for_target_registers_p (int regs_saved,
|
6412 |
|
|
HARD_REG_SET *live_regs_mask)
|
6413 |
|
|
{
|
6414 |
|
|
if (optimize_size)
|
6415 |
|
|
return 0;
|
6416 |
|
|
return shmedia_target_regs_stack_space (live_regs_mask) <= regs_saved;
|
6417 |
|
|
}
|
6418 |
|
|
|
6419 |
|
|
/* Decide how much space to reserve for callee-save target registers
|
6420 |
|
|
in case target register allocation wants to use them.
|
6421 |
|
|
LIVE_REGS_MASK is the register mask calculated by calc_live_regs. */
|
6422 |
|
|
|
6423 |
|
|
static int
|
6424 |
|
|
shmedia_target_regs_stack_adjust (HARD_REG_SET *live_regs_mask)
|
6425 |
|
|
{
|
6426 |
|
|
if (shmedia_space_reserved_for_target_registers)
|
6427 |
|
|
return shmedia_target_regs_stack_space (live_regs_mask);
|
6428 |
|
|
else
|
6429 |
|
|
return 0;
|
6430 |
|
|
}
|
6431 |
|
|
|
6432 |
|
|
/* Work out the registers which need to be saved, both as a mask and a
|
6433 |
|
|
count of saved words. Return the count.
|
6434 |
|
|
|
6435 |
|
|
If doing a pragma interrupt function, then push all regs used by the
|
6436 |
|
|
function, and if we call another function (we can tell by looking at PR),
|
6437 |
|
|
make sure that all the regs it clobbers are safe too. */
|
6438 |
|
|
|
6439 |
|
|
static int
|
6440 |
|
|
calc_live_regs (HARD_REG_SET *live_regs_mask)
|
6441 |
|
|
{
|
6442 |
|
|
unsigned int reg;
|
6443 |
|
|
int count;
|
6444 |
|
|
tree attrs;
|
6445 |
|
|
bool interrupt_or_trapa_handler, trapa_handler, interrupt_handler;
|
6446 |
|
|
bool nosave_low_regs;
|
6447 |
|
|
int pr_live, has_call;
|
6448 |
|
|
|
6449 |
|
|
attrs = DECL_ATTRIBUTES (current_function_decl);
|
6450 |
|
|
interrupt_or_trapa_handler = sh_cfun_interrupt_handler_p ();
|
6451 |
|
|
trapa_handler = lookup_attribute ("trapa_handler", attrs) != NULL_TREE;
|
6452 |
|
|
interrupt_handler = interrupt_or_trapa_handler && ! trapa_handler;
|
6453 |
|
|
nosave_low_regs = lookup_attribute ("nosave_low_regs", attrs) != NULL_TREE;
|
6454 |
|
|
|
6455 |
|
|
CLEAR_HARD_REG_SET (*live_regs_mask);
|
6456 |
|
|
if ((TARGET_SH4 || TARGET_SH2A_DOUBLE) && TARGET_FMOVD && interrupt_handler
|
6457 |
|
|
&& df_regs_ever_live_p (FPSCR_REG))
|
6458 |
|
|
target_flags &= ~MASK_FPU_SINGLE;
|
6459 |
|
|
/* If we can save a lot of saves by switching to double mode, do that. */
|
6460 |
|
|
else if ((TARGET_SH4 || TARGET_SH2A_DOUBLE) && TARGET_FMOVD && TARGET_FPU_SINGLE)
|
6461 |
|
|
for (count = 0, reg = FIRST_FP_REG; reg <= LAST_FP_REG; reg += 2)
|
6462 |
|
|
if (df_regs_ever_live_p (reg) && df_regs_ever_live_p (reg+1)
|
6463 |
|
|
&& (! call_really_used_regs[reg]
|
6464 |
|
|
|| interrupt_handler)
|
6465 |
|
|
&& ++count > 2)
|
6466 |
|
|
{
|
6467 |
|
|
target_flags &= ~MASK_FPU_SINGLE;
|
6468 |
|
|
break;
|
6469 |
|
|
}
|
6470 |
|
|
/* PR_MEDIA_REG is a general purpose register, thus global_alloc already
|
6471 |
|
|
knows how to use it. That means the pseudo originally allocated for
|
6472 |
|
|
the initial value can become the PR_MEDIA_REG hard register, as seen for
|
6473 |
|
|
execute/20010122-1.c:test9. */
|
6474 |
|
|
if (TARGET_SHMEDIA)
|
6475 |
|
|
/* ??? this function is called from initial_elimination_offset, hence we
|
6476 |
|
|
can't use the result of sh_media_register_for_return here. */
|
6477 |
|
|
pr_live = sh_pr_n_sets ();
|
6478 |
|
|
else
|
6479 |
|
|
{
|
6480 |
|
|
rtx pr_initial = has_hard_reg_initial_val (Pmode, PR_REG);
|
6481 |
|
|
pr_live = (pr_initial
|
6482 |
|
|
? (!REG_P (pr_initial)
|
6483 |
|
|
|| REGNO (pr_initial) != (PR_REG))
|
6484 |
|
|
: df_regs_ever_live_p (PR_REG));
|
6485 |
|
|
/* For Shcompact, if not optimizing, we end up with a memory reference
|
6486 |
|
|
using the return address pointer for __builtin_return_address even
|
6487 |
|
|
though there is no actual need to put the PR register on the stack. */
|
6488 |
|
|
pr_live |= df_regs_ever_live_p (RETURN_ADDRESS_POINTER_REGNUM);
|
6489 |
|
|
}
|
6490 |
|
|
/* Force PR to be live if the prologue has to call the SHmedia
|
6491 |
|
|
argument decoder or register saver. */
|
6492 |
|
|
if (TARGET_SHCOMPACT
|
6493 |
|
|
&& ((crtl->args.info.call_cookie
|
6494 |
|
|
& ~ CALL_COOKIE_RET_TRAMP (1))
|
6495 |
|
|
|| crtl->saves_all_registers))
|
6496 |
|
|
pr_live = 1;
|
6497 |
|
|
has_call = TARGET_SHMEDIA ? ! leaf_function_p () : pr_live;
|
6498 |
|
|
for (count = 0, reg = FIRST_PSEUDO_REGISTER; reg-- != 0; )
|
6499 |
|
|
{
|
6500 |
|
|
if (reg == (TARGET_SHMEDIA ? PR_MEDIA_REG : PR_REG)
|
6501 |
|
|
? pr_live
|
6502 |
|
|
: interrupt_handler
|
6503 |
|
|
? (/* Need to save all the regs ever live. */
|
6504 |
|
|
(df_regs_ever_live_p (reg)
|
6505 |
|
|
|| (call_really_used_regs[reg]
|
6506 |
|
|
&& (! fixed_regs[reg] || reg == MACH_REG || reg == MACL_REG
|
6507 |
|
|
|| reg == PIC_OFFSET_TABLE_REGNUM)
|
6508 |
|
|
&& has_call)
|
6509 |
|
|
|| (TARGET_SHMEDIA && has_call
|
6510 |
|
|
&& REGISTER_NATURAL_MODE (reg) == SImode
|
6511 |
|
|
&& (GENERAL_REGISTER_P (reg) || TARGET_REGISTER_P (reg))))
|
6512 |
|
|
&& reg != STACK_POINTER_REGNUM && reg != ARG_POINTER_REGNUM
|
6513 |
|
|
&& reg != RETURN_ADDRESS_POINTER_REGNUM
|
6514 |
|
|
&& reg != T_REG && reg != GBR_REG
|
6515 |
|
|
/* Push fpscr only on targets which have FPU */
|
6516 |
|
|
&& (reg != FPSCR_REG || TARGET_FPU_ANY))
|
6517 |
|
|
: (/* Only push those regs which are used and need to be saved. */
|
6518 |
|
|
(TARGET_SHCOMPACT
|
6519 |
|
|
&& flag_pic
|
6520 |
|
|
&& crtl->args.info.call_cookie
|
6521 |
|
|
&& reg == PIC_OFFSET_TABLE_REGNUM)
|
6522 |
|
|
|| (df_regs_ever_live_p (reg)
|
6523 |
|
|
&& ((!call_really_used_regs[reg]
|
6524 |
|
|
&& !(reg != PIC_OFFSET_TABLE_REGNUM
|
6525 |
|
|
&& fixed_regs[reg] && call_used_regs[reg]))
|
6526 |
|
|
|| (trapa_handler && reg == FPSCR_REG && TARGET_FPU_ANY)))
|
6527 |
|
|
|| (crtl->calls_eh_return
|
6528 |
|
|
&& (reg == EH_RETURN_DATA_REGNO (0)
|
6529 |
|
|
|| reg == EH_RETURN_DATA_REGNO (1)
|
6530 |
|
|
|| reg == EH_RETURN_DATA_REGNO (2)
|
6531 |
|
|
|| reg == EH_RETURN_DATA_REGNO (3)))
|
6532 |
|
|
|| ((reg == MACL_REG || reg == MACH_REG)
|
6533 |
|
|
&& df_regs_ever_live_p (reg)
|
6534 |
|
|
&& sh_cfun_attr_renesas_p ())
|
6535 |
|
|
))
|
6536 |
|
|
{
|
6537 |
|
|
SET_HARD_REG_BIT (*live_regs_mask, reg);
|
6538 |
|
|
count += GET_MODE_SIZE (REGISTER_NATURAL_MODE (reg));
|
6539 |
|
|
|
6540 |
|
|
if ((TARGET_SH4 || TARGET_SH2A_DOUBLE || TARGET_SH5) && TARGET_FMOVD
|
6541 |
|
|
&& GET_MODE_CLASS (REGISTER_NATURAL_MODE (reg)) == MODE_FLOAT)
|
6542 |
|
|
{
|
6543 |
|
|
if (FP_REGISTER_P (reg))
|
6544 |
|
|
{
|
6545 |
|
|
if (! TARGET_FPU_SINGLE && ! df_regs_ever_live_p (reg ^ 1))
|
6546 |
|
|
{
|
6547 |
|
|
SET_HARD_REG_BIT (*live_regs_mask, (reg ^ 1));
|
6548 |
|
|
count += GET_MODE_SIZE (REGISTER_NATURAL_MODE (reg ^ 1));
|
6549 |
|
|
}
|
6550 |
|
|
}
|
6551 |
|
|
else if (XD_REGISTER_P (reg))
|
6552 |
|
|
{
|
6553 |
|
|
/* Must switch to double mode to access these registers. */
|
6554 |
|
|
target_flags &= ~MASK_FPU_SINGLE;
|
6555 |
|
|
}
|
6556 |
|
|
}
|
6557 |
|
|
}
|
6558 |
|
|
if (nosave_low_regs && reg == R8_REG)
|
6559 |
|
|
break;
|
6560 |
|
|
}
|
6561 |
|
|
/* If we have a target register optimization pass after prologue / epilogue
|
6562 |
|
|
threading, we need to assume all target registers will be live even if
|
6563 |
|
|
they aren't now. */
|
6564 |
|
|
if (flag_branch_target_load_optimize2
|
6565 |
|
|
&& TARGET_SAVE_ALL_TARGET_REGS
|
6566 |
|
|
&& shmedia_space_reserved_for_target_registers)
|
6567 |
|
|
for (reg = LAST_TARGET_REG; reg >= FIRST_TARGET_REG; reg--)
|
6568 |
|
|
if ((! call_really_used_regs[reg] || interrupt_handler)
|
6569 |
|
|
&& ! TEST_HARD_REG_BIT (*live_regs_mask, reg))
|
6570 |
|
|
{
|
6571 |
|
|
SET_HARD_REG_BIT (*live_regs_mask, reg);
|
6572 |
|
|
count += GET_MODE_SIZE (REGISTER_NATURAL_MODE (reg));
|
6573 |
|
|
}
|
6574 |
|
|
/* If this is an interrupt handler, we don't have any call-clobbered
|
6575 |
|
|
registers we can conveniently use for target register save/restore.
|
6576 |
|
|
Make sure we save at least one general purpose register when we need
|
6577 |
|
|
to save target registers. */
|
6578 |
|
|
if (interrupt_handler
|
6579 |
|
|
&& hard_reg_set_intersect_p (*live_regs_mask,
|
6580 |
|
|
reg_class_contents[TARGET_REGS])
|
6581 |
|
|
&& ! hard_reg_set_intersect_p (*live_regs_mask,
|
6582 |
|
|
reg_class_contents[GENERAL_REGS]))
|
6583 |
|
|
{
|
6584 |
|
|
SET_HARD_REG_BIT (*live_regs_mask, R0_REG);
|
6585 |
|
|
count += GET_MODE_SIZE (REGISTER_NATURAL_MODE (R0_REG));
|
6586 |
|
|
}
|
6587 |
|
|
|
6588 |
|
|
return count;
|
6589 |
|
|
}
|
6590 |
|
|
|
6591 |
|
|
/* Code to generate prologue and epilogue sequences */
|
6592 |
|
|
|
6593 |
|
|
/* PUSHED is the number of bytes that are being pushed on the
|
6594 |
|
|
stack for register saves. Return the frame size, padded
|
6595 |
|
|
appropriately so that the stack stays properly aligned. */
|
6596 |
|
|
static HOST_WIDE_INT
|
6597 |
|
|
rounded_frame_size (int pushed)
|
6598 |
|
|
{
|
6599 |
|
|
HOST_WIDE_INT size = get_frame_size ();
|
6600 |
|
|
HOST_WIDE_INT align = STACK_BOUNDARY / BITS_PER_UNIT;
|
6601 |
|
|
|
6602 |
|
|
return ((size + pushed + align - 1) & -align) - pushed;
|
6603 |
|
|
}
|
6604 |
|
|
|
6605 |
|
|
/* Choose a call-clobbered target-branch register that remains
|
6606 |
|
|
unchanged along the whole function. We set it up as the return
|
6607 |
|
|
value in the prologue. */
|
6608 |
|
|
int
|
6609 |
|
|
sh_media_register_for_return (void)
|
6610 |
|
|
{
|
6611 |
|
|
int regno;
|
6612 |
|
|
int tr0_used;
|
6613 |
|
|
|
6614 |
|
|
if (! current_function_is_leaf)
|
6615 |
|
|
return -1;
|
6616 |
|
|
if (lookup_attribute ("interrupt_handler",
|
6617 |
|
|
DECL_ATTRIBUTES (current_function_decl)))
|
6618 |
|
|
return -1;
|
6619 |
|
|
if (sh_cfun_interrupt_handler_p ())
|
6620 |
|
|
return -1;
|
6621 |
|
|
|
6622 |
|
|
tr0_used = flag_pic && df_regs_ever_live_p (PIC_OFFSET_TABLE_REGNUM);
|
6623 |
|
|
|
6624 |
|
|
for (regno = FIRST_TARGET_REG + tr0_used; regno <= LAST_TARGET_REG; regno++)
|
6625 |
|
|
if (call_really_used_regs[regno] && ! df_regs_ever_live_p (regno))
|
6626 |
|
|
return regno;
|
6627 |
|
|
|
6628 |
|
|
return -1;
|
6629 |
|
|
}
|
6630 |
|
|
|
6631 |
|
|
/* The maximum registers we need to save are:
|
6632 |
|
|
- 62 general purpose registers (r15 is stack pointer, r63 is zero)
|
6633 |
|
|
- 32 floating point registers (for each pair, we save none,
|
6634 |
|
|
one single precision value, or a double precision value).
|
6635 |
|
|
- 8 target registers
|
6636 |
|
|
- add 1 entry for a delimiter. */
|
6637 |
|
|
#define MAX_SAVED_REGS (62+32+8)
|
6638 |
|
|
|
6639 |
|
|
typedef struct save_entry_s
|
6640 |
|
|
{
|
6641 |
|
|
unsigned char reg;
|
6642 |
|
|
unsigned char mode;
|
6643 |
|
|
short offset;
|
6644 |
|
|
} save_entry;
|
6645 |
|
|
|
6646 |
|
|
#define MAX_TEMPS 4
|
6647 |
|
|
|
6648 |
|
|
/* There will be a delimiter entry with VOIDmode both at the start and the
|
6649 |
|
|
end of a filled in schedule. The end delimiter has the offset of the
|
6650 |
|
|
save with the smallest (i.e. most negative) offset. */
|
6651 |
|
|
typedef struct save_schedule_s
|
6652 |
|
|
{
|
6653 |
|
|
save_entry entries[MAX_SAVED_REGS + 2];
|
6654 |
|
|
int temps[MAX_TEMPS+1];
|
6655 |
|
|
} save_schedule;
|
6656 |
|
|
|
6657 |
|
|
/* Fill in SCHEDULE according to LIVE_REGS_MASK. If RESTORE is nonzero,
|
6658 |
|
|
use reverse order. Returns the last entry written to (not counting
|
6659 |
|
|
the delimiter). OFFSET_BASE is a number to be added to all offset
|
6660 |
|
|
entries. */
|
6661 |
|
|
|
6662 |
|
|
static save_entry *
|
6663 |
|
|
sh5_schedule_saves (HARD_REG_SET *live_regs_mask, save_schedule *schedule,
|
6664 |
|
|
int offset_base)
|
6665 |
|
|
{
|
6666 |
|
|
int align, i;
|
6667 |
|
|
save_entry *entry = schedule->entries;
|
6668 |
|
|
int tmpx = 0;
|
6669 |
|
|
int offset;
|
6670 |
|
|
|
6671 |
|
|
if (! current_function_interrupt)
|
6672 |
|
|
for (i = FIRST_GENERAL_REG; tmpx < MAX_TEMPS && i <= LAST_GENERAL_REG; i++)
|
6673 |
|
|
if (call_really_used_regs[i] && ! fixed_regs[i] && i != PR_MEDIA_REG
|
6674 |
|
|
&& ! FUNCTION_ARG_REGNO_P (i)
|
6675 |
|
|
&& i != FIRST_RET_REG
|
6676 |
|
|
&& ! (cfun->static_chain_decl != NULL && i == STATIC_CHAIN_REGNUM)
|
6677 |
|
|
&& ! (crtl->calls_eh_return
|
6678 |
|
|
&& (i == EH_RETURN_STACKADJ_REGNO
|
6679 |
|
|
|| ((unsigned) i >= EH_RETURN_DATA_REGNO (0)
|
6680 |
|
|
&& (unsigned) i <= EH_RETURN_DATA_REGNO (3)))))
|
6681 |
|
|
schedule->temps[tmpx++] = i;
|
6682 |
|
|
entry->reg = -1;
|
6683 |
|
|
entry->mode = VOIDmode;
|
6684 |
|
|
entry->offset = offset_base;
|
6685 |
|
|
entry++;
|
6686 |
|
|
/* We loop twice: first, we save 8-byte aligned registers in the
|
6687 |
|
|
higher addresses, that are known to be aligned. Then, we
|
6688 |
|
|
proceed to saving 32-bit registers that don't need 8-byte
|
6689 |
|
|
alignment.
|
6690 |
|
|
If this is an interrupt function, all registers that need saving
|
6691 |
|
|
need to be saved in full. moreover, we need to postpone saving
|
6692 |
|
|
target registers till we have saved some general purpose registers
|
6693 |
|
|
we can then use as scratch registers. */
|
6694 |
|
|
offset = offset_base;
|
6695 |
|
|
for (align = 1; align >= 0; align--)
|
6696 |
|
|
{
|
6697 |
|
|
for (i = FIRST_PSEUDO_REGISTER - 1; i >= 0; i--)
|
6698 |
|
|
if (TEST_HARD_REG_BIT (*live_regs_mask, i))
|
6699 |
|
|
{
|
6700 |
|
|
enum machine_mode mode = REGISTER_NATURAL_MODE (i);
|
6701 |
|
|
int reg = i;
|
6702 |
|
|
|
6703 |
|
|
if (current_function_interrupt)
|
6704 |
|
|
{
|
6705 |
|
|
if (TARGET_REGISTER_P (i))
|
6706 |
|
|
continue;
|
6707 |
|
|
if (GENERAL_REGISTER_P (i))
|
6708 |
|
|
mode = DImode;
|
6709 |
|
|
}
|
6710 |
|
|
if (mode == SFmode && (i % 2) == 1
|
6711 |
|
|
&& ! TARGET_FPU_SINGLE && FP_REGISTER_P (i)
|
6712 |
|
|
&& (TEST_HARD_REG_BIT (*live_regs_mask, (i ^ 1))))
|
6713 |
|
|
{
|
6714 |
|
|
mode = DFmode;
|
6715 |
|
|
i--;
|
6716 |
|
|
reg--;
|
6717 |
|
|
}
|
6718 |
|
|
|
6719 |
|
|
/* If we're doing the aligned pass and this is not aligned,
|
6720 |
|
|
or we're doing the unaligned pass and this is aligned,
|
6721 |
|
|
skip it. */
|
6722 |
|
|
if ((GET_MODE_SIZE (mode) % (STACK_BOUNDARY / BITS_PER_UNIT) == 0)
|
6723 |
|
|
!= align)
|
6724 |
|
|
continue;
|
6725 |
|
|
|
6726 |
|
|
if (current_function_interrupt
|
6727 |
|
|
&& GENERAL_REGISTER_P (i)
|
6728 |
|
|
&& tmpx < MAX_TEMPS)
|
6729 |
|
|
schedule->temps[tmpx++] = i;
|
6730 |
|
|
|
6731 |
|
|
offset -= GET_MODE_SIZE (mode);
|
6732 |
|
|
entry->reg = i;
|
6733 |
|
|
entry->mode = mode;
|
6734 |
|
|
entry->offset = offset;
|
6735 |
|
|
entry++;
|
6736 |
|
|
}
|
6737 |
|
|
if (align && current_function_interrupt)
|
6738 |
|
|
for (i = LAST_TARGET_REG; i >= FIRST_TARGET_REG; i--)
|
6739 |
|
|
if (TEST_HARD_REG_BIT (*live_regs_mask, i))
|
6740 |
|
|
{
|
6741 |
|
|
offset -= GET_MODE_SIZE (DImode);
|
6742 |
|
|
entry->reg = i;
|
6743 |
|
|
entry->mode = DImode;
|
6744 |
|
|
entry->offset = offset;
|
6745 |
|
|
entry++;
|
6746 |
|
|
}
|
6747 |
|
|
}
|
6748 |
|
|
entry->reg = -1;
|
6749 |
|
|
entry->mode = VOIDmode;
|
6750 |
|
|
entry->offset = offset;
|
6751 |
|
|
schedule->temps[tmpx] = -1;
|
6752 |
|
|
return entry - 1;
|
6753 |
|
|
}
|
6754 |
|
|
|
6755 |
|
|
void
|
6756 |
|
|
sh_expand_prologue (void)
|
6757 |
|
|
{
|
6758 |
|
|
HARD_REG_SET live_regs_mask;
|
6759 |
|
|
int d, i;
|
6760 |
|
|
int d_rounding = 0;
|
6761 |
|
|
int save_flags = target_flags;
|
6762 |
|
|
int pretend_args;
|
6763 |
|
|
tree sp_switch_attr
|
6764 |
|
|
= lookup_attribute ("sp_switch", DECL_ATTRIBUTES (current_function_decl));
|
6765 |
|
|
|
6766 |
|
|
current_function_interrupt = sh_cfun_interrupt_handler_p ();
|
6767 |
|
|
|
6768 |
|
|
/* We have pretend args if we had an object sent partially in registers
|
6769 |
|
|
and partially on the stack, e.g. a large structure. */
|
6770 |
|
|
pretend_args = crtl->args.pretend_args_size;
|
6771 |
|
|
if (TARGET_VARARGS_PRETEND_ARGS (current_function_decl)
|
6772 |
|
|
&& (NPARM_REGS(SImode)
|
6773 |
|
|
> crtl->args.info.arg_count[(int) SH_ARG_INT]))
|
6774 |
|
|
pretend_args = 0;
|
6775 |
|
|
/* Dwarf2 module doesn't expect frame related insns here. */
|
6776 |
|
|
output_stack_adjust (-pretend_args
|
6777 |
|
|
- crtl->args.info.stack_regs * 8,
|
6778 |
|
|
stack_pointer_rtx, 0, NULL, false);
|
6779 |
|
|
|
6780 |
|
|
if (TARGET_SHCOMPACT && flag_pic && crtl->args.info.call_cookie)
|
6781 |
|
|
/* We're going to use the PIC register to load the address of the
|
6782 |
|
|
incoming-argument decoder and/or of the return trampoline from
|
6783 |
|
|
the GOT, so make sure the PIC register is preserved and
|
6784 |
|
|
initialized. */
|
6785 |
|
|
df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM, true);
|
6786 |
|
|
|
6787 |
|
|
if (TARGET_SHCOMPACT
|
6788 |
|
|
&& (crtl->args.info.call_cookie & ~ CALL_COOKIE_RET_TRAMP(1)))
|
6789 |
|
|
{
|
6790 |
|
|
int reg;
|
6791 |
|
|
|
6792 |
|
|
/* First, make all registers with incoming arguments that will
|
6793 |
|
|
be pushed onto the stack live, so that register renaming
|
6794 |
|
|
doesn't overwrite them. */
|
6795 |
|
|
for (reg = 0; reg < NPARM_REGS (SImode); reg++)
|
6796 |
|
|
if (CALL_COOKIE_STACKSEQ_GET (crtl->args.info.call_cookie)
|
6797 |
|
|
>= NPARM_REGS (SImode) - reg)
|
6798 |
|
|
for (; reg < NPARM_REGS (SImode); reg++)
|
6799 |
|
|
emit_insn (gen_shcompact_preserve_incoming_args
|
6800 |
|
|
(gen_rtx_REG (SImode, FIRST_PARM_REG + reg)));
|
6801 |
|
|
else if (CALL_COOKIE_INT_REG_GET
|
6802 |
|
|
(crtl->args.info.call_cookie, reg) == 1)
|
6803 |
|
|
emit_insn (gen_shcompact_preserve_incoming_args
|
6804 |
|
|
(gen_rtx_REG (SImode, FIRST_PARM_REG + reg)));
|
6805 |
|
|
|
6806 |
|
|
emit_move_insn (gen_rtx_REG (Pmode, MACL_REG),
|
6807 |
|
|
stack_pointer_rtx);
|
6808 |
|
|
emit_move_insn (gen_rtx_REG (SImode, R0_REG),
|
6809 |
|
|
GEN_INT (crtl->args.info.call_cookie));
|
6810 |
|
|
emit_move_insn (gen_rtx_REG (SImode, MACH_REG),
|
6811 |
|
|
gen_rtx_REG (SImode, R0_REG));
|
6812 |
|
|
}
|
6813 |
|
|
else if (TARGET_SHMEDIA)
|
6814 |
|
|
{
|
6815 |
|
|
int tr = sh_media_register_for_return ();
|
6816 |
|
|
|
6817 |
|
|
if (tr >= 0)
|
6818 |
|
|
emit_move_insn (gen_rtx_REG (DImode, tr),
|
6819 |
|
|
gen_rtx_REG (DImode, PR_MEDIA_REG));
|
6820 |
|
|
}
|
6821 |
|
|
|
6822 |
|
|
/* Emit the code for SETUP_VARARGS. */
|
6823 |
|
|
if (cfun->stdarg)
|
6824 |
|
|
{
|
6825 |
|
|
if (TARGET_VARARGS_PRETEND_ARGS (current_function_decl))
|
6826 |
|
|
{
|
6827 |
|
|
/* Push arg regs as if they'd been provided by caller in stack. */
|
6828 |
|
|
for (i = 0; i < NPARM_REGS(SImode); i++)
|
6829 |
|
|
{
|
6830 |
|
|
int rn = NPARM_REGS(SImode) + FIRST_PARM_REG - i - 1;
|
6831 |
|
|
rtx insn;
|
6832 |
|
|
|
6833 |
|
|
if (i >= (NPARM_REGS(SImode)
|
6834 |
|
|
- crtl->args.info.arg_count[(int) SH_ARG_INT]
|
6835 |
|
|
))
|
6836 |
|
|
break;
|
6837 |
|
|
insn = push (rn);
|
6838 |
|
|
}
|
6839 |
|
|
}
|
6840 |
|
|
}
|
6841 |
|
|
|
6842 |
|
|
/* If we're supposed to switch stacks at function entry, do so now. */
|
6843 |
|
|
if (sp_switch_attr)
|
6844 |
|
|
{
|
6845 |
|
|
rtx lab, newsrc;
|
6846 |
|
|
/* The argument specifies a variable holding the address of the
|
6847 |
|
|
stack the interrupt function should switch to/from at entry/exit. */
|
6848 |
|
|
tree arg = TREE_VALUE ( TREE_VALUE (sp_switch_attr));
|
6849 |
|
|
const char *s
|
6850 |
|
|
= ggc_strdup (TREE_STRING_POINTER (arg));
|
6851 |
|
|
rtx sp_switch = gen_rtx_SYMBOL_REF (Pmode, s);
|
6852 |
|
|
|
6853 |
|
|
lab = add_constant (sp_switch, SImode, 0);
|
6854 |
|
|
newsrc = gen_rtx_LABEL_REF (VOIDmode, lab);
|
6855 |
|
|
newsrc = gen_const_mem (SImode, newsrc);
|
6856 |
|
|
|
6857 |
|
|
emit_insn (gen_sp_switch_1 (newsrc));
|
6858 |
|
|
}
|
6859 |
|
|
|
6860 |
|
|
d = calc_live_regs (&live_regs_mask);
|
6861 |
|
|
/* ??? Maybe we could save some switching if we can move a mode switch
|
6862 |
|
|
that already happens to be at the function start into the prologue. */
|
6863 |
|
|
if (target_flags != save_flags && ! current_function_interrupt)
|
6864 |
|
|
emit_insn (gen_toggle_sz ());
|
6865 |
|
|
|
6866 |
|
|
if (TARGET_SH5)
|
6867 |
|
|
{
|
6868 |
|
|
int offset_base, offset;
|
6869 |
|
|
rtx r0 = NULL_RTX;
|
6870 |
|
|
int offset_in_r0 = -1;
|
6871 |
|
|
int sp_in_r0 = 0;
|
6872 |
|
|
int tregs_space = shmedia_target_regs_stack_adjust (&live_regs_mask);
|
6873 |
|
|
int total_size, save_size;
|
6874 |
|
|
save_schedule schedule;
|
6875 |
|
|
save_entry *entry;
|
6876 |
|
|
int *tmp_pnt;
|
6877 |
|
|
|
6878 |
|
|
if (call_really_used_regs[R0_REG] && ! fixed_regs[R0_REG]
|
6879 |
|
|
&& ! current_function_interrupt)
|
6880 |
|
|
r0 = gen_rtx_REG (Pmode, R0_REG);
|
6881 |
|
|
|
6882 |
|
|
/* D is the actual number of bytes that we need for saving registers,
|
6883 |
|
|
however, in initial_elimination_offset we have committed to using
|
6884 |
|
|
an additional TREGS_SPACE amount of bytes - in order to keep both
|
6885 |
|
|
addresses to arguments supplied by the caller and local variables
|
6886 |
|
|
valid, we must keep this gap. Place it between the incoming
|
6887 |
|
|
arguments and the actually saved registers in a bid to optimize
|
6888 |
|
|
locality of reference. */
|
6889 |
|
|
total_size = d + tregs_space;
|
6890 |
|
|
total_size += rounded_frame_size (total_size);
|
6891 |
|
|
save_size = total_size - rounded_frame_size (d);
|
6892 |
|
|
if (save_size % (STACK_BOUNDARY / BITS_PER_UNIT))
|
6893 |
|
|
d_rounding = ((STACK_BOUNDARY / BITS_PER_UNIT)
|
6894 |
|
|
- save_size % (STACK_BOUNDARY / BITS_PER_UNIT));
|
6895 |
|
|
|
6896 |
|
|
/* If adjusting the stack in a single step costs nothing extra, do so.
|
6897 |
|
|
I.e. either if a single addi is enough, or we need a movi anyway,
|
6898 |
|
|
and we don't exceed the maximum offset range (the test for the
|
6899 |
|
|
latter is conservative for simplicity). */
|
6900 |
|
|
if (TARGET_SHMEDIA
|
6901 |
|
|
&& (CONST_OK_FOR_I10 (-total_size)
|
6902 |
|
|
|| (! CONST_OK_FOR_I10 (-(save_size + d_rounding))
|
6903 |
|
|
&& total_size <= 2044)))
|
6904 |
|
|
d_rounding = total_size - save_size;
|
6905 |
|
|
|
6906 |
|
|
offset_base = d + d_rounding;
|
6907 |
|
|
|
6908 |
|
|
output_stack_adjust (-(save_size + d_rounding), stack_pointer_rtx,
|
6909 |
|
|
0, NULL, true);
|
6910 |
|
|
|
6911 |
|
|
sh5_schedule_saves (&live_regs_mask, &schedule, offset_base);
|
6912 |
|
|
tmp_pnt = schedule.temps;
|
6913 |
|
|
for (entry = &schedule.entries[1]; entry->mode != VOIDmode; entry++)
|
6914 |
|
|
{
|
6915 |
|
|
enum machine_mode mode = (enum machine_mode) entry->mode;
|
6916 |
|
|
unsigned int reg = entry->reg;
|
6917 |
|
|
rtx reg_rtx, mem_rtx, pre_dec = NULL_RTX;
|
6918 |
|
|
rtx orig_reg_rtx;
|
6919 |
|
|
|
6920 |
|
|
offset = entry->offset;
|
6921 |
|
|
|
6922 |
|
|
reg_rtx = gen_rtx_REG (mode, reg);
|
6923 |
|
|
|
6924 |
|
|
mem_rtx = gen_frame_mem (mode,
|
6925 |
|
|
gen_rtx_PLUS (Pmode,
|
6926 |
|
|
stack_pointer_rtx,
|
6927 |
|
|
GEN_INT (offset)));
|
6928 |
|
|
|
6929 |
|
|
if (!memory_address_p (mode, XEXP (mem_rtx, 0)))
|
6930 |
|
|
{
|
6931 |
|
|
gcc_assert (r0);
|
6932 |
|
|
mem_rtx = NULL_RTX;
|
6933 |
|
|
}
|
6934 |
|
|
|
6935 |
|
|
if (HAVE_PRE_DECREMENT
|
6936 |
|
|
&& (offset_in_r0 - offset == GET_MODE_SIZE (mode)
|
6937 |
|
|
|| mem_rtx == NULL_RTX
|
6938 |
|
|
|| reg == PR_REG || SPECIAL_REGISTER_P (reg)))
|
6939 |
|
|
{
|
6940 |
|
|
pre_dec = gen_frame_mem (mode, gen_rtx_PRE_DEC (Pmode, r0));
|
6941 |
|
|
|
6942 |
|
|
if (!memory_address_p (mode, XEXP (pre_dec, 0)))
|
6943 |
|
|
pre_dec = NULL_RTX;
|
6944 |
|
|
else
|
6945 |
|
|
{
|
6946 |
|
|
mem_rtx = NULL_RTX;
|
6947 |
|
|
offset += GET_MODE_SIZE (mode);
|
6948 |
|
|
}
|
6949 |
|
|
}
|
6950 |
|
|
|
6951 |
|
|
if (mem_rtx != NULL_RTX)
|
6952 |
|
|
goto addr_ok;
|
6953 |
|
|
|
6954 |
|
|
if (offset_in_r0 == -1)
|
6955 |
|
|
{
|
6956 |
|
|
emit_move_insn (r0, GEN_INT (offset));
|
6957 |
|
|
offset_in_r0 = offset;
|
6958 |
|
|
}
|
6959 |
|
|
else if (offset != offset_in_r0)
|
6960 |
|
|
{
|
6961 |
|
|
emit_move_insn (r0,
|
6962 |
|
|
gen_rtx_PLUS
|
6963 |
|
|
(Pmode, r0,
|
6964 |
|
|
GEN_INT (offset - offset_in_r0)));
|
6965 |
|
|
offset_in_r0 += offset - offset_in_r0;
|
6966 |
|
|
}
|
6967 |
|
|
|
6968 |
|
|
if (pre_dec != NULL_RTX)
|
6969 |
|
|
{
|
6970 |
|
|
if (! sp_in_r0)
|
6971 |
|
|
{
|
6972 |
|
|
emit_move_insn (r0,
|
6973 |
|
|
gen_rtx_PLUS
|
6974 |
|
|
(Pmode, r0, stack_pointer_rtx));
|
6975 |
|
|
sp_in_r0 = 1;
|
6976 |
|
|
}
|
6977 |
|
|
|
6978 |
|
|
offset -= GET_MODE_SIZE (mode);
|
6979 |
|
|
offset_in_r0 -= GET_MODE_SIZE (mode);
|
6980 |
|
|
|
6981 |
|
|
mem_rtx = pre_dec;
|
6982 |
|
|
}
|
6983 |
|
|
else if (sp_in_r0)
|
6984 |
|
|
mem_rtx = gen_frame_mem (mode, r0);
|
6985 |
|
|
else
|
6986 |
|
|
mem_rtx = gen_frame_mem (mode,
|
6987 |
|
|
gen_rtx_PLUS (Pmode,
|
6988 |
|
|
stack_pointer_rtx,
|
6989 |
|
|
r0));
|
6990 |
|
|
|
6991 |
|
|
/* We must not use an r0-based address for target-branch
|
6992 |
|
|
registers or for special registers without pre-dec
|
6993 |
|
|
memory addresses, since we store their values in r0
|
6994 |
|
|
first. */
|
6995 |
|
|
gcc_assert (!TARGET_REGISTER_P (reg)
|
6996 |
|
|
&& ((reg != PR_REG && !SPECIAL_REGISTER_P (reg))
|
6997 |
|
|
|| mem_rtx == pre_dec));
|
6998 |
|
|
|
6999 |
|
|
addr_ok:
|
7000 |
|
|
orig_reg_rtx = reg_rtx;
|
7001 |
|
|
if (TARGET_REGISTER_P (reg)
|
7002 |
|
|
|| ((reg == PR_REG || SPECIAL_REGISTER_P (reg))
|
7003 |
|
|
&& mem_rtx != pre_dec))
|
7004 |
|
|
{
|
7005 |
|
|
rtx tmp_reg = gen_rtx_REG (GET_MODE (reg_rtx), *tmp_pnt);
|
7006 |
|
|
|
7007 |
|
|
emit_move_insn (tmp_reg, reg_rtx);
|
7008 |
|
|
|
7009 |
|
|
if (REGNO (tmp_reg) == R0_REG)
|
7010 |
|
|
{
|
7011 |
|
|
offset_in_r0 = -1;
|
7012 |
|
|
sp_in_r0 = 0;
|
7013 |
|
|
gcc_assert (!refers_to_regno_p
|
7014 |
|
|
(R0_REG, R0_REG+1, mem_rtx, (rtx *) 0));
|
7015 |
|
|
}
|
7016 |
|
|
|
7017 |
|
|
if (*++tmp_pnt <= 0)
|
7018 |
|
|
tmp_pnt = schedule.temps;
|
7019 |
|
|
|
7020 |
|
|
reg_rtx = tmp_reg;
|
7021 |
|
|
}
|
7022 |
|
|
{
|
7023 |
|
|
rtx insn;
|
7024 |
|
|
|
7025 |
|
|
/* Mark as interesting for dwarf cfi generator */
|
7026 |
|
|
insn = emit_move_insn (mem_rtx, reg_rtx);
|
7027 |
|
|
RTX_FRAME_RELATED_P (insn) = 1;
|
7028 |
|
|
/* If we use an intermediate register for the save, we can't
|
7029 |
|
|
describe this exactly in cfi as a copy of the to-be-saved
|
7030 |
|
|
register into the temporary register and then the temporary
|
7031 |
|
|
register on the stack, because the temporary register can
|
7032 |
|
|
have a different natural size than the to-be-saved register.
|
7033 |
|
|
Thus, we gloss over the intermediate copy and pretend we do
|
7034 |
|
|
a direct save from the to-be-saved register. */
|
7035 |
|
|
if (REGNO (reg_rtx) != reg)
|
7036 |
|
|
{
|
7037 |
|
|
rtx set;
|
7038 |
|
|
|
7039 |
|
|
set = gen_rtx_SET (VOIDmode, mem_rtx, orig_reg_rtx);
|
7040 |
|
|
add_reg_note (insn, REG_FRAME_RELATED_EXPR, set);
|
7041 |
|
|
}
|
7042 |
|
|
|
7043 |
|
|
if (TARGET_SHCOMPACT && (offset_in_r0 != -1))
|
7044 |
|
|
{
|
7045 |
|
|
rtx reg_rtx = gen_rtx_REG (mode, reg);
|
7046 |
|
|
rtx set;
|
7047 |
|
|
rtx mem_rtx = gen_frame_mem (mode,
|
7048 |
|
|
gen_rtx_PLUS (Pmode,
|
7049 |
|
|
stack_pointer_rtx,
|
7050 |
|
|
GEN_INT (offset)));
|
7051 |
|
|
|
7052 |
|
|
set = gen_rtx_SET (VOIDmode, mem_rtx, reg_rtx);
|
7053 |
|
|
add_reg_note (insn, REG_FRAME_RELATED_EXPR, set);
|
7054 |
|
|
}
|
7055 |
|
|
}
|
7056 |
|
|
}
|
7057 |
|
|
|
7058 |
|
|
gcc_assert (entry->offset == d_rounding);
|
7059 |
|
|
}
|
7060 |
|
|
else
|
7061 |
|
|
push_regs (&live_regs_mask, current_function_interrupt);
|
7062 |
|
|
|
7063 |
|
|
if (flag_pic && df_regs_ever_live_p (PIC_OFFSET_TABLE_REGNUM))
|
7064 |
|
|
emit_insn (gen_GOTaddr2picreg ());
|
7065 |
|
|
|
7066 |
|
|
if (SHMEDIA_REGS_STACK_ADJUST ())
|
7067 |
|
|
{
|
7068 |
|
|
/* This must NOT go through the PLT, otherwise mach and macl
|
7069 |
|
|
may be clobbered. */
|
7070 |
|
|
function_symbol (gen_rtx_REG (Pmode, R0_REG),
|
7071 |
|
|
(TARGET_FPU_ANY
|
7072 |
|
|
? "__GCC_push_shmedia_regs"
|
7073 |
|
|
: "__GCC_push_shmedia_regs_nofpu"), SFUNC_GOT);
|
7074 |
|
|
emit_insn (gen_shmedia_save_restore_regs_compact
|
7075 |
|
|
(GEN_INT (-SHMEDIA_REGS_STACK_ADJUST ())));
|
7076 |
|
|
}
|
7077 |
|
|
|
7078 |
|
|
if (target_flags != save_flags && ! current_function_interrupt)
|
7079 |
|
|
emit_insn (gen_toggle_sz ());
|
7080 |
|
|
|
7081 |
|
|
target_flags = save_flags;
|
7082 |
|
|
|
7083 |
|
|
output_stack_adjust (-rounded_frame_size (d) + d_rounding,
|
7084 |
|
|
stack_pointer_rtx, 0, NULL, true);
|
7085 |
|
|
|
7086 |
|
|
if (frame_pointer_needed)
|
7087 |
|
|
frame_insn (GEN_MOV (hard_frame_pointer_rtx, stack_pointer_rtx));
|
7088 |
|
|
|
7089 |
|
|
if (TARGET_SHCOMPACT
|
7090 |
|
|
&& (crtl->args.info.call_cookie & ~ CALL_COOKIE_RET_TRAMP(1)))
|
7091 |
|
|
{
|
7092 |
|
|
/* This must NOT go through the PLT, otherwise mach and macl
|
7093 |
|
|
may be clobbered. */
|
7094 |
|
|
function_symbol (gen_rtx_REG (Pmode, R0_REG),
|
7095 |
|
|
"__GCC_shcompact_incoming_args", SFUNC_GOT);
|
7096 |
|
|
emit_insn (gen_shcompact_incoming_args ());
|
7097 |
|
|
}
|
7098 |
|
|
}
|
7099 |
|
|
|
7100 |
|
|
void
|
7101 |
|
|
sh_expand_epilogue (bool sibcall_p)
|
7102 |
|
|
{
|
7103 |
|
|
HARD_REG_SET live_regs_mask;
|
7104 |
|
|
int d, i;
|
7105 |
|
|
int d_rounding = 0;
|
7106 |
|
|
|
7107 |
|
|
int save_flags = target_flags;
|
7108 |
|
|
int frame_size, save_size;
|
7109 |
|
|
int fpscr_deferred = 0;
|
7110 |
|
|
int e = sibcall_p ? -1 : 1;
|
7111 |
|
|
|
7112 |
|
|
d = calc_live_regs (&live_regs_mask);
|
7113 |
|
|
|
7114 |
|
|
save_size = d;
|
7115 |
|
|
frame_size = rounded_frame_size (d);
|
7116 |
|
|
|
7117 |
|
|
if (TARGET_SH5)
|
7118 |
|
|
{
|
7119 |
|
|
int tregs_space = shmedia_target_regs_stack_adjust (&live_regs_mask);
|
7120 |
|
|
int total_size;
|
7121 |
|
|
if (d % (STACK_BOUNDARY / BITS_PER_UNIT))
|
7122 |
|
|
d_rounding = ((STACK_BOUNDARY / BITS_PER_UNIT)
|
7123 |
|
|
- d % (STACK_BOUNDARY / BITS_PER_UNIT));
|
7124 |
|
|
|
7125 |
|
|
total_size = d + tregs_space;
|
7126 |
|
|
total_size += rounded_frame_size (total_size);
|
7127 |
|
|
save_size = total_size - frame_size;
|
7128 |
|
|
|
7129 |
|
|
/* If adjusting the stack in a single step costs nothing extra, do so.
|
7130 |
|
|
I.e. either if a single addi is enough, or we need a movi anyway,
|
7131 |
|
|
and we don't exceed the maximum offset range (the test for the
|
7132 |
|
|
latter is conservative for simplicity). */
|
7133 |
|
|
if (TARGET_SHMEDIA
|
7134 |
|
|
&& ! frame_pointer_needed
|
7135 |
|
|
&& (CONST_OK_FOR_I10 (total_size)
|
7136 |
|
|
|| (! CONST_OK_FOR_I10 (save_size + d_rounding)
|
7137 |
|
|
&& total_size <= 2044)))
|
7138 |
|
|
d_rounding = frame_size;
|
7139 |
|
|
|
7140 |
|
|
frame_size -= d_rounding;
|
7141 |
|
|
}
|
7142 |
|
|
|
7143 |
|
|
if (frame_pointer_needed)
|
7144 |
|
|
{
|
7145 |
|
|
/* We must avoid scheduling the epilogue with previous basic blocks.
|
7146 |
|
|
See PR/18032 and PR/40313. */
|
7147 |
|
|
emit_insn (gen_blockage ());
|
7148 |
|
|
output_stack_adjust (frame_size, hard_frame_pointer_rtx, e,
|
7149 |
|
|
&live_regs_mask, false);
|
7150 |
|
|
|
7151 |
|
|
/* We must avoid moving the stack pointer adjustment past code
|
7152 |
|
|
which reads from the local frame, else an interrupt could
|
7153 |
|
|
occur after the SP adjustment and clobber data in the local
|
7154 |
|
|
frame. */
|
7155 |
|
|
emit_insn (gen_blockage ());
|
7156 |
|
|
emit_insn (GEN_MOV (stack_pointer_rtx, hard_frame_pointer_rtx));
|
7157 |
|
|
}
|
7158 |
|
|
else if (frame_size)
|
7159 |
|
|
{
|
7160 |
|
|
/* We must avoid moving the stack pointer adjustment past code
|
7161 |
|
|
which reads from the local frame, else an interrupt could
|
7162 |
|
|
occur after the SP adjustment and clobber data in the local
|
7163 |
|
|
frame. */
|
7164 |
|
|
emit_insn (gen_blockage ());
|
7165 |
|
|
output_stack_adjust (frame_size, stack_pointer_rtx, e,
|
7166 |
|
|
&live_regs_mask, false);
|
7167 |
|
|
}
|
7168 |
|
|
|
7169 |
|
|
if (SHMEDIA_REGS_STACK_ADJUST ())
|
7170 |
|
|
{
|
7171 |
|
|
function_symbol (gen_rtx_REG (Pmode, R0_REG),
|
7172 |
|
|
(TARGET_FPU_ANY
|
7173 |
|
|
? "__GCC_pop_shmedia_regs"
|
7174 |
|
|
: "__GCC_pop_shmedia_regs_nofpu"), SFUNC_GOT);
|
7175 |
|
|
/* This must NOT go through the PLT, otherwise mach and macl
|
7176 |
|
|
may be clobbered. */
|
7177 |
|
|
emit_insn (gen_shmedia_save_restore_regs_compact
|
7178 |
|
|
(GEN_INT (SHMEDIA_REGS_STACK_ADJUST ())));
|
7179 |
|
|
}
|
7180 |
|
|
|
7181 |
|
|
/* Pop all the registers. */
|
7182 |
|
|
|
7183 |
|
|
if (target_flags != save_flags && ! current_function_interrupt)
|
7184 |
|
|
emit_insn (gen_toggle_sz ());
|
7185 |
|
|
if (TARGET_SH5)
|
7186 |
|
|
{
|
7187 |
|
|
int offset_base, offset;
|
7188 |
|
|
int offset_in_r0 = -1;
|
7189 |
|
|
int sp_in_r0 = 0;
|
7190 |
|
|
rtx r0 = gen_rtx_REG (Pmode, R0_REG);
|
7191 |
|
|
save_schedule schedule;
|
7192 |
|
|
save_entry *entry;
|
7193 |
|
|
int *tmp_pnt;
|
7194 |
|
|
|
7195 |
|
|
entry = sh5_schedule_saves (&live_regs_mask, &schedule, d_rounding);
|
7196 |
|
|
offset_base = -entry[1].offset + d_rounding;
|
7197 |
|
|
tmp_pnt = schedule.temps;
|
7198 |
|
|
for (; entry->mode != VOIDmode; entry--)
|
7199 |
|
|
{
|
7200 |
|
|
enum machine_mode mode = (enum machine_mode) entry->mode;
|
7201 |
|
|
int reg = entry->reg;
|
7202 |
|
|
rtx reg_rtx, mem_rtx, post_inc = NULL_RTX, insn;
|
7203 |
|
|
|
7204 |
|
|
offset = offset_base + entry->offset;
|
7205 |
|
|
reg_rtx = gen_rtx_REG (mode, reg);
|
7206 |
|
|
|
7207 |
|
|
mem_rtx = gen_frame_mem (mode,
|
7208 |
|
|
gen_rtx_PLUS (Pmode,
|
7209 |
|
|
stack_pointer_rtx,
|
7210 |
|
|
GEN_INT (offset)));
|
7211 |
|
|
|
7212 |
|
|
if (!memory_address_p (mode, XEXP (mem_rtx, 0)))
|
7213 |
|
|
mem_rtx = NULL_RTX;
|
7214 |
|
|
|
7215 |
|
|
if (HAVE_POST_INCREMENT
|
7216 |
|
|
&& (offset == offset_in_r0
|
7217 |
|
|
|| (offset + GET_MODE_SIZE (mode) != d + d_rounding
|
7218 |
|
|
&& mem_rtx == NULL_RTX)
|
7219 |
|
|
|| reg == PR_REG || SPECIAL_REGISTER_P (reg)))
|
7220 |
|
|
{
|
7221 |
|
|
post_inc = gen_frame_mem (mode, gen_rtx_POST_INC (Pmode, r0));
|
7222 |
|
|
|
7223 |
|
|
if (!memory_address_p (mode, XEXP (post_inc, 0)))
|
7224 |
|
|
post_inc = NULL_RTX;
|
7225 |
|
|
else
|
7226 |
|
|
mem_rtx = NULL_RTX;
|
7227 |
|
|
}
|
7228 |
|
|
|
7229 |
|
|
if (mem_rtx != NULL_RTX)
|
7230 |
|
|
goto addr_ok;
|
7231 |
|
|
|
7232 |
|
|
if (offset_in_r0 == -1)
|
7233 |
|
|
{
|
7234 |
|
|
emit_move_insn (r0, GEN_INT (offset));
|
7235 |
|
|
offset_in_r0 = offset;
|
7236 |
|
|
}
|
7237 |
|
|
else if (offset != offset_in_r0)
|
7238 |
|
|
{
|
7239 |
|
|
emit_move_insn (r0,
|
7240 |
|
|
gen_rtx_PLUS
|
7241 |
|
|
(Pmode, r0,
|
7242 |
|
|
GEN_INT (offset - offset_in_r0)));
|
7243 |
|
|
offset_in_r0 += offset - offset_in_r0;
|
7244 |
|
|
}
|
7245 |
|
|
|
7246 |
|
|
if (post_inc != NULL_RTX)
|
7247 |
|
|
{
|
7248 |
|
|
if (! sp_in_r0)
|
7249 |
|
|
{
|
7250 |
|
|
emit_move_insn (r0,
|
7251 |
|
|
gen_rtx_PLUS
|
7252 |
|
|
(Pmode, r0, stack_pointer_rtx));
|
7253 |
|
|
sp_in_r0 = 1;
|
7254 |
|
|
}
|
7255 |
|
|
|
7256 |
|
|
mem_rtx = post_inc;
|
7257 |
|
|
|
7258 |
|
|
offset_in_r0 += GET_MODE_SIZE (mode);
|
7259 |
|
|
}
|
7260 |
|
|
else if (sp_in_r0)
|
7261 |
|
|
mem_rtx = gen_frame_mem (mode, r0);
|
7262 |
|
|
else
|
7263 |
|
|
mem_rtx = gen_frame_mem (mode,
|
7264 |
|
|
gen_rtx_PLUS (Pmode,
|
7265 |
|
|
stack_pointer_rtx,
|
7266 |
|
|
r0));
|
7267 |
|
|
|
7268 |
|
|
gcc_assert ((reg != PR_REG && !SPECIAL_REGISTER_P (reg))
|
7269 |
|
|
|| mem_rtx == post_inc);
|
7270 |
|
|
|
7271 |
|
|
addr_ok:
|
7272 |
|
|
if ((reg == PR_REG || SPECIAL_REGISTER_P (reg))
|
7273 |
|
|
&& mem_rtx != post_inc)
|
7274 |
|
|
{
|
7275 |
|
|
insn = emit_move_insn (r0, mem_rtx);
|
7276 |
|
|
mem_rtx = r0;
|
7277 |
|
|
}
|
7278 |
|
|
else if (TARGET_REGISTER_P (reg))
|
7279 |
|
|
{
|
7280 |
|
|
rtx tmp_reg = gen_rtx_REG (mode, *tmp_pnt);
|
7281 |
|
|
|
7282 |
|
|
/* Give the scheduler a bit of freedom by using up to
|
7283 |
|
|
MAX_TEMPS registers in a round-robin fashion. */
|
7284 |
|
|
insn = emit_move_insn (tmp_reg, mem_rtx);
|
7285 |
|
|
mem_rtx = tmp_reg;
|
7286 |
|
|
if (*++tmp_pnt < 0)
|
7287 |
|
|
tmp_pnt = schedule.temps;
|
7288 |
|
|
}
|
7289 |
|
|
|
7290 |
|
|
insn = emit_move_insn (reg_rtx, mem_rtx);
|
7291 |
|
|
}
|
7292 |
|
|
|
7293 |
|
|
gcc_assert (entry->offset + offset_base == d + d_rounding);
|
7294 |
|
|
}
|
7295 |
|
|
else /* ! TARGET_SH5 */
|
7296 |
|
|
{
|
7297 |
|
|
int last_reg;
|
7298 |
|
|
|
7299 |
|
|
save_size = 0;
|
7300 |
|
|
/* For an ISR with RESBANK attribute assigned, don't pop PR
|
7301 |
|
|
register. */
|
7302 |
|
|
if (TEST_HARD_REG_BIT (live_regs_mask, PR_REG)
|
7303 |
|
|
&& !sh_cfun_resbank_handler_p ())
|
7304 |
|
|
{
|
7305 |
|
|
if (!frame_pointer_needed)
|
7306 |
|
|
emit_insn (gen_blockage ());
|
7307 |
|
|
pop (PR_REG);
|
7308 |
|
|
}
|
7309 |
|
|
|
7310 |
|
|
/* Banked registers are popped first to avoid being scheduled in the
|
7311 |
|
|
delay slot. RTE switches banks before the ds instruction. */
|
7312 |
|
|
if (current_function_interrupt)
|
7313 |
|
|
{
|
7314 |
|
|
for (i = LAST_BANKED_REG; i >= FIRST_BANKED_REG; i--)
|
7315 |
|
|
if (TEST_HARD_REG_BIT (live_regs_mask, i))
|
7316 |
|
|
pop (i);
|
7317 |
|
|
|
7318 |
|
|
last_reg = FIRST_PSEUDO_REGISTER - LAST_BANKED_REG - 1;
|
7319 |
|
|
}
|
7320 |
|
|
else
|
7321 |
|
|
last_reg = FIRST_PSEUDO_REGISTER;
|
7322 |
|
|
|
7323 |
|
|
for (i = 0; i < last_reg; i++)
|
7324 |
|
|
{
|
7325 |
|
|
int j = (FIRST_PSEUDO_REGISTER - 1) - i;
|
7326 |
|
|
|
7327 |
|
|
if (j == FPSCR_REG && current_function_interrupt && TARGET_FMOVD
|
7328 |
|
|
&& hard_reg_set_intersect_p (live_regs_mask,
|
7329 |
|
|
reg_class_contents[DF_REGS]))
|
7330 |
|
|
fpscr_deferred = 1;
|
7331 |
|
|
/* For an ISR with RESBANK attribute assigned, don't pop
|
7332 |
|
|
following registers, R0-R14, MACH, MACL and GBR. */
|
7333 |
|
|
else if (j != PR_REG && TEST_HARD_REG_BIT (live_regs_mask, j)
|
7334 |
|
|
&& ! (sh_cfun_resbank_handler_p ()
|
7335 |
|
|
&& ((j >= FIRST_GENERAL_REG
|
7336 |
|
|
&& j < LAST_GENERAL_REG)
|
7337 |
|
|
|| j == MACH_REG
|
7338 |
|
|
|| j == MACL_REG
|
7339 |
|
|
|| j == GBR_REG)))
|
7340 |
|
|
pop (j);
|
7341 |
|
|
|
7342 |
|
|
if (j == FIRST_FP_REG && fpscr_deferred)
|
7343 |
|
|
pop (FPSCR_REG);
|
7344 |
|
|
}
|
7345 |
|
|
}
|
7346 |
|
|
if (target_flags != save_flags && ! current_function_interrupt)
|
7347 |
|
|
emit_insn (gen_toggle_sz ());
|
7348 |
|
|
target_flags = save_flags;
|
7349 |
|
|
|
7350 |
|
|
output_stack_adjust (crtl->args.pretend_args_size
|
7351 |
|
|
+ save_size + d_rounding
|
7352 |
|
|
+ crtl->args.info.stack_regs * 8,
|
7353 |
|
|
stack_pointer_rtx, e, NULL, false);
|
7354 |
|
|
|
7355 |
|
|
if (crtl->calls_eh_return)
|
7356 |
|
|
emit_insn (GEN_ADD3 (stack_pointer_rtx, stack_pointer_rtx,
|
7357 |
|
|
EH_RETURN_STACKADJ_RTX));
|
7358 |
|
|
|
7359 |
|
|
/* Switch back to the normal stack if necessary. */
|
7360 |
|
|
if (lookup_attribute ("sp_switch", DECL_ATTRIBUTES (current_function_decl)))
|
7361 |
|
|
emit_insn (gen_sp_switch_2 ());
|
7362 |
|
|
|
7363 |
|
|
/* Tell flow the insn that pops PR isn't dead. */
|
7364 |
|
|
/* PR_REG will never be live in SHmedia mode, and we don't need to
|
7365 |
|
|
USE PR_MEDIA_REG, since it will be explicitly copied to TR0_REG
|
7366 |
|
|
by the return pattern. */
|
7367 |
|
|
if (TEST_HARD_REG_BIT (live_regs_mask, PR_REG))
|
7368 |
|
|
emit_use (gen_rtx_REG (SImode, PR_REG));
|
7369 |
|
|
}
|
7370 |
|
|
|
7371 |
|
|
static int sh_need_epilogue_known = 0;
|
7372 |
|
|
|
7373 |
|
|
int
|
7374 |
|
|
sh_need_epilogue (void)
|
7375 |
|
|
{
|
7376 |
|
|
if (! sh_need_epilogue_known)
|
7377 |
|
|
{
|
7378 |
|
|
rtx epilogue;
|
7379 |
|
|
|
7380 |
|
|
start_sequence ();
|
7381 |
|
|
sh_expand_epilogue (0);
|
7382 |
|
|
epilogue = get_insns ();
|
7383 |
|
|
end_sequence ();
|
7384 |
|
|
sh_need_epilogue_known = (epilogue == NULL ? -1 : 1);
|
7385 |
|
|
}
|
7386 |
|
|
return sh_need_epilogue_known > 0;
|
7387 |
|
|
}
|
7388 |
|
|
|
7389 |
|
|
/* Emit code to change the current function's return address to RA.
|
7390 |
|
|
TEMP is available as a scratch register, if needed. */
|
7391 |
|
|
|
7392 |
|
|
void
|
7393 |
|
|
sh_set_return_address (rtx ra, rtx tmp)
|
7394 |
|
|
{
|
7395 |
|
|
HARD_REG_SET live_regs_mask;
|
7396 |
|
|
int d;
|
7397 |
|
|
int pr_reg = TARGET_SHMEDIA ? PR_MEDIA_REG : PR_REG;
|
7398 |
|
|
int pr_offset;
|
7399 |
|
|
|
7400 |
|
|
d = calc_live_regs (&live_regs_mask);
|
7401 |
|
|
|
7402 |
|
|
/* If pr_reg isn't life, we can set it (or the register given in
|
7403 |
|
|
sh_media_register_for_return) directly. */
|
7404 |
|
|
if (! TEST_HARD_REG_BIT (live_regs_mask, pr_reg))
|
7405 |
|
|
{
|
7406 |
|
|
rtx rr;
|
7407 |
|
|
|
7408 |
|
|
if (TARGET_SHMEDIA)
|
7409 |
|
|
{
|
7410 |
|
|
int rr_regno = sh_media_register_for_return ();
|
7411 |
|
|
|
7412 |
|
|
if (rr_regno < 0)
|
7413 |
|
|
rr_regno = pr_reg;
|
7414 |
|
|
|
7415 |
|
|
rr = gen_rtx_REG (DImode, rr_regno);
|
7416 |
|
|
}
|
7417 |
|
|
else
|
7418 |
|
|
rr = gen_rtx_REG (SImode, pr_reg);
|
7419 |
|
|
|
7420 |
|
|
emit_insn (GEN_MOV (rr, ra));
|
7421 |
|
|
/* Tell flow the register for return isn't dead. */
|
7422 |
|
|
emit_use (rr);
|
7423 |
|
|
return;
|
7424 |
|
|
}
|
7425 |
|
|
|
7426 |
|
|
if (TARGET_SH5)
|
7427 |
|
|
{
|
7428 |
|
|
int offset;
|
7429 |
|
|
save_schedule schedule;
|
7430 |
|
|
save_entry *entry;
|
7431 |
|
|
|
7432 |
|
|
entry = sh5_schedule_saves (&live_regs_mask, &schedule, 0);
|
7433 |
|
|
offset = entry[1].offset;
|
7434 |
|
|
for (; entry->mode != VOIDmode; entry--)
|
7435 |
|
|
if (entry->reg == pr_reg)
|
7436 |
|
|
goto found;
|
7437 |
|
|
|
7438 |
|
|
/* We can't find pr register. */
|
7439 |
|
|
gcc_unreachable ();
|
7440 |
|
|
|
7441 |
|
|
found:
|
7442 |
|
|
offset = entry->offset - offset;
|
7443 |
|
|
pr_offset = (rounded_frame_size (d) + offset
|
7444 |
|
|
+ SHMEDIA_REGS_STACK_ADJUST ());
|
7445 |
|
|
}
|
7446 |
|
|
else
|
7447 |
|
|
pr_offset = rounded_frame_size (d);
|
7448 |
|
|
|
7449 |
|
|
emit_insn (GEN_MOV (tmp, GEN_INT (pr_offset)));
|
7450 |
|
|
emit_insn (GEN_ADD3 (tmp, tmp, hard_frame_pointer_rtx));
|
7451 |
|
|
|
7452 |
|
|
tmp = gen_frame_mem (Pmode, tmp);
|
7453 |
|
|
emit_insn (GEN_MOV (tmp, ra));
|
7454 |
|
|
/* Tell this store isn't dead. */
|
7455 |
|
|
emit_use (tmp);
|
7456 |
|
|
}
|
7457 |
|
|
|
7458 |
|
|
/* Clear variables at function end. */
|
7459 |
|
|
|
7460 |
|
|
static void
|
7461 |
|
|
sh_output_function_epilogue (FILE *file ATTRIBUTE_UNUSED,
|
7462 |
|
|
HOST_WIDE_INT size ATTRIBUTE_UNUSED)
|
7463 |
|
|
{
|
7464 |
|
|
sh_need_epilogue_known = 0;
|
7465 |
|
|
}
|
7466 |
|
|
|
7467 |
|
|
static rtx
|
7468 |
|
|
sh_builtin_saveregs (void)
|
7469 |
|
|
{
|
7470 |
|
|
/* First unnamed integer register. */
|
7471 |
|
|
int first_intreg = crtl->args.info.arg_count[(int) SH_ARG_INT];
|
7472 |
|
|
/* Number of integer registers we need to save. */
|
7473 |
|
|
int n_intregs = MAX (0, NPARM_REGS (SImode) - first_intreg);
|
7474 |
|
|
/* First unnamed SFmode float reg */
|
7475 |
|
|
int first_floatreg = crtl->args.info.arg_count[(int) SH_ARG_FLOAT];
|
7476 |
|
|
/* Number of SFmode float regs to save. */
|
7477 |
|
|
int n_floatregs = MAX (0, NPARM_REGS (SFmode) - first_floatreg);
|
7478 |
|
|
rtx regbuf, fpregs;
|
7479 |
|
|
int bufsize, regno;
|
7480 |
|
|
alias_set_type alias_set;
|
7481 |
|
|
|
7482 |
|
|
if (TARGET_SH5)
|
7483 |
|
|
{
|
7484 |
|
|
if (n_intregs)
|
7485 |
|
|
{
|
7486 |
|
|
int pushregs = n_intregs;
|
7487 |
|
|
|
7488 |
|
|
while (pushregs < NPARM_REGS (SImode) - 1
|
7489 |
|
|
&& (CALL_COOKIE_INT_REG_GET
|
7490 |
|
|
(crtl->args.info.call_cookie,
|
7491 |
|
|
NPARM_REGS (SImode) - pushregs)
|
7492 |
|
|
== 1))
|
7493 |
|
|
{
|
7494 |
|
|
crtl->args.info.call_cookie
|
7495 |
|
|
&= ~ CALL_COOKIE_INT_REG (NPARM_REGS (SImode)
|
7496 |
|
|
- pushregs, 1);
|
7497 |
|
|
pushregs++;
|
7498 |
|
|
}
|
7499 |
|
|
|
7500 |
|
|
if (pushregs == NPARM_REGS (SImode))
|
7501 |
|
|
crtl->args.info.call_cookie
|
7502 |
|
|
|= (CALL_COOKIE_INT_REG (0, 1)
|
7503 |
|
|
| CALL_COOKIE_STACKSEQ (pushregs - 1));
|
7504 |
|
|
else
|
7505 |
|
|
crtl->args.info.call_cookie
|
7506 |
|
|
|= CALL_COOKIE_STACKSEQ (pushregs);
|
7507 |
|
|
|
7508 |
|
|
crtl->args.pretend_args_size += 8 * n_intregs;
|
7509 |
|
|
}
|
7510 |
|
|
if (TARGET_SHCOMPACT)
|
7511 |
|
|
return const0_rtx;
|
7512 |
|
|
}
|
7513 |
|
|
|
7514 |
|
|
if (! TARGET_SH2E && ! TARGET_SH4 && ! TARGET_SH5)
|
7515 |
|
|
{
|
7516 |
|
|
error ("__builtin_saveregs not supported by this subtarget");
|
7517 |
|
|
return const0_rtx;
|
7518 |
|
|
}
|
7519 |
|
|
|
7520 |
|
|
if (TARGET_SHMEDIA)
|
7521 |
|
|
n_floatregs = 0;
|
7522 |
|
|
|
7523 |
|
|
/* Allocate block of memory for the regs. */
|
7524 |
|
|
/* ??? If n_intregs + n_floatregs == 0, should we allocate at least 1 byte?
|
7525 |
|
|
Or can assign_stack_local accept a 0 SIZE argument? */
|
7526 |
|
|
bufsize = (n_intregs * UNITS_PER_WORD) + (n_floatregs * UNITS_PER_WORD);
|
7527 |
|
|
|
7528 |
|
|
if (TARGET_SHMEDIA)
|
7529 |
|
|
regbuf = gen_frame_mem (BLKmode, gen_rtx_REG (Pmode, ARG_POINTER_REGNUM));
|
7530 |
|
|
else if (n_floatregs & 1)
|
7531 |
|
|
{
|
7532 |
|
|
rtx addr;
|
7533 |
|
|
|
7534 |
|
|
regbuf = assign_stack_local (BLKmode, bufsize + UNITS_PER_WORD, 0);
|
7535 |
|
|
addr = copy_to_mode_reg (Pmode, XEXP (regbuf, 0));
|
7536 |
|
|
emit_insn (gen_iorsi3 (addr, addr, GEN_INT (UNITS_PER_WORD)));
|
7537 |
|
|
regbuf = change_address (regbuf, BLKmode, addr);
|
7538 |
|
|
}
|
7539 |
|
|
else if (STACK_BOUNDARY < 64 && TARGET_FPU_DOUBLE && n_floatregs)
|
7540 |
|
|
{
|
7541 |
|
|
rtx addr, mask;
|
7542 |
|
|
|
7543 |
|
|
regbuf = assign_stack_local (BLKmode, bufsize + UNITS_PER_WORD, 0);
|
7544 |
|
|
addr = copy_to_mode_reg (Pmode, plus_constant (XEXP (regbuf, 0), 4));
|
7545 |
|
|
mask = copy_to_mode_reg (Pmode, GEN_INT (-8));
|
7546 |
|
|
emit_insn (gen_andsi3 (addr, addr, mask));
|
7547 |
|
|
regbuf = change_address (regbuf, BLKmode, addr);
|
7548 |
|
|
}
|
7549 |
|
|
else
|
7550 |
|
|
regbuf = assign_stack_local (BLKmode, bufsize, TARGET_FPU_DOUBLE ? 64 : 0);
|
7551 |
|
|
alias_set = get_varargs_alias_set ();
|
7552 |
|
|
set_mem_alias_set (regbuf, alias_set);
|
7553 |
|
|
|
7554 |
|
|
/* Save int args.
|
7555 |
|
|
This is optimized to only save the regs that are necessary. Explicitly
|
7556 |
|
|
named args need not be saved. */
|
7557 |
|
|
if (n_intregs > 0)
|
7558 |
|
|
move_block_from_reg (BASE_ARG_REG (SImode) + first_intreg,
|
7559 |
|
|
adjust_address (regbuf, BLKmode,
|
7560 |
|
|
n_floatregs * UNITS_PER_WORD),
|
7561 |
|
|
n_intregs);
|
7562 |
|
|
|
7563 |
|
|
if (TARGET_SHMEDIA)
|
7564 |
|
|
/* Return the address of the regbuf. */
|
7565 |
|
|
return XEXP (regbuf, 0);
|
7566 |
|
|
|
7567 |
|
|
/* Save float args.
|
7568 |
|
|
This is optimized to only save the regs that are necessary. Explicitly
|
7569 |
|
|
named args need not be saved.
|
7570 |
|
|
We explicitly build a pointer to the buffer because it halves the insn
|
7571 |
|
|
count when not optimizing (otherwise the pointer is built for each reg
|
7572 |
|
|
saved).
|
7573 |
|
|
We emit the moves in reverse order so that we can use predecrement. */
|
7574 |
|
|
|
7575 |
|
|
fpregs = copy_to_mode_reg (Pmode,
|
7576 |
|
|
plus_constant (XEXP (regbuf, 0),
|
7577 |
|
|
n_floatregs * UNITS_PER_WORD));
|
7578 |
|
|
if (TARGET_SH4 || TARGET_SH2A_DOUBLE)
|
7579 |
|
|
{
|
7580 |
|
|
rtx mem;
|
7581 |
|
|
for (regno = NPARM_REGS (DFmode) - 2; regno >= first_floatreg; regno -= 2)
|
7582 |
|
|
{
|
7583 |
|
|
emit_insn (gen_addsi3 (fpregs, fpregs,
|
7584 |
|
|
GEN_INT (-2 * UNITS_PER_WORD)));
|
7585 |
|
|
mem = change_address (regbuf, DFmode, fpregs);
|
7586 |
|
|
emit_move_insn (mem,
|
7587 |
|
|
gen_rtx_REG (DFmode, BASE_ARG_REG (DFmode) + regno));
|
7588 |
|
|
}
|
7589 |
|
|
regno = first_floatreg;
|
7590 |
|
|
if (regno & 1)
|
7591 |
|
|
{
|
7592 |
|
|
emit_insn (gen_addsi3 (fpregs, fpregs, GEN_INT (-UNITS_PER_WORD)));
|
7593 |
|
|
mem = change_address (regbuf, SFmode, fpregs);
|
7594 |
|
|
emit_move_insn (mem,
|
7595 |
|
|
gen_rtx_REG (SFmode, BASE_ARG_REG (SFmode) + regno
|
7596 |
|
|
- (TARGET_LITTLE_ENDIAN != 0)));
|
7597 |
|
|
}
|
7598 |
|
|
}
|
7599 |
|
|
else
|
7600 |
|
|
for (regno = NPARM_REGS (SFmode) - 1; regno >= first_floatreg; regno--)
|
7601 |
|
|
{
|
7602 |
|
|
rtx mem;
|
7603 |
|
|
|
7604 |
|
|
emit_insn (gen_addsi3 (fpregs, fpregs, GEN_INT (-UNITS_PER_WORD)));
|
7605 |
|
|
mem = change_address (regbuf, SFmode, fpregs);
|
7606 |
|
|
emit_move_insn (mem,
|
7607 |
|
|
gen_rtx_REG (SFmode, BASE_ARG_REG (SFmode) + regno));
|
7608 |
|
|
}
|
7609 |
|
|
|
7610 |
|
|
/* Return the address of the regbuf. */
|
7611 |
|
|
return XEXP (regbuf, 0);
|
7612 |
|
|
}
|
7613 |
|
|
|
7614 |
|
|
/* Define the `__builtin_va_list' type for the ABI. */
|
7615 |
|
|
|
7616 |
|
|
static tree
|
7617 |
|
|
sh_build_builtin_va_list (void)
|
7618 |
|
|
{
|
7619 |
|
|
tree f_next_o, f_next_o_limit, f_next_fp, f_next_fp_limit, f_next_stack;
|
7620 |
|
|
tree record;
|
7621 |
|
|
|
7622 |
|
|
if (TARGET_SH5 || (! TARGET_SH2E && ! TARGET_SH4)
|
7623 |
|
|
|| TARGET_HITACHI || sh_cfun_attr_renesas_p ())
|
7624 |
|
|
return ptr_type_node;
|
7625 |
|
|
|
7626 |
|
|
record = (*lang_hooks.types.make_type) (RECORD_TYPE);
|
7627 |
|
|
|
7628 |
|
|
f_next_o = build_decl (BUILTINS_LOCATION,
|
7629 |
|
|
FIELD_DECL, get_identifier ("__va_next_o"),
|
7630 |
|
|
ptr_type_node);
|
7631 |
|
|
f_next_o_limit = build_decl (BUILTINS_LOCATION,
|
7632 |
|
|
FIELD_DECL,
|
7633 |
|
|
get_identifier ("__va_next_o_limit"),
|
7634 |
|
|
ptr_type_node);
|
7635 |
|
|
f_next_fp = build_decl (BUILTINS_LOCATION,
|
7636 |
|
|
FIELD_DECL, get_identifier ("__va_next_fp"),
|
7637 |
|
|
ptr_type_node);
|
7638 |
|
|
f_next_fp_limit = build_decl (BUILTINS_LOCATION,
|
7639 |
|
|
FIELD_DECL,
|
7640 |
|
|
get_identifier ("__va_next_fp_limit"),
|
7641 |
|
|
ptr_type_node);
|
7642 |
|
|
f_next_stack = build_decl (BUILTINS_LOCATION,
|
7643 |
|
|
FIELD_DECL, get_identifier ("__va_next_stack"),
|
7644 |
|
|
ptr_type_node);
|
7645 |
|
|
|
7646 |
|
|
DECL_FIELD_CONTEXT (f_next_o) = record;
|
7647 |
|
|
DECL_FIELD_CONTEXT (f_next_o_limit) = record;
|
7648 |
|
|
DECL_FIELD_CONTEXT (f_next_fp) = record;
|
7649 |
|
|
DECL_FIELD_CONTEXT (f_next_fp_limit) = record;
|
7650 |
|
|
DECL_FIELD_CONTEXT (f_next_stack) = record;
|
7651 |
|
|
|
7652 |
|
|
TYPE_FIELDS (record) = f_next_o;
|
7653 |
|
|
TREE_CHAIN (f_next_o) = f_next_o_limit;
|
7654 |
|
|
TREE_CHAIN (f_next_o_limit) = f_next_fp;
|
7655 |
|
|
TREE_CHAIN (f_next_fp) = f_next_fp_limit;
|
7656 |
|
|
TREE_CHAIN (f_next_fp_limit) = f_next_stack;
|
7657 |
|
|
|
7658 |
|
|
layout_type (record);
|
7659 |
|
|
|
7660 |
|
|
return record;
|
7661 |
|
|
}
|
7662 |
|
|
|
7663 |
|
|
/* Implement `va_start' for varargs and stdarg. */
|
7664 |
|
|
|
7665 |
|
|
static void
|
7666 |
|
|
sh_va_start (tree valist, rtx nextarg)
|
7667 |
|
|
{
|
7668 |
|
|
tree f_next_o, f_next_o_limit, f_next_fp, f_next_fp_limit, f_next_stack;
|
7669 |
|
|
tree next_o, next_o_limit, next_fp, next_fp_limit, next_stack;
|
7670 |
|
|
tree t, u;
|
7671 |
|
|
int nfp, nint;
|
7672 |
|
|
|
7673 |
|
|
if (TARGET_SH5)
|
7674 |
|
|
{
|
7675 |
|
|
expand_builtin_saveregs ();
|
7676 |
|
|
std_expand_builtin_va_start (valist, nextarg);
|
7677 |
|
|
return;
|
7678 |
|
|
}
|
7679 |
|
|
|
7680 |
|
|
if ((! TARGET_SH2E && ! TARGET_SH4)
|
7681 |
|
|
|| TARGET_HITACHI || sh_cfun_attr_renesas_p ())
|
7682 |
|
|
{
|
7683 |
|
|
std_expand_builtin_va_start (valist, nextarg);
|
7684 |
|
|
return;
|
7685 |
|
|
}
|
7686 |
|
|
|
7687 |
|
|
f_next_o = TYPE_FIELDS (va_list_type_node);
|
7688 |
|
|
f_next_o_limit = TREE_CHAIN (f_next_o);
|
7689 |
|
|
f_next_fp = TREE_CHAIN (f_next_o_limit);
|
7690 |
|
|
f_next_fp_limit = TREE_CHAIN (f_next_fp);
|
7691 |
|
|
f_next_stack = TREE_CHAIN (f_next_fp_limit);
|
7692 |
|
|
|
7693 |
|
|
next_o = build3 (COMPONENT_REF, TREE_TYPE (f_next_o), valist, f_next_o,
|
7694 |
|
|
NULL_TREE);
|
7695 |
|
|
next_o_limit = build3 (COMPONENT_REF, TREE_TYPE (f_next_o_limit),
|
7696 |
|
|
valist, f_next_o_limit, NULL_TREE);
|
7697 |
|
|
next_fp = build3 (COMPONENT_REF, TREE_TYPE (f_next_fp), valist, f_next_fp,
|
7698 |
|
|
NULL_TREE);
|
7699 |
|
|
next_fp_limit = build3 (COMPONENT_REF, TREE_TYPE (f_next_fp_limit),
|
7700 |
|
|
valist, f_next_fp_limit, NULL_TREE);
|
7701 |
|
|
next_stack = build3 (COMPONENT_REF, TREE_TYPE (f_next_stack),
|
7702 |
|
|
valist, f_next_stack, NULL_TREE);
|
7703 |
|
|
|
7704 |
|
|
/* Call __builtin_saveregs. */
|
7705 |
|
|
u = make_tree (sizetype, expand_builtin_saveregs ());
|
7706 |
|
|
u = fold_convert (ptr_type_node, u);
|
7707 |
|
|
t = build2 (MODIFY_EXPR, ptr_type_node, next_fp, u);
|
7708 |
|
|
TREE_SIDE_EFFECTS (t) = 1;
|
7709 |
|
|
expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
7710 |
|
|
|
7711 |
|
|
nfp = crtl->args.info.arg_count[SH_ARG_FLOAT];
|
7712 |
|
|
if (nfp < 8)
|
7713 |
|
|
nfp = 8 - nfp;
|
7714 |
|
|
else
|
7715 |
|
|
nfp = 0;
|
7716 |
|
|
u = fold_build2 (POINTER_PLUS_EXPR, ptr_type_node, u,
|
7717 |
|
|
size_int (UNITS_PER_WORD * nfp));
|
7718 |
|
|
t = build2 (MODIFY_EXPR, ptr_type_node, next_fp_limit, u);
|
7719 |
|
|
TREE_SIDE_EFFECTS (t) = 1;
|
7720 |
|
|
expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
7721 |
|
|
|
7722 |
|
|
t = build2 (MODIFY_EXPR, ptr_type_node, next_o, u);
|
7723 |
|
|
TREE_SIDE_EFFECTS (t) = 1;
|
7724 |
|
|
expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
7725 |
|
|
|
7726 |
|
|
nint = crtl->args.info.arg_count[SH_ARG_INT];
|
7727 |
|
|
if (nint < 4)
|
7728 |
|
|
nint = 4 - nint;
|
7729 |
|
|
else
|
7730 |
|
|
nint = 0;
|
7731 |
|
|
u = fold_build2 (POINTER_PLUS_EXPR, ptr_type_node, u,
|
7732 |
|
|
size_int (UNITS_PER_WORD * nint));
|
7733 |
|
|
t = build2 (MODIFY_EXPR, ptr_type_node, next_o_limit, u);
|
7734 |
|
|
TREE_SIDE_EFFECTS (t) = 1;
|
7735 |
|
|
expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
7736 |
|
|
|
7737 |
|
|
u = make_tree (ptr_type_node, nextarg);
|
7738 |
|
|
t = build2 (MODIFY_EXPR, ptr_type_node, next_stack, u);
|
7739 |
|
|
TREE_SIDE_EFFECTS (t) = 1;
|
7740 |
|
|
expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
|
7741 |
|
|
}
|
7742 |
|
|
|
7743 |
|
|
/* TYPE is a RECORD_TYPE. If there is only a single nonzero-sized
|
7744 |
|
|
member, return it. */
|
7745 |
|
|
static tree
|
7746 |
|
|
find_sole_member (tree type)
|
7747 |
|
|
{
|
7748 |
|
|
tree field, member = NULL_TREE;
|
7749 |
|
|
|
7750 |
|
|
for (field = TYPE_FIELDS (type); field; field = TREE_CHAIN (field))
|
7751 |
|
|
{
|
7752 |
|
|
if (TREE_CODE (field) != FIELD_DECL)
|
7753 |
|
|
continue;
|
7754 |
|
|
if (!DECL_SIZE (field))
|
7755 |
|
|
return NULL_TREE;
|
7756 |
|
|
if (integer_zerop (DECL_SIZE (field)))
|
7757 |
|
|
continue;
|
7758 |
|
|
if (member)
|
7759 |
|
|
return NULL_TREE;
|
7760 |
|
|
member = field;
|
7761 |
|
|
}
|
7762 |
|
|
return member;
|
7763 |
|
|
}
|
7764 |
|
|
/* Implement `va_arg'. */
|
7765 |
|
|
|
7766 |
|
|
static tree
|
7767 |
|
|
sh_gimplify_va_arg_expr (tree valist, tree type, gimple_seq *pre_p,
|
7768 |
|
|
gimple_seq *post_p ATTRIBUTE_UNUSED)
|
7769 |
|
|
{
|
7770 |
|
|
HOST_WIDE_INT size, rsize;
|
7771 |
|
|
tree tmp, pptr_type_node;
|
7772 |
|
|
tree addr, lab_over = NULL, result = NULL;
|
7773 |
|
|
int pass_by_ref = targetm.calls.must_pass_in_stack (TYPE_MODE (type), type);
|
7774 |
|
|
tree eff_type;
|
7775 |
|
|
|
7776 |
|
|
if (pass_by_ref)
|
7777 |
|
|
type = build_pointer_type (type);
|
7778 |
|
|
|
7779 |
|
|
size = int_size_in_bytes (type);
|
7780 |
|
|
rsize = (size + UNITS_PER_WORD - 1) & -UNITS_PER_WORD;
|
7781 |
|
|
pptr_type_node = build_pointer_type (ptr_type_node);
|
7782 |
|
|
|
7783 |
|
|
if (! TARGET_SH5 && (TARGET_SH2E || TARGET_SH4)
|
7784 |
|
|
&& ! (TARGET_HITACHI || sh_cfun_attr_renesas_p ()))
|
7785 |
|
|
{
|
7786 |
|
|
tree f_next_o, f_next_o_limit, f_next_fp, f_next_fp_limit, f_next_stack;
|
7787 |
|
|
tree next_o, next_o_limit, next_fp, next_fp_limit, next_stack;
|
7788 |
|
|
int pass_as_float;
|
7789 |
|
|
tree lab_false;
|
7790 |
|
|
tree member;
|
7791 |
|
|
|
7792 |
|
|
f_next_o = TYPE_FIELDS (va_list_type_node);
|
7793 |
|
|
f_next_o_limit = TREE_CHAIN (f_next_o);
|
7794 |
|
|
f_next_fp = TREE_CHAIN (f_next_o_limit);
|
7795 |
|
|
f_next_fp_limit = TREE_CHAIN (f_next_fp);
|
7796 |
|
|
f_next_stack = TREE_CHAIN (f_next_fp_limit);
|
7797 |
|
|
|
7798 |
|
|
next_o = build3 (COMPONENT_REF, TREE_TYPE (f_next_o), valist, f_next_o,
|
7799 |
|
|
NULL_TREE);
|
7800 |
|
|
next_o_limit = build3 (COMPONENT_REF, TREE_TYPE (f_next_o_limit),
|
7801 |
|
|
valist, f_next_o_limit, NULL_TREE);
|
7802 |
|
|
next_fp = build3 (COMPONENT_REF, TREE_TYPE (f_next_fp),
|
7803 |
|
|
valist, f_next_fp, NULL_TREE);
|
7804 |
|
|
next_fp_limit = build3 (COMPONENT_REF, TREE_TYPE (f_next_fp_limit),
|
7805 |
|
|
valist, f_next_fp_limit, NULL_TREE);
|
7806 |
|
|
next_stack = build3 (COMPONENT_REF, TREE_TYPE (f_next_stack),
|
7807 |
|
|
valist, f_next_stack, NULL_TREE);
|
7808 |
|
|
|
7809 |
|
|
/* Structures with a single member with a distinct mode are passed
|
7810 |
|
|
like their member. This is relevant if the latter has a REAL_TYPE
|
7811 |
|
|
or COMPLEX_TYPE type. */
|
7812 |
|
|
eff_type = type;
|
7813 |
|
|
while (TREE_CODE (eff_type) == RECORD_TYPE
|
7814 |
|
|
&& (member = find_sole_member (eff_type))
|
7815 |
|
|
&& (TREE_CODE (TREE_TYPE (member)) == REAL_TYPE
|
7816 |
|
|
|| TREE_CODE (TREE_TYPE (member)) == COMPLEX_TYPE
|
7817 |
|
|
|| TREE_CODE (TREE_TYPE (member)) == RECORD_TYPE))
|
7818 |
|
|
{
|
7819 |
|
|
tree field_type = TREE_TYPE (member);
|
7820 |
|
|
|
7821 |
|
|
if (TYPE_MODE (eff_type) == TYPE_MODE (field_type))
|
7822 |
|
|
eff_type = field_type;
|
7823 |
|
|
else
|
7824 |
|
|
{
|
7825 |
|
|
gcc_assert ((TYPE_ALIGN (eff_type)
|
7826 |
|
|
< GET_MODE_ALIGNMENT (TYPE_MODE (field_type)))
|
7827 |
|
|
|| (TYPE_ALIGN (eff_type)
|
7828 |
|
|
> GET_MODE_BITSIZE (TYPE_MODE (field_type))));
|
7829 |
|
|
break;
|
7830 |
|
|
}
|
7831 |
|
|
}
|
7832 |
|
|
|
7833 |
|
|
if (TARGET_SH4 || TARGET_SH2A_DOUBLE)
|
7834 |
|
|
{
|
7835 |
|
|
pass_as_float = ((TREE_CODE (eff_type) == REAL_TYPE && size <= 8)
|
7836 |
|
|
|| (TREE_CODE (eff_type) == COMPLEX_TYPE
|
7837 |
|
|
&& TREE_CODE (TREE_TYPE (eff_type)) == REAL_TYPE
|
7838 |
|
|
&& size <= 16));
|
7839 |
|
|
}
|
7840 |
|
|
else
|
7841 |
|
|
{
|
7842 |
|
|
pass_as_float = (TREE_CODE (eff_type) == REAL_TYPE && size == 4);
|
7843 |
|
|
}
|
7844 |
|
|
|
7845 |
|
|
addr = create_tmp_var (pptr_type_node, NULL);
|
7846 |
|
|
lab_false = create_artificial_label (UNKNOWN_LOCATION);
|
7847 |
|
|
lab_over = create_artificial_label (UNKNOWN_LOCATION);
|
7848 |
|
|
|
7849 |
|
|
valist = build1 (INDIRECT_REF, ptr_type_node, addr);
|
7850 |
|
|
|
7851 |
|
|
if (pass_as_float)
|
7852 |
|
|
{
|
7853 |
|
|
tree next_fp_tmp = create_tmp_var (TREE_TYPE (f_next_fp), NULL);
|
7854 |
|
|
tree cmp;
|
7855 |
|
|
bool is_double = size == 8 && TREE_CODE (eff_type) == REAL_TYPE;
|
7856 |
|
|
|
7857 |
|
|
tmp = build1 (ADDR_EXPR, pptr_type_node, unshare_expr (next_fp));
|
7858 |
|
|
gimplify_assign (unshare_expr (addr), tmp, pre_p);
|
7859 |
|
|
|
7860 |
|
|
gimplify_assign (unshare_expr (next_fp_tmp), valist, pre_p);
|
7861 |
|
|
tmp = next_fp_limit;
|
7862 |
|
|
if (size > 4 && !is_double)
|
7863 |
|
|
tmp = build2 (POINTER_PLUS_EXPR, TREE_TYPE (tmp),
|
7864 |
|
|
unshare_expr (tmp), size_int (4 - size));
|
7865 |
|
|
tmp = build2 (GE_EXPR, boolean_type_node,
|
7866 |
|
|
unshare_expr (next_fp_tmp), unshare_expr (tmp));
|
7867 |
|
|
cmp = build3 (COND_EXPR, void_type_node, tmp,
|
7868 |
|
|
build1 (GOTO_EXPR, void_type_node,
|
7869 |
|
|
unshare_expr (lab_false)), NULL_TREE);
|
7870 |
|
|
if (!is_double)
|
7871 |
|
|
gimplify_and_add (cmp, pre_p);
|
7872 |
|
|
|
7873 |
|
|
if (TYPE_ALIGN (eff_type) > BITS_PER_WORD
|
7874 |
|
|
|| (is_double || size == 16))
|
7875 |
|
|
{
|
7876 |
|
|
tmp = fold_convert (sizetype, next_fp_tmp);
|
7877 |
|
|
tmp = build2 (BIT_AND_EXPR, sizetype, tmp,
|
7878 |
|
|
size_int (UNITS_PER_WORD));
|
7879 |
|
|
tmp = build2 (POINTER_PLUS_EXPR, ptr_type_node,
|
7880 |
|
|
unshare_expr (next_fp_tmp), tmp);
|
7881 |
|
|
gimplify_assign (unshare_expr (next_fp_tmp), tmp, pre_p);
|
7882 |
|
|
}
|
7883 |
|
|
if (is_double)
|
7884 |
|
|
gimplify_and_add (cmp, pre_p);
|
7885 |
|
|
|
7886 |
|
|
#ifdef FUNCTION_ARG_SCmode_WART
|
7887 |
|
|
if (TYPE_MODE (eff_type) == SCmode
|
7888 |
|
|
&& TARGET_SH4 && TARGET_LITTLE_ENDIAN)
|
7889 |
|
|
{
|
7890 |
|
|
tree subtype = TREE_TYPE (eff_type);
|
7891 |
|
|
tree real, imag;
|
7892 |
|
|
|
7893 |
|
|
imag
|
7894 |
|
|
= std_gimplify_va_arg_expr (next_fp_tmp, subtype, pre_p, NULL);
|
7895 |
|
|
imag = get_initialized_tmp_var (imag, pre_p, NULL);
|
7896 |
|
|
|
7897 |
|
|
real
|
7898 |
|
|
= std_gimplify_va_arg_expr (next_fp_tmp, subtype, pre_p, NULL);
|
7899 |
|
|
real = get_initialized_tmp_var (real, pre_p, NULL);
|
7900 |
|
|
|
7901 |
|
|
result = build2 (COMPLEX_EXPR, eff_type, real, imag);
|
7902 |
|
|
if (type != eff_type)
|
7903 |
|
|
result = build1 (VIEW_CONVERT_EXPR, type, result);
|
7904 |
|
|
result = get_initialized_tmp_var (result, pre_p, NULL);
|
7905 |
|
|
}
|
7906 |
|
|
#endif /* FUNCTION_ARG_SCmode_WART */
|
7907 |
|
|
|
7908 |
|
|
tmp = build1 (GOTO_EXPR, void_type_node, unshare_expr (lab_over));
|
7909 |
|
|
gimplify_and_add (tmp, pre_p);
|
7910 |
|
|
|
7911 |
|
|
tmp = build1 (LABEL_EXPR, void_type_node, unshare_expr (lab_false));
|
7912 |
|
|
gimplify_and_add (tmp, pre_p);
|
7913 |
|
|
|
7914 |
|
|
tmp = build1 (ADDR_EXPR, pptr_type_node, unshare_expr (next_stack));
|
7915 |
|
|
gimplify_assign (unshare_expr (addr), tmp, pre_p);
|
7916 |
|
|
gimplify_assign (unshare_expr (next_fp_tmp),
|
7917 |
|
|
unshare_expr (valist), pre_p);
|
7918 |
|
|
|
7919 |
|
|
gimplify_assign (unshare_expr (valist),
|
7920 |
|
|
unshare_expr (next_fp_tmp), post_p);
|
7921 |
|
|
valist = next_fp_tmp;
|
7922 |
|
|
}
|
7923 |
|
|
else
|
7924 |
|
|
{
|
7925 |
|
|
tmp = build2 (POINTER_PLUS_EXPR, ptr_type_node,
|
7926 |
|
|
unshare_expr (next_o), size_int (rsize));
|
7927 |
|
|
tmp = build2 (GT_EXPR, boolean_type_node, tmp,
|
7928 |
|
|
unshare_expr (next_o_limit));
|
7929 |
|
|
tmp = build3 (COND_EXPR, void_type_node, tmp,
|
7930 |
|
|
build1 (GOTO_EXPR, void_type_node,
|
7931 |
|
|
unshare_expr (lab_false)),
|
7932 |
|
|
NULL_TREE);
|
7933 |
|
|
gimplify_and_add (tmp, pre_p);
|
7934 |
|
|
|
7935 |
|
|
tmp = build1 (ADDR_EXPR, pptr_type_node, unshare_expr (next_o));
|
7936 |
|
|
gimplify_assign (unshare_expr (addr), tmp, pre_p);
|
7937 |
|
|
|
7938 |
|
|
tmp = build1 (GOTO_EXPR, void_type_node, unshare_expr (lab_over));
|
7939 |
|
|
gimplify_and_add (tmp, pre_p);
|
7940 |
|
|
|
7941 |
|
|
tmp = build1 (LABEL_EXPR, void_type_node, unshare_expr (lab_false));
|
7942 |
|
|
gimplify_and_add (tmp, pre_p);
|
7943 |
|
|
|
7944 |
|
|
if (size > 4 && ! (TARGET_SH4 || TARGET_SH2A))
|
7945 |
|
|
gimplify_assign (unshare_expr (next_o),
|
7946 |
|
|
unshare_expr (next_o_limit), pre_p);
|
7947 |
|
|
|
7948 |
|
|
tmp = build1 (ADDR_EXPR, pptr_type_node, unshare_expr (next_stack));
|
7949 |
|
|
gimplify_assign (unshare_expr (addr), tmp, pre_p);
|
7950 |
|
|
}
|
7951 |
|
|
|
7952 |
|
|
if (!result)
|
7953 |
|
|
{
|
7954 |
|
|
tmp = build1 (LABEL_EXPR, void_type_node, unshare_expr (lab_over));
|
7955 |
|
|
gimplify_and_add (tmp, pre_p);
|
7956 |
|
|
}
|
7957 |
|
|
}
|
7958 |
|
|
|
7959 |
|
|
/* ??? In va-sh.h, there had been code to make values larger than
|
7960 |
|
|
size 8 indirect. This does not match the FUNCTION_ARG macros. */
|
7961 |
|
|
|
7962 |
|
|
tmp = std_gimplify_va_arg_expr (valist, type, pre_p, NULL);
|
7963 |
|
|
if (result)
|
7964 |
|
|
{
|
7965 |
|
|
gimplify_assign (result, tmp, pre_p);
|
7966 |
|
|
result = build1 (NOP_EXPR, TREE_TYPE (result), result);
|
7967 |
|
|
tmp = build1 (LABEL_EXPR, void_type_node, unshare_expr (lab_over));
|
7968 |
|
|
gimplify_and_add (tmp, pre_p);
|
7969 |
|
|
}
|
7970 |
|
|
else
|
7971 |
|
|
result = tmp;
|
7972 |
|
|
|
7973 |
|
|
if (pass_by_ref)
|
7974 |
|
|
result = build_va_arg_indirect_ref (result);
|
7975 |
|
|
|
7976 |
|
|
return result;
|
7977 |
|
|
}
|
7978 |
|
|
|
7979 |
|
|
/* 64 bit floating points memory transfers are paired single precision loads
|
7980 |
|
|
or store. So DWARF information needs fixing in little endian (unless
|
7981 |
|
|
PR=SZ=1 in FPSCR). */
|
7982 |
|
|
rtx
|
7983 |
|
|
sh_dwarf_register_span (rtx reg)
|
7984 |
|
|
{
|
7985 |
|
|
unsigned regno = REGNO (reg);
|
7986 |
|
|
|
7987 |
|
|
if (WORDS_BIG_ENDIAN || GET_MODE (reg) != DFmode)
|
7988 |
|
|
return NULL_RTX;
|
7989 |
|
|
|
7990 |
|
|
return
|
7991 |
|
|
gen_rtx_PARALLEL (VOIDmode,
|
7992 |
|
|
gen_rtvec (2,
|
7993 |
|
|
gen_rtx_REG (SFmode,
|
7994 |
|
|
DBX_REGISTER_NUMBER (regno+1)),
|
7995 |
|
|
gen_rtx_REG (SFmode,
|
7996 |
|
|
DBX_REGISTER_NUMBER (regno))));
|
7997 |
|
|
}
|
7998 |
|
|
|
7999 |
|
|
static enum machine_mode
|
8000 |
|
|
sh_promote_function_mode (const_tree type, enum machine_mode mode,
|
8001 |
|
|
int *punsignedp, const_tree funtype,
|
8002 |
|
|
int for_return ATTRIBUTE_UNUSED)
|
8003 |
|
|
{
|
8004 |
|
|
if (sh_promote_prototypes (funtype))
|
8005 |
|
|
return promote_mode (type, mode, punsignedp);
|
8006 |
|
|
else
|
8007 |
|
|
return mode;
|
8008 |
|
|
}
|
8009 |
|
|
|
8010 |
|
|
static bool
|
8011 |
|
|
sh_promote_prototypes (const_tree type)
|
8012 |
|
|
{
|
8013 |
|
|
if (TARGET_HITACHI)
|
8014 |
|
|
return 0;
|
8015 |
|
|
if (! type)
|
8016 |
|
|
return 1;
|
8017 |
|
|
return ! sh_attr_renesas_p (type);
|
8018 |
|
|
}
|
8019 |
|
|
|
8020 |
|
|
/* Whether an argument must be passed by reference. On SHcompact, we
|
8021 |
|
|
pretend arguments wider than 32-bits that would have been passed in
|
8022 |
|
|
registers are passed by reference, so that an SHmedia trampoline
|
8023 |
|
|
loads them into the full 64-bits registers. */
|
8024 |
|
|
|
8025 |
|
|
static int
|
8026 |
|
|
shcompact_byref (const CUMULATIVE_ARGS *cum, enum machine_mode mode,
|
8027 |
|
|
const_tree type, bool named)
|
8028 |
|
|
{
|
8029 |
|
|
unsigned HOST_WIDE_INT size;
|
8030 |
|
|
|
8031 |
|
|
if (type)
|
8032 |
|
|
size = int_size_in_bytes (type);
|
8033 |
|
|
else
|
8034 |
|
|
size = GET_MODE_SIZE (mode);
|
8035 |
|
|
|
8036 |
|
|
if (cum->arg_count[SH_ARG_INT] < NPARM_REGS (SImode)
|
8037 |
|
|
&& (!named
|
8038 |
|
|
|| GET_SH_ARG_CLASS (mode) == SH_ARG_INT
|
8039 |
|
|
|| (GET_SH_ARG_CLASS (mode) == SH_ARG_FLOAT
|
8040 |
|
|
&& cum->arg_count[SH_ARG_FLOAT] >= NPARM_REGS (SFmode)))
|
8041 |
|
|
&& size > 4
|
8042 |
|
|
&& !SHCOMPACT_FORCE_ON_STACK (mode, type)
|
8043 |
|
|
&& !SH5_WOULD_BE_PARTIAL_NREGS (*cum, mode, type, named))
|
8044 |
|
|
return size;
|
8045 |
|
|
else
|
8046 |
|
|
return 0;
|
8047 |
|
|
}
|
8048 |
|
|
|
8049 |
|
|
static bool
|
8050 |
|
|
sh_pass_by_reference (CUMULATIVE_ARGS *cum, enum machine_mode mode,
|
8051 |
|
|
const_tree type, bool named)
|
8052 |
|
|
{
|
8053 |
|
|
if (targetm.calls.must_pass_in_stack (mode, type))
|
8054 |
|
|
return true;
|
8055 |
|
|
|
8056 |
|
|
/* ??? std_gimplify_va_arg_expr passes NULL for cum. That function
|
8057 |
|
|
wants to know about pass-by-reference semantics for incoming
|
8058 |
|
|
arguments. */
|
8059 |
|
|
if (! cum)
|
8060 |
|
|
return false;
|
8061 |
|
|
|
8062 |
|
|
if (TARGET_SHCOMPACT)
|
8063 |
|
|
{
|
8064 |
|
|
cum->byref = shcompact_byref (cum, mode, type, named);
|
8065 |
|
|
return cum->byref != 0;
|
8066 |
|
|
}
|
8067 |
|
|
|
8068 |
|
|
return false;
|
8069 |
|
|
}
|
8070 |
|
|
|
8071 |
|
|
static bool
|
8072 |
|
|
sh_callee_copies (CUMULATIVE_ARGS *cum, enum machine_mode mode,
|
8073 |
|
|
const_tree type, bool named ATTRIBUTE_UNUSED)
|
8074 |
|
|
{
|
8075 |
|
|
/* ??? How can it possibly be correct to return true only on the
|
8076 |
|
|
caller side of the equation? Is there someplace else in the
|
8077 |
|
|
sh backend that's magically producing the copies? */
|
8078 |
|
|
return (cum->outgoing
|
8079 |
|
|
&& ((mode == BLKmode ? TYPE_ALIGN (type) : GET_MODE_ALIGNMENT (mode))
|
8080 |
|
|
% SH_MIN_ALIGN_FOR_CALLEE_COPY == 0));
|
8081 |
|
|
}
|
8082 |
|
|
|
8083 |
|
|
static int
|
8084 |
|
|
sh_arg_partial_bytes (CUMULATIVE_ARGS *cum, enum machine_mode mode,
|
8085 |
|
|
tree type, bool named ATTRIBUTE_UNUSED)
|
8086 |
|
|
{
|
8087 |
|
|
int words = 0;
|
8088 |
|
|
|
8089 |
|
|
if (!TARGET_SH5
|
8090 |
|
|
&& PASS_IN_REG_P (*cum, mode, type)
|
8091 |
|
|
&& !(TARGET_SH4 || TARGET_SH2A_DOUBLE)
|
8092 |
|
|
&& (ROUND_REG (*cum, mode)
|
8093 |
|
|
+ (mode != BLKmode
|
8094 |
|
|
? ROUND_ADVANCE (GET_MODE_SIZE (mode))
|
8095 |
|
|
: ROUND_ADVANCE (int_size_in_bytes (type)))
|
8096 |
|
|
> NPARM_REGS (mode)))
|
8097 |
|
|
words = NPARM_REGS (mode) - ROUND_REG (*cum, mode);
|
8098 |
|
|
|
8099 |
|
|
else if (!TARGET_SHCOMPACT
|
8100 |
|
|
&& SH5_WOULD_BE_PARTIAL_NREGS (*cum, mode, type, named))
|
8101 |
|
|
words = NPARM_REGS (SImode) - cum->arg_count[SH_ARG_INT];
|
8102 |
|
|
|
8103 |
|
|
return words * UNITS_PER_WORD;
|
8104 |
|
|
}
|
8105 |
|
|
|
8106 |
|
|
|
8107 |
|
|
/* Define where to put the arguments to a function.
|
8108 |
|
|
Value is zero to push the argument on the stack,
|
8109 |
|
|
or a hard register in which to store the argument.
|
8110 |
|
|
|
8111 |
|
|
MODE is the argument's machine mode.
|
8112 |
|
|
TYPE is the data type of the argument (as a tree).
|
8113 |
|
|
This is null for libcalls where that information may
|
8114 |
|
|
not be available.
|
8115 |
|
|
CUM is a variable of type CUMULATIVE_ARGS which gives info about
|
8116 |
|
|
the preceding args and about the function being called.
|
8117 |
|
|
NAMED is nonzero if this argument is a named parameter
|
8118 |
|
|
(otherwise it is an extra parameter matching an ellipsis).
|
8119 |
|
|
|
8120 |
|
|
On SH the first args are normally in registers
|
8121 |
|
|
and the rest are pushed. Any arg that starts within the first
|
8122 |
|
|
NPARM_REGS words is at least partially passed in a register unless
|
8123 |
|
|
its data type forbids. */
|
8124 |
|
|
|
8125 |
|
|
|
8126 |
|
|
rtx
|
8127 |
|
|
sh_function_arg (CUMULATIVE_ARGS *ca, enum machine_mode mode,
|
8128 |
|
|
tree type, int named)
|
8129 |
|
|
{
|
8130 |
|
|
if (! TARGET_SH5 && mode == VOIDmode)
|
8131 |
|
|
return GEN_INT (ca->renesas_abi ? 1 : 0);
|
8132 |
|
|
|
8133 |
|
|
if (! TARGET_SH5
|
8134 |
|
|
&& PASS_IN_REG_P (*ca, mode, type)
|
8135 |
|
|
&& (named || ! (TARGET_HITACHI || ca->renesas_abi)))
|
8136 |
|
|
{
|
8137 |
|
|
int regno;
|
8138 |
|
|
|
8139 |
|
|
if (mode == SCmode && TARGET_SH4 && TARGET_LITTLE_ENDIAN
|
8140 |
|
|
&& (! FUNCTION_ARG_SCmode_WART || (ROUND_REG (*ca, mode) & 1)))
|
8141 |
|
|
{
|
8142 |
|
|
rtx r1 = gen_rtx_EXPR_LIST (VOIDmode,
|
8143 |
|
|
gen_rtx_REG (SFmode,
|
8144 |
|
|
BASE_ARG_REG (mode)
|
8145 |
|
|
+ (ROUND_REG (*ca, mode) ^ 1)),
|
8146 |
|
|
const0_rtx);
|
8147 |
|
|
rtx r2 = gen_rtx_EXPR_LIST (VOIDmode,
|
8148 |
|
|
gen_rtx_REG (SFmode,
|
8149 |
|
|
BASE_ARG_REG (mode)
|
8150 |
|
|
+ ((ROUND_REG (*ca, mode) + 1) ^ 1)),
|
8151 |
|
|
GEN_INT (4));
|
8152 |
|
|
return gen_rtx_PARALLEL(SCmode, gen_rtvec(2, r1, r2));
|
8153 |
|
|
}
|
8154 |
|
|
|
8155 |
|
|
/* If the alignment of a DF value causes an SF register to be
|
8156 |
|
|
skipped, we will use that skipped register for the next SF
|
8157 |
|
|
value. */
|
8158 |
|
|
if ((TARGET_HITACHI || ca->renesas_abi)
|
8159 |
|
|
&& ca->free_single_fp_reg
|
8160 |
|
|
&& mode == SFmode)
|
8161 |
|
|
return gen_rtx_REG (mode, ca->free_single_fp_reg);
|
8162 |
|
|
|
8163 |
|
|
regno = (BASE_ARG_REG (mode) + ROUND_REG (*ca, mode))
|
8164 |
|
|
^ (mode == SFmode && TARGET_SH4
|
8165 |
|
|
&& TARGET_LITTLE_ENDIAN != 0
|
8166 |
|
|
&& ! TARGET_HITACHI && ! ca->renesas_abi);
|
8167 |
|
|
return gen_rtx_REG (mode, regno);
|
8168 |
|
|
|
8169 |
|
|
}
|
8170 |
|
|
|
8171 |
|
|
if (TARGET_SH5)
|
8172 |
|
|
{
|
8173 |
|
|
if (mode == VOIDmode && TARGET_SHCOMPACT)
|
8174 |
|
|
return GEN_INT (ca->call_cookie);
|
8175 |
|
|
|
8176 |
|
|
/* The following test assumes unnamed arguments are promoted to
|
8177 |
|
|
DFmode. */
|
8178 |
|
|
if (mode == SFmode && ca->free_single_fp_reg)
|
8179 |
|
|
return SH5_PROTOTYPED_FLOAT_ARG (*ca, mode, ca->free_single_fp_reg);
|
8180 |
|
|
|
8181 |
|
|
if ((GET_SH_ARG_CLASS (mode) == SH_ARG_FLOAT)
|
8182 |
|
|
&& (named || ! ca->prototype_p)
|
8183 |
|
|
&& ca->arg_count[(int) SH_ARG_FLOAT] < NPARM_REGS (SFmode))
|
8184 |
|
|
{
|
8185 |
|
|
if (! ca->prototype_p && TARGET_SHMEDIA)
|
8186 |
|
|
return SH5_PROTOTYPELESS_FLOAT_ARG (*ca, mode);
|
8187 |
|
|
|
8188 |
|
|
return SH5_PROTOTYPED_FLOAT_ARG (*ca, mode,
|
8189 |
|
|
FIRST_FP_PARM_REG
|
8190 |
|
|
+ ca->arg_count[(int) SH_ARG_FLOAT]);
|
8191 |
|
|
}
|
8192 |
|
|
|
8193 |
|
|
if (ca->arg_count[(int) SH_ARG_INT] < NPARM_REGS (SImode)
|
8194 |
|
|
&& (! TARGET_SHCOMPACT
|
8195 |
|
|
|| (! SHCOMPACT_FORCE_ON_STACK (mode, type)
|
8196 |
|
|
&& ! SH5_WOULD_BE_PARTIAL_NREGS (*ca, mode,
|
8197 |
|
|
type, named))))
|
8198 |
|
|
{
|
8199 |
|
|
return gen_rtx_REG (mode, (FIRST_PARM_REG
|
8200 |
|
|
+ ca->arg_count[(int) SH_ARG_INT]));
|
8201 |
|
|
}
|
8202 |
|
|
|
8203 |
|
|
return 0;
|
8204 |
|
|
}
|
8205 |
|
|
|
8206 |
|
|
return 0;
|
8207 |
|
|
}
|
8208 |
|
|
|
8209 |
|
|
/* Update the data in CUM to advance over an argument
|
8210 |
|
|
of mode MODE and data type TYPE.
|
8211 |
|
|
(TYPE is null for libcalls where that information may not be
|
8212 |
|
|
available.) */
|
8213 |
|
|
|
8214 |
|
|
void
|
8215 |
|
|
sh_function_arg_advance (CUMULATIVE_ARGS *ca, enum machine_mode mode,
|
8216 |
|
|
tree type, int named)
|
8217 |
|
|
{
|
8218 |
|
|
if (ca->force_mem)
|
8219 |
|
|
ca->force_mem = 0;
|
8220 |
|
|
else if (TARGET_SH5)
|
8221 |
|
|
{
|
8222 |
|
|
tree type2 = (ca->byref && type
|
8223 |
|
|
? TREE_TYPE (type)
|
8224 |
|
|
: type);
|
8225 |
|
|
enum machine_mode mode2 = (ca->byref && type
|
8226 |
|
|
? TYPE_MODE (type2)
|
8227 |
|
|
: mode);
|
8228 |
|
|
int dwords = ((ca->byref
|
8229 |
|
|
? ca->byref
|
8230 |
|
|
: mode2 == BLKmode
|
8231 |
|
|
? int_size_in_bytes (type2)
|
8232 |
|
|
: GET_MODE_SIZE (mode2)) + 7) / 8;
|
8233 |
|
|
int numregs = MIN (dwords, NPARM_REGS (SImode)
|
8234 |
|
|
- ca->arg_count[(int) SH_ARG_INT]);
|
8235 |
|
|
|
8236 |
|
|
if (numregs)
|
8237 |
|
|
{
|
8238 |
|
|
ca->arg_count[(int) SH_ARG_INT] += numregs;
|
8239 |
|
|
if (TARGET_SHCOMPACT
|
8240 |
|
|
&& SHCOMPACT_FORCE_ON_STACK (mode2, type2))
|
8241 |
|
|
{
|
8242 |
|
|
ca->call_cookie
|
8243 |
|
|
|= CALL_COOKIE_INT_REG (ca->arg_count[(int) SH_ARG_INT]
|
8244 |
|
|
- numregs, 1);
|
8245 |
|
|
/* N.B. We want this also for outgoing. */
|
8246 |
|
|
ca->stack_regs += numregs;
|
8247 |
|
|
}
|
8248 |
|
|
else if (ca->byref)
|
8249 |
|
|
{
|
8250 |
|
|
if (! ca->outgoing)
|
8251 |
|
|
ca->stack_regs += numregs;
|
8252 |
|
|
ca->byref_regs += numregs;
|
8253 |
|
|
ca->byref = 0;
|
8254 |
|
|
do
|
8255 |
|
|
ca->call_cookie
|
8256 |
|
|
|= CALL_COOKIE_INT_REG (ca->arg_count[(int) SH_ARG_INT]
|
8257 |
|
|
- numregs, 2);
|
8258 |
|
|
while (--numregs);
|
8259 |
|
|
ca->call_cookie
|
8260 |
|
|
|= CALL_COOKIE_INT_REG (ca->arg_count[(int) SH_ARG_INT]
|
8261 |
|
|
- 1, 1);
|
8262 |
|
|
}
|
8263 |
|
|
else if (dwords > numregs)
|
8264 |
|
|
{
|
8265 |
|
|
int pushregs = numregs;
|
8266 |
|
|
|
8267 |
|
|
if (TARGET_SHCOMPACT)
|
8268 |
|
|
ca->stack_regs += numregs;
|
8269 |
|
|
while (pushregs < NPARM_REGS (SImode) - 1
|
8270 |
|
|
&& (CALL_COOKIE_INT_REG_GET
|
8271 |
|
|
(ca->call_cookie,
|
8272 |
|
|
NPARM_REGS (SImode) - pushregs)
|
8273 |
|
|
== 1))
|
8274 |
|
|
{
|
8275 |
|
|
ca->call_cookie
|
8276 |
|
|
&= ~ CALL_COOKIE_INT_REG (NPARM_REGS (SImode)
|
8277 |
|
|
- pushregs, 1);
|
8278 |
|
|
pushregs++;
|
8279 |
|
|
}
|
8280 |
|
|
if (numregs == NPARM_REGS (SImode))
|
8281 |
|
|
ca->call_cookie
|
8282 |
|
|
|= CALL_COOKIE_INT_REG (0, 1)
|
8283 |
|
|
| CALL_COOKIE_STACKSEQ (numregs - 1);
|
8284 |
|
|
else
|
8285 |
|
|
ca->call_cookie
|
8286 |
|
|
|= CALL_COOKIE_STACKSEQ (numregs);
|
8287 |
|
|
}
|
8288 |
|
|
}
|
8289 |
|
|
if (GET_SH_ARG_CLASS (mode2) == SH_ARG_FLOAT
|
8290 |
|
|
&& (named || ! ca->prototype_p))
|
8291 |
|
|
{
|
8292 |
|
|
if (mode2 == SFmode && ca->free_single_fp_reg)
|
8293 |
|
|
ca->free_single_fp_reg = 0;
|
8294 |
|
|
else if (ca->arg_count[(int) SH_ARG_FLOAT]
|
8295 |
|
|
< NPARM_REGS (SFmode))
|
8296 |
|
|
{
|
8297 |
|
|
int numfpregs
|
8298 |
|
|
= MIN ((GET_MODE_SIZE (mode2) + 7) / 8 * 2,
|
8299 |
|
|
NPARM_REGS (SFmode)
|
8300 |
|
|
- ca->arg_count[(int) SH_ARG_FLOAT]);
|
8301 |
|
|
|
8302 |
|
|
ca->arg_count[(int) SH_ARG_FLOAT] += numfpregs;
|
8303 |
|
|
|
8304 |
|
|
if (TARGET_SHCOMPACT && ! ca->prototype_p)
|
8305 |
|
|
{
|
8306 |
|
|
if (ca->outgoing && numregs > 0)
|
8307 |
|
|
do
|
8308 |
|
|
{
|
8309 |
|
|
ca->call_cookie
|
8310 |
|
|
|= (CALL_COOKIE_INT_REG
|
8311 |
|
|
(ca->arg_count[(int) SH_ARG_INT]
|
8312 |
|
|
- numregs + ((numfpregs - 2) / 2),
|
8313 |
|
|
4 + (ca->arg_count[(int) SH_ARG_FLOAT]
|
8314 |
|
|
- numfpregs) / 2));
|
8315 |
|
|
}
|
8316 |
|
|
while (numfpregs -= 2);
|
8317 |
|
|
}
|
8318 |
|
|
else if (mode2 == SFmode && (named)
|
8319 |
|
|
&& (ca->arg_count[(int) SH_ARG_FLOAT]
|
8320 |
|
|
< NPARM_REGS (SFmode)))
|
8321 |
|
|
ca->free_single_fp_reg
|
8322 |
|
|
= FIRST_FP_PARM_REG - numfpregs
|
8323 |
|
|
+ ca->arg_count[(int) SH_ARG_FLOAT] + 1;
|
8324 |
|
|
}
|
8325 |
|
|
}
|
8326 |
|
|
return;
|
8327 |
|
|
}
|
8328 |
|
|
|
8329 |
|
|
if ((TARGET_HITACHI || ca->renesas_abi) && TARGET_FPU_DOUBLE)
|
8330 |
|
|
{
|
8331 |
|
|
/* Note that we've used the skipped register. */
|
8332 |
|
|
if (mode == SFmode && ca->free_single_fp_reg)
|
8333 |
|
|
{
|
8334 |
|
|
ca->free_single_fp_reg = 0;
|
8335 |
|
|
return;
|
8336 |
|
|
}
|
8337 |
|
|
/* When we have a DF after an SF, there's an SF register that get
|
8338 |
|
|
skipped in order to align the DF value. We note this skipped
|
8339 |
|
|
register, because the next SF value will use it, and not the
|
8340 |
|
|
SF that follows the DF. */
|
8341 |
|
|
if (mode == DFmode
|
8342 |
|
|
&& ROUND_REG (*ca, DFmode) != ROUND_REG (*ca, SFmode))
|
8343 |
|
|
{
|
8344 |
|
|
ca->free_single_fp_reg = (ROUND_REG (*ca, SFmode)
|
8345 |
|
|
+ BASE_ARG_REG (mode));
|
8346 |
|
|
}
|
8347 |
|
|
}
|
8348 |
|
|
|
8349 |
|
|
if (! ((TARGET_SH4 || TARGET_SH2A) || ca->renesas_abi)
|
8350 |
|
|
|| PASS_IN_REG_P (*ca, mode, type))
|
8351 |
|
|
(ca->arg_count[(int) GET_SH_ARG_CLASS (mode)]
|
8352 |
|
|
= (ROUND_REG (*ca, mode)
|
8353 |
|
|
+ (mode == BLKmode
|
8354 |
|
|
? ROUND_ADVANCE (int_size_in_bytes (type))
|
8355 |
|
|
: ROUND_ADVANCE (GET_MODE_SIZE (mode)))));
|
8356 |
|
|
}
|
8357 |
|
|
|
8358 |
|
|
/* The Renesas calling convention doesn't quite fit into this scheme since
|
8359 |
|
|
the address is passed like an invisible argument, but one that is always
|
8360 |
|
|
passed in memory. */
|
8361 |
|
|
static rtx
|
8362 |
|
|
sh_struct_value_rtx (tree fndecl, int incoming ATTRIBUTE_UNUSED)
|
8363 |
|
|
{
|
8364 |
|
|
if (TARGET_HITACHI || sh_attr_renesas_p (fndecl))
|
8365 |
|
|
return 0;
|
8366 |
|
|
return gen_rtx_REG (Pmode, 2);
|
8367 |
|
|
}
|
8368 |
|
|
|
8369 |
|
|
/* Worker function for TARGET_FUNCTION_VALUE.
|
8370 |
|
|
|
8371 |
|
|
For the SH, this is like LIBCALL_VALUE, except that we must change the
|
8372 |
|
|
mode like PROMOTE_MODE does.
|
8373 |
|
|
??? PROMOTE_MODE is ignored for non-scalar types. The set of types
|
8374 |
|
|
tested here has to be kept in sync with the one in explow.c:promote_mode.
|
8375 |
|
|
*/
|
8376 |
|
|
|
8377 |
|
|
static rtx
|
8378 |
|
|
sh_function_value (const_tree valtype,
|
8379 |
|
|
const_tree fn_decl_or_type,
|
8380 |
|
|
bool outgoing ATTRIBUTE_UNUSED)
|
8381 |
|
|
{
|
8382 |
|
|
if (fn_decl_or_type
|
8383 |
|
|
&& !DECL_P (fn_decl_or_type))
|
8384 |
|
|
fn_decl_or_type = NULL;
|
8385 |
|
|
|
8386 |
|
|
return gen_rtx_REG (
|
8387 |
|
|
((GET_MODE_CLASS (TYPE_MODE (valtype)) == MODE_INT
|
8388 |
|
|
&& GET_MODE_SIZE (TYPE_MODE (valtype)) < 4
|
8389 |
|
|
&& (TREE_CODE (valtype) == INTEGER_TYPE
|
8390 |
|
|
|| TREE_CODE (valtype) == ENUMERAL_TYPE
|
8391 |
|
|
|| TREE_CODE (valtype) == BOOLEAN_TYPE
|
8392 |
|
|
|| TREE_CODE (valtype) == REAL_TYPE
|
8393 |
|
|
|| TREE_CODE (valtype) == OFFSET_TYPE))
|
8394 |
|
|
&& sh_promote_prototypes (fn_decl_or_type)
|
8395 |
|
|
? (TARGET_SHMEDIA64 ? DImode : SImode) : TYPE_MODE (valtype)),
|
8396 |
|
|
BASE_RETURN_VALUE_REG (TYPE_MODE (valtype)));
|
8397 |
|
|
}
|
8398 |
|
|
|
8399 |
|
|
/* Worker function for TARGET_LIBCALL_VALUE. */
|
8400 |
|
|
|
8401 |
|
|
static rtx
|
8402 |
|
|
sh_libcall_value (enum machine_mode mode, const_rtx fun ATTRIBUTE_UNUSED)
|
8403 |
|
|
{
|
8404 |
|
|
return gen_rtx_REG (mode, BASE_RETURN_VALUE_REG (mode));
|
8405 |
|
|
}
|
8406 |
|
|
|
8407 |
|
|
/* Worker function for FUNCTION_VALUE_REGNO_P. */
|
8408 |
|
|
|
8409 |
|
|
bool
|
8410 |
|
|
sh_function_value_regno_p (const unsigned int regno)
|
8411 |
|
|
{
|
8412 |
|
|
return ((regno) == FIRST_RET_REG
|
8413 |
|
|
|| (TARGET_SH2E && (regno) == FIRST_FP_RET_REG)
|
8414 |
|
|
|| (TARGET_SHMEDIA_FPU && (regno) == FIRST_FP_RET_REG));
|
8415 |
|
|
}
|
8416 |
|
|
|
8417 |
|
|
/* Worker function for TARGET_RETURN_IN_MEMORY. */
|
8418 |
|
|
|
8419 |
|
|
static bool
|
8420 |
|
|
sh_return_in_memory (const_tree type, const_tree fndecl)
|
8421 |
|
|
{
|
8422 |
|
|
if (TARGET_SH5)
|
8423 |
|
|
{
|
8424 |
|
|
if (TYPE_MODE (type) == BLKmode)
|
8425 |
|
|
return ((unsigned HOST_WIDE_INT) int_size_in_bytes (type)) > 8;
|
8426 |
|
|
else
|
8427 |
|
|
return GET_MODE_SIZE (TYPE_MODE (type)) > 8;
|
8428 |
|
|
}
|
8429 |
|
|
else
|
8430 |
|
|
{
|
8431 |
|
|
return (TYPE_MODE (type) == BLKmode
|
8432 |
|
|
|| ((TARGET_HITACHI || sh_attr_renesas_p (fndecl))
|
8433 |
|
|
&& TREE_CODE (type) == RECORD_TYPE));
|
8434 |
|
|
}
|
8435 |
|
|
}
|
8436 |
|
|
|
8437 |
|
|
/* We actually emit the code in sh_expand_prologue. We used to use
|
8438 |
|
|
a static variable to flag that we need to emit this code, but that
|
8439 |
|
|
doesn't when inlining, when functions are deferred and then emitted
|
8440 |
|
|
later. Fortunately, we already have two flags that are part of struct
|
8441 |
|
|
function that tell if a function uses varargs or stdarg. */
|
8442 |
|
|
static void
|
8443 |
|
|
sh_setup_incoming_varargs (CUMULATIVE_ARGS *ca,
|
8444 |
|
|
enum machine_mode mode,
|
8445 |
|
|
tree type,
|
8446 |
|
|
int *pretend_arg_size,
|
8447 |
|
|
int second_time ATTRIBUTE_UNUSED)
|
8448 |
|
|
{
|
8449 |
|
|
gcc_assert (cfun->stdarg);
|
8450 |
|
|
if (TARGET_VARARGS_PRETEND_ARGS (current_function_decl))
|
8451 |
|
|
{
|
8452 |
|
|
int named_parm_regs, anon_parm_regs;
|
8453 |
|
|
|
8454 |
|
|
named_parm_regs = (ROUND_REG (*ca, mode)
|
8455 |
|
|
+ (mode == BLKmode
|
8456 |
|
|
? ROUND_ADVANCE (int_size_in_bytes (type))
|
8457 |
|
|
: ROUND_ADVANCE (GET_MODE_SIZE (mode))));
|
8458 |
|
|
anon_parm_regs = NPARM_REGS (SImode) - named_parm_regs;
|
8459 |
|
|
if (anon_parm_regs > 0)
|
8460 |
|
|
*pretend_arg_size = anon_parm_regs * 4;
|
8461 |
|
|
}
|
8462 |
|
|
}
|
8463 |
|
|
|
8464 |
|
|
static bool
|
8465 |
|
|
sh_strict_argument_naming (CUMULATIVE_ARGS *ca ATTRIBUTE_UNUSED)
|
8466 |
|
|
{
|
8467 |
|
|
return TARGET_SH5;
|
8468 |
|
|
}
|
8469 |
|
|
|
8470 |
|
|
static bool
|
8471 |
|
|
sh_pretend_outgoing_varargs_named (CUMULATIVE_ARGS *ca)
|
8472 |
|
|
{
|
8473 |
|
|
return ! (TARGET_HITACHI || ca->renesas_abi) && ! TARGET_SH5;
|
8474 |
|
|
}
|
8475 |
|
|
|
8476 |
|
|
|
8477 |
|
|
/* Define the offset between two registers, one to be eliminated, and
|
8478 |
|
|
the other its replacement, at the start of a routine. */
|
8479 |
|
|
|
8480 |
|
|
int
|
8481 |
|
|
initial_elimination_offset (int from, int to)
|
8482 |
|
|
{
|
8483 |
|
|
int regs_saved;
|
8484 |
|
|
int regs_saved_rounding = 0;
|
8485 |
|
|
int total_saved_regs_space;
|
8486 |
|
|
int total_auto_space;
|
8487 |
|
|
int save_flags = target_flags;
|
8488 |
|
|
int copy_flags;
|
8489 |
|
|
HARD_REG_SET live_regs_mask;
|
8490 |
|
|
|
8491 |
|
|
shmedia_space_reserved_for_target_registers = false;
|
8492 |
|
|
regs_saved = calc_live_regs (&live_regs_mask);
|
8493 |
|
|
regs_saved += SHMEDIA_REGS_STACK_ADJUST ();
|
8494 |
|
|
|
8495 |
|
|
if (shmedia_reserve_space_for_target_registers_p (regs_saved, &live_regs_mask))
|
8496 |
|
|
{
|
8497 |
|
|
shmedia_space_reserved_for_target_registers = true;
|
8498 |
|
|
regs_saved += shmedia_target_regs_stack_adjust (&live_regs_mask);
|
8499 |
|
|
}
|
8500 |
|
|
|
8501 |
|
|
if (TARGET_SH5 && regs_saved % (STACK_BOUNDARY / BITS_PER_UNIT))
|
8502 |
|
|
regs_saved_rounding = ((STACK_BOUNDARY / BITS_PER_UNIT)
|
8503 |
|
|
- regs_saved % (STACK_BOUNDARY / BITS_PER_UNIT));
|
8504 |
|
|
|
8505 |
|
|
total_auto_space = rounded_frame_size (regs_saved) - regs_saved_rounding;
|
8506 |
|
|
copy_flags = target_flags;
|
8507 |
|
|
target_flags = save_flags;
|
8508 |
|
|
|
8509 |
|
|
total_saved_regs_space = regs_saved + regs_saved_rounding;
|
8510 |
|
|
|
8511 |
|
|
if (from == ARG_POINTER_REGNUM && to == HARD_FRAME_POINTER_REGNUM)
|
8512 |
|
|
return total_saved_regs_space + total_auto_space
|
8513 |
|
|
+ crtl->args.info.byref_regs * 8;
|
8514 |
|
|
|
8515 |
|
|
if (from == ARG_POINTER_REGNUM && to == STACK_POINTER_REGNUM)
|
8516 |
|
|
return total_saved_regs_space + total_auto_space
|
8517 |
|
|
+ crtl->args.info.byref_regs * 8;
|
8518 |
|
|
|
8519 |
|
|
/* Initial gap between fp and sp is 0. */
|
8520 |
|
|
if (from == HARD_FRAME_POINTER_REGNUM && to == STACK_POINTER_REGNUM)
|
8521 |
|
|
return 0;
|
8522 |
|
|
|
8523 |
|
|
if (from == FRAME_POINTER_REGNUM && to == STACK_POINTER_REGNUM)
|
8524 |
|
|
return rounded_frame_size (0);
|
8525 |
|
|
|
8526 |
|
|
if (from == FRAME_POINTER_REGNUM && to == HARD_FRAME_POINTER_REGNUM)
|
8527 |
|
|
return rounded_frame_size (0);
|
8528 |
|
|
|
8529 |
|
|
gcc_assert (from == RETURN_ADDRESS_POINTER_REGNUM
|
8530 |
|
|
&& (to == HARD_FRAME_POINTER_REGNUM
|
8531 |
|
|
|| to == STACK_POINTER_REGNUM));
|
8532 |
|
|
if (TARGET_SH5)
|
8533 |
|
|
{
|
8534 |
|
|
int n = total_saved_regs_space;
|
8535 |
|
|
int pr_reg = TARGET_SHMEDIA ? PR_MEDIA_REG : PR_REG;
|
8536 |
|
|
save_schedule schedule;
|
8537 |
|
|
save_entry *entry;
|
8538 |
|
|
|
8539 |
|
|
n += total_auto_space;
|
8540 |
|
|
|
8541 |
|
|
/* If it wasn't saved, there's not much we can do. */
|
8542 |
|
|
if (! TEST_HARD_REG_BIT (live_regs_mask, pr_reg))
|
8543 |
|
|
return n;
|
8544 |
|
|
|
8545 |
|
|
target_flags = copy_flags;
|
8546 |
|
|
|
8547 |
|
|
sh5_schedule_saves (&live_regs_mask, &schedule, n);
|
8548 |
|
|
for (entry = &schedule.entries[1]; entry->mode != VOIDmode; entry++)
|
8549 |
|
|
if (entry->reg == pr_reg)
|
8550 |
|
|
{
|
8551 |
|
|
target_flags = save_flags;
|
8552 |
|
|
return entry->offset;
|
8553 |
|
|
}
|
8554 |
|
|
gcc_unreachable ();
|
8555 |
|
|
}
|
8556 |
|
|
else
|
8557 |
|
|
return total_auto_space;
|
8558 |
|
|
}
|
8559 |
|
|
|
8560 |
|
|
/* Parse the -mfixed-range= option string. */
|
8561 |
|
|
void
|
8562 |
|
|
sh_fix_range (const char *const_str)
|
8563 |
|
|
{
|
8564 |
|
|
int i, first, last;
|
8565 |
|
|
char *str, *dash, *comma;
|
8566 |
|
|
|
8567 |
|
|
/* str must be of the form REG1'-'REG2{,REG1'-'REG} where REG1 and
|
8568 |
|
|
REG2 are either register names or register numbers. The effect
|
8569 |
|
|
of this option is to mark the registers in the range from REG1 to
|
8570 |
|
|
REG2 as ``fixed'' so they won't be used by the compiler. */
|
8571 |
|
|
|
8572 |
|
|
i = strlen (const_str);
|
8573 |
|
|
str = (char *) alloca (i + 1);
|
8574 |
|
|
memcpy (str, const_str, i + 1);
|
8575 |
|
|
|
8576 |
|
|
while (1)
|
8577 |
|
|
{
|
8578 |
|
|
dash = strchr (str, '-');
|
8579 |
|
|
if (!dash)
|
8580 |
|
|
{
|
8581 |
|
|
warning (0, "value of -mfixed-range must have form REG1-REG2");
|
8582 |
|
|
return;
|
8583 |
|
|
}
|
8584 |
|
|
*dash = '\0';
|
8585 |
|
|
comma = strchr (dash + 1, ',');
|
8586 |
|
|
if (comma)
|
8587 |
|
|
*comma = '\0';
|
8588 |
|
|
|
8589 |
|
|
first = decode_reg_name (str);
|
8590 |
|
|
if (first < 0)
|
8591 |
|
|
{
|
8592 |
|
|
warning (0, "unknown register name: %s", str);
|
8593 |
|
|
return;
|
8594 |
|
|
}
|
8595 |
|
|
|
8596 |
|
|
last = decode_reg_name (dash + 1);
|
8597 |
|
|
if (last < 0)
|
8598 |
|
|
{
|
8599 |
|
|
warning (0, "unknown register name: %s", dash + 1);
|
8600 |
|
|
return;
|
8601 |
|
|
}
|
8602 |
|
|
|
8603 |
|
|
*dash = '-';
|
8604 |
|
|
|
8605 |
|
|
if (first > last)
|
8606 |
|
|
{
|
8607 |
|
|
warning (0, "%s-%s is an empty range", str, dash + 1);
|
8608 |
|
|
return;
|
8609 |
|
|
}
|
8610 |
|
|
|
8611 |
|
|
for (i = first; i <= last; ++i)
|
8612 |
|
|
fixed_regs[i] = call_used_regs[i] = 1;
|
8613 |
|
|
|
8614 |
|
|
if (!comma)
|
8615 |
|
|
break;
|
8616 |
|
|
|
8617 |
|
|
*comma = ',';
|
8618 |
|
|
str = comma + 1;
|
8619 |
|
|
}
|
8620 |
|
|
}
|
8621 |
|
|
|
8622 |
|
|
/* Insert any deferred function attributes from earlier pragmas. */
|
8623 |
|
|
static void
|
8624 |
|
|
sh_insert_attributes (tree node, tree *attributes)
|
8625 |
|
|
{
|
8626 |
|
|
tree attrs;
|
8627 |
|
|
|
8628 |
|
|
if (TREE_CODE (node) != FUNCTION_DECL)
|
8629 |
|
|
return;
|
8630 |
|
|
|
8631 |
|
|
/* We are only interested in fields. */
|
8632 |
|
|
if (!DECL_P (node))
|
8633 |
|
|
return;
|
8634 |
|
|
|
8635 |
|
|
/* Append the attributes to the deferred attributes. */
|
8636 |
|
|
*sh_deferred_function_attributes_tail = *attributes;
|
8637 |
|
|
attrs = sh_deferred_function_attributes;
|
8638 |
|
|
if (!attrs)
|
8639 |
|
|
return;
|
8640 |
|
|
|
8641 |
|
|
/* Some attributes imply or require the interrupt attribute. */
|
8642 |
|
|
if (!lookup_attribute ("interrupt_handler", attrs)
|
8643 |
|
|
&& !lookup_attribute ("interrupt_handler", DECL_ATTRIBUTES (node)))
|
8644 |
|
|
{
|
8645 |
|
|
/* If we have a trapa_handler, but no interrupt_handler attribute,
|
8646 |
|
|
insert an interrupt_handler attribute. */
|
8647 |
|
|
if (lookup_attribute ("trapa_handler", attrs) != NULL_TREE)
|
8648 |
|
|
/* We can't use sh_pr_interrupt here because that's not in the
|
8649 |
|
|
java frontend. */
|
8650 |
|
|
attrs
|
8651 |
|
|
= tree_cons (get_identifier("interrupt_handler"), NULL_TREE, attrs);
|
8652 |
|
|
/* However, for sp_switch, trap_exit, nosave_low_regs and resbank,
|
8653 |
|
|
if the interrupt attribute is missing, we ignore the attribute
|
8654 |
|
|
and warn. */
|
8655 |
|
|
else if (lookup_attribute ("sp_switch", attrs)
|
8656 |
|
|
|| lookup_attribute ("trap_exit", attrs)
|
8657 |
|
|
|| lookup_attribute ("nosave_low_regs", attrs)
|
8658 |
|
|
|| lookup_attribute ("resbank", attrs))
|
8659 |
|
|
{
|
8660 |
|
|
tree *tail;
|
8661 |
|
|
|
8662 |
|
|
for (tail = attributes; attrs; attrs = TREE_CHAIN (attrs))
|
8663 |
|
|
{
|
8664 |
|
|
if (is_attribute_p ("sp_switch", TREE_PURPOSE (attrs))
|
8665 |
|
|
|| is_attribute_p ("trap_exit", TREE_PURPOSE (attrs))
|
8666 |
|
|
|| is_attribute_p ("nosave_low_regs", TREE_PURPOSE (attrs))
|
8667 |
|
|
|| is_attribute_p ("resbank", TREE_PURPOSE (attrs)))
|
8668 |
|
|
warning (OPT_Wattributes,
|
8669 |
|
|
"%qE attribute only applies to interrupt functions",
|
8670 |
|
|
TREE_PURPOSE (attrs));
|
8671 |
|
|
else
|
8672 |
|
|
{
|
8673 |
|
|
*tail = tree_cons (TREE_PURPOSE (attrs), NULL_TREE,
|
8674 |
|
|
NULL_TREE);
|
8675 |
|
|
tail = &TREE_CHAIN (*tail);
|
8676 |
|
|
}
|
8677 |
|
|
}
|
8678 |
|
|
attrs = *attributes;
|
8679 |
|
|
}
|
8680 |
|
|
}
|
8681 |
|
|
|
8682 |
|
|
/* Install the processed list. */
|
8683 |
|
|
*attributes = attrs;
|
8684 |
|
|
|
8685 |
|
|
/* Clear deferred attributes. */
|
8686 |
|
|
sh_deferred_function_attributes = NULL_TREE;
|
8687 |
|
|
sh_deferred_function_attributes_tail = &sh_deferred_function_attributes;
|
8688 |
|
|
|
8689 |
|
|
return;
|
8690 |
|
|
}
|
8691 |
|
|
|
8692 |
|
|
/* Supported attributes:
|
8693 |
|
|
|
8694 |
|
|
interrupt_handler -- specifies this function is an interrupt handler.
|
8695 |
|
|
|
8696 |
|
|
trapa_handler - like above, but don't save all registers.
|
8697 |
|
|
|
8698 |
|
|
sp_switch -- specifies an alternate stack for an interrupt handler
|
8699 |
|
|
to run on.
|
8700 |
|
|
|
8701 |
|
|
trap_exit -- use a trapa to exit an interrupt function instead of
|
8702 |
|
|
an rte instruction.
|
8703 |
|
|
|
8704 |
|
|
nosave_low_regs - don't save r0..r7 in an interrupt handler.
|
8705 |
|
|
This is useful on the SH3 and upwards,
|
8706 |
|
|
which has a separate set of low regs for User and Supervisor modes.
|
8707 |
|
|
This should only be used for the lowest level of interrupts. Higher levels
|
8708 |
|
|
of interrupts must save the registers in case they themselves are
|
8709 |
|
|
interrupted.
|
8710 |
|
|
|
8711 |
|
|
renesas -- use Renesas calling/layout conventions (functions and
|
8712 |
|
|
structures).
|
8713 |
|
|
|
8714 |
|
|
resbank -- In case of an ISR, use a register bank to save registers
|
8715 |
|
|
R0-R14, MACH, MACL, GBR and PR. This is useful only on SH2A targets.
|
8716 |
|
|
*/
|
8717 |
|
|
|
8718 |
|
|
/* Handle a 'resbank' attribute. */
|
8719 |
|
|
static tree
|
8720 |
|
|
sh_handle_resbank_handler_attribute (tree * node, tree name,
|
8721 |
|
|
tree args ATTRIBUTE_UNUSED,
|
8722 |
|
|
int flags ATTRIBUTE_UNUSED,
|
8723 |
|
|
bool * no_add_attrs)
|
8724 |
|
|
{
|
8725 |
|
|
if (!TARGET_SH2A)
|
8726 |
|
|
{
|
8727 |
|
|
warning (OPT_Wattributes, "%qE attribute is supported only for SH2A",
|
8728 |
|
|
name);
|
8729 |
|
|
*no_add_attrs = true;
|
8730 |
|
|
}
|
8731 |
|
|
if (TREE_CODE (*node) != FUNCTION_DECL)
|
8732 |
|
|
{
|
8733 |
|
|
warning (OPT_Wattributes, "%qE attribute only applies to functions",
|
8734 |
|
|
name);
|
8735 |
|
|
*no_add_attrs = true;
|
8736 |
|
|
}
|
8737 |
|
|
|
8738 |
|
|
return NULL_TREE;
|
8739 |
|
|
}
|
8740 |
|
|
|
8741 |
|
|
/* Handle an "interrupt_handler" attribute; arguments as in
|
8742 |
|
|
struct attribute_spec.handler. */
|
8743 |
|
|
static tree
|
8744 |
|
|
sh_handle_interrupt_handler_attribute (tree *node, tree name,
|
8745 |
|
|
tree args ATTRIBUTE_UNUSED,
|
8746 |
|
|
int flags ATTRIBUTE_UNUSED,
|
8747 |
|
|
bool *no_add_attrs)
|
8748 |
|
|
{
|
8749 |
|
|
if (TREE_CODE (*node) != FUNCTION_DECL)
|
8750 |
|
|
{
|
8751 |
|
|
warning (OPT_Wattributes, "%qE attribute only applies to functions",
|
8752 |
|
|
name);
|
8753 |
|
|
*no_add_attrs = true;
|
8754 |
|
|
}
|
8755 |
|
|
else if (TARGET_SHCOMPACT)
|
8756 |
|
|
{
|
8757 |
|
|
error ("attribute interrupt_handler is not compatible with -m5-compact");
|
8758 |
|
|
*no_add_attrs = true;
|
8759 |
|
|
}
|
8760 |
|
|
|
8761 |
|
|
return NULL_TREE;
|
8762 |
|
|
}
|
8763 |
|
|
|
8764 |
|
|
/* Handle an 'function_vector' attribute; arguments as in
|
8765 |
|
|
struct attribute_spec.handler. */
|
8766 |
|
|
static tree
|
8767 |
|
|
sh2a_handle_function_vector_handler_attribute (tree * node, tree name,
|
8768 |
|
|
tree args ATTRIBUTE_UNUSED,
|
8769 |
|
|
int flags ATTRIBUTE_UNUSED,
|
8770 |
|
|
bool * no_add_attrs)
|
8771 |
|
|
{
|
8772 |
|
|
if (!TARGET_SH2A)
|
8773 |
|
|
{
|
8774 |
|
|
warning (OPT_Wattributes, "%qE attribute only applies to SH2A",
|
8775 |
|
|
name);
|
8776 |
|
|
*no_add_attrs = true;
|
8777 |
|
|
}
|
8778 |
|
|
else if (TREE_CODE (*node) != FUNCTION_DECL)
|
8779 |
|
|
{
|
8780 |
|
|
warning (OPT_Wattributes, "%qE attribute only applies to functions",
|
8781 |
|
|
name);
|
8782 |
|
|
*no_add_attrs = true;
|
8783 |
|
|
}
|
8784 |
|
|
else if (TREE_CODE (TREE_VALUE (args)) != INTEGER_CST)
|
8785 |
|
|
{
|
8786 |
|
|
/* The argument must be a constant integer. */
|
8787 |
|
|
warning (OPT_Wattributes,
|
8788 |
|
|
"%qE attribute argument not an integer constant",
|
8789 |
|
|
name);
|
8790 |
|
|
*no_add_attrs = true;
|
8791 |
|
|
}
|
8792 |
|
|
else if (TREE_INT_CST_LOW (TREE_VALUE (args)) > 255)
|
8793 |
|
|
{
|
8794 |
|
|
/* The argument value must be between 0 to 255. */
|
8795 |
|
|
warning (OPT_Wattributes,
|
8796 |
|
|
"%qE attribute argument should be between 0 to 255",
|
8797 |
|
|
name);
|
8798 |
|
|
*no_add_attrs = true;
|
8799 |
|
|
}
|
8800 |
|
|
return NULL_TREE;
|
8801 |
|
|
}
|
8802 |
|
|
|
8803 |
|
|
/* Returns 1 if current function has been assigned the attribute
|
8804 |
|
|
'function_vector'. */
|
8805 |
|
|
int
|
8806 |
|
|
sh2a_is_function_vector_call (rtx x)
|
8807 |
|
|
{
|
8808 |
|
|
if (GET_CODE (x) == SYMBOL_REF
|
8809 |
|
|
&& (SYMBOL_REF_FLAGS (x) & SYMBOL_FLAG_FUNCVEC_FUNCTION))
|
8810 |
|
|
{
|
8811 |
|
|
tree tr = SYMBOL_REF_DECL (x);
|
8812 |
|
|
|
8813 |
|
|
if (sh2a_function_vector_p (tr))
|
8814 |
|
|
return 1;
|
8815 |
|
|
}
|
8816 |
|
|
|
8817 |
|
|
return 0;
|
8818 |
|
|
}
|
8819 |
|
|
|
8820 |
|
|
/* Returns the function vector number, if the the attribute
|
8821 |
|
|
'function_vector' is assigned, otherwise returns zero. */
|
8822 |
|
|
int
|
8823 |
|
|
sh2a_get_function_vector_number (rtx x)
|
8824 |
|
|
{
|
8825 |
|
|
int num;
|
8826 |
|
|
tree list, t;
|
8827 |
|
|
|
8828 |
|
|
if ((GET_CODE (x) == SYMBOL_REF)
|
8829 |
|
|
&& (SYMBOL_REF_FLAGS (x) & SYMBOL_FLAG_FUNCVEC_FUNCTION))
|
8830 |
|
|
{
|
8831 |
|
|
t = SYMBOL_REF_DECL (x);
|
8832 |
|
|
|
8833 |
|
|
if (TREE_CODE (t) != FUNCTION_DECL)
|
8834 |
|
|
return 0;
|
8835 |
|
|
|
8836 |
|
|
list = SH_ATTRIBUTES (t);
|
8837 |
|
|
while (list)
|
8838 |
|
|
{
|
8839 |
|
|
if (is_attribute_p ("function_vector", TREE_PURPOSE (list)))
|
8840 |
|
|
{
|
8841 |
|
|
num = TREE_INT_CST_LOW (TREE_VALUE (TREE_VALUE (list)));
|
8842 |
|
|
return num;
|
8843 |
|
|
}
|
8844 |
|
|
|
8845 |
|
|
list = TREE_CHAIN (list);
|
8846 |
|
|
}
|
8847 |
|
|
|
8848 |
|
|
return 0;
|
8849 |
|
|
}
|
8850 |
|
|
else
|
8851 |
|
|
return 0;
|
8852 |
|
|
}
|
8853 |
|
|
|
8854 |
|
|
/* Handle an "sp_switch" attribute; arguments as in
|
8855 |
|
|
struct attribute_spec.handler. */
|
8856 |
|
|
static tree
|
8857 |
|
|
sh_handle_sp_switch_attribute (tree *node, tree name, tree args,
|
8858 |
|
|
int flags ATTRIBUTE_UNUSED, bool *no_add_attrs)
|
8859 |
|
|
{
|
8860 |
|
|
if (TREE_CODE (*node) != FUNCTION_DECL)
|
8861 |
|
|
{
|
8862 |
|
|
warning (OPT_Wattributes, "%qE attribute only applies to functions",
|
8863 |
|
|
name);
|
8864 |
|
|
*no_add_attrs = true;
|
8865 |
|
|
}
|
8866 |
|
|
else if (TREE_CODE (TREE_VALUE (args)) != STRING_CST)
|
8867 |
|
|
{
|
8868 |
|
|
/* The argument must be a constant string. */
|
8869 |
|
|
warning (OPT_Wattributes, "%qE attribute argument not a string constant",
|
8870 |
|
|
name);
|
8871 |
|
|
*no_add_attrs = true;
|
8872 |
|
|
}
|
8873 |
|
|
|
8874 |
|
|
return NULL_TREE;
|
8875 |
|
|
}
|
8876 |
|
|
|
8877 |
|
|
/* Handle an "trap_exit" attribute; arguments as in
|
8878 |
|
|
struct attribute_spec.handler. */
|
8879 |
|
|
static tree
|
8880 |
|
|
sh_handle_trap_exit_attribute (tree *node, tree name, tree args,
|
8881 |
|
|
int flags ATTRIBUTE_UNUSED, bool *no_add_attrs)
|
8882 |
|
|
{
|
8883 |
|
|
if (TREE_CODE (*node) != FUNCTION_DECL)
|
8884 |
|
|
{
|
8885 |
|
|
warning (OPT_Wattributes, "%qE attribute only applies to functions",
|
8886 |
|
|
name);
|
8887 |
|
|
*no_add_attrs = true;
|
8888 |
|
|
}
|
8889 |
|
|
/* The argument specifies a trap number to be used in a trapa instruction
|
8890 |
|
|
at function exit (instead of an rte instruction). */
|
8891 |
|
|
else if (TREE_CODE (TREE_VALUE (args)) != INTEGER_CST)
|
8892 |
|
|
{
|
8893 |
|
|
/* The argument must be a constant integer. */
|
8894 |
|
|
warning (OPT_Wattributes, "%qE attribute argument not an "
|
8895 |
|
|
"integer constant", name);
|
8896 |
|
|
*no_add_attrs = true;
|
8897 |
|
|
}
|
8898 |
|
|
|
8899 |
|
|
return NULL_TREE;
|
8900 |
|
|
}
|
8901 |
|
|
|
8902 |
|
|
static tree
|
8903 |
|
|
sh_handle_renesas_attribute (tree *node ATTRIBUTE_UNUSED,
|
8904 |
|
|
tree name ATTRIBUTE_UNUSED,
|
8905 |
|
|
tree args ATTRIBUTE_UNUSED,
|
8906 |
|
|
int flags ATTRIBUTE_UNUSED,
|
8907 |
|
|
bool *no_add_attrs ATTRIBUTE_UNUSED)
|
8908 |
|
|
{
|
8909 |
|
|
return NULL_TREE;
|
8910 |
|
|
}
|
8911 |
|
|
|
8912 |
|
|
/* True if __attribute__((renesas)) or -mrenesas. */
|
8913 |
|
|
int
|
8914 |
|
|
sh_attr_renesas_p (const_tree td)
|
8915 |
|
|
{
|
8916 |
|
|
if (TARGET_HITACHI)
|
8917 |
|
|
return 1;
|
8918 |
|
|
if (td == 0)
|
8919 |
|
|
return 0;
|
8920 |
|
|
if (DECL_P (td))
|
8921 |
|
|
td = TREE_TYPE (td);
|
8922 |
|
|
if (td == error_mark_node)
|
8923 |
|
|
return 0;
|
8924 |
|
|
return (lookup_attribute ("renesas", TYPE_ATTRIBUTES (td))
|
8925 |
|
|
!= NULL_TREE);
|
8926 |
|
|
}
|
8927 |
|
|
|
8928 |
|
|
/* True if __attribute__((renesas)) or -mrenesas, for the current
|
8929 |
|
|
function. */
|
8930 |
|
|
int
|
8931 |
|
|
sh_cfun_attr_renesas_p (void)
|
8932 |
|
|
{
|
8933 |
|
|
return sh_attr_renesas_p (current_function_decl);
|
8934 |
|
|
}
|
8935 |
|
|
|
8936 |
|
|
int
|
8937 |
|
|
sh_cfun_interrupt_handler_p (void)
|
8938 |
|
|
{
|
8939 |
|
|
return (lookup_attribute ("interrupt_handler",
|
8940 |
|
|
DECL_ATTRIBUTES (current_function_decl))
|
8941 |
|
|
!= NULL_TREE);
|
8942 |
|
|
}
|
8943 |
|
|
|
8944 |
|
|
/* Returns 1 if FUNC has been assigned the attribute
|
8945 |
|
|
"function_vector". */
|
8946 |
|
|
int
|
8947 |
|
|
sh2a_function_vector_p (tree func)
|
8948 |
|
|
{
|
8949 |
|
|
tree list;
|
8950 |
|
|
if (TREE_CODE (func) != FUNCTION_DECL)
|
8951 |
|
|
return 0;
|
8952 |
|
|
|
8953 |
|
|
list = SH_ATTRIBUTES (func);
|
8954 |
|
|
while (list)
|
8955 |
|
|
{
|
8956 |
|
|
if (is_attribute_p ("function_vector", TREE_PURPOSE (list)))
|
8957 |
|
|
return 1;
|
8958 |
|
|
|
8959 |
|
|
list = TREE_CHAIN (list);
|
8960 |
|
|
}
|
8961 |
|
|
return 0;
|
8962 |
|
|
}
|
8963 |
|
|
|
8964 |
|
|
/* Returns TRUE if given tree has the "resbank" attribute. */
|
8965 |
|
|
|
8966 |
|
|
int
|
8967 |
|
|
sh_cfun_resbank_handler_p (void)
|
8968 |
|
|
{
|
8969 |
|
|
return ((lookup_attribute ("resbank",
|
8970 |
|
|
DECL_ATTRIBUTES (current_function_decl))
|
8971 |
|
|
!= NULL_TREE)
|
8972 |
|
|
&& (lookup_attribute ("interrupt_handler",
|
8973 |
|
|
DECL_ATTRIBUTES (current_function_decl))
|
8974 |
|
|
!= NULL_TREE) && TARGET_SH2A);
|
8975 |
|
|
}
|
8976 |
|
|
|
8977 |
|
|
/* Implement TARGET_CHECK_PCH_TARGET_FLAGS. */
|
8978 |
|
|
|
8979 |
|
|
static const char *
|
8980 |
|
|
sh_check_pch_target_flags (int old_flags)
|
8981 |
|
|
{
|
8982 |
|
|
if ((old_flags ^ target_flags) & (MASK_SH1 | MASK_SH2 | MASK_SH3
|
8983 |
|
|
| MASK_SH_E | MASK_HARD_SH4
|
8984 |
|
|
| MASK_FPU_SINGLE | MASK_SH4))
|
8985 |
|
|
return _("created and used with different architectures / ABIs");
|
8986 |
|
|
if ((old_flags ^ target_flags) & MASK_HITACHI)
|
8987 |
|
|
return _("created and used with different ABIs");
|
8988 |
|
|
if ((old_flags ^ target_flags) & MASK_LITTLE_ENDIAN)
|
8989 |
|
|
return _("created and used with different endianness");
|
8990 |
|
|
return NULL;
|
8991 |
|
|
}
|
8992 |
|
|
|
8993 |
|
|
/* Predicates used by the templates. */
|
8994 |
|
|
|
8995 |
|
|
/* Returns 1 if OP is MACL, MACH or PR. The input must be a REG rtx.
|
8996 |
|
|
Used only in general_movsrc_operand. */
|
8997 |
|
|
|
8998 |
|
|
int
|
8999 |
|
|
system_reg_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
|
9000 |
|
|
{
|
9001 |
|
|
switch (REGNO (op))
|
9002 |
|
|
{
|
9003 |
|
|
case PR_REG:
|
9004 |
|
|
case MACL_REG:
|
9005 |
|
|
case MACH_REG:
|
9006 |
|
|
return 1;
|
9007 |
|
|
}
|
9008 |
|
|
return 0;
|
9009 |
|
|
}
|
9010 |
|
|
|
9011 |
|
|
/* Nonzero if OP is a floating point value with value 0.0. */
|
9012 |
|
|
|
9013 |
|
|
int
|
9014 |
|
|
fp_zero_operand (rtx op)
|
9015 |
|
|
{
|
9016 |
|
|
REAL_VALUE_TYPE r;
|
9017 |
|
|
|
9018 |
|
|
if (GET_MODE (op) != SFmode)
|
9019 |
|
|
return 0;
|
9020 |
|
|
|
9021 |
|
|
REAL_VALUE_FROM_CONST_DOUBLE (r, op);
|
9022 |
|
|
return REAL_VALUES_EQUAL (r, dconst0) && ! REAL_VALUE_MINUS_ZERO (r);
|
9023 |
|
|
}
|
9024 |
|
|
|
9025 |
|
|
/* Nonzero if OP is a floating point value with value 1.0. */
|
9026 |
|
|
|
9027 |
|
|
int
|
9028 |
|
|
fp_one_operand (rtx op)
|
9029 |
|
|
{
|
9030 |
|
|
REAL_VALUE_TYPE r;
|
9031 |
|
|
|
9032 |
|
|
if (GET_MODE (op) != SFmode)
|
9033 |
|
|
return 0;
|
9034 |
|
|
|
9035 |
|
|
REAL_VALUE_FROM_CONST_DOUBLE (r, op);
|
9036 |
|
|
return REAL_VALUES_EQUAL (r, dconst1);
|
9037 |
|
|
}
|
9038 |
|
|
|
9039 |
|
|
/* In general mode switching is used. If we are
|
9040 |
|
|
compiling without -mfmovd, movsf_ie isn't taken into account for
|
9041 |
|
|
mode switching. We could check in machine_dependent_reorg for
|
9042 |
|
|
cases where we know we are in single precision mode, but there is
|
9043 |
|
|
interface to find that out during reload, so we must avoid
|
9044 |
|
|
choosing an fldi alternative during reload and thus failing to
|
9045 |
|
|
allocate a scratch register for the constant loading. */
|
9046 |
|
|
int
|
9047 |
|
|
fldi_ok (void)
|
9048 |
|
|
{
|
9049 |
|
|
return 1;
|
9050 |
|
|
}
|
9051 |
|
|
|
9052 |
|
|
int
|
9053 |
|
|
tertiary_reload_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
|
9054 |
|
|
{
|
9055 |
|
|
enum rtx_code code = GET_CODE (op);
|
9056 |
|
|
return code == MEM || (TARGET_SH4 && code == CONST_DOUBLE);
|
9057 |
|
|
}
|
9058 |
|
|
|
9059 |
|
|
/* Return the TLS type for TLS symbols, 0 for otherwise. */
|
9060 |
|
|
enum tls_model
|
9061 |
|
|
tls_symbolic_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
|
9062 |
|
|
{
|
9063 |
|
|
if (GET_CODE (op) != SYMBOL_REF)
|
9064 |
|
|
return TLS_MODEL_NONE;
|
9065 |
|
|
return SYMBOL_REF_TLS_MODEL (op);
|
9066 |
|
|
}
|
9067 |
|
|
|
9068 |
|
|
/* Return the destination address of a branch. */
|
9069 |
|
|
|
9070 |
|
|
static int
|
9071 |
|
|
branch_dest (rtx branch)
|
9072 |
|
|
{
|
9073 |
|
|
rtx dest = SET_SRC (PATTERN (branch));
|
9074 |
|
|
int dest_uid;
|
9075 |
|
|
|
9076 |
|
|
if (GET_CODE (dest) == IF_THEN_ELSE)
|
9077 |
|
|
dest = XEXP (dest, 1);
|
9078 |
|
|
dest = XEXP (dest, 0);
|
9079 |
|
|
dest_uid = INSN_UID (dest);
|
9080 |
|
|
return INSN_ADDRESSES (dest_uid);
|
9081 |
|
|
}
|
9082 |
|
|
|
9083 |
|
|
/* Return nonzero if REG is not used after INSN.
|
9084 |
|
|
We assume REG is a reload reg, and therefore does
|
9085 |
|
|
not live past labels. It may live past calls or jumps though. */
|
9086 |
|
|
int
|
9087 |
|
|
reg_unused_after (rtx reg, rtx insn)
|
9088 |
|
|
{
|
9089 |
|
|
enum rtx_code code;
|
9090 |
|
|
rtx set;
|
9091 |
|
|
|
9092 |
|
|
/* If the reg is set by this instruction, then it is safe for our
|
9093 |
|
|
case. Disregard the case where this is a store to memory, since
|
9094 |
|
|
we are checking a register used in the store address. */
|
9095 |
|
|
set = single_set (insn);
|
9096 |
|
|
if (set && !MEM_P (SET_DEST (set))
|
9097 |
|
|
&& reg_overlap_mentioned_p (reg, SET_DEST (set)))
|
9098 |
|
|
return 1;
|
9099 |
|
|
|
9100 |
|
|
while ((insn = NEXT_INSN (insn)))
|
9101 |
|
|
{
|
9102 |
|
|
rtx set;
|
9103 |
|
|
if (!INSN_P (insn))
|
9104 |
|
|
continue;
|
9105 |
|
|
|
9106 |
|
|
code = GET_CODE (insn);
|
9107 |
|
|
|
9108 |
|
|
#if 0
|
9109 |
|
|
/* If this is a label that existed before reload, then the register
|
9110 |
|
|
if dead here. However, if this is a label added by reorg, then
|
9111 |
|
|
the register may still be live here. We can't tell the difference,
|
9112 |
|
|
so we just ignore labels completely. */
|
9113 |
|
|
if (code == CODE_LABEL)
|
9114 |
|
|
return 1;
|
9115 |
|
|
/* else */
|
9116 |
|
|
#endif
|
9117 |
|
|
|
9118 |
|
|
if (code == JUMP_INSN)
|
9119 |
|
|
return 0;
|
9120 |
|
|
|
9121 |
|
|
/* If this is a sequence, we must handle them all at once.
|
9122 |
|
|
We could have for instance a call that sets the target register,
|
9123 |
|
|
and an insn in a delay slot that uses the register. In this case,
|
9124 |
|
|
we must return 0. */
|
9125 |
|
|
else if (code == INSN && GET_CODE (PATTERN (insn)) == SEQUENCE)
|
9126 |
|
|
{
|
9127 |
|
|
int i;
|
9128 |
|
|
int retval = 0;
|
9129 |
|
|
|
9130 |
|
|
for (i = 0; i < XVECLEN (PATTERN (insn), 0); i++)
|
9131 |
|
|
{
|
9132 |
|
|
rtx this_insn = XVECEXP (PATTERN (insn), 0, i);
|
9133 |
|
|
rtx set = single_set (this_insn);
|
9134 |
|
|
|
9135 |
|
|
if (CALL_P (this_insn))
|
9136 |
|
|
code = CALL_INSN;
|
9137 |
|
|
else if (JUMP_P (this_insn))
|
9138 |
|
|
{
|
9139 |
|
|
if (INSN_ANNULLED_BRANCH_P (this_insn))
|
9140 |
|
|
return 0;
|
9141 |
|
|
code = JUMP_INSN;
|
9142 |
|
|
}
|
9143 |
|
|
|
9144 |
|
|
if (set && reg_overlap_mentioned_p (reg, SET_SRC (set)))
|
9145 |
|
|
return 0;
|
9146 |
|
|
if (set && reg_overlap_mentioned_p (reg, SET_DEST (set)))
|
9147 |
|
|
{
|
9148 |
|
|
if (!MEM_P (SET_DEST (set)))
|
9149 |
|
|
retval = 1;
|
9150 |
|
|
else
|
9151 |
|
|
return 0;
|
9152 |
|
|
}
|
9153 |
|
|
if (set == 0
|
9154 |
|
|
&& reg_overlap_mentioned_p (reg, PATTERN (this_insn)))
|
9155 |
|
|
return 0;
|
9156 |
|
|
}
|
9157 |
|
|
if (retval == 1)
|
9158 |
|
|
return 1;
|
9159 |
|
|
else if (code == JUMP_INSN)
|
9160 |
|
|
return 0;
|
9161 |
|
|
}
|
9162 |
|
|
|
9163 |
|
|
set = single_set (insn);
|
9164 |
|
|
if (set && reg_overlap_mentioned_p (reg, SET_SRC (set)))
|
9165 |
|
|
return 0;
|
9166 |
|
|
if (set && reg_overlap_mentioned_p (reg, SET_DEST (set)))
|
9167 |
|
|
return !MEM_P (SET_DEST (set));
|
9168 |
|
|
if (set == 0 && reg_overlap_mentioned_p (reg, PATTERN (insn)))
|
9169 |
|
|
return 0;
|
9170 |
|
|
|
9171 |
|
|
if (code == CALL_INSN && call_really_used_regs[REGNO (reg)])
|
9172 |
|
|
return 1;
|
9173 |
|
|
}
|
9174 |
|
|
return 1;
|
9175 |
|
|
}
|
9176 |
|
|
|
9177 |
|
|
#include "ggc.h"
|
9178 |
|
|
|
9179 |
|
|
static GTY(()) rtx fpscr_rtx;
|
9180 |
|
|
rtx
|
9181 |
|
|
get_fpscr_rtx (void)
|
9182 |
|
|
{
|
9183 |
|
|
if (! fpscr_rtx)
|
9184 |
|
|
{
|
9185 |
|
|
fpscr_rtx = gen_rtx_REG (PSImode, FPSCR_REG);
|
9186 |
|
|
REG_USERVAR_P (fpscr_rtx) = 1;
|
9187 |
|
|
mark_user_reg (fpscr_rtx);
|
9188 |
|
|
}
|
9189 |
|
|
if (! reload_completed || mdep_reorg_phase != SH_AFTER_MDEP_REORG)
|
9190 |
|
|
mark_user_reg (fpscr_rtx);
|
9191 |
|
|
return fpscr_rtx;
|
9192 |
|
|
}
|
9193 |
|
|
|
9194 |
|
|
static GTY(()) tree fpscr_values;
|
9195 |
|
|
|
9196 |
|
|
static void
|
9197 |
|
|
emit_fpu_switch (rtx scratch, int index)
|
9198 |
|
|
{
|
9199 |
|
|
rtx dst, src;
|
9200 |
|
|
|
9201 |
|
|
if (fpscr_values == NULL)
|
9202 |
|
|
{
|
9203 |
|
|
tree t;
|
9204 |
|
|
|
9205 |
|
|
t = build_index_type (integer_one_node);
|
9206 |
|
|
t = build_array_type (integer_type_node, t);
|
9207 |
|
|
t = build_decl (BUILTINS_LOCATION,
|
9208 |
|
|
VAR_DECL, get_identifier ("__fpscr_values"), t);
|
9209 |
|
|
DECL_ARTIFICIAL (t) = 1;
|
9210 |
|
|
DECL_IGNORED_P (t) = 1;
|
9211 |
|
|
DECL_EXTERNAL (t) = 1;
|
9212 |
|
|
TREE_STATIC (t) = 1;
|
9213 |
|
|
TREE_PUBLIC (t) = 1;
|
9214 |
|
|
TREE_USED (t) = 1;
|
9215 |
|
|
|
9216 |
|
|
fpscr_values = t;
|
9217 |
|
|
}
|
9218 |
|
|
|
9219 |
|
|
src = DECL_RTL (fpscr_values);
|
9220 |
|
|
if (!can_create_pseudo_p ())
|
9221 |
|
|
{
|
9222 |
|
|
emit_move_insn (scratch, XEXP (src, 0));
|
9223 |
|
|
if (index != 0)
|
9224 |
|
|
emit_insn (gen_addsi3 (scratch, scratch, GEN_INT (index * 4)));
|
9225 |
|
|
src = adjust_automodify_address (src, PSImode, scratch, index * 4);
|
9226 |
|
|
}
|
9227 |
|
|
else
|
9228 |
|
|
src = adjust_address (src, PSImode, index * 4);
|
9229 |
|
|
|
9230 |
|
|
dst = get_fpscr_rtx ();
|
9231 |
|
|
emit_move_insn (dst, src);
|
9232 |
|
|
}
|
9233 |
|
|
|
9234 |
|
|
void
|
9235 |
|
|
emit_sf_insn (rtx pat)
|
9236 |
|
|
{
|
9237 |
|
|
emit_insn (pat);
|
9238 |
|
|
}
|
9239 |
|
|
|
9240 |
|
|
void
|
9241 |
|
|
emit_df_insn (rtx pat)
|
9242 |
|
|
{
|
9243 |
|
|
emit_insn (pat);
|
9244 |
|
|
}
|
9245 |
|
|
|
9246 |
|
|
void
|
9247 |
|
|
expand_sf_unop (rtx (*fun) (rtx, rtx, rtx), rtx *operands)
|
9248 |
|
|
{
|
9249 |
|
|
emit_sf_insn ((*fun) (operands[0], operands[1], get_fpscr_rtx ()));
|
9250 |
|
|
}
|
9251 |
|
|
|
9252 |
|
|
void
|
9253 |
|
|
expand_sf_binop (rtx (*fun) (rtx, rtx, rtx, rtx), rtx *operands)
|
9254 |
|
|
{
|
9255 |
|
|
emit_sf_insn ((*fun) (operands[0], operands[1], operands[2],
|
9256 |
|
|
get_fpscr_rtx ()));
|
9257 |
|
|
}
|
9258 |
|
|
|
9259 |
|
|
void
|
9260 |
|
|
expand_df_unop (rtx (*fun) (rtx, rtx, rtx), rtx *operands)
|
9261 |
|
|
{
|
9262 |
|
|
emit_df_insn ((*fun) (operands[0], operands[1], get_fpscr_rtx ()));
|
9263 |
|
|
}
|
9264 |
|
|
|
9265 |
|
|
void
|
9266 |
|
|
expand_df_binop (rtx (*fun) (rtx, rtx, rtx, rtx), rtx *operands)
|
9267 |
|
|
{
|
9268 |
|
|
emit_df_insn ((*fun) (operands[0], operands[1], operands[2],
|
9269 |
|
|
get_fpscr_rtx ()));
|
9270 |
|
|
}
|
9271 |
|
|
|
9272 |
|
|
static rtx get_free_reg (HARD_REG_SET);
|
9273 |
|
|
|
9274 |
|
|
/* This function returns a register to use to load the address to load
|
9275 |
|
|
the fpscr from. Currently it always returns r1 or r7, but when we are
|
9276 |
|
|
able to use pseudo registers after combine, or have a better mechanism
|
9277 |
|
|
for choosing a register, it should be done here. */
|
9278 |
|
|
/* REGS_LIVE is the liveness information for the point for which we
|
9279 |
|
|
need this allocation. In some bare-bones exit blocks, r1 is live at the
|
9280 |
|
|
start. We can even have all of r0..r3 being live:
|
9281 |
|
|
__complex__ long long f (double d) { if (d == 0) return 2; else return 3; }
|
9282 |
|
|
INSN before which new insns are placed with will clobber the register
|
9283 |
|
|
we return. If a basic block consists only of setting the return value
|
9284 |
|
|
register to a pseudo and using that register, the return value is not
|
9285 |
|
|
live before or after this block, yet we we'll insert our insns right in
|
9286 |
|
|
the middle. */
|
9287 |
|
|
|
9288 |
|
|
static rtx
|
9289 |
|
|
get_free_reg (HARD_REG_SET regs_live)
|
9290 |
|
|
{
|
9291 |
|
|
if (! TEST_HARD_REG_BIT (regs_live, 1))
|
9292 |
|
|
return gen_rtx_REG (Pmode, 1);
|
9293 |
|
|
|
9294 |
|
|
/* Hard reg 1 is live; since this is a SMALL_REGISTER_CLASSES target,
|
9295 |
|
|
there shouldn't be anything but a jump before the function end. */
|
9296 |
|
|
gcc_assert (!TEST_HARD_REG_BIT (regs_live, 7));
|
9297 |
|
|
return gen_rtx_REG (Pmode, 7);
|
9298 |
|
|
}
|
9299 |
|
|
|
9300 |
|
|
/* This function will set the fpscr from memory.
|
9301 |
|
|
MODE is the mode we are setting it to. */
|
9302 |
|
|
void
|
9303 |
|
|
fpscr_set_from_mem (int mode, HARD_REG_SET regs_live)
|
9304 |
|
|
{
|
9305 |
|
|
enum attr_fp_mode fp_mode = (enum attr_fp_mode) mode;
|
9306 |
|
|
enum attr_fp_mode norm_mode = ACTUAL_NORMAL_MODE (FP_MODE);
|
9307 |
|
|
rtx addr_reg;
|
9308 |
|
|
|
9309 |
|
|
addr_reg = !can_create_pseudo_p () ? get_free_reg (regs_live) : NULL_RTX;
|
9310 |
|
|
emit_fpu_switch (addr_reg, fp_mode == norm_mode);
|
9311 |
|
|
}
|
9312 |
|
|
|
9313 |
|
|
/* Is the given character a logical line separator for the assembler? */
|
9314 |
|
|
#ifndef IS_ASM_LOGICAL_LINE_SEPARATOR
|
9315 |
|
|
#define IS_ASM_LOGICAL_LINE_SEPARATOR(C, STR) ((C) == ';')
|
9316 |
|
|
#endif
|
9317 |
|
|
|
9318 |
|
|
int
|
9319 |
|
|
sh_insn_length_adjustment (rtx insn)
|
9320 |
|
|
{
|
9321 |
|
|
/* Instructions with unfilled delay slots take up an extra two bytes for
|
9322 |
|
|
the nop in the delay slot. */
|
9323 |
|
|
if (((NONJUMP_INSN_P (insn)
|
9324 |
|
|
&& GET_CODE (PATTERN (insn)) != USE
|
9325 |
|
|
&& GET_CODE (PATTERN (insn)) != CLOBBER)
|
9326 |
|
|
|| CALL_P (insn)
|
9327 |
|
|
|| (JUMP_P (insn) && !JUMP_TABLE_DATA_P (insn)))
|
9328 |
|
|
&& GET_CODE (PATTERN (NEXT_INSN (PREV_INSN (insn)))) != SEQUENCE
|
9329 |
|
|
&& get_attr_needs_delay_slot (insn) == NEEDS_DELAY_SLOT_YES)
|
9330 |
|
|
return 2;
|
9331 |
|
|
|
9332 |
|
|
/* SH2e has a bug that prevents the use of annulled branches, so if
|
9333 |
|
|
the delay slot is not filled, we'll have to put a NOP in it. */
|
9334 |
|
|
if (sh_cpu_attr == CPU_SH2E
|
9335 |
|
|
&& JUMP_P (insn) && !JUMP_TABLE_DATA_P (insn)
|
9336 |
|
|
&& get_attr_type (insn) == TYPE_CBRANCH
|
9337 |
|
|
&& GET_CODE (PATTERN (NEXT_INSN (PREV_INSN (insn)))) != SEQUENCE)
|
9338 |
|
|
return 2;
|
9339 |
|
|
|
9340 |
|
|
/* sh-dsp parallel processing insn take four bytes instead of two. */
|
9341 |
|
|
|
9342 |
|
|
if (NONJUMP_INSN_P (insn))
|
9343 |
|
|
{
|
9344 |
|
|
int sum = 0;
|
9345 |
|
|
rtx body = PATTERN (insn);
|
9346 |
|
|
const char *templ;
|
9347 |
|
|
char c;
|
9348 |
|
|
int maybe_label = 1;
|
9349 |
|
|
|
9350 |
|
|
if (GET_CODE (body) == ASM_INPUT)
|
9351 |
|
|
templ = XSTR (body, 0);
|
9352 |
|
|
else if (asm_noperands (body) >= 0)
|
9353 |
|
|
templ
|
9354 |
|
|
= decode_asm_operands (body, NULL, NULL, NULL, NULL, NULL);
|
9355 |
|
|
else
|
9356 |
|
|
return 0;
|
9357 |
|
|
do
|
9358 |
|
|
{
|
9359 |
|
|
int ppi_adjust = 0;
|
9360 |
|
|
|
9361 |
|
|
do
|
9362 |
|
|
c = *templ++;
|
9363 |
|
|
while (c == ' ' || c == '\t');
|
9364 |
|
|
/* all sh-dsp parallel-processing insns start with p.
|
9365 |
|
|
The only non-ppi sh insn starting with p is pref.
|
9366 |
|
|
The only ppi starting with pr is prnd. */
|
9367 |
|
|
if ((c == 'p' || c == 'P') && strncasecmp ("re", templ, 2))
|
9368 |
|
|
ppi_adjust = 2;
|
9369 |
|
|
/* The repeat pseudo-insn expands two three insns, a total of
|
9370 |
|
|
six bytes in size. */
|
9371 |
|
|
else if ((c == 'r' || c == 'R')
|
9372 |
|
|
&& ! strncasecmp ("epeat", templ, 5))
|
9373 |
|
|
ppi_adjust = 4;
|
9374 |
|
|
while (c && c != '\n'
|
9375 |
|
|
&& ! IS_ASM_LOGICAL_LINE_SEPARATOR (c, templ))
|
9376 |
|
|
{
|
9377 |
|
|
/* If this is a label, it is obviously not a ppi insn. */
|
9378 |
|
|
if (c == ':' && maybe_label)
|
9379 |
|
|
{
|
9380 |
|
|
ppi_adjust = 0;
|
9381 |
|
|
break;
|
9382 |
|
|
}
|
9383 |
|
|
else if (c == '\'' || c == '"')
|
9384 |
|
|
maybe_label = 0;
|
9385 |
|
|
c = *templ++;
|
9386 |
|
|
}
|
9387 |
|
|
sum += ppi_adjust;
|
9388 |
|
|
maybe_label = c != ':';
|
9389 |
|
|
}
|
9390 |
|
|
while (c);
|
9391 |
|
|
return sum;
|
9392 |
|
|
}
|
9393 |
|
|
return 0;
|
9394 |
|
|
}
|
9395 |
|
|
|
9396 |
|
|
/* Return TRUE for a valid displacement for the REG+disp addressing
|
9397 |
|
|
with MODE. */
|
9398 |
|
|
|
9399 |
|
|
/* ??? The SH2e does not have the REG+disp addressing mode when loading values
|
9400 |
|
|
into the FRx registers. We implement this by setting the maximum offset
|
9401 |
|
|
to zero when the value is SFmode. This also restricts loading of SFmode
|
9402 |
|
|
values into the integer registers, but that can't be helped. */
|
9403 |
|
|
|
9404 |
|
|
/* The SH allows a displacement in a QI or HI amode, but only when the
|
9405 |
|
|
other operand is R0. GCC doesn't handle this very well, so we forgot
|
9406 |
|
|
all of that.
|
9407 |
|
|
|
9408 |
|
|
A legitimate index for a QI or HI is 0, SI can be any number 0..63,
|
9409 |
|
|
DI can be any number 0..60. */
|
9410 |
|
|
|
9411 |
|
|
bool
|
9412 |
|
|
sh_legitimate_index_p (enum machine_mode mode, rtx op)
|
9413 |
|
|
{
|
9414 |
|
|
if (CONST_INT_P (op))
|
9415 |
|
|
{
|
9416 |
|
|
if (TARGET_SHMEDIA)
|
9417 |
|
|
{
|
9418 |
|
|
int size;
|
9419 |
|
|
|
9420 |
|
|
/* Check if this the address of an unaligned load / store. */
|
9421 |
|
|
if (mode == VOIDmode)
|
9422 |
|
|
return CONST_OK_FOR_I06 (INTVAL (op));
|
9423 |
|
|
|
9424 |
|
|
size = GET_MODE_SIZE (mode);
|
9425 |
|
|
return (!(INTVAL (op) & (size - 1))
|
9426 |
|
|
&& INTVAL (op) >= -512 * size
|
9427 |
|
|
&& INTVAL (op) < 512 * size);
|
9428 |
|
|
}
|
9429 |
|
|
|
9430 |
|
|
if (TARGET_SH2A)
|
9431 |
|
|
{
|
9432 |
|
|
if (GET_MODE_SIZE (mode) == 1
|
9433 |
|
|
&& (unsigned) INTVAL (op) < 4096)
|
9434 |
|
|
return true;
|
9435 |
|
|
}
|
9436 |
|
|
|
9437 |
|
|
if ((GET_MODE_SIZE (mode) == 4
|
9438 |
|
|
&& (unsigned) INTVAL (op) < 64
|
9439 |
|
|
&& !(INTVAL (op) & 3)
|
9440 |
|
|
&& !(TARGET_SH2E && mode == SFmode))
|
9441 |
|
|
|| (GET_MODE_SIZE (mode) == 4
|
9442 |
|
|
&& (unsigned) INTVAL (op) < 16383
|
9443 |
|
|
&& !(INTVAL (op) & 3) && TARGET_SH2A))
|
9444 |
|
|
return true;
|
9445 |
|
|
|
9446 |
|
|
if ((GET_MODE_SIZE (mode) == 8
|
9447 |
|
|
&& (unsigned) INTVAL (op) < 60
|
9448 |
|
|
&& !(INTVAL (op) & 3)
|
9449 |
|
|
&& !((TARGET_SH4 || TARGET_SH2A) && mode == DFmode))
|
9450 |
|
|
|| ((GET_MODE_SIZE (mode)==8)
|
9451 |
|
|
&& (unsigned) INTVAL (op) < 8192
|
9452 |
|
|
&& !(INTVAL (op) & (TARGET_SH2A_DOUBLE ? 7 : 3))
|
9453 |
|
|
&& (TARGET_SH2A && mode == DFmode)))
|
9454 |
|
|
return true;
|
9455 |
|
|
}
|
9456 |
|
|
|
9457 |
|
|
return false;
|
9458 |
|
|
}
|
9459 |
|
|
|
9460 |
|
|
/* Recognize an RTL expression that is a valid memory address for
|
9461 |
|
|
an instruction.
|
9462 |
|
|
The MODE argument is the machine mode for the MEM expression
|
9463 |
|
|
that wants to use this address.
|
9464 |
|
|
Allow REG
|
9465 |
|
|
REG+disp
|
9466 |
|
|
REG+r0
|
9467 |
|
|
REG++
|
9468 |
|
|
--REG */
|
9469 |
|
|
|
9470 |
|
|
static bool
|
9471 |
|
|
sh_legitimate_address_p (enum machine_mode mode, rtx x, bool strict)
|
9472 |
|
|
{
|
9473 |
|
|
if (MAYBE_BASE_REGISTER_RTX_P (x, strict))
|
9474 |
|
|
return true;
|
9475 |
|
|
else if ((GET_CODE (x) == POST_INC || GET_CODE (x) == PRE_DEC)
|
9476 |
|
|
&& ! TARGET_SHMEDIA
|
9477 |
|
|
&& MAYBE_BASE_REGISTER_RTX_P (XEXP (x, 0), strict))
|
9478 |
|
|
return true;
|
9479 |
|
|
else if (GET_CODE (x) == PLUS
|
9480 |
|
|
&& (mode != PSImode || reload_completed))
|
9481 |
|
|
{
|
9482 |
|
|
rtx xop0 = XEXP (x, 0);
|
9483 |
|
|
rtx xop1 = XEXP (x, 1);
|
9484 |
|
|
|
9485 |
|
|
if (GET_MODE_SIZE (mode) <= 8
|
9486 |
|
|
&& MAYBE_BASE_REGISTER_RTX_P (xop0, strict)
|
9487 |
|
|
&& sh_legitimate_index_p (mode, xop1))
|
9488 |
|
|
return true;
|
9489 |
|
|
|
9490 |
|
|
if ((ALLOW_INDEXED_ADDRESS || GET_MODE (x) == DImode
|
9491 |
|
|
|| ((xop0 == stack_pointer_rtx
|
9492 |
|
|
|| xop0 == hard_frame_pointer_rtx)
|
9493 |
|
|
&& REG_P (xop1) && REGNO (xop1) == R0_REG)
|
9494 |
|
|
|| ((xop1 == stack_pointer_rtx
|
9495 |
|
|
|| xop1 == hard_frame_pointer_rtx)
|
9496 |
|
|
&& REG_P (xop0) && REGNO (xop0) == R0_REG))
|
9497 |
|
|
&& ((!TARGET_SHMEDIA && GET_MODE_SIZE (mode) <= 4)
|
9498 |
|
|
|| (TARGET_SHMEDIA && GET_MODE_SIZE (mode) <= 8)
|
9499 |
|
|
|| ((TARGET_SH4 || TARGET_SH2A_DOUBLE)
|
9500 |
|
|
&& TARGET_FMOVD && mode == DFmode)))
|
9501 |
|
|
{
|
9502 |
|
|
if (MAYBE_BASE_REGISTER_RTX_P (xop1, strict)
|
9503 |
|
|
&& MAYBE_INDEX_REGISTER_RTX_P (xop0, strict))
|
9504 |
|
|
return true;
|
9505 |
|
|
if (MAYBE_INDEX_REGISTER_RTX_P (xop1, strict)
|
9506 |
|
|
&& MAYBE_BASE_REGISTER_RTX_P (xop0, strict))
|
9507 |
|
|
return true;
|
9508 |
|
|
}
|
9509 |
|
|
}
|
9510 |
|
|
|
9511 |
|
|
return false;
|
9512 |
|
|
}
|
9513 |
|
|
|
9514 |
|
|
/* Return TRUE if X references a SYMBOL_REF or LABEL_REF whose symbol
|
9515 |
|
|
isn't protected by a PIC unspec. */
|
9516 |
|
|
int
|
9517 |
|
|
nonpic_symbol_mentioned_p (rtx x)
|
9518 |
|
|
{
|
9519 |
|
|
register const char *fmt;
|
9520 |
|
|
register int i;
|
9521 |
|
|
|
9522 |
|
|
if (GET_CODE (x) == SYMBOL_REF || GET_CODE (x) == LABEL_REF
|
9523 |
|
|
|| GET_CODE (x) == PC)
|
9524 |
|
|
return 1;
|
9525 |
|
|
|
9526 |
|
|
/* We don't want to look into the possible MEM location of a
|
9527 |
|
|
CONST_DOUBLE, since we're not going to use it, in general. */
|
9528 |
|
|
if (GET_CODE (x) == CONST_DOUBLE)
|
9529 |
|
|
return 0;
|
9530 |
|
|
|
9531 |
|
|
if (GET_CODE (x) == UNSPEC
|
9532 |
|
|
&& (XINT (x, 1) == UNSPEC_PIC
|
9533 |
|
|
|| XINT (x, 1) == UNSPEC_GOT
|
9534 |
|
|
|| XINT (x, 1) == UNSPEC_GOTOFF
|
9535 |
|
|
|| XINT (x, 1) == UNSPEC_GOTPLT
|
9536 |
|
|
|| XINT (x, 1) == UNSPEC_GOTTPOFF
|
9537 |
|
|
|| XINT (x, 1) == UNSPEC_DTPOFF
|
9538 |
|
|
|| XINT (x, 1) == UNSPEC_TPOFF
|
9539 |
|
|
|| XINT (x, 1) == UNSPEC_PLT
|
9540 |
|
|
|| XINT (x, 1) == UNSPEC_SYMOFF
|
9541 |
|
|
|| XINT (x, 1) == UNSPEC_PCREL_SYMOFF))
|
9542 |
|
|
return 0;
|
9543 |
|
|
|
9544 |
|
|
fmt = GET_RTX_FORMAT (GET_CODE (x));
|
9545 |
|
|
for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--)
|
9546 |
|
|
{
|
9547 |
|
|
if (fmt[i] == 'E')
|
9548 |
|
|
{
|
9549 |
|
|
register int j;
|
9550 |
|
|
|
9551 |
|
|
for (j = XVECLEN (x, i) - 1; j >= 0; j--)
|
9552 |
|
|
if (nonpic_symbol_mentioned_p (XVECEXP (x, i, j)))
|
9553 |
|
|
return 1;
|
9554 |
|
|
}
|
9555 |
|
|
else if (fmt[i] == 'e' && nonpic_symbol_mentioned_p (XEXP (x, i)))
|
9556 |
|
|
return 1;
|
9557 |
|
|
}
|
9558 |
|
|
|
9559 |
|
|
return 0;
|
9560 |
|
|
}
|
9561 |
|
|
|
9562 |
|
|
/* Convert a non-PIC address in `orig' to a PIC address using @GOT or
|
9563 |
|
|
@GOTOFF in `reg'. */
|
9564 |
|
|
rtx
|
9565 |
|
|
legitimize_pic_address (rtx orig, enum machine_mode mode ATTRIBUTE_UNUSED,
|
9566 |
|
|
rtx reg)
|
9567 |
|
|
{
|
9568 |
|
|
if (tls_symbolic_operand (orig, Pmode) != TLS_MODEL_NONE)
|
9569 |
|
|
return orig;
|
9570 |
|
|
|
9571 |
|
|
if (GET_CODE (orig) == LABEL_REF
|
9572 |
|
|
|| (GET_CODE (orig) == SYMBOL_REF && SYMBOL_REF_LOCAL_P (orig)))
|
9573 |
|
|
{
|
9574 |
|
|
if (reg == 0)
|
9575 |
|
|
reg = gen_reg_rtx (Pmode);
|
9576 |
|
|
|
9577 |
|
|
emit_insn (gen_symGOTOFF2reg (reg, orig));
|
9578 |
|
|
return reg;
|
9579 |
|
|
}
|
9580 |
|
|
else if (GET_CODE (orig) == SYMBOL_REF)
|
9581 |
|
|
{
|
9582 |
|
|
if (reg == 0)
|
9583 |
|
|
reg = gen_reg_rtx (Pmode);
|
9584 |
|
|
|
9585 |
|
|
emit_insn (gen_symGOT2reg (reg, orig));
|
9586 |
|
|
return reg;
|
9587 |
|
|
}
|
9588 |
|
|
return orig;
|
9589 |
|
|
}
|
9590 |
|
|
|
9591 |
|
|
/* Try machine-dependent ways of modifying an illegitimate address
|
9592 |
|
|
to be legitimate. If we find one, return the new, valid address.
|
9593 |
|
|
Otherwise, return X.
|
9594 |
|
|
|
9595 |
|
|
For the SH, if X is almost suitable for indexing, but the offset is
|
9596 |
|
|
out of range, convert it into a normal form so that CSE has a chance
|
9597 |
|
|
of reducing the number of address registers used. */
|
9598 |
|
|
|
9599 |
|
|
static rtx
|
9600 |
|
|
sh_legitimize_address (rtx x, rtx oldx, enum machine_mode mode)
|
9601 |
|
|
{
|
9602 |
|
|
if (flag_pic)
|
9603 |
|
|
x = legitimize_pic_address (oldx, mode, NULL_RTX);
|
9604 |
|
|
|
9605 |
|
|
if (GET_CODE (x) == PLUS
|
9606 |
|
|
&& (GET_MODE_SIZE (mode) == 4
|
9607 |
|
|
|| GET_MODE_SIZE (mode) == 8)
|
9608 |
|
|
&& CONST_INT_P (XEXP (x, 1))
|
9609 |
|
|
&& BASE_REGISTER_RTX_P (XEXP (x, 0))
|
9610 |
|
|
&& ! TARGET_SHMEDIA
|
9611 |
|
|
&& ! ((TARGET_SH4 || TARGET_SH2A_DOUBLE) && mode == DFmode)
|
9612 |
|
|
&& ! (TARGET_SH2E && mode == SFmode))
|
9613 |
|
|
{
|
9614 |
|
|
rtx index_rtx = XEXP (x, 1);
|
9615 |
|
|
HOST_WIDE_INT offset = INTVAL (index_rtx), offset_base;
|
9616 |
|
|
rtx sum;
|
9617 |
|
|
|
9618 |
|
|
/* On rare occasions, we might get an unaligned pointer
|
9619 |
|
|
that is indexed in a way to give an aligned address.
|
9620 |
|
|
Therefore, keep the lower two bits in offset_base. */
|
9621 |
|
|
/* Instead of offset_base 128..131 use 124..127, so that
|
9622 |
|
|
simple add suffices. */
|
9623 |
|
|
if (offset > 127)
|
9624 |
|
|
offset_base = ((offset + 4) & ~60) - 4;
|
9625 |
|
|
else
|
9626 |
|
|
offset_base = offset & ~60;
|
9627 |
|
|
|
9628 |
|
|
/* Sometimes the normal form does not suit DImode. We
|
9629 |
|
|
could avoid that by using smaller ranges, but that
|
9630 |
|
|
would give less optimized code when SImode is
|
9631 |
|
|
prevalent. */
|
9632 |
|
|
if (GET_MODE_SIZE (mode) + offset - offset_base <= 64)
|
9633 |
|
|
{
|
9634 |
|
|
sum = expand_binop (Pmode, add_optab, XEXP (x, 0),
|
9635 |
|
|
GEN_INT (offset_base), NULL_RTX, 0,
|
9636 |
|
|
OPTAB_LIB_WIDEN);
|
9637 |
|
|
|
9638 |
|
|
return gen_rtx_PLUS (Pmode, sum, GEN_INT (offset - offset_base));
|
9639 |
|
|
}
|
9640 |
|
|
}
|
9641 |
|
|
|
9642 |
|
|
return x;
|
9643 |
|
|
}
|
9644 |
|
|
|
9645 |
|
|
/* Mark the use of a constant in the literal table. If the constant
|
9646 |
|
|
has multiple labels, make it unique. */
|
9647 |
|
|
static rtx
|
9648 |
|
|
mark_constant_pool_use (rtx x)
|
9649 |
|
|
{
|
9650 |
|
|
rtx insn, lab, pattern;
|
9651 |
|
|
|
9652 |
|
|
if (x == NULL)
|
9653 |
|
|
return x;
|
9654 |
|
|
|
9655 |
|
|
switch (GET_CODE (x))
|
9656 |
|
|
{
|
9657 |
|
|
case LABEL_REF:
|
9658 |
|
|
x = XEXP (x, 0);
|
9659 |
|
|
case CODE_LABEL:
|
9660 |
|
|
break;
|
9661 |
|
|
default:
|
9662 |
|
|
return x;
|
9663 |
|
|
}
|
9664 |
|
|
|
9665 |
|
|
/* Get the first label in the list of labels for the same constant
|
9666 |
|
|
and delete another labels in the list. */
|
9667 |
|
|
lab = x;
|
9668 |
|
|
for (insn = PREV_INSN (x); insn; insn = PREV_INSN (insn))
|
9669 |
|
|
{
|
9670 |
|
|
if (!LABEL_P (insn)
|
9671 |
|
|
|| LABEL_REFS (insn) != NEXT_INSN (insn))
|
9672 |
|
|
break;
|
9673 |
|
|
lab = insn;
|
9674 |
|
|
}
|
9675 |
|
|
|
9676 |
|
|
for (insn = LABEL_REFS (lab); insn; insn = LABEL_REFS (insn))
|
9677 |
|
|
INSN_DELETED_P (insn) = 1;
|
9678 |
|
|
|
9679 |
|
|
/* Mark constants in a window. */
|
9680 |
|
|
for (insn = NEXT_INSN (x); insn; insn = NEXT_INSN (insn))
|
9681 |
|
|
{
|
9682 |
|
|
if (!NONJUMP_INSN_P (insn))
|
9683 |
|
|
continue;
|
9684 |
|
|
|
9685 |
|
|
pattern = PATTERN (insn);
|
9686 |
|
|
if (GET_CODE (pattern) != UNSPEC_VOLATILE)
|
9687 |
|
|
continue;
|
9688 |
|
|
|
9689 |
|
|
switch (XINT (pattern, 1))
|
9690 |
|
|
{
|
9691 |
|
|
case UNSPECV_CONST2:
|
9692 |
|
|
case UNSPECV_CONST4:
|
9693 |
|
|
case UNSPECV_CONST8:
|
9694 |
|
|
XVECEXP (pattern, 0, 1) = const1_rtx;
|
9695 |
|
|
break;
|
9696 |
|
|
case UNSPECV_WINDOW_END:
|
9697 |
|
|
if (XVECEXP (pattern, 0, 0) == x)
|
9698 |
|
|
return lab;
|
9699 |
|
|
break;
|
9700 |
|
|
case UNSPECV_CONST_END:
|
9701 |
|
|
return lab;
|
9702 |
|
|
default:
|
9703 |
|
|
break;
|
9704 |
|
|
}
|
9705 |
|
|
}
|
9706 |
|
|
|
9707 |
|
|
return lab;
|
9708 |
|
|
}
|
9709 |
|
|
|
9710 |
|
|
/* Return true if it's possible to redirect BRANCH1 to the destination
|
9711 |
|
|
of an unconditional jump BRANCH2. We only want to do this if the
|
9712 |
|
|
resulting branch will have a short displacement. */
|
9713 |
|
|
int
|
9714 |
|
|
sh_can_redirect_branch (rtx branch1, rtx branch2)
|
9715 |
|
|
{
|
9716 |
|
|
if (flag_expensive_optimizations && simplejump_p (branch2))
|
9717 |
|
|
{
|
9718 |
|
|
rtx dest = XEXP (SET_SRC (single_set (branch2)), 0);
|
9719 |
|
|
rtx insn;
|
9720 |
|
|
int distance;
|
9721 |
|
|
|
9722 |
|
|
for (distance = 0, insn = NEXT_INSN (branch1);
|
9723 |
|
|
insn && distance < 256;
|
9724 |
|
|
insn = PREV_INSN (insn))
|
9725 |
|
|
{
|
9726 |
|
|
if (insn == dest)
|
9727 |
|
|
return 1;
|
9728 |
|
|
else
|
9729 |
|
|
distance += get_attr_length (insn);
|
9730 |
|
|
}
|
9731 |
|
|
for (distance = 0, insn = NEXT_INSN (branch1);
|
9732 |
|
|
insn && distance < 256;
|
9733 |
|
|
insn = NEXT_INSN (insn))
|
9734 |
|
|
{
|
9735 |
|
|
if (insn == dest)
|
9736 |
|
|
return 1;
|
9737 |
|
|
else
|
9738 |
|
|
distance += get_attr_length (insn);
|
9739 |
|
|
}
|
9740 |
|
|
}
|
9741 |
|
|
return 0;
|
9742 |
|
|
}
|
9743 |
|
|
|
9744 |
|
|
/* Return nonzero if register old_reg can be renamed to register new_reg. */
|
9745 |
|
|
int
|
9746 |
|
|
sh_hard_regno_rename_ok (unsigned int old_reg ATTRIBUTE_UNUSED,
|
9747 |
|
|
unsigned int new_reg)
|
9748 |
|
|
{
|
9749 |
|
|
/* Interrupt functions can only use registers that have already been
|
9750 |
|
|
saved by the prologue, even if they would normally be
|
9751 |
|
|
call-clobbered. */
|
9752 |
|
|
|
9753 |
|
|
if (sh_cfun_interrupt_handler_p () && !df_regs_ever_live_p (new_reg))
|
9754 |
|
|
return 0;
|
9755 |
|
|
|
9756 |
|
|
return 1;
|
9757 |
|
|
}
|
9758 |
|
|
|
9759 |
|
|
/* Function to update the integer COST
|
9760 |
|
|
based on the relationship between INSN that is dependent on
|
9761 |
|
|
DEP_INSN through the dependence LINK. The default is to make no
|
9762 |
|
|
adjustment to COST. This can be used for example to specify to
|
9763 |
|
|
the scheduler that an output- or anti-dependence does not incur
|
9764 |
|
|
the same cost as a data-dependence. The return value should be
|
9765 |
|
|
the new value for COST. */
|
9766 |
|
|
static int
|
9767 |
|
|
sh_adjust_cost (rtx insn, rtx link ATTRIBUTE_UNUSED, rtx dep_insn, int cost)
|
9768 |
|
|
{
|
9769 |
|
|
rtx reg, use_pat;
|
9770 |
|
|
|
9771 |
|
|
if (TARGET_SHMEDIA)
|
9772 |
|
|
{
|
9773 |
|
|
/* On SHmedia, if the dependence is an anti-dependence or
|
9774 |
|
|
output-dependence, there is no cost. */
|
9775 |
|
|
if (REG_NOTE_KIND (link) != 0)
|
9776 |
|
|
{
|
9777 |
|
|
/* However, dependencies between target register loads and
|
9778 |
|
|
uses of the register in a subsequent block that are separated
|
9779 |
|
|
by a conditional branch are not modelled - we have to do with
|
9780 |
|
|
the anti-dependency between the target register load and the
|
9781 |
|
|
conditional branch that ends the current block. */
|
9782 |
|
|
if (REG_NOTE_KIND (link) == REG_DEP_ANTI
|
9783 |
|
|
&& GET_CODE (PATTERN (dep_insn)) == SET
|
9784 |
|
|
&& (get_attr_type (dep_insn) == TYPE_PT_MEDIA
|
9785 |
|
|
|| get_attr_type (dep_insn) == TYPE_PTABS_MEDIA)
|
9786 |
|
|
&& get_attr_type (insn) == TYPE_CBRANCH_MEDIA)
|
9787 |
|
|
{
|
9788 |
|
|
int orig_cost = cost;
|
9789 |
|
|
rtx note = find_reg_note (insn, REG_BR_PROB, 0);
|
9790 |
|
|
rtx target = ((! note
|
9791 |
|
|
|| INTVAL (XEXP (note, 0)) * 2 < REG_BR_PROB_BASE)
|
9792 |
|
|
? insn : JUMP_LABEL (insn));
|
9793 |
|
|
/* On the likely path, the branch costs 1, on the unlikely path,
|
9794 |
|
|
it costs 3. */
|
9795 |
|
|
cost--;
|
9796 |
|
|
do
|
9797 |
|
|
target = next_active_insn (target);
|
9798 |
|
|
while (target && ! flow_dependent_p (target, dep_insn)
|
9799 |
|
|
&& --cost > 0);
|
9800 |
|
|
/* If two branches are executed in immediate succession, with the
|
9801 |
|
|
first branch properly predicted, this causes a stall at the
|
9802 |
|
|
second branch, hence we won't need the target for the
|
9803 |
|
|
second branch for two cycles after the launch of the first
|
9804 |
|
|
branch. */
|
9805 |
|
|
if (cost > orig_cost - 2)
|
9806 |
|
|
cost = orig_cost - 2;
|
9807 |
|
|
}
|
9808 |
|
|
else
|
9809 |
|
|
cost = 0;
|
9810 |
|
|
}
|
9811 |
|
|
|
9812 |
|
|
else if (get_attr_is_mac_media (insn)
|
9813 |
|
|
&& get_attr_is_mac_media (dep_insn))
|
9814 |
|
|
cost = 1;
|
9815 |
|
|
|
9816 |
|
|
else if (! reload_completed
|
9817 |
|
|
&& GET_CODE (PATTERN (insn)) == SET
|
9818 |
|
|
&& GET_CODE (SET_SRC (PATTERN (insn))) == FLOAT
|
9819 |
|
|
&& GET_CODE (PATTERN (dep_insn)) == SET
|
9820 |
|
|
&& fp_arith_reg_operand (SET_SRC (PATTERN (dep_insn)), VOIDmode)
|
9821 |
|
|
&& cost < 4)
|
9822 |
|
|
cost = 4;
|
9823 |
|
|
/* Schedule the ptabs for a casesi_jump_media in preference to stuff
|
9824 |
|
|
that is needed at the target. */
|
9825 |
|
|
else if (get_attr_type (insn) == TYPE_JUMP_MEDIA
|
9826 |
|
|
&& ! flow_dependent_p (insn, dep_insn))
|
9827 |
|
|
cost--;
|
9828 |
|
|
}
|
9829 |
|
|
else if (REG_NOTE_KIND (link) == 0)
|
9830 |
|
|
{
|
9831 |
|
|
enum attr_type type;
|
9832 |
|
|
rtx dep_set;
|
9833 |
|
|
|
9834 |
|
|
if (recog_memoized (insn) < 0
|
9835 |
|
|
|| recog_memoized (dep_insn) < 0)
|
9836 |
|
|
return cost;
|
9837 |
|
|
|
9838 |
|
|
dep_set = single_set (dep_insn);
|
9839 |
|
|
|
9840 |
|
|
/* The latency that we specify in the scheduling description refers
|
9841 |
|
|
to the actual output, not to an auto-increment register; for that,
|
9842 |
|
|
the latency is one. */
|
9843 |
|
|
if (dep_set && MEM_P (SET_SRC (dep_set)) && cost > 1)
|
9844 |
|
|
{
|
9845 |
|
|
rtx set = single_set (insn);
|
9846 |
|
|
|
9847 |
|
|
if (set
|
9848 |
|
|
&& !reg_mentioned_p (SET_DEST (dep_set), SET_SRC (set))
|
9849 |
|
|
&& (!MEM_P (SET_DEST (set))
|
9850 |
|
|
|| !reg_mentioned_p (SET_DEST (dep_set),
|
9851 |
|
|
XEXP (SET_DEST (set), 0))))
|
9852 |
|
|
cost = 1;
|
9853 |
|
|
}
|
9854 |
|
|
/* The only input for a call that is timing-critical is the
|
9855 |
|
|
function's address. */
|
9856 |
|
|
if (CALL_P (insn))
|
9857 |
|
|
{
|
9858 |
|
|
rtx call = PATTERN (insn);
|
9859 |
|
|
|
9860 |
|
|
if (GET_CODE (call) == PARALLEL)
|
9861 |
|
|
call = XVECEXP (call, 0 ,0);
|
9862 |
|
|
if (GET_CODE (call) == SET)
|
9863 |
|
|
call = SET_SRC (call);
|
9864 |
|
|
if (GET_CODE (call) == CALL && MEM_P (XEXP (call, 0))
|
9865 |
|
|
/* sibcalli_thunk uses a symbol_ref in an unspec. */
|
9866 |
|
|
&& (GET_CODE (XEXP (XEXP (call, 0), 0)) == UNSPEC
|
9867 |
|
|
|| ! reg_set_p (XEXP (XEXP (call, 0), 0), dep_insn)))
|
9868 |
|
|
cost -= TARGET_SH4_300 ? 3 : 6;
|
9869 |
|
|
}
|
9870 |
|
|
/* Likewise, the most timing critical input for an sfuncs call
|
9871 |
|
|
is the function address. However, sfuncs typically start
|
9872 |
|
|
using their arguments pretty quickly.
|
9873 |
|
|
Assume a four cycle delay for SH4 before they are needed.
|
9874 |
|
|
Cached ST40-300 calls are quicker, so assume only a one
|
9875 |
|
|
cycle delay there.
|
9876 |
|
|
??? Maybe we should encode the delays till input registers
|
9877 |
|
|
are needed by sfuncs into the sfunc call insn. */
|
9878 |
|
|
/* All sfunc calls are parallels with at least four components.
|
9879 |
|
|
Exploit this to avoid unnecessary calls to sfunc_uses_reg. */
|
9880 |
|
|
else if (GET_CODE (PATTERN (insn)) == PARALLEL
|
9881 |
|
|
&& XVECLEN (PATTERN (insn), 0) >= 4
|
9882 |
|
|
&& (reg = sfunc_uses_reg (insn)))
|
9883 |
|
|
{
|
9884 |
|
|
if (! reg_set_p (reg, dep_insn))
|
9885 |
|
|
cost -= TARGET_SH4_300 ? 1 : 4;
|
9886 |
|
|
}
|
9887 |
|
|
if (TARGET_HARD_SH4 && !TARGET_SH4_300)
|
9888 |
|
|
{
|
9889 |
|
|
enum attr_type dep_type = get_attr_type (dep_insn);
|
9890 |
|
|
|
9891 |
|
|
if (dep_type == TYPE_FLOAD || dep_type == TYPE_PCFLOAD)
|
9892 |
|
|
cost--;
|
9893 |
|
|
else if ((dep_type == TYPE_LOAD_SI || dep_type == TYPE_PCLOAD_SI)
|
9894 |
|
|
&& (type = get_attr_type (insn)) != TYPE_CALL
|
9895 |
|
|
&& type != TYPE_SFUNC)
|
9896 |
|
|
cost--;
|
9897 |
|
|
/* When the preceding instruction loads the shift amount of
|
9898 |
|
|
the following SHAD/SHLD, the latency of the load is increased
|
9899 |
|
|
by 1 cycle. */
|
9900 |
|
|
if (get_attr_type (insn) == TYPE_DYN_SHIFT
|
9901 |
|
|
&& get_attr_any_int_load (dep_insn) == ANY_INT_LOAD_YES
|
9902 |
|
|
&& reg_overlap_mentioned_p (SET_DEST (dep_set),
|
9903 |
|
|
XEXP (SET_SRC (single_set (insn)),
|
9904 |
|
|
1)))
|
9905 |
|
|
cost++;
|
9906 |
|
|
/* When an LS group instruction with a latency of less than
|
9907 |
|
|
3 cycles is followed by a double-precision floating-point
|
9908 |
|
|
instruction, FIPR, or FTRV, the latency of the first
|
9909 |
|
|
instruction is increased to 3 cycles. */
|
9910 |
|
|
else if (cost < 3
|
9911 |
|
|
&& get_attr_insn_class (dep_insn) == INSN_CLASS_LS_GROUP
|
9912 |
|
|
&& get_attr_dfp_comp (insn) == DFP_COMP_YES)
|
9913 |
|
|
cost = 3;
|
9914 |
|
|
/* The lsw register of a double-precision computation is ready one
|
9915 |
|
|
cycle earlier. */
|
9916 |
|
|
else if (reload_completed
|
9917 |
|
|
&& get_attr_dfp_comp (dep_insn) == DFP_COMP_YES
|
9918 |
|
|
&& (use_pat = single_set (insn))
|
9919 |
|
|
&& ! regno_use_in (REGNO (SET_DEST (single_set (dep_insn))),
|
9920 |
|
|
SET_SRC (use_pat)))
|
9921 |
|
|
cost -= 1;
|
9922 |
|
|
|
9923 |
|
|
if (get_attr_any_fp_comp (dep_insn) == ANY_FP_COMP_YES
|
9924 |
|
|
&& get_attr_late_fp_use (insn) == LATE_FP_USE_YES)
|
9925 |
|
|
cost -= 1;
|
9926 |
|
|
}
|
9927 |
|
|
else if (TARGET_SH4_300)
|
9928 |
|
|
{
|
9929 |
|
|
/* Stores need their input register two cycles later. */
|
9930 |
|
|
if (dep_set && cost >= 1
|
9931 |
|
|
&& ((type = get_attr_type (insn)) == TYPE_STORE
|
9932 |
|
|
|| type == TYPE_PSTORE
|
9933 |
|
|
|| type == TYPE_FSTORE || type == TYPE_MAC_MEM))
|
9934 |
|
|
{
|
9935 |
|
|
rtx set = single_set (insn);
|
9936 |
|
|
|
9937 |
|
|
if (!reg_mentioned_p (SET_SRC (set), XEXP (SET_DEST (set), 0))
|
9938 |
|
|
&& rtx_equal_p (SET_SRC (set), SET_DEST (dep_set)))
|
9939 |
|
|
{
|
9940 |
|
|
cost -= 2;
|
9941 |
|
|
/* But don't reduce the cost below 1 if the address depends
|
9942 |
|
|
on a side effect of dep_insn. */
|
9943 |
|
|
if (cost < 1
|
9944 |
|
|
&& modified_in_p (XEXP (SET_DEST (set), 0), dep_insn))
|
9945 |
|
|
cost = 1;
|
9946 |
|
|
}
|
9947 |
|
|
}
|
9948 |
|
|
}
|
9949 |
|
|
}
|
9950 |
|
|
/* An anti-dependence penalty of two applies if the first insn is a double
|
9951 |
|
|
precision fadd / fsub / fmul. */
|
9952 |
|
|
else if (!TARGET_SH4_300
|
9953 |
|
|
&& REG_NOTE_KIND (link) == REG_DEP_ANTI
|
9954 |
|
|
&& recog_memoized (dep_insn) >= 0
|
9955 |
|
|
&& (get_attr_type (dep_insn) == TYPE_DFP_ARITH
|
9956 |
|
|
|| get_attr_type (dep_insn) == TYPE_DFP_MUL)
|
9957 |
|
|
/* A lot of alleged anti-flow dependences are fake,
|
9958 |
|
|
so check this one is real. */
|
9959 |
|
|
&& flow_dependent_p (dep_insn, insn))
|
9960 |
|
|
cost = 2;
|
9961 |
|
|
|
9962 |
|
|
return cost;
|
9963 |
|
|
}
|
9964 |
|
|
|
9965 |
|
|
/* Check if INSN is flow-dependent on DEP_INSN. Can also be used to check
|
9966 |
|
|
if DEP_INSN is anti-flow dependent on INSN. */
|
9967 |
|
|
static int
|
9968 |
|
|
flow_dependent_p (rtx insn, rtx dep_insn)
|
9969 |
|
|
{
|
9970 |
|
|
rtx tmp = PATTERN (insn);
|
9971 |
|
|
|
9972 |
|
|
note_stores (PATTERN (dep_insn), flow_dependent_p_1, &tmp);
|
9973 |
|
|
return tmp == NULL_RTX;
|
9974 |
|
|
}
|
9975 |
|
|
|
9976 |
|
|
/* A helper function for flow_dependent_p called through note_stores. */
|
9977 |
|
|
static void
|
9978 |
|
|
flow_dependent_p_1 (rtx x, const_rtx pat ATTRIBUTE_UNUSED, void *data)
|
9979 |
|
|
{
|
9980 |
|
|
rtx * pinsn = (rtx *) data;
|
9981 |
|
|
|
9982 |
|
|
if (*pinsn && reg_referenced_p (x, *pinsn))
|
9983 |
|
|
*pinsn = NULL_RTX;
|
9984 |
|
|
}
|
9985 |
|
|
|
9986 |
|
|
/* For use by sh_allocate_initial_value. Note that sh.md contains some
|
9987 |
|
|
'special function' patterns (type sfunc) that clobber pr, but that
|
9988 |
|
|
do not look like function calls to leaf_function_p. Hence we must
|
9989 |
|
|
do this extra check. */
|
9990 |
|
|
static int
|
9991 |
|
|
sh_pr_n_sets (void)
|
9992 |
|
|
{
|
9993 |
|
|
return DF_REG_DEF_COUNT (TARGET_SHMEDIA ? PR_MEDIA_REG : PR_REG);
|
9994 |
|
|
}
|
9995 |
|
|
|
9996 |
|
|
/* Return where to allocate pseudo for a given hard register initial
|
9997 |
|
|
value. */
|
9998 |
|
|
static rtx
|
9999 |
|
|
sh_allocate_initial_value (rtx hard_reg)
|
10000 |
|
|
{
|
10001 |
|
|
rtx x;
|
10002 |
|
|
|
10003 |
|
|
if (REGNO (hard_reg) == (TARGET_SHMEDIA ? PR_MEDIA_REG : PR_REG))
|
10004 |
|
|
{
|
10005 |
|
|
if (current_function_is_leaf
|
10006 |
|
|
&& ! sh_pr_n_sets ()
|
10007 |
|
|
&& ! (TARGET_SHCOMPACT
|
10008 |
|
|
&& ((crtl->args.info.call_cookie
|
10009 |
|
|
& ~ CALL_COOKIE_RET_TRAMP (1))
|
10010 |
|
|
|| crtl->saves_all_registers)))
|
10011 |
|
|
x = hard_reg;
|
10012 |
|
|
else
|
10013 |
|
|
x = gen_frame_mem (Pmode, return_address_pointer_rtx);
|
10014 |
|
|
}
|
10015 |
|
|
else
|
10016 |
|
|
x = NULL_RTX;
|
10017 |
|
|
|
10018 |
|
|
return x;
|
10019 |
|
|
}
|
10020 |
|
|
|
10021 |
|
|
/* This function returns "2" to indicate dual issue for the SH4
|
10022 |
|
|
processor. To be used by the DFA pipeline description. */
|
10023 |
|
|
static int
|
10024 |
|
|
sh_issue_rate (void)
|
10025 |
|
|
{
|
10026 |
|
|
if (TARGET_SUPERSCALAR)
|
10027 |
|
|
return 2;
|
10028 |
|
|
else
|
10029 |
|
|
return 1;
|
10030 |
|
|
}
|
10031 |
|
|
|
10032 |
|
|
/* Functions for ready queue reordering for sched1. */
|
10033 |
|
|
|
10034 |
|
|
/* Get weight for mode for a set x. */
|
10035 |
|
|
static short
|
10036 |
|
|
find_set_regmode_weight (rtx x, enum machine_mode mode)
|
10037 |
|
|
{
|
10038 |
|
|
if (GET_CODE (x) == CLOBBER && register_operand (SET_DEST (x), mode))
|
10039 |
|
|
return 1;
|
10040 |
|
|
if (GET_CODE (x) == SET && register_operand (SET_DEST (x), mode))
|
10041 |
|
|
{
|
10042 |
|
|
if (REG_P (SET_DEST (x)))
|
10043 |
|
|
{
|
10044 |
|
|
if (!reg_mentioned_p (SET_DEST (x), SET_SRC (x)))
|
10045 |
|
|
return 1;
|
10046 |
|
|
else
|
10047 |
|
|
return 0;
|
10048 |
|
|
}
|
10049 |
|
|
return 1;
|
10050 |
|
|
}
|
10051 |
|
|
return 0;
|
10052 |
|
|
}
|
10053 |
|
|
|
10054 |
|
|
/* Get regmode weight for insn. */
|
10055 |
|
|
static short
|
10056 |
|
|
find_insn_regmode_weight (rtx insn, enum machine_mode mode)
|
10057 |
|
|
{
|
10058 |
|
|
short reg_weight = 0;
|
10059 |
|
|
rtx x;
|
10060 |
|
|
|
10061 |
|
|
/* Increment weight for each register born here. */
|
10062 |
|
|
x = PATTERN (insn);
|
10063 |
|
|
reg_weight += find_set_regmode_weight (x, mode);
|
10064 |
|
|
if (GET_CODE (x) == PARALLEL)
|
10065 |
|
|
{
|
10066 |
|
|
int j;
|
10067 |
|
|
for (j = XVECLEN (x, 0) - 1; j >= 0; j--)
|
10068 |
|
|
{
|
10069 |
|
|
x = XVECEXP (PATTERN (insn), 0, j);
|
10070 |
|
|
reg_weight += find_set_regmode_weight (x, mode);
|
10071 |
|
|
}
|
10072 |
|
|
}
|
10073 |
|
|
/* Decrement weight for each register that dies here. */
|
10074 |
|
|
for (x = REG_NOTES (insn); x; x = XEXP (x, 1))
|
10075 |
|
|
{
|
10076 |
|
|
if (REG_NOTE_KIND (x) == REG_DEAD || REG_NOTE_KIND (x) == REG_UNUSED)
|
10077 |
|
|
{
|
10078 |
|
|
rtx note = XEXP (x, 0);
|
10079 |
|
|
if (REG_P (note) && GET_MODE (note) == mode)
|
10080 |
|
|
reg_weight--;
|
10081 |
|
|
}
|
10082 |
|
|
}
|
10083 |
|
|
return reg_weight;
|
10084 |
|
|
}
|
10085 |
|
|
|
10086 |
|
|
/* Calculate regmode weights for all insns of a basic block. */
|
10087 |
|
|
static void
|
10088 |
|
|
find_regmode_weight (basic_block b, enum machine_mode mode)
|
10089 |
|
|
{
|
10090 |
|
|
rtx insn, next_tail, head, tail;
|
10091 |
|
|
|
10092 |
|
|
get_ebb_head_tail (b, b, &head, &tail);
|
10093 |
|
|
next_tail = NEXT_INSN (tail);
|
10094 |
|
|
|
10095 |
|
|
for (insn = head; insn != next_tail; insn = NEXT_INSN (insn))
|
10096 |
|
|
{
|
10097 |
|
|
/* Handle register life information. */
|
10098 |
|
|
if (!INSN_P (insn))
|
10099 |
|
|
continue;
|
10100 |
|
|
|
10101 |
|
|
if (mode == SFmode)
|
10102 |
|
|
INSN_REGMODE_WEIGHT (insn, mode) =
|
10103 |
|
|
find_insn_regmode_weight (insn, mode) + 2 * find_insn_regmode_weight (insn, DFmode);
|
10104 |
|
|
else if (mode == SImode)
|
10105 |
|
|
INSN_REGMODE_WEIGHT (insn, mode) =
|
10106 |
|
|
find_insn_regmode_weight (insn, mode) + 2 * find_insn_regmode_weight (insn, DImode);
|
10107 |
|
|
}
|
10108 |
|
|
}
|
10109 |
|
|
|
10110 |
|
|
/* Comparison function for ready queue sorting. */
|
10111 |
|
|
static int
|
10112 |
|
|
rank_for_reorder (const void *x, const void *y)
|
10113 |
|
|
{
|
10114 |
|
|
rtx tmp = *(const rtx *) y;
|
10115 |
|
|
rtx tmp2 = *(const rtx *) x;
|
10116 |
|
|
|
10117 |
|
|
/* The insn in a schedule group should be issued the first. */
|
10118 |
|
|
if (SCHED_GROUP_P (tmp) != SCHED_GROUP_P (tmp2))
|
10119 |
|
|
return SCHED_GROUP_P (tmp2) ? 1 : -1;
|
10120 |
|
|
|
10121 |
|
|
/* If insns are equally good, sort by INSN_LUID (original insn order), This
|
10122 |
|
|
minimizes instruction movement, thus minimizing sched's effect on
|
10123 |
|
|
register pressure. */
|
10124 |
|
|
return INSN_LUID (tmp) - INSN_LUID (tmp2);
|
10125 |
|
|
}
|
10126 |
|
|
|
10127 |
|
|
/* Resort the array A in which only element at index N may be out of order. */
|
10128 |
|
|
static void
|
10129 |
|
|
swap_reorder (rtx *a, int n)
|
10130 |
|
|
{
|
10131 |
|
|
rtx insn = a[n - 1];
|
10132 |
|
|
int i = n - 2;
|
10133 |
|
|
|
10134 |
|
|
while (i >= 0 && rank_for_reorder (a + i, &insn) >= 0)
|
10135 |
|
|
{
|
10136 |
|
|
a[i + 1] = a[i];
|
10137 |
|
|
i -= 1;
|
10138 |
|
|
}
|
10139 |
|
|
a[i + 1] = insn;
|
10140 |
|
|
}
|
10141 |
|
|
|
10142 |
|
|
#define SCHED_REORDER(READY, N_READY) \
|
10143 |
|
|
do \
|
10144 |
|
|
{ \
|
10145 |
|
|
if ((N_READY) == 2) \
|
10146 |
|
|
swap_reorder (READY, N_READY); \
|
10147 |
|
|
else if ((N_READY) > 2) \
|
10148 |
|
|
qsort (READY, N_READY, sizeof (rtx), rank_for_reorder); \
|
10149 |
|
|
} \
|
10150 |
|
|
while (0)
|
10151 |
|
|
|
10152 |
|
|
/* Sort the ready list READY by ascending priority, using the SCHED_REORDER
|
10153 |
|
|
macro. */
|
10154 |
|
|
static void
|
10155 |
|
|
ready_reorder (rtx *ready, int nready)
|
10156 |
|
|
{
|
10157 |
|
|
SCHED_REORDER (ready, nready);
|
10158 |
|
|
}
|
10159 |
|
|
|
10160 |
|
|
/* Count life regions of r0 for a block. */
|
10161 |
|
|
static int
|
10162 |
|
|
find_r0_life_regions (basic_block b)
|
10163 |
|
|
{
|
10164 |
|
|
rtx end, insn;
|
10165 |
|
|
rtx pset;
|
10166 |
|
|
rtx r0_reg;
|
10167 |
|
|
int live;
|
10168 |
|
|
int set;
|
10169 |
|
|
int death = 0;
|
10170 |
|
|
|
10171 |
|
|
if (REGNO_REG_SET_P (df_get_live_in (b), R0_REG))
|
10172 |
|
|
{
|
10173 |
|
|
set = 1;
|
10174 |
|
|
live = 1;
|
10175 |
|
|
}
|
10176 |
|
|
else
|
10177 |
|
|
{
|
10178 |
|
|
set = 0;
|
10179 |
|
|
live = 0;
|
10180 |
|
|
}
|
10181 |
|
|
|
10182 |
|
|
insn = BB_HEAD (b);
|
10183 |
|
|
end = BB_END (b);
|
10184 |
|
|
r0_reg = gen_rtx_REG (SImode, R0_REG);
|
10185 |
|
|
while (1)
|
10186 |
|
|
{
|
10187 |
|
|
if (INSN_P (insn))
|
10188 |
|
|
{
|
10189 |
|
|
if (find_regno_note (insn, REG_DEAD, R0_REG))
|
10190 |
|
|
{
|
10191 |
|
|
death++;
|
10192 |
|
|
live = 0;
|
10193 |
|
|
}
|
10194 |
|
|
if (!live
|
10195 |
|
|
&& (pset = single_set (insn))
|
10196 |
|
|
&& reg_overlap_mentioned_p (r0_reg, SET_DEST (pset))
|
10197 |
|
|
&& !find_regno_note (insn, REG_UNUSED, R0_REG))
|
10198 |
|
|
{
|
10199 |
|
|
set++;
|
10200 |
|
|
live = 1;
|
10201 |
|
|
}
|
10202 |
|
|
}
|
10203 |
|
|
if (insn == end)
|
10204 |
|
|
break;
|
10205 |
|
|
insn = NEXT_INSN (insn);
|
10206 |
|
|
}
|
10207 |
|
|
return set - death;
|
10208 |
|
|
}
|
10209 |
|
|
|
10210 |
|
|
/* Calculate regmode weights for all insns of all basic block. */
|
10211 |
|
|
static void
|
10212 |
|
|
sh_md_init_global (FILE *dump ATTRIBUTE_UNUSED,
|
10213 |
|
|
int verbose ATTRIBUTE_UNUSED,
|
10214 |
|
|
int old_max_uid)
|
10215 |
|
|
{
|
10216 |
|
|
basic_block b;
|
10217 |
|
|
|
10218 |
|
|
regmode_weight[0] = (short *) xcalloc (old_max_uid, sizeof (short));
|
10219 |
|
|
regmode_weight[1] = (short *) xcalloc (old_max_uid, sizeof (short));
|
10220 |
|
|
r0_life_regions = 0;
|
10221 |
|
|
|
10222 |
|
|
FOR_EACH_BB_REVERSE (b)
|
10223 |
|
|
{
|
10224 |
|
|
find_regmode_weight (b, SImode);
|
10225 |
|
|
find_regmode_weight (b, SFmode);
|
10226 |
|
|
if (!reload_completed)
|
10227 |
|
|
r0_life_regions += find_r0_life_regions (b);
|
10228 |
|
|
}
|
10229 |
|
|
|
10230 |
|
|
CURR_REGMODE_PRESSURE (SImode) = 0;
|
10231 |
|
|
CURR_REGMODE_PRESSURE (SFmode) = 0;
|
10232 |
|
|
|
10233 |
|
|
}
|
10234 |
|
|
|
10235 |
|
|
/* Cleanup. */
|
10236 |
|
|
static void
|
10237 |
|
|
sh_md_finish_global (FILE *dump ATTRIBUTE_UNUSED,
|
10238 |
|
|
int verbose ATTRIBUTE_UNUSED)
|
10239 |
|
|
{
|
10240 |
|
|
if (regmode_weight[0])
|
10241 |
|
|
{
|
10242 |
|
|
free (regmode_weight[0]);
|
10243 |
|
|
regmode_weight[0] = NULL;
|
10244 |
|
|
}
|
10245 |
|
|
if (regmode_weight[1])
|
10246 |
|
|
{
|
10247 |
|
|
free (regmode_weight[1]);
|
10248 |
|
|
regmode_weight[1] = NULL;
|
10249 |
|
|
}
|
10250 |
|
|
}
|
10251 |
|
|
|
10252 |
|
|
/* The scalar modes supported differs from the default version in TImode
|
10253 |
|
|
for 32-bit SHMEDIA. */
|
10254 |
|
|
static bool
|
10255 |
|
|
sh_scalar_mode_supported_p (enum machine_mode mode)
|
10256 |
|
|
{
|
10257 |
|
|
if (TARGET_SHMEDIA32 && mode == TImode)
|
10258 |
|
|
return false;
|
10259 |
|
|
|
10260 |
|
|
return default_scalar_mode_supported_p (mode);
|
10261 |
|
|
}
|
10262 |
|
|
|
10263 |
|
|
/* Cache the can_issue_more so that we can return it from reorder2. Also,
|
10264 |
|
|
keep count of register pressures on SImode and SFmode. */
|
10265 |
|
|
static int
|
10266 |
|
|
sh_variable_issue (FILE *dump ATTRIBUTE_UNUSED,
|
10267 |
|
|
int sched_verbose ATTRIBUTE_UNUSED,
|
10268 |
|
|
rtx insn,
|
10269 |
|
|
int can_issue_more)
|
10270 |
|
|
{
|
10271 |
|
|
if (GET_CODE (PATTERN (insn)) != USE
|
10272 |
|
|
&& GET_CODE (PATTERN (insn)) != CLOBBER)
|
10273 |
|
|
cached_can_issue_more = can_issue_more - 1;
|
10274 |
|
|
else
|
10275 |
|
|
cached_can_issue_more = can_issue_more;
|
10276 |
|
|
|
10277 |
|
|
if (reload_completed)
|
10278 |
|
|
return cached_can_issue_more;
|
10279 |
|
|
|
10280 |
|
|
CURR_REGMODE_PRESSURE (SImode) += INSN_REGMODE_WEIGHT (insn, SImode);
|
10281 |
|
|
CURR_REGMODE_PRESSURE (SFmode) += INSN_REGMODE_WEIGHT (insn, SFmode);
|
10282 |
|
|
|
10283 |
|
|
return cached_can_issue_more;
|
10284 |
|
|
}
|
10285 |
|
|
|
10286 |
|
|
static void
|
10287 |
|
|
sh_md_init (FILE *dump ATTRIBUTE_UNUSED,
|
10288 |
|
|
int verbose ATTRIBUTE_UNUSED,
|
10289 |
|
|
int veclen ATTRIBUTE_UNUSED)
|
10290 |
|
|
{
|
10291 |
|
|
CURR_REGMODE_PRESSURE (SImode) = 0;
|
10292 |
|
|
CURR_REGMODE_PRESSURE (SFmode) = 0;
|
10293 |
|
|
}
|
10294 |
|
|
|
10295 |
|
|
/* Some magic numbers. */
|
10296 |
|
|
/* Pressure on register r0 can lead to spill failures. so avoid sched1 for
|
10297 |
|
|
functions that already have high pressure on r0. */
|
10298 |
|
|
#define R0_MAX_LIFE_REGIONS 2
|
10299 |
|
|
/* Register Pressure thresholds for SImode and SFmode registers. */
|
10300 |
|
|
#define SIMODE_MAX_WEIGHT 5
|
10301 |
|
|
#define SFMODE_MAX_WEIGHT 10
|
10302 |
|
|
|
10303 |
|
|
/* Return true if the pressure is high for MODE. */
|
10304 |
|
|
static short
|
10305 |
|
|
high_pressure (enum machine_mode mode)
|
10306 |
|
|
{
|
10307 |
|
|
/* Pressure on register r0 can lead to spill failures. so avoid sched1 for
|
10308 |
|
|
functions that already have high pressure on r0. */
|
10309 |
|
|
if (r0_life_regions >= R0_MAX_LIFE_REGIONS)
|
10310 |
|
|
return 1;
|
10311 |
|
|
|
10312 |
|
|
if (mode == SFmode)
|
10313 |
|
|
return (CURR_REGMODE_PRESSURE (SFmode) > SFMODE_MAX_WEIGHT);
|
10314 |
|
|
else
|
10315 |
|
|
return (CURR_REGMODE_PRESSURE (SImode) > SIMODE_MAX_WEIGHT);
|
10316 |
|
|
}
|
10317 |
|
|
|
10318 |
|
|
/* Reorder ready queue if register pressure is high. */
|
10319 |
|
|
static int
|
10320 |
|
|
sh_reorder (FILE *dump ATTRIBUTE_UNUSED,
|
10321 |
|
|
int sched_verbose ATTRIBUTE_UNUSED,
|
10322 |
|
|
rtx *ready,
|
10323 |
|
|
int *n_readyp,
|
10324 |
|
|
int clock_var ATTRIBUTE_UNUSED)
|
10325 |
|
|
{
|
10326 |
|
|
if (reload_completed)
|
10327 |
|
|
return sh_issue_rate ();
|
10328 |
|
|
|
10329 |
|
|
if (high_pressure (SFmode) || high_pressure (SImode))
|
10330 |
|
|
{
|
10331 |
|
|
ready_reorder (ready, *n_readyp);
|
10332 |
|
|
}
|
10333 |
|
|
|
10334 |
|
|
return sh_issue_rate ();
|
10335 |
|
|
}
|
10336 |
|
|
|
10337 |
|
|
/* Skip cycles if the current register pressure is high. */
|
10338 |
|
|
static int
|
10339 |
|
|
sh_reorder2 (FILE *dump ATTRIBUTE_UNUSED,
|
10340 |
|
|
int sched_verbose ATTRIBUTE_UNUSED,
|
10341 |
|
|
rtx *ready ATTRIBUTE_UNUSED,
|
10342 |
|
|
int *n_readyp ATTRIBUTE_UNUSED,
|
10343 |
|
|
int clock_var ATTRIBUTE_UNUSED)
|
10344 |
|
|
{
|
10345 |
|
|
if (reload_completed)
|
10346 |
|
|
return cached_can_issue_more;
|
10347 |
|
|
|
10348 |
|
|
if (high_pressure(SFmode) || high_pressure (SImode))
|
10349 |
|
|
skip_cycles = 1;
|
10350 |
|
|
|
10351 |
|
|
return cached_can_issue_more;
|
10352 |
|
|
}
|
10353 |
|
|
|
10354 |
|
|
/* Skip cycles without sorting the ready queue. This will move insn from
|
10355 |
|
|
Q->R. If this is the last cycle we are skipping; allow sorting of ready
|
10356 |
|
|
queue by sh_reorder. */
|
10357 |
|
|
|
10358 |
|
|
/* Generally, skipping these many cycles are sufficient for all insns to move
|
10359 |
|
|
from Q -> R. */
|
10360 |
|
|
#define MAX_SKIPS 8
|
10361 |
|
|
|
10362 |
|
|
static int
|
10363 |
|
|
sh_dfa_new_cycle (FILE *sched_dump ATTRIBUTE_UNUSED,
|
10364 |
|
|
int sched_verbose ATTRIBUTE_UNUSED,
|
10365 |
|
|
rtx insn ATTRIBUTE_UNUSED,
|
10366 |
|
|
int last_clock_var,
|
10367 |
|
|
int clock_var,
|
10368 |
|
|
int *sort_p)
|
10369 |
|
|
{
|
10370 |
|
|
if (reload_completed)
|
10371 |
|
|
return 0;
|
10372 |
|
|
|
10373 |
|
|
if (skip_cycles)
|
10374 |
|
|
{
|
10375 |
|
|
if ((clock_var - last_clock_var) < MAX_SKIPS)
|
10376 |
|
|
{
|
10377 |
|
|
*sort_p = 0;
|
10378 |
|
|
return 1;
|
10379 |
|
|
}
|
10380 |
|
|
/* If this is the last cycle we are skipping, allow reordering of R. */
|
10381 |
|
|
if ((clock_var - last_clock_var) == MAX_SKIPS)
|
10382 |
|
|
{
|
10383 |
|
|
*sort_p = 1;
|
10384 |
|
|
return 1;
|
10385 |
|
|
}
|
10386 |
|
|
}
|
10387 |
|
|
|
10388 |
|
|
skip_cycles = 0;
|
10389 |
|
|
|
10390 |
|
|
return 0;
|
10391 |
|
|
}
|
10392 |
|
|
|
10393 |
|
|
/* SHmedia requires registers for branches, so we can't generate new
|
10394 |
|
|
branches past reload. */
|
10395 |
|
|
static bool
|
10396 |
|
|
sh_cannot_modify_jumps_p (void)
|
10397 |
|
|
{
|
10398 |
|
|
return (TARGET_SHMEDIA && (reload_in_progress || reload_completed));
|
10399 |
|
|
}
|
10400 |
|
|
|
10401 |
|
|
static enum reg_class
|
10402 |
|
|
sh_target_reg_class (void)
|
10403 |
|
|
{
|
10404 |
|
|
return TARGET_SHMEDIA ? TARGET_REGS : NO_REGS;
|
10405 |
|
|
}
|
10406 |
|
|
|
10407 |
|
|
static bool
|
10408 |
|
|
sh_optimize_target_register_callee_saved (bool after_prologue_epilogue_gen)
|
10409 |
|
|
{
|
10410 |
|
|
HARD_REG_SET dummy;
|
10411 |
|
|
#if 0
|
10412 |
|
|
rtx insn;
|
10413 |
|
|
#endif
|
10414 |
|
|
|
10415 |
|
|
if (! shmedia_space_reserved_for_target_registers)
|
10416 |
|
|
return 0;
|
10417 |
|
|
if (after_prologue_epilogue_gen && ! TARGET_SAVE_ALL_TARGET_REGS)
|
10418 |
|
|
return 0;
|
10419 |
|
|
if (calc_live_regs (&dummy) >= 6 * 8)
|
10420 |
|
|
return 1;
|
10421 |
|
|
return 0;
|
10422 |
|
|
}
|
10423 |
|
|
|
10424 |
|
|
static bool
|
10425 |
|
|
sh_ms_bitfield_layout_p (const_tree record_type ATTRIBUTE_UNUSED)
|
10426 |
|
|
{
|
10427 |
|
|
return (TARGET_SH5 || TARGET_HITACHI || sh_attr_renesas_p (record_type));
|
10428 |
|
|
}
|
10429 |
|
|
|
10430 |
|
|
/*
|
10431 |
|
|
On the SH1..SH4, the trampoline looks like
|
10432 |
|
|
2 0002 D202 mov.l l2,r2
|
10433 |
|
|
1 0000 D301 mov.l l1,r3
|
10434 |
|
|
3 0004 422B jmp @r2
|
10435 |
|
|
4 0006 0009 nop
|
10436 |
|
|
5 0008 00000000 l1: .long area
|
10437 |
|
|
6 000c 00000000 l2: .long function
|
10438 |
|
|
|
10439 |
|
|
SH5 (compact) uses r1 instead of r3 for the static chain. */
|
10440 |
|
|
|
10441 |
|
|
|
10442 |
|
|
/* Emit RTL insns to initialize the variable parts of a trampoline.
|
10443 |
|
|
FNADDR is an RTX for the address of the function's pure code.
|
10444 |
|
|
CXT is an RTX for the static chain value for the function. */
|
10445 |
|
|
|
10446 |
|
|
static void
|
10447 |
|
|
sh_trampoline_init (rtx tramp_mem, tree fndecl, rtx cxt)
|
10448 |
|
|
{
|
10449 |
|
|
rtx fnaddr = XEXP (DECL_RTL (fndecl), 0);
|
10450 |
|
|
rtx tramp = force_reg (Pmode, XEXP (tramp_mem, 0));
|
10451 |
|
|
|
10452 |
|
|
if (TARGET_SHMEDIA64)
|
10453 |
|
|
{
|
10454 |
|
|
rtx tramp_templ;
|
10455 |
|
|
int fixed_len;
|
10456 |
|
|
|
10457 |
|
|
rtx movi1 = GEN_INT (0xcc000010);
|
10458 |
|
|
rtx shori1 = GEN_INT (0xc8000010);
|
10459 |
|
|
rtx src, dst;
|
10460 |
|
|
|
10461 |
|
|
/* The following trampoline works within a +- 128 KB range for cxt:
|
10462 |
|
|
ptb/u cxt,tr1; movi fnaddr >> 48,r0; shori fnaddr >> 32,r0;
|
10463 |
|
|
shori fnaddr >> 16,r0; shori fnaddr,r0; ptabs/l r0,tr0
|
10464 |
|
|
gettr tr1,r1; blink tr0,r63 */
|
10465 |
|
|
/* Address rounding makes it hard to compute the exact bounds of the
|
10466 |
|
|
offset for this trampoline, but we have a rather generous offset
|
10467 |
|
|
range, so frame_offset should do fine as an upper bound. */
|
10468 |
|
|
if (cxt == virtual_stack_vars_rtx && frame_offset < 0x20000)
|
10469 |
|
|
{
|
10470 |
|
|
/* ??? could optimize this trampoline initialization
|
10471 |
|
|
by writing DImode words with two insns each. */
|
10472 |
|
|
rtx mask = force_reg (DImode, GEN_INT (0x3fffc00));
|
10473 |
|
|
rtx insn = gen_rtx_MINUS (DImode, cxt, tramp);
|
10474 |
|
|
insn = gen_rtx_ASHIFT (DImode, insn, GEN_INT (10-2));
|
10475 |
|
|
insn = gen_rtx_AND (DImode, insn, mask);
|
10476 |
|
|
/* Or in ptb/u .,tr1 pattern */
|
10477 |
|
|
insn = gen_rtx_IOR (DImode, insn, gen_int_mode (0xec000010, SImode));
|
10478 |
|
|
insn = force_operand (insn, NULL_RTX);
|
10479 |
|
|
insn = gen_lowpart (SImode, insn);
|
10480 |
|
|
emit_move_insn (change_address (tramp_mem, SImode, NULL_RTX), insn);
|
10481 |
|
|
insn = gen_rtx_LSHIFTRT (DImode, fnaddr, GEN_INT (38));
|
10482 |
|
|
insn = gen_rtx_AND (DImode, insn, mask);
|
10483 |
|
|
insn = force_operand (gen_rtx_IOR (DImode, movi1, insn), NULL_RTX);
|
10484 |
|
|
insn = gen_lowpart (SImode, insn);
|
10485 |
|
|
emit_move_insn (adjust_address (tramp_mem, SImode, 4), insn);
|
10486 |
|
|
insn = gen_rtx_LSHIFTRT (DImode, fnaddr, GEN_INT (22));
|
10487 |
|
|
insn = gen_rtx_AND (DImode, insn, mask);
|
10488 |
|
|
insn = force_operand (gen_rtx_IOR (DImode, shori1, insn), NULL_RTX);
|
10489 |
|
|
insn = gen_lowpart (SImode, insn);
|
10490 |
|
|
emit_move_insn (adjust_address (tramp_mem, SImode, 8), insn);
|
10491 |
|
|
insn = gen_rtx_LSHIFTRT (DImode, fnaddr, GEN_INT (6));
|
10492 |
|
|
insn = gen_rtx_AND (DImode, insn, mask);
|
10493 |
|
|
insn = force_operand (gen_rtx_IOR (DImode, shori1, insn), NULL_RTX);
|
10494 |
|
|
insn = gen_lowpart (SImode, insn);
|
10495 |
|
|
emit_move_insn (adjust_address (tramp_mem, SImode, 12), insn);
|
10496 |
|
|
insn = gen_rtx_ASHIFT (DImode, fnaddr, GEN_INT (10));
|
10497 |
|
|
insn = gen_rtx_AND (DImode, insn, mask);
|
10498 |
|
|
insn = force_operand (gen_rtx_IOR (DImode, shori1, insn), NULL_RTX);
|
10499 |
|
|
insn = gen_lowpart (SImode, insn);
|
10500 |
|
|
emit_move_insn (adjust_address (tramp_mem, SImode, 16), insn);
|
10501 |
|
|
emit_move_insn (adjust_address (tramp_mem, SImode, 20),
|
10502 |
|
|
GEN_INT (0x6bf10600));
|
10503 |
|
|
emit_move_insn (adjust_address (tramp_mem, SImode, 24),
|
10504 |
|
|
GEN_INT (0x4415fc10));
|
10505 |
|
|
emit_move_insn (adjust_address (tramp_mem, SImode, 28),
|
10506 |
|
|
GEN_INT (0x4401fff0));
|
10507 |
|
|
emit_insn (gen_ic_invalidate_line (tramp));
|
10508 |
|
|
return;
|
10509 |
|
|
}
|
10510 |
|
|
tramp_templ = gen_rtx_SYMBOL_REF (Pmode,"__GCC_nested_trampoline");
|
10511 |
|
|
fixed_len = TRAMPOLINE_SIZE - 2 * GET_MODE_SIZE (Pmode);
|
10512 |
|
|
|
10513 |
|
|
tramp_templ = gen_datalabel_ref (tramp_templ);
|
10514 |
|
|
dst = tramp_mem;
|
10515 |
|
|
src = gen_const_mem (BLKmode, tramp_templ);
|
10516 |
|
|
set_mem_align (dst, 256);
|
10517 |
|
|
set_mem_align (src, 64);
|
10518 |
|
|
emit_block_move (dst, src, GEN_INT (fixed_len), BLOCK_OP_NORMAL);
|
10519 |
|
|
|
10520 |
|
|
emit_move_insn (adjust_address (tramp_mem, Pmode, fixed_len), fnaddr);
|
10521 |
|
|
emit_move_insn (adjust_address (tramp_mem, Pmode,
|
10522 |
|
|
fixed_len + GET_MODE_SIZE (Pmode)),
|
10523 |
|
|
cxt);
|
10524 |
|
|
emit_insn (gen_ic_invalidate_line (tramp));
|
10525 |
|
|
return;
|
10526 |
|
|
}
|
10527 |
|
|
else if (TARGET_SHMEDIA)
|
10528 |
|
|
{
|
10529 |
|
|
/* movi fnaddr >> 16,r1; shori fnaddr,r1; ptabs/l r1,tr0
|
10530 |
|
|
movi cxt >> 16,r1; shori cxt,r1; blink tr0,r63 */
|
10531 |
|
|
rtx quad0 = gen_reg_rtx (DImode), cxtload = gen_reg_rtx (DImode);
|
10532 |
|
|
rtx quad1 = gen_reg_rtx (DImode), quad2 = gen_reg_rtx (DImode);
|
10533 |
|
|
/* movi 0,r1: 0xcc000010 shori 0,r1: c8000010 concatenated,
|
10534 |
|
|
rotated 10 right, and higher 16 bit of every 32 selected. */
|
10535 |
|
|
rtx movishori
|
10536 |
|
|
= force_reg (V2HImode, (simplify_gen_subreg
|
10537 |
|
|
(V2HImode, GEN_INT (0x4330432), SImode, 0)));
|
10538 |
|
|
rtx ptabs = force_reg (DImode, GEN_INT (0x6bf10600));
|
10539 |
|
|
rtx blink = force_reg (DImode, GEN_INT (0x4401fff0));
|
10540 |
|
|
|
10541 |
|
|
fnaddr = force_reg (SImode, fnaddr);
|
10542 |
|
|
cxt = force_reg (SImode, cxt);
|
10543 |
|
|
emit_insn (gen_mshflo_w_x (gen_rtx_SUBREG (V4HImode, quad0, 0),
|
10544 |
|
|
gen_rtx_SUBREG (V2HImode, fnaddr, 0),
|
10545 |
|
|
movishori));
|
10546 |
|
|
emit_insn (gen_rotrdi3_mextr (quad0, quad0,
|
10547 |
|
|
GEN_INT (TARGET_LITTLE_ENDIAN ? 24 : 56)));
|
10548 |
|
|
emit_insn (gen_ashldi3_media (quad0, quad0, const2_rtx));
|
10549 |
|
|
emit_move_insn (change_address (tramp_mem, DImode, NULL_RTX), quad0);
|
10550 |
|
|
emit_insn (gen_mshflo_w_x (gen_rtx_SUBREG (V4HImode, cxtload, 0),
|
10551 |
|
|
gen_rtx_SUBREG (V2HImode, cxt, 0),
|
10552 |
|
|
movishori));
|
10553 |
|
|
emit_insn (gen_rotrdi3_mextr (cxtload, cxtload,
|
10554 |
|
|
GEN_INT (TARGET_LITTLE_ENDIAN ? 24 : 56)));
|
10555 |
|
|
emit_insn (gen_ashldi3_media (cxtload, cxtload, const2_rtx));
|
10556 |
|
|
if (TARGET_LITTLE_ENDIAN)
|
10557 |
|
|
{
|
10558 |
|
|
emit_insn (gen_mshflo_l_di (quad1, ptabs, cxtload));
|
10559 |
|
|
emit_insn (gen_mextr4 (quad2, cxtload, blink));
|
10560 |
|
|
}
|
10561 |
|
|
else
|
10562 |
|
|
{
|
10563 |
|
|
emit_insn (gen_mextr4 (quad1, cxtload, ptabs));
|
10564 |
|
|
emit_insn (gen_mshflo_l_di (quad2, blink, cxtload));
|
10565 |
|
|
}
|
10566 |
|
|
emit_move_insn (adjust_address (tramp_mem, DImode, 8), quad1);
|
10567 |
|
|
emit_move_insn (adjust_address (tramp_mem, DImode, 16), quad2);
|
10568 |
|
|
emit_insn (gen_ic_invalidate_line (tramp));
|
10569 |
|
|
return;
|
10570 |
|
|
}
|
10571 |
|
|
else if (TARGET_SHCOMPACT)
|
10572 |
|
|
{
|
10573 |
|
|
emit_insn (gen_initialize_trampoline (tramp, cxt, fnaddr));
|
10574 |
|
|
return;
|
10575 |
|
|
}
|
10576 |
|
|
emit_move_insn (change_address (tramp_mem, SImode, NULL_RTX),
|
10577 |
|
|
gen_int_mode (TARGET_LITTLE_ENDIAN ? 0xd301d202 : 0xd202d301,
|
10578 |
|
|
SImode));
|
10579 |
|
|
emit_move_insn (adjust_address (tramp_mem, SImode, 4),
|
10580 |
|
|
gen_int_mode (TARGET_LITTLE_ENDIAN ? 0x0009422b : 0x422b0009,
|
10581 |
|
|
SImode));
|
10582 |
|
|
emit_move_insn (adjust_address (tramp_mem, SImode, 8), cxt);
|
10583 |
|
|
emit_move_insn (adjust_address (tramp_mem, SImode, 12), fnaddr);
|
10584 |
|
|
if (TARGET_HARVARD)
|
10585 |
|
|
{
|
10586 |
|
|
if (!TARGET_INLINE_IC_INVALIDATE
|
10587 |
|
|
|| (!(TARGET_SH4A_ARCH || TARGET_SH4_300) && TARGET_USERMODE))
|
10588 |
|
|
emit_library_call (function_symbol (NULL, "__ic_invalidate",
|
10589 |
|
|
FUNCTION_ORDINARY),
|
10590 |
|
|
LCT_NORMAL, VOIDmode, 1, tramp, SImode);
|
10591 |
|
|
else
|
10592 |
|
|
emit_insn (gen_ic_invalidate_line (tramp));
|
10593 |
|
|
}
|
10594 |
|
|
}
|
10595 |
|
|
|
10596 |
|
|
/* On SH5, trampolines are SHmedia code, so add 1 to the address. */
|
10597 |
|
|
|
10598 |
|
|
static rtx
|
10599 |
|
|
sh_trampoline_adjust_address (rtx tramp)
|
10600 |
|
|
{
|
10601 |
|
|
if (TARGET_SHMEDIA)
|
10602 |
|
|
tramp = expand_simple_binop (Pmode, PLUS, tramp, const1_rtx,
|
10603 |
|
|
gen_reg_rtx (Pmode), 0, OPTAB_LIB_WIDEN);
|
10604 |
|
|
return tramp;
|
10605 |
|
|
}
|
10606 |
|
|
|
10607 |
|
|
/* FIXME: This is overly conservative. A SHcompact function that
|
10608 |
|
|
receives arguments ``by reference'' will have them stored in its
|
10609 |
|
|
own stack frame, so it must not pass pointers or references to
|
10610 |
|
|
these arguments to other functions by means of sibling calls. */
|
10611 |
|
|
/* If PIC, we cannot make sibling calls to global functions
|
10612 |
|
|
because the PLT requires r12 to be live. */
|
10613 |
|
|
static bool
|
10614 |
|
|
sh_function_ok_for_sibcall (tree decl, tree exp ATTRIBUTE_UNUSED)
|
10615 |
|
|
{
|
10616 |
|
|
return (1
|
10617 |
|
|
&& (! TARGET_SHCOMPACT
|
10618 |
|
|
|| crtl->args.info.stack_regs == 0)
|
10619 |
|
|
&& ! sh_cfun_interrupt_handler_p ()
|
10620 |
|
|
&& (! flag_pic
|
10621 |
|
|
|| (decl && ! TREE_PUBLIC (decl))
|
10622 |
|
|
|| (decl && DECL_VISIBILITY (decl) != VISIBILITY_DEFAULT)));
|
10623 |
|
|
}
|
10624 |
|
|
|
10625 |
|
|
/* Machine specific built-in functions. */
|
10626 |
|
|
|
10627 |
|
|
struct builtin_description
|
10628 |
|
|
{
|
10629 |
|
|
const enum insn_code icode;
|
10630 |
|
|
const char *const name;
|
10631 |
|
|
int signature;
|
10632 |
|
|
tree fndecl;
|
10633 |
|
|
};
|
10634 |
|
|
|
10635 |
|
|
/* describe number and signedness of arguments; arg[0] == result
|
10636 |
|
|
(1: unsigned, 2: signed, 4: don't care, 8: pointer 0: no argument */
|
10637 |
|
|
/* 9: 64-bit pointer, 10: 32-bit pointer */
|
10638 |
|
|
static const char signature_args[][4] =
|
10639 |
|
|
{
|
10640 |
|
|
#define SH_BLTIN_V2SI2 0
|
10641 |
|
|
{ 4, 4 },
|
10642 |
|
|
#define SH_BLTIN_V4HI2 1
|
10643 |
|
|
{ 4, 4 },
|
10644 |
|
|
#define SH_BLTIN_V2SI3 2
|
10645 |
|
|
{ 4, 4, 4 },
|
10646 |
|
|
#define SH_BLTIN_V4HI3 3
|
10647 |
|
|
{ 4, 4, 4 },
|
10648 |
|
|
#define SH_BLTIN_V8QI3 4
|
10649 |
|
|
{ 4, 4, 4 },
|
10650 |
|
|
#define SH_BLTIN_MAC_HISI 5
|
10651 |
|
|
{ 1, 4, 4, 1 },
|
10652 |
|
|
#define SH_BLTIN_SH_HI 6
|
10653 |
|
|
{ 4, 4, 1 },
|
10654 |
|
|
#define SH_BLTIN_SH_SI 7
|
10655 |
|
|
{ 4, 4, 1 },
|
10656 |
|
|
#define SH_BLTIN_V4HI2V2SI 8
|
10657 |
|
|
{ 4, 4, 4 },
|
10658 |
|
|
#define SH_BLTIN_V4HI2V8QI 9
|
10659 |
|
|
{ 4, 4, 4 },
|
10660 |
|
|
#define SH_BLTIN_SISF 10
|
10661 |
|
|
{ 4, 2 },
|
10662 |
|
|
#define SH_BLTIN_LDUA_L 11
|
10663 |
|
|
{ 2, 10 },
|
10664 |
|
|
#define SH_BLTIN_LDUA_Q 12
|
10665 |
|
|
{ 1, 10 },
|
10666 |
|
|
#define SH_BLTIN_STUA_L 13
|
10667 |
|
|
{ 0, 10, 2 },
|
10668 |
|
|
#define SH_BLTIN_STUA_Q 14
|
10669 |
|
|
{ 0, 10, 1 },
|
10670 |
|
|
#define SH_BLTIN_LDUA_L64 15
|
10671 |
|
|
{ 2, 9 },
|
10672 |
|
|
#define SH_BLTIN_LDUA_Q64 16
|
10673 |
|
|
{ 1, 9 },
|
10674 |
|
|
#define SH_BLTIN_STUA_L64 17
|
10675 |
|
|
{ 0, 9, 2 },
|
10676 |
|
|
#define SH_BLTIN_STUA_Q64 18
|
10677 |
|
|
{ 0, 9, 1 },
|
10678 |
|
|
#define SH_BLTIN_NUM_SHARED_SIGNATURES 19
|
10679 |
|
|
#define SH_BLTIN_2 19
|
10680 |
|
|
#define SH_BLTIN_SU 19
|
10681 |
|
|
{ 1, 2 },
|
10682 |
|
|
#define SH_BLTIN_3 20
|
10683 |
|
|
#define SH_BLTIN_SUS 20
|
10684 |
|
|
{ 2, 2, 1 },
|
10685 |
|
|
#define SH_BLTIN_PSSV 21
|
10686 |
|
|
{ 0, 8, 2, 2 },
|
10687 |
|
|
#define SH_BLTIN_XXUU 22
|
10688 |
|
|
#define SH_BLTIN_UUUU 22
|
10689 |
|
|
{ 1, 1, 1, 1 },
|
10690 |
|
|
#define SH_BLTIN_PV 23
|
10691 |
|
|
{ 0, 8 },
|
10692 |
|
|
};
|
10693 |
|
|
/* mcmv: operands considered unsigned. */
|
10694 |
|
|
/* mmulsum_wq, msad_ubq: result considered unsigned long long. */
|
10695 |
|
|
/* mperm: control value considered unsigned int. */
|
10696 |
|
|
/* mshalds, mshard, mshards, mshlld, mshlrd: shift count is unsigned int. */
|
10697 |
|
|
/* mshards_q: returns signed short. */
|
10698 |
|
|
/* nsb: takes long long arg, returns unsigned char. */
|
10699 |
|
|
static struct builtin_description bdesc[] =
|
10700 |
|
|
{
|
10701 |
|
|
{ CODE_FOR_absv2si2, "__builtin_absv2si2", SH_BLTIN_V2SI2, 0 },
|
10702 |
|
|
{ CODE_FOR_absv4hi2, "__builtin_absv4hi2", SH_BLTIN_V4HI2, 0 },
|
10703 |
|
|
{ CODE_FOR_addv2si3, "__builtin_addv2si3", SH_BLTIN_V2SI3, 0 },
|
10704 |
|
|
{ CODE_FOR_addv4hi3, "__builtin_addv4hi3", SH_BLTIN_V4HI3, 0 },
|
10705 |
|
|
{ CODE_FOR_ssaddv2si3,"__builtin_ssaddv2si3", SH_BLTIN_V2SI3, 0 },
|
10706 |
|
|
{ CODE_FOR_usaddv8qi3,"__builtin_usaddv8qi3", SH_BLTIN_V8QI3, 0 },
|
10707 |
|
|
{ CODE_FOR_ssaddv4hi3,"__builtin_ssaddv4hi3", SH_BLTIN_V4HI3, 0 },
|
10708 |
|
|
{ CODE_FOR_alloco_i, "__builtin_sh_media_ALLOCO", SH_BLTIN_PV, 0 },
|
10709 |
|
|
{ CODE_FOR_negcmpeqv8qi,"__builtin_sh_media_MCMPEQ_B", SH_BLTIN_V8QI3, 0 },
|
10710 |
|
|
{ CODE_FOR_negcmpeqv2si,"__builtin_sh_media_MCMPEQ_L", SH_BLTIN_V2SI3, 0 },
|
10711 |
|
|
{ CODE_FOR_negcmpeqv4hi,"__builtin_sh_media_MCMPEQ_W", SH_BLTIN_V4HI3, 0 },
|
10712 |
|
|
{ CODE_FOR_negcmpgtuv8qi,"__builtin_sh_media_MCMPGT_UB", SH_BLTIN_V8QI3, 0 },
|
10713 |
|
|
{ CODE_FOR_negcmpgtv2si,"__builtin_sh_media_MCMPGT_L", SH_BLTIN_V2SI3, 0 },
|
10714 |
|
|
{ CODE_FOR_negcmpgtv4hi,"__builtin_sh_media_MCMPGT_W", SH_BLTIN_V4HI3, 0 },
|
10715 |
|
|
{ CODE_FOR_mcmv, "__builtin_sh_media_MCMV", SH_BLTIN_UUUU, 0 },
|
10716 |
|
|
{ CODE_FOR_mcnvs_lw, "__builtin_sh_media_MCNVS_LW", SH_BLTIN_3, 0 },
|
10717 |
|
|
{ CODE_FOR_mcnvs_wb, "__builtin_sh_media_MCNVS_WB", SH_BLTIN_V4HI2V8QI, 0 },
|
10718 |
|
|
{ CODE_FOR_mcnvs_wub, "__builtin_sh_media_MCNVS_WUB", SH_BLTIN_V4HI2V8QI, 0 },
|
10719 |
|
|
{ CODE_FOR_mextr1, "__builtin_sh_media_MEXTR1", SH_BLTIN_V8QI3, 0 },
|
10720 |
|
|
{ CODE_FOR_mextr2, "__builtin_sh_media_MEXTR2", SH_BLTIN_V8QI3, 0 },
|
10721 |
|
|
{ CODE_FOR_mextr3, "__builtin_sh_media_MEXTR3", SH_BLTIN_V8QI3, 0 },
|
10722 |
|
|
{ CODE_FOR_mextr4, "__builtin_sh_media_MEXTR4", SH_BLTIN_V8QI3, 0 },
|
10723 |
|
|
{ CODE_FOR_mextr5, "__builtin_sh_media_MEXTR5", SH_BLTIN_V8QI3, 0 },
|
10724 |
|
|
{ CODE_FOR_mextr6, "__builtin_sh_media_MEXTR6", SH_BLTIN_V8QI3, 0 },
|
10725 |
|
|
{ CODE_FOR_mextr7, "__builtin_sh_media_MEXTR7", SH_BLTIN_V8QI3, 0 },
|
10726 |
|
|
{ CODE_FOR_mmacfx_wl, "__builtin_sh_media_MMACFX_WL", SH_BLTIN_MAC_HISI, 0 },
|
10727 |
|
|
{ CODE_FOR_mmacnfx_wl,"__builtin_sh_media_MMACNFX_WL", SH_BLTIN_MAC_HISI, 0 },
|
10728 |
|
|
{ CODE_FOR_mulv2si3, "__builtin_mulv2si3", SH_BLTIN_V2SI3, 0 },
|
10729 |
|
|
{ CODE_FOR_mulv4hi3, "__builtin_mulv4hi3", SH_BLTIN_V4HI3, 0 },
|
10730 |
|
|
{ CODE_FOR_mmulfx_l, "__builtin_sh_media_MMULFX_L", SH_BLTIN_V2SI3, 0 },
|
10731 |
|
|
{ CODE_FOR_mmulfx_w, "__builtin_sh_media_MMULFX_W", SH_BLTIN_V4HI3, 0 },
|
10732 |
|
|
{ CODE_FOR_mmulfxrp_w,"__builtin_sh_media_MMULFXRP_W", SH_BLTIN_V4HI3, 0 },
|
10733 |
|
|
{ CODE_FOR_mmulhi_wl, "__builtin_sh_media_MMULHI_WL", SH_BLTIN_V4HI2V2SI, 0 },
|
10734 |
|
|
{ CODE_FOR_mmullo_wl, "__builtin_sh_media_MMULLO_WL", SH_BLTIN_V4HI2V2SI, 0 },
|
10735 |
|
|
{ CODE_FOR_mmulsum_wq,"__builtin_sh_media_MMULSUM_WQ", SH_BLTIN_XXUU, 0 },
|
10736 |
|
|
{ CODE_FOR_mperm_w, "__builtin_sh_media_MPERM_W", SH_BLTIN_SH_HI, 0 },
|
10737 |
|
|
{ CODE_FOR_msad_ubq, "__builtin_sh_media_MSAD_UBQ", SH_BLTIN_XXUU, 0 },
|
10738 |
|
|
{ CODE_FOR_mshalds_l, "__builtin_sh_media_MSHALDS_L", SH_BLTIN_SH_SI, 0 },
|
10739 |
|
|
{ CODE_FOR_mshalds_w, "__builtin_sh_media_MSHALDS_W", SH_BLTIN_SH_HI, 0 },
|
10740 |
|
|
{ CODE_FOR_ashrv2si3, "__builtin_ashrv2si3", SH_BLTIN_SH_SI, 0 },
|
10741 |
|
|
{ CODE_FOR_ashrv4hi3, "__builtin_ashrv4hi3", SH_BLTIN_SH_HI, 0 },
|
10742 |
|
|
{ CODE_FOR_mshards_q, "__builtin_sh_media_MSHARDS_Q", SH_BLTIN_SUS, 0 },
|
10743 |
|
|
{ CODE_FOR_mshfhi_b, "__builtin_sh_media_MSHFHI_B", SH_BLTIN_V8QI3, 0 },
|
10744 |
|
|
{ CODE_FOR_mshfhi_l, "__builtin_sh_media_MSHFHI_L", SH_BLTIN_V2SI3, 0 },
|
10745 |
|
|
{ CODE_FOR_mshfhi_w, "__builtin_sh_media_MSHFHI_W", SH_BLTIN_V4HI3, 0 },
|
10746 |
|
|
{ CODE_FOR_mshflo_b, "__builtin_sh_media_MSHFLO_B", SH_BLTIN_V8QI3, 0 },
|
10747 |
|
|
{ CODE_FOR_mshflo_l, "__builtin_sh_media_MSHFLO_L", SH_BLTIN_V2SI3, 0 },
|
10748 |
|
|
{ CODE_FOR_mshflo_w, "__builtin_sh_media_MSHFLO_W", SH_BLTIN_V4HI3, 0 },
|
10749 |
|
|
{ CODE_FOR_ashlv2si3, "__builtin_ashlv2si3", SH_BLTIN_SH_SI, 0 },
|
10750 |
|
|
{ CODE_FOR_ashlv4hi3, "__builtin_ashlv4hi3", SH_BLTIN_SH_HI, 0 },
|
10751 |
|
|
{ CODE_FOR_lshrv2si3, "__builtin_lshrv2si3", SH_BLTIN_SH_SI, 0 },
|
10752 |
|
|
{ CODE_FOR_lshrv4hi3, "__builtin_lshrv4hi3", SH_BLTIN_SH_HI, 0 },
|
10753 |
|
|
{ CODE_FOR_subv2si3, "__builtin_subv2si3", SH_BLTIN_V2SI3, 0 },
|
10754 |
|
|
{ CODE_FOR_subv4hi3, "__builtin_subv4hi3", SH_BLTIN_V4HI3, 0 },
|
10755 |
|
|
{ CODE_FOR_sssubv2si3,"__builtin_sssubv2si3", SH_BLTIN_V2SI3, 0 },
|
10756 |
|
|
{ CODE_FOR_ussubv8qi3,"__builtin_ussubv8qi3", SH_BLTIN_V8QI3, 0 },
|
10757 |
|
|
{ CODE_FOR_sssubv4hi3,"__builtin_sssubv4hi3", SH_BLTIN_V4HI3, 0 },
|
10758 |
|
|
{ CODE_FOR_fcosa_s, "__builtin_sh_media_FCOSA_S", SH_BLTIN_SISF, 0 },
|
10759 |
|
|
{ CODE_FOR_fsina_s, "__builtin_sh_media_FSINA_S", SH_BLTIN_SISF, 0 },
|
10760 |
|
|
{ CODE_FOR_fipr, "__builtin_sh_media_FIPR_S", SH_BLTIN_3, 0 },
|
10761 |
|
|
{ CODE_FOR_ftrv, "__builtin_sh_media_FTRV_S", SH_BLTIN_3, 0 },
|
10762 |
|
|
{ CODE_FOR_mac_media, "__builtin_sh_media_FMAC_S", SH_BLTIN_3, 0 },
|
10763 |
|
|
{ CODE_FOR_sqrtdf2, "__builtin_sh_media_FSQRT_D", SH_BLTIN_2, 0 },
|
10764 |
|
|
{ CODE_FOR_sqrtsf2, "__builtin_sh_media_FSQRT_S", SH_BLTIN_2, 0 },
|
10765 |
|
|
{ CODE_FOR_fsrra_s, "__builtin_sh_media_FSRRA_S", SH_BLTIN_2, 0 },
|
10766 |
|
|
{ CODE_FOR_ldhi_l, "__builtin_sh_media_LDHI_L", SH_BLTIN_LDUA_L, 0 },
|
10767 |
|
|
{ CODE_FOR_ldhi_q, "__builtin_sh_media_LDHI_Q", SH_BLTIN_LDUA_Q, 0 },
|
10768 |
|
|
{ CODE_FOR_ldlo_l, "__builtin_sh_media_LDLO_L", SH_BLTIN_LDUA_L, 0 },
|
10769 |
|
|
{ CODE_FOR_ldlo_q, "__builtin_sh_media_LDLO_Q", SH_BLTIN_LDUA_Q, 0 },
|
10770 |
|
|
{ CODE_FOR_sthi_l, "__builtin_sh_media_STHI_L", SH_BLTIN_STUA_L, 0 },
|
10771 |
|
|
{ CODE_FOR_sthi_q, "__builtin_sh_media_STHI_Q", SH_BLTIN_STUA_Q, 0 },
|
10772 |
|
|
{ CODE_FOR_stlo_l, "__builtin_sh_media_STLO_L", SH_BLTIN_STUA_L, 0 },
|
10773 |
|
|
{ CODE_FOR_stlo_q, "__builtin_sh_media_STLO_Q", SH_BLTIN_STUA_Q, 0 },
|
10774 |
|
|
{ CODE_FOR_ldhi_l64, "__builtin_sh_media_LDHI_L", SH_BLTIN_LDUA_L64, 0 },
|
10775 |
|
|
{ CODE_FOR_ldhi_q64, "__builtin_sh_media_LDHI_Q", SH_BLTIN_LDUA_Q64, 0 },
|
10776 |
|
|
{ CODE_FOR_ldlo_l64, "__builtin_sh_media_LDLO_L", SH_BLTIN_LDUA_L64, 0 },
|
10777 |
|
|
{ CODE_FOR_ldlo_q64, "__builtin_sh_media_LDLO_Q", SH_BLTIN_LDUA_Q64, 0 },
|
10778 |
|
|
{ CODE_FOR_sthi_l64, "__builtin_sh_media_STHI_L", SH_BLTIN_STUA_L64, 0 },
|
10779 |
|
|
{ CODE_FOR_sthi_q64, "__builtin_sh_media_STHI_Q", SH_BLTIN_STUA_Q64, 0 },
|
10780 |
|
|
{ CODE_FOR_stlo_l64, "__builtin_sh_media_STLO_L", SH_BLTIN_STUA_L64, 0 },
|
10781 |
|
|
{ CODE_FOR_stlo_q64, "__builtin_sh_media_STLO_Q", SH_BLTIN_STUA_Q64, 0 },
|
10782 |
|
|
{ CODE_FOR_nsb, "__builtin_sh_media_NSB", SH_BLTIN_SU, 0 },
|
10783 |
|
|
{ CODE_FOR_byterev, "__builtin_sh_media_BYTEREV", SH_BLTIN_2, 0 },
|
10784 |
|
|
{ CODE_FOR_prefetch, "__builtin_sh_media_PREFO", SH_BLTIN_PSSV, 0 },
|
10785 |
|
|
};
|
10786 |
|
|
|
10787 |
|
|
static void
|
10788 |
|
|
sh_media_init_builtins (void)
|
10789 |
|
|
{
|
10790 |
|
|
tree shared[SH_BLTIN_NUM_SHARED_SIGNATURES];
|
10791 |
|
|
struct builtin_description *d;
|
10792 |
|
|
|
10793 |
|
|
memset (shared, 0, sizeof shared);
|
10794 |
|
|
for (d = bdesc; d - bdesc < (int) ARRAY_SIZE (bdesc); d++)
|
10795 |
|
|
{
|
10796 |
|
|
tree type, arg_type = 0;
|
10797 |
|
|
int signature = d->signature;
|
10798 |
|
|
int i;
|
10799 |
|
|
|
10800 |
|
|
if (signature < SH_BLTIN_NUM_SHARED_SIGNATURES && shared[signature])
|
10801 |
|
|
type = shared[signature];
|
10802 |
|
|
else
|
10803 |
|
|
{
|
10804 |
|
|
int has_result = signature_args[signature][0] != 0;
|
10805 |
|
|
|
10806 |
|
|
if ((signature_args[signature][1] & 8)
|
10807 |
|
|
&& (((signature_args[signature][1] & 1) && TARGET_SHMEDIA32)
|
10808 |
|
|
|| ((signature_args[signature][1] & 2) && TARGET_SHMEDIA64)))
|
10809 |
|
|
continue;
|
10810 |
|
|
if (! TARGET_FPU_ANY
|
10811 |
|
|
&& FLOAT_MODE_P (insn_data[d->icode].operand[0].mode))
|
10812 |
|
|
continue;
|
10813 |
|
|
type = void_list_node;
|
10814 |
|
|
for (i = 3; ; i--)
|
10815 |
|
|
{
|
10816 |
|
|
int arg = signature_args[signature][i];
|
10817 |
|
|
int opno = i - 1 + has_result;
|
10818 |
|
|
|
10819 |
|
|
if (arg & 8)
|
10820 |
|
|
arg_type = ptr_type_node;
|
10821 |
|
|
else if (arg)
|
10822 |
|
|
arg_type = (*lang_hooks.types.type_for_mode)
|
10823 |
|
|
(insn_data[d->icode].operand[opno].mode,
|
10824 |
|
|
(arg & 1));
|
10825 |
|
|
else if (i)
|
10826 |
|
|
continue;
|
10827 |
|
|
else
|
10828 |
|
|
arg_type = void_type_node;
|
10829 |
|
|
if (i == 0)
|
10830 |
|
|
break;
|
10831 |
|
|
type = tree_cons (NULL_TREE, arg_type, type);
|
10832 |
|
|
}
|
10833 |
|
|
type = build_function_type (arg_type, type);
|
10834 |
|
|
if (signature < SH_BLTIN_NUM_SHARED_SIGNATURES)
|
10835 |
|
|
shared[signature] = type;
|
10836 |
|
|
}
|
10837 |
|
|
d->fndecl =
|
10838 |
|
|
add_builtin_function (d->name, type, d - bdesc, BUILT_IN_MD,
|
10839 |
|
|
NULL, NULL_TREE);
|
10840 |
|
|
}
|
10841 |
|
|
}
|
10842 |
|
|
|
10843 |
|
|
/* Returns the shmedia builtin decl for CODE. */
|
10844 |
|
|
|
10845 |
|
|
static tree
|
10846 |
|
|
sh_media_builtin_decl (unsigned code, bool initialize_p ATTRIBUTE_UNUSED)
|
10847 |
|
|
{
|
10848 |
|
|
if (code >= ARRAY_SIZE (bdesc))
|
10849 |
|
|
return error_mark_node;
|
10850 |
|
|
|
10851 |
|
|
return bdesc[code].fndecl;
|
10852 |
|
|
}
|
10853 |
|
|
|
10854 |
|
|
/* Implements target hook vector_mode_supported_p. */
|
10855 |
|
|
bool
|
10856 |
|
|
sh_vector_mode_supported_p (enum machine_mode mode)
|
10857 |
|
|
{
|
10858 |
|
|
if (TARGET_FPU_ANY
|
10859 |
|
|
&& ((mode == V2SFmode)
|
10860 |
|
|
|| (mode == V4SFmode)
|
10861 |
|
|
|| (mode == V16SFmode)))
|
10862 |
|
|
return true;
|
10863 |
|
|
|
10864 |
|
|
else if (TARGET_SHMEDIA
|
10865 |
|
|
&& ((mode == V8QImode)
|
10866 |
|
|
|| (mode == V2HImode)
|
10867 |
|
|
|| (mode == V4HImode)
|
10868 |
|
|
|| (mode == V2SImode)))
|
10869 |
|
|
return true;
|
10870 |
|
|
|
10871 |
|
|
return false;
|
10872 |
|
|
}
|
10873 |
|
|
|
10874 |
|
|
/* Implements target hook dwarf_calling_convention. Return an enum
|
10875 |
|
|
of dwarf_calling_convention. */
|
10876 |
|
|
int
|
10877 |
|
|
sh_dwarf_calling_convention (const_tree func)
|
10878 |
|
|
{
|
10879 |
|
|
if (sh_attr_renesas_p (func))
|
10880 |
|
|
return DW_CC_GNU_renesas_sh;
|
10881 |
|
|
|
10882 |
|
|
return DW_CC_normal;
|
10883 |
|
|
}
|
10884 |
|
|
|
10885 |
|
|
static void
|
10886 |
|
|
sh_init_builtins (void)
|
10887 |
|
|
{
|
10888 |
|
|
if (TARGET_SHMEDIA)
|
10889 |
|
|
sh_media_init_builtins ();
|
10890 |
|
|
}
|
10891 |
|
|
|
10892 |
|
|
/* Returns the sh builtin decl for CODE. */
|
10893 |
|
|
|
10894 |
|
|
static tree
|
10895 |
|
|
sh_builtin_decl (unsigned code, bool initialize_p ATTRIBUTE_UNUSED)
|
10896 |
|
|
{
|
10897 |
|
|
if (TARGET_SHMEDIA)
|
10898 |
|
|
return sh_media_builtin_decl (code, initialize_p);
|
10899 |
|
|
|
10900 |
|
|
return error_mark_node;
|
10901 |
|
|
}
|
10902 |
|
|
|
10903 |
|
|
/* Expand an expression EXP that calls a built-in function,
|
10904 |
|
|
with result going to TARGET if that's convenient
|
10905 |
|
|
(and in mode MODE if that's convenient).
|
10906 |
|
|
SUBTARGET may be used as the target for computing one of EXP's operands.
|
10907 |
|
|
IGNORE is nonzero if the value is to be ignored. */
|
10908 |
|
|
|
10909 |
|
|
static rtx
|
10910 |
|
|
sh_expand_builtin (tree exp, rtx target, rtx subtarget ATTRIBUTE_UNUSED,
|
10911 |
|
|
enum machine_mode mode ATTRIBUTE_UNUSED, int ignore)
|
10912 |
|
|
{
|
10913 |
|
|
tree fndecl = TREE_OPERAND (CALL_EXPR_FN (exp), 0);
|
10914 |
|
|
unsigned int fcode = DECL_FUNCTION_CODE (fndecl);
|
10915 |
|
|
const struct builtin_description *d = &bdesc[fcode];
|
10916 |
|
|
enum insn_code icode = d->icode;
|
10917 |
|
|
int signature = d->signature;
|
10918 |
|
|
enum machine_mode tmode = VOIDmode;
|
10919 |
|
|
int nop = 0, i;
|
10920 |
|
|
rtx op[4];
|
10921 |
|
|
rtx pat = 0;
|
10922 |
|
|
|
10923 |
|
|
if (signature_args[signature][0])
|
10924 |
|
|
{
|
10925 |
|
|
if (ignore)
|
10926 |
|
|
return 0;
|
10927 |
|
|
|
10928 |
|
|
tmode = insn_data[icode].operand[0].mode;
|
10929 |
|
|
if (! target
|
10930 |
|
|
|| GET_MODE (target) != tmode
|
10931 |
|
|
|| ! (*insn_data[icode].operand[0].predicate) (target, tmode))
|
10932 |
|
|
target = gen_reg_rtx (tmode);
|
10933 |
|
|
op[nop++] = target;
|
10934 |
|
|
}
|
10935 |
|
|
else
|
10936 |
|
|
target = 0;
|
10937 |
|
|
|
10938 |
|
|
for (i = 1; i <= 3; i++, nop++)
|
10939 |
|
|
{
|
10940 |
|
|
tree arg;
|
10941 |
|
|
enum machine_mode opmode, argmode;
|
10942 |
|
|
tree optype;
|
10943 |
|
|
|
10944 |
|
|
if (! signature_args[signature][i])
|
10945 |
|
|
break;
|
10946 |
|
|
arg = CALL_EXPR_ARG (exp, i - 1);
|
10947 |
|
|
if (arg == error_mark_node)
|
10948 |
|
|
return const0_rtx;
|
10949 |
|
|
if (signature_args[signature][i] & 8)
|
10950 |
|
|
{
|
10951 |
|
|
opmode = ptr_mode;
|
10952 |
|
|
optype = ptr_type_node;
|
10953 |
|
|
}
|
10954 |
|
|
else
|
10955 |
|
|
{
|
10956 |
|
|
opmode = insn_data[icode].operand[nop].mode;
|
10957 |
|
|
optype = (*lang_hooks.types.type_for_mode) (opmode, 0);
|
10958 |
|
|
}
|
10959 |
|
|
argmode = TYPE_MODE (TREE_TYPE (arg));
|
10960 |
|
|
if (argmode != opmode)
|
10961 |
|
|
arg = build1 (NOP_EXPR, optype, arg);
|
10962 |
|
|
op[nop] = expand_expr (arg, NULL_RTX, opmode, EXPAND_NORMAL);
|
10963 |
|
|
if (! (*insn_data[icode].operand[nop].predicate) (op[nop], opmode))
|
10964 |
|
|
op[nop] = copy_to_mode_reg (opmode, op[nop]);
|
10965 |
|
|
}
|
10966 |
|
|
|
10967 |
|
|
switch (nop)
|
10968 |
|
|
{
|
10969 |
|
|
case 1:
|
10970 |
|
|
pat = (*insn_data[d->icode].genfun) (op[0]);
|
10971 |
|
|
break;
|
10972 |
|
|
case 2:
|
10973 |
|
|
pat = (*insn_data[d->icode].genfun) (op[0], op[1]);
|
10974 |
|
|
break;
|
10975 |
|
|
case 3:
|
10976 |
|
|
pat = (*insn_data[d->icode].genfun) (op[0], op[1], op[2]);
|
10977 |
|
|
break;
|
10978 |
|
|
case 4:
|
10979 |
|
|
pat = (*insn_data[d->icode].genfun) (op[0], op[1], op[2], op[3]);
|
10980 |
|
|
break;
|
10981 |
|
|
default:
|
10982 |
|
|
gcc_unreachable ();
|
10983 |
|
|
}
|
10984 |
|
|
if (! pat)
|
10985 |
|
|
return 0;
|
10986 |
|
|
emit_insn (pat);
|
10987 |
|
|
return target;
|
10988 |
|
|
}
|
10989 |
|
|
|
10990 |
|
|
void
|
10991 |
|
|
sh_expand_unop_v2sf (enum rtx_code code, rtx op0, rtx op1)
|
10992 |
|
|
{
|
10993 |
|
|
rtx sel0 = const0_rtx;
|
10994 |
|
|
rtx sel1 = const1_rtx;
|
10995 |
|
|
rtx (*fn) (rtx, rtx, rtx, rtx, rtx) = gen_unary_sf_op;
|
10996 |
|
|
rtx op = gen_rtx_fmt_e (code, SFmode, op1);
|
10997 |
|
|
|
10998 |
|
|
emit_insn ((*fn) (op0, op1, op, sel0, sel0));
|
10999 |
|
|
emit_insn ((*fn) (op0, op1, op, sel1, sel1));
|
11000 |
|
|
}
|
11001 |
|
|
|
11002 |
|
|
void
|
11003 |
|
|
sh_expand_binop_v2sf (enum rtx_code code, rtx op0, rtx op1, rtx op2)
|
11004 |
|
|
{
|
11005 |
|
|
rtx op = gen_rtx_fmt_ee (code, SFmode, op1, op2);
|
11006 |
|
|
|
11007 |
|
|
emit_insn (gen_binary_sf_op0 (op0, op1, op2, op));
|
11008 |
|
|
emit_insn (gen_binary_sf_op1 (op0, op1, op2, op));
|
11009 |
|
|
}
|
11010 |
|
|
|
11011 |
|
|
/* Return true if hard register REGNO can hold a value of machine-mode MODE.
|
11012 |
|
|
We can allow any mode in any general register. The special registers
|
11013 |
|
|
only allow SImode. Don't allow any mode in the PR.
|
11014 |
|
|
|
11015 |
|
|
We cannot hold DCmode values in the XD registers because alter_reg
|
11016 |
|
|
handles subregs of them incorrectly. We could work around this by
|
11017 |
|
|
spacing the XD registers like the DR registers, but this would require
|
11018 |
|
|
additional memory in every compilation to hold larger register vectors.
|
11019 |
|
|
We could hold SFmode / SCmode values in XD registers, but that
|
11020 |
|
|
would require a tertiary reload when reloading from / to memory,
|
11021 |
|
|
and a secondary reload to reload from / to general regs; that
|
11022 |
|
|
seems to be a loosing proposition.
|
11023 |
|
|
|
11024 |
|
|
We want to allow TImode FP regs so that when V4SFmode is loaded as TImode,
|
11025 |
|
|
it won't be ferried through GP registers first. */
|
11026 |
|
|
|
11027 |
|
|
bool
|
11028 |
|
|
sh_hard_regno_mode_ok (unsigned int regno, enum machine_mode mode)
|
11029 |
|
|
{
|
11030 |
|
|
if (SPECIAL_REGISTER_P (regno))
|
11031 |
|
|
return mode == SImode;
|
11032 |
|
|
|
11033 |
|
|
if (regno == FPUL_REG)
|
11034 |
|
|
return (mode == SImode || mode == SFmode);
|
11035 |
|
|
|
11036 |
|
|
if (FP_REGISTER_P (regno) && mode == SFmode)
|
11037 |
|
|
return true;
|
11038 |
|
|
|
11039 |
|
|
if (mode == V2SFmode)
|
11040 |
|
|
{
|
11041 |
|
|
if (((FP_REGISTER_P (regno) && (regno - FIRST_FP_REG) % 2 == 0)
|
11042 |
|
|
|| GENERAL_REGISTER_P (regno)))
|
11043 |
|
|
return true;
|
11044 |
|
|
else
|
11045 |
|
|
return false;
|
11046 |
|
|
}
|
11047 |
|
|
|
11048 |
|
|
if (mode == V4SFmode)
|
11049 |
|
|
{
|
11050 |
|
|
if ((FP_REGISTER_P (regno) && (regno - FIRST_FP_REG) % 4 == 0)
|
11051 |
|
|
|| GENERAL_REGISTER_P (regno))
|
11052 |
|
|
return true;
|
11053 |
|
|
else
|
11054 |
|
|
return false;
|
11055 |
|
|
}
|
11056 |
|
|
|
11057 |
|
|
if (mode == V16SFmode)
|
11058 |
|
|
{
|
11059 |
|
|
if (TARGET_SHMEDIA)
|
11060 |
|
|
{
|
11061 |
|
|
if (FP_REGISTER_P (regno) && (regno - FIRST_FP_REG) % 16 == 0)
|
11062 |
|
|
return true;
|
11063 |
|
|
else
|
11064 |
|
|
return false;
|
11065 |
|
|
}
|
11066 |
|
|
else
|
11067 |
|
|
return regno == FIRST_XD_REG;
|
11068 |
|
|
}
|
11069 |
|
|
|
11070 |
|
|
if (FP_REGISTER_P (regno))
|
11071 |
|
|
{
|
11072 |
|
|
if (mode == SFmode
|
11073 |
|
|
|| mode == SImode
|
11074 |
|
|
|| ((TARGET_SH2E || TARGET_SHMEDIA) && mode == SCmode)
|
11075 |
|
|
|| ((((TARGET_SH4 || TARGET_SH2A_DOUBLE) && mode == DFmode)
|
11076 |
|
|
|| mode == DCmode
|
11077 |
|
|
|| (TARGET_SHMEDIA
|
11078 |
|
|
&& (mode == DFmode || mode == DImode
|
11079 |
|
|
|| mode == V2SFmode || mode == TImode)))
|
11080 |
|
|
&& ((regno - FIRST_FP_REG) & 1) == 0)
|
11081 |
|
|
|| ((TARGET_SH4 || TARGET_SHMEDIA) && mode == TImode
|
11082 |
|
|
&& ((regno - FIRST_FP_REG) & 3) == 0))
|
11083 |
|
|
return true;
|
11084 |
|
|
else
|
11085 |
|
|
return false;
|
11086 |
|
|
}
|
11087 |
|
|
|
11088 |
|
|
if (XD_REGISTER_P (regno))
|
11089 |
|
|
return mode == DFmode;
|
11090 |
|
|
|
11091 |
|
|
if (TARGET_REGISTER_P (regno))
|
11092 |
|
|
return (mode == DImode || mode == SImode || mode == PDImode);
|
11093 |
|
|
|
11094 |
|
|
if (regno == PR_REG)
|
11095 |
|
|
return mode == SImode;
|
11096 |
|
|
|
11097 |
|
|
if (regno == FPSCR_REG)
|
11098 |
|
|
return mode == PSImode;
|
11099 |
|
|
|
11100 |
|
|
/* FIXME. This works around PR target/37633 for -O0. */
|
11101 |
|
|
if (!optimize && TARGET_SHMEDIA32 && GET_MODE_SIZE (mode) > 4)
|
11102 |
|
|
{
|
11103 |
|
|
unsigned int n = GET_MODE_SIZE (mode) / 8;
|
11104 |
|
|
|
11105 |
|
|
if (regno >= FIRST_GENERAL_REG + 10 - n + 1
|
11106 |
|
|
&& regno <= FIRST_GENERAL_REG + 14)
|
11107 |
|
|
return false;
|
11108 |
|
|
}
|
11109 |
|
|
|
11110 |
|
|
return true;
|
11111 |
|
|
}
|
11112 |
|
|
|
11113 |
|
|
/* Return the class of registers for which a mode change from FROM to TO
|
11114 |
|
|
is invalid. */
|
11115 |
|
|
bool
|
11116 |
|
|
sh_cannot_change_mode_class (enum machine_mode from, enum machine_mode to,
|
11117 |
|
|
enum reg_class rclass)
|
11118 |
|
|
{
|
11119 |
|
|
/* We want to enable the use of SUBREGs as a means to
|
11120 |
|
|
VEC_SELECT a single element of a vector. */
|
11121 |
|
|
if (to == SFmode && VECTOR_MODE_P (from) && GET_MODE_INNER (from) == SFmode)
|
11122 |
|
|
return (reg_classes_intersect_p (GENERAL_REGS, rclass));
|
11123 |
|
|
|
11124 |
|
|
if (GET_MODE_SIZE (from) != GET_MODE_SIZE (to))
|
11125 |
|
|
{
|
11126 |
|
|
if (TARGET_LITTLE_ENDIAN)
|
11127 |
|
|
{
|
11128 |
|
|
if (GET_MODE_SIZE (to) < 8 || GET_MODE_SIZE (from) < 8)
|
11129 |
|
|
return reg_classes_intersect_p (DF_REGS, rclass);
|
11130 |
|
|
}
|
11131 |
|
|
else
|
11132 |
|
|
{
|
11133 |
|
|
if (GET_MODE_SIZE (from) < 8)
|
11134 |
|
|
return reg_classes_intersect_p (DF_HI_REGS, rclass);
|
11135 |
|
|
}
|
11136 |
|
|
}
|
11137 |
|
|
return 0;
|
11138 |
|
|
}
|
11139 |
|
|
|
11140 |
|
|
|
11141 |
|
|
/* If ADDRESS refers to a CODE_LABEL, add NUSES to the number of times
|
11142 |
|
|
that label is used. */
|
11143 |
|
|
|
11144 |
|
|
void
|
11145 |
|
|
sh_mark_label (rtx address, int nuses)
|
11146 |
|
|
{
|
11147 |
|
|
if (GOTOFF_P (address))
|
11148 |
|
|
{
|
11149 |
|
|
/* Extract the label or symbol. */
|
11150 |
|
|
address = XEXP (address, 0);
|
11151 |
|
|
if (GET_CODE (address) == PLUS)
|
11152 |
|
|
address = XEXP (address, 0);
|
11153 |
|
|
address = XVECEXP (address, 0, 0);
|
11154 |
|
|
}
|
11155 |
|
|
if (GET_CODE (address) == LABEL_REF
|
11156 |
|
|
&& LABEL_P (XEXP (address, 0)))
|
11157 |
|
|
LABEL_NUSES (XEXP (address, 0)) += nuses;
|
11158 |
|
|
}
|
11159 |
|
|
|
11160 |
|
|
/* Compute extra cost of moving data between one register class
|
11161 |
|
|
and another. */
|
11162 |
|
|
|
11163 |
|
|
/* If SECONDARY*_RELOAD_CLASS says something about the src/dst pair, regclass
|
11164 |
|
|
uses this information. Hence, the general register <-> floating point
|
11165 |
|
|
register information here is not used for SFmode. */
|
11166 |
|
|
|
11167 |
|
|
int
|
11168 |
|
|
sh_register_move_cost (enum machine_mode mode,
|
11169 |
|
|
enum reg_class srcclass, enum reg_class dstclass)
|
11170 |
|
|
{
|
11171 |
|
|
if (dstclass == T_REGS || dstclass == PR_REGS)
|
11172 |
|
|
return 10;
|
11173 |
|
|
|
11174 |
|
|
if (dstclass == MAC_REGS && srcclass == MAC_REGS)
|
11175 |
|
|
return 4;
|
11176 |
|
|
|
11177 |
|
|
if (mode == SImode && ! TARGET_SHMEDIA && TARGET_FMOVD
|
11178 |
|
|
&& REGCLASS_HAS_FP_REG (srcclass)
|
11179 |
|
|
&& REGCLASS_HAS_FP_REG (dstclass))
|
11180 |
|
|
return 4;
|
11181 |
|
|
|
11182 |
|
|
if (REGCLASS_HAS_FP_REG (dstclass) && srcclass == T_REGS)
|
11183 |
|
|
return ((TARGET_HARD_SH4 && !optimize_size) ? 10 : 7);
|
11184 |
|
|
|
11185 |
|
|
if ((REGCLASS_HAS_FP_REG (dstclass) && srcclass == MAC_REGS)
|
11186 |
|
|
|| (dstclass == MAC_REGS && REGCLASS_HAS_FP_REG (srcclass)))
|
11187 |
|
|
return 9;
|
11188 |
|
|
|
11189 |
|
|
if ((REGCLASS_HAS_FP_REG (dstclass)
|
11190 |
|
|
&& REGCLASS_HAS_GENERAL_REG (srcclass))
|
11191 |
|
|
|| (REGCLASS_HAS_GENERAL_REG (dstclass)
|
11192 |
|
|
&& REGCLASS_HAS_FP_REG (srcclass)))
|
11193 |
|
|
return ((TARGET_SHMEDIA ? 4 : TARGET_FMOVD ? 8 : 12)
|
11194 |
|
|
* ((GET_MODE_SIZE (mode) + 7) / 8U));
|
11195 |
|
|
|
11196 |
|
|
if ((dstclass == FPUL_REGS
|
11197 |
|
|
&& REGCLASS_HAS_GENERAL_REG (srcclass))
|
11198 |
|
|
|| (srcclass == FPUL_REGS
|
11199 |
|
|
&& REGCLASS_HAS_GENERAL_REG (dstclass)))
|
11200 |
|
|
return 5;
|
11201 |
|
|
|
11202 |
|
|
if ((dstclass == FPUL_REGS
|
11203 |
|
|
&& (srcclass == PR_REGS || srcclass == MAC_REGS || srcclass == T_REGS))
|
11204 |
|
|
|| (srcclass == FPUL_REGS
|
11205 |
|
|
&& (dstclass == PR_REGS || dstclass == MAC_REGS)))
|
11206 |
|
|
return 7;
|
11207 |
|
|
|
11208 |
|
|
if ((srcclass == TARGET_REGS && ! REGCLASS_HAS_GENERAL_REG (dstclass))
|
11209 |
|
|
|| ((dstclass) == TARGET_REGS && ! REGCLASS_HAS_GENERAL_REG (srcclass)))
|
11210 |
|
|
return 20;
|
11211 |
|
|
|
11212 |
|
|
/* ??? ptabs faults on (value & 0x3) == 0x3 */
|
11213 |
|
|
if (TARGET_SHMEDIA
|
11214 |
|
|
&& ((srcclass) == TARGET_REGS || (srcclass) == SIBCALL_REGS))
|
11215 |
|
|
{
|
11216 |
|
|
if (sh_gettrcost >= 0)
|
11217 |
|
|
return sh_gettrcost;
|
11218 |
|
|
else if (!TARGET_PT_FIXED)
|
11219 |
|
|
return 100;
|
11220 |
|
|
}
|
11221 |
|
|
|
11222 |
|
|
if ((srcclass == FPSCR_REGS && ! REGCLASS_HAS_GENERAL_REG (dstclass))
|
11223 |
|
|
|| (dstclass == FPSCR_REGS && ! REGCLASS_HAS_GENERAL_REG (srcclass)))
|
11224 |
|
|
return 4;
|
11225 |
|
|
|
11226 |
|
|
if (TARGET_SHMEDIA
|
11227 |
|
|
|| (TARGET_FMOVD
|
11228 |
|
|
&& ! REGCLASS_HAS_GENERAL_REG (srcclass)
|
11229 |
|
|
&& ! REGCLASS_HAS_GENERAL_REG (dstclass)))
|
11230 |
|
|
return 2 * ((GET_MODE_SIZE (mode) + 7) / 8U);
|
11231 |
|
|
|
11232 |
|
|
return 2 * ((GET_MODE_SIZE (mode) + 3) / 4U);
|
11233 |
|
|
}
|
11234 |
|
|
|
11235 |
|
|
static rtx emit_load_ptr (rtx, rtx);
|
11236 |
|
|
|
11237 |
|
|
static rtx
|
11238 |
|
|
emit_load_ptr (rtx reg, rtx addr)
|
11239 |
|
|
{
|
11240 |
|
|
rtx mem = gen_const_mem (ptr_mode, addr);
|
11241 |
|
|
|
11242 |
|
|
if (Pmode != ptr_mode)
|
11243 |
|
|
mem = gen_rtx_SIGN_EXTEND (Pmode, mem);
|
11244 |
|
|
return emit_move_insn (reg, mem);
|
11245 |
|
|
}
|
11246 |
|
|
|
11247 |
|
|
static void
|
11248 |
|
|
sh_output_mi_thunk (FILE *file, tree thunk_fndecl ATTRIBUTE_UNUSED,
|
11249 |
|
|
HOST_WIDE_INT delta, HOST_WIDE_INT vcall_offset,
|
11250 |
|
|
tree function)
|
11251 |
|
|
{
|
11252 |
|
|
CUMULATIVE_ARGS cum;
|
11253 |
|
|
int structure_value_byref = 0;
|
11254 |
|
|
rtx this_rtx, this_value, sibcall, insns, funexp;
|
11255 |
|
|
tree funtype = TREE_TYPE (function);
|
11256 |
|
|
int simple_add = CONST_OK_FOR_ADD (delta);
|
11257 |
|
|
int did_load = 0;
|
11258 |
|
|
rtx scratch0, scratch1, scratch2;
|
11259 |
|
|
unsigned i;
|
11260 |
|
|
|
11261 |
|
|
reload_completed = 1;
|
11262 |
|
|
epilogue_completed = 1;
|
11263 |
|
|
current_function_uses_only_leaf_regs = 1;
|
11264 |
|
|
|
11265 |
|
|
emit_note (NOTE_INSN_PROLOGUE_END);
|
11266 |
|
|
|
11267 |
|
|
/* Find the "this" pointer. We have such a wide range of ABIs for the
|
11268 |
|
|
SH that it's best to do this completely machine independently.
|
11269 |
|
|
"this" is passed as first argument, unless a structure return pointer
|
11270 |
|
|
comes first, in which case "this" comes second. */
|
11271 |
|
|
INIT_CUMULATIVE_ARGS (cum, funtype, NULL_RTX, 0, 1);
|
11272 |
|
|
#ifndef PCC_STATIC_STRUCT_RETURN
|
11273 |
|
|
if (aggregate_value_p (TREE_TYPE (TREE_TYPE (function)), function))
|
11274 |
|
|
structure_value_byref = 1;
|
11275 |
|
|
#endif /* not PCC_STATIC_STRUCT_RETURN */
|
11276 |
|
|
if (structure_value_byref && sh_struct_value_rtx (function, 0) == 0)
|
11277 |
|
|
{
|
11278 |
|
|
tree ptype = build_pointer_type (TREE_TYPE (funtype));
|
11279 |
|
|
|
11280 |
|
|
FUNCTION_ARG_ADVANCE (cum, Pmode, ptype, 1);
|
11281 |
|
|
}
|
11282 |
|
|
this_rtx = FUNCTION_ARG (cum, Pmode, ptr_type_node, 1);
|
11283 |
|
|
|
11284 |
|
|
/* For SHcompact, we only have r0 for a scratch register: r1 is the
|
11285 |
|
|
static chain pointer (even if you can't have nested virtual functions
|
11286 |
|
|
right now, someone might implement them sometime), and the rest of the
|
11287 |
|
|
registers are used for argument passing, are callee-saved, or reserved. */
|
11288 |
|
|
/* We need to check call_used_regs / fixed_regs in case -fcall_saved-reg /
|
11289 |
|
|
-ffixed-reg has been used. */
|
11290 |
|
|
if (! call_used_regs[0] || fixed_regs[0])
|
11291 |
|
|
error ("r0 needs to be available as a call-clobbered register");
|
11292 |
|
|
scratch0 = scratch1 = scratch2 = gen_rtx_REG (Pmode, 0);
|
11293 |
|
|
if (! TARGET_SH5)
|
11294 |
|
|
{
|
11295 |
|
|
if (call_used_regs[1] && ! fixed_regs[1])
|
11296 |
|
|
scratch1 = gen_rtx_REG (ptr_mode, 1);
|
11297 |
|
|
/* N.B., if not TARGET_HITACHI, register 2 is used to pass the pointer
|
11298 |
|
|
pointing where to return struct values. */
|
11299 |
|
|
if (call_used_regs[3] && ! fixed_regs[3])
|
11300 |
|
|
scratch2 = gen_rtx_REG (Pmode, 3);
|
11301 |
|
|
}
|
11302 |
|
|
else if (TARGET_SHMEDIA)
|
11303 |
|
|
{
|
11304 |
|
|
for (i = FIRST_GENERAL_REG; i <= LAST_GENERAL_REG; i++)
|
11305 |
|
|
if (i != REGNO (scratch0) &&
|
11306 |
|
|
call_used_regs[i] && ! fixed_regs[i] && ! FUNCTION_ARG_REGNO_P (i))
|
11307 |
|
|
{
|
11308 |
|
|
scratch1 = gen_rtx_REG (ptr_mode, i);
|
11309 |
|
|
break;
|
11310 |
|
|
}
|
11311 |
|
|
if (scratch1 == scratch0)
|
11312 |
|
|
error ("Need a second call-clobbered general purpose register");
|
11313 |
|
|
for (i = FIRST_TARGET_REG; i <= LAST_TARGET_REG; i++)
|
11314 |
|
|
if (call_used_regs[i] && ! fixed_regs[i])
|
11315 |
|
|
{
|
11316 |
|
|
scratch2 = gen_rtx_REG (Pmode, i);
|
11317 |
|
|
break;
|
11318 |
|
|
}
|
11319 |
|
|
if (scratch2 == scratch0)
|
11320 |
|
|
error ("Need a call-clobbered target register");
|
11321 |
|
|
}
|
11322 |
|
|
|
11323 |
|
|
this_value = plus_constant (this_rtx, delta);
|
11324 |
|
|
if (vcall_offset
|
11325 |
|
|
&& (simple_add || scratch0 != scratch1)
|
11326 |
|
|
&& strict_memory_address_p (ptr_mode, this_value))
|
11327 |
|
|
{
|
11328 |
|
|
emit_load_ptr (scratch0, this_value);
|
11329 |
|
|
did_load = 1;
|
11330 |
|
|
}
|
11331 |
|
|
|
11332 |
|
|
if (!delta)
|
11333 |
|
|
; /* Do nothing. */
|
11334 |
|
|
else if (simple_add)
|
11335 |
|
|
emit_move_insn (this_rtx, this_value);
|
11336 |
|
|
else
|
11337 |
|
|
{
|
11338 |
|
|
emit_move_insn (scratch1, GEN_INT (delta));
|
11339 |
|
|
emit_insn (gen_add2_insn (this_rtx, scratch1));
|
11340 |
|
|
}
|
11341 |
|
|
|
11342 |
|
|
if (vcall_offset)
|
11343 |
|
|
{
|
11344 |
|
|
rtx offset_addr;
|
11345 |
|
|
|
11346 |
|
|
if (!did_load)
|
11347 |
|
|
emit_load_ptr (scratch0, this_rtx);
|
11348 |
|
|
|
11349 |
|
|
offset_addr = plus_constant (scratch0, vcall_offset);
|
11350 |
|
|
if (strict_memory_address_p (ptr_mode, offset_addr))
|
11351 |
|
|
; /* Do nothing. */
|
11352 |
|
|
else if (! TARGET_SH5 && scratch0 != scratch1)
|
11353 |
|
|
{
|
11354 |
|
|
/* scratch0 != scratch1, and we have indexed loads. Get better
|
11355 |
|
|
schedule by loading the offset into r1 and using an indexed
|
11356 |
|
|
load - then the load of r1 can issue before the load from
|
11357 |
|
|
(this_rtx + delta) finishes. */
|
11358 |
|
|
emit_move_insn (scratch1, GEN_INT (vcall_offset));
|
11359 |
|
|
offset_addr = gen_rtx_PLUS (Pmode, scratch0, scratch1);
|
11360 |
|
|
}
|
11361 |
|
|
else if (CONST_OK_FOR_ADD (vcall_offset))
|
11362 |
|
|
{
|
11363 |
|
|
emit_insn (gen_add2_insn (scratch0, GEN_INT (vcall_offset)));
|
11364 |
|
|
offset_addr = scratch0;
|
11365 |
|
|
}
|
11366 |
|
|
else if (scratch0 != scratch1)
|
11367 |
|
|
{
|
11368 |
|
|
emit_move_insn (scratch1, GEN_INT (vcall_offset));
|
11369 |
|
|
emit_insn (gen_add2_insn (scratch0, scratch1));
|
11370 |
|
|
offset_addr = scratch0;
|
11371 |
|
|
}
|
11372 |
|
|
else
|
11373 |
|
|
gcc_unreachable (); /* FIXME */
|
11374 |
|
|
emit_load_ptr (scratch0, offset_addr);
|
11375 |
|
|
|
11376 |
|
|
if (Pmode != ptr_mode)
|
11377 |
|
|
scratch0 = gen_rtx_TRUNCATE (ptr_mode, scratch0);
|
11378 |
|
|
emit_insn (gen_add2_insn (this_rtx, scratch0));
|
11379 |
|
|
}
|
11380 |
|
|
|
11381 |
|
|
/* Generate a tail call to the target function. */
|
11382 |
|
|
if (! TREE_USED (function))
|
11383 |
|
|
{
|
11384 |
|
|
assemble_external (function);
|
11385 |
|
|
TREE_USED (function) = 1;
|
11386 |
|
|
}
|
11387 |
|
|
funexp = XEXP (DECL_RTL (function), 0);
|
11388 |
|
|
/* If the function is overridden, so is the thunk, hence we don't
|
11389 |
|
|
need GOT addressing even if this is a public symbol. */
|
11390 |
|
|
#if 0
|
11391 |
|
|
if (TARGET_SH1 && ! flag_weak)
|
11392 |
|
|
sibcall = gen_sibcalli_thunk (funexp, const0_rtx);
|
11393 |
|
|
else
|
11394 |
|
|
#endif
|
11395 |
|
|
if (TARGET_SH2 && flag_pic)
|
11396 |
|
|
{
|
11397 |
|
|
sibcall = gen_sibcall_pcrel (funexp, const0_rtx);
|
11398 |
|
|
XEXP (XVECEXP (sibcall, 0, 2), 0) = scratch2;
|
11399 |
|
|
}
|
11400 |
|
|
else
|
11401 |
|
|
{
|
11402 |
|
|
if (TARGET_SHMEDIA && flag_pic)
|
11403 |
|
|
{
|
11404 |
|
|
funexp = gen_sym2PIC (funexp);
|
11405 |
|
|
PUT_MODE (funexp, Pmode);
|
11406 |
|
|
}
|
11407 |
|
|
emit_move_insn (scratch2, funexp);
|
11408 |
|
|
funexp = gen_rtx_MEM (FUNCTION_MODE, scratch2);
|
11409 |
|
|
sibcall = gen_sibcall (funexp, const0_rtx, NULL_RTX);
|
11410 |
|
|
}
|
11411 |
|
|
sibcall = emit_call_insn (sibcall);
|
11412 |
|
|
SIBLING_CALL_P (sibcall) = 1;
|
11413 |
|
|
use_reg (&CALL_INSN_FUNCTION_USAGE (sibcall), this_rtx);
|
11414 |
|
|
emit_barrier ();
|
11415 |
|
|
|
11416 |
|
|
/* Run just enough of rest_of_compilation to do scheduling and get
|
11417 |
|
|
the insns emitted. Note that use_thunk calls
|
11418 |
|
|
assemble_start_function and assemble_end_function. */
|
11419 |
|
|
|
11420 |
|
|
insn_locators_alloc ();
|
11421 |
|
|
insns = get_insns ();
|
11422 |
|
|
|
11423 |
|
|
if (optimize > 0)
|
11424 |
|
|
{
|
11425 |
|
|
if (! cfun->cfg)
|
11426 |
|
|
init_flow (cfun);
|
11427 |
|
|
split_all_insns_noflow ();
|
11428 |
|
|
}
|
11429 |
|
|
|
11430 |
|
|
sh_reorg ();
|
11431 |
|
|
|
11432 |
|
|
if (optimize > 0 && flag_delayed_branch)
|
11433 |
|
|
dbr_schedule (insns);
|
11434 |
|
|
|
11435 |
|
|
shorten_branches (insns);
|
11436 |
|
|
final_start_function (insns, file, 1);
|
11437 |
|
|
final (insns, file, 1);
|
11438 |
|
|
final_end_function ();
|
11439 |
|
|
|
11440 |
|
|
reload_completed = 0;
|
11441 |
|
|
epilogue_completed = 0;
|
11442 |
|
|
}
|
11443 |
|
|
|
11444 |
|
|
rtx
|
11445 |
|
|
function_symbol (rtx target, const char *name, enum sh_function_kind kind)
|
11446 |
|
|
{
|
11447 |
|
|
rtx sym;
|
11448 |
|
|
|
11449 |
|
|
/* If this is not an ordinary function, the name usually comes from a
|
11450 |
|
|
string literal or an sprintf buffer. Make sure we use the same
|
11451 |
|
|
string consistently, so that cse will be able to unify address loads. */
|
11452 |
|
|
if (kind != FUNCTION_ORDINARY)
|
11453 |
|
|
name = IDENTIFIER_POINTER (get_identifier (name));
|
11454 |
|
|
sym = gen_rtx_SYMBOL_REF (Pmode, name);
|
11455 |
|
|
SYMBOL_REF_FLAGS (sym) = SYMBOL_FLAG_FUNCTION;
|
11456 |
|
|
if (flag_pic)
|
11457 |
|
|
switch (kind)
|
11458 |
|
|
{
|
11459 |
|
|
case FUNCTION_ORDINARY:
|
11460 |
|
|
break;
|
11461 |
|
|
case SFUNC_GOT:
|
11462 |
|
|
{
|
11463 |
|
|
rtx reg = target ? target : gen_reg_rtx (Pmode);
|
11464 |
|
|
|
11465 |
|
|
emit_insn (gen_symGOT2reg (reg, sym));
|
11466 |
|
|
sym = reg;
|
11467 |
|
|
break;
|
11468 |
|
|
}
|
11469 |
|
|
case SFUNC_STATIC:
|
11470 |
|
|
{
|
11471 |
|
|
/* ??? To allow cse to work, we use GOTOFF relocations.
|
11472 |
|
|
we could add combiner patterns to transform this into
|
11473 |
|
|
straight pc-relative calls with sym2PIC / bsrf when
|
11474 |
|
|
label load and function call are still 1:1 and in the
|
11475 |
|
|
same basic block during combine. */
|
11476 |
|
|
rtx reg = target ? target : gen_reg_rtx (Pmode);
|
11477 |
|
|
|
11478 |
|
|
emit_insn (gen_symGOTOFF2reg (reg, sym));
|
11479 |
|
|
sym = reg;
|
11480 |
|
|
break;
|
11481 |
|
|
}
|
11482 |
|
|
}
|
11483 |
|
|
if (target && sym != target)
|
11484 |
|
|
{
|
11485 |
|
|
emit_move_insn (target, sym);
|
11486 |
|
|
return target;
|
11487 |
|
|
}
|
11488 |
|
|
return sym;
|
11489 |
|
|
}
|
11490 |
|
|
|
11491 |
|
|
/* Find the number of a general purpose register in S. */
|
11492 |
|
|
static int
|
11493 |
|
|
scavenge_reg (HARD_REG_SET *s)
|
11494 |
|
|
{
|
11495 |
|
|
int r;
|
11496 |
|
|
for (r = FIRST_GENERAL_REG; r <= LAST_GENERAL_REG; r++)
|
11497 |
|
|
if (TEST_HARD_REG_BIT (*s, r))
|
11498 |
|
|
return r;
|
11499 |
|
|
return -1;
|
11500 |
|
|
}
|
11501 |
|
|
|
11502 |
|
|
rtx
|
11503 |
|
|
sh_get_pr_initial_val (void)
|
11504 |
|
|
{
|
11505 |
|
|
rtx val;
|
11506 |
|
|
|
11507 |
|
|
/* ??? Unfortunately, get_hard_reg_initial_val doesn't always work for the
|
11508 |
|
|
PR register on SHcompact, because it might be clobbered by the prologue.
|
11509 |
|
|
We check first if that is known to be the case. */
|
11510 |
|
|
if (TARGET_SHCOMPACT
|
11511 |
|
|
&& ((crtl->args.info.call_cookie
|
11512 |
|
|
& ~ CALL_COOKIE_RET_TRAMP (1))
|
11513 |
|
|
|| crtl->saves_all_registers))
|
11514 |
|
|
return gen_frame_mem (SImode, return_address_pointer_rtx);
|
11515 |
|
|
|
11516 |
|
|
/* If we haven't finished rtl generation, there might be a nonlocal label
|
11517 |
|
|
that we haven't seen yet.
|
11518 |
|
|
??? get_hard_reg_initial_val fails if it is called after register
|
11519 |
|
|
allocation has started, unless it has been called before for the
|
11520 |
|
|
same register. And even then, we end in trouble if we didn't use
|
11521 |
|
|
the register in the same basic block before. So call
|
11522 |
|
|
get_hard_reg_initial_val now and wrap it in an unspec if we might
|
11523 |
|
|
need to replace it. */
|
11524 |
|
|
/* ??? We also must do this for TARGET_SH1 in general, because otherwise
|
11525 |
|
|
combine can put the pseudo returned by get_hard_reg_initial_val into
|
11526 |
|
|
instructions that need a general purpose registers, which will fail to
|
11527 |
|
|
be recognized when the pseudo becomes allocated to PR. */
|
11528 |
|
|
val
|
11529 |
|
|
= get_hard_reg_initial_val (Pmode, TARGET_SHMEDIA ? PR_MEDIA_REG : PR_REG);
|
11530 |
|
|
if (TARGET_SH1)
|
11531 |
|
|
return gen_rtx_UNSPEC (SImode, gen_rtvec (1, val), UNSPEC_RA);
|
11532 |
|
|
return val;
|
11533 |
|
|
}
|
11534 |
|
|
|
11535 |
|
|
int
|
11536 |
|
|
sh_expand_t_scc (rtx operands[])
|
11537 |
|
|
{
|
11538 |
|
|
enum rtx_code code = GET_CODE (operands[1]);
|
11539 |
|
|
rtx target = operands[0];
|
11540 |
|
|
rtx op0 = operands[2];
|
11541 |
|
|
rtx op1 = operands[3];
|
11542 |
|
|
rtx result = target;
|
11543 |
|
|
HOST_WIDE_INT val;
|
11544 |
|
|
|
11545 |
|
|
if (!REG_P (op0) || REGNO (op0) != T_REG
|
11546 |
|
|
|| !CONST_INT_P (op1))
|
11547 |
|
|
return 0;
|
11548 |
|
|
if (!REG_P (result))
|
11549 |
|
|
result = gen_reg_rtx (SImode);
|
11550 |
|
|
val = INTVAL (op1);
|
11551 |
|
|
if ((code == EQ && val == 1) || (code == NE && val == 0))
|
11552 |
|
|
emit_insn (gen_movt (result));
|
11553 |
|
|
else if (TARGET_SH2A && ((code == EQ && val == 0)
|
11554 |
|
|
|| (code == NE && val == 1)))
|
11555 |
|
|
emit_insn (gen_xorsi3_movrt (result));
|
11556 |
|
|
else if ((code == EQ && val == 0) || (code == NE && val == 1))
|
11557 |
|
|
{
|
11558 |
|
|
emit_clobber (result);
|
11559 |
|
|
emit_insn (gen_subc (result, result, result));
|
11560 |
|
|
emit_insn (gen_addsi3 (result, result, const1_rtx));
|
11561 |
|
|
}
|
11562 |
|
|
else if (code == EQ || code == NE)
|
11563 |
|
|
emit_insn (gen_move_insn (result, GEN_INT (code == NE)));
|
11564 |
|
|
else
|
11565 |
|
|
return 0;
|
11566 |
|
|
if (result != target)
|
11567 |
|
|
emit_move_insn (target, result);
|
11568 |
|
|
return 1;
|
11569 |
|
|
}
|
11570 |
|
|
|
11571 |
|
|
/* INSN is an sfunc; return the rtx that describes the address used. */
|
11572 |
|
|
static rtx
|
11573 |
|
|
extract_sfunc_addr (rtx insn)
|
11574 |
|
|
{
|
11575 |
|
|
rtx pattern, part = NULL_RTX;
|
11576 |
|
|
int len, i;
|
11577 |
|
|
|
11578 |
|
|
pattern = PATTERN (insn);
|
11579 |
|
|
len = XVECLEN (pattern, 0);
|
11580 |
|
|
for (i = 0; i < len; i++)
|
11581 |
|
|
{
|
11582 |
|
|
part = XVECEXP (pattern, 0, i);
|
11583 |
|
|
if (GET_CODE (part) == USE && GET_MODE (XEXP (part, 0)) == Pmode
|
11584 |
|
|
&& GENERAL_REGISTER_P (true_regnum (XEXP (part, 0))))
|
11585 |
|
|
return XEXP (part, 0);
|
11586 |
|
|
}
|
11587 |
|
|
gcc_assert (GET_CODE (XVECEXP (pattern, 0, 0)) == UNSPEC_VOLATILE);
|
11588 |
|
|
return XVECEXP (XVECEXP (pattern, 0, 0), 0, 1);
|
11589 |
|
|
}
|
11590 |
|
|
|
11591 |
|
|
/* Verify that the register in use_sfunc_addr still agrees with the address
|
11592 |
|
|
used in the sfunc. This prevents fill_slots_from_thread from changing
|
11593 |
|
|
use_sfunc_addr.
|
11594 |
|
|
INSN is the use_sfunc_addr instruction, and REG is the register it
|
11595 |
|
|
guards. */
|
11596 |
|
|
int
|
11597 |
|
|
check_use_sfunc_addr (rtx insn, rtx reg)
|
11598 |
|
|
{
|
11599 |
|
|
/* Search for the sfunc. It should really come right after INSN. */
|
11600 |
|
|
while ((insn = NEXT_INSN (insn)))
|
11601 |
|
|
{
|
11602 |
|
|
if (LABEL_P (insn) || JUMP_P (insn))
|
11603 |
|
|
break;
|
11604 |
|
|
if (! INSN_P (insn))
|
11605 |
|
|
continue;
|
11606 |
|
|
|
11607 |
|
|
if (GET_CODE (PATTERN (insn)) == SEQUENCE)
|
11608 |
|
|
insn = XVECEXP (PATTERN (insn), 0, 0);
|
11609 |
|
|
if (GET_CODE (PATTERN (insn)) != PARALLEL
|
11610 |
|
|
|| get_attr_type (insn) != TYPE_SFUNC)
|
11611 |
|
|
continue;
|
11612 |
|
|
return rtx_equal_p (extract_sfunc_addr (insn), reg);
|
11613 |
|
|
}
|
11614 |
|
|
gcc_unreachable ();
|
11615 |
|
|
}
|
11616 |
|
|
|
11617 |
|
|
/* This function returns a constant rtx that represents pi / 2**15 in
|
11618 |
|
|
SFmode. it's used to scale SFmode angles, in radians, to a
|
11619 |
|
|
fixed-point signed 16.16-bit fraction of a full circle, i.e., 2*pi
|
11620 |
|
|
maps to 0x10000). */
|
11621 |
|
|
|
11622 |
|
|
static GTY(()) rtx sh_fsca_sf2int_rtx;
|
11623 |
|
|
|
11624 |
|
|
rtx
|
11625 |
|
|
sh_fsca_sf2int (void)
|
11626 |
|
|
{
|
11627 |
|
|
if (! sh_fsca_sf2int_rtx)
|
11628 |
|
|
{
|
11629 |
|
|
REAL_VALUE_TYPE rv;
|
11630 |
|
|
|
11631 |
|
|
real_from_string (&rv, "10430.378350470453");
|
11632 |
|
|
sh_fsca_sf2int_rtx = const_double_from_real_value (rv, SFmode);
|
11633 |
|
|
}
|
11634 |
|
|
|
11635 |
|
|
return sh_fsca_sf2int_rtx;
|
11636 |
|
|
}
|
11637 |
|
|
|
11638 |
|
|
/* This function returns a constant rtx that represents pi / 2**15 in
|
11639 |
|
|
DFmode. it's used to scale DFmode angles, in radians, to a
|
11640 |
|
|
fixed-point signed 16.16-bit fraction of a full circle, i.e., 2*pi
|
11641 |
|
|
maps to 0x10000). */
|
11642 |
|
|
|
11643 |
|
|
static GTY(()) rtx sh_fsca_df2int_rtx;
|
11644 |
|
|
|
11645 |
|
|
rtx
|
11646 |
|
|
sh_fsca_df2int (void)
|
11647 |
|
|
{
|
11648 |
|
|
if (! sh_fsca_df2int_rtx)
|
11649 |
|
|
{
|
11650 |
|
|
REAL_VALUE_TYPE rv;
|
11651 |
|
|
|
11652 |
|
|
real_from_string (&rv, "10430.378350470453");
|
11653 |
|
|
sh_fsca_df2int_rtx = const_double_from_real_value (rv, DFmode);
|
11654 |
|
|
}
|
11655 |
|
|
|
11656 |
|
|
return sh_fsca_df2int_rtx;
|
11657 |
|
|
}
|
11658 |
|
|
|
11659 |
|
|
/* This function returns a constant rtx that represents 2**15 / pi in
|
11660 |
|
|
SFmode. it's used to scale a fixed-point signed 16.16-bit fraction
|
11661 |
|
|
of a full circle back to a SFmode value, i.e., 0x10000 maps to
|
11662 |
|
|
2*pi). */
|
11663 |
|
|
|
11664 |
|
|
static GTY(()) rtx sh_fsca_int2sf_rtx;
|
11665 |
|
|
|
11666 |
|
|
rtx
|
11667 |
|
|
sh_fsca_int2sf (void)
|
11668 |
|
|
{
|
11669 |
|
|
if (! sh_fsca_int2sf_rtx)
|
11670 |
|
|
{
|
11671 |
|
|
REAL_VALUE_TYPE rv;
|
11672 |
|
|
|
11673 |
|
|
real_from_string (&rv, "9.587379924285257e-5");
|
11674 |
|
|
sh_fsca_int2sf_rtx = const_double_from_real_value (rv, SFmode);
|
11675 |
|
|
}
|
11676 |
|
|
|
11677 |
|
|
return sh_fsca_int2sf_rtx;
|
11678 |
|
|
}
|
11679 |
|
|
|
11680 |
|
|
/* Initialize the CUMULATIVE_ARGS structure. */
|
11681 |
|
|
|
11682 |
|
|
void
|
11683 |
|
|
sh_init_cumulative_args (CUMULATIVE_ARGS * pcum,
|
11684 |
|
|
tree fntype,
|
11685 |
|
|
rtx libname ATTRIBUTE_UNUSED,
|
11686 |
|
|
tree fndecl,
|
11687 |
|
|
signed int n_named_args,
|
11688 |
|
|
enum machine_mode mode)
|
11689 |
|
|
{
|
11690 |
|
|
pcum->arg_count [(int) SH_ARG_FLOAT] = 0;
|
11691 |
|
|
pcum->free_single_fp_reg = 0;
|
11692 |
|
|
pcum->stack_regs = 0;
|
11693 |
|
|
pcum->byref_regs = 0;
|
11694 |
|
|
pcum->byref = 0;
|
11695 |
|
|
pcum->outgoing = (n_named_args == -1) ? 0 : 1;
|
11696 |
|
|
|
11697 |
|
|
/* XXX - Should we check TARGET_HITACHI here ??? */
|
11698 |
|
|
pcum->renesas_abi = sh_attr_renesas_p (fntype) ? 1 : 0;
|
11699 |
|
|
|
11700 |
|
|
if (fntype)
|
11701 |
|
|
{
|
11702 |
|
|
pcum->force_mem = ((TARGET_HITACHI || pcum->renesas_abi)
|
11703 |
|
|
&& aggregate_value_p (TREE_TYPE (fntype), fndecl));
|
11704 |
|
|
pcum->prototype_p = TYPE_ARG_TYPES (fntype) ? TRUE : FALSE;
|
11705 |
|
|
pcum->arg_count [(int) SH_ARG_INT]
|
11706 |
|
|
= TARGET_SH5 && aggregate_value_p (TREE_TYPE (fntype), fndecl);
|
11707 |
|
|
|
11708 |
|
|
pcum->call_cookie
|
11709 |
|
|
= CALL_COOKIE_RET_TRAMP (TARGET_SHCOMPACT
|
11710 |
|
|
&& pcum->arg_count [(int) SH_ARG_INT] == 0
|
11711 |
|
|
&& (TYPE_MODE (TREE_TYPE (fntype)) == BLKmode
|
11712 |
|
|
? int_size_in_bytes (TREE_TYPE (fntype))
|
11713 |
|
|
: GET_MODE_SIZE (TYPE_MODE (TREE_TYPE (fntype)))) > 4
|
11714 |
|
|
&& (BASE_RETURN_VALUE_REG (TYPE_MODE (TREE_TYPE (fntype)))
|
11715 |
|
|
== FIRST_RET_REG));
|
11716 |
|
|
}
|
11717 |
|
|
else
|
11718 |
|
|
{
|
11719 |
|
|
pcum->arg_count [(int) SH_ARG_INT] = 0;
|
11720 |
|
|
pcum->prototype_p = FALSE;
|
11721 |
|
|
if (mode != VOIDmode)
|
11722 |
|
|
{
|
11723 |
|
|
pcum->call_cookie =
|
11724 |
|
|
CALL_COOKIE_RET_TRAMP (TARGET_SHCOMPACT
|
11725 |
|
|
&& GET_MODE_SIZE (mode) > 4
|
11726 |
|
|
&& BASE_RETURN_VALUE_REG (mode) == FIRST_RET_REG);
|
11727 |
|
|
|
11728 |
|
|
/* If the default ABI is the Renesas ABI then all library
|
11729 |
|
|
calls must assume that the library will be using the
|
11730 |
|
|
Renesas ABI. So if the function would return its result
|
11731 |
|
|
in memory then we must force the address of this memory
|
11732 |
|
|
block onto the stack. Ideally we would like to call
|
11733 |
|
|
targetm.calls.return_in_memory() here but we do not have
|
11734 |
|
|
the TYPE or the FNDECL available so we synthesize the
|
11735 |
|
|
contents of that function as best we can. */
|
11736 |
|
|
pcum->force_mem =
|
11737 |
|
|
(TARGET_DEFAULT & MASK_HITACHI)
|
11738 |
|
|
&& (mode == BLKmode
|
11739 |
|
|
|| (GET_MODE_SIZE (mode) > 4
|
11740 |
|
|
&& !(mode == DFmode
|
11741 |
|
|
&& TARGET_FPU_DOUBLE)));
|
11742 |
|
|
}
|
11743 |
|
|
else
|
11744 |
|
|
{
|
11745 |
|
|
pcum->call_cookie = 0;
|
11746 |
|
|
pcum->force_mem = FALSE;
|
11747 |
|
|
}
|
11748 |
|
|
}
|
11749 |
|
|
}
|
11750 |
|
|
|
11751 |
|
|
/* Replace any occurrence of FROM(n) in X with TO(n). The function does
|
11752 |
|
|
not enter into CONST_DOUBLE for the replace.
|
11753 |
|
|
|
11754 |
|
|
Note that copying is not done so X must not be shared unless all copies
|
11755 |
|
|
are to be modified.
|
11756 |
|
|
|
11757 |
|
|
This is like replace_rtx, except that we operate on N_REPLACEMENTS
|
11758 |
|
|
replacements simultaneously - FROM(n) is replacements[n*2] and to(n) is
|
11759 |
|
|
replacements[n*2+1] - and that we take mode changes into account.
|
11760 |
|
|
|
11761 |
|
|
If a replacement is ambiguous, return NULL_RTX.
|
11762 |
|
|
|
11763 |
|
|
If MODIFY is zero, don't modify any rtl in place,
|
11764 |
|
|
just return zero or nonzero for failure / success. */
|
11765 |
|
|
|
11766 |
|
|
rtx
|
11767 |
|
|
replace_n_hard_rtx (rtx x, rtx *replacements, int n_replacements, int modify)
|
11768 |
|
|
{
|
11769 |
|
|
int i, j;
|
11770 |
|
|
const char *fmt;
|
11771 |
|
|
|
11772 |
|
|
/* The following prevents loops occurrence when we change MEM in
|
11773 |
|
|
CONST_DOUBLE onto the same CONST_DOUBLE. */
|
11774 |
|
|
if (x != 0 && GET_CODE (x) == CONST_DOUBLE)
|
11775 |
|
|
return x;
|
11776 |
|
|
|
11777 |
|
|
for (i = n_replacements - 1; i >= 0 ; i--)
|
11778 |
|
|
if (x == replacements[i*2] && GET_MODE (x) == GET_MODE (replacements[i*2+1]))
|
11779 |
|
|
return replacements[i*2+1];
|
11780 |
|
|
|
11781 |
|
|
/* Allow this function to make replacements in EXPR_LISTs. */
|
11782 |
|
|
if (x == 0)
|
11783 |
|
|
return 0;
|
11784 |
|
|
|
11785 |
|
|
if (GET_CODE (x) == SUBREG)
|
11786 |
|
|
{
|
11787 |
|
|
rtx new_rtx = replace_n_hard_rtx (SUBREG_REG (x), replacements,
|
11788 |
|
|
n_replacements, modify);
|
11789 |
|
|
|
11790 |
|
|
if (CONST_INT_P (new_rtx))
|
11791 |
|
|
{
|
11792 |
|
|
x = simplify_subreg (GET_MODE (x), new_rtx,
|
11793 |
|
|
GET_MODE (SUBREG_REG (x)),
|
11794 |
|
|
SUBREG_BYTE (x));
|
11795 |
|
|
if (! x)
|
11796 |
|
|
abort ();
|
11797 |
|
|
}
|
11798 |
|
|
else if (modify)
|
11799 |
|
|
SUBREG_REG (x) = new_rtx;
|
11800 |
|
|
|
11801 |
|
|
return x;
|
11802 |
|
|
}
|
11803 |
|
|
else if (REG_P (x))
|
11804 |
|
|
{
|
11805 |
|
|
unsigned regno = REGNO (x);
|
11806 |
|
|
unsigned nregs = (regno < FIRST_PSEUDO_REGISTER
|
11807 |
|
|
? HARD_REGNO_NREGS (regno, GET_MODE (x)) : 1);
|
11808 |
|
|
rtx result = NULL_RTX;
|
11809 |
|
|
|
11810 |
|
|
for (i = n_replacements - 1; i >= 0; i--)
|
11811 |
|
|
{
|
11812 |
|
|
rtx from = replacements[i*2];
|
11813 |
|
|
rtx to = replacements[i*2+1];
|
11814 |
|
|
unsigned from_regno, from_nregs, to_regno, new_regno;
|
11815 |
|
|
|
11816 |
|
|
if (!REG_P (from))
|
11817 |
|
|
continue;
|
11818 |
|
|
from_regno = REGNO (from);
|
11819 |
|
|
from_nregs = (from_regno < FIRST_PSEUDO_REGISTER
|
11820 |
|
|
? HARD_REGNO_NREGS (from_regno, GET_MODE (from)) : 1);
|
11821 |
|
|
if (regno < from_regno + from_nregs && regno + nregs > from_regno)
|
11822 |
|
|
{
|
11823 |
|
|
if (regno < from_regno
|
11824 |
|
|
|| regno + nregs > from_regno + nregs
|
11825 |
|
|
|| !REG_P (to)
|
11826 |
|
|
|| result)
|
11827 |
|
|
return NULL_RTX;
|
11828 |
|
|
to_regno = REGNO (to);
|
11829 |
|
|
if (to_regno < FIRST_PSEUDO_REGISTER)
|
11830 |
|
|
{
|
11831 |
|
|
new_regno = regno + to_regno - from_regno;
|
11832 |
|
|
if ((unsigned) HARD_REGNO_NREGS (new_regno, GET_MODE (x))
|
11833 |
|
|
!= nregs)
|
11834 |
|
|
return NULL_RTX;
|
11835 |
|
|
result = gen_rtx_REG (GET_MODE (x), new_regno);
|
11836 |
|
|
}
|
11837 |
|
|
else if (GET_MODE (x) <= GET_MODE (to))
|
11838 |
|
|
result = gen_lowpart_common (GET_MODE (x), to);
|
11839 |
|
|
else
|
11840 |
|
|
result = gen_lowpart_SUBREG (GET_MODE (x), to);
|
11841 |
|
|
}
|
11842 |
|
|
}
|
11843 |
|
|
return result ? result : x;
|
11844 |
|
|
}
|
11845 |
|
|
else if (GET_CODE (x) == ZERO_EXTEND)
|
11846 |
|
|
{
|
11847 |
|
|
rtx new_rtx = replace_n_hard_rtx (XEXP (x, 0), replacements,
|
11848 |
|
|
n_replacements, modify);
|
11849 |
|
|
|
11850 |
|
|
if (CONST_INT_P (new_rtx))
|
11851 |
|
|
{
|
11852 |
|
|
x = simplify_unary_operation (ZERO_EXTEND, GET_MODE (x),
|
11853 |
|
|
new_rtx, GET_MODE (XEXP (x, 0)));
|
11854 |
|
|
if (! x)
|
11855 |
|
|
abort ();
|
11856 |
|
|
}
|
11857 |
|
|
else if (modify)
|
11858 |
|
|
XEXP (x, 0) = new_rtx;
|
11859 |
|
|
|
11860 |
|
|
return x;
|
11861 |
|
|
}
|
11862 |
|
|
|
11863 |
|
|
fmt = GET_RTX_FORMAT (GET_CODE (x));
|
11864 |
|
|
for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--)
|
11865 |
|
|
{
|
11866 |
|
|
rtx new_rtx;
|
11867 |
|
|
|
11868 |
|
|
if (fmt[i] == 'e')
|
11869 |
|
|
{
|
11870 |
|
|
new_rtx = replace_n_hard_rtx (XEXP (x, i), replacements,
|
11871 |
|
|
n_replacements, modify);
|
11872 |
|
|
if (!new_rtx)
|
11873 |
|
|
return NULL_RTX;
|
11874 |
|
|
if (modify)
|
11875 |
|
|
XEXP (x, i) = new_rtx;
|
11876 |
|
|
}
|
11877 |
|
|
else if (fmt[i] == 'E')
|
11878 |
|
|
for (j = XVECLEN (x, i) - 1; j >= 0; j--)
|
11879 |
|
|
{
|
11880 |
|
|
new_rtx = replace_n_hard_rtx (XVECEXP (x, i, j), replacements,
|
11881 |
|
|
n_replacements, modify);
|
11882 |
|
|
if (!new_rtx)
|
11883 |
|
|
return NULL_RTX;
|
11884 |
|
|
if (modify)
|
11885 |
|
|
XVECEXP (x, i, j) = new_rtx;
|
11886 |
|
|
}
|
11887 |
|
|
}
|
11888 |
|
|
|
11889 |
|
|
return x;
|
11890 |
|
|
}
|
11891 |
|
|
|
11892 |
|
|
rtx
|
11893 |
|
|
sh_gen_truncate (enum machine_mode mode, rtx x, int need_sign_ext)
|
11894 |
|
|
{
|
11895 |
|
|
enum rtx_code code = TRUNCATE;
|
11896 |
|
|
|
11897 |
|
|
if (GET_CODE (x) == ZERO_EXTEND || GET_CODE (x) == SIGN_EXTEND)
|
11898 |
|
|
{
|
11899 |
|
|
rtx inner = XEXP (x, 0);
|
11900 |
|
|
enum machine_mode inner_mode = GET_MODE (inner);
|
11901 |
|
|
|
11902 |
|
|
if (inner_mode == mode)
|
11903 |
|
|
return inner;
|
11904 |
|
|
else if (GET_MODE_SIZE (inner_mode) >= GET_MODE_SIZE (mode))
|
11905 |
|
|
x = inner;
|
11906 |
|
|
else if (GET_MODE_SIZE (inner_mode) < GET_MODE_SIZE (mode)
|
11907 |
|
|
&& (! need_sign_ext || GET_CODE (x) == SIGN_EXTEND))
|
11908 |
|
|
{
|
11909 |
|
|
code = GET_CODE (x);
|
11910 |
|
|
x = inner;
|
11911 |
|
|
}
|
11912 |
|
|
}
|
11913 |
|
|
return gen_rtx_fmt_e (code, mode, x);
|
11914 |
|
|
}
|
11915 |
|
|
|
11916 |
|
|
/* called via for_each_rtx after reload, to clean up truncates of
|
11917 |
|
|
registers that span multiple actual hard registers. */
|
11918 |
|
|
int
|
11919 |
|
|
shmedia_cleanup_truncate (rtx *p, void *n_changes)
|
11920 |
|
|
{
|
11921 |
|
|
rtx x = *p, reg;
|
11922 |
|
|
|
11923 |
|
|
if (GET_CODE (x) != TRUNCATE)
|
11924 |
|
|
return 0;
|
11925 |
|
|
reg = XEXP (x, 0);
|
11926 |
|
|
if (GET_MODE_SIZE (GET_MODE (reg)) > 8 && REG_P (reg))
|
11927 |
|
|
{
|
11928 |
|
|
enum machine_mode reg_mode = GET_MODE (reg);
|
11929 |
|
|
XEXP (x, 0) = simplify_subreg (DImode, reg, reg_mode,
|
11930 |
|
|
subreg_lowpart_offset (DImode, reg_mode));
|
11931 |
|
|
*(int*) n_changes += 1;
|
11932 |
|
|
return -1;
|
11933 |
|
|
}
|
11934 |
|
|
return 0;
|
11935 |
|
|
}
|
11936 |
|
|
|
11937 |
|
|
/* Load and store depend on the highpart of the address. However,
|
11938 |
|
|
set_attr_alternative does not give well-defined results before reload,
|
11939 |
|
|
so we must look at the rtl ourselves to see if any of the feeding
|
11940 |
|
|
registers is used in a memref. */
|
11941 |
|
|
|
11942 |
|
|
/* Called by sh_contains_memref_p via for_each_rtx. */
|
11943 |
|
|
static int
|
11944 |
|
|
sh_contains_memref_p_1 (rtx *loc, void *data ATTRIBUTE_UNUSED)
|
11945 |
|
|
{
|
11946 |
|
|
return (MEM_P (*loc));
|
11947 |
|
|
}
|
11948 |
|
|
|
11949 |
|
|
/* Return nonzero iff INSN contains a MEM. */
|
11950 |
|
|
int
|
11951 |
|
|
sh_contains_memref_p (rtx insn)
|
11952 |
|
|
{
|
11953 |
|
|
return for_each_rtx (&PATTERN (insn), &sh_contains_memref_p_1, NULL);
|
11954 |
|
|
}
|
11955 |
|
|
|
11956 |
|
|
/* Return nonzero iff INSN loads a banked register. */
|
11957 |
|
|
int
|
11958 |
|
|
sh_loads_bankedreg_p (rtx insn)
|
11959 |
|
|
{
|
11960 |
|
|
if (GET_CODE (PATTERN (insn)) == SET)
|
11961 |
|
|
{
|
11962 |
|
|
rtx op = SET_DEST (PATTERN(insn));
|
11963 |
|
|
if (REG_P (op) && BANKED_REGISTER_P (REGNO (op)))
|
11964 |
|
|
return 1;
|
11965 |
|
|
}
|
11966 |
|
|
|
11967 |
|
|
return 0;
|
11968 |
|
|
}
|
11969 |
|
|
|
11970 |
|
|
/* FNADDR is the MEM expression from a call expander. Return an address
|
11971 |
|
|
to use in an SHmedia insn pattern. */
|
11972 |
|
|
rtx
|
11973 |
|
|
shmedia_prepare_call_address (rtx fnaddr, int is_sibcall)
|
11974 |
|
|
{
|
11975 |
|
|
int is_sym;
|
11976 |
|
|
|
11977 |
|
|
fnaddr = XEXP (fnaddr, 0);
|
11978 |
|
|
is_sym = GET_CODE (fnaddr) == SYMBOL_REF;
|
11979 |
|
|
if (flag_pic && is_sym)
|
11980 |
|
|
{
|
11981 |
|
|
if (! SYMBOL_REF_LOCAL_P (fnaddr))
|
11982 |
|
|
{
|
11983 |
|
|
rtx reg = gen_reg_rtx (Pmode);
|
11984 |
|
|
|
11985 |
|
|
/* We must not use GOTPLT for sibcalls, because PIC_REG
|
11986 |
|
|
must be restored before the PLT code gets to run. */
|
11987 |
|
|
if (is_sibcall)
|
11988 |
|
|
emit_insn (gen_symGOT2reg (reg, fnaddr));
|
11989 |
|
|
else
|
11990 |
|
|
emit_insn (gen_symGOTPLT2reg (reg, fnaddr));
|
11991 |
|
|
fnaddr = reg;
|
11992 |
|
|
}
|
11993 |
|
|
else
|
11994 |
|
|
{
|
11995 |
|
|
fnaddr = gen_sym2PIC (fnaddr);
|
11996 |
|
|
PUT_MODE (fnaddr, Pmode);
|
11997 |
|
|
}
|
11998 |
|
|
}
|
11999 |
|
|
/* If ptabs might trap, make this visible to the rest of the compiler.
|
12000 |
|
|
We generally assume that symbols pertain to valid locations, but
|
12001 |
|
|
it is possible to generate invalid symbols with asm or linker tricks.
|
12002 |
|
|
In a list of functions where each returns its successor, an invalid
|
12003 |
|
|
symbol might denote an empty list. */
|
12004 |
|
|
if (!TARGET_PT_FIXED
|
12005 |
|
|
&& (!is_sym || TARGET_INVALID_SYMBOLS)
|
12006 |
|
|
&& (!REG_P (fnaddr) || ! TARGET_REGISTER_P (REGNO (fnaddr))))
|
12007 |
|
|
{
|
12008 |
|
|
rtx tr = gen_reg_rtx (PDImode);
|
12009 |
|
|
|
12010 |
|
|
emit_insn (gen_ptabs (tr, fnaddr));
|
12011 |
|
|
fnaddr = tr;
|
12012 |
|
|
}
|
12013 |
|
|
else if (! target_reg_operand (fnaddr, Pmode))
|
12014 |
|
|
fnaddr = copy_to_mode_reg (Pmode, fnaddr);
|
12015 |
|
|
return fnaddr;
|
12016 |
|
|
}
|
12017 |
|
|
|
12018 |
|
|
enum reg_class
|
12019 |
|
|
sh_secondary_reload (bool in_p, rtx x, enum reg_class rclass,
|
12020 |
|
|
enum machine_mode mode, secondary_reload_info *sri)
|
12021 |
|
|
{
|
12022 |
|
|
if (in_p)
|
12023 |
|
|
{
|
12024 |
|
|
if (REGCLASS_HAS_FP_REG (rclass)
|
12025 |
|
|
&& ! TARGET_SHMEDIA
|
12026 |
|
|
&& immediate_operand ((x), mode)
|
12027 |
|
|
&& ! ((fp_zero_operand (x) || fp_one_operand (x))
|
12028 |
|
|
&& mode == SFmode && fldi_ok ()))
|
12029 |
|
|
switch (mode)
|
12030 |
|
|
{
|
12031 |
|
|
case SFmode:
|
12032 |
|
|
sri->icode = CODE_FOR_reload_insf__frn;
|
12033 |
|
|
return NO_REGS;
|
12034 |
|
|
case DFmode:
|
12035 |
|
|
sri->icode = CODE_FOR_reload_indf__frn;
|
12036 |
|
|
return NO_REGS;
|
12037 |
|
|
case SImode:
|
12038 |
|
|
/* ??? If we knew that we are in the appropriate mode -
|
12039 |
|
|
single precision - we could use a reload pattern directly. */
|
12040 |
|
|
return FPUL_REGS;
|
12041 |
|
|
default:
|
12042 |
|
|
abort ();
|
12043 |
|
|
}
|
12044 |
|
|
if (rclass == FPUL_REGS
|
12045 |
|
|
&& ((REG_P (x)
|
12046 |
|
|
&& (REGNO (x) == MACL_REG || REGNO (x) == MACH_REG
|
12047 |
|
|
|| REGNO (x) == T_REG))
|
12048 |
|
|
|| GET_CODE (x) == PLUS))
|
12049 |
|
|
return GENERAL_REGS;
|
12050 |
|
|
if (rclass == FPUL_REGS && immediate_operand (x, mode))
|
12051 |
|
|
{
|
12052 |
|
|
if (satisfies_constraint_I08 (x) || fp_zero_operand (x))
|
12053 |
|
|
return GENERAL_REGS;
|
12054 |
|
|
else if (mode == SFmode)
|
12055 |
|
|
return FP_REGS;
|
12056 |
|
|
sri->icode = CODE_FOR_reload_insi__i_fpul;
|
12057 |
|
|
return NO_REGS;
|
12058 |
|
|
}
|
12059 |
|
|
if (rclass == FPSCR_REGS
|
12060 |
|
|
&& ((REG_P (x) && REGNO (x) >= FIRST_PSEUDO_REGISTER)
|
12061 |
|
|
|| (MEM_P (x) && GET_CODE (XEXP (x, 0)) == PLUS)))
|
12062 |
|
|
return GENERAL_REGS;
|
12063 |
|
|
if (REGCLASS_HAS_FP_REG (rclass)
|
12064 |
|
|
&& TARGET_SHMEDIA
|
12065 |
|
|
&& immediate_operand (x, mode)
|
12066 |
|
|
&& x != CONST0_RTX (GET_MODE (x))
|
12067 |
|
|
&& GET_MODE (x) != V4SFmode)
|
12068 |
|
|
return GENERAL_REGS;
|
12069 |
|
|
if ((mode == QImode || mode == HImode)
|
12070 |
|
|
&& TARGET_SHMEDIA && inqhi_operand (x, mode))
|
12071 |
|
|
{
|
12072 |
|
|
sri->icode = ((mode == QImode)
|
12073 |
|
|
? CODE_FOR_reload_inqi : CODE_FOR_reload_inhi);
|
12074 |
|
|
return NO_REGS;
|
12075 |
|
|
}
|
12076 |
|
|
if (TARGET_SHMEDIA && rclass == GENERAL_REGS
|
12077 |
|
|
&& (GET_CODE (x) == LABEL_REF || PIC_ADDR_P (x)))
|
12078 |
|
|
return TARGET_REGS;
|
12079 |
|
|
} /* end of input-only processing. */
|
12080 |
|
|
|
12081 |
|
|
if (((REGCLASS_HAS_FP_REG (rclass)
|
12082 |
|
|
&& (REG_P (x)
|
12083 |
|
|
&& (GENERAL_OR_AP_REGISTER_P (REGNO (x))
|
12084 |
|
|
|| (FP_REGISTER_P (REGNO (x)) && mode == SImode
|
12085 |
|
|
&& TARGET_FMOVD))))
|
12086 |
|
|
|| (REGCLASS_HAS_GENERAL_REG (rclass)
|
12087 |
|
|
&& REG_P (x)
|
12088 |
|
|
&& FP_REGISTER_P (REGNO (x))))
|
12089 |
|
|
&& ! TARGET_SHMEDIA
|
12090 |
|
|
&& (mode == SFmode || mode == SImode))
|
12091 |
|
|
return FPUL_REGS;
|
12092 |
|
|
if ((rclass == FPUL_REGS
|
12093 |
|
|
|| (REGCLASS_HAS_FP_REG (rclass)
|
12094 |
|
|
&& ! TARGET_SHMEDIA && mode == SImode))
|
12095 |
|
|
&& (MEM_P (x)
|
12096 |
|
|
|| (REG_P (x)
|
12097 |
|
|
&& (REGNO (x) >= FIRST_PSEUDO_REGISTER
|
12098 |
|
|
|| REGNO (x) == T_REG
|
12099 |
|
|
|| system_reg_operand (x, VOIDmode)))))
|
12100 |
|
|
{
|
12101 |
|
|
if (rclass == FPUL_REGS)
|
12102 |
|
|
return GENERAL_REGS;
|
12103 |
|
|
return FPUL_REGS;
|
12104 |
|
|
}
|
12105 |
|
|
if ((rclass == TARGET_REGS
|
12106 |
|
|
|| (TARGET_SHMEDIA && rclass == SIBCALL_REGS))
|
12107 |
|
|
&& !satisfies_constraint_Csy (x)
|
12108 |
|
|
&& (!REG_P (x) || ! GENERAL_REGISTER_P (REGNO (x))))
|
12109 |
|
|
return GENERAL_REGS;
|
12110 |
|
|
if ((rclass == MAC_REGS || rclass == PR_REGS)
|
12111 |
|
|
&& REG_P (x) && ! GENERAL_REGISTER_P (REGNO (x))
|
12112 |
|
|
&& rclass != REGNO_REG_CLASS (REGNO (x)))
|
12113 |
|
|
return GENERAL_REGS;
|
12114 |
|
|
if (rclass != GENERAL_REGS && REG_P (x)
|
12115 |
|
|
&& TARGET_REGISTER_P (REGNO (x)))
|
12116 |
|
|
return GENERAL_REGS;
|
12117 |
|
|
return NO_REGS;
|
12118 |
|
|
}
|
12119 |
|
|
|
12120 |
|
|
enum sh_divide_strategy_e sh_div_strategy = SH_DIV_STRATEGY_DEFAULT;
|
12121 |
|
|
|
12122 |
|
|
#include "gt-sh.h"
|