OpenCores
URL https://opencores.org/ocsvn/openrisc/openrisc/trunk

Subversion Repositories openrisc

[/] [openrisc/] [trunk/] [gnu-dev/] [or1k-gcc/] [gcc/] [tree-vectorizer.h] - Blame information for rev 762

Go to most recent revision | Details | Compare with Previous | View Log

Line No. Rev Author Line
1 684 jeremybenn
/* Vectorizer
2
   Copyright (C) 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011, 2012
3
   Free Software Foundation, Inc.
4
   Contributed by Dorit Naishlos <dorit@il.ibm.com>
5
 
6
This file is part of GCC.
7
 
8
GCC is free software; you can redistribute it and/or modify it under
9
the terms of the GNU General Public License as published by the Free
10
Software Foundation; either version 3, or (at your option) any later
11
version.
12
 
13
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
14
WARRANTY; without even the implied warranty of MERCHANTABILITY or
15
FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
16
for more details.
17
 
18
You should have received a copy of the GNU General Public License
19
along with GCC; see the file COPYING3.  If not see
20
<http://www.gnu.org/licenses/>.  */
21
 
22
#ifndef GCC_TREE_VECTORIZER_H
23
#define GCC_TREE_VECTORIZER_H
24
 
25
#include "tree-data-ref.h"
26
 
27
typedef source_location LOC;
28
#define UNKNOWN_LOC UNKNOWN_LOCATION
29
#define EXPR_LOC(e) EXPR_LOCATION(e)
30
#define LOC_FILE(l) LOCATION_FILE (l)
31
#define LOC_LINE(l) LOCATION_LINE (l)
32
 
33
/* Used for naming of new temporaries.  */
34
enum vect_var_kind {
35
  vect_simple_var,
36
  vect_pointer_var,
37
  vect_scalar_var
38
};
39
 
40
/* Defines type of operation.  */
41
enum operation_type {
42
  unary_op = 1,
43
  binary_op,
44
  ternary_op
45
};
46
 
47
/* Define type of available alignment support.  */
48
enum dr_alignment_support {
49
  dr_unaligned_unsupported,
50
  dr_unaligned_supported,
51
  dr_explicit_realign,
52
  dr_explicit_realign_optimized,
53
  dr_aligned
54
};
55
 
56
/* Define type of def-use cross-iteration cycle.  */
57
enum vect_def_type {
58
  vect_uninitialized_def = 0,
59
  vect_constant_def = 1,
60
  vect_external_def,
61
  vect_internal_def,
62
  vect_induction_def,
63
  vect_reduction_def,
64
  vect_double_reduction_def,
65
  vect_nested_cycle,
66
  vect_unknown_def_type
67
};
68
 
69
#define VECTORIZABLE_CYCLE_DEF(D) (((D) == vect_reduction_def)           \
70
                                   || ((D) == vect_double_reduction_def) \
71
                                   || ((D) == vect_nested_cycle))
72
 
73
/************************************************************************
74
  SLP
75
 ************************************************************************/
76
typedef void *slp_void_p;
77
DEF_VEC_P (slp_void_p);
78
DEF_VEC_ALLOC_P (slp_void_p, heap);
79
 
80
/* A computation tree of an SLP instance.  Each node corresponds to a group of
81
   stmts to be packed in a SIMD stmt.  */
82
typedef struct _slp_tree {
83
  /* Nodes that contain def-stmts of this node statements operands.  */
84
  VEC (slp_void_p, heap) *children;
85
  /* A group of scalar stmts to be vectorized together.  */
86
  VEC (gimple, heap) *stmts;
87
  /* Vectorized stmt/s.  */
88
  VEC (gimple, heap) *vec_stmts;
89
  /* Number of vector stmts that are created to replace the group of scalar
90
     stmts. It is calculated during the transformation phase as the number of
91
     scalar elements in one scalar iteration (GROUP_SIZE) multiplied by VF
92
     divided by vector size.  */
93
  unsigned int vec_stmts_size;
94
  /* Vectorization costs associated with SLP node.  */
95
  struct
96
  {
97
    int outside_of_loop;     /* Statements generated outside loop.  */
98
    int inside_of_loop;      /* Statements generated inside loop.  */
99
  } cost;
100
} *slp_tree;
101
 
102
DEF_VEC_P(slp_tree);
103
DEF_VEC_ALLOC_P(slp_tree, heap);
104
 
105
/* SLP instance is a sequence of stmts in a loop that can be packed into
106
   SIMD stmts.  */
107
typedef struct _slp_instance {
108
  /* The root of SLP tree.  */
109
  slp_tree root;
110
 
111
  /* Size of groups of scalar stmts that will be replaced by SIMD stmt/s.  */
112
  unsigned int group_size;
113
 
114
  /* The unrolling factor required to vectorized this SLP instance.  */
115
  unsigned int unrolling_factor;
116
 
117
  /* Vectorization costs associated with SLP instance.  */
118
  struct
119
  {
120
    int outside_of_loop;     /* Statements generated outside loop.  */
121
    int inside_of_loop;      /* Statements generated inside loop.  */
122
  } cost;
123
 
124
  /* Loads permutation relatively to the stores, NULL if there is no
125
     permutation.  */
126
  VEC (int, heap) *load_permutation;
127
 
128
  /* The group of nodes that contain loads of this SLP instance.  */
129
  VEC (slp_tree, heap) *loads;
130
 
131
  /* The first scalar load of the instance. The created vector loads will be
132
     inserted before this statement.  */
133
  gimple first_load;
134
} *slp_instance;
135
 
136
DEF_VEC_P(slp_instance);
137
DEF_VEC_ALLOC_P(slp_instance, heap);
138
 
139
/* Access Functions.  */
140
#define SLP_INSTANCE_TREE(S)                     (S)->root
141
#define SLP_INSTANCE_GROUP_SIZE(S)               (S)->group_size
142
#define SLP_INSTANCE_UNROLLING_FACTOR(S)         (S)->unrolling_factor
143
#define SLP_INSTANCE_OUTSIDE_OF_LOOP_COST(S)     (S)->cost.outside_of_loop
144
#define SLP_INSTANCE_INSIDE_OF_LOOP_COST(S)      (S)->cost.inside_of_loop
145
#define SLP_INSTANCE_LOAD_PERMUTATION(S)         (S)->load_permutation
146
#define SLP_INSTANCE_LOADS(S)                    (S)->loads
147
#define SLP_INSTANCE_FIRST_LOAD_STMT(S)          (S)->first_load
148
 
149
#define SLP_TREE_CHILDREN(S)                     (S)->children
150
#define SLP_TREE_SCALAR_STMTS(S)                 (S)->stmts
151
#define SLP_TREE_VEC_STMTS(S)                    (S)->vec_stmts
152
#define SLP_TREE_NUMBER_OF_VEC_STMTS(S)          (S)->vec_stmts_size
153
#define SLP_TREE_OUTSIDE_OF_LOOP_COST(S)         (S)->cost.outside_of_loop
154
#define SLP_TREE_INSIDE_OF_LOOP_COST(S)          (S)->cost.inside_of_loop
155
 
156
/* This structure is used in creation of an SLP tree.  Each instance
157
   corresponds to the same operand in a group of scalar stmts in an SLP
158
   node.  */
159
typedef struct _slp_oprnd_info
160
{
161
  /* Def-stmts for the operands.  */
162
  VEC (gimple, heap) *def_stmts;
163
  /* Information about the first statement, its vector def-type, type, the
164
     operand itself in case it's constant, and an indication if it's a pattern
165
     stmt.  */
166
  enum vect_def_type first_dt;
167
  tree first_def_type;
168
  tree first_const_oprnd;
169
  bool first_pattern;
170
} *slp_oprnd_info;
171
 
172
DEF_VEC_P(slp_oprnd_info);
173
DEF_VEC_ALLOC_P(slp_oprnd_info, heap);
174
 
175
 
176
typedef struct _vect_peel_info
177
{
178
  int npeel;
179
  struct data_reference *dr;
180
  unsigned int count;
181
} *vect_peel_info;
182
 
183
typedef struct _vect_peel_extended_info
184
{
185
  struct _vect_peel_info peel_info;
186
  unsigned int inside_cost;
187
  unsigned int outside_cost;
188
} *vect_peel_extended_info;
189
 
190
/*-----------------------------------------------------------------*/
191
/* Info on vectorized loops.                                       */
192
/*-----------------------------------------------------------------*/
193
typedef struct _loop_vec_info {
194
 
195
  /* The loop to which this info struct refers to.  */
196
  struct loop *loop;
197
 
198
  /* The loop basic blocks.  */
199
  basic_block *bbs;
200
 
201
  /* Number of iterations.  */
202
  tree num_iters;
203
  tree num_iters_unchanged;
204
 
205
  /* Minimum number of iterations below which vectorization is expected to
206
     not be profitable (as estimated by the cost model).
207
     -1 indicates that vectorization will not be profitable.
208
     FORNOW: This field is an int. Will be a tree in the future, to represent
209
             values unknown at compile time.  */
210
  int min_profitable_iters;
211
 
212
  /* Is the loop vectorizable? */
213
  bool vectorizable;
214
 
215
  /* Unrolling factor  */
216
  int vectorization_factor;
217
 
218
  /* The loop location in the source.  */
219
  LOC loop_line_number;
220
 
221
  /* Unknown DRs according to which loop was peeled.  */
222
  struct data_reference *unaligned_dr;
223
 
224
  /* peeling_for_alignment indicates whether peeling for alignment will take
225
     place, and what the peeling factor should be:
226
     peeling_for_alignment = X means:
227
        If X=0: Peeling for alignment will not be applied.
228
        If X>0: Peel first X iterations.
229
        If X=-1: Generate a runtime test to calculate the number of iterations
230
                 to be peeled, using the dataref recorded in the field
231
                 unaligned_dr.  */
232
  int peeling_for_alignment;
233
 
234
  /* The mask used to check the alignment of pointers or arrays.  */
235
  int ptr_mask;
236
 
237
  /* The loop nest in which the data dependences are computed.  */
238
  VEC (loop_p, heap) *loop_nest;
239
 
240
  /* All data references in the loop.  */
241
  VEC (data_reference_p, heap) *datarefs;
242
 
243
  /* All data dependences in the loop.  */
244
  VEC (ddr_p, heap) *ddrs;
245
 
246
  /* Data Dependence Relations defining address ranges that are candidates
247
     for a run-time aliasing check.  */
248
  VEC (ddr_p, heap) *may_alias_ddrs;
249
 
250
  /* Statements in the loop that have data references that are candidates for a
251
     runtime (loop versioning) misalignment check.  */
252
  VEC(gimple,heap) *may_misalign_stmts;
253
 
254
  /* All interleaving chains of stores in the loop, represented by the first
255
     stmt in the chain.  */
256
  VEC(gimple, heap) *strided_stores;
257
 
258
  /* All SLP instances in the loop. This is a subset of the set of STRIDED_STORES
259
     of the loop.  */
260
  VEC(slp_instance, heap) *slp_instances;
261
 
262
  /* The unrolling factor needed to SLP the loop. In case of that pure SLP is
263
     applied to the loop, i.e., no unrolling is needed, this is 1.  */
264
  unsigned slp_unrolling_factor;
265
 
266
  /* Reduction cycles detected in the loop. Used in loop-aware SLP.  */
267
  VEC (gimple, heap) *reductions;
268
 
269
  /* All reduction chains in the loop, represented by the first
270
     stmt in the chain.  */
271
  VEC (gimple, heap) *reduction_chains;
272
 
273
  /* Hash table used to choose the best peeling option.  */
274
  htab_t peeling_htab;
275
 
276
  /* When we have strided data accesses with gaps, we may introduce invalid
277
     memory accesses.  We peel the last iteration of the loop to prevent
278
     this.  */
279
  bool peeling_for_gaps;
280
 
281
} *loop_vec_info;
282
 
283
/* Access Functions.  */
284
#define LOOP_VINFO_LOOP(L)                 (L)->loop
285
#define LOOP_VINFO_BBS(L)                  (L)->bbs
286
#define LOOP_VINFO_NITERS(L)               (L)->num_iters
287
/* Since LOOP_VINFO_NITERS can change after prologue peeling
288
   retain total unchanged scalar loop iterations for cost model.  */
289
#define LOOP_VINFO_NITERS_UNCHANGED(L)     (L)->num_iters_unchanged
290
#define LOOP_VINFO_COST_MODEL_MIN_ITERS(L) (L)->min_profitable_iters
291
#define LOOP_VINFO_VECTORIZABLE_P(L)       (L)->vectorizable
292
#define LOOP_VINFO_VECT_FACTOR(L)          (L)->vectorization_factor
293
#define LOOP_VINFO_PTR_MASK(L)             (L)->ptr_mask
294
#define LOOP_VINFO_LOOP_NEST(L)            (L)->loop_nest
295
#define LOOP_VINFO_DATAREFS(L)             (L)->datarefs
296
#define LOOP_VINFO_DDRS(L)                 (L)->ddrs
297
#define LOOP_VINFO_INT_NITERS(L)           (TREE_INT_CST_LOW ((L)->num_iters))
298
#define LOOP_PEELING_FOR_ALIGNMENT(L)      (L)->peeling_for_alignment
299
#define LOOP_VINFO_UNALIGNED_DR(L)         (L)->unaligned_dr
300
#define LOOP_VINFO_MAY_MISALIGN_STMTS(L)   (L)->may_misalign_stmts
301
#define LOOP_VINFO_LOC(L)                  (L)->loop_line_number
302
#define LOOP_VINFO_MAY_ALIAS_DDRS(L)       (L)->may_alias_ddrs
303
#define LOOP_VINFO_STRIDED_STORES(L)       (L)->strided_stores
304
#define LOOP_VINFO_SLP_INSTANCES(L)        (L)->slp_instances
305
#define LOOP_VINFO_SLP_UNROLLING_FACTOR(L) (L)->slp_unrolling_factor
306
#define LOOP_VINFO_REDUCTIONS(L)           (L)->reductions
307
#define LOOP_VINFO_REDUCTION_CHAINS(L)     (L)->reduction_chains
308
#define LOOP_VINFO_PEELING_HTAB(L)         (L)->peeling_htab
309
#define LOOP_VINFO_PEELING_FOR_GAPS(L)     (L)->peeling_for_gaps
310
 
311
#define LOOP_REQUIRES_VERSIONING_FOR_ALIGNMENT(L) \
312
VEC_length (gimple, (L)->may_misalign_stmts) > 0
313
#define LOOP_REQUIRES_VERSIONING_FOR_ALIAS(L)     \
314
VEC_length (ddr_p, (L)->may_alias_ddrs) > 0
315
 
316
#define NITERS_KNOWN_P(n)                     \
317
(host_integerp ((n),0)                        \
318
&& TREE_INT_CST_LOW ((n)) > 0)
319
 
320
#define LOOP_VINFO_NITERS_KNOWN_P(L)          \
321
NITERS_KNOWN_P((L)->num_iters)
322
 
323
static inline loop_vec_info
324
loop_vec_info_for_loop (struct loop *loop)
325
{
326
  return (loop_vec_info) loop->aux;
327
}
328
 
329
static inline bool
330
nested_in_vect_loop_p (struct loop *loop, gimple stmt)
331
{
332
  return (loop->inner
333
          && (loop->inner == (gimple_bb (stmt))->loop_father));
334
}
335
 
336
typedef struct _bb_vec_info {
337
 
338
  basic_block bb;
339
  /* All interleaving chains of stores in the basic block, represented by the
340
     first stmt in the chain.  */
341
  VEC(gimple, heap) *strided_stores;
342
 
343
  /* All SLP instances in the basic block. This is a subset of the set of
344
     STRIDED_STORES of the basic block.  */
345
  VEC(slp_instance, heap) *slp_instances;
346
 
347
  /* All data references in the basic block.  */
348
  VEC (data_reference_p, heap) *datarefs;
349
 
350
  /* All data dependences in the basic block.  */
351
  VEC (ddr_p, heap) *ddrs;
352
} *bb_vec_info;
353
 
354
#define BB_VINFO_BB(B)              (B)->bb
355
#define BB_VINFO_STRIDED_STORES(B)  (B)->strided_stores
356
#define BB_VINFO_SLP_INSTANCES(B)   (B)->slp_instances
357
#define BB_VINFO_DATAREFS(B)        (B)->datarefs
358
#define BB_VINFO_DDRS(B)            (B)->ddrs
359
 
360
static inline bb_vec_info
361
vec_info_for_bb (basic_block bb)
362
{
363
  return (bb_vec_info) bb->aux;
364
}
365
 
366
/*-----------------------------------------------------------------*/
367
/* Info on vectorized defs.                                        */
368
/*-----------------------------------------------------------------*/
369
enum stmt_vec_info_type {
370
  undef_vec_info_type = 0,
371
  load_vec_info_type,
372
  store_vec_info_type,
373
  shift_vec_info_type,
374
  op_vec_info_type,
375
  call_vec_info_type,
376
  assignment_vec_info_type,
377
  condition_vec_info_type,
378
  reduc_vec_info_type,
379
  induc_vec_info_type,
380
  type_promotion_vec_info_type,
381
  type_demotion_vec_info_type,
382
  type_conversion_vec_info_type,
383
  loop_exit_ctrl_vec_info_type
384
};
385
 
386
/* Indicates whether/how a variable is used in the scope of loop/basic
387
   block.  */
388
enum vect_relevant {
389
  vect_unused_in_scope = 0,
390
  /* The def is in the inner loop, and the use is in the outer loop, and the
391
     use is a reduction stmt.  */
392
  vect_used_in_outer_by_reduction,
393
  /* The def is in the inner loop, and the use is in the outer loop (and is
394
     not part of reduction).  */
395
  vect_used_in_outer,
396
 
397
  /* defs that feed computations that end up (only) in a reduction. These
398
     defs may be used by non-reduction stmts, but eventually, any
399
     computations/values that are affected by these defs are used to compute
400
     a reduction (i.e. don't get stored to memory, for example). We use this
401
     to identify computations that we can change the order in which they are
402
     computed.  */
403
  vect_used_by_reduction,
404
 
405
  vect_used_in_scope
406
};
407
 
408
/* The type of vectorization that can be applied to the stmt: regular loop-based
409
   vectorization; pure SLP - the stmt is a part of SLP instances and does not
410
   have uses outside SLP instances; or hybrid SLP and loop-based - the stmt is
411
   a part of SLP instance and also must be loop-based vectorized, since it has
412
   uses outside SLP sequences.
413
 
414
   In the loop context the meanings of pure and hybrid SLP are slightly
415
   different. By saying that pure SLP is applied to the loop, we mean that we
416
   exploit only intra-iteration parallelism in the loop; i.e., the loop can be
417
   vectorized without doing any conceptual unrolling, cause we don't pack
418
   together stmts from different iterations, only within a single iteration.
419
   Loop hybrid SLP means that we exploit both intra-iteration and
420
   inter-iteration parallelism (e.g., number of elements in the vector is 4
421
   and the slp-group-size is 2, in which case we don't have enough parallelism
422
   within an iteration, so we obtain the rest of the parallelism from subsequent
423
   iterations by unrolling the loop by 2).  */
424
enum slp_vect_type {
425
  loop_vect = 0,
426
  pure_slp,
427
  hybrid
428
};
429
 
430
 
431
typedef struct data_reference *dr_p;
432
DEF_VEC_P(dr_p);
433
DEF_VEC_ALLOC_P(dr_p,heap);
434
 
435
typedef struct _stmt_vec_info {
436
 
437
  enum stmt_vec_info_type type;
438
 
439
  /* Indicates whether this stmts is part of a computation whose result is
440
     used outside the loop.  */
441
  bool live;
442
 
443
  /* Stmt is part of some pattern (computation idiom)  */
444
  bool in_pattern_p;
445
 
446
  /* For loads only, if there is a store with the same location, this field is
447
     TRUE.  */
448
  bool read_write_dep;
449
 
450
  /* The stmt to which this info struct refers to.  */
451
  gimple stmt;
452
 
453
  /* The loop_vec_info with respect to which STMT is vectorized.  */
454
  loop_vec_info loop_vinfo;
455
 
456
  /* The vector type to be used for the LHS of this statement.  */
457
  tree vectype;
458
 
459
  /* The vectorized version of the stmt.  */
460
  gimple vectorized_stmt;
461
 
462
 
463
  /** The following is relevant only for stmts that contain a non-scalar
464
     data-ref (array/pointer/struct access). A GIMPLE stmt is expected to have
465
     at most one such data-ref.  **/
466
 
467
  /* Information about the data-ref (access function, etc),
468
     relative to the inner-most containing loop.  */
469
  struct data_reference *data_ref_info;
470
 
471
  /* Information about the data-ref relative to this loop
472
     nest (the loop that is being considered for vectorization).  */
473
  tree dr_base_address;
474
  tree dr_init;
475
  tree dr_offset;
476
  tree dr_step;
477
  tree dr_aligned_to;
478
 
479
  /* Used for various bookkeeping purposes, generally holding a pointer to
480
     some other stmt S that is in some way "related" to this stmt.
481
     Current use of this field is:
482
        If this stmt is part of a pattern (i.e. the field 'in_pattern_p' is
483
        true): S is the "pattern stmt" that represents (and replaces) the
484
        sequence of stmts that constitutes the pattern.  Similarly, the
485
        related_stmt of the "pattern stmt" points back to this stmt (which is
486
        the last stmt in the original sequence of stmts that constitutes the
487
        pattern).  */
488
  gimple related_stmt;
489
 
490
  /* Used to keep a sequence of def stmts of a pattern stmt if such exists.  */
491
  gimple_seq pattern_def_seq;
492
 
493
  /* List of datarefs that are known to have the same alignment as the dataref
494
     of this stmt.  */
495
  VEC(dr_p,heap) *same_align_refs;
496
 
497
  /* Classify the def of this stmt.  */
498
  enum vect_def_type def_type;
499
 
500
  /*  Whether the stmt is SLPed, loop-based vectorized, or both.  */
501
  enum slp_vect_type slp_type;
502
 
503
  /* Interleaving and reduction chains info.  */
504
  /* First element in the group.  */
505
  gimple first_element;
506
  /* Pointer to the next element in the group.  */
507
  gimple next_element;
508
  /* For data-refs, in case that two or more stmts share data-ref, this is the
509
     pointer to the previously detected stmt with the same dr.  */
510
  gimple same_dr_stmt;
511
  /* The size of the group.  */
512
  unsigned int size;
513
  /* For stores, number of stores from this group seen. We vectorize the last
514
     one.  */
515
  unsigned int store_count;
516
  /* For loads only, the gap from the previous load. For consecutive loads, GAP
517
     is 1.  */
518
  unsigned int gap;
519
 
520
  /* Not all stmts in the loop need to be vectorized. e.g, the increment
521
     of the loop induction variable and computation of array indexes. relevant
522
     indicates whether the stmt needs to be vectorized.  */
523
  enum vect_relevant relevant;
524
 
525
  /* Vectorization costs associated with statement.  */
526
  struct
527
  {
528
    int outside_of_loop;     /* Statements generated outside loop.  */
529
    int inside_of_loop;      /* Statements generated inside loop.  */
530
  } cost;
531
 
532
  /* The bb_vec_info with respect to which STMT is vectorized.  */
533
  bb_vec_info bb_vinfo;
534
 
535
  /* Is this statement vectorizable or should it be skipped in (partial)
536
     vectorization.  */
537
  bool vectorizable;
538
 
539
  /* For loads only, true if this is a gather load.  */
540
  bool gather_p;
541
} *stmt_vec_info;
542
 
543
/* Access Functions.  */
544
#define STMT_VINFO_TYPE(S)                 (S)->type
545
#define STMT_VINFO_STMT(S)                 (S)->stmt
546
#define STMT_VINFO_LOOP_VINFO(S)           (S)->loop_vinfo
547
#define STMT_VINFO_BB_VINFO(S)             (S)->bb_vinfo
548
#define STMT_VINFO_RELEVANT(S)             (S)->relevant
549
#define STMT_VINFO_LIVE_P(S)               (S)->live
550
#define STMT_VINFO_VECTYPE(S)              (S)->vectype
551
#define STMT_VINFO_VEC_STMT(S)             (S)->vectorized_stmt
552
#define STMT_VINFO_VECTORIZABLE(S)         (S)->vectorizable
553
#define STMT_VINFO_DATA_REF(S)             (S)->data_ref_info
554
#define STMT_VINFO_GATHER_P(S)             (S)->gather_p
555
 
556
#define STMT_VINFO_DR_BASE_ADDRESS(S)      (S)->dr_base_address
557
#define STMT_VINFO_DR_INIT(S)              (S)->dr_init
558
#define STMT_VINFO_DR_OFFSET(S)            (S)->dr_offset
559
#define STMT_VINFO_DR_STEP(S)              (S)->dr_step
560
#define STMT_VINFO_DR_ALIGNED_TO(S)        (S)->dr_aligned_to
561
 
562
#define STMT_VINFO_IN_PATTERN_P(S)         (S)->in_pattern_p
563
#define STMT_VINFO_RELATED_STMT(S)         (S)->related_stmt
564
#define STMT_VINFO_PATTERN_DEF_SEQ(S)      (S)->pattern_def_seq
565
#define STMT_VINFO_SAME_ALIGN_REFS(S)      (S)->same_align_refs
566
#define STMT_VINFO_DEF_TYPE(S)             (S)->def_type
567
#define STMT_VINFO_GROUP_FIRST_ELEMENT(S)  (S)->first_element
568
#define STMT_VINFO_GROUP_NEXT_ELEMENT(S)   (S)->next_element
569
#define STMT_VINFO_GROUP_SIZE(S)           (S)->size
570
#define STMT_VINFO_GROUP_STORE_COUNT(S)    (S)->store_count
571
#define STMT_VINFO_GROUP_GAP(S)            (S)->gap
572
#define STMT_VINFO_GROUP_SAME_DR_STMT(S)   (S)->same_dr_stmt
573
#define STMT_VINFO_GROUP_READ_WRITE_DEPENDENCE(S)  (S)->read_write_dep
574
#define STMT_VINFO_STRIDED_ACCESS(S)      ((S)->first_element != NULL && (S)->data_ref_info)
575
 
576
#define GROUP_FIRST_ELEMENT(S)          (S)->first_element
577
#define GROUP_NEXT_ELEMENT(S)           (S)->next_element
578
#define GROUP_SIZE(S)                   (S)->size
579
#define GROUP_STORE_COUNT(S)            (S)->store_count
580
#define GROUP_GAP(S)                    (S)->gap
581
#define GROUP_SAME_DR_STMT(S)           (S)->same_dr_stmt
582
#define GROUP_READ_WRITE_DEPENDENCE(S)  (S)->read_write_dep
583
 
584
#define STMT_VINFO_RELEVANT_P(S)          ((S)->relevant != vect_unused_in_scope)
585
#define STMT_VINFO_OUTSIDE_OF_LOOP_COST(S) (S)->cost.outside_of_loop
586
#define STMT_VINFO_INSIDE_OF_LOOP_COST(S)  (S)->cost.inside_of_loop
587
 
588
#define HYBRID_SLP_STMT(S)                ((S)->slp_type == hybrid)
589
#define PURE_SLP_STMT(S)                  ((S)->slp_type == pure_slp)
590
#define STMT_SLP_TYPE(S)                   (S)->slp_type
591
 
592
#define VECT_MAX_COST 1000
593
 
594
/* The maximum number of intermediate steps required in multi-step type
595
   conversion.  */
596
#define MAX_INTERM_CVT_STEPS         3
597
 
598
/* The maximum vectorization factor supported by any target (V32QI).  */
599
#define MAX_VECTORIZATION_FACTOR 32
600
 
601
/* Avoid GTY(()) on stmt_vec_info.  */
602
typedef void *vec_void_p;
603
DEF_VEC_P (vec_void_p);
604
DEF_VEC_ALLOC_P (vec_void_p, heap);
605
 
606
extern VEC(vec_void_p,heap) *stmt_vec_info_vec;
607
 
608
void init_stmt_vec_info_vec (void);
609
void free_stmt_vec_info_vec (void);
610
 
611
/* Return a stmt_vec_info corresponding to STMT.  */
612
 
613
static inline stmt_vec_info
614
vinfo_for_stmt (gimple stmt)
615
{
616
  unsigned int uid = gimple_uid (stmt);
617
  if (uid == 0)
618
    return NULL;
619
 
620
  return (stmt_vec_info) VEC_index (vec_void_p, stmt_vec_info_vec, uid - 1);
621
}
622
 
623
/* Set vectorizer information INFO for STMT.  */
624
 
625
static inline void
626
set_vinfo_for_stmt (gimple stmt, stmt_vec_info info)
627
{
628
  unsigned int uid = gimple_uid (stmt);
629
  if (uid == 0)
630
    {
631
      gcc_checking_assert (info);
632
      uid = VEC_length (vec_void_p, stmt_vec_info_vec) + 1;
633
      gimple_set_uid (stmt, uid);
634
      VEC_safe_push (vec_void_p, heap, stmt_vec_info_vec, (vec_void_p) info);
635
    }
636
  else
637
    VEC_replace (vec_void_p, stmt_vec_info_vec, uid - 1, (vec_void_p) info);
638
}
639
 
640
/* Return the earlier statement between STMT1 and STMT2.  */
641
 
642
static inline gimple
643
get_earlier_stmt (gimple stmt1, gimple stmt2)
644
{
645
  unsigned int uid1, uid2;
646
 
647
  if (stmt1 == NULL)
648
    return stmt2;
649
 
650
  if (stmt2 == NULL)
651
    return stmt1;
652
 
653
  uid1 = gimple_uid (stmt1);
654
  uid2 = gimple_uid (stmt2);
655
 
656
  if (uid1 == 0 || uid2 == 0)
657
    return NULL;
658
 
659
  gcc_checking_assert (uid1 <= VEC_length (vec_void_p, stmt_vec_info_vec)
660
                       && uid2 <= VEC_length (vec_void_p, stmt_vec_info_vec));
661
 
662
  if (uid1 < uid2)
663
    return stmt1;
664
  else
665
    return stmt2;
666
}
667
 
668
/* Return the later statement between STMT1 and STMT2.  */
669
 
670
static inline gimple
671
get_later_stmt (gimple stmt1, gimple stmt2)
672
{
673
  unsigned int uid1, uid2;
674
 
675
  if (stmt1 == NULL)
676
    return stmt2;
677
 
678
  if (stmt2 == NULL)
679
    return stmt1;
680
 
681
  uid1 = gimple_uid (stmt1);
682
  uid2 = gimple_uid (stmt2);
683
 
684
  if (uid1 == 0 || uid2 == 0)
685
    return NULL;
686
 
687
  gcc_assert (uid1 <= VEC_length (vec_void_p, stmt_vec_info_vec));
688
  gcc_assert (uid2 <= VEC_length (vec_void_p, stmt_vec_info_vec));
689
 
690
  if (uid1 > uid2)
691
    return stmt1;
692
  else
693
    return stmt2;
694
}
695
 
696
/* Return TRUE if a statement represented by STMT_INFO is a part of a
697
   pattern.  */
698
 
699
static inline bool
700
is_pattern_stmt_p (stmt_vec_info stmt_info)
701
{
702
  gimple related_stmt;
703
  stmt_vec_info related_stmt_info;
704
 
705
  related_stmt = STMT_VINFO_RELATED_STMT (stmt_info);
706
  if (related_stmt
707
      && (related_stmt_info = vinfo_for_stmt (related_stmt))
708
      && STMT_VINFO_IN_PATTERN_P (related_stmt_info))
709
    return true;
710
 
711
  return false;
712
}
713
 
714
/* Return true if BB is a loop header.  */
715
 
716
static inline bool
717
is_loop_header_bb_p (basic_block bb)
718
{
719
  if (bb == (bb->loop_father)->header)
720
    return true;
721
  gcc_checking_assert (EDGE_COUNT (bb->preds) == 1);
722
  return false;
723
}
724
 
725
/* Set inside loop vectorization cost.  */
726
 
727
static inline void
728
stmt_vinfo_set_inside_of_loop_cost (stmt_vec_info stmt_info, slp_tree slp_node,
729
                                    int cost)
730
{
731
  if (slp_node)
732
    SLP_TREE_INSIDE_OF_LOOP_COST (slp_node) = cost;
733
  else
734
    STMT_VINFO_INSIDE_OF_LOOP_COST (stmt_info) = cost;
735
}
736
 
737
/* Set inside loop vectorization cost.  */
738
 
739
static inline void
740
stmt_vinfo_set_outside_of_loop_cost (stmt_vec_info stmt_info, slp_tree slp_node,
741
                                     int cost)
742
{
743
  if (slp_node)
744
    SLP_TREE_OUTSIDE_OF_LOOP_COST (slp_node) = cost;
745
  else
746
    STMT_VINFO_OUTSIDE_OF_LOOP_COST (stmt_info) = cost;
747
}
748
 
749
/* Return pow2 (X).  */
750
 
751
static inline int
752
vect_pow2 (int x)
753
{
754
  int i, res = 1;
755
 
756
  for (i = 0; i < x; i++)
757
    res *= 2;
758
 
759
  return res;
760
}
761
 
762
/*-----------------------------------------------------------------*/
763
/* Info on data references alignment.                              */
764
/*-----------------------------------------------------------------*/
765
 
766
/* Reflects actual alignment of first access in the vectorized loop,
767
   taking into account peeling/versioning if applied.  */
768
#define DR_MISALIGNMENT(DR)   ((int) (size_t) (DR)->aux)
769
#define SET_DR_MISALIGNMENT(DR, VAL)   ((DR)->aux = (void *) (size_t) (VAL))
770
 
771
/* Return TRUE if the data access is aligned, and FALSE otherwise.  */
772
 
773
static inline bool
774
aligned_access_p (struct data_reference *data_ref_info)
775
{
776
  return (DR_MISALIGNMENT (data_ref_info) == 0);
777
}
778
 
779
/* Return TRUE if the alignment of the data access is known, and FALSE
780
   otherwise.  */
781
 
782
static inline bool
783
known_alignment_for_access_p (struct data_reference *data_ref_info)
784
{
785
  return (DR_MISALIGNMENT (data_ref_info) != -1);
786
}
787
 
788
/* vect_dump will be set to stderr or dump_file if exist.  */
789
extern FILE *vect_dump;
790
extern LOC vect_loop_location;
791
 
792
/*-----------------------------------------------------------------*/
793
/* Function prototypes.                                            */
794
/*-----------------------------------------------------------------*/
795
 
796
/* Simple loop peeling and versioning utilities for vectorizer's purposes -
797
   in tree-vect-loop-manip.c.  */
798
extern void slpeel_make_loop_iterate_ntimes (struct loop *, tree);
799
extern bool slpeel_can_duplicate_loop_p (const struct loop *, const_edge);
800
extern void vect_loop_versioning (loop_vec_info, bool, tree *, gimple_seq *);
801
extern void vect_do_peeling_for_loop_bound (loop_vec_info, tree *,
802
                                            tree, gimple_seq);
803
extern void vect_do_peeling_for_alignment (loop_vec_info);
804
extern LOC find_loop_location (struct loop *);
805
extern bool vect_can_advance_ivs_p (loop_vec_info);
806
 
807
/* In tree-vect-stmts.c.  */
808
extern unsigned int current_vector_size;
809
extern tree get_vectype_for_scalar_type (tree);
810
extern tree get_same_sized_vectype (tree, tree);
811
extern bool vect_is_simple_use (tree, gimple, loop_vec_info,
812
                                bb_vec_info, gimple *,
813
                                tree *,  enum vect_def_type *);
814
extern bool vect_is_simple_use_1 (tree, gimple, loop_vec_info,
815
                                  bb_vec_info, gimple *,
816
                                  tree *,  enum vect_def_type *, tree *);
817
extern bool supportable_widening_operation (enum tree_code, gimple, tree, tree,
818
                                            tree *, tree *, enum tree_code *,
819
                                            enum tree_code *, int *,
820
                                            VEC (tree, heap) **);
821
extern bool supportable_narrowing_operation (enum tree_code, tree, tree,
822
                                             enum tree_code *,
823
                                             int *, VEC (tree, heap) **);
824
extern stmt_vec_info new_stmt_vec_info (gimple stmt, loop_vec_info,
825
                                        bb_vec_info);
826
extern void free_stmt_vec_info (gimple stmt);
827
extern tree vectorizable_function (gimple, tree, tree);
828
extern void vect_model_simple_cost (stmt_vec_info, int, enum vect_def_type *,
829
                                    slp_tree);
830
extern void vect_model_store_cost (stmt_vec_info, int, bool,
831
                                   enum vect_def_type, slp_tree);
832
extern void vect_model_load_cost (stmt_vec_info, int, bool, slp_tree);
833
extern void vect_finish_stmt_generation (gimple, gimple,
834
                                         gimple_stmt_iterator *);
835
extern bool vect_mark_stmts_to_be_vectorized (loop_vec_info);
836
extern int cost_for_stmt (gimple);
837
extern tree vect_get_vec_def_for_operand (tree, gimple, tree *);
838
extern tree vect_init_vector (gimple, tree, tree,
839
                              gimple_stmt_iterator *);
840
extern tree vect_get_vec_def_for_stmt_copy (enum vect_def_type, tree);
841
extern bool vect_transform_stmt (gimple, gimple_stmt_iterator *,
842
                                 bool *, slp_tree, slp_instance);
843
extern void vect_remove_stores (gimple);
844
extern bool vect_analyze_stmt (gimple, bool *, slp_tree);
845
extern bool vectorizable_condition (gimple, gimple_stmt_iterator *, gimple *,
846
                                    tree, int, slp_tree);
847
extern void vect_get_load_cost (struct data_reference *, int, bool,
848
                                unsigned int *, unsigned int *);
849
extern void vect_get_store_cost (struct data_reference *, int, unsigned int *);
850
extern bool vect_supportable_shift (enum tree_code, tree);
851
extern void vect_get_vec_defs (tree, tree, gimple, VEC (tree, heap) **,
852
                               VEC (tree, heap) **, slp_tree, int);
853
extern tree vect_gen_perm_mask (tree, unsigned char *);
854
 
855
/* In tree-vect-data-refs.c.  */
856
extern bool vect_can_force_dr_alignment_p (const_tree, unsigned int);
857
extern enum dr_alignment_support vect_supportable_dr_alignment
858
                                           (struct data_reference *, bool);
859
extern tree vect_get_smallest_scalar_type (gimple, HOST_WIDE_INT *,
860
                                           HOST_WIDE_INT *);
861
extern bool vect_analyze_data_ref_dependences (loop_vec_info, bb_vec_info,
862
                                               int *);
863
extern bool vect_enhance_data_refs_alignment (loop_vec_info);
864
extern bool vect_analyze_data_refs_alignment (loop_vec_info, bb_vec_info);
865
extern bool vect_verify_datarefs_alignment (loop_vec_info, bb_vec_info);
866
extern bool vect_analyze_data_ref_accesses (loop_vec_info, bb_vec_info);
867
extern bool vect_prune_runtime_alias_test_list (loop_vec_info);
868
extern tree vect_check_gather (gimple, loop_vec_info, tree *, tree *,
869
                               int *);
870
extern bool vect_analyze_data_refs (loop_vec_info, bb_vec_info, int *);
871
extern tree vect_create_data_ref_ptr (gimple, tree, struct loop *, tree,
872
                                      tree *, gimple_stmt_iterator *,
873
                                      gimple *, bool, bool *);
874
extern tree bump_vector_ptr (tree, gimple, gimple_stmt_iterator *, gimple, tree);
875
extern tree vect_create_destination_var (tree, tree);
876
extern bool vect_strided_store_supported (tree, unsigned HOST_WIDE_INT);
877
extern bool vect_store_lanes_supported (tree, unsigned HOST_WIDE_INT);
878
extern bool vect_strided_load_supported (tree, unsigned HOST_WIDE_INT);
879
extern bool vect_load_lanes_supported (tree, unsigned HOST_WIDE_INT);
880
extern void vect_permute_store_chain (VEC(tree,heap) *,unsigned int, gimple,
881
                                    gimple_stmt_iterator *, VEC(tree,heap) **);
882
extern tree vect_setup_realignment (gimple, gimple_stmt_iterator *, tree *,
883
                                    enum dr_alignment_support, tree,
884
                                    struct loop **);
885
extern void vect_transform_strided_load (gimple, VEC(tree,heap) *, int,
886
                                         gimple_stmt_iterator *);
887
extern void vect_record_strided_load_vectors (gimple, VEC(tree,heap) *);
888
extern int vect_get_place_in_interleaving_chain (gimple, gimple);
889
extern tree vect_get_new_vect_var (tree, enum vect_var_kind, const char *);
890
extern tree vect_create_addr_base_for_vector_ref (gimple, gimple_seq *,
891
                                                  tree, struct loop *);
892
 
893
/* In tree-vect-loop.c.  */
894
/* FORNOW: Used in tree-parloops.c.  */
895
extern void destroy_loop_vec_info (loop_vec_info, bool);
896
extern gimple vect_force_simple_reduction (loop_vec_info, gimple, bool, bool *);
897
/* Drive for loop analysis stage.  */
898
extern loop_vec_info vect_analyze_loop (struct loop *);
899
/* Drive for loop transformation stage.  */
900
extern void vect_transform_loop (loop_vec_info);
901
extern loop_vec_info vect_analyze_loop_form (struct loop *);
902
extern bool vectorizable_live_operation (gimple, gimple_stmt_iterator *,
903
                                         gimple *);
904
extern bool vectorizable_reduction (gimple, gimple_stmt_iterator *, gimple *,
905
                                    slp_tree);
906
extern bool vectorizable_induction (gimple, gimple_stmt_iterator *, gimple *);
907
extern int vect_estimate_min_profitable_iters (loop_vec_info);
908
extern tree get_initial_def_for_reduction (gimple, tree, tree *);
909
extern int vect_min_worthwhile_factor (enum tree_code);
910
extern int vect_get_known_peeling_cost (loop_vec_info, int, int *, int);
911
extern int vect_get_single_scalar_iteraion_cost (loop_vec_info);
912
 
913
/* In tree-vect-slp.c.  */
914
extern void vect_free_slp_instance (slp_instance);
915
extern bool vect_transform_slp_perm_load (gimple, VEC (tree, heap) *,
916
                                          gimple_stmt_iterator *, int,
917
                                          slp_instance, bool);
918
extern bool vect_schedule_slp (loop_vec_info, bb_vec_info);
919
extern void vect_update_slp_costs_according_to_vf (loop_vec_info);
920
extern bool vect_analyze_slp (loop_vec_info, bb_vec_info);
921
extern bool vect_make_slp_decision (loop_vec_info);
922
extern void vect_detect_hybrid_slp (loop_vec_info);
923
extern void vect_get_slp_defs (VEC (tree, heap) *, slp_tree,
924
                               VEC (slp_void_p, heap) **, int);
925
 
926
extern LOC find_bb_location (basic_block);
927
extern bb_vec_info vect_slp_analyze_bb (basic_block);
928
extern void vect_slp_transform_bb (basic_block);
929
 
930
/* In tree-vect-patterns.c.  */
931
/* Pattern recognition functions.
932
   Additional pattern recognition functions can (and will) be added
933
   in the future.  */
934
typedef gimple (* vect_recog_func_ptr) (VEC (gimple, heap) **, tree *, tree *);
935
#define NUM_PATTERNS 10
936
void vect_pattern_recog (loop_vec_info);
937
 
938
/* In tree-vectorizer.c.  */
939
unsigned vectorize_loops (void);
940
/* Vectorization debug information */
941
extern bool vect_print_dump_info (enum vect_verbosity_levels);
942
 
943
#endif  /* GCC_TREE_VECTORIZER_H  */

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.