OpenCores
URL https://opencores.org/ocsvn/openrisc/openrisc/trunk

Subversion Repositories openrisc

[/] [openrisc/] [trunk/] [gnu-stable/] [gcc-4.5.1/] [libgomp/] [loop.c] - Blame information for rev 850

Go to most recent revision | Details | Compare with Previous | View Log

Line No. Rev Author Line
1 273 jeremybenn
/* Copyright (C) 2005, 2008, 2009 Free Software Foundation, Inc.
2
   Contributed by Richard Henderson <rth@redhat.com>.
3
 
4
   This file is part of the GNU OpenMP Library (libgomp).
5
 
6
   Libgomp is free software; you can redistribute it and/or modify it
7
   under the terms of the GNU General Public License as published by
8
   the Free Software Foundation; either version 3, or (at your option)
9
   any later version.
10
 
11
   Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
12
   WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
13
   FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
14
   more details.
15
 
16
   Under Section 7 of GPL version 3, you are granted additional
17
   permissions described in the GCC Runtime Library Exception, version
18
   3.1, as published by the Free Software Foundation.
19
 
20
   You should have received a copy of the GNU General Public License and
21
   a copy of the GCC Runtime Library Exception along with this program;
22
   see the files COPYING3 and COPYING.RUNTIME respectively.  If not, see
23
   <http://www.gnu.org/licenses/>.  */
24
 
25
/* This file handles the LOOP (FOR/DO) construct.  */
26
 
27
#include <limits.h>
28
#include <stdlib.h>
29
#include "libgomp.h"
30
 
31
 
32
/* Initialize the given work share construct from the given arguments.  */
33
 
34
static inline void
35
gomp_loop_init (struct gomp_work_share *ws, long start, long end, long incr,
36
                enum gomp_schedule_type sched, long chunk_size)
37
{
38
  ws->sched = sched;
39
  ws->chunk_size = chunk_size;
40
  /* Canonicalize loops that have zero iterations to ->next == ->end.  */
41
  ws->end = ((incr > 0 && start > end) || (incr < 0 && start < end))
42
            ? start : end;
43
  ws->incr = incr;
44
  ws->next = start;
45
  if (sched == GFS_DYNAMIC)
46
    {
47
      ws->chunk_size *= incr;
48
 
49
#ifdef HAVE_SYNC_BUILTINS
50
      {
51
        /* For dynamic scheduling prepare things to make each iteration
52
           faster.  */
53
        struct gomp_thread *thr = gomp_thread ();
54
        struct gomp_team *team = thr->ts.team;
55
        long nthreads = team ? team->nthreads : 1;
56
 
57
        if (__builtin_expect (incr > 0, 1))
58
          {
59
            /* Cheap overflow protection.  */
60
            if (__builtin_expect ((nthreads | ws->chunk_size)
61
                                  >= 1UL << (sizeof (long)
62
                                             * __CHAR_BIT__ / 2 - 1), 0))
63
              ws->mode = 0;
64
            else
65
              ws->mode = ws->end < (LONG_MAX
66
                                    - (nthreads + 1) * ws->chunk_size);
67
          }
68
        /* Cheap overflow protection.  */
69
        else if (__builtin_expect ((nthreads | -ws->chunk_size)
70
                                   >= 1UL << (sizeof (long)
71
                                              * __CHAR_BIT__ / 2 - 1), 0))
72
          ws->mode = 0;
73
        else
74
          ws->mode = ws->end > (nthreads + 1) * -ws->chunk_size - LONG_MAX;
75
      }
76
#endif
77
    }
78
}
79
 
80
/* The *_start routines are called when first encountering a loop construct
81
   that is not bound directly to a parallel construct.  The first thread
82
   that arrives will create the work-share construct; subsequent threads
83
   will see the construct exists and allocate work from it.
84
 
85
   START, END, INCR are the bounds of the loop; due to the restrictions of
86
   OpenMP, these values must be the same in every thread.  This is not
87
   verified (nor is it entirely verifiable, since START is not necessarily
88
   retained intact in the work-share data structure).  CHUNK_SIZE is the
89
   scheduling parameter; again this must be identical in all threads.
90
 
91
   Returns true if there's any work for this thread to perform.  If so,
92
   *ISTART and *IEND are filled with the bounds of the iteration block
93
   allocated to this thread.  Returns false if all work was assigned to
94
   other threads prior to this thread's arrival.  */
95
 
96
static bool
97
gomp_loop_static_start (long start, long end, long incr, long chunk_size,
98
                        long *istart, long *iend)
99
{
100
  struct gomp_thread *thr = gomp_thread ();
101
 
102
  thr->ts.static_trip = 0;
103
  if (gomp_work_share_start (false))
104
    {
105
      gomp_loop_init (thr->ts.work_share, start, end, incr,
106
                      GFS_STATIC, chunk_size);
107
      gomp_work_share_init_done ();
108
    }
109
 
110
  return !gomp_iter_static_next (istart, iend);
111
}
112
 
113
static bool
114
gomp_loop_dynamic_start (long start, long end, long incr, long chunk_size,
115
                         long *istart, long *iend)
116
{
117
  struct gomp_thread *thr = gomp_thread ();
118
  bool ret;
119
 
120
  if (gomp_work_share_start (false))
121
    {
122
      gomp_loop_init (thr->ts.work_share, start, end, incr,
123
                      GFS_DYNAMIC, chunk_size);
124
      gomp_work_share_init_done ();
125
    }
126
 
127
#ifdef HAVE_SYNC_BUILTINS
128
  ret = gomp_iter_dynamic_next (istart, iend);
129
#else
130
  gomp_mutex_lock (&thr->ts.work_share->lock);
131
  ret = gomp_iter_dynamic_next_locked (istart, iend);
132
  gomp_mutex_unlock (&thr->ts.work_share->lock);
133
#endif
134
 
135
  return ret;
136
}
137
 
138
static bool
139
gomp_loop_guided_start (long start, long end, long incr, long chunk_size,
140
                        long *istart, long *iend)
141
{
142
  struct gomp_thread *thr = gomp_thread ();
143
  bool ret;
144
 
145
  if (gomp_work_share_start (false))
146
    {
147
      gomp_loop_init (thr->ts.work_share, start, end, incr,
148
                      GFS_GUIDED, chunk_size);
149
      gomp_work_share_init_done ();
150
    }
151
 
152
#ifdef HAVE_SYNC_BUILTINS
153
  ret = gomp_iter_guided_next (istart, iend);
154
#else
155
  gomp_mutex_lock (&thr->ts.work_share->lock);
156
  ret = gomp_iter_guided_next_locked (istart, iend);
157
  gomp_mutex_unlock (&thr->ts.work_share->lock);
158
#endif
159
 
160
  return ret;
161
}
162
 
163
bool
164
GOMP_loop_runtime_start (long start, long end, long incr,
165
                         long *istart, long *iend)
166
{
167
  struct gomp_task_icv *icv = gomp_icv (false);
168
  switch (icv->run_sched_var)
169
    {
170
    case GFS_STATIC:
171
      return gomp_loop_static_start (start, end, incr, icv->run_sched_modifier,
172
                                     istart, iend);
173
    case GFS_DYNAMIC:
174
      return gomp_loop_dynamic_start (start, end, incr, icv->run_sched_modifier,
175
                                      istart, iend);
176
    case GFS_GUIDED:
177
      return gomp_loop_guided_start (start, end, incr, icv->run_sched_modifier,
178
                                     istart, iend);
179
    case GFS_AUTO:
180
      /* For now map to schedule(static), later on we could play with feedback
181
         driven choice.  */
182
      return gomp_loop_static_start (start, end, incr, 0, istart, iend);
183
    default:
184
      abort ();
185
    }
186
}
187
 
188
/* The *_ordered_*_start routines are similar.  The only difference is that
189
   this work-share construct is initialized to expect an ORDERED section.  */
190
 
191
static bool
192
gomp_loop_ordered_static_start (long start, long end, long incr,
193
                                long chunk_size, long *istart, long *iend)
194
{
195
  struct gomp_thread *thr = gomp_thread ();
196
 
197
  thr->ts.static_trip = 0;
198
  if (gomp_work_share_start (true))
199
    {
200
      gomp_loop_init (thr->ts.work_share, start, end, incr,
201
                      GFS_STATIC, chunk_size);
202
      gomp_ordered_static_init ();
203
      gomp_work_share_init_done ();
204
    }
205
 
206
  return !gomp_iter_static_next (istart, iend);
207
}
208
 
209
static bool
210
gomp_loop_ordered_dynamic_start (long start, long end, long incr,
211
                                 long chunk_size, long *istart, long *iend)
212
{
213
  struct gomp_thread *thr = gomp_thread ();
214
  bool ret;
215
 
216
  if (gomp_work_share_start (true))
217
    {
218
      gomp_loop_init (thr->ts.work_share, start, end, incr,
219
                      GFS_DYNAMIC, chunk_size);
220
      gomp_mutex_lock (&thr->ts.work_share->lock);
221
      gomp_work_share_init_done ();
222
    }
223
  else
224
    gomp_mutex_lock (&thr->ts.work_share->lock);
225
 
226
  ret = gomp_iter_dynamic_next_locked (istart, iend);
227
  if (ret)
228
    gomp_ordered_first ();
229
  gomp_mutex_unlock (&thr->ts.work_share->lock);
230
 
231
  return ret;
232
}
233
 
234
static bool
235
gomp_loop_ordered_guided_start (long start, long end, long incr,
236
                                long chunk_size, long *istart, long *iend)
237
{
238
  struct gomp_thread *thr = gomp_thread ();
239
  bool ret;
240
 
241
  if (gomp_work_share_start (true))
242
    {
243
      gomp_loop_init (thr->ts.work_share, start, end, incr,
244
                      GFS_GUIDED, chunk_size);
245
      gomp_mutex_lock (&thr->ts.work_share->lock);
246
      gomp_work_share_init_done ();
247
    }
248
  else
249
    gomp_mutex_lock (&thr->ts.work_share->lock);
250
 
251
  ret = gomp_iter_guided_next_locked (istart, iend);
252
  if (ret)
253
    gomp_ordered_first ();
254
  gomp_mutex_unlock (&thr->ts.work_share->lock);
255
 
256
  return ret;
257
}
258
 
259
bool
260
GOMP_loop_ordered_runtime_start (long start, long end, long incr,
261
                                 long *istart, long *iend)
262
{
263
  struct gomp_task_icv *icv = gomp_icv (false);
264
  switch (icv->run_sched_var)
265
    {
266
    case GFS_STATIC:
267
      return gomp_loop_ordered_static_start (start, end, incr,
268
                                             icv->run_sched_modifier,
269
                                             istart, iend);
270
    case GFS_DYNAMIC:
271
      return gomp_loop_ordered_dynamic_start (start, end, incr,
272
                                              icv->run_sched_modifier,
273
                                              istart, iend);
274
    case GFS_GUIDED:
275
      return gomp_loop_ordered_guided_start (start, end, incr,
276
                                             icv->run_sched_modifier,
277
                                             istart, iend);
278
    case GFS_AUTO:
279
      /* For now map to schedule(static), later on we could play with feedback
280
         driven choice.  */
281
      return gomp_loop_ordered_static_start (start, end, incr,
282
                                             0, istart, iend);
283
    default:
284
      abort ();
285
    }
286
}
287
 
288
/* The *_next routines are called when the thread completes processing of
289
   the iteration block currently assigned to it.  If the work-share
290
   construct is bound directly to a parallel construct, then the iteration
291
   bounds may have been set up before the parallel.  In which case, this
292
   may be the first iteration for the thread.
293
 
294
   Returns true if there is work remaining to be performed; *ISTART and
295
   *IEND are filled with a new iteration block.  Returns false if all work
296
   has been assigned.  */
297
 
298
static bool
299
gomp_loop_static_next (long *istart, long *iend)
300
{
301
  return !gomp_iter_static_next (istart, iend);
302
}
303
 
304
static bool
305
gomp_loop_dynamic_next (long *istart, long *iend)
306
{
307
  bool ret;
308
 
309
#ifdef HAVE_SYNC_BUILTINS
310
  ret = gomp_iter_dynamic_next (istart, iend);
311
#else
312
  struct gomp_thread *thr = gomp_thread ();
313
  gomp_mutex_lock (&thr->ts.work_share->lock);
314
  ret = gomp_iter_dynamic_next_locked (istart, iend);
315
  gomp_mutex_unlock (&thr->ts.work_share->lock);
316
#endif
317
 
318
  return ret;
319
}
320
 
321
static bool
322
gomp_loop_guided_next (long *istart, long *iend)
323
{
324
  bool ret;
325
 
326
#ifdef HAVE_SYNC_BUILTINS
327
  ret = gomp_iter_guided_next (istart, iend);
328
#else
329
  struct gomp_thread *thr = gomp_thread ();
330
  gomp_mutex_lock (&thr->ts.work_share->lock);
331
  ret = gomp_iter_guided_next_locked (istart, iend);
332
  gomp_mutex_unlock (&thr->ts.work_share->lock);
333
#endif
334
 
335
  return ret;
336
}
337
 
338
bool
339
GOMP_loop_runtime_next (long *istart, long *iend)
340
{
341
  struct gomp_thread *thr = gomp_thread ();
342
 
343
  switch (thr->ts.work_share->sched)
344
    {
345
    case GFS_STATIC:
346
    case GFS_AUTO:
347
      return gomp_loop_static_next (istart, iend);
348
    case GFS_DYNAMIC:
349
      return gomp_loop_dynamic_next (istart, iend);
350
    case GFS_GUIDED:
351
      return gomp_loop_guided_next (istart, iend);
352
    default:
353
      abort ();
354
    }
355
}
356
 
357
/* The *_ordered_*_next routines are called when the thread completes
358
   processing of the iteration block currently assigned to it.
359
 
360
   Returns true if there is work remaining to be performed; *ISTART and
361
   *IEND are filled with a new iteration block.  Returns false if all work
362
   has been assigned.  */
363
 
364
static bool
365
gomp_loop_ordered_static_next (long *istart, long *iend)
366
{
367
  struct gomp_thread *thr = gomp_thread ();
368
  int test;
369
 
370
  gomp_ordered_sync ();
371
  gomp_mutex_lock (&thr->ts.work_share->lock);
372
  test = gomp_iter_static_next (istart, iend);
373
  if (test >= 0)
374
    gomp_ordered_static_next ();
375
  gomp_mutex_unlock (&thr->ts.work_share->lock);
376
 
377
  return test == 0;
378
}
379
 
380
static bool
381
gomp_loop_ordered_dynamic_next (long *istart, long *iend)
382
{
383
  struct gomp_thread *thr = gomp_thread ();
384
  bool ret;
385
 
386
  gomp_ordered_sync ();
387
  gomp_mutex_lock (&thr->ts.work_share->lock);
388
  ret = gomp_iter_dynamic_next_locked (istart, iend);
389
  if (ret)
390
    gomp_ordered_next ();
391
  else
392
    gomp_ordered_last ();
393
  gomp_mutex_unlock (&thr->ts.work_share->lock);
394
 
395
  return ret;
396
}
397
 
398
static bool
399
gomp_loop_ordered_guided_next (long *istart, long *iend)
400
{
401
  struct gomp_thread *thr = gomp_thread ();
402
  bool ret;
403
 
404
  gomp_ordered_sync ();
405
  gomp_mutex_lock (&thr->ts.work_share->lock);
406
  ret = gomp_iter_guided_next_locked (istart, iend);
407
  if (ret)
408
    gomp_ordered_next ();
409
  else
410
    gomp_ordered_last ();
411
  gomp_mutex_unlock (&thr->ts.work_share->lock);
412
 
413
  return ret;
414
}
415
 
416
bool
417
GOMP_loop_ordered_runtime_next (long *istart, long *iend)
418
{
419
  struct gomp_thread *thr = gomp_thread ();
420
 
421
  switch (thr->ts.work_share->sched)
422
    {
423
    case GFS_STATIC:
424
    case GFS_AUTO:
425
      return gomp_loop_ordered_static_next (istart, iend);
426
    case GFS_DYNAMIC:
427
      return gomp_loop_ordered_dynamic_next (istart, iend);
428
    case GFS_GUIDED:
429
      return gomp_loop_ordered_guided_next (istart, iend);
430
    default:
431
      abort ();
432
    }
433
}
434
 
435
/* The GOMP_parallel_loop_* routines pre-initialize a work-share construct
436
   to avoid one synchronization once we get into the loop.  */
437
 
438
static void
439
gomp_parallel_loop_start (void (*fn) (void *), void *data,
440
                          unsigned num_threads, long start, long end,
441
                          long incr, enum gomp_schedule_type sched,
442
                          long chunk_size)
443
{
444
  struct gomp_team *team;
445
 
446
  num_threads = gomp_resolve_num_threads (num_threads, 0);
447
  team = gomp_new_team (num_threads);
448
  gomp_loop_init (&team->work_shares[0], start, end, incr, sched, chunk_size);
449
  gomp_team_start (fn, data, num_threads, team);
450
}
451
 
452
void
453
GOMP_parallel_loop_static_start (void (*fn) (void *), void *data,
454
                                 unsigned num_threads, long start, long end,
455
                                 long incr, long chunk_size)
456
{
457
  gomp_parallel_loop_start (fn, data, num_threads, start, end, incr,
458
                            GFS_STATIC, chunk_size);
459
}
460
 
461
void
462
GOMP_parallel_loop_dynamic_start (void (*fn) (void *), void *data,
463
                                  unsigned num_threads, long start, long end,
464
                                  long incr, long chunk_size)
465
{
466
  gomp_parallel_loop_start (fn, data, num_threads, start, end, incr,
467
                            GFS_DYNAMIC, chunk_size);
468
}
469
 
470
void
471
GOMP_parallel_loop_guided_start (void (*fn) (void *), void *data,
472
                                 unsigned num_threads, long start, long end,
473
                                 long incr, long chunk_size)
474
{
475
  gomp_parallel_loop_start (fn, data, num_threads, start, end, incr,
476
                            GFS_GUIDED, chunk_size);
477
}
478
 
479
void
480
GOMP_parallel_loop_runtime_start (void (*fn) (void *), void *data,
481
                                  unsigned num_threads, long start, long end,
482
                                  long incr)
483
{
484
  struct gomp_task_icv *icv = gomp_icv (false);
485
  gomp_parallel_loop_start (fn, data, num_threads, start, end, incr,
486
                            icv->run_sched_var, icv->run_sched_modifier);
487
}
488
 
489
/* The GOMP_loop_end* routines are called after the thread is told that
490
   all loop iterations are complete.  This first version synchronizes
491
   all threads; the nowait version does not.  */
492
 
493
void
494
GOMP_loop_end (void)
495
{
496
  gomp_work_share_end ();
497
}
498
 
499
void
500
GOMP_loop_end_nowait (void)
501
{
502
  gomp_work_share_end_nowait ();
503
}
504
 
505
 
506
/* We use static functions above so that we're sure that the "runtime"
507
   function can defer to the proper routine without interposition.  We
508
   export the static function with a strong alias when possible, or with
509
   a wrapper function otherwise.  */
510
 
511
#ifdef HAVE_ATTRIBUTE_ALIAS
512
extern __typeof(gomp_loop_static_start) GOMP_loop_static_start
513
        __attribute__((alias ("gomp_loop_static_start")));
514
extern __typeof(gomp_loop_dynamic_start) GOMP_loop_dynamic_start
515
        __attribute__((alias ("gomp_loop_dynamic_start")));
516
extern __typeof(gomp_loop_guided_start) GOMP_loop_guided_start
517
        __attribute__((alias ("gomp_loop_guided_start")));
518
 
519
extern __typeof(gomp_loop_ordered_static_start) GOMP_loop_ordered_static_start
520
        __attribute__((alias ("gomp_loop_ordered_static_start")));
521
extern __typeof(gomp_loop_ordered_dynamic_start) GOMP_loop_ordered_dynamic_start
522
        __attribute__((alias ("gomp_loop_ordered_dynamic_start")));
523
extern __typeof(gomp_loop_ordered_guided_start) GOMP_loop_ordered_guided_start
524
        __attribute__((alias ("gomp_loop_ordered_guided_start")));
525
 
526
extern __typeof(gomp_loop_static_next) GOMP_loop_static_next
527
        __attribute__((alias ("gomp_loop_static_next")));
528
extern __typeof(gomp_loop_dynamic_next) GOMP_loop_dynamic_next
529
        __attribute__((alias ("gomp_loop_dynamic_next")));
530
extern __typeof(gomp_loop_guided_next) GOMP_loop_guided_next
531
        __attribute__((alias ("gomp_loop_guided_next")));
532
 
533
extern __typeof(gomp_loop_ordered_static_next) GOMP_loop_ordered_static_next
534
        __attribute__((alias ("gomp_loop_ordered_static_next")));
535
extern __typeof(gomp_loop_ordered_dynamic_next) GOMP_loop_ordered_dynamic_next
536
        __attribute__((alias ("gomp_loop_ordered_dynamic_next")));
537
extern __typeof(gomp_loop_ordered_guided_next) GOMP_loop_ordered_guided_next
538
        __attribute__((alias ("gomp_loop_ordered_guided_next")));
539
#else
540
bool
541
GOMP_loop_static_start (long start, long end, long incr, long chunk_size,
542
                        long *istart, long *iend)
543
{
544
  return gomp_loop_static_start (start, end, incr, chunk_size, istart, iend);
545
}
546
 
547
bool
548
GOMP_loop_dynamic_start (long start, long end, long incr, long chunk_size,
549
                         long *istart, long *iend)
550
{
551
  return gomp_loop_dynamic_start (start, end, incr, chunk_size, istart, iend);
552
}
553
 
554
bool
555
GOMP_loop_guided_start (long start, long end, long incr, long chunk_size,
556
                        long *istart, long *iend)
557
{
558
  return gomp_loop_guided_start (start, end, incr, chunk_size, istart, iend);
559
}
560
 
561
bool
562
GOMP_loop_ordered_static_start (long start, long end, long incr,
563
                                long chunk_size, long *istart, long *iend)
564
{
565
  return gomp_loop_ordered_static_start (start, end, incr, chunk_size,
566
                                         istart, iend);
567
}
568
 
569
bool
570
GOMP_loop_ordered_dynamic_start (long start, long end, long incr,
571
                                 long chunk_size, long *istart, long *iend)
572
{
573
  return gomp_loop_ordered_dynamic_start (start, end, incr, chunk_size,
574
                                          istart, iend);
575
}
576
 
577
bool
578
GOMP_loop_ordered_guided_start (long start, long end, long incr,
579
                                long chunk_size, long *istart, long *iend)
580
{
581
  return gomp_loop_ordered_guided_start (start, end, incr, chunk_size,
582
                                         istart, iend);
583
}
584
 
585
bool
586
GOMP_loop_static_next (long *istart, long *iend)
587
{
588
  return gomp_loop_static_next (istart, iend);
589
}
590
 
591
bool
592
GOMP_loop_dynamic_next (long *istart, long *iend)
593
{
594
  return gomp_loop_dynamic_next (istart, iend);
595
}
596
 
597
bool
598
GOMP_loop_guided_next (long *istart, long *iend)
599
{
600
  return gomp_loop_guided_next (istart, iend);
601
}
602
 
603
bool
604
GOMP_loop_ordered_static_next (long *istart, long *iend)
605
{
606
  return gomp_loop_ordered_static_next (istart, iend);
607
}
608
 
609
bool
610
GOMP_loop_ordered_dynamic_next (long *istart, long *iend)
611
{
612
  return gomp_loop_ordered_dynamic_next (istart, iend);
613
}
614
 
615
bool
616
GOMP_loop_ordered_guided_next (long *istart, long *iend)
617
{
618
  return gomp_loop_ordered_guided_next (istart, iend);
619
}
620
#endif

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.