OpenCores
URL https://opencores.org/ocsvn/openrisc/openrisc/trunk

Subversion Repositories openrisc

[/] [openrisc/] [trunk/] [gnu-dev/] [or1k-gcc/] [libgomp/] [team.c] - Blame information for rev 735

Details | Compare with Previous | View Log

Line No. Rev Author Line
1 735 jeremybenn
/* Copyright (C) 2005, 2006, 2007, 2008, 2009, 2011
2
   Free Software Foundation, Inc.
3
   Contributed by Richard Henderson <rth@redhat.com>.
4
 
5
   This file is part of the GNU OpenMP Library (libgomp).
6
 
7
   Libgomp is free software; you can redistribute it and/or modify it
8
   under the terms of the GNU General Public License as published by
9
   the Free Software Foundation; either version 3, or (at your option)
10
   any later version.
11
 
12
   Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
13
   WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
14
   FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
15
   more details.
16
 
17
   Under Section 7 of GPL version 3, you are granted additional
18
   permissions described in the GCC Runtime Library Exception, version
19
   3.1, as published by the Free Software Foundation.
20
 
21
   You should have received a copy of the GNU General Public License and
22
   a copy of the GCC Runtime Library Exception along with this program;
23
   see the files COPYING3 and COPYING.RUNTIME respectively.  If not, see
24
   <http://www.gnu.org/licenses/>.  */
25
 
26
/* This file handles the maintainence of threads in response to team
27
   creation and termination.  */
28
 
29
#include "libgomp.h"
30
#include <stdlib.h>
31
#include <string.h>
32
 
33
/* This attribute contains PTHREAD_CREATE_DETACHED.  */
34
pthread_attr_t gomp_thread_attr;
35
 
36
/* This key is for the thread destructor.  */
37
pthread_key_t gomp_thread_destructor;
38
 
39
 
40
/* This is the libgomp per-thread data structure.  */
41
#ifdef HAVE_TLS
42
__thread struct gomp_thread gomp_tls_data;
43
#else
44
pthread_key_t gomp_tls_key;
45
#endif
46
 
47
 
48
/* This structure is used to communicate across pthread_create.  */
49
 
50
struct gomp_thread_start_data
51
{
52
  void (*fn) (void *);
53
  void *fn_data;
54
  struct gomp_team_state ts;
55
  struct gomp_task *task;
56
  struct gomp_thread_pool *thread_pool;
57
  bool nested;
58
};
59
 
60
 
61
/* This function is a pthread_create entry point.  This contains the idle
62
   loop in which a thread waits to be called up to become part of a team.  */
63
 
64
static void *
65
gomp_thread_start (void *xdata)
66
{
67
  struct gomp_thread_start_data *data = xdata;
68
  struct gomp_thread *thr;
69
  struct gomp_thread_pool *pool;
70
  void (*local_fn) (void *);
71
  void *local_data;
72
 
73
#ifdef HAVE_TLS
74
  thr = &gomp_tls_data;
75
#else
76
  struct gomp_thread local_thr;
77
  thr = &local_thr;
78
  pthread_setspecific (gomp_tls_key, thr);
79
#endif
80
  gomp_sem_init (&thr->release, 0);
81
 
82
  /* Extract what we need from data.  */
83
  local_fn = data->fn;
84
  local_data = data->fn_data;
85
  thr->thread_pool = data->thread_pool;
86
  thr->ts = data->ts;
87
  thr->task = data->task;
88
 
89
  thr->ts.team->ordered_release[thr->ts.team_id] = &thr->release;
90
 
91
  /* Make thread pool local. */
92
  pool = thr->thread_pool;
93
 
94
  if (data->nested)
95
    {
96
      struct gomp_team *team = thr->ts.team;
97
      struct gomp_task *task = thr->task;
98
 
99
      gomp_barrier_wait (&team->barrier);
100
 
101
      local_fn (local_data);
102
      gomp_team_barrier_wait (&team->barrier);
103
      gomp_finish_task (task);
104
      gomp_barrier_wait_last (&team->barrier);
105
    }
106
  else
107
    {
108
      pool->threads[thr->ts.team_id] = thr;
109
 
110
      gomp_barrier_wait (&pool->threads_dock);
111
      do
112
        {
113
          struct gomp_team *team = thr->ts.team;
114
          struct gomp_task *task = thr->task;
115
 
116
          local_fn (local_data);
117
          gomp_team_barrier_wait (&team->barrier);
118
          gomp_finish_task (task);
119
 
120
          gomp_barrier_wait (&pool->threads_dock);
121
 
122
          local_fn = thr->fn;
123
          local_data = thr->data;
124
          thr->fn = NULL;
125
        }
126
      while (local_fn);
127
    }
128
 
129
  gomp_sem_destroy (&thr->release);
130
  return NULL;
131
}
132
 
133
 
134
/* Create a new team data structure.  */
135
 
136
struct gomp_team *
137
gomp_new_team (unsigned nthreads)
138
{
139
  struct gomp_team *team;
140
  size_t size;
141
  int i;
142
 
143
  size = sizeof (*team) + nthreads * (sizeof (team->ordered_release[0])
144
                                      + sizeof (team->implicit_task[0]));
145
  team = gomp_malloc (size);
146
 
147
  team->work_share_chunk = 8;
148
#ifdef HAVE_SYNC_BUILTINS
149
  team->single_count = 0;
150
#else
151
  gomp_mutex_init (&team->work_share_list_free_lock);
152
#endif
153
  gomp_init_work_share (&team->work_shares[0], false, nthreads);
154
  team->work_shares[0].next_alloc = NULL;
155
  team->work_share_list_free = NULL;
156
  team->work_share_list_alloc = &team->work_shares[1];
157
  for (i = 1; i < 7; i++)
158
    team->work_shares[i].next_free = &team->work_shares[i + 1];
159
  team->work_shares[i].next_free = NULL;
160
 
161
  team->nthreads = nthreads;
162
  gomp_barrier_init (&team->barrier, nthreads);
163
 
164
  gomp_sem_init (&team->master_release, 0);
165
  team->ordered_release = (void *) &team->implicit_task[nthreads];
166
  team->ordered_release[0] = &team->master_release;
167
 
168
  gomp_mutex_init (&team->task_lock);
169
  team->task_queue = NULL;
170
  team->task_count = 0;
171
  team->task_running_count = 0;
172
 
173
  return team;
174
}
175
 
176
 
177
/* Free a team data structure.  */
178
 
179
static void
180
free_team (struct gomp_team *team)
181
{
182
  gomp_barrier_destroy (&team->barrier);
183
  gomp_mutex_destroy (&team->task_lock);
184
  free (team);
185
}
186
 
187
/* Allocate and initialize a thread pool. */
188
 
189
static struct gomp_thread_pool *gomp_new_thread_pool (void)
190
{
191
  struct gomp_thread_pool *pool
192
    = gomp_malloc (sizeof(struct gomp_thread_pool));
193
  pool->threads = NULL;
194
  pool->threads_size = 0;
195
  pool->threads_used = 0;
196
  pool->last_team = NULL;
197
  return pool;
198
}
199
 
200
static void
201
gomp_free_pool_helper (void *thread_pool)
202
{
203
  struct gomp_thread_pool *pool
204
    = (struct gomp_thread_pool *) thread_pool;
205
  gomp_barrier_wait_last (&pool->threads_dock);
206
  gomp_sem_destroy (&gomp_thread ()->release);
207
  pthread_exit (NULL);
208
}
209
 
210
/* Free a thread pool and release its threads. */
211
 
212
static void
213
gomp_free_thread (void *arg __attribute__((unused)))
214
{
215
  struct gomp_thread *thr = gomp_thread ();
216
  struct gomp_thread_pool *pool = thr->thread_pool;
217
  if (pool)
218
    {
219
      if (pool->threads_used > 0)
220
        {
221
          int i;
222
          for (i = 1; i < pool->threads_used; i++)
223
            {
224
              struct gomp_thread *nthr = pool->threads[i];
225
              nthr->fn = gomp_free_pool_helper;
226
              nthr->data = pool;
227
            }
228
          /* This barrier undocks threads docked on pool->threads_dock.  */
229
          gomp_barrier_wait (&pool->threads_dock);
230
          /* And this waits till all threads have called gomp_barrier_wait_last
231
             in gomp_free_pool_helper.  */
232
          gomp_barrier_wait (&pool->threads_dock);
233
          /* Now it is safe to destroy the barrier and free the pool.  */
234
          gomp_barrier_destroy (&pool->threads_dock);
235
        }
236
      free (pool->threads);
237
      if (pool->last_team)
238
        free_team (pool->last_team);
239
      free (pool);
240
      thr->thread_pool = NULL;
241
    }
242
  if (thr->task != NULL)
243
    {
244
      struct gomp_task *task = thr->task;
245
      gomp_end_task ();
246
      free (task);
247
    }
248
}
249
 
250
/* Launch a team.  */
251
 
252
void
253
gomp_team_start (void (*fn) (void *), void *data, unsigned nthreads,
254
                 struct gomp_team *team)
255
{
256
  struct gomp_thread_start_data *start_data;
257
  struct gomp_thread *thr, *nthr;
258
  struct gomp_task *task;
259
  struct gomp_task_icv *icv;
260
  bool nested;
261
  struct gomp_thread_pool *pool;
262
  unsigned i, n, old_threads_used = 0;
263
  pthread_attr_t thread_attr, *attr;
264
  unsigned long nthreads_var;
265
 
266
  thr = gomp_thread ();
267
  nested = thr->ts.team != NULL;
268
  if (__builtin_expect (thr->thread_pool == NULL, 0))
269
    {
270
      thr->thread_pool = gomp_new_thread_pool ();
271
      pthread_setspecific (gomp_thread_destructor, thr);
272
    }
273
  pool = thr->thread_pool;
274
  task = thr->task;
275
  icv = task ? &task->icv : &gomp_global_icv;
276
 
277
  /* Always save the previous state, even if this isn't a nested team.
278
     In particular, we should save any work share state from an outer
279
     orphaned work share construct.  */
280
  team->prev_ts = thr->ts;
281
 
282
  thr->ts.team = team;
283
  thr->ts.team_id = 0;
284
  ++thr->ts.level;
285
  if (nthreads > 1)
286
    ++thr->ts.active_level;
287
  thr->ts.work_share = &team->work_shares[0];
288
  thr->ts.last_work_share = NULL;
289
#ifdef HAVE_SYNC_BUILTINS
290
  thr->ts.single_count = 0;
291
#endif
292
  thr->ts.static_trip = 0;
293
  thr->task = &team->implicit_task[0];
294
  nthreads_var = icv->nthreads_var;
295
  if (__builtin_expect (gomp_nthreads_var_list != NULL, 0)
296
      && thr->ts.level < gomp_nthreads_var_list_len)
297
    nthreads_var = gomp_nthreads_var_list[thr->ts.level];
298
  gomp_init_task (thr->task, task, icv);
299
  team->implicit_task[0].icv.nthreads_var = nthreads_var;
300
 
301
  if (nthreads == 1)
302
    return;
303
 
304
  i = 1;
305
 
306
  /* We only allow the reuse of idle threads for non-nested PARALLEL
307
     regions.  This appears to be implied by the semantics of
308
     threadprivate variables, but perhaps that's reading too much into
309
     things.  Certainly it does prevent any locking problems, since
310
     only the initial program thread will modify gomp_threads.  */
311
  if (!nested)
312
    {
313
      old_threads_used = pool->threads_used;
314
 
315
      if (nthreads <= old_threads_used)
316
        n = nthreads;
317
      else if (old_threads_used == 0)
318
        {
319
          n = 0;
320
          gomp_barrier_init (&pool->threads_dock, nthreads);
321
        }
322
      else
323
        {
324
          n = old_threads_used;
325
 
326
          /* Increase the barrier threshold to make sure all new
327
             threads arrive before the team is released.  */
328
          gomp_barrier_reinit (&pool->threads_dock, nthreads);
329
        }
330
 
331
      /* Not true yet, but soon will be.  We're going to release all
332
         threads from the dock, and those that aren't part of the
333
         team will exit.  */
334
      pool->threads_used = nthreads;
335
 
336
      /* Release existing idle threads.  */
337
      for (; i < n; ++i)
338
        {
339
          nthr = pool->threads[i];
340
          nthr->ts.team = team;
341
          nthr->ts.work_share = &team->work_shares[0];
342
          nthr->ts.last_work_share = NULL;
343
          nthr->ts.team_id = i;
344
          nthr->ts.level = team->prev_ts.level + 1;
345
          nthr->ts.active_level = thr->ts.active_level;
346
#ifdef HAVE_SYNC_BUILTINS
347
          nthr->ts.single_count = 0;
348
#endif
349
          nthr->ts.static_trip = 0;
350
          nthr->task = &team->implicit_task[i];
351
          gomp_init_task (nthr->task, task, icv);
352
          team->implicit_task[i].icv.nthreads_var = nthreads_var;
353
          nthr->fn = fn;
354
          nthr->data = data;
355
          team->ordered_release[i] = &nthr->release;
356
        }
357
 
358
      if (i == nthreads)
359
        goto do_release;
360
 
361
      /* If necessary, expand the size of the gomp_threads array.  It is
362
         expected that changes in the number of threads are rare, thus we
363
         make no effort to expand gomp_threads_size geometrically.  */
364
      if (nthreads >= pool->threads_size)
365
        {
366
          pool->threads_size = nthreads + 1;
367
          pool->threads
368
            = gomp_realloc (pool->threads,
369
                            pool->threads_size
370
                            * sizeof (struct gomp_thread_data *));
371
        }
372
    }
373
 
374
  if (__builtin_expect (nthreads > old_threads_used, 0))
375
    {
376
      long diff = (long) nthreads - (long) old_threads_used;
377
 
378
      if (old_threads_used == 0)
379
        --diff;
380
 
381
#ifdef HAVE_SYNC_BUILTINS
382
      __sync_fetch_and_add (&gomp_managed_threads, diff);
383
#else
384
      gomp_mutex_lock (&gomp_remaining_threads_lock);
385
      gomp_managed_threads += diff;
386
      gomp_mutex_unlock (&gomp_remaining_threads_lock);
387
#endif
388
    }
389
 
390
  attr = &gomp_thread_attr;
391
  if (__builtin_expect (gomp_cpu_affinity != NULL, 0))
392
    {
393
      size_t stacksize;
394
      pthread_attr_init (&thread_attr);
395
      pthread_attr_setdetachstate (&thread_attr, PTHREAD_CREATE_DETACHED);
396
      if (! pthread_attr_getstacksize (&gomp_thread_attr, &stacksize))
397
        pthread_attr_setstacksize (&thread_attr, stacksize);
398
      attr = &thread_attr;
399
    }
400
 
401
  start_data = gomp_alloca (sizeof (struct gomp_thread_start_data)
402
                            * (nthreads-i));
403
 
404
  /* Launch new threads.  */
405
  for (; i < nthreads; ++i, ++start_data)
406
    {
407
      pthread_t pt;
408
      int err;
409
 
410
      start_data->fn = fn;
411
      start_data->fn_data = data;
412
      start_data->ts.team = team;
413
      start_data->ts.work_share = &team->work_shares[0];
414
      start_data->ts.last_work_share = NULL;
415
      start_data->ts.team_id = i;
416
      start_data->ts.level = team->prev_ts.level + 1;
417
      start_data->ts.active_level = thr->ts.active_level;
418
#ifdef HAVE_SYNC_BUILTINS
419
      start_data->ts.single_count = 0;
420
#endif
421
      start_data->ts.static_trip = 0;
422
      start_data->task = &team->implicit_task[i];
423
      gomp_init_task (start_data->task, task, icv);
424
      team->implicit_task[i].icv.nthreads_var = nthreads_var;
425
      start_data->thread_pool = pool;
426
      start_data->nested = nested;
427
 
428
      if (gomp_cpu_affinity != NULL)
429
        gomp_init_thread_affinity (attr);
430
 
431
      err = pthread_create (&pt, attr, gomp_thread_start, start_data);
432
      if (err != 0)
433
        gomp_fatal ("Thread creation failed: %s", strerror (err));
434
    }
435
 
436
  if (__builtin_expect (gomp_cpu_affinity != NULL, 0))
437
    pthread_attr_destroy (&thread_attr);
438
 
439
 do_release:
440
  gomp_barrier_wait (nested ? &team->barrier : &pool->threads_dock);
441
 
442
  /* Decrease the barrier threshold to match the number of threads
443
     that should arrive back at the end of this team.  The extra
444
     threads should be exiting.  Note that we arrange for this test
445
     to never be true for nested teams.  */
446
  if (__builtin_expect (nthreads < old_threads_used, 0))
447
    {
448
      long diff = (long) nthreads - (long) old_threads_used;
449
 
450
      gomp_barrier_reinit (&pool->threads_dock, nthreads);
451
 
452
#ifdef HAVE_SYNC_BUILTINS
453
      __sync_fetch_and_add (&gomp_managed_threads, diff);
454
#else
455
      gomp_mutex_lock (&gomp_remaining_threads_lock);
456
      gomp_managed_threads += diff;
457
      gomp_mutex_unlock (&gomp_remaining_threads_lock);
458
#endif
459
    }
460
}
461
 
462
 
463
/* Terminate the current team.  This is only to be called by the master
464
   thread.  We assume that we must wait for the other threads.  */
465
 
466
void
467
gomp_team_end (void)
468
{
469
  struct gomp_thread *thr = gomp_thread ();
470
  struct gomp_team *team = thr->ts.team;
471
 
472
  /* This barrier handles all pending explicit threads.  */
473
  gomp_team_barrier_wait (&team->barrier);
474
  gomp_fini_work_share (thr->ts.work_share);
475
 
476
  gomp_end_task ();
477
  thr->ts = team->prev_ts;
478
 
479
  if (__builtin_expect (thr->ts.team != NULL, 0))
480
    {
481
#ifdef HAVE_SYNC_BUILTINS
482
      __sync_fetch_and_add (&gomp_managed_threads, 1L - team->nthreads);
483
#else
484
      gomp_mutex_lock (&gomp_remaining_threads_lock);
485
      gomp_managed_threads -= team->nthreads - 1L;
486
      gomp_mutex_unlock (&gomp_remaining_threads_lock);
487
#endif
488
      /* This barrier has gomp_barrier_wait_last counterparts
489
         and ensures the team can be safely destroyed.  */
490
      gomp_barrier_wait (&team->barrier);
491
    }
492
 
493
  if (__builtin_expect (team->work_shares[0].next_alloc != NULL, 0))
494
    {
495
      struct gomp_work_share *ws = team->work_shares[0].next_alloc;
496
      do
497
        {
498
          struct gomp_work_share *next_ws = ws->next_alloc;
499
          free (ws);
500
          ws = next_ws;
501
        }
502
      while (ws != NULL);
503
    }
504
  gomp_sem_destroy (&team->master_release);
505
#ifndef HAVE_SYNC_BUILTINS
506
  gomp_mutex_destroy (&team->work_share_list_free_lock);
507
#endif
508
 
509
  if (__builtin_expect (thr->ts.team != NULL, 0)
510
      || __builtin_expect (team->nthreads == 1, 0))
511
    free_team (team);
512
  else
513
    {
514
      struct gomp_thread_pool *pool = thr->thread_pool;
515
      if (pool->last_team)
516
        free_team (pool->last_team);
517
      pool->last_team = team;
518
    }
519
}
520
 
521
 
522
/* Constructors for this file.  */
523
 
524
static void __attribute__((constructor))
525
initialize_team (void)
526
{
527
  struct gomp_thread *thr;
528
 
529
#ifndef HAVE_TLS
530
  static struct gomp_thread initial_thread_tls_data;
531
 
532
  pthread_key_create (&gomp_tls_key, NULL);
533
  pthread_setspecific (gomp_tls_key, &initial_thread_tls_data);
534
#endif
535
 
536
  if (pthread_key_create (&gomp_thread_destructor, gomp_free_thread) != 0)
537
    gomp_fatal ("could not create thread pool destructor.");
538
 
539
#ifdef HAVE_TLS
540
  thr = &gomp_tls_data;
541
#else
542
  thr = &initial_thread_tls_data;
543
#endif
544
  gomp_sem_init (&thr->release, 0);
545
}
546
 
547
static void __attribute__((destructor))
548
team_destructor (void)
549
{
550
  /* Without this dlclose on libgomp could lead to subsequent
551
     crashes.  */
552
  pthread_key_delete (gomp_thread_destructor);
553
}
554
 
555
struct gomp_task_icv *
556
gomp_new_icv (void)
557
{
558
  struct gomp_thread *thr = gomp_thread ();
559
  struct gomp_task *task = gomp_malloc (sizeof (struct gomp_task));
560
  gomp_init_task (task, NULL, &gomp_global_icv);
561
  thr->task = task;
562
  pthread_setspecific (gomp_thread_destructor, thr);
563
  return &task->icv;
564
}

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.