OpenCores
URL https://opencores.org/ocsvn/openrisc/openrisc/trunk

Subversion Repositories openrisc

[/] [openrisc/] [trunk/] [gnu-dev/] [or1k-gcc/] [boehm-gc/] [include/] [private/] [specific.h] - Blame information for rev 830

Go to most recent revision | Details | Compare with Previous | View Log

Line No. Rev Author Line
1 721 jeremybenn
/*
2
 * This is a reimplementation of a subset of the pthread_getspecific/setspecific
3
 * interface. This appears to outperform the standard linuxthreads one
4
 * by a significant margin.
5
 * The major restriction is that each thread may only make a single
6
 * pthread_setspecific call on a single key.  (The current data structure
7
 * doesn't really require that.  The restriction should be easily removable.)
8
 * We don't currently support the destruction functions, though that
9
 * could be done.
10
 * We also currently assume that only one pthread_setspecific call
11
 * can be executed at a time, though that assumption would be easy to remove
12
 * by adding a lock.
13
 */
14
 
15
#include <errno.h>
16
 
17
/* Called during key creation or setspecific.           */
18
/* For the GC we already hold lock.                     */
19
/* Currently allocated objects leak on thread exit.     */
20
/* That's hard to fix, but OK if we allocate garbage    */
21
/* collected memory.                                    */
22
#define MALLOC_CLEAR(n) GC_INTERNAL_MALLOC(n, NORMAL)
23
#define PREFIXED(name) GC_##name
24
 
25
#define TS_CACHE_SIZE 1024
26
#define CACHE_HASH(n) (((((long)n) >> 8) ^ (long)n) & (TS_CACHE_SIZE - 1))
27
#define TS_HASH_SIZE 1024
28
#define HASH(n) (((((long)n) >> 8) ^ (long)n) & (TS_HASH_SIZE - 1))
29
 
30
/* An entry describing a thread-specific value for a given thread.      */
31
/* All such accessible structures preserve the invariant that if either */
32
/* thread is a valid pthread id or qtid is a valid "quick tread id"     */
33
/* for a thread, then value holds the corresponding thread specific     */
34
/* value.  This invariant must be preserved at ALL times, since         */
35
/* asynchronous reads are allowed.                                      */
36
typedef struct thread_specific_entry {
37
        unsigned long qtid;     /* quick thread id, only for cache */
38
        void * value;
39
        struct thread_specific_entry *next;
40
        pthread_t thread;
41
} tse;
42
 
43
 
44
/* We represent each thread-specific datum as two tables.  The first is */
45
/* a cache, indexed by a "quick thread identifier".  The "quick" thread */
46
/* identifier is an easy to compute value, which is guaranteed to       */
47
/* determine the thread, though a thread may correspond to more than    */
48
/* one value.  We typically use the address of a page in the stack.     */
49
/* The second is a hash table, indexed by pthread_self().  It is used   */
50
/* only as a backup.                                                    */
51
 
52
/* Return the "quick thread id".  Default version.  Assumes page size,  */
53
/* or at least thread stack separation, is at least 4K.                 */
54
/* Must be defined so that it never returns 0.  (Page 0 can't really    */
55
/* be part of any stack, since that would make 0 a valid stack pointer.)*/
56
static __inline__ unsigned long quick_thread_id() {
57
    int dummy;
58
    return (unsigned long)(&dummy) >> 12;
59
}
60
 
61
#define INVALID_QTID ((unsigned long)0)
62
#define INVALID_THREADID ((pthread_t)0)
63
 
64
typedef struct thread_specific_data {
65
    tse * volatile cache[TS_CACHE_SIZE];
66
                        /* A faster index to the hash table */
67
    tse * hash[TS_HASH_SIZE];
68
    pthread_mutex_t lock;
69
} tsd;
70
 
71
typedef tsd * PREFIXED(key_t);
72
 
73
extern int PREFIXED(key_create) (tsd ** key_ptr, void (* destructor)(void *));
74
 
75
extern int PREFIXED(setspecific) (tsd * key, void * value);
76
 
77
extern void PREFIXED(remove_specific) (tsd * key);
78
 
79
/* An internal version of getspecific that assumes a cache miss.        */
80
void * PREFIXED(slow_getspecific) (tsd * key, unsigned long qtid,
81
                                   tse * volatile * cache_entry);
82
 
83
static __inline__ void * PREFIXED(getspecific) (tsd * key) {
84
    long qtid = quick_thread_id();
85
    unsigned hash_val = CACHE_HASH(qtid);
86
    tse * volatile * entry_ptr = key -> cache + hash_val;
87
    tse * entry = *entry_ptr;   /* Must be loaded only once.    */
88
    if (EXPECT(entry -> qtid == qtid, 1)) {
89
      GC_ASSERT(entry -> thread == pthread_self());
90
      return entry -> value;
91
    }
92
    return PREFIXED(slow_getspecific) (key, qtid, entry_ptr);
93
}
94
 
95
 

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.