OpenCores
URL https://opencores.org/ocsvn/or1k_soc_on_altera_embedded_dev_kit/or1k_soc_on_altera_embedded_dev_kit/trunk

Subversion Repositories or1k_soc_on_altera_embedded_dev_kit

[/] [or1k_soc_on_altera_embedded_dev_kit/] [trunk/] [linux-2.6/] [linux-2.6.24/] [arch/] [mips/] [kernel/] [semaphore.c] - Blame information for rev 3

Details | Compare with Previous | View Log

Line No. Rev Author Line
1 3 xianfeng
/*
2
 * MIPS-specific semaphore code.
3
 *
4
 * Copyright (C) 1999 Cort Dougan <cort@cs.nmt.edu>
5
 * Copyright (C) 2004 Ralf Baechle <ralf@linux-mips.org>
6
 *
7
 * This program is free software; you can redistribute it and/or
8
 * modify it under the terms of the GNU General Public License
9
 * as published by the Free Software Foundation; either version
10
 * 2 of the License, or (at your option) any later version.
11
 *
12
 * April 2001 - Reworked by Paul Mackerras <paulus@samba.org>
13
 * to eliminate the SMP races in the old version between the updates
14
 * of `count' and `waking'.  Now we use negative `count' values to
15
 * indicate that some process(es) are waiting for the semaphore.
16
 */
17
 
18
#include <linux/module.h>
19
#include <linux/sched.h>
20
#include <linux/init.h>
21
#include <asm/atomic.h>
22
#include <asm/cpu-features.h>
23
#include <asm/errno.h>
24
#include <asm/semaphore.h>
25
#include <asm/war.h>
26
/*
27
 * Atomically update sem->count.
28
 * This does the equivalent of the following:
29
 *
30
 *      old_count = sem->count;
31
 *      tmp = MAX(old_count, 0) + incr;
32
 *      sem->count = tmp;
33
 *      return old_count;
34
 *
35
 * On machines without lld/scd we need a spinlock to make the manipulation of
36
 * sem->count and sem->waking atomic.  Scalability isn't an issue because
37
 * this lock is used on UP only so it's just an empty variable.
38
 */
39
static inline int __sem_update_count(struct semaphore *sem, int incr)
40
{
41
        int old_count, tmp;
42
 
43
        if (cpu_has_llsc && R10000_LLSC_WAR) {
44
                __asm__ __volatile__(
45
                "       .set    mips3                                   \n"
46
                "1:     ll      %0, %2          # __sem_update_count    \n"
47
                "       sra     %1, %0, 31                              \n"
48
                "       not     %1                                      \n"
49
                "       and     %1, %0, %1                              \n"
50
                "       addu    %1, %1, %3                              \n"
51
                "       sc      %1, %2                                  \n"
52
                "       beqzl   %1, 1b                                  \n"
53
                "       .set    mips0                                   \n"
54
                : "=&r" (old_count), "=&r" (tmp), "=m" (sem->count)
55
                : "r" (incr), "m" (sem->count));
56
        } else if (cpu_has_llsc) {
57
                __asm__ __volatile__(
58
                "       .set    mips3                                   \n"
59
                "1:     ll      %0, %2          # __sem_update_count    \n"
60
                "       sra     %1, %0, 31                              \n"
61
                "       not     %1                                      \n"
62
                "       and     %1, %0, %1                              \n"
63
                "       addu    %1, %1, %3                              \n"
64
                "       sc      %1, %2                                  \n"
65
                "       beqz    %1, 1b                                  \n"
66
                "       .set    mips0                                   \n"
67
                : "=&r" (old_count), "=&r" (tmp), "=m" (sem->count)
68
                : "r" (incr), "m" (sem->count));
69
        } else {
70
                static DEFINE_SPINLOCK(semaphore_lock);
71
                unsigned long flags;
72
 
73
                spin_lock_irqsave(&semaphore_lock, flags);
74
                old_count = atomic_read(&sem->count);
75
                tmp = max_t(int, old_count, 0) + incr;
76
                atomic_set(&sem->count, tmp);
77
                spin_unlock_irqrestore(&semaphore_lock, flags);
78
        }
79
 
80
        return old_count;
81
}
82
 
83
void __up(struct semaphore *sem)
84
{
85
        /*
86
         * Note that we incremented count in up() before we came here,
87
         * but that was ineffective since the result was <= 0, and
88
         * any negative value of count is equivalent to 0.
89
         * This ends up setting count to 1, unless count is now > 0
90
         * (i.e. because some other cpu has called up() in the meantime),
91
         * in which case we just increment count.
92
         */
93
        __sem_update_count(sem, 1);
94
        wake_up(&sem->wait);
95
}
96
 
97
EXPORT_SYMBOL(__up);
98
 
99
/*
100
 * Note that when we come in to __down or __down_interruptible,
101
 * we have already decremented count, but that decrement was
102
 * ineffective since the result was < 0, and any negative value
103
 * of count is equivalent to 0.
104
 * Thus it is only when we decrement count from some value > 0
105
 * that we have actually got the semaphore.
106
 */
107
void __sched __down(struct semaphore *sem)
108
{
109
        struct task_struct *tsk = current;
110
        DECLARE_WAITQUEUE(wait, tsk);
111
 
112
        __set_task_state(tsk, TASK_UNINTERRUPTIBLE);
113
        add_wait_queue_exclusive(&sem->wait, &wait);
114
 
115
        /*
116
         * Try to get the semaphore.  If the count is > 0, then we've
117
         * got the semaphore; we decrement count and exit the loop.
118
         * If the count is 0 or negative, we set it to -1, indicating
119
         * that we are asleep, and then sleep.
120
         */
121
        while (__sem_update_count(sem, -1) <= 0) {
122
                schedule();
123
                set_task_state(tsk, TASK_UNINTERRUPTIBLE);
124
        }
125
        remove_wait_queue(&sem->wait, &wait);
126
        __set_task_state(tsk, TASK_RUNNING);
127
 
128
        /*
129
         * If there are any more sleepers, wake one of them up so
130
         * that it can either get the semaphore, or set count to -1
131
         * indicating that there are still processes sleeping.
132
         */
133
        wake_up(&sem->wait);
134
}
135
 
136
EXPORT_SYMBOL(__down);
137
 
138
int __sched __down_interruptible(struct semaphore * sem)
139
{
140
        int retval = 0;
141
        struct task_struct *tsk = current;
142
        DECLARE_WAITQUEUE(wait, tsk);
143
 
144
        __set_task_state(tsk, TASK_INTERRUPTIBLE);
145
        add_wait_queue_exclusive(&sem->wait, &wait);
146
 
147
        while (__sem_update_count(sem, -1) <= 0) {
148
                if (signal_pending(current)) {
149
                        /*
150
                         * A signal is pending - give up trying.
151
                         * Set sem->count to 0 if it is negative,
152
                         * since we are no longer sleeping.
153
                         */
154
                        __sem_update_count(sem, 0);
155
                        retval = -EINTR;
156
                        break;
157
                }
158
                schedule();
159
                set_task_state(tsk, TASK_INTERRUPTIBLE);
160
        }
161
        remove_wait_queue(&sem->wait, &wait);
162
        __set_task_state(tsk, TASK_RUNNING);
163
 
164
        wake_up(&sem->wait);
165
        return retval;
166
}
167
 
168
EXPORT_SYMBOL(__down_interruptible);

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.