URL
https://opencores.org/ocsvn/openrisc/openrisc/trunk
Subversion Repositories openrisc
[/] [openrisc/] [trunk/] [gnu-dev/] [or1k-gcc/] [libitm/] [config/] [linux/] [rwlock.cc] - Rev 834
Go to most recent revision | Compare with Previous | Blame | View Log
/* Copyright (C) 2011 Free Software Foundation, Inc. Contributed by Torvald Riegel <triegel@redhat.com>. This file is part of the GNU Transactional Memory Library (libitm). Libitm is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3 of the License, or (at your option) any later version. Libitm is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. Under Section 7 of GPL version 3, you are granted additional permissions described in the GCC Runtime Library Exception, version 3.1, as published by the Free Software Foundation. You should have received a copy of the GNU General Public License and a copy of the GCC Runtime Library Exception along with this program; see the files COPYING3 and COPYING.RUNTIME respectively. If not, see <http://www.gnu.org/licenses/>. */ #include "libitm_i.h" #include "futex.h" #include <limits.h> namespace GTM HIDDEN { // Acquire a RW lock for reading. void gtm_rwlock::read_lock (gtm_thread *tx) { for (;;) { // Fast path: first announce our intent to read, then check for // conflicting intents to write. The fence ensures that this happens // in exactly this order. tx->shared_state.store (0, memory_order_relaxed); atomic_thread_fence (memory_order_seq_cst); if (likely (writers.load (memory_order_relaxed) == 0)) return; // There seems to be an active, waiting, or confirmed writer, so enter // the futex-based slow path. // Before waiting, we clear our read intent check whether there are any // writers that might potentially wait for readers. If so, wake them. // We need the barrier here for the same reason that we need it in // read_unlock(). // TODO Potentially too many wake-ups. See comments in read_unlock(). tx->shared_state.store (-1, memory_order_relaxed); atomic_thread_fence (memory_order_seq_cst); if (writer_readers.load (memory_order_relaxed) > 0) { writer_readers.store (0, memory_order_relaxed); futex_wake(&writer_readers, 1); } // Signal that there are waiting readers and wait until there is no // writer anymore. // TODO Spin here on writers for a while. Consider whether we woke // any writers before? while (writers.load (memory_order_relaxed)) { // An active writer. Wait until it has finished. To avoid lost // wake-ups, we need to use Dekker-like synchronization. // Note that we cannot reset readers to zero when we see that there // are no writers anymore after the barrier because this pending // store could then lead to lost wake-ups at other readers. readers.store (1, memory_order_relaxed); atomic_thread_fence (memory_order_seq_cst); if (writers.load (memory_order_relaxed)) futex_wait(&readers, 1); } // And we try again to acquire a read lock. } } // Acquire a RW lock for writing. Generic version that also works for // upgrades. // Note that an upgrade might fail (and thus waste previous work done during // this transaction) if there is another thread that tried to go into serial // mode earlier (i.e., upgrades do not have higher priority than pure writers). // However, this seems rare enough to not consider it further as we need both // a non-upgrade writer and a writer to happen to switch to serial mode // concurrently. If we'd want to handle this, a writer waiting for readers // would have to coordinate with later arriving upgrades and hand over the // lock to them, including the the reader-waiting state. We can try to support // this if this will actually happen often enough in real workloads. bool gtm_rwlock::write_lock_generic (gtm_thread *tx) { // Try to acquire the write lock. int w = 0; if (unlikely (!writers.compare_exchange_strong (w, 1))) { // If this is an upgrade, we must not wait for other writers or // upgrades. if (tx != 0) return false; // There is already a writer. If there are no other waiting writers, // switch to contended mode. We need seq_cst memory order to make the // Dekker-style synchronization work. if (w != 2) w = writers.exchange (2); while (w != 0) { futex_wait(&writers, 2); w = writers.exchange (2); } } // We have acquired the writer side of the R/W lock. Now wait for any // readers that might still be active. // We don't need an extra barrier here because the CAS and the xchg // operations have full barrier semantics already. // TODO In the worst case, this requires one wait/wake pair for each // active reader. Reduce this! for (gtm_thread *it = gtm_thread::list_of_threads; it != 0; it = it->next_thread) { if (it == tx) continue; // Use a loop here to check reader flags again after waiting. while (it->shared_state.load (memory_order_relaxed) != ~(typeof it->shared_state)0) { // An active reader. Wait until it has finished. To avoid lost // wake-ups, we need to use Dekker-like synchronization. // Note that we can reset writer_readers to zero when we see after // the barrier that the reader has finished in the meantime; // however, this is only possible because we are the only writer. // TODO Spin for a while on this reader flag. writer_readers.store (1, memory_order_relaxed); atomic_thread_fence (memory_order_seq_cst); if (it->shared_state.load (memory_order_relaxed) != ~(typeof it->shared_state)0) futex_wait(&writer_readers, 1); else writer_readers.store (0, memory_order_relaxed); } } return true; } // Acquire a RW lock for writing. void gtm_rwlock::write_lock () { write_lock_generic (0); } // Upgrade a RW lock that has been locked for reading to a writing lock. // Do this without possibility of another writer incoming. Return false // if this attempt fails (i.e. another thread also upgraded). bool gtm_rwlock::write_upgrade (gtm_thread *tx) { return write_lock_generic (tx); } // Has to be called iff the previous upgrade was successful and after it is // safe for the transaction to not be marked as a reader anymore. void gtm_rwlock::write_upgrade_finish (gtm_thread *tx) { // We are not a reader anymore. This is only safe to do after we have // acquired the writer lock. tx->shared_state.store (-1, memory_order_release); } // Release a RW lock from reading. void gtm_rwlock::read_unlock (gtm_thread *tx) { // We only need release memory order here because of privatization safety // (this ensures that marking the transaction as inactive happens after // any prior data accesses by this transaction, and that neither the // compiler nor the hardware order this store earlier). // ??? We might be able to avoid this release here if the compiler can't // merge the release fence with the subsequent seq_cst fence. tx->shared_state.store (-1, memory_order_release); // If there is a writer waiting for readers, wake it up. We need the fence // to avoid lost wake-ups. Furthermore, the privatization safety // implementation in gtm_thread::try_commit() relies on the existence of // this seq_cst fence. // ??? We might not be the last active reader, so the wake-up might happen // too early. How do we avoid this without slowing down readers too much? // Each reader could scan the list of txns for other active readers but // this can result in many cache misses. Use combining instead? // TODO Sends out one wake-up for each reader in the worst case. atomic_thread_fence (memory_order_seq_cst); if (unlikely (writer_readers.load (memory_order_relaxed) > 0)) { // No additional barrier needed here (see write_unlock()). writer_readers.store (0, memory_order_relaxed); futex_wake(&writer_readers, 1); } } // Release a RW lock from writing. void gtm_rwlock::write_unlock () { // This needs to have seq_cst memory order. if (writers.fetch_sub (1) == 2) { // There might be waiting writers, so wake them. writers.store (0, memory_order_relaxed); if (futex_wake(&writers, 1) == 0) { // If we did not wake any waiting writers, we might indeed be the // last writer (this can happen because write_lock_generic() // exchanges 0 or 1 to 2 and thus might go to contended mode even if // no other thread holds the write lock currently). Therefore, we // have to wake up readers here as well. Execute a barrier after // the previous relaxed reset of writers (Dekker-style), and fall // through to the normal reader wake-up code. atomic_thread_fence (memory_order_seq_cst); } else return; } // No waiting writers, so wake up all waiting readers. // Because the fetch_and_sub is a full barrier already, we don't need // another barrier here (as in read_unlock()). if (readers.load (memory_order_relaxed) > 0) { // No additional barrier needed here. The previous load must be in // modification order because of the coherency constraints. Late stores // by a reader are not a problem because readers do Dekker-style // synchronization on writers. readers.store (0, memory_order_relaxed); futex_wake(&readers, INT_MAX); } } } // namespace GTM
Go to most recent revision | Compare with Previous | Blame | View Log