1
0
Fork 0
arangodb/lib/Basics/ReadWriteSpinLock.h

155 lines
5.2 KiB
C++

////////////////////////////////////////////////////////////////////////////////
/// DISCLAIMER
///
/// Copyright 2014-2018 ArangoDB GmbH, Cologne, Germany
/// Copyright 2004-2014 triAGENS GmbH, Cologne, Germany
///
/// Licensed under the Apache License, Version 2.0 (the "License");
/// you may not use this file except in compliance with the License.
/// You may obtain a copy of the License at
///
/// http://www.apache.org/licenses/LICENSE-2.0
///
/// Unless required by applicable law or agreed to in writing, software
/// distributed under the License is distributed on an "AS IS" BASIS,
/// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
/// See the License for the specific language governing permissions and
/// limitations under the License.
///
/// Copyright holder is ArangoDB GmbH, Cologne, Germany
///
/// @author Daniel H. Larkin
////////////////////////////////////////////////////////////////////////////////
#ifndef ARANGO_READ_WRITE_SPIN_LOCK_H
#define ARANGO_READ_WRITE_SPIN_LOCK_H 1
#include "Basics/Common.h"
#include "Basics/cpu-relax.h"
#include <atomic>
namespace arangodb {
namespace basics {
class ReadWriteSpinLock {
public:
ReadWriteSpinLock() : _state(0) {}
// TODO: remove these - copyable locks are potentially flawed!
ReadWriteSpinLock(ReadWriteSpinLock const& other)
{
auto val = other._state.load(std::memory_order_relaxed);
TRI_ASSERT(val == 0);
_state.store(val, std::memory_order_relaxed);
}
ReadWriteSpinLock& operator=(ReadWriteSpinLock const& other) {
auto val = other._state.load(std::memory_order_relaxed);
TRI_ASSERT(val == 0);
val = _state.exchange(val, std::memory_order_relaxed);
TRI_ASSERT(val == 0);
return *this;
}
bool tryWriteLock() {
// order_relaxed is an optimization, cmpxchg will synchronize side-effects
auto state = _state.load(std::memory_order_relaxed);
// try to acquire write lock as long as no readers or writers are active,
// we might "overtake" other queued writers though.
while ((state & ~QUEUED_WRITER_MASK) == 0) {
if (_state.compare_exchange_weak(state, state | WRITE_LOCK, std::memory_order_acquire)) {
return true; // we successfully acquired the write lock!
}
}
return false;
}
bool writeLock(uint64_t maxAttempts = UINT64_MAX) {
if (tryWriteLock()) {
return true;
}
uint64_t attempts = 0;
// the lock is either hold by another writer or we have active readers
// -> announce that we want to write
auto state = _state.fetch_add(QUEUED_WRITER_INC, std::memory_order_relaxed);
while (++attempts < maxAttempts) {
while ((state & ~QUEUED_WRITER_MASK) == 0) {
// try to acquire lock and perform queued writer decrement in one step
if (_state.compare_exchange_weak(state, (state - QUEUED_WRITER_INC) | WRITE_LOCK, std::memory_order_acquire)) {
return true;
}
if (++attempts > maxAttempts) {
return false;
}
}
cpu_relax();
state = _state.load(std::memory_order_relaxed);
}
return false;
}
bool tryReadLock() {
// order_relaxed is an optimization, cmpxchg will synchronize side-effects
auto state = _state.load(std::memory_order_relaxed);
// try to acquire read lock as long as no writers are active or queued
while ((state & ~READER_MASK) == 0) {
if (_state.compare_exchange_weak(state, state + READER_INC, std::memory_order_acquire)) {
return true;
}
}
return false;
}
bool readLock(uint64_t maxAttempts = UINT64_MAX) {
uint64_t attempts = 0;
while (attempts++ < maxAttempts) {
if (tryReadLock()) {
return true;
}
cpu_relax();
}
return false;
}
void readUnlock() { unlockRead(); }
void unlockRead() {
_state.fetch_sub(READER_INC, std::memory_order_release);
}
void writeUnlock() { unlockWrite(); }
void unlockWrite() {
_state.fetch_sub(WRITE_LOCK, std::memory_order_release);
}
bool isLocked() const { return (_state.load(std::memory_order_relaxed) & ~QUEUED_WRITER_MASK) != 0; }
bool isWriteLocked() const { return _state.load(std::memory_order_relaxed) & WRITE_LOCK; }
private:
std::atomic<uint32_t> _state;
static const unsigned WRITE_LOCK = 1;
static const unsigned READER_INC = 1 << 16;
static const unsigned READER_MASK = ~(READER_INC - 1);
static const unsigned QUEUED_WRITER_INC = 1 << 1;
static const unsigned QUEUED_WRITER_MASK = (READER_INC - 1) & ~WRITE_LOCK;
static_assert((READER_MASK & WRITE_LOCK) == 0, "READER_MASK and WRITE_LOCK conflict");
static_assert((READER_MASK & QUEUED_WRITER_MASK) == 0, "READER_MASK and QUEUED_WRITER_MASK conflict");
static_assert((QUEUED_WRITER_MASK & WRITE_LOCK) == 0, "QUEUED_WRITER_MASK and WRITE_LOCK conflict");
static_assert((READER_MASK & READER_INC) != 0 &&
(READER_MASK & (READER_INC >> 1)) == 0,
"READER_INC must be first bit in READER_MASK");
static_assert((QUEUED_WRITER_MASK & QUEUED_WRITER_INC) != 0 &&
(QUEUED_WRITER_MASK & (QUEUED_WRITER_INC >> 1)) == 0,
"QUEUED_WRITER_INC must be first bit in QUEUED_WRITER_MASK");
};
} // namespace basics
} // namespace arangodb
#endif