1
0
Fork 0
arangodb/arangod/RestServer/FlushFeature.cpp

833 lines
26 KiB
C++

////////////////////////////////////////////////////////////////////////////////
/// DISCLAIMER
///
/// Copyright 2016 ArangoDB GmbH, Cologne, Germany
///
/// Licensed under the Apache License, Version 2.0 (the "License");
/// you may not use this file except in compliance with the License.
/// You may obtain a copy of the License at
///
/// http://www.apache.org/licenses/LICENSE-2.0
///
/// Unless required by applicable law or agreed to in writing, software
/// distributed under the License is distributed on an "AS IS" BASIS,
/// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
/// See the License for the specific language governing permissions and
/// limitations under the License.
///
/// Copyright holder is ArangoDB GmbH, Cologne, Germany
///
/// @author Jan Steemann
////////////////////////////////////////////////////////////////////////////////
#include <rocksdb/utilities/transaction_db.h>
#include "FlushFeature.h"
#include "ApplicationFeatures/ApplicationServer.h"
#include "Aql/QueryCache.h"
#include "Basics/encoding.h"
#include "Basics/ReadLocker.h"
#include "Basics/RocksDBUtils.h"
#include "Basics/WriteLocker.h"
#include "Cluster/ServerState.h"
#include "Logger/Logger.h"
#include "MMFiles/MMFilesDatafile.h"
#include "MMFiles/MMFilesEngine.h"
#include "MMFiles/MMFilesLogfileManager.h"
#include "MMFiles/MMFilesWalMarker.h"
#include "ProgramOptions/ProgramOptions.h"
#include "ProgramOptions/Section.h"
#include "RestServer/DatabaseFeature.h"
#include "RocksDBEngine/RocksDBEngine.h"
#include "RocksDBEngine/RocksDBFormat.h"
#include "RocksDBEngine/RocksDBLogValue.h"
#include "RocksDBEngine/RocksDBRecoveryHelper.h"
#include "StorageEngine/EngineSelectorFeature.h"
#include "Utils/FlushThread.h"
namespace {
const std::string DATA_ATTRIBUTE("data"); // attribute inside the flush marker storing custom data body
const std::string TYPE_ATTRIBUTE("type"); // attribute inside the flush marker storing custom data type
/// @brief base class for FlushSubscription implementations
class FlushSubscriptionBase : public arangodb::FlushFeature::FlushSubscription {
public:
virtual arangodb::Result commit(VPackSlice data, TRI_voc_tick_t tick) override final {
if (data.isNone()) {
// upgrade tick without commiting actual marker
resetCurrentTick(tick);
return {};
}
return commitImpl(data, tick);
}
/// @brief earliest tick that can be released
TRI_voc_tick_t tick() const noexcept final {
return _tickPrevious.load(std::memory_order_acquire);
}
protected:
FlushSubscriptionBase(std::string const& type, TRI_voc_tick_t databaseId)
: _databaseId(databaseId),
_tickCurrent(0), // default (smallest) tick for StorageEngine
_tickPrevious(0), // default (smallest) tick for StorageEngine
_type(type) {
}
void resetCurrentTick(TRI_voc_tick_t tick) noexcept {
// the whole method isn't intended to be atomic, only
// '_tickPrevious' can be accessed from 2 different
// threads concurrently:
// - FlushThread (consumer)
// - IResearchLink commit thread (producer)
_tickPrevious.store(_tickCurrent, std::memory_order_release);
_tickCurrent = tick;
}
virtual arangodb::Result commitImpl(VPackSlice data, TRI_voc_tick_t tick) = 0;
TRI_voc_tick_t const _databaseId;
TRI_voc_tick_t _tickCurrent; // last successful tick, should be replayed
std::atomic<TRI_voc_tick_t> _tickPrevious; // previous successful tick, should be replayed
std::string const _type;
};
// wrap vector inside a static function to ensure proper initialization order
std::unordered_map<std::string, arangodb::FlushFeature::FlushRecoveryCallback>& getFlushRecoveryCallbacks() {
static std::unordered_map<std::string, arangodb::FlushFeature::FlushRecoveryCallback> callbacks;
return callbacks;
}
arangodb::Result applyRecoveryMarker(
TRI_voc_tick_t dbId,
arangodb::velocypack::Slice const& slice
) {
if (!slice.isObject()) {
return arangodb::Result(
TRI_ERROR_INTERNAL,
"failed to find a JSON object inside the supplied "
"marker body while applying 'Flush' recovery marker");
}
if (!slice.hasKey(TYPE_ATTRIBUTE)) {
return arangodb::Result(
TRI_ERROR_INTERNAL,
"failed to find a 'type' attribute inside the supplied "
"marker body while applying 'Flush' recovery marker");
}
auto typeSlice = slice.get(TYPE_ATTRIBUTE);
if (!typeSlice.isString()) {
return arangodb::Result(
TRI_ERROR_INTERNAL,
"failed to find a string 'type' attribute inside the supplied "
"marker body while applying 'Flush' recovery marker");
}
auto type = typeSlice.copyString();
auto& callbacks = getFlushRecoveryCallbacks();
auto itr = callbacks.find(type);
if (itr == callbacks.end()) {
return arangodb::Result(
TRI_ERROR_INTERNAL,
"failed to find handler while applying 'Flush' recovery marker of type '" + type + "'");
}
TRI_ASSERT(itr->second); // non-nullptr ensured by registerFlushRecoveryCallback(...)
auto* dbFeature = arangodb::application_features::ApplicationServer::lookupFeature<
arangodb::DatabaseFeature
>("Database");
if (!dbFeature) {
return arangodb::Result(
TRI_ERROR_INTERNAL,
"failure to find feature 'Database' while applying 'Flush' recovery marker of type '" +
type + "'");
}
auto* vocbase = dbFeature->useDatabase(dbId);
if (!vocbase) {
return arangodb::Result(
TRI_ERROR_INTERNAL,
"failed to find database '" + std::to_string(dbId) +
"' while applying 'Flush' recovery marker of type '" + type + "'");
}
TRI_DEFER(vocbase->release());
if (!slice.hasKey(TYPE_ATTRIBUTE)) {
return arangodb::Result(
TRI_ERROR_INTERNAL,
"failed to find a 'data' attribute inside the supplied marker body while "
"applying 'Flush' recovery marker of type '" + type + "'"
);
}
auto data = slice.get(DATA_ATTRIBUTE);
try {
return itr->second(*vocbase, data);
} catch (arangodb::basics::Exception const& e) {
return arangodb::Result(
e.code(),
"caught exception while applying 'Flush' recovery marker of type '" + type + "': " + e.what());
} catch (std::exception const& e) {
return arangodb::Result(
TRI_ERROR_INTERNAL,
"caught exception while applying 'Flush' recovery marker of type '" + type + "': " + e.what());
} catch (...) {
return arangodb::Result(
TRI_ERROR_INTERNAL,
"caught exception while applying 'Flush' recovery marker of type '" + type + "'");
}
}
class MMFilesFlushMarker final: public arangodb::MMFilesWalMarker {
public:
/// @brief read constructor
explicit MMFilesFlushMarker(MMFilesMarker const& marker) {
if (type() != marker.getType()) {
THROW_ARANGO_EXCEPTION(arangodb::Result(
TRI_ERROR_BAD_PARAMETER,
"invalid marker type supplied while parsing 'Flush' recovery marker of type '"
+ std::to_string(marker.getType()) + "'"));
}
auto* data = reinterpret_cast<uint8_t const*>(&marker);
auto* ptr = data + sizeof(MMFilesMarker);
auto* end = ptr + marker.getSize() - sizeof(MMFilesMarker);
if (sizeof(TRI_voc_tick_t) > size_t(end - ptr)) {
THROW_ARANGO_EXCEPTION(arangodb::Result(
TRI_ERROR_BAD_PARAMETER,
"marker remaining size smaller than sizeof(TRI_voc_tick_t) "
"while parsing 'Flush' recovery marker of type '" +
std::to_string(marker.getType()) + "', remaining size '" +
std::to_string(size_t(end - ptr)) + "'"));
}
_databaseId = arangodb::encoding::readNumber<TRI_voc_tick_t>(
ptr, sizeof(TRI_voc_tick_t)
);
ptr += sizeof(TRI_voc_tick_t);
_slice = arangodb::velocypack::Slice(ptr);
if (_slice.byteSize() != size_t(end - ptr)) {
THROW_ARANGO_EXCEPTION(arangodb::Result(
TRI_ERROR_BAD_PARAMETER,
"marker remaining size not equal to the expected body size '" + std::to_string(_slice.byteSize()) +
"' while parsing 'Flush' recovery marker of type '" + std::to_string(marker.getType()) +
"', remaining size '" + std::to_string(size_t(end - ptr)) + "'"));
}
}
/// @brief write constructor
MMFilesFlushMarker(
TRI_voc_tick_t databaseId,
arangodb::velocypack::Slice const& slice
): _databaseId(databaseId), _slice(slice) {
}
TRI_voc_tick_t databaseId() const noexcept { return _databaseId; }
// always 0 since not yet in any datafile
TRI_voc_fid_t fid() const override final { return 0; }
uint32_t size() const override final {
return static_cast<uint32_t>(sizeof(MMFilesMarker) + sizeof(TRI_voc_tick_t) + _slice.byteSize());
}
arangodb::velocypack::Slice const& slice() const noexcept { return _slice; }
void store(char* mem) const override final {
auto* ptr = reinterpret_cast<uint8_t*>(mem) + sizeof(MMFilesMarker);
arangodb::encoding::storeNumber<TRI_voc_tick_t>(
ptr, _databaseId, sizeof(TRI_voc_tick_t)
);
ptr += sizeof(TRI_voc_tick_t);
std::memcpy(ptr, _slice.begin(), size_t(_slice.byteSize()));
}
MMFilesMarkerType type() const override final {
return MMFilesMarkerType::TRI_DF_MARKER_VPACK_FLUSH_SYNC;
}
private:
TRI_voc_tick_t _databaseId;
arangodb::velocypack::Slice _slice;
};
/// @note0 in MMFiles WAL file removal is based on:
/// min(releaseTick(...), addLogfileBarrier(...))
/// @note1 releaseTick(...) also controls WAL collection/compaction/flush, thus
/// it must always be released up to the currentTick() or the
/// aforementioned will wait indefinitely
class MMFilesFlushSubscription final : public FlushSubscriptionBase {
public:
MMFilesFlushSubscription(
std::string const& type, // subscription type
TRI_voc_tick_t databaseId, // vocbase id
arangodb::MMFilesLogfileManager& wal // marker write destination
): FlushSubscriptionBase(type, databaseId),
_barrier(wal.addLogfileBarrier( // earliest possible barrier
databaseId, 0, std::numeric_limits<double>::infinity() // args
)),
_wal(wal) {
}
~MMFilesFlushSubscription() {
if (!arangodb::MMFilesLogfileManager::instance(true)) { // true to avoid assertion failure
LOG_TOPIC("f7bea", ERR, arangodb::Logger::FLUSH)
<< "failed to remove MMFiles Logfile barrier from "
"subscription due to missing LogFileManager";
return; // ignore (probably already deallocated)
}
try {
_wal.removeLogfileBarrier(_barrier);
} catch (...) {
// NOOP
}
}
arangodb::Result commitImpl(VPackSlice data, TRI_voc_tick_t tick) override {
TRI_ASSERT(!data.isNone()); // ensured by 'FlushSubscriptionBase::commit(...)'
// must be present for WAL write to succeed or '_wal' is a dangling instance
// guard against scenario: FlushFeature::stop() + MMFilesEngine::stop() and later commit()
// since subscription could survive after FlushFeature::stop(), e.g. DatabaseFeature::unprepare()
TRI_ASSERT(arangodb::MMFilesLogfileManager::instance(true)); // true to avoid assertion failure
arangodb::velocypack::Builder builder;
builder.openObject();
builder.add(TYPE_ATTRIBUTE, arangodb::velocypack::Value(_type));
builder.add(DATA_ATTRIBUTE, data);
builder.close();
MMFilesFlushMarker marker(_databaseId, builder.slice());
auto res = arangodb::Result(_wal.allocateAndWrite(marker, true).errorCode); // will check for allowWalWrites()
if (res.ok()) {
auto barrier = _wal.addLogfileBarrier( // barier starting from previous marker
_databaseId, _tickPrevious, std::numeric_limits<double>::infinity() // args
);
try {
_wal.removeLogfileBarrier(_barrier);
_barrier = barrier;
} catch (...) {
_wal.removeLogfileBarrier(barrier);
}
resetCurrentTick(tick);
}
return res;
}
private:
TRI_voc_tick_t _barrier;
arangodb::MMFilesLogfileManager& _wal;
};
class MMFilesRecoveryHelper final: public arangodb::MMFilesRecoveryHelper {
virtual arangodb::Result replayMarker(
MMFilesMarker const& marker
) const override {
switch (marker.getType()) {
case MMFilesMarkerType::TRI_DF_MARKER_VPACK_FLUSH_SYNC: {
auto* dbFeature = arangodb::application_features::ApplicationServer::lookupFeature<
arangodb::DatabaseFeature
>("Database");
if (!dbFeature) {
return arangodb::Result(
TRI_ERROR_INTERNAL,
"failure to find feature 'Database' while applying 'Flush' recovery marker");
}
MMFilesFlushMarker flushMarker(marker);
return applyRecoveryMarker(flushMarker.databaseId(), flushMarker.slice());
}
default: {
// NOOP
}
};
return arangodb::Result();
}
};
class RocksDBFlushMarker {
public:
/// @brief read constructor
explicit RocksDBFlushMarker(rocksdb::Slice const& marker) {
if (arangodb::RocksDBLogType::FlushSync != arangodb::RocksDBLogValue::type(marker)) {
THROW_ARANGO_EXCEPTION(arangodb::Result(
TRI_ERROR_BAD_PARAMETER,
"invalid marker type supplied while parsing 'Flush' recovery marker of type '"
+ std::to_string(size_t(arangodb::RocksDBLogValue::type(marker))) + "'"));
}
auto* data = marker.data();
auto* ptr = data + sizeof(arangodb::RocksDBLogType);
auto* end = ptr + marker.size() - sizeof(arangodb::RocksDBLogType);
if (sizeof(uint64_t) > size_t(end - ptr)) {
THROW_ARANGO_EXCEPTION(arangodb::Result(
TRI_ERROR_BAD_PARAMETER,
"marker size smaller than sizeof(uint64_t) while parsing 'Flush' recovery marker of type '" +
std::to_string(size_t(arangodb::RocksDBLogValue::type(marker))) +
"', remaining size '" + std::to_string(size_t(end - ptr)) + "'"));
}
_databaseId = arangodb::rocksutils::uint64FromPersistent(ptr);
ptr += sizeof(uint64_t);
_slice = arangodb::velocypack::Slice(reinterpret_cast<uint8_t const*>(ptr));
if (_slice.byteSize() != size_t(end - ptr)) {
THROW_ARANGO_EXCEPTION(arangodb::Result(
TRI_ERROR_BAD_PARAMETER,
"marker remaining size not equal to the expected body size '" +
std::to_string(_slice.byteSize()) +
"' while parsing 'Flush' recovery marker of type '" +
std::to_string(size_t(arangodb::RocksDBLogValue::type(marker))) +
"', remaining size '" + std::to_string(size_t(end - ptr)) + "'"));
}
}
/// @brief write constructor
RocksDBFlushMarker(
TRI_voc_tick_t databaseId,
arangodb::velocypack::Slice const& slice
): _databaseId(databaseId), _slice(slice) {
}
TRI_voc_tick_t databaseId() const noexcept { return _databaseId; }
arangodb::velocypack::Slice const& slice() const noexcept { return _slice; }
void store(std::string& buf) const {
buf.reserve(
buf.size()
+ sizeof(arangodb::RocksDBLogType)
+ sizeof(uint64_t)
+ _slice.byteSize()
);
buf.push_back(static_cast<char>(arangodb::RocksDBLogType::FlushSync));
arangodb::rocksutils::uint64ToPersistent(buf, _databaseId);
buf.append(_slice.startAs<char>(), _slice.byteSize());
}
private:
TRI_voc_tick_t _databaseId;
arangodb::velocypack::Slice _slice;
};
/// @note in RocksDB WAL file removal is based on:
/// min(releaseTick(...)) only (contrary to MMFiles)
class RocksDBFlushSubscription final : public FlushSubscriptionBase {
public:
RocksDBFlushSubscription(
std::string const& type, // subscription type
TRI_voc_tick_t databaseId, // vocbase id
rocksdb::DB& wal // marker write destination
): FlushSubscriptionBase(type, databaseId),
_wal(wal) {
}
arangodb::Result commitImpl(VPackSlice data, TRI_voc_tick_t tick) override {
TRI_ASSERT(!data.isNone()); // ensured by 'FlushSubscriptionBase::commit(...)'
// must be present for WAL write to succeed or '_wal' is a dangling instance
// guard against scenario: FlushFeature::stop() + RocksDBEngine::stop() and later commit()
// since subscription could survive after FlushFeature::stop(), e.g. DatabaseFeature::unprepare()
TRI_ASSERT( // precondition
arangodb::EngineSelectorFeature::ENGINE // ensure have engine
&& static_cast<arangodb::RocksDBEngine*>(arangodb::EngineSelectorFeature::ENGINE)->db() // not stopped
);
arangodb::velocypack::Builder builder;
builder.openObject();
builder.add(TYPE_ATTRIBUTE, arangodb::velocypack::Value(_type));
builder.add(DATA_ATTRIBUTE, data);
builder.close();
rocksdb::WriteBatch batch;
std::string buf;
RocksDBFlushMarker marker(_databaseId, builder.slice());
marker.store(buf);
batch.PutLogData(rocksdb::Slice(buf));
static const rocksdb::WriteOptions op;
auto res = arangodb::rocksutils::convertStatus(_wal.Write(op, &batch));
if (res.ok()) {
resetCurrentTick(tick);
}
return res;
}
private:
rocksdb::DB& _wal;
};
class RocksDBRecoveryHelper final: public arangodb::RocksDBRecoveryHelper {
virtual void LogData(rocksdb::Slice const& marker) override {
switch (arangodb::RocksDBLogValue::type(marker)) {
case arangodb::RocksDBLogType::FlushSync: {
auto* dbFeature = arangodb::application_features::ApplicationServer::lookupFeature<
arangodb::DatabaseFeature
>("Database");
if (!dbFeature) {
THROW_ARANGO_EXCEPTION(arangodb::Result(
TRI_ERROR_INTERNAL,
"failure to find feature 'Database' while applying 'Flush' recovery marker"));
}
RocksDBFlushMarker flushMarker(marker);
auto res = applyRecoveryMarker(flushMarker.databaseId(), flushMarker.slice());
if (!res.ok()) {
THROW_ARANGO_EXCEPTION(res);
}
return;
}
default: {
// NOOP
}
};
}
};
void registerRecoveryHelper() {
static bool done(false);
static const MMFilesRecoveryHelper mmfilesHelper;
static const RocksDBRecoveryHelper rocksDBHelper;
if (done) {
return; // already registered previously, avoid duplicate registration (yes this is possible)
}
auto res = arangodb::MMFilesEngine::registerRecoveryHelper(mmfilesHelper);
if (!res.ok()) {
THROW_ARANGO_EXCEPTION(res);
}
res = arangodb::RocksDBEngine::registerRecoveryHelper(
std::shared_ptr<RocksDBRecoveryHelper>(std::shared_ptr<RocksDBRecoveryHelper>(),
const_cast<RocksDBRecoveryHelper*>(&rocksDBHelper))
);
if (!res.ok()) {
THROW_ARANGO_EXCEPTION(res);
}
done = true;
}
}
using namespace arangodb::application_features;
using namespace arangodb::basics;
using namespace arangodb::options;
namespace arangodb {
// used by catch tests
#ifdef ARANGODB_USE_GOOGLE_TESTS
/*static*/ FlushFeature::DefaultFlushSubscription FlushFeature::_defaultFlushSubscription;
#endif
std::atomic<bool> FlushFeature::_isRunning(false);
FlushFeature::FlushFeature(application_features::ApplicationServer& server)
: ApplicationFeature(server, "Flush"),
_flushInterval(1000000),
_stopped(false) {
setOptional(true);
startsAfter("BasicsPhase");
startsAfter("StorageEngine");
startsAfter("MMFilesLogfileManager");
}
void FlushFeature::collectOptions(std::shared_ptr<ProgramOptions> options) {
options->addSection("server", "Server features");
options->addOption("--server.flush-interval",
"interval (in microseconds) for flushing data",
new UInt64Parameter(&_flushInterval),
arangodb::options::makeFlags(arangodb::options::Flags::Hidden));
}
/*static*/ bool FlushFeature::registerFlushRecoveryCallback(
std::string const& type,
FlushRecoveryCallback const& callback) {
return !callback // skip empty callbacks
|| getFlushRecoveryCallbacks().emplace(type, callback).second;
}
std::shared_ptr<FlushFeature::FlushSubscription> FlushFeature::registerFlushSubscription(
std::string const& type,
TRI_vocbase_t const& vocbase) {
auto* engine = EngineSelectorFeature::ENGINE;
if (!engine) {
LOG_TOPIC("683b1", ERR, Logger::FLUSH)
<< "failed to find a storage engine while registering "
"'Flush' marker subscription for type '" << type << "'";
return nullptr;
}
std::shared_ptr<FlushFeature::FlushSubscription> subscription;
if (EngineSelectorFeature::isMMFiles()) {
auto* logFileManager = MMFilesLogfileManager::instance(true); // true to avoid assertion failure
if (!logFileManager) {
LOG_TOPIC("038b3", ERR, Logger::FLUSH)
<< "failed to find an MMFiles log file manager instance while registering "
"'Flush' marker subscription for type '" << type << "'";
return nullptr;
}
subscription = std::make_shared<MMFilesFlushSubscription>(type, vocbase.id(), *logFileManager);
} else if (EngineSelectorFeature::isRocksDB()) {
#ifdef ARANGODB_ENABLE_MAINTAINER_MODE
auto* rocksdbEngine = dynamic_cast<RocksDBEngine*>(engine);
#else
auto* rocksdbEngine = static_cast<RocksDBEngine*>(engine);
#endif
TRI_ASSERT(rocksdbEngine);
auto* db = rocksdbEngine->db();
if (!db) {
LOG_TOPIC("0f0f6", ERR, Logger::FLUSH)
<< "failed to find a RocksDB engine db while registering "
"'Flush' marker subscription for type '" << type << "'";
return nullptr;
}
auto* rootDb = db->GetRootDB();
if (!rootDb) {
LOG_TOPIC("26c23", ERR, Logger::FLUSH)
<< "failed to find a RocksDB engine root db while registering "
"'Flush' marker subscription for type '" << type << "'";
return nullptr;
}
subscription = std::make_shared<RocksDBFlushSubscription>(type, vocbase.id(), *rootDb);
}
#ifdef ARANGODB_USE_GOOGLE_TESTS
else if (_defaultFlushSubscription) {
struct DelegatingFlushSubscription final: public FlushSubscriptionBase {
DelegatingFlushSubscription(
std::string const& type,
TRI_vocbase_t const& vocbase,
DefaultFlushSubscription& delegate)
: FlushSubscriptionBase(type, vocbase.id()),
_delegate(delegate),
_vocbase(vocbase) {
}
Result commitImpl(VPackSlice data, TRI_voc_tick_t tick) override {
auto const res = _delegate(_type, _vocbase, data, tick);
if (res.ok()) {
resetCurrentTick(tick);
}
return res;
}
DefaultFlushSubscription _delegate;
TRI_vocbase_t const& _vocbase;
};
subscription = std::make_shared<DelegatingFlushSubscription>(type, vocbase, _defaultFlushSubscription);
}
#endif
if (!subscription) {
LOG_TOPIC("53c4e", ERR, Logger::FLUSH)
<< "failed to identify storage engine while registering "
"'Flush' marker subscription for type '" << type << "'";
return nullptr;
}
std::lock_guard<std::mutex> lock(_flushSubscriptionsMutex);
if (_stopped) {
LOG_TOPIC("798c4", ERR, Logger::FLUSH) << "FlushFeature not running";
return nullptr;
}
_flushSubscriptions.emplace_back(subscription);
return subscription;
}
arangodb::Result FlushFeature::releaseUnusedTicks(size_t& count, TRI_voc_tick_t& minTick) {
count = 0;
auto* engine = EngineSelectorFeature::ENGINE;
if (!engine) {
return Result(
TRI_ERROR_INTERNAL,
"failed to find a storage engine while releasing unused ticks"
);
}
minTick = engine->currentTick();
{
std::lock_guard<std::mutex> lock(_flushSubscriptionsMutex);
// find min tick and remove stale subscriptions
for (auto itr = _flushSubscriptions.begin(); itr != _flushSubscriptions.end();) {
auto entry = itr->lock();
if (!entry) {
// remove stale
itr = _flushSubscriptions.erase(itr);
++count;
} else {
minTick = std::min(minTick, entry->tick());
++itr;
}
}
}
TRI_ASSERT(minTick <= engine->currentTick());
TRI_IF_FAILURE("FlushCrashBeforeSyncingMinTick") {
TRI_SegfaultDebugging("crashing before syncing min tick");
}
engine->waitForSyncTick(minTick);
TRI_IF_FAILURE("FlushCrashAfterSyncingMinTick") {
TRI_SegfaultDebugging("crashing after syncing min tick");
}
engine->releaseTick(minTick);
TRI_IF_FAILURE("FlushCrashAfterReleasingMinTick") {
TRI_SegfaultDebugging("crashing after releasing min tick");
}
return {};
}
void FlushFeature::validateOptions(std::shared_ptr<options::ProgramOptions> options) {
if (_flushInterval < 1000) {
// do not go below 1000 microseconds
_flushInterval = 1000;
}
}
void FlushFeature::prepare() {
// At least for now we need FlushThread for ArangoSearch views
// on a DB/Single server only, so we avoid starting FlushThread on
// a coordinator and on agency nodes.
setEnabled(!arangodb::ServerState::instance()->isCoordinator() &&
!arangodb::ServerState::instance()->isAgent());
if (!isEnabled()) {
return;
}
registerRecoveryHelper();
}
void FlushFeature::start() {
{
WRITE_LOCKER(lock, _threadLock);
_flushThread.reset(new FlushThread(_flushInterval));
}
DatabaseFeature* dbFeature = DatabaseFeature::DATABASE;
dbFeature->registerPostRecoveryCallback([this]() -> Result {
READ_LOCKER(lock, _threadLock);
if (!this->_flushThread->start()) {
LOG_TOPIC("bdc3c", FATAL, Logger::FLUSH) << "unable to start FlushThread";
FATAL_ERROR_ABORT();
} else {
LOG_TOPIC("ed9cd", DEBUG, Logger::FLUSH) << "started FlushThread";
}
this->_isRunning.store(true);
return {TRI_ERROR_NO_ERROR};
});
}
void FlushFeature::beginShutdown() {
// pass on the shutdown signal
READ_LOCKER(lock, _threadLock);
if (_flushThread != nullptr) {
_flushThread->beginShutdown();
}
}
void FlushFeature::stop() {
LOG_TOPIC("2b0a6", TRACE, arangodb::Logger::FLUSH) << "stopping FlushThread";
// wait until thread is fully finished
FlushThread* thread = nullptr;
{
READ_LOCKER(lock, _threadLock);
thread = _flushThread.get();
}
if (thread != nullptr) {
while (thread->isRunning()) {
std::this_thread::sleep_for(std::chrono::microseconds(10000));
}
{
WRITE_LOCKER(wlock, _threadLock);
_isRunning.store(false);
_flushThread.reset();
}
{
std::lock_guard<std::mutex> lock(_flushSubscriptionsMutex);
// release any remaining flush subscriptions so that they may get deallocated ASAP
// subscriptions could survive after FlushFeature::stop(), e.g. DatabaseFeature::unprepare()
_flushSubscriptions.clear();
_stopped = true;
}
}
}
void FlushFeature::unprepare() {
}
} // namespace arangodb