1
0
Fork 0
arangodb/arangod/RocksDBEngine/RocksDBCommon.cpp

275 lines
9.7 KiB
C++

////////////////////////////////////////////////////////////////////////////////
/// DISCLAIMER
///
/// Copyright 2014-2017 ArangoDB GmbH, Cologne, Germany
/// Copyright 2004-2014 triAGENS GmbH, Cologne, Germany
///
/// Licensed under the Apache License, Version 2.0 (the "License");
/// you may not use this file except in compliance with the License.
/// You may obtain a copy of the License at
///
/// http://www.apache.org/licenses/LICENSE-2.0
///
/// Unless required by applicable law or agreed to in writing, software
/// distributed under the License is distributed on an "AS IS" BASIS,
/// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
/// See the License for the specific language governing permissions and
/// limitations under the License.
///
/// Copyright holder is ArangoDB GmbH, Cologne, Germany
///
/// @author Daniel H. Larkin
/// @author Jan Steemann
/// @author Jan Christoph Uhde
////////////////////////////////////////////////////////////////////////////////
#include "RocksDBCommon.h"
#include "Basics/RocksDBUtils.h"
#include "Basics/StringRef.h"
#include "Logger/Logger.h"
#include "RocksDBEngine/RocksDBColumnFamily.h"
#include "RocksDBEngine/RocksDBComparator.h"
#include "RocksDBEngine/RocksDBEngine.h"
#include "RocksDBEngine/RocksDBKey.h"
#include "RocksDBEngine/RocksDBKeyBounds.h"
#include "RocksDBEngine/RocksDBTransactionState.h"
#include "StorageEngine/EngineSelectorFeature.h"
#include "Transaction/Methods.h"
#include <rocksdb/comparator.h>
#include <rocksdb/convenience.h>
#include <rocksdb/utilities/transaction_db.h>
#include <velocypack/Iterator.h>
namespace arangodb {
namespace rocksutils {
rocksdb::TransactionDB* globalRocksDB() {
StorageEngine* engine = EngineSelectorFeature::ENGINE;
TRI_ASSERT(engine != nullptr);
RocksDBEngine* rocks = static_cast<RocksDBEngine*>(engine);
TRI_ASSERT(rocks->db() != nullptr);
return rocks->db();
}
rocksdb::ColumnFamilyHandle* defaultCF() {
auto db = globalRocksDB();
TRI_ASSERT(db != nullptr);
return db->DefaultColumnFamily();
}
RocksDBEngine* globalRocksEngine() {
StorageEngine* engine = EngineSelectorFeature::ENGINE;
TRI_ASSERT(engine != nullptr);
return static_cast<RocksDBEngine*>(engine);
}
arangodb::Result globalRocksDBPut(rocksdb::ColumnFamilyHandle* cf,
rocksdb::Slice const& key,
rocksdb::Slice const& val,
rocksdb::WriteOptions const& options) {
TRI_ASSERT(cf != nullptr);
auto status = globalRocksDB()->Put(options, cf, key, val);
return convertStatus(status);
}
arangodb::Result globalRocksDBRemove(rocksdb::ColumnFamilyHandle* cf,
rocksdb::Slice const& key,
rocksdb::WriteOptions const& options) {
TRI_ASSERT(cf != nullptr);
auto status = globalRocksDB()->Delete(options, cf, key);
return convertStatus(status);
}
uint64_t latestSequenceNumber() {
auto seq = globalRocksDB()->GetLatestSequenceNumber();
return static_cast<uint64_t>(seq);
}
void addCollectionMapping(uint64_t objectId, TRI_voc_tick_t did,
TRI_voc_cid_t cid) {
StorageEngine* engine = EngineSelectorFeature::ENGINE;
TRI_ASSERT(engine != nullptr);
RocksDBEngine* rocks = static_cast<RocksDBEngine*>(engine);
TRI_ASSERT(rocks->db() != nullptr);
return rocks->addCollectionMapping(objectId, did, cid);
}
std::pair<TRI_voc_tick_t, TRI_voc_cid_t> mapObjectToCollection(
uint64_t objectId) {
StorageEngine* engine = EngineSelectorFeature::ENGINE;
TRI_ASSERT(engine != nullptr);
RocksDBEngine* rocks = static_cast<RocksDBEngine*>(engine);
TRI_ASSERT(rocks->db() != nullptr);
return rocks->mapObjectToCollection(objectId);
}
/// @brief count all keys in the given column family
std::size_t countKeys(rocksdb::DB* db, rocksdb::ColumnFamilyHandle* cf) {
TRI_ASSERT(cf != nullptr);
rocksdb::ReadOptions opts;
opts.fill_cache = false;
opts.total_order_seek = true;
std::unique_ptr<rocksdb::Iterator> it(db->NewIterator(opts, cf));
std::size_t count = 0;
rocksdb::Slice lower("\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00", 16);
it->Seek(lower);
while (it->Valid()) {
++count;
it->Next();
}
return count;
}
/// @brief iterate over all keys in range and count them
std::size_t countKeyRange(rocksdb::DB* db,
RocksDBKeyBounds const& bounds,
bool prefix_same_as_start) {
rocksdb::Slice lower(bounds.start());
rocksdb::Slice upper(bounds.end());
rocksdb::ReadOptions readOptions;
readOptions.prefix_same_as_start = prefix_same_as_start;
readOptions.iterate_upper_bound = &upper;
readOptions.total_order_seek = !prefix_same_as_start;
readOptions.verify_checksums = false;
readOptions.fill_cache = false;
rocksdb::ColumnFamilyHandle* cf = bounds.columnFamily();
rocksdb::Comparator const* cmp = cf->GetComparator();
std::unique_ptr<rocksdb::Iterator> it(db->NewIterator(readOptions, cf));
std::size_t count = 0;
it->Seek(lower);
while (it->Valid() && cmp->Compare(it->key(), upper) < 0) {
++count;
it->Next();
}
return count;
}
/// @brief helper method to remove large ranges of data
/// Should mainly be used to implement the drop() call
Result removeLargeRange(rocksdb::TransactionDB* db,
RocksDBKeyBounds const& bounds,
bool prefix_same_as_start) {
LOG_TOPIC(DEBUG, Logger::ROCKSDB) << "removing large range: " << bounds;
rocksdb::ColumnFamilyHandle* cf = bounds.columnFamily();
rocksdb::DB* bDB = db->GetBaseDB();
TRI_ASSERT(bDB != nullptr);
try {
// delete files in range lower..upper
rocksdb::Slice lower(bounds.start());
rocksdb::Slice upper(bounds.end());
{
rocksdb::Status status =
rocksdb::DeleteFilesInRange(bDB, cf, &lower, &upper);
if (!status.ok()) {
// if file deletion failed, we will still iterate over the remaining
// keys, so we don't need to abort and raise an error here
arangodb::Result r = rocksutils::convertStatus(status);
LOG_TOPIC(WARN, arangodb::Logger::ROCKSDB)
<< "RocksDB file deletion failed: " << r.errorMessage();
}
}
// go on and delete the remaining keys (delete files in range does not
// necessarily find them all, just complete files)
rocksdb::Comparator const* cmp = cf->GetComparator();
rocksdb::WriteBatch batch;
rocksdb::ReadOptions readOptions;
readOptions.iterate_upper_bound = &upper;
readOptions.prefix_same_as_start = prefix_same_as_start;
readOptions.total_order_seek = !prefix_same_as_start;
readOptions.verify_checksums = false;
readOptions.fill_cache = false;
std::unique_ptr<rocksdb::Iterator> it(bDB->NewIterator(readOptions, cf));
size_t total = 0;
size_t counter = 0;
for (it->Seek(lower); it->Valid(); it->Next()) {
TRI_ASSERT(cmp->Compare(it->key(), lower) > 0);
TRI_ASSERT(cmp->Compare(it->key(), upper) < 0);
++total;
++counter;
batch.Delete(cf, it->key());
if (counter == 1000) {
LOG_TOPIC(DEBUG, Logger::FIXME) << "intermediate delete write";
// Persist deletes all 1000 documents
rocksdb::Status status = bDB->Write(rocksdb::WriteOptions(), &batch);
if (!status.ok()) {
LOG_TOPIC(WARN, arangodb::Logger::FIXME)
<< "RocksDB key deletion failed: " << status.ToString();
return TRI_ERROR_INTERNAL;
}
batch.Clear();
counter = 0;
}
}
LOG_TOPIC(DEBUG, Logger::ROCKSDB) << "removing large range, deleted in total: " << total;
if (counter > 0) {
LOG_TOPIC(DEBUG, Logger::FIXME) << "intermediate delete write";
// We still have sth to write
// now apply deletion batch
rocksdb::Status status = bDB->Write(rocksdb::WriteOptions(), &batch);
if (!status.ok()) {
LOG_TOPIC(WARN, arangodb::Logger::FIXME)
<< "RocksDB key deletion failed: " << status.ToString();
return TRI_ERROR_INTERNAL;
}
}
return TRI_ERROR_NO_ERROR;
} catch (arangodb::basics::Exception const& ex) {
LOG_TOPIC(ERR, arangodb::Logger::FIXME)
<< "caught exception during RocksDB key prefix deletion: " << ex.what();
return ex.code();
} catch (std::exception const& ex) {
LOG_TOPIC(ERR, arangodb::Logger::FIXME)
<< "caught exception during RocksDB key prefix deletion: " << ex.what();
return TRI_ERROR_INTERNAL;
} catch (...) {
LOG_TOPIC(ERR, arangodb::Logger::FIXME)
<< "caught unknown exception during RocksDB key prefix deletion";
return TRI_ERROR_INTERNAL;
}
}
std::vector<std::pair<RocksDBKey, RocksDBValue>> collectionKVPairs(
TRI_voc_tick_t databaseId) {
std::vector<std::pair<RocksDBKey, RocksDBValue>> rv;
RocksDBKeyBounds bounds = RocksDBKeyBounds::DatabaseCollections(databaseId);
iterateBounds(bounds,
[&rv](rocksdb::Iterator* it) {
rv.emplace_back(
RocksDBKey(it->key()),
RocksDBValue(RocksDBEntryType::Collection, it->value()));
},
arangodb::RocksDBColumnFamily::definitions());
return rv;
}
std::vector<std::pair<RocksDBKey, RocksDBValue>> viewKVPairs(
TRI_voc_tick_t databaseId) {
std::vector<std::pair<RocksDBKey, RocksDBValue>> rv;
RocksDBKeyBounds bounds = RocksDBKeyBounds::DatabaseViews(databaseId);
iterateBounds(bounds,
[&rv](rocksdb::Iterator* it) {
rv.emplace_back(
RocksDBKey(it->key()),
RocksDBValue(RocksDBEntryType::View, it->value()));
},
arangodb::RocksDBColumnFamily::definitions());
return rv;
}
} // namespace rocksutils
} // namespace arangodb