mirror of https://gitee.com/bigwinds/arangodb
241 lines
7.8 KiB
C++
241 lines
7.8 KiB
C++
////////////////////////////////////////////////////////////////////////////////
|
|
/// DISCLAIMER
|
|
///
|
|
/// Copyright 2014-2016 ArangoDB GmbH, Cologne, Germany
|
|
/// Copyright 2004-2014 triAGENS GmbH, Cologne, Germany
|
|
///
|
|
/// Licensed under the Apache License, Version 2.0 (the "License");
|
|
/// you may not use this file except in compliance with the License.
|
|
/// You may obtain a copy of the License at
|
|
///
|
|
/// http://www.apache.org/licenses/LICENSE-2.0
|
|
///
|
|
/// Unless required by applicable law or agreed to in writing, software
|
|
/// distributed under the License is distributed on an "AS IS" BASIS,
|
|
/// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
/// See the License for the specific language governing permissions and
|
|
/// limitations under the License.
|
|
///
|
|
/// Copyright holder is ArangoDB GmbH, Cologne, Germany
|
|
///
|
|
/// @author Jan Steemann
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
|
|
#include "CollectionKeys.h"
|
|
#include "Utils/CollectionGuard.h"
|
|
#include "Utils/SingleCollectionTransaction.h"
|
|
#include "Utils/StandaloneTransactionContext.h"
|
|
#include "VocBase/compactor.h"
|
|
#include "VocBase/DatafileHelper.h"
|
|
#include "VocBase/Ditch.h"
|
|
#include "VocBase/document-collection.h"
|
|
#include "VocBase/server.h"
|
|
#include "VocBase/vocbase.h"
|
|
#include "Wal/LogfileManager.h"
|
|
|
|
#include <velocypack/Builder.h>
|
|
#include <velocypack/Iterator.h>
|
|
#include <velocypack/velocypack-aliases.h>
|
|
|
|
using namespace arangodb;
|
|
|
|
CollectionKeys::CollectionKeys(TRI_vocbase_t* vocbase, std::string const& name,
|
|
TRI_voc_tick_t blockerId, double ttl)
|
|
: _vocbase(vocbase),
|
|
_guard(nullptr),
|
|
_document(nullptr),
|
|
_ditch(nullptr),
|
|
_name(name),
|
|
_resolver(vocbase),
|
|
_blockerId(blockerId),
|
|
_markers(nullptr),
|
|
_id(0),
|
|
_ttl(ttl),
|
|
_expires(0.0),
|
|
_isDeleted(false),
|
|
_isUsed(false) {
|
|
_id = TRI_NewTickServer();
|
|
_expires = TRI_microtime() + _ttl;
|
|
TRI_ASSERT(_blockerId > 0);
|
|
|
|
// prevent the collection from being unloaded while the export is ongoing
|
|
// this may throw
|
|
_guard = new arangodb::CollectionGuard(vocbase, _name.c_str(), false);
|
|
|
|
_document = _guard->collection()->_collection;
|
|
TRI_ASSERT(_document != nullptr);
|
|
}
|
|
|
|
CollectionKeys::~CollectionKeys() {
|
|
// remove compaction blocker
|
|
TRI_RemoveBlockerCompactorVocBase(_vocbase, _blockerId);
|
|
|
|
delete _markers;
|
|
|
|
if (_ditch != nullptr) {
|
|
_ditch->ditches()->freeDocumentDitch(_ditch, false);
|
|
}
|
|
|
|
delete _guard;
|
|
}
|
|
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
/// @brief initially creates the list of keys
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
|
|
void CollectionKeys::create(TRI_voc_tick_t maxTick) {
|
|
arangodb::wal::LogfileManager::instance()->waitForCollectorQueue(
|
|
_document->_info.id(), 30.0);
|
|
|
|
// try to acquire the exclusive lock on the compaction
|
|
while (!TRI_CheckAndLockCompactorVocBase(_document->_vocbase)) {
|
|
// didn't get it. try again...
|
|
usleep(5000);
|
|
}
|
|
|
|
// create a ditch under the compaction lock
|
|
_ditch = _document->ditches()->createDocumentDitch(false, __FILE__, __LINE__);
|
|
|
|
// release the lock
|
|
TRI_UnlockCompactorVocBase(_document->_vocbase);
|
|
|
|
// now we either have a ditch or not
|
|
if (_ditch == nullptr) {
|
|
THROW_ARANGO_EXCEPTION(TRI_ERROR_OUT_OF_MEMORY);
|
|
}
|
|
|
|
TRI_ASSERT(_markers == nullptr);
|
|
_markers = new std::vector<TRI_df_marker_t const*>();
|
|
|
|
// copy all datafile markers into the result under the read-lock
|
|
{
|
|
SingleCollectionTransaction trx(StandaloneTransactionContext::Create(_document->_vocbase),
|
|
_name, TRI_TRANSACTION_READ);
|
|
|
|
int res = trx.begin();
|
|
|
|
if (res != TRI_ERROR_NO_ERROR) {
|
|
THROW_ARANGO_EXCEPTION(res);
|
|
}
|
|
|
|
trx.invokeOnAllElements(_document->_info.name(), [this, &maxTick](TRI_doc_mptr_t const* mptr) {
|
|
// only use those markers that point into datafiles
|
|
if (!mptr->pointsToWal()) {
|
|
auto marker = mptr->getMarkerPtr();
|
|
|
|
if (marker->getTick() <= maxTick) {
|
|
_markers->emplace_back(marker);
|
|
}
|
|
}
|
|
|
|
return true;
|
|
});
|
|
|
|
trx.finish(res);
|
|
}
|
|
|
|
// now sort all markers without the read-lock
|
|
std::sort(_markers->begin(), _markers->end(),
|
|
[](TRI_df_marker_t const* lhs, TRI_df_marker_t const* rhs) -> bool {
|
|
VPackSlice l(reinterpret_cast<char const*>(lhs) + DatafileHelper::VPackOffset(TRI_DF_MARKER_VPACK_DOCUMENT));
|
|
VPackSlice r(reinterpret_cast<char const*>(rhs) + DatafileHelper::VPackOffset(TRI_DF_MARKER_VPACK_DOCUMENT));
|
|
|
|
return (l.get(TRI_VOC_ATTRIBUTE_KEY).copyString() < r.get(TRI_VOC_ATTRIBUTE_KEY).copyString());
|
|
});
|
|
}
|
|
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
/// @brief hashes a chunk of keys
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
|
|
std::tuple<std::string, std::string, uint64_t> CollectionKeys::hashChunk(
|
|
size_t from, size_t to) const {
|
|
if (from >= _markers->size() || to > _markers->size() || from >= to ||
|
|
to == 0) {
|
|
THROW_ARANGO_EXCEPTION(TRI_ERROR_BAD_PARAMETER);
|
|
}
|
|
|
|
size_t const offset = DatafileHelper::VPackOffset(TRI_DF_MARKER_VPACK_DOCUMENT);
|
|
VPackSlice first(reinterpret_cast<char const*>(_markers->at(from)) + offset);
|
|
VPackSlice last(reinterpret_cast<char const*>(_markers->at(to - 1)) + offset);
|
|
|
|
TRI_ASSERT(first.isObject());
|
|
TRI_ASSERT(last.isObject());
|
|
|
|
uint64_t hash = 0x012345678;
|
|
|
|
for (size_t i = from; i < to; ++i) {
|
|
VPackSlice current(reinterpret_cast<char const*>(_markers->at(i)) + offset);
|
|
TRI_ASSERT(current.isObject());
|
|
|
|
hash ^= current.get(TRI_VOC_ATTRIBUTE_KEY).hash();
|
|
hash ^= current.get(TRI_VOC_ATTRIBUTE_REV).hash();
|
|
}
|
|
|
|
return std::make_tuple(
|
|
first.get(TRI_VOC_ATTRIBUTE_KEY).copyString(),
|
|
last.get(TRI_VOC_ATTRIBUTE_KEY).copyString(),
|
|
hash);
|
|
}
|
|
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
/// @brief dumps keys into the result
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
|
|
void CollectionKeys::dumpKeys(VPackBuilder& result, size_t chunk,
|
|
size_t chunkSize) const {
|
|
size_t from = chunk * chunkSize;
|
|
size_t to = (chunk + 1) * chunkSize;
|
|
|
|
if (to > _markers->size()) {
|
|
to = _markers->size();
|
|
}
|
|
|
|
if (from >= _markers->size() || from >= to || to == 0) {
|
|
THROW_ARANGO_EXCEPTION(TRI_ERROR_BAD_PARAMETER);
|
|
}
|
|
|
|
size_t const offset = DatafileHelper::VPackOffset(TRI_DF_MARKER_VPACK_DOCUMENT);
|
|
|
|
for (size_t i = from; i < to; ++i) {
|
|
VPackSlice current(reinterpret_cast<char const*>(_markers->at(i)) + offset);
|
|
TRI_ASSERT(current.isObject());
|
|
|
|
result.openArray();
|
|
result.add(current.get(TRI_VOC_ATTRIBUTE_KEY));
|
|
result.add(current.get(TRI_VOC_ATTRIBUTE_REV));
|
|
result.close();
|
|
}
|
|
}
|
|
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
/// @brief dumps documents into the result
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
|
|
void CollectionKeys::dumpDocs(arangodb::velocypack::Builder& result, size_t chunk,
|
|
size_t chunkSize, VPackSlice const& ids) const {
|
|
if (!ids.isArray()) {
|
|
THROW_ARANGO_EXCEPTION(TRI_ERROR_BAD_PARAMETER);
|
|
}
|
|
|
|
size_t const offset = DatafileHelper::VPackOffset(TRI_DF_MARKER_VPACK_DOCUMENT);
|
|
|
|
for (auto const& it : VPackArrayIterator(ids)) {
|
|
if (!it.isNumber()) {
|
|
THROW_ARANGO_EXCEPTION(TRI_ERROR_BAD_PARAMETER);
|
|
}
|
|
|
|
size_t position = chunk * chunkSize + it.getNumber<size_t>();
|
|
|
|
if (position >= _markers->size()) {
|
|
THROW_ARANGO_EXCEPTION(TRI_ERROR_BAD_PARAMETER);
|
|
}
|
|
|
|
VPackSlice current(reinterpret_cast<char const*>(_markers->at(position)) + offset);
|
|
TRI_ASSERT(current.isObject());
|
|
|
|
result.add(current);
|
|
}
|
|
}
|
|
|