//////////////////////////////////////////////////////////////////////////////// /// DISCLAIMER /// /// Copyright 2014-2016 ArangoDB GmbH, Cologne, Germany /// Copyright 2004-2014 triAGENS GmbH, Cologne, Germany /// /// Licensed under the Apache License, Version 2.0 (the "License"); /// you may not use this file except in compliance with the License. /// You may obtain a copy of the License at /// /// http://www.apache.org/licenses/LICENSE-2.0 /// /// Unless required by applicable law or agreed to in writing, software /// distributed under the License is distributed on an "AS IS" BASIS, /// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. /// See the License for the specific language governing permissions and /// limitations under the License. /// /// Copyright holder is ArangoDB GmbH, Cologne, Germany /// /// @author Jan Steemann //////////////////////////////////////////////////////////////////////////////// #include "CollectionExport.h" #include "Basics/JsonHelper.h" #include "Indexes/PrimaryIndex.h" #include "Utils/CollectionGuard.h" #include "Utils/SingleCollectionTransaction.h" #include "Utils/StandaloneTransactionContext.h" #include "VocBase/compactor.h" #include "VocBase/Ditch.h" #include "VocBase/vocbase.h" using namespace arangodb; CollectionExport::CollectionExport(TRI_vocbase_t* vocbase, std::string const& name, Restrictions const& restrictions) : _guard(nullptr), _document(nullptr), _ditch(nullptr), _name(name), _resolver(vocbase), _restrictions(restrictions), _documents(nullptr) { // prevent the collection from being unloaded while the export is ongoing // this may throw _guard = new arangodb::CollectionGuard(vocbase, _name.c_str(), false); _document = _guard->collection()->_collection; TRI_ASSERT(_document != nullptr); } CollectionExport::~CollectionExport() { delete _documents; if (_ditch != nullptr) { _ditch->ditches()->freeDocumentDitch(_ditch, false); } // if not already done delete _guard; } void CollectionExport::run(uint64_t maxWaitTime, size_t limit) { // try to acquire the exclusive lock on the compaction while (!TRI_CheckAndLockCompactorVocBase(_document->_vocbase)) { // didn't get it. try again... usleep(5000); } // create a ditch under the compaction lock _ditch = _document->ditches()->createDocumentDitch(false, __FILE__, __LINE__); // release the lock TRI_UnlockCompactorVocBase(_document->_vocbase); // now we either have a ditch or not if (_ditch == nullptr) { THROW_ARANGO_EXCEPTION(TRI_ERROR_OUT_OF_MEMORY); } TRI_ASSERT(_documents == nullptr); _documents = new std::vector(); { static uint64_t const SleepTime = 10000; uint64_t tries = 0; uint64_t const maxTries = maxWaitTime / SleepTime; while (++tries < maxTries) { if (_document->isFullyCollected()) { break; } usleep(SleepTime); } } { SingleCollectionTransaction trx(StandaloneTransactionContext::Create(_document->_vocbase), _name, TRI_TRANSACTION_READ); // already locked by guard above trx.addHint(TRI_TRANSACTION_HINT_NO_USAGE_LOCK, true); int res = trx.begin(); if (res != TRI_ERROR_NO_ERROR) { THROW_ARANGO_EXCEPTION(res); } size_t maxDocuments = _document->primaryIndex()->size(); if (limit > 0 && limit < maxDocuments) { maxDocuments = limit; } else { limit = maxDocuments; } _documents->reserve(maxDocuments); trx.invokeOnAllElements(_document->_info.name(), [this, &limit](TRI_doc_mptr_t const* mptr) { if (limit == 0) { return false; } if (!mptr->pointsToWal()) { _documents->emplace_back(mptr->getMarkerPtr()); --limit; } return true; }); trx.finish(res); } // delete guard right now as we're about to return // if we would continue holding the guard's collection lock and return, // and the export object gets later freed in a different thread, then all // would be lost. so we'll release the lock here and rely on the cleanup // thread not unloading the collection (as we've acquired a document ditch // for the collection already - this will prevent unloading of the collection's // datafiles etc.) delete _guard; _guard = nullptr; }