mirror of https://gitee.com/bigwinds/arangodb
Merge branch 'devel' of https://github.com/arangodb/arangodb into devel
This commit is contained in:
commit
af393605db
|
@ -2284,7 +2284,7 @@ ClusterMethods::persistCollectionInAgency(
|
||||||
std::string distributeShardsLike = col->distributeShardsLike();
|
std::string distributeShardsLike = col->distributeShardsLike();
|
||||||
std::vector<std::string> dbServers;
|
std::vector<std::string> dbServers;
|
||||||
std::vector<std::string> avoid = col->avoidServers();
|
std::vector<std::string> avoid = col->avoidServers();
|
||||||
|
|
||||||
ClusterInfo* ci = ClusterInfo::instance();
|
ClusterInfo* ci = ClusterInfo::instance();
|
||||||
if (!distributeShardsLike.empty()) {
|
if (!distributeShardsLike.empty()) {
|
||||||
CollectionNameResolver resolver(col->vocbase());
|
CollectionNameResolver resolver(col->vocbase());
|
||||||
|
@ -2320,7 +2320,6 @@ ClusterMethods::persistCollectionInAgency(
|
||||||
}
|
}
|
||||||
col->distributeShardsLike(otherCidString);
|
col->distributeShardsLike(otherCidString);
|
||||||
} else {
|
} else {
|
||||||
LOG_TOPIC(WARN, Logger::CLUSTER) << "WTF? " << ignoreDistributeShardsLikeErrors;
|
|
||||||
if (ignoreDistributeShardsLikeErrors) {
|
if (ignoreDistributeShardsLikeErrors) {
|
||||||
col->distributeShardsLike(std::string());
|
col->distributeShardsLike(std::string());
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -258,8 +258,8 @@ class ClusterMethods {
|
||||||
static std::unique_ptr<LogicalCollection> createCollectionOnCoordinator(
|
static std::unique_ptr<LogicalCollection> createCollectionOnCoordinator(
|
||||||
TRI_col_type_e collectionType, TRI_vocbase_t* vocbase,
|
TRI_col_type_e collectionType, TRI_vocbase_t* vocbase,
|
||||||
arangodb::velocypack::Slice parameters,
|
arangodb::velocypack::Slice parameters,
|
||||||
bool ignoreDistributeShardsLikeErrors,
|
bool ignoreDistributeShardsLikeErrors = true,
|
||||||
bool waitForSyncReplication);
|
bool waitForSyncReplication = true);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
|
|
||||||
|
@ -268,7 +268,8 @@ class ClusterMethods {
|
||||||
////////////////////////////////////////////////////////////////////////////////
|
////////////////////////////////////////////////////////////////////////////////
|
||||||
|
|
||||||
static std::unique_ptr<LogicalCollection> persistCollectionInAgency(
|
static std::unique_ptr<LogicalCollection> persistCollectionInAgency(
|
||||||
LogicalCollection* col, bool ignoreDistributeShardsLikeErrors, bool waitForSyncReplication);
|
LogicalCollection* col, bool ignoreDistributeShardsLikeErrors = true,
|
||||||
|
bool waitForSyncReplication = true);
|
||||||
};
|
};
|
||||||
|
|
||||||
} // namespace arangodb
|
} // namespace arangodb
|
||||||
|
|
|
@ -1265,8 +1265,7 @@ int InitialSyncer::handleSyncKeysRocksDB(arangodb::LogicalCollection* col,
|
||||||
// smaller values than lowKey mean they don't exist remotely
|
// smaller values than lowKey mean they don't exist remotely
|
||||||
trx.remove(collectionName, key, options);
|
trx.remove(collectionName, key, options);
|
||||||
return;
|
return;
|
||||||
}
|
} else if (cmp1 >= 0 && cmp2 <= 0) {
|
||||||
if (cmp1 >= 0 && cmp2 <= 0) {
|
|
||||||
// we only need to hash we are in the range
|
// we only need to hash we are in the range
|
||||||
if (cmp1 == 0) {
|
if (cmp1 == 0) {
|
||||||
foundLowKey = true;
|
foundLowKey = true;
|
||||||
|
@ -1294,21 +1293,23 @@ int InitialSyncer::handleSyncKeysRocksDB(arangodb::LogicalCollection* col,
|
||||||
nextChunk = true;
|
nextChunk = true;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (rangeUnequal) {
|
|
||||||
int res = syncChunkRocksDB(&trx, keysId, currentChunkId, lowKey,
|
|
||||||
highKey, markers, errorMsg);
|
|
||||||
if (res != TRI_ERROR_NO_ERROR) {
|
|
||||||
THROW_ARANGO_EXCEPTION(res);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
TRI_ASSERT(!rangeUnequal || nextChunk); // A => B
|
TRI_ASSERT(!rangeUnequal || nextChunk); // A => B
|
||||||
if (nextChunk && currentChunkId + 1 < numChunks) {
|
if (nextChunk) {// we are out of range, see next chunk
|
||||||
currentChunkId++; // we are out of range, see next chunk
|
if (rangeUnequal && currentChunkId < numChunks) {
|
||||||
resetChunk();
|
int res = syncChunkRocksDB(&trx, keysId, currentChunkId, lowKey,
|
||||||
|
highKey, markers, errorMsg);
|
||||||
// key is higher than upper bound, recheck the current document
|
if (res != TRI_ERROR_NO_ERROR) {
|
||||||
if (cmp2 > 0) {
|
THROW_ARANGO_EXCEPTION(res);
|
||||||
parseDoc(doc, key);
|
}
|
||||||
|
}
|
||||||
|
currentChunkId++;
|
||||||
|
if (currentChunkId < numChunks) {
|
||||||
|
resetChunk();
|
||||||
|
// key is higher than upper bound, recheck the current document
|
||||||
|
if (cmp2 > 0) {
|
||||||
|
parseDoc(doc, key);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
@ -1325,6 +1326,19 @@ int InitialSyncer::handleSyncKeysRocksDB(arangodb::LogicalCollection* col,
|
||||||
parseDoc(doc, key);
|
parseDoc(doc, key);
|
||||||
},
|
},
|
||||||
UINT64_MAX);
|
UINT64_MAX);
|
||||||
|
|
||||||
|
// we might have missed chunks, if the keys don't exist at all locally
|
||||||
|
while (currentChunkId < numChunks) {
|
||||||
|
int res = syncChunkRocksDB(&trx, keysId, currentChunkId, lowKey,
|
||||||
|
highKey, markers, errorMsg);
|
||||||
|
if (res != TRI_ERROR_NO_ERROR) {
|
||||||
|
THROW_ARANGO_EXCEPTION(res);
|
||||||
|
}
|
||||||
|
currentChunkId++;
|
||||||
|
if (currentChunkId < numChunks) {
|
||||||
|
resetChunk();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
res = trx.commit();
|
res = trx.commit();
|
||||||
if (!res.ok()) {
|
if (!res.ok()) {
|
||||||
|
@ -1423,8 +1437,8 @@ int InitialSyncer::syncChunkRocksDB(
|
||||||
size_t const numKeys = static_cast<size_t>(responseBody.length());
|
size_t const numKeys = static_cast<size_t>(responseBody.length());
|
||||||
if (numKeys == 0) {
|
if (numKeys == 0) {
|
||||||
errorMsg = "got invalid response from master at " + _masterInfo._endpoint +
|
errorMsg = "got invalid response from master at " + _masterInfo._endpoint +
|
||||||
": response contains an empty chunk. ChunkId: " +
|
": response contains an empty chunk. Collection: " + collectionName +
|
||||||
std::to_string(chunkId);
|
" Chunk: " + std::to_string(chunkId);
|
||||||
return TRI_ERROR_REPLICATION_INVALID_RESPONSE;
|
return TRI_ERROR_REPLICATION_INVALID_RESPONSE;
|
||||||
}
|
}
|
||||||
TRI_ASSERT(numKeys > 0);
|
TRI_ASSERT(numKeys > 0);
|
||||||
|
@ -1494,6 +1508,23 @@ int InitialSyncer::syncChunkRocksDB(
|
||||||
|
|
||||||
i++;
|
i++;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// delete all keys at end of the range
|
||||||
|
while (nextStart < markers.size()) {
|
||||||
|
std::string const& localKey = markers[nextStart].first;
|
||||||
|
|
||||||
|
TRI_ASSERT(localKey.compare(highString) > 0);
|
||||||
|
//if (localKey.compare(highString) > 0) {
|
||||||
|
// we have a local key that is not present remotely
|
||||||
|
keyBuilder->clear();
|
||||||
|
keyBuilder->openObject();
|
||||||
|
keyBuilder->add(StaticStrings::KeyString, VPackValue(localKey));
|
||||||
|
keyBuilder->close();
|
||||||
|
|
||||||
|
trx->remove(collectionName, keyBuilder->slice(), options);
|
||||||
|
//}
|
||||||
|
++nextStart;
|
||||||
|
}
|
||||||
|
|
||||||
if (!toFetch.empty()) {
|
if (!toFetch.empty()) {
|
||||||
VPackBuilder keysBuilder;
|
VPackBuilder keysBuilder;
|
||||||
|
|
|
@ -71,7 +71,7 @@ static inline rocksdb::Transaction* rocksTransaction(
|
||||||
return static_cast<RocksDBTransactionState*>(trx->state())
|
return static_cast<RocksDBTransactionState*>(trx->state())
|
||||||
->rocksTransaction();
|
->rocksTransaction();
|
||||||
}
|
}
|
||||||
}
|
} // namespace
|
||||||
|
|
||||||
RocksDBCollection::RocksDBCollection(LogicalCollection* collection,
|
RocksDBCollection::RocksDBCollection(LogicalCollection* collection,
|
||||||
VPackSlice const& info)
|
VPackSlice const& info)
|
||||||
|
@ -196,8 +196,8 @@ void RocksDBCollection::open(bool ignoreErrors) {
|
||||||
RocksDBEngine* engine =
|
RocksDBEngine* engine =
|
||||||
static_cast<RocksDBEngine*>(EngineSelectorFeature::ENGINE);
|
static_cast<RocksDBEngine*>(EngineSelectorFeature::ENGINE);
|
||||||
auto counterValue = engine->counterManager()->loadCounter(this->objectId());
|
auto counterValue = engine->counterManager()->loadCounter(this->objectId());
|
||||||
LOG_TOPIC(ERR, Logger::DEVEL) << " number of documents: "
|
LOG_TOPIC(ERR, Logger::DEVEL)
|
||||||
<< counterValue.added();
|
<< " number of documents: " << counterValue.added();
|
||||||
_numberDocuments = counterValue.added() - counterValue.removed();
|
_numberDocuments = counterValue.added() - counterValue.removed();
|
||||||
_revisionId = counterValue.revisionId();
|
_revisionId = counterValue.revisionId();
|
||||||
//_numberDocuments = countKeyRange(db, readOptions,
|
//_numberDocuments = countKeyRange(db, readOptions,
|
||||||
|
@ -443,19 +443,21 @@ void RocksDBCollection::truncate(transaction::Methods* trx,
|
||||||
iter->Seek(documentBounds.start());
|
iter->Seek(documentBounds.start());
|
||||||
|
|
||||||
while (iter->Valid() && cmp->Compare(iter->key(), documentBounds.end()) < 0) {
|
while (iter->Valid() && cmp->Compare(iter->key(), documentBounds.end()) < 0) {
|
||||||
|
TRI_voc_rid_t revisionId = RocksDBKey::revisionId(iter->key());
|
||||||
|
|
||||||
|
// add possible log statement
|
||||||
|
state->prepareOperation(cid, revisionId, TRI_VOC_DOCUMENT_OPERATION_REMOVE);
|
||||||
rocksdb::Status s = rtrx->Delete(iter->key());
|
rocksdb::Status s = rtrx->Delete(iter->key());
|
||||||
if (!s.ok()) {
|
if (!s.ok()) {
|
||||||
auto converted = convertStatus(s);
|
auto converted = convertStatus(s);
|
||||||
THROW_ARANGO_EXCEPTION(converted);
|
THROW_ARANGO_EXCEPTION(converted);
|
||||||
}
|
}
|
||||||
|
|
||||||
// transaction size limit reached -- fail
|
|
||||||
TRI_voc_rid_t revisionId = RocksDBKey::revisionId(iter->key());
|
|
||||||
// report size of key
|
// report size of key
|
||||||
RocksDBOperationResult result =
|
RocksDBOperationResult result =
|
||||||
state->addOperation(cid, revisionId, TRI_VOC_DOCUMENT_OPERATION_REMOVE,
|
state->addOperation(cid, revisionId, TRI_VOC_DOCUMENT_OPERATION_REMOVE,
|
||||||
0, iter->key().size());
|
0, iter->key().size());
|
||||||
|
|
||||||
|
// transaction size limit reached -- fail
|
||||||
if (result.fail()) {
|
if (result.fail()) {
|
||||||
THROW_ARANGO_EXCEPTION(result);
|
THROW_ARANGO_EXCEPTION(result);
|
||||||
}
|
}
|
||||||
|
@ -615,6 +617,11 @@ int RocksDBCollection::insert(arangodb::transaction::Methods* trx,
|
||||||
RocksDBSavePoint guard(rocksTransaction(trx),
|
RocksDBSavePoint guard(rocksTransaction(trx),
|
||||||
trx->isSingleOperationTransaction());
|
trx->isSingleOperationTransaction());
|
||||||
|
|
||||||
|
RocksDBTransactionState* state =
|
||||||
|
static_cast<RocksDBTransactionState*>(trx->state());
|
||||||
|
state->prepareOperation(_logicalCollection->cid(), revisionId,
|
||||||
|
TRI_VOC_DOCUMENT_OPERATION_INSERT);
|
||||||
|
|
||||||
res = insertDocument(trx, revisionId, newSlice, options.waitForSync);
|
res = insertDocument(trx, revisionId, newSlice, options.waitForSync);
|
||||||
if (res.ok()) {
|
if (res.ok()) {
|
||||||
Result lookupResult = lookupRevisionVPack(revisionId, trx, mdr);
|
Result lookupResult = lookupRevisionVPack(revisionId, trx, mdr);
|
||||||
|
@ -624,11 +631,9 @@ int RocksDBCollection::insert(arangodb::transaction::Methods* trx,
|
||||||
}
|
}
|
||||||
|
|
||||||
// report document and key size
|
// report document and key size
|
||||||
RocksDBOperationResult result =
|
RocksDBOperationResult result = state->addOperation(
|
||||||
static_cast<RocksDBTransactionState*>(trx->state())
|
_logicalCollection->cid(), revisionId,
|
||||||
->addOperation(_logicalCollection->cid(), revisionId,
|
TRI_VOC_DOCUMENT_OPERATION_INSERT, newSlice.byteSize(), res.keySize());
|
||||||
TRI_VOC_DOCUMENT_OPERATION_INSERT,
|
|
||||||
newSlice.byteSize(), res.keySize());
|
|
||||||
|
|
||||||
// transaction size limit reached -- fail
|
// transaction size limit reached -- fail
|
||||||
if (result.fail()) {
|
if (result.fail()) {
|
||||||
|
@ -703,8 +708,9 @@ int RocksDBCollection::update(arangodb::transaction::Methods* trx,
|
||||||
mergeObjectsForUpdate(trx, oldDoc, newSlice, isEdgeCollection,
|
mergeObjectsForUpdate(trx, oldDoc, newSlice, isEdgeCollection,
|
||||||
TRI_RidToString(revisionId), options.mergeObjects,
|
TRI_RidToString(revisionId), options.mergeObjects,
|
||||||
options.keepNull, *builder.get());
|
options.keepNull, *builder.get());
|
||||||
|
RocksDBTransactionState* state =
|
||||||
if (trx->state()->isDBServer()) {
|
static_cast<RocksDBTransactionState*>(trx->state());
|
||||||
|
if (state->isDBServer()) {
|
||||||
// Need to check that no sharding keys have changed:
|
// Need to check that no sharding keys have changed:
|
||||||
if (arangodb::shardKeysChanged(_logicalCollection->dbName(),
|
if (arangodb::shardKeysChanged(_logicalCollection->dbName(),
|
||||||
trx->resolver()->getCollectionNameCluster(
|
trx->resolver()->getCollectionNameCluster(
|
||||||
|
@ -717,8 +723,10 @@ int RocksDBCollection::update(arangodb::transaction::Methods* trx,
|
||||||
RocksDBSavePoint guard(rocksTransaction(trx),
|
RocksDBSavePoint guard(rocksTransaction(trx),
|
||||||
trx->isSingleOperationTransaction());
|
trx->isSingleOperationTransaction());
|
||||||
|
|
||||||
|
// add possible log statement under guard
|
||||||
|
state->prepareOperation(_logicalCollection->cid(), revisionId,
|
||||||
|
TRI_VOC_DOCUMENT_OPERATION_UPDATE);
|
||||||
VPackSlice const newDoc(builder->slice());
|
VPackSlice const newDoc(builder->slice());
|
||||||
|
|
||||||
res = updateDocument(trx, oldRevisionId, oldDoc, revisionId, newDoc,
|
res = updateDocument(trx, oldRevisionId, oldDoc, revisionId, newDoc,
|
||||||
options.waitForSync);
|
options.waitForSync);
|
||||||
|
|
||||||
|
@ -731,10 +739,9 @@ int RocksDBCollection::update(arangodb::transaction::Methods* trx,
|
||||||
TRI_ASSERT(!mdr.empty());
|
TRI_ASSERT(!mdr.empty());
|
||||||
|
|
||||||
// report document and key size
|
// report document and key size
|
||||||
result = static_cast<RocksDBTransactionState*>(trx->state())
|
result = state->addOperation(_logicalCollection->cid(), revisionId,
|
||||||
->addOperation(_logicalCollection->cid(), revisionId,
|
TRI_VOC_DOCUMENT_OPERATION_UPDATE,
|
||||||
TRI_VOC_DOCUMENT_OPERATION_UPDATE,
|
newDoc.byteSize(), res.keySize());
|
||||||
newDoc.byteSize(), res.keySize());
|
|
||||||
|
|
||||||
// transaction size limit reached -- fail
|
// transaction size limit reached -- fail
|
||||||
if (result.fail()) {
|
if (result.fail()) {
|
||||||
|
@ -803,7 +810,9 @@ int RocksDBCollection::replace(
|
||||||
isEdgeCollection, TRI_RidToString(revisionId),
|
isEdgeCollection, TRI_RidToString(revisionId),
|
||||||
*builder.get());
|
*builder.get());
|
||||||
|
|
||||||
if (trx->state()->isDBServer()) {
|
RocksDBTransactionState* state =
|
||||||
|
static_cast<RocksDBTransactionState*>(trx->state());
|
||||||
|
if (state->isDBServer()) {
|
||||||
// Need to check that no sharding keys have changed:
|
// Need to check that no sharding keys have changed:
|
||||||
if (arangodb::shardKeysChanged(_logicalCollection->dbName(),
|
if (arangodb::shardKeysChanged(_logicalCollection->dbName(),
|
||||||
trx->resolver()->getCollectionNameCluster(
|
trx->resolver()->getCollectionNameCluster(
|
||||||
|
@ -816,6 +825,10 @@ int RocksDBCollection::replace(
|
||||||
RocksDBSavePoint guard(rocksTransaction(trx),
|
RocksDBSavePoint guard(rocksTransaction(trx),
|
||||||
trx->isSingleOperationTransaction());
|
trx->isSingleOperationTransaction());
|
||||||
|
|
||||||
|
// add possible log statement under guard
|
||||||
|
state->prepareOperation(_logicalCollection->cid(), revisionId,
|
||||||
|
TRI_VOC_DOCUMENT_OPERATION_REPLACE);
|
||||||
|
|
||||||
RocksDBOperationResult opResult =
|
RocksDBOperationResult opResult =
|
||||||
updateDocument(trx, oldRevisionId, oldDoc, revisionId,
|
updateDocument(trx, oldRevisionId, oldDoc, revisionId,
|
||||||
VPackSlice(builder->slice()), options.waitForSync);
|
VPackSlice(builder->slice()), options.waitForSync);
|
||||||
|
@ -829,11 +842,10 @@ int RocksDBCollection::replace(
|
||||||
TRI_ASSERT(!mdr.empty());
|
TRI_ASSERT(!mdr.empty());
|
||||||
|
|
||||||
// report document and key size
|
// report document and key size
|
||||||
result = static_cast<RocksDBTransactionState*>(trx->state())
|
result = state->addOperation(_logicalCollection->cid(), revisionId,
|
||||||
->addOperation(_logicalCollection->cid(), revisionId,
|
TRI_VOC_DOCUMENT_OPERATION_REPLACE,
|
||||||
TRI_VOC_DOCUMENT_OPERATION_REPLACE,
|
VPackSlice(builder->slice()).byteSize(),
|
||||||
VPackSlice(builder->slice()).byteSize(),
|
opResult.keySize());
|
||||||
opResult.keySize());
|
|
||||||
|
|
||||||
// transaction size limit reached -- fail
|
// transaction size limit reached -- fail
|
||||||
if (result.fail()) {
|
if (result.fail()) {
|
||||||
|
@ -901,13 +913,19 @@ int RocksDBCollection::remove(arangodb::transaction::Methods* trx,
|
||||||
RocksDBSavePoint guard(rocksTransaction(trx),
|
RocksDBSavePoint guard(rocksTransaction(trx),
|
||||||
trx->isSingleOperationTransaction());
|
trx->isSingleOperationTransaction());
|
||||||
|
|
||||||
|
// add possible log statement under guard
|
||||||
|
RocksDBTransactionState* state =
|
||||||
|
static_cast<RocksDBTransactionState*>(trx->state());
|
||||||
|
state->prepareOperation(_logicalCollection->cid(), revisionId,
|
||||||
|
TRI_VOC_DOCUMENT_OPERATION_REMOVE);
|
||||||
|
// RocksDBLogValue val = RocksDBLogValue::DocumentRemove(StringRef(key));
|
||||||
|
// state->rocksTransaction()->PutLogData(val.slice());
|
||||||
res = removeDocument(trx, oldRevisionId, oldDoc, options.waitForSync);
|
res = removeDocument(trx, oldRevisionId, oldDoc, options.waitForSync);
|
||||||
if (res.ok()) {
|
if (res.ok()) {
|
||||||
// report key size
|
// report key size
|
||||||
res =
|
res = state->addOperation(_logicalCollection->cid(), revisionId,
|
||||||
static_cast<RocksDBTransactionState*>(trx->state())
|
TRI_VOC_DOCUMENT_OPERATION_REMOVE, 0,
|
||||||
->addOperation(_logicalCollection->cid(), revisionId,
|
res.keySize());
|
||||||
TRI_VOC_DOCUMENT_OPERATION_REMOVE, 0, res.keySize());
|
|
||||||
// transaction size limit reached -- fail
|
// transaction size limit reached -- fail
|
||||||
if (res.fail()) {
|
if (res.fail()) {
|
||||||
THROW_ARANGO_EXCEPTION(res);
|
THROW_ARANGO_EXCEPTION(res);
|
||||||
|
|
|
@ -85,6 +85,10 @@ RocksDBLogValue RocksDBLogValue::ViewDrop(TRI_voc_cid_t cid,
|
||||||
return RocksDBLogValue(RocksDBLogType::ViewDrop, cid, iid);
|
return RocksDBLogValue(RocksDBLogType::ViewDrop, cid, iid);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
RocksDBLogValue RocksDBLogValue::DocumentOpsPrologue(TRI_voc_cid_t cid) {
|
||||||
|
return RocksDBLogValue(RocksDBLogType::DocumentOperationsPrologue, cid);
|
||||||
|
}
|
||||||
|
|
||||||
RocksDBLogValue RocksDBLogValue::DocumentRemove(
|
RocksDBLogValue RocksDBLogValue::DocumentRemove(
|
||||||
arangodb::StringRef const& key) {
|
arangodb::StringRef const& key) {
|
||||||
return RocksDBLogValue(RocksDBLogType::DocumentRemove, key);
|
return RocksDBLogValue(RocksDBLogType::DocumentRemove, key);
|
||||||
|
@ -105,7 +109,8 @@ RocksDBLogValue::RocksDBLogValue(RocksDBLogType type, uint64_t val)
|
||||||
: _buffer() {
|
: _buffer() {
|
||||||
switch (type) {
|
switch (type) {
|
||||||
case RocksDBLogType::DatabaseDrop:
|
case RocksDBLogType::DatabaseDrop:
|
||||||
case RocksDBLogType::CollectionCreate: {
|
case RocksDBLogType::CollectionCreate:
|
||||||
|
case RocksDBLogType::DocumentOperationsPrologue: {
|
||||||
_buffer.reserve(sizeof(RocksDBLogType) + sizeof(uint64_t));
|
_buffer.reserve(sizeof(RocksDBLogType) + sizeof(uint64_t));
|
||||||
_buffer += static_cast<char>(type);
|
_buffer += static_cast<char>(type);
|
||||||
uint64ToPersistent(_buffer, val); // database or collection ID
|
uint64ToPersistent(_buffer, val); // database or collection ID
|
||||||
|
|
|
@ -64,7 +64,7 @@ class RocksDBLogValue {
|
||||||
|
|
||||||
static RocksDBLogValue ViewCreate(TRI_voc_cid_t, TRI_idx_iid_t);
|
static RocksDBLogValue ViewCreate(TRI_voc_cid_t, TRI_idx_iid_t);
|
||||||
static RocksDBLogValue ViewDrop(TRI_voc_cid_t, TRI_idx_iid_t);
|
static RocksDBLogValue ViewDrop(TRI_voc_cid_t, TRI_idx_iid_t);
|
||||||
|
static RocksDBLogValue DocumentOpsPrologue(TRI_voc_cid_t cid);
|
||||||
static RocksDBLogValue DocumentRemove(arangodb::StringRef const&);
|
static RocksDBLogValue DocumentRemove(arangodb::StringRef const&);
|
||||||
|
|
||||||
public:
|
public:
|
||||||
|
|
|
@ -52,7 +52,7 @@
|
||||||
|
|
||||||
using namespace arangodb;
|
using namespace arangodb;
|
||||||
|
|
||||||
// for the RocksDB engine we do not need any additional data
|
// for the RocksDB engine we do not need any additional data
|
||||||
struct RocksDBTransactionData final : public TransactionData {};
|
struct RocksDBTransactionData final : public TransactionData {};
|
||||||
|
|
||||||
RocksDBSavePoint::RocksDBSavePoint(rocksdb::Transaction* trx)
|
RocksDBSavePoint::RocksDBSavePoint(rocksdb::Transaction* trx)
|
||||||
|
@ -98,7 +98,8 @@ RocksDBTransactionState::RocksDBTransactionState(
|
||||||
_numInserts(0),
|
_numInserts(0),
|
||||||
_numUpdates(0),
|
_numUpdates(0),
|
||||||
_numRemoves(0),
|
_numRemoves(0),
|
||||||
_intermediateTransactionEnabled(intermediateTransactionEnabled) {}
|
_intermediateTransactionEnabled(intermediateTransactionEnabled),
|
||||||
|
_lastUsedCollection(UINT64_MAX) {}
|
||||||
|
|
||||||
/// @brief free a transaction container
|
/// @brief free a transaction container
|
||||||
RocksDBTransactionState::~RocksDBTransactionState() {
|
RocksDBTransactionState::~RocksDBTransactionState() {
|
||||||
|
@ -111,9 +112,9 @@ RocksDBTransactionState::~RocksDBTransactionState() {
|
||||||
|
|
||||||
/// @brief start a transaction
|
/// @brief start a transaction
|
||||||
Result RocksDBTransactionState::beginTransaction(transaction::Hints hints) {
|
Result RocksDBTransactionState::beginTransaction(transaction::Hints hints) {
|
||||||
LOG_TRX(this, _nestingLevel) << "beginning " << AccessMode::typeString(_type)
|
LOG_TRX(this, _nestingLevel)
|
||||||
<< " transaction";
|
<< "beginning " << AccessMode::typeString(_type) << " transaction";
|
||||||
|
|
||||||
Result result = useCollections(_nestingLevel);
|
Result result = useCollections(_nestingLevel);
|
||||||
|
|
||||||
if (result.ok()) {
|
if (result.ok()) {
|
||||||
|
@ -157,23 +158,23 @@ Result RocksDBTransactionState::beginTransaction(transaction::Hints hints) {
|
||||||
_rocksWriteOptions, rocksdb::TransactionOptions()));
|
_rocksWriteOptions, rocksdb::TransactionOptions()));
|
||||||
_rocksTransaction->SetSnapshot();
|
_rocksTransaction->SetSnapshot();
|
||||||
_rocksReadOptions.snapshot = _rocksTransaction->GetSnapshot();
|
_rocksReadOptions.snapshot = _rocksTransaction->GetSnapshot();
|
||||||
|
|
||||||
RocksDBLogValue header = RocksDBLogValue::BeginTransaction(_vocbase->id(),
|
RocksDBLogValue header =
|
||||||
_id);
|
RocksDBLogValue::BeginTransaction(_vocbase->id(), _id);
|
||||||
_rocksTransaction->PutLogData(header.slice());
|
_rocksTransaction->PutLogData(header.slice());
|
||||||
|
|
||||||
} else {
|
} else {
|
||||||
TRI_ASSERT(_status == transaction::Status::RUNNING);
|
TRI_ASSERT(_status == transaction::Status::RUNNING);
|
||||||
}
|
}
|
||||||
|
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// @brief commit a transaction
|
/// @brief commit a transaction
|
||||||
Result RocksDBTransactionState::commitTransaction(
|
Result RocksDBTransactionState::commitTransaction(
|
||||||
transaction::Methods* activeTrx) {
|
transaction::Methods* activeTrx) {
|
||||||
LOG_TRX(this, _nestingLevel) << "committing " << AccessMode::typeString(_type)
|
LOG_TRX(this, _nestingLevel)
|
||||||
<< " transaction";
|
<< "committing " << AccessMode::typeString(_type) << " transaction";
|
||||||
|
|
||||||
TRI_ASSERT(_status == transaction::Status::RUNNING);
|
TRI_ASSERT(_status == transaction::Status::RUNNING);
|
||||||
TRI_IF_FAILURE("TransactionWriteCommitMarker") {
|
TRI_IF_FAILURE("TransactionWriteCommitMarker") {
|
||||||
|
@ -189,7 +190,7 @@ Result RocksDBTransactionState::commitTransaction(
|
||||||
_rocksWriteOptions.sync = true;
|
_rocksWriteOptions.sync = true;
|
||||||
_rocksTransaction->SetWriteOptions(_rocksWriteOptions);
|
_rocksTransaction->SetWriteOptions(_rocksWriteOptions);
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO wait for response on github issue to see how we can use the
|
// TODO wait for response on github issue to see how we can use the
|
||||||
// sequence number
|
// sequence number
|
||||||
result = rocksutils::convertStatus(_rocksTransaction->Commit());
|
result = rocksutils::convertStatus(_rocksTransaction->Commit());
|
||||||
|
@ -245,8 +246,8 @@ Result RocksDBTransactionState::commitTransaction(
|
||||||
/// @brief abort and rollback a transaction
|
/// @brief abort and rollback a transaction
|
||||||
Result RocksDBTransactionState::abortTransaction(
|
Result RocksDBTransactionState::abortTransaction(
|
||||||
transaction::Methods* activeTrx) {
|
transaction::Methods* activeTrx) {
|
||||||
LOG_TRX(this, _nestingLevel) << "aborting " << AccessMode::typeString(_type)
|
LOG_TRX(this, _nestingLevel)
|
||||||
<< " transaction";
|
<< "aborting " << AccessMode::typeString(_type) << " transaction";
|
||||||
TRI_ASSERT(_status == transaction::Status::RUNNING);
|
TRI_ASSERT(_status == transaction::Status::RUNNING);
|
||||||
Result result;
|
Result result;
|
||||||
|
|
||||||
|
@ -277,6 +278,26 @@ Result RocksDBTransactionState::abortTransaction(
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void RocksDBTransactionState::prepareOperation(
|
||||||
|
TRI_voc_cid_t collectionId, TRI_voc_rid_t revisionId,
|
||||||
|
TRI_voc_document_operation_e operationType) {
|
||||||
|
switch (operationType) {
|
||||||
|
case TRI_VOC_DOCUMENT_OPERATION_UNKNOWN:
|
||||||
|
break;
|
||||||
|
case TRI_VOC_DOCUMENT_OPERATION_INSERT:
|
||||||
|
case TRI_VOC_DOCUMENT_OPERATION_UPDATE:
|
||||||
|
case TRI_VOC_DOCUMENT_OPERATION_REPLACE:
|
||||||
|
case TRI_VOC_DOCUMENT_OPERATION_REMOVE: {
|
||||||
|
if (collectionId != _lastUsedCollection) {
|
||||||
|
RocksDBLogValue logValue =
|
||||||
|
RocksDBLogValue::DocumentOpsPrologue(collectionId);
|
||||||
|
//_rocksTransaction->PutLogData(logValue.slice());
|
||||||
|
_lastUsedCollection = collectionId;
|
||||||
|
}
|
||||||
|
} break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// @brief add an operation for a transaction collection
|
/// @brief add an operation for a transaction collection
|
||||||
RocksDBOperationResult RocksDBTransactionState::addOperation(
|
RocksDBOperationResult RocksDBTransactionState::addOperation(
|
||||||
TRI_voc_cid_t cid, TRI_voc_rid_t revisionId,
|
TRI_voc_cid_t cid, TRI_voc_rid_t revisionId,
|
||||||
|
@ -298,7 +319,7 @@ RocksDBOperationResult RocksDBTransactionState::addOperation(
|
||||||
static_cast<RocksDBTransactionCollection*>(findCollection(cid));
|
static_cast<RocksDBTransactionCollection*>(findCollection(cid));
|
||||||
|
|
||||||
if (collection == nullptr) {
|
if (collection == nullptr) {
|
||||||
std::string message = "collection '" + std::to_string(cid) +
|
std::string message = "collection '" + std::to_string(cid) +
|
||||||
"' not found in transaction state";
|
"' not found in transaction state";
|
||||||
THROW_ARANGO_EXCEPTION_MESSAGE(TRI_ERROR_INTERNAL, message);
|
THROW_ARANGO_EXCEPTION_MESSAGE(TRI_ERROR_INTERNAL, message);
|
||||||
}
|
}
|
||||||
|
|
|
@ -42,7 +42,7 @@ namespace rocksdb {
|
||||||
class Transaction;
|
class Transaction;
|
||||||
class Slice;
|
class Slice;
|
||||||
class Iterator;
|
class Iterator;
|
||||||
}
|
} // namespace rocksdb
|
||||||
|
|
||||||
namespace arangodb {
|
namespace arangodb {
|
||||||
namespace cache {
|
namespace cache {
|
||||||
|
@ -62,6 +62,7 @@ class RocksDBSavePoint {
|
||||||
~RocksDBSavePoint();
|
~RocksDBSavePoint();
|
||||||
|
|
||||||
void commit();
|
void commit();
|
||||||
|
|
||||||
private:
|
private:
|
||||||
void rollback();
|
void rollback();
|
||||||
|
|
||||||
|
@ -101,6 +102,9 @@ class RocksDBTransactionState final : public TransactionState {
|
||||||
return (_status == transaction::Status::ABORTED) && hasOperations();
|
return (_status == transaction::Status::ABORTED) && hasOperations();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void prepareOperation(TRI_voc_cid_t collectionId, TRI_voc_rid_t revisionId,
|
||||||
|
TRI_voc_document_operation_e operationType);
|
||||||
|
|
||||||
/// @brief add an operation for a transaction collection
|
/// @brief add an operation for a transaction collection
|
||||||
RocksDBOperationResult addOperation(
|
RocksDBOperationResult addOperation(
|
||||||
TRI_voc_cid_t collectionId, TRI_voc_rid_t revisionId,
|
TRI_voc_cid_t collectionId, TRI_voc_rid_t revisionId,
|
||||||
|
@ -135,7 +139,10 @@ class RocksDBTransactionState final : public TransactionState {
|
||||||
uint64_t _numUpdates;
|
uint64_t _numUpdates;
|
||||||
uint64_t _numRemoves;
|
uint64_t _numRemoves;
|
||||||
bool _intermediateTransactionEnabled;
|
bool _intermediateTransactionEnabled;
|
||||||
|
|
||||||
|
/// Last collection used for transaction
|
||||||
|
TRI_voc_cid_t _lastUsedCollection;
|
||||||
};
|
};
|
||||||
}
|
} // namespace arangodb
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -57,7 +57,8 @@ enum class RocksDBLogType : char {
|
||||||
ViewCreate = '9',
|
ViewCreate = '9',
|
||||||
ViewDrop = ':',
|
ViewDrop = ':',
|
||||||
ViewChange = ';',
|
ViewChange = ';',
|
||||||
DocumentRemove = '<'
|
DocumentOperationsPrologue = '<',
|
||||||
|
DocumentRemove = '='
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -734,8 +734,9 @@ static void CreateVocBase(v8::FunctionCallbackInfo<v8::Value> const& args,
|
||||||
}
|
}
|
||||||
|
|
||||||
std::unique_ptr<LogicalCollection> col =
|
std::unique_ptr<LogicalCollection> col =
|
||||||
ClusterMethods::createCollectionOnCoordinator(collectionType, vocbase,
|
ClusterMethods::createCollectionOnCoordinator(
|
||||||
infoSlice, true, createWaitsForSyncReplication);
|
collectionType, vocbase, infoSlice, false,
|
||||||
|
createWaitsForSyncReplication);
|
||||||
TRI_V8_RETURN(WrapCollection(isolate, col.release()));
|
TRI_V8_RETURN(WrapCollection(isolate, col.release()));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -499,15 +499,7 @@ std::string LogicalCollection::name() const {
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string const LogicalCollection::distributeShardsLike() const {
|
std::string const LogicalCollection::distributeShardsLike() const {
|
||||||
if (!_distributeShardsLike.empty()) {
|
return _distributeShardsLike;
|
||||||
CollectionNameResolver resolver(_vocbase);
|
|
||||||
TRI_voc_cid_t shardLike =
|
|
||||||
resolver.getCollectionIdCluster(_distributeShardsLike);
|
|
||||||
if (shardLike != 0) {
|
|
||||||
return basics::StringUtils::itoa(shardLike);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return "";
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void LogicalCollection::distributeShardsLike(std::string const& cid) {
|
void LogicalCollection::distributeShardsLike(std::string const& cid) {
|
||||||
|
|
|
@ -1687,7 +1687,7 @@ int32_t int32(std::string const& str) {
|
||||||
struct reent buffer;
|
struct reent buffer;
|
||||||
return _strtol_r(&buffer, str.c_str(), 0, 10);
|
return _strtol_r(&buffer, str.c_str(), 0, 10);
|
||||||
#else
|
#else
|
||||||
return strtol(str.c_str(), 0, 10);
|
return (int32_t)strtol(str.c_str(), 0, 10);
|
||||||
#endif
|
#endif
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
@ -1713,7 +1713,7 @@ int32_t int32(char const* value, size_t size) {
|
||||||
struct reent buffer;
|
struct reent buffer;
|
||||||
return _strtol_r(&buffer, value, 0, 10);
|
return _strtol_r(&buffer, value, 0, 10);
|
||||||
#else
|
#else
|
||||||
return strtol(value, 0, 10);
|
return (int32_t)strtol(value, 0, 10);
|
||||||
#endif
|
#endif
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
@ -1727,7 +1727,7 @@ uint32_t uint32(std::string const& str) {
|
||||||
struct reent buffer;
|
struct reent buffer;
|
||||||
return _strtoul_r(&buffer, str.c_str(), 0, 10);
|
return _strtoul_r(&buffer, str.c_str(), 0, 10);
|
||||||
#else
|
#else
|
||||||
return strtoul(str.c_str(), 0, 10);
|
return (uint32_t)strtoul(str.c_str(), 0, 10);
|
||||||
#endif
|
#endif
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
@ -1741,7 +1741,7 @@ uint32_t unhexUint32(std::string const& str) {
|
||||||
struct reent buffer;
|
struct reent buffer;
|
||||||
return _strtoul_r(&buffer, str.c_str(), 0, 16);
|
return _strtoul_r(&buffer, str.c_str(), 0, 16);
|
||||||
#else
|
#else
|
||||||
return strtoul(str.c_str(), 0, 16);
|
return (uint32_t)strtoul(str.c_str(), 0, 16);
|
||||||
#endif
|
#endif
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
@ -1767,7 +1767,7 @@ uint32_t uint32(char const* value, size_t size) {
|
||||||
struct reent buffer;
|
struct reent buffer;
|
||||||
return _strtoul_r(&buffer, value, 0, 10);
|
return _strtoul_r(&buffer, value, 0, 10);
|
||||||
#else
|
#else
|
||||||
return strtoul(value, 0, 10);
|
return (uint32_t)strtoul(value, 0, 10);
|
||||||
#endif
|
#endif
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
@ -1793,7 +1793,7 @@ uint32_t unhexUint32(char const* value, size_t size) {
|
||||||
struct reent buffer;
|
struct reent buffer;
|
||||||
return _strtoul_r(&buffer, value, 0, 16);
|
return _strtoul_r(&buffer, value, 0, 16);
|
||||||
#else
|
#else
|
||||||
return strtoul(value, 0, 16);
|
return (uint32_t)strtoul(value, 0, 16);
|
||||||
#endif
|
#endif
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
|
@ -105,11 +105,11 @@ static inline int TRI_bind(TRI_socket_t s, const struct sockaddr* address,
|
||||||
////////////////////////////////////////////////////////////////////////////////
|
////////////////////////////////////////////////////////////////////////////////
|
||||||
|
|
||||||
static inline int TRI_connect(TRI_socket_t s, const struct sockaddr* address,
|
static inline int TRI_connect(TRI_socket_t s, const struct sockaddr* address,
|
||||||
int addr_len) {
|
size_t addr_len) {
|
||||||
#ifdef _WIN32
|
#ifdef _WIN32
|
||||||
return connect(s.fileHandle, address, addr_len);
|
return connect(s.fileHandle, address, (int)addr_len);
|
||||||
#else
|
#else
|
||||||
return connect(s.fileDescriptor, address, addr_len);
|
return connect(s.fileDescriptor, address, (socklen_t)addr_len);
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -117,7 +117,7 @@ static inline int TRI_connect(TRI_socket_t s, const struct sockaddr* address,
|
||||||
/// @brief send abstraction for different OSes
|
/// @brief send abstraction for different OSes
|
||||||
////////////////////////////////////////////////////////////////////////////////
|
////////////////////////////////////////////////////////////////////////////////
|
||||||
|
|
||||||
static inline int TRI_send(TRI_socket_t s, const void* buffer, size_t length,
|
static inline long TRI_send(TRI_socket_t s, const void* buffer, size_t length,
|
||||||
int flags) {
|
int flags) {
|
||||||
#ifdef _WIN32
|
#ifdef _WIN32
|
||||||
return send(s.fileHandle, (char*)buffer, (int)length, flags);
|
return send(s.fileHandle, (char*)buffer, (int)length, flags);
|
||||||
|
|
|
@ -89,7 +89,7 @@ int gettimeofday(struct timeval* tv, void* tz) {
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
void TRI_localtime(time_t tt, struct tm* tb) {
|
void TRI_localtime(time_t tt, struct tm* tb) {
|
||||||
#ifdef TRI_HAVE_LOCALTIME_R
|
#ifdef ARANGODB_HAVE_LOCALTIME_R
|
||||||
localtime_r(&tt, tb);
|
localtime_r(&tt, tb);
|
||||||
#else
|
#else
|
||||||
#ifdef ARANGODB_HAVE_LOCALTIME_S
|
#ifdef ARANGODB_HAVE_LOCALTIME_S
|
||||||
|
@ -105,10 +105,10 @@ void TRI_localtime(time_t tt, struct tm* tb) {
|
||||||
}
|
}
|
||||||
|
|
||||||
void TRI_gmtime(time_t tt, struct tm* tb) {
|
void TRI_gmtime(time_t tt, struct tm* tb) {
|
||||||
#ifdef TRI_HAVE_GMTIME_R
|
#ifdef ARANGODB_HAVE_GMTIME_R
|
||||||
gmtime_r(&tt, tb);
|
gmtime_r(&tt, tb);
|
||||||
#else
|
#else
|
||||||
#ifdef TRI_HAVE_GMTIME_S
|
#ifdef ARANGODB_HAVE_GMTIME_S
|
||||||
gmtime_s(tb, &tt);
|
gmtime_s(tb, &tt);
|
||||||
#else
|
#else
|
||||||
struct tm* tp = gmtime(&tt);
|
struct tm* tp = gmtime(&tt);
|
||||||
|
|
|
@ -228,7 +228,7 @@ TRI_socket_t EndpointIp::connectSocket(const struct addrinfo* aip,
|
||||||
setTimeout(listenSocket, connectTimeout);
|
setTimeout(listenSocket, connectTimeout);
|
||||||
|
|
||||||
int result = TRI_connect(listenSocket, (const struct sockaddr*)aip->ai_addr,
|
int result = TRI_connect(listenSocket, (const struct sockaddr*)aip->ai_addr,
|
||||||
(int)aip->ai_addrlen);
|
aip->ai_addrlen);
|
||||||
|
|
||||||
if (result != 0) {
|
if (result != 0) {
|
||||||
pErr = STR_ERROR();
|
pErr = STR_ERROR();
|
||||||
|
|
|
@ -326,15 +326,15 @@ bool ClientConnection::writeClientConnection(void const* buffer, size_t length,
|
||||||
|
|
||||||
#if defined(__APPLE__)
|
#if defined(__APPLE__)
|
||||||
// MSG_NOSIGNAL not supported on apple platform
|
// MSG_NOSIGNAL not supported on apple platform
|
||||||
int status = TRI_send(_socket, buffer, length, 0);
|
long status = TRI_send(_socket, buffer, length, 0);
|
||||||
#elif defined(_WIN32)
|
#elif defined(_WIN32)
|
||||||
// MSG_NOSIGNAL not supported on windows platform
|
// MSG_NOSIGNAL not supported on windows platform
|
||||||
int status = TRI_send(_socket, buffer, length, 0);
|
long status = TRI_send(_socket, buffer, length, 0);
|
||||||
#elif defined(__sun)
|
#elif defined(__sun)
|
||||||
// MSG_NOSIGNAL not supported on solaris platform
|
// MSG_NOSIGNAL not supported on solaris platform
|
||||||
int status = TRI_send(_socket, buffer, length, 0);
|
long status = TRI_send(_socket, buffer, length, 0);
|
||||||
#else
|
#else
|
||||||
int status = TRI_send(_socket, buffer, length, MSG_NOSIGNAL);
|
long status = TRI_send(_socket, buffer, length, MSG_NOSIGNAL);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
if (status < 0) {
|
if (status < 0) {
|
||||||
|
|
Loading…
Reference in New Issue