* added proper UI support for collections using smartGraphAttribute or smartGraphAttribute
* changelog
* make key selection more clear
* added awesome ' : ' div
* fixed smart doc creation logic
* spawn an arangosh that checks cluster nodes are responsive
* attempting to kill 0 may end up bad for us, prohibit it
* don't trip over this optional stuff.
* fix killing of spectator
* only write one line per minute
* fix function name
* print sHitlist of tests, add resource usage
* fix lint
* start refactoring result processing into its own library
* /proc reading only works on linux
* new result processing library
* clean up test loading flow, remove unused blacklist implementation
* measure SUT start/stop + test time
* lint
* more non-test variables
* improve test runner naming
* finish gathering statistics about start/run/stop
* use internal stats code for external processes too
* fix status in sample testcase
* tell that procdump is gone - it seems this happenes in reality without coredumps being written
* refactor test result analyzing, add utility to work with analyzers of json dumps from CI systems
* fix testcase error handling
* also run watcher for agency
* lint
* fix default options for test added default options
* if arguments occur multiple times, update value to an array
* start implementing some test analyzers
* fix color
* write json report unconditional
* add analyzer that searches for long setup/teardown tests
* enable thread dump; log error if we fail to acquire the threads
* disable procdump for the agency
* output error if we fail to get process stats
* fix json invocation
* add debug logging
* only add buildType if that directory actually exists
* trap sleepers
* trap sleepers
* trap sleepers
* trap sleepers
* trap sleepers
* disable thread counting on the wintendo
* disable thread counting on the wintendo
* more measurements
* more measurements
* one more place
* one more place
* remove debugging code
* Update js/client/modules/@arangodb/process-utils.js
Co-Authored-By: Dan Larkin-York <danielhlarkin@users.noreply.github.com>
* Update js/client/modules/@arangodb/process-utils.js
Co-Authored-By: Dan Larkin-York <danielhlarkin@users.noreply.github.com>
* Apply suggestions from code review
Co-Authored-By: Dan Larkin-York <danielhlarkin@users.noreply.github.com>
* rename as sugested by @dan
* undo debug changes
* lint, make cluster health monitor optional per default
* fix spawning of active failover SUT, fix cluster health monitor shutdown
* use std::find_if (as @dan sugested), fix log ids
* fix scope of before-time
* Consolidated _servers and _serverAdvertisedEndpoints, added rebootId, prepared change notifications
* Cleanup
* Added a RebootId type
* Began implementing RebootTracker (still WIP)
* Moved RebootId operators into the class
* Removed RebootId operator<< again
* Added tests, added CallbackGuard, removed/commented old RebootTracker code
* Fix: do not try to call unset callbacks
* Split one test, added another
* Added more tests
* Renamed tests, added more tests
* Fixed missing variable declarations
* Let MockServer appear to be started
* Reorded test, fixed naming
* Implemented callMeOnChange()
* Re-implemented RebootTracker (not yet working)
* Resolved a TODO, updated a test, added comments
* Call old callbacks immediately
* Fixed tests
* Use EXPECT_* instead of ASSERT_*
* Suppress a log message
* Resolved TODOs
* Reverted changes on reading ServersRegistered
* Update RebootTracker
* Introduce `rebootId` into ServerState for Cluster
* A server *boots* if it is started on a previously non-existing data
directory and hence does not have a UUID yet.
* A server *reboots* if it is started on a pre-existing data directory
We keep the rebootId in the cluster's agency under
Current/ServersKnown/$uuid/rebootId.
When rebooting (and subsequently re-joining a cluster), the server increments
its rebootId in Phase 2 of registration. This way it can be detected within the
cluster whether a server was restarted.
This information will later be used to handle cases where server restarts can
lead to problems, for example with transactions or in-progress queries.
* Move rebootId into Current/ServersKnown/
* Fixed typo
* Fixed log ids
* Add deletion of ServersKnown/UUID from agency
* Add deletion of Current/ServersKnown/UUID to removeServer
* Clean up readRebootIdFromAgency and add retry loop around it
* Bugfix
* Added nolint comments
* Fixed initialization order
* Fixed ClusterInfo-test
* Added log messages
* Revert "Fixed ClusterInfo-test"
This reverts commit d983596979.
* Disabled assertion for google tests
* Ignore windows compile warning
* Always call loadServers in loadCurrent
* Fix really subtle bug when not returning a value
* Introduce `rebootId` into ServerState for Cluster
* A server *boots* if it is started on a previously non-existing data
directory and hence does not have a UUID yet.
* A server *reboots* if it is started on a pre-existing data directory
We keep the rebootId in the cluster's agency under
Current/ServersKnown/$uuid/rebootId.
When rebooting (and subsequently re-joining a cluster), the server increments
its rebootId in Phase 2 of registration. This way it can be detected within the
cluster whether a server was restarted.
This information will later be used to handle cases where server restarts can
lead to problems, for example with transactions or in-progress queries.
* Move rebootId into Current/ServersKnown/
* Add deletion of ServersKnown/UUID from agency
* Add deletion of Current/ServersKnown/UUID to removeServer
* Clean up readRebootIdFromAgency and add retry loop around it
* Fixed compile error due to forbidden implicit cast
* Fixed compile error on windows
* Fixed compile error due to devel merge
* Removed dead comment
* Removed TODO note
* Extended comment
* Removed TODO note
* Fixed using an invalidated iterator
* Copy string only if necessary
* Fixed compile error
* First version of ResignLeadership Job.
* Port some performance optimizations from CleanOutServerJob.
* Draft of resigning leadership on shutdown.
* Moved code into Maintenance Feature. Fixed beginShutdown.
* remove http so https arangos will work
* verify that query parameters are proper swagger data types, fix offending documentation files
* return the actual type - not the list of available ones
* check formats
* there is no uint64 in swagger
* fresh swagger
* port of feature-3.4/mv-gzip-export to devel branch
* add explicit namespaces so gcc 6.3.0 would successfully compile
* add conditional cleanup of ENCRYPTION file from another branch
* use _lseek() for Windows build to avoid deprecated warning.
* change from ifdef for lseek variants to TRI_LSEEK.
* force Windows lseek to return Linux expected type.
* Added startup error for bad temporary directory setting.
If the temporary directory (--temp.path) setting is identical to the database
directory (`--database.directory`) this can eventually lead to data loss, as
temporary files may be created inside the temporary directory, causing overwrites of
existing database files/directories with the same names.
Additionally the temporary directory may be cleaned at some point, and this would lead
to an unintended cleanup of the database files/directories as well.
Now, if the database directory and temporary directory are set to the same path, there
will be a startup warning about potential data loss (though in ArangoDB 3.4 allowing to
continue the startup - in 3.5 and higher we will abort the startup).
* Bug fix 3.5/min replication factor (#9524)
* Cherry-pick minReplicationFactor
* Bug fix/failover with min replication factor (#9486)
* Improve collection time of IResearchQueryOptimizationTest
* Added a minReplicationFactor field in Collections. It is not possible to modify it yet and noone cares for it
* Added some assertion son minReplicationFactor
* Transaction API will now reject writes as soon as minimal replication factor is NOT fulfilled
* added minReplicationFactor to the user interface, preparation for the collection api changes
* added minReplicationFactor to VocBaseCollection, RestReplicationHandler, RestCollectionHandler, ClusterMethods, ClusterInfo and ClusterCollectionCreationInfo
* added minReplicationFactor usage to tests
* TODO TEMOPORARY COMMIT FOR TESTING PLEASE REVERT ME
* minReplicationFactor now able to change via collection properties route
* fixed wrongly assert
* added minReplicationFactor to the graph management ui
* added minReplicationFactor to the gharial api
* Fixed off-by-one error in minReplicationFactor. We actually enforced one more.
* adjusted description of minReplicationFactor
* FollowerInfo Refactoring
* added gharial api graph creation tests with minimal replication factor
* proper cleanup of shell collection tests, removed lots of duplicate code, preparation for some new tests
* added collection create tests using invalid/valid names, replicationFactor and minReplicationFactor
* Debug logging
* MORE Debug logging
* Included replication fast lane
* Use correct minreplicationfactor
* modified debug logging
* Fixed compileissues
* MORE Debug logging
* MORE Debug logging
* MORE Debug logging
* MORE Debug logging
* MORE Debug logging
* MORE Debug logging
* MORE Debug logging
* Revert "MORE Debug logging"
This reverts commit dab5af28c0.
* Revert "MORE Debug logging"
This reverts commit 6134b664bd.
* Revert "MORE Debug logging"
This reverts commit 80160bdf3b.
* Revert "MORE Debug logging"
This reverts commit 06aabcdfe1.
* Removed debug output
* Added replication fast lane. Also refactored the commands as i cannot take it any more...
* Put some requests of RocksDBReplication onto CATCHUP Lane.
* Put some requests of MMFilesReplication onto CATCHUP Lane.
* Adjusted Fast and MED lane usage in Supervised scheduler
* Added changelog entry
* Added new features entry
* A new leader will now keep old followers in case of failover
* Update arangod/Cluster/ClusterCollectionCreationInfo.cpp
Co-Authored-By: Tobias Gödderz <tobias@arangodb.com>
* Fixed JSLINT
* Unified lane handling of replication handlers
* Sorry forgotten in last commit
* replaced strings with static strings
* more use of static strings
* optimized min repl description in the ui
* decr initial loop variable
* clean up of the createWithId test
* more use of static strings
* Update js/apps/system/_admin/aardvark/APP/frontend/js/views/collectionsView.js
Co-Authored-By: Tobias Gödderz <tobias@arangodb.com>
* Added some comments on condition, renamed variable as suggested in review
* Added check for min replicationFactor to be non-zero
* Added assertion
* Added function to modify min and max replication factor in one go
* added missing semicolon
* rm log devel
* Added a second information to follower info that can keep track of followers that have been in sync before a failover has taken place
* Maintenance reports previous version now to follower info. instead of lying by itself. The Follower Info now gets a failover save mode to report insync followers
* check replFactor against nr dbservers
* Add lie reporting in CURRENT
* Reverted most of my recent commits about Failover situation. The intended plan simply does not work out
* move replication checks from logical collection to rest collection handler
* added more replication tests
* Include assert only if we are not in gtest
* jslint
* set min repl factor to zero if satellite collection
* check replication attributes in v8 collection
* Initial commit, old plan, does not yet work
* fixed ires tests
* Included FailoverCandidates key. Not fully implemented
* fixed wrong assert
* unified in sync follower reporting
* fixed compiler errors
* Cleanup locking, and fixed potential deadlocks
* Comments about locking order in FollowerInfo.
* properly check uint
* Keep old leader as potential failover candidate
* Transaction methods now use followerInfo to check if the leader can write, this might have the sideeffect that 'failoverCandidates' are updated
* Let agency check failoverCandidates if possible
* Initialize member variables
* Use unified follower reporting in DBServerAgencySync
* Removed obsolete variable, collecting it somewhere else
* repl factor attr check
* Reimplemented previous followers, second attempt now. PhaseOne and PhaseTwo can now synchronize on current.
* Fixed assertion, forgot an off-by-one
* adjusted test to be more preciese now
* Fixed failove candidates list
* Disable write on dropping too many followers
* Allow to run updateFailoerCandidates multiple times with same leader.
* Final fixes, resilience tests now green, crossing fingers for jenkins
* Fixed race on atomics comparison
* Fixed invalid number type
* added nullptr handling
* added nullptr handling
* Removed invalid assert
* Make takeover of leadership an atomic operation
* Update tests/js/common/shell/shell-cluster-collection.js
Co-Authored-By: Tobias Gödderz <tobias@arangodb.com>
* Review fixes
* Fixed creation code to use takeoverLeadership
* Update arangod/Cluster/FollowerInfo.h
Co-Authored-By: Tobias Gödderz <tobias@arangodb.com>
* Applied review fixes
* There is no timeout
* Moved AQL + Pregel to INTERNAL_AQL lane, which is medium priority, to avoid deadlocks with Sync replication
* More review fixes
* Use difference if you want to compare two vectors...
* Use std::string ...
* Now check if we are in recovery mode
* Added documentation for minReplicationFactor
* Added readme update as well in documenation
* Removed merge conflict leftovers 0o, i should not trust the IDE
* Update js/apps/system/_admin/aardvark/APP/frontend/js/views/collectionsView.js
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Update js/apps/system/_admin/aardvark/APP/frontend/js/views/collectionsView.js
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Update Documentation/Books/Manual/Architecture/Replication/README.md
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Update CHANGELOG
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Update Documentation/Books/Manual/DataModeling/Collections/DatabaseMethods.md
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Update Documentation/Books/Manual/ReleaseNotes/NewFeatures35.md
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Update Documentation/DocuBlocks/Rest/Collections/1_structs.md
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Update js/apps/system/_admin/aardvark/APP/frontend/js/views/graphManagementView.js
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Update js/apps/system/_admin/aardvark/APP/frontend/js/views/graphManagementView.js
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Update Documentation/DocuBlocks/Rest/Graph/1_structs.md
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Apply suggestions from code review
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Adepted review requests, thanks for finding!
* Removed unnecessary const
* Apply suggestions from code review
Co-Authored-By: Jan <jsteemann@users.noreply.github.com>
* Moved initilization of variable more downwards
* Apply lock before notify_all()
* Remove documentation except DocuBlocks, covered by PR in docs repo
* Remove accidental indent
* Removed leftover merge conflict in documentation block
* Added a minReplicationFactor field in Collections. It is not possible to modify it yet and noone cares for it
* Added some assertion son minReplicationFactor
* Transaction API will now reject writes as soon as minimal replication factor is NOT fulfilled
* added minReplicationFactor to the user interface, preparation for the collection api changes
* added minReplicationFactor to VocBaseCollection, RestReplicationHandler, RestCollectionHandler, ClusterMethods, ClusterInfo and ClusterCollectionCreationInfo
* added minReplicationFactor usage to tests
* TODO TEMOPORARY COMMIT FOR TESTING PLEASE REVERT ME
* minReplicationFactor now able to change via collection properties route
* fixed wrongly assert
* added minReplicationFactor to the graph management ui
* added minReplicationFactor to the gharial api
* Fixed off-by-one error in minReplicationFactor. We actually enforced one more.
* adjusted description of minReplicationFactor
* FollowerInfo Refactoring
* added gharial api graph creation tests with minimal replication factor
* proper cleanup of shell collection tests, removed lots of duplicate code, preparation for some new tests
* added collection create tests using invalid/valid names, replicationFactor and minReplicationFactor
* Debug logging
* MORE Debug logging
* Included replication fast lane
* Use correct minreplicationfactor
* modified debug logging
* Fixed compileissues
* MORE Debug logging
* MORE Debug logging
* MORE Debug logging
* MORE Debug logging
* MORE Debug logging
* MORE Debug logging
* MORE Debug logging
* Revert "MORE Debug logging"
This reverts commit dab5af28c0.
* Revert "MORE Debug logging"
This reverts commit 6134b664bd.
* Revert "MORE Debug logging"
This reverts commit 80160bdf3b.
* Revert "MORE Debug logging"
This reverts commit 06aabcdfe1.
* Removed debug output
* Added replication fast lane. Also refactored the commands as i cannot take it any more...
* Put some requests of RocksDBReplication onto CATCHUP Lane.
* Put some requests of MMFilesReplication onto CATCHUP Lane.
* Adjusted Fast and MED lane usage in Supervised scheduler
* Added changelog entry
* Added new features entry
* A new leader will now keep old followers in case of failover
* Update arangod/Cluster/ClusterCollectionCreationInfo.cpp
Co-Authored-By: Tobias Gödderz <tobias@arangodb.com>
* Fixed JSLINT
* Unified lane handling of replication handlers
* Sorry forgotten in last commit
* replaced strings with static strings
* more use of static strings
* optimized min repl description in the ui
* decr initial loop variable
* clean up of the createWithId test
* more use of static strings
* Update js/apps/system/_admin/aardvark/APP/frontend/js/views/collectionsView.js
Co-Authored-By: Tobias Gödderz <tobias@arangodb.com>
* Added some comments on condition, renamed variable as suggested in review
* Added check for min replicationFactor to be non-zero
* Added assertion
* Added function to modify min and max replication factor in one go
* added missing semicolon
* rm log devel
* Added a second information to follower info that can keep track of followers that have been in sync before a failover has taken place
* Maintenance reports previous version now to follower info. instead of lying by itself. The Follower Info now gets a failover save mode to report insync followers
* check replFactor against nr dbservers
* Add lie reporting in CURRENT
* Reverted most of my recent commits about Failover situation. The intended plan simply does not work out
* move replication checks from logical collection to rest collection handler
* added more replication tests
* Include assert only if we are not in gtest
* jslint
* set min repl factor to zero if satellite collection
* check replication attributes in v8 collection
* fixed ires tests
* fixed wrong assert
* properly check uint
* repl factor attr check
* adjusted test to be more preciese now
* Fixed race on atomics comparison
* Fixed invalid number type
* Update tests/js/common/shell/shell-cluster-collection.js
Co-Authored-By: Tobias Gödderz <tobias@arangodb.com>
* Review fixes
* More review fixes
* migrate mochaGrep into --testCase parametrizing
* fix filter forwarding for spawned arangoshs - add mocha version
* sometimes we have the string undefined to ignore
* fix location - while they work with the server these tests are run in arangosh so they need to be in client/
* rename testsuite
* rename file as its testsuite
* Rename permissions_server to server_permissions
* apply filters before starting the server, so we can detect whether no test would be executed
* skipTimecritial and skipNondeterministic also qualify for non-errnous empty testcase
* simplify code, if is issued no matching testcase is an error
* Update test-utils.js
* add analyser http API tests
* add more tests
* add tests and lint
* remove print
* add another test
* fix formattig
* add e6 expresseion as key in object
* WIP: add analyzer permissions test
* fix test
* lint
* WIP: try to add analyzer auth-test
* add sleepBeforeStart option
* use basic auth
* fix tests
* change test so it would work if the api was as expected
* fix authentication test
* fit shell client analyzer test
* fix shell-analyzer-rest-api.js
* fix auth-analyzer.js
* fix user-access-right-update-view-arangosearch.js
* remove print statements
* Fixed analyzer parameters for test
* make rspec timeouteable (on windows for now)
* add timeout state, fix windows implementation
* abort testrun if a timeout in rspec occured
* lint
* fix scope of procdump stopper
* fix include scope
* fix invokation
* use proper variable to wait for
* add output before we attempt to talk to the SUT, so we can better insulate startup lockups
* when we timeout a process dump out all we know about the SUT.
* when we timeout a process dump out all we know about the SUT.
* make rspec timeouteable (on windows for now)
* add timeout state, fix windows implementation
* abort testrun if a timeout in rspec occured
* lint
* fix scope of procdump stopper
* fix include scope
* fix invokation
* use proper variable to wait for
* Add failing tests
* Add tests for Buffer#slice
* Add tests for Buffer#new
* Skip failing tests
* Implement changes
* Unskip failing tests
This reverts commit 9f8a34bd74e152b7b3158fea79b12e6da8343001.
* Fix implementation
* Remove meaningless test
This test only succeeded because of a broken implementation of the Buffer constructor. It doesn't really do anything useful.
* Add note to CHANGELOG
* preserve the instance shutdown state, and if its bad, don't clean up the test directory
* Update js/client/modules/@arangodb/testing.js
Co-Authored-By: Tobias Gödderz <github@tobias.goedderz.info>
* Update js/client/modules/@arangodb/testsuites/permissions_server.js
Co-Authored-By: Tobias Gödderz <github@tobias.goedderz.info>
* Update js/client/modules/@arangodb/testsuites/permissions_server.js
Co-Authored-By: Tobias Gödderz <github@tobias.goedderz.info>
* Update js/client/modules/@arangodb/testsuites/single.js
Co-Authored-By: Tobias Gödderz <github@tobias.goedderz.info>
* fix handling in custom start/stop handlers
* remove shutdown state from reporting structure
* fix shutdown of second instance
* one more to adjust
* only add jwt if we run the agency with authentification
* lint
* Began work on generic traversal testing framework
* Translate shard ids to smart attributes
* Minor fixes
* Added a TODO note
* fixed tests
* cleanup
* first tests inside new traversal framework working
* added single server diamond tests
* Enable sharding in aql-graph-traversal-generic-graphs.js
* Tried to fix smart sharding
* Fixed smart graph name clashes, some cleanup
* Added meta tests
* Added TODO note
* Fixed single server
* Added a BFS order check
* Renamed test
* added more dfs and bfs tests for the openDiamond graph
* Added new graph "largeBinTree"
* Fix Map usage
* added smallCircle graph, added three dfs tests
* added smallCircle graph, added three dfs tests
* more smallCircle dfs tests
* Fixed largeBinTree shardings, and long graph names
* Fix long suffixes
* Provided names for shardings
* Added a log message
* more bfs smallCircle tests
* more bfs smallCircle tests - global
* smallCircle picture
* added completeGraph to graph test suite
* added more dfs tests for the complete graph up to depth 3
* Removed duplicate function definitions
* added depth 2 dfs complete graph test, added missing test to the exports
* Added test for largeBinTree
* added three path graphs, easy, advanced and more advanced
* Added a working check for bfs with global uniqueness
* Fixed usage of bfs checks
* Fixed expected input for two tests
* Added some static assertions to largeBinTree
* Oops
* added more advancedPath dfs tests
* better formatting of the visual graph
* more dfs advanced graph tests
* Implemented all easy path test combinations
* Fixed testAqlGraphQuery
* Fix in bfs check error reports
* Removed duplicate test
* Added advancedPath BFS tests
* Updated comments
* Fixed new tests
* Fixed graph description
* added more complete graph bfs tests
* added more complete graph bfs tests
* added testCompleteGraphBfsUniqueEdgesNoneD1 and testCompleteGraphBfsUniqueEdgesNoneD2
* added testCompleteGraphBfsUniqueVerticesUniqueEdgesPathD3 test
* added a few more complete graph bfs checks
* Extended a comment
* Unbreak aardvark when --foxx.api false
The easiest way to disable the aardvark foxx routes is to just skip the
import and mount altogether. The previous approach caused the service to
throw an error because the module doesn't actually export anything.
There's no need to prevent the export if we just don't import the module
to begin with.
* Add an error code for service api disabled
* Throw an actual error when accessing Foxx API routes
* Throw an error when using legacy Foxx API
There's no easy way to use a catch-all here so disable them one by one. No new routes will be added to this API and it'll go away in 4.0 so there's no risk of forgetting to add the throw-check in new routes.
* Render subnav without delay
The subnav doesn't need the perm check results but we need the subnav to exist so we can populate it in other parts of the app. This fixes the subnav not showing up when reloading certain views (e.g. install service via upload).
* Hide ONLY the Foxx store when it is disabled
We still want to allow installing and upgrading/replacing services. We just don't want to show the store.
* Hide Foxx services in frontend if Foxx API is disabled
* Build aardvark
* fixed typo in disable check, fixed tests