1
0
Fork 0

Doc - Fix build (whitespaces, triple-backticks, comments) (#7857)

This commit is contained in:
Simran 2018-12-26 12:41:19 -08:00 committed by sleto-it
parent 3bf3818ba9
commit f7b9733a84
11 changed files with 253 additions and 247 deletions

View File

@ -277,7 +277,7 @@ the vertex IDs:
@startDocuBlockInline GRAPHTRAV_02_traverse_all_a @startDocuBlockInline GRAPHTRAV_02_traverse_all_a
@EXAMPLE_AQL{GRAPHTRAV_02_traverse_all_a} @EXAMPLE_AQL{GRAPHTRAV_02_traverse_all_a}
@DATASET{traversalGraph} @DATASET{traversalGraph}
FOR v IN 1..3 OUTBOUND 'circles/A' GRAPH 'traversalGraph' FOR v IN 1..3 OUTBOUND 'circles/A' GRAPH 'traversalGraph'
RETURN v._key RETURN v._key
@END_EXAMPLE_AQL @END_EXAMPLE_AQL
@endDocuBlock GRAPHTRAV_02_traverse_all_a @endDocuBlock GRAPHTRAV_02_traverse_all_a

View File

@ -70,7 +70,6 @@ and maximum tick values per logfile:
require("@arangodb/replication").logger.tickRanges(); require("@arangodb/replication").logger.tickRanges();
### Replication Applier ### Replication Applier
**Purpose** **Purpose**

View File

@ -40,31 +40,31 @@ So in summary these are the commands to start an _Agency_ of size 3:
``` ```
arangod --server.endpoint tcp://0.0.0.0:5001 \ arangod --server.endpoint tcp://0.0.0.0:5001 \
--agency.my-address=tcp://127.0.0.1:5001 \ --agency.my-address=tcp://127.0.0.1:5001 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.endpoint tcp://127.0.0.1:5001 \ --agency.endpoint tcp://127.0.0.1:5001 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent1 & --database.directory agent1 &
arangod --server.endpoint tcp://0.0.0.0:5002 \ arangod --server.endpoint tcp://0.0.0.0:5002 \
--agency.my-address=tcp://127.0.0.1:5002 \ --agency.my-address=tcp://127.0.0.1:5002 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.endpoint tcp://127.0.0.1:5001 \ --agency.endpoint tcp://127.0.0.1:5001 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent2 & --database.directory agent2 &
arangod --server.endpoint tcp://0.0.0.0:5003 \ arangod --server.endpoint tcp://0.0.0.0:5003 \
--agency.my-address=tcp://127.0.0.1:5003 \ --agency.my-address=tcp://127.0.0.1:5003 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.endpoint tcp://127.0.0.1:5001 \ --agency.endpoint tcp://127.0.0.1:5001 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent3 & --database.directory agent3 &
``` ```
### Single Server Test Instances ### Single Server Test Instances
@ -73,24 +73,24 @@ To start the two single server instances, you can use the following commands:
``` ```
arangod --server.authentication false \ arangod --server.authentication false \
--server.endpoint tcp://127.0.0.1:6001 \ --server.endpoint tcp://127.0.0.1:6001 \
--cluster.my-address tcp://127.0.0.1:6001 \ --cluster.my-address tcp://127.0.0.1:6001 \
--cluster.my-role SINGLE \ --cluster.my-role SINGLE \
--cluster.agency-endpoint tcp://127.0.0.1:5001 \ --cluster.agency-endpoint tcp://127.0.0.1:5001 \
--cluster.agency-endpoint tcp://127.0.0.1:5002 \ --cluster.agency-endpoint tcp://127.0.0.1:5002 \
--cluster.agency-endpoint tcp://127.0.0.1:5003 \ --cluster.agency-endpoint tcp://127.0.0.1:5003 \
--replication.automatic-failover true \ --replication.automatic-failover true \
--database.directory singleserver6001 & --database.directory singleserver6001 &
arangod --server.authentication false \ arangod --server.authentication false \
--server.endpoint tcp://127.0.0.1:6002 \ --server.endpoint tcp://127.0.0.1:6002 \
--cluster.my-address tcp://127.0.0.1:6002 \ --cluster.my-address tcp://127.0.0.1:6002 \
--cluster.my-role SINGLE \ --cluster.my-role SINGLE \
--cluster.agency-endpoint tcp://127.0.0.1:5001 \ --cluster.agency-endpoint tcp://127.0.0.1:5001 \
--cluster.agency-endpoint tcp://127.0.0.1:5002 \ --cluster.agency-endpoint tcp://127.0.0.1:5002 \
--cluster.agency-endpoint tcp://127.0.0.1:5003 \ --cluster.agency-endpoint tcp://127.0.0.1:5003 \
--replication.automatic-failover true \ --replication.automatic-failover true \
--database.directory singleserver6002 & --database.directory singleserver6002 &
``` ```
Multiple Machines Multiple Machines
@ -127,39 +127,39 @@ On 192.168.1.1:
``` ```
arangod --server.endpoint tcp://0.0.0.0:8531 \ arangod --server.endpoint tcp://0.0.0.0:8531 \
--agency.my-address tcp://192.168.1.1:8531 \ --agency.my-address tcp://192.168.1.1:8531 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent --database.directory agent
``` ```
On 192.168.1.2: On 192.168.1.2:
``` ```
arangod --server.endpoint tcp://0.0.0.0:8531 \ arangod --server.endpoint tcp://0.0.0.0:8531 \
--agency.my-address tcp://192.168.1.2:8531 \ --agency.my-address tcp://192.168.1.2:8531 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent --database.directory agent
``` ```
On 192.168.1.3: On 192.168.1.3:
``` ```
arangod --server.endpoint tcp://0.0.0.0:8531 \ arangod --server.endpoint tcp://0.0.0.0:8531 \
--agency.my-address tcp://192.168.1.3:8531 \ --agency.my-address tcp://192.168.1.3:8531 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.endpoint tcp://192.168.1.1:8531 \ --agency.endpoint tcp://192.168.1.1:8531 \
--agency.endpoint tcp://192.168.1.2:8531 \ --agency.endpoint tcp://192.168.1.2:8531 \
--agency.endpoint tcp://192.168.1.3:8531 \ --agency.endpoint tcp://192.168.1.3:8531 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent --database.directory agent
``` ```
### Single Server Instances ### Single Server Instances
@ -168,14 +168,14 @@ On 192.168.1.1:
``` ```
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:8529 \ --server.endpoint tcp://0.0.0.0:8529 \
--cluster.my-address tcp://192.168.1.1:8529 \ --cluster.my-address tcp://192.168.1.1:8529 \
--cluster.my-role SINGLE \ --cluster.my-role SINGLE \
--cluster.agency-endpoint tcp://192.168.1.1:8531 \ --cluster.agency-endpoint tcp://192.168.1.1:8531 \
--cluster.agency-endpoint tcp://192.168.1.2:8531 \ --cluster.agency-endpoint tcp://192.168.1.2:8531 \
--cluster.agency-endpoint tcp://192.168.1.3:8531 \ --cluster.agency-endpoint tcp://192.168.1.3:8531 \
--replication.automatic-failover true \ --replication.automatic-failover true \
--database.directory singleserver & --database.directory singleserver &
``` ```
On 192.168.1.2: On 192.168.1.2:
@ -185,14 +185,14 @@ instance:
``` ```
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:8529 \ --server.endpoint tcp://0.0.0.0:8529 \
--cluster.my-address tcp://192.168.1.2:8529 \ --cluster.my-address tcp://192.168.1.2:8529 \
--cluster.my-role SINGLE \ --cluster.my-role SINGLE \
--cluster.agency-endpoint tcp://192.168.1.1:8531 \ --cluster.agency-endpoint tcp://192.168.1.1:8531 \
--cluster.agency-endpoint tcp://192.168.1.2:8531 \ --cluster.agency-endpoint tcp://192.168.1.2:8531 \
--cluster.agency-endpoint tcp://192.168.1.3:8531 \ --cluster.agency-endpoint tcp://192.168.1.3:8531 \
--replication.automatic-failover true \ --replication.automatic-failover true \
--database.directory singleserver & --database.directory singleserver &
``` ```
**Note:** in the above commands, you can use host names, if they can be resolved, **Note:** in the above commands, you can use host names, if they can be resolved,
@ -220,13 +220,13 @@ An example configuration might look like this:
``` ```
docker run -e ARANGO_NO_AUTH=1 -p 192.168.1.1:10000:8529 arangodb/arangodb arangod \ docker run -e ARANGO_NO_AUTH=1 -p 192.168.1.1:10000:8529 arangodb/arangodb arangod \
--server.endpoint tcp://0.0.0.0:8529\ --server.endpoint tcp://0.0.0.0:8529\
--cluster.my-address tcp://192.168.1.1:10000 \ --cluster.my-address tcp://192.168.1.1:10000 \
--cluster.my-role SINGLE \ --cluster.my-role SINGLE \
--cluster.agency-endpoint tcp://192.168.1.1:9001 \ --cluster.agency-endpoint tcp://192.168.1.1:9001 \
--cluster.agency-endpoint tcp://192.168.1.2:9001 \ --cluster.agency-endpoint tcp://192.168.1.2:9001 \
--cluster.agency-endpoint tcp://192.168.1.3:9001 \ --cluster.agency-endpoint tcp://192.168.1.3:9001 \
--replication.automatic-failover true --replication.automatic-failover true
``` ```
This will start a single server within a Docker container with an isolated network. This will start a single server within a Docker container with an isolated network.

View File

@ -41,31 +41,31 @@ So in summary these are the commands to start an _Agency_ of size 3:
``` ```
arangod --server.endpoint tcp://0.0.0.0:5001 \ arangod --server.endpoint tcp://0.0.0.0:5001 \
--agency.my-address=tcp://127.0.0.1:5001 \ --agency.my-address=tcp://127.0.0.1:5001 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.endpoint tcp://127.0.0.1:5001 \ --agency.endpoint tcp://127.0.0.1:5001 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent1 & --database.directory agent1 &
arangod --server.endpoint tcp://0.0.0.0:5002 \ arangod --server.endpoint tcp://0.0.0.0:5002 \
--agency.my-address=tcp://127.0.0.1:5002 \ --agency.my-address=tcp://127.0.0.1:5002 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.endpoint tcp://127.0.0.1:5001 \ --agency.endpoint tcp://127.0.0.1:5001 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent2 & --database.directory agent2 &
arangod --server.endpoint tcp://0.0.0.0:5003 \ arangod --server.endpoint tcp://0.0.0.0:5003 \
--agency.my-address=tcp://127.0.0.1:5003 \ --agency.my-address=tcp://127.0.0.1:5003 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.endpoint tcp://127.0.0.1:5001 \ --agency.endpoint tcp://127.0.0.1:5001 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent3 & --database.directory agent3 &
``` ```
### Local Test DBServers and Coordinators ### Local Test DBServers and Coordinators
@ -82,46 +82,46 @@ The following is a full example of what it might look like.
``` ```
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:6001 \ --server.endpoint tcp://0.0.0.0:6001 \
--cluster.my-address tcp://127.0.0.1:6001 \ --cluster.my-address tcp://127.0.0.1:6001 \
--cluster.my-role DBSERVER \ --cluster.my-role DBSERVER \
--cluster.agency-endpoint tcp://127.0.0.1:5001 \ --cluster.agency-endpoint tcp://127.0.0.1:5001 \
--cluster.agency-endpoint tcp://127.0.0.1:5002 \ --cluster.agency-endpoint tcp://127.0.0.1:5002 \
--cluster.agency-endpoint tcp://127.0.0.1:5003 \ --cluster.agency-endpoint tcp://127.0.0.1:5003 \
--database.directory dbserver1 & --database.directory dbserver1 &
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:6002 \ --server.endpoint tcp://0.0.0.0:6002 \
--cluster.my-address tcp://127.0.0.1:6002 \ --cluster.my-address tcp://127.0.0.1:6002 \
--cluster.my-role DBSERVER \ --cluster.my-role DBSERVER \
--cluster.agency-endpoint tcp://127.0.0.1:5001 \ --cluster.agency-endpoint tcp://127.0.0.1:5001 \
--cluster.agency-endpoint tcp://127.0.0.1:5002 \ --cluster.agency-endpoint tcp://127.0.0.1:5002 \
--cluster.agency-endpoint tcp://127.0.0.1:5003 \ --cluster.agency-endpoint tcp://127.0.0.1:5003 \
--database.directory dbserver2 & --database.directory dbserver2 &
``` ```
**Coordinators:** **Coordinators:**
``` ```
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:7001 \ --server.endpoint tcp://0.0.0.0:7001 \
--cluster.my-address tcp://127.0.0.1:7001 \ --cluster.my-address tcp://127.0.0.1:7001 \
--cluster.my-role COORDINATOR \ --cluster.my-role COORDINATOR \
--cluster.agency-endpoint tcp://127.0.0.1:5001 \ --cluster.agency-endpoint tcp://127.0.0.1:5001 \
--cluster.agency-endpoint tcp://127.0.0.1:5002 \ --cluster.agency-endpoint tcp://127.0.0.1:5002 \
--cluster.agency-endpoint tcp://127.0.0.1:5003 \ --cluster.agency-endpoint tcp://127.0.0.1:5003 \
--database.directory coordinator1 & --database.directory coordinator1 &
``` ```
``` ```
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:7002 \ --server.endpoint tcp://0.0.0.0:7002 \
--cluster.my-address tcp://127.0.0.1:7002 \ --cluster.my-address tcp://127.0.0.1:7002 \
--cluster.my-role COORDINATOR \ --cluster.my-role COORDINATOR \
--cluster.agency-endpoint tcp://127.0.0.1:5001 \ --cluster.agency-endpoint tcp://127.0.0.1:5001 \
--cluster.agency-endpoint tcp://127.0.0.1:5002 \ --cluster.agency-endpoint tcp://127.0.0.1:5002 \
--cluster.agency-endpoint tcp://127.0.0.1:5003 \ --cluster.agency-endpoint tcp://127.0.0.1:5003 \
--database.directory coordinator2 & --database.directory coordinator2 &
``` ```
Note in particular that the endpoint descriptions given under `--cluster.my-address` Note in particular that the endpoint descriptions given under `--cluster.my-address`
@ -172,39 +172,39 @@ On 192.168.1.1:
``` ```
arangod --server.endpoint tcp://0.0.0.0:8531 \ arangod --server.endpoint tcp://0.0.0.0:8531 \
--agency.my-address tcp://192.168.1.1:8531 \ --agency.my-address tcp://192.168.1.1:8531 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent --database.directory agent
``` ```
On 192.168.1.2: On 192.168.1.2:
``` ```
arangod --server.endpoint tcp://0.0.0.0:8531 \ arangod --server.endpoint tcp://0.0.0.0:8531 \
--agency.my-address tcp://192.168.1.2:8531 \ --agency.my-address tcp://192.168.1.2:8531 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent --database.directory agent
``` ```
On 192.168.1.3: On 192.168.1.3:
``` ```
arangod --server.endpoint tcp://0.0.0.0:8531 \ arangod --server.endpoint tcp://0.0.0.0:8531 \
--agency.my-address tcp://192.168.1.3:8531 \ --agency.my-address tcp://192.168.1.3:8531 \
--server.authentication false \ --server.authentication false \
--agency.activate true \ --agency.activate true \
--agency.size 3 \ --agency.size 3 \
--agency.endpoint tcp://192.168.1.1:8531 \ --agency.endpoint tcp://192.168.1.1:8531 \
--agency.endpoint tcp://192.168.1.2:8531 \ --agency.endpoint tcp://192.168.1.2:8531 \
--agency.endpoint tcp://192.168.1.3:8531 \ --agency.endpoint tcp://192.168.1.3:8531 \
--agency.supervision true \ --agency.supervision true \
--database.directory agent --database.directory agent
``` ```
### DBServers ### DBServers
@ -217,39 +217,39 @@ On 192.168.1.1:
``` ```
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:8530 \ --server.endpoint tcp://0.0.0.0:8530 \
--cluster.my-address tcp://192.168.1.1:8530 \ --cluster.my-address tcp://192.168.1.1:8530 \
--cluster.my-role DBSERVER \ --cluster.my-role DBSERVER \
--cluster.agency-endpoint tcp://192.168.1.1:8531 \ --cluster.agency-endpoint tcp://192.168.1.1:8531 \
--cluster.agency-endpoint tcp://192.168.1.2:8531 \ --cluster.agency-endpoint tcp://192.168.1.2:8531 \
--cluster.agency-endpoint tcp://192.168.1.3:8531 \ --cluster.agency-endpoint tcp://192.168.1.3:8531 \
--database.directory dbserver & --database.directory dbserver &
``` ```
On 192.168.1.2: On 192.168.1.2:
``` ```
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:8530 \ --server.endpoint tcp://0.0.0.0:8530 \
--cluster.my-address tcp://192.168.1.2:8530 \ --cluster.my-address tcp://192.168.1.2:8530 \
--cluster.my-role DBSERVER \ --cluster.my-role DBSERVER \
--cluster.agency-endpoint tcp://192.168.1.1:8531 \ --cluster.agency-endpoint tcp://192.168.1.1:8531 \
--cluster.agency-endpoint tcp://192.168.1.2:8531 \ --cluster.agency-endpoint tcp://192.168.1.2:8531 \
--cluster.agency-endpoint tcp://192.168.1.3:8531 \ --cluster.agency-endpoint tcp://192.168.1.3:8531 \
--database.directory dbserver & --database.directory dbserver &
``` ```
On 192.168.1.3: On 192.168.1.3:
``` ```
sudo arangod --server.authentication=false \ sudo arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:8530 \ --server.endpoint tcp://0.0.0.0:8530 \
--cluster.my-address tcp://192.168.1.3:8530 \ --cluster.my-address tcp://192.168.1.3:8530 \
--cluster.my-role DBSERVER \ --cluster.my-role DBSERVER \
--cluster.agency-endpoint tcp://192.168.1.1:8531 \ --cluster.agency-endpoint tcp://192.168.1.1:8531 \
--cluster.agency-endpoint tcp://192.168.1.2:8531 \ --cluster.agency-endpoint tcp://192.168.1.2:8531 \
--cluster.agency-endpoint tcp://192.168.1.3:8531 \ --cluster.agency-endpoint tcp://192.168.1.3:8531 \
--database.directory dbserver & --database.directory dbserver &
``` ```
### Coordinators ### Coordinators
@ -258,39 +258,39 @@ On 192.168.1.1:
``` ```
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:8529 \ --server.endpoint tcp://0.0.0.0:8529 \
--cluster.my-address tcp://192.168.1.1:8529 \ --cluster.my-address tcp://192.168.1.1:8529 \
--cluster.my-role COORDINATOR \ --cluster.my-role COORDINATOR \
--cluster.agency-endpoint tcp://192.168.1.1:8531 \ --cluster.agency-endpoint tcp://192.168.1.1:8531 \
--cluster.agency-endpoint tcp://192.168.1.2:8531 \ --cluster.agency-endpoint tcp://192.168.1.2:8531 \
--cluster.agency-endpoint tcp://192.168.1.3:8531 \ --cluster.agency-endpoint tcp://192.168.1.3:8531 \
--database.directory coordinator & --database.directory coordinator &
``` ```
On 192.168.1.2: On 192.168.1.2:
``` ```
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:8529 \ --server.endpoint tcp://0.0.0.0:8529 \
--cluster.my-address tcp://192.168.1.2:8529 \ --cluster.my-address tcp://192.168.1.2:8529 \
--cluster.my-role COORDINATOR \ --cluster.my-role COORDINATOR \
--cluster.agency-endpoint tcp://192.168.1.1:8531 \ --cluster.agency-endpoint tcp://192.168.1.1:8531 \
--cluster.agency-endpoint tcp://192.168.1.2:8531 \ --cluster.agency-endpoint tcp://192.168.1.2:8531 \
--cluster.agency-endpoint tcp://192.168.1.3:8531 \ --cluster.agency-endpoint tcp://192.168.1.3:8531 \
--database.directory coordinator & --database.directory coordinator &
``` ```
On 192.168.1.3: On 192.168.1.3:
``` ```
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:8529 \ --server.endpoint tcp://0.0.0.0:8529 \
--cluster.my-address tcp://192.168.1.3:8529 \ --cluster.my-address tcp://192.168.1.3:8529 \
--cluster.my-role COORDINATOR \ --cluster.my-role COORDINATOR \
--cluster.agency-endpoint tcp://192.168.1.1:8531 \ --cluster.agency-endpoint tcp://192.168.1.1:8531 \
--cluster.agency-endpoint tcp://192.168.1.2:8531 \ --cluster.agency-endpoint tcp://192.168.1.2:8531 \
--cluster.agency-endpoint tcp://192.168.1.3:8531 \ --cluster.agency-endpoint tcp://192.168.1.3:8531 \
--database.directory coordinator & --database.directory coordinator &
``` ```
**Note:** in the above commands, you can use host names, if they can be resolved, **Note:** in the above commands, you can use host names, if they can be resolved,
@ -304,22 +304,22 @@ On 192.168.1.4:
``` ```
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:8530 \ --server.endpoint tcp://0.0.0.0:8530 \
--cluster.my-address tcp://192.168.4.1:8530 \ --cluster.my-address tcp://192.168.4.1:8530 \
--cluster.my-role DBSERVER \ --cluster.my-role DBSERVER \
--cluster.agency-endpoint tcp://192.168.1.1:8531 \ --cluster.agency-endpoint tcp://192.168.1.1:8531 \
--cluster.agency-endpoint tcp://192.168.1.2:8531 \ --cluster.agency-endpoint tcp://192.168.1.2:8531 \
--cluster.agency-endpoint tcp://192.168.1.3:8531 \ --cluster.agency-endpoint tcp://192.168.1.3:8531 \
--database.directory dbserver & --database.directory dbserver &
arangod --server.authentication=false \ arangod --server.authentication=false \
--server.endpoint tcp://0.0.0.0:8529 \ --server.endpoint tcp://0.0.0.0:8529 \
--cluster.my-address tcp://192.168.1.4:8529 \ --cluster.my-address tcp://192.168.1.4:8529 \
--cluster.my-role COORDINATOR \ --cluster.my-role COORDINATOR \
--cluster.agency-endpoint tcp://192.168.1.1:8531 \ --cluster.agency-endpoint tcp://192.168.1.1:8531 \
--cluster.agency-endpoint tcp://192.168.1.2:8531 \ --cluster.agency-endpoint tcp://192.168.1.2:8531 \
--cluster.agency-endpoint tcp://192.168.1.3:8531 \ --cluster.agency-endpoint tcp://192.168.1.3:8531 \
--database.directory coordinator & --database.directory coordinator &
``` ```
Manual Start in Docker Manual Start in Docker
@ -344,12 +344,12 @@ An example configuration might look like this:
``` ```
docker run -e ARANGO_NO_AUTH=1 -p 192.168.1.1:10000:8530 arangodb/arangodb arangod \ docker run -e ARANGO_NO_AUTH=1 -p 192.168.1.1:10000:8530 arangodb/arangodb arangod \
--server.endpoint tcp://0.0.0.0:8530 \ --server.endpoint tcp://0.0.0.0:8530 \
--cluster.my-address tcp://192.168.1.1:10000 \ --cluster.my-address tcp://192.168.1.1:10000 \
--cluster.my-role DBSERVER \ --cluster.my-role DBSERVER \
--cluster.agency-endpoint tcp://192.168.1.1:9001 \ --cluster.agency-endpoint tcp://192.168.1.1:9001 \
--cluster.agency-endpoint tcp://192.168.1.2:9001 \ --cluster.agency-endpoint tcp://192.168.1.2:9001 \
--cluster.agency-endpoint tcp://192.168.1.3:9001 --cluster.agency-endpoint tcp://192.168.1.3:9001
``` ```
This will start a _DBServer_ within a Docker container with an isolated network. This will start a _DBServer_ within a Docker container with an isolated network.

View File

@ -11,7 +11,7 @@ We will assume that your IP is 127.0.0.1 and that the port 8529 is free:
``` ```
arangod --server.endpoint tcp://0.0.0.0:8529 \ arangod --server.endpoint tcp://0.0.0.0:8529 \
--database.directory standalone & --database.directory standalone &
``` ```
Manual Start in Docker Manual Start in Docker
@ -36,7 +36,7 @@ An example configuration might look like this:
``` ```
docker run -e ARANGO_NO_AUTH=1 -p 192.168.1.1:10000:8529 arangodb/arangodb arangod \ docker run -e ARANGO_NO_AUTH=1 -p 192.168.1.1:10000:8529 arangodb/arangodb arangod \
--server.endpoint tcp://0.0.0.0:8529\ --server.endpoint tcp://0.0.0.0:8529\
``` ```
This will start a single server within a Docker container with an isolated network. This will start a single server within a Docker container with an isolated network.

View File

@ -118,7 +118,8 @@ details, including the index-identifier, is returned.
@endDocuBlock ensureHashIndexArray @endDocuBlock ensureHashIndexArray
<!-- Creating Hash Index in Background {###
Creating Hash Index in Background
--------------------------------- ---------------------------------
{% hint 'info' %} {% hint 'info' %}
@ -127,7 +128,7 @@ This section only applies to the *rocksdb* storage engine
Creating new indexes is by default done under an exclusive collection lock. This means Creating new indexes is by default done under an exclusive collection lock. This means
that the collection (or the respective shards) are not available as long as the index that the collection (or the respective shards) are not available as long as the index
is created. This "foreground" index creation can be undesireable, if you have to perform it is created. This "foreground" index creation can be undesirable, if you have to perform it
on a live system without a dedicated maintenance window. on a live system without a dedicated maintenance window.
Indexes can also be created in "background", not using an exclusive lock during the creation. Indexes can also be created in "background", not using an exclusive lock during the creation.
@ -138,9 +139,10 @@ To create an hash index in the background in *arangosh* just specify `inBackgrou
```js ```js
db.collection.ensureIndex({ type: "hash", fields: [ "value" ], inBackground: true }); db.collection.ensureIndex({ type: "hash", fields: [ "value" ], inBackground: true });
``` --> ```
For more information see [Creating Indexes in Background](IndexBasics.md#creating-indexes-in-background) For more information see "Creating Indexes in Background" in the [Index basics](IndexBasics.md) page.
###}
Ensure uniqueness of relations in edge collections Ensure uniqueness of relations in edge collections
-------------------------------------------------- --------------------------------------------------

View File

@ -22,13 +22,15 @@ are covered by an edge collection's edge index automatically.
Using the system attribute `_id` in user-defined indexes is not possible, but Using the system attribute `_id` in user-defined indexes is not possible, but
indexing `_key`, `_rev`, `_from`, and `_to` is. indexing `_key`, `_rev`, `_from`, and `_to` is.
<!-- Creating new indexes is usually done under an exclusive collection lock. The collection is not {###
available as long as the index is created. This "foreground" index creation can be undesireable, Creating new indexes is usually done under an exclusive collection lock. The collection is not
available as long as the index is created. This "foreground" index creation can be undesirable,
if you have to perform it on a live system without a dedicated maintenance window. if you have to perform it on a live system without a dedicated maintenance window.
For potentially long running index creation operations the _rocksdb_ storage-engine also supports For potentially long running index creation operations the _rocksdb_ storage-engine also supports
creating indexes in "background". The colletion remains available durint the index creation, creating indexes in "background". The collection remains available during the index creation,
see the section [Creating Indexes in Background](#creating-indexes-in-background) for more information. --> see the section "Creating Indexes in Background" for more information.
###}
ArangoDB provides the following index types: ArangoDB provides the following index types:
@ -549,7 +551,7 @@ based on the costs it estimates, even if a vertex centric index might
in fact be faster. Vertex centric indexes are more likely to be chosen in fact be faster. Vertex centric indexes are more likely to be chosen
for highly connected graphs and with RocksDB storage engine. for highly connected graphs and with RocksDB storage engine.
<!-- {###
Creating Indexes in Background Creating Indexes in Background
------------------------------ ------------------------------
@ -559,10 +561,10 @@ This section only applies to the *rocksdb* storage engine
Creating new indexes is by default done under an exclusive collection lock. This means Creating new indexes is by default done under an exclusive collection lock. This means
that the collection (or the respective shards) are not available as long as the index that the collection (or the respective shards) are not available as long as the index
is created. This "foreground" index creation can be undesireable, if you have to perform it is created. This "foreground" index creation can be undesirable, if you have to perform it
on a live system without a dedicated maintenance window. on a live system without a dedicated maintenance window.
Indexes can also be created in "background", not using an exclusive lock during the creation. **STARTING FROM VERSION vX.Y.Z**, indexes can also be created in "background", not using an exclusive lock during the creation.
The collection remains available, other CRUD operations can run on the collection while the index is created. The collection remains available, other CRUD operations can run on the collection while the index is created.
This can be achieved by using the *inBackground* option. This can be achieved by using the *inBackground* option.
@ -584,7 +586,7 @@ db.collection.ensureIndex({ type: "geo", fields: [ "latitude", "longitude"], inB
db.collection.ensureIndex({ type: "fulltext", fields: [ "text" ], minLength: 4, inBackground: true }) db.collection.ensureIndex({ type: "fulltext", fields: [ "text" ], minLength: 4, inBackground: true })
``` ```
### Behaviour ### Behavior
Indexes that are still in the build process will not be visible via the ArangoDB API. Nevertheless it is not Indexes that are still in the build process will not be visible via the ArangoDB API. Nevertheless it is not
possible to create the same index twice via the *ensureIndex* API. AQL Queries will not use these indexes either possible to create the same index twice via the *ensureIndex* API. AQL Queries will not use these indexes either
@ -601,10 +603,11 @@ In the ArangoDB cluster the index might then be automatically recreated on affec
### Performance ### Performance
The backround index creation might be slower than the "foreground" index creation and require more RAM. The background index creation might be slower than the "foreground" index creation and require more RAM.
Under a write heavy load (specifically many remove, update or replace) operations, Under a write heavy load (specifically many remove, update or replace) operations,
the background index creation needs to keep a list of removed documents in RAM. This might become unsuistainable the background index creation needs to keep a list of removed documents in RAM. This might become unsustainable
if this list grows to tens of millions of entries. if this list grows to tens of millions of entries.
Building an index is always a write heavy operation (internally), it is alsways a good idea to build indexes Building an index is always a write heavy operation (internally), it is always a good idea to build indexes
during times with less load. --> during times with less load.
###}

View File

@ -187,8 +187,9 @@ and
will match. will match.
<!-- Creating Skiplist Index in Background {###
--------------------------------- Creating Skiplist Index in Background
-------------------------------------
{% hint 'info' %} {% hint 'info' %}
This section only applies to the *rocksdb* storage engine This section only applies to the *rocksdb* storage engine
@ -196,7 +197,7 @@ This section only applies to the *rocksdb* storage engine
Creating new indexes is by default done under an exclusive collection lock. This means Creating new indexes is by default done under an exclusive collection lock. This means
that the collection (or the respective shards) are not available as long as the index that the collection (or the respective shards) are not available as long as the index
is created. This "foreground" index creation can be undesireable, if you have to perform it is created. This "foreground" index creation can be undesirable, if you have to perform it
on a live system without a dedicated maintenance window. on a live system without a dedicated maintenance window.
Indexes can also be created in "background", not using an exclusive lock during the creation. Indexes can also be created in "background", not using an exclusive lock during the creation.
@ -209,4 +210,5 @@ To create a Skiplist index in the background in *arangosh* just specify `inBackg
db.collection.ensureIndex({ type: "skiplist", fields: [ "value" ], inBackground: true }); db.collection.ensureIndex({ type: "skiplist", fields: [ "value" ], inBackground: true });
``` ```
For more information see [Creating Indexes in Background](IndexBasics.md#creating-indexes-in-background) --> For more information see "Creating Indexes in Background" in the [Index basics](IndexBasics.md#) page.
###}

View File

@ -760,15 +760,15 @@ For example, the query
FOR doc1 IN collection1 FOR doc1 IN collection1
FOR doc2 IN collection2 FOR doc2 IN collection2
COLLECT x = doc1.x INTO g COLLECT x = doc1.x INTO g
RETURN { x, all: g[*].doc1.y } RETURN { x, all: g[*].doc1.y }
will automatically be turned into will automatically be turned into
FOR doc1 IN collection1 FOR doc1 IN collection1
FOR doc2 IN collection2 FOR doc2 IN collection2
COLLECT x = doc1.x INTO g KEEP doc1 COLLECT x = doc1.x INTO g KEEP doc1
RETURN { x, all: g[*].doc1.y } RETURN { x, all: g[*].doc1.y }
This prevents variable `doc2` from being temporarily stored in the variable `g`, This prevents variable `doc2` from being temporarily stored in the variable `g`,
which saves processing time and memory, especially for big result sets. which saves processing time and memory, especially for big result sets.

View File

@ -74,7 +74,7 @@ a descriptive error message
var body = { var body = {
name: "myfunctions::temperature::celsiustofahrenheit", name: "myfunctions::temperature::celsiustofahrenheit",
code : "function (celsius) { return celsius * 1.8 + 32; }", code : "function (celsius) { return celsius * 1.8 + 32; }",
isDeterministic: true isDeterministic: true
}; };
var response = logCurlRequest('POST', url, body); var response = logCurlRequest('POST', url, body);