1
0
Fork 0
arangodb/Documentation/UserManual/Sharding.md

9.0 KiB

Sharding

@NAVIGATE_Sharding @EMBEDTOC{ShardingTOC}

Introduction

Sharding allows to use multiple machines to run a cluster of ArangoDB instances that together appear to be a single database. This enables you to store much more data, since ArangoDB distributes the data automatically to the different servers.

In a cluster there are essentially two types of processes: "DBservers" and "coordinators". The former actually store the data, the latter expose the database to the outside world. The clients talk to the coordinators exactly as they would talk to a single ArangoDB instance via the REST interface. The coordinators know about the configuration of the cluster and automatically forward the incoming requests to the right DBservers.

As a central highly available service to hold the cluster configuration and to synchronise reconfiguration and failover operations we currently use a an external program called etcd. It provides a hierarchical key value store with strong consistency and reliability promises.

All this is admittedly a relatively complicated setup and involves a lot of steps for startup and shutdown of clusters. Therefore we have created convenience functionality to plan, setup, start and shutdown clusters.

The whole process works in two phases, first the "planning" phase and then the "running" phase. In the planning phase it is decided which processes with which roles run on which machine, which ports they use, where the central agency resides and what ports its agents use. The result of the planning phase is a "cluster plan", which is just a relatively big datastructure in JSON format. You can then use this cluster plan to startup, shutdown, check and cleanup your cluster.

This latter phase uses so-called "dispatchers". A dispatcher is yet another ArangoDB instance and you have to install exactly one such instance on every machine that will take part in your cluster. No special configuration whatsoever is needed and you can organise authentication exactly as you would in a normal ArangoDB instance.

However, you can use any of these dispatchers to plan and start your cluster. In the planning phase, you have tell the planner about all dispatchers in your cluster and it will automatically distribute your agency, DBserver and coordinator processes amonst the dispatchers. The result is the cluster plan which you feed into the kickstarter. The kickstarter is a program that actually uses the dispatchers to manipulate the processes in your cluster. It runs on one of the dispatchers, which analyses the cluster plan, and executes all actions, for which it is personally responsible, and forwards all other actions to the corresponding dispatcher. This is possible, because the cluster plan incorporates the information about all dispatchers.

How to try it out

In this text we assume that you are working with a standard installation of ArangoDB with version at least 2.0. This means that everything is compiled for cluster operation and that etcd is compiled and the executable is installed in the bin directory of your ArangoDB installation.

We will first plan and launch a cluster, such that all your servers run on the local machine.

Start up a regular ArangoDB, either in console mode or connect to it with the Arango shell arangosh. Then you can ask it to plan a cluster for you:

arangodb> var Planner = require("org/arangodb/cluster").Planner;
arangodb> p = new Planner({numberOfDBservers:3, numberOfCoordinators:2});
[object Object]

If you are curious you can look at the plan of your cluster:

arangodb> p.getPlan();

will show you a huge JSON document. More interestingly, some further components tell you more about the layout of your cluster:

arangodb> p.DBservers;
[ 
  { 
    "id" : "Pavel", 
    "dispatcher" : "me", 
    "port" : 8629 
  }, 
  { 
    "id" : "Perry", 
    "dispatcher" : "me", 
    "port" : 8630 
  }, 
  { 
    "id" : "Pancho", 
    "dispatcher" : "me", 
    "port" : 8631 
  } 
]
arangodb> p.coordinators;
[ 
  { 
    "id" : "Claus", 
    "dispatcher" : "me", 
    "port" : 8530 
  }, 
  { 
    "id" : "Chantalle", 
    "dispatcher" : "me", 
    "port" : 8531 
  } 
]

This for example tells you the ports on which your ArangoDB processes will run. We will need the 8530 (or whatever appears on your machine) for the coordinators below.

More interesting is that such a cluster plan document can be used to start up the cluster conveniently using a Kickstarter object. Please note that the launch method of the kickstarter shown below initialises all data directories and log files, so if you have previously used the same cluster plan you will lose all your data. Use the relaunch method described below instead in that case.

arangodb> var Kickstarter = require("org/arangodb/cluster").Kickstarter;
arangodb> k = new Kickstarter(p.getPlan());
arangodb> k.launch();

is all you have to do to fire up your first cluster. You will see some output, which you can safely ignore (as long as no error happens).

From then on you can contact one of the coordinators and use the cluster as if it were a single ArangoDB instance (use the port number from above instead of 8530, if you get a different one) (probably from another shell window):

$ arangosh --server.endpoint tcp://localhost:8530
[... some output omitted]
arangosh [_system]> db._listDatabases();
[ 
  "_system" 
]

This for example lists the cluster wide databases. Now let us create a sharded collection, we only have to specify the number of shards to use in addition to the usual command. The shards are automatically distributed amongst your DBservers:

arangosh [_system]> example = db._create("example",{numberOfShards:6});
[ArangoCollection 1000001, "example" (type document, status loaded)]
arangosh [_system]> x = example.save({"name":"Hans", "age":44});
{ 
  "error" : false, 
  "_id" : "example/1000008", 
  "_rev" : "13460426", 
  "_key" : "1000008" 
}
arangosh [_system]> example.document(x._key);
{ 
  "age" : 44, 
  "name" : "Hans", 
  "_id" : "example/1000008", 
  "_rev" : "13460426", 
  "_key" : "1000008" 
}

You can shut down your cluster by using the following Kickstarter method (in the original ArangoDB console):

arangodb> k.shutdown();

If you want to start your cluster again without losing the data you have previosly stored in it, you can use the relaunch method in exactly the same way as you previously used the launch method:

arangodb> k.relaunch();

You can now check whether or not all your cluster processes are still running by issueing:

arangodb> k.isHealthy();

which will show you the status of all the processes in the cluster. You should see "RUNNING" there in all relevant places.

Finally, to clean up the whole cluster (losing all the data stored in it), do:

arangodb> k.shutdown();
arangodb> k.cleanup();

We conclude this section with another example using two machines, that is two different dispatchers. We start from scratch using two machines, running on the network addresses tcp://192.168.173.78:8529 and tcp://192.168.173.13:6789, say. Both need to have a regular ArangoDB instance installed and running, please make sure that both bind to all network devices, such that they can talk to each other.

arangodb> var Planner = require("org/arangodb/cluster").Planner;
arangodb> var p = new Planner({
     dispatchers: {"me":{"endpoint":"tcp://192.168.173.78:8529"},
                   "theother":{"endpoint":"tcp://192.168.173.13:6789"}}, 
     "numberOfCoordinators":2, "numberOfDBservers": 2});

With this you create a cluster plan involving two machines. The planner will put one DBserver and one coordinator on each machine. You can now launch this cluster exactly as above:

arangodb> var Kickstarter = require("org/arangodb/cluster").Kickstarter;
arangodb> k = new Kickstarter(p.getPlan());
arangodb> k.launch();

Likewise, the methods shutdown, relaunch, isHealthy and cleanup work exactly as in the single server case.

See @ref JSModuleCluster "the corresponding chapter of the reference manual" for detailed information about the Planner and Kickstarter classes.

Status of the implementation

At this stage all the basic CRUD operation for single documents or edges and all simple queries are implemented. AQL queries work but not with the best possible performance. Creating collections, dropping collections, creating databases, dropping databases works, however these operations are relatively slow at this stage. This is mostly due to the etcd program we are currently using. We will probably see a speedup in a later release. The basic CRUD operations already have a relatively good performance, the simple queries are in need of further optimisations as well.

@NAVIGATE_Sharding