mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-25 09:02:00 +00:00
112 lines
11 KiB
ReStructuredText
112 lines
11 KiB
ReStructuredText
Distributed
|
|
-----------
|
|
|
|
**The Distributed engine does not store data itself**, but allows distributed query processing on multiple servers.
|
|
Reading is automatically parallelized. During a read, the table indexes on remote servers are used, if there are any.
|
|
The Distributed engine accepts parameters: the cluster name in the server's config file, the name of a remote database, the name of a remote table, and (optionally) a sharding key.
|
|
Example:
|
|
::
|
|
Distributed(logs, default, hits[, sharding_key])
|
|
|
|
- Data will be read from all servers in the 'logs' cluster, from the 'default.hits' table located on every server in the cluster.
|
|
Data is not only read, but is partially processed on the remote servers (to the extent that this is possible).
|
|
For example, for a query with GROUP BY, data will be aggregated on remote servers, and the intermediate states of aggregate functions will be sent to the requestor server. Then data will be further aggregated.
|
|
|
|
Instead of the database name, you can use a constant expression that returns a string. For example, ``currentDatabase()``.
|
|
|
|
logs - The cluster name in the server's config file.
|
|
|
|
Clusters are set like this:
|
|
|
|
.. code-block:: xml
|
|
|
|
<remote_servers>
|
|
<logs>
|
|
<shard>
|
|
<!-- Optional. Shard weight when writing data. By default, 1. -->
|
|
<weight>1</weight>
|
|
<!-- Optional. Whether to write data to just one of the replicas. By default, false - write data to all of the replicas. -->
|
|
<internal_replication>false</internal_replication>
|
|
<replica>
|
|
<host>example01-01-1</host>
|
|
<port>9000</port>
|
|
</replica>
|
|
<replica>
|
|
<host>example01-01-2</host>
|
|
<port>9000</port>
|
|
</replica>
|
|
</shard>
|
|
<shard>
|
|
<weight>2</weight>
|
|
<internal_replication>false</internal_replication>
|
|
<replica>
|
|
<host>example01-02-1</host>
|
|
<port>9000</port>
|
|
</replica>
|
|
<replica>
|
|
<host>example01-02-2</host>
|
|
<port>9000</port>
|
|
</replica>
|
|
</shard>
|
|
</logs>
|
|
</remote_servers>
|
|
|
|
Here a cluster is defined with the name 'logs' that consists of two shards, each of which contains two replicas. Shards refer to the servers that contain different parts of the data (in order to read all the data, you must access all the shards).
|
|
Replicas are duplicating servers (in order to read all the data, you can access the data on any one of the replicas).
|
|
|
|
For each server, there are several parameters: mandatory: ``'host'``, ``'port'``, and optional: ``'user'``, ``'password'``.
|
|
* ``host`` - address of remote server. May be specified as domain name or IPv4 or IPv6 address. If you specify domain, server will perform DNS lookup at startup, and result will be cached till server shutdown. If DNS request is failed, server won't start. If you are changing DNS records, restart the server for new records to take effect.
|
|
* ``port`` - TCP-port for interserver communication (tcp_port in configuration file, usually 9000). Don't get confused with http_port.
|
|
* ``user`` - user name to connect to remote server. By default user is 'default'. This user must have access rights to connect to remote server. Access rights are managed in users.xml configuration file. For additional info, consider "Access rights" section.
|
|
* ``password`` - password to log in to remote server, in plaintext. Default is empty string.
|
|
|
|
When specifying replicas, one of the available replicas will be selected for each of the shards when reading. You can configure the algorithm for load balancing (the preference for which replica to access) - see the 'load_balancing' setting.
|
|
If the connection with the server is not established, there will be an attempt to connect with a short timeout. If the connection failed, the next replica will be selected, and so on for all the replicas. If the connection attempt failed for all the replicas, the attempt will be repeated the same way, several times.
|
|
This works in favor of resiliency, but does not provide complete fault tolerance: a remote server might accept the connection, but might not work, or work poorly.
|
|
|
|
You can specify just one of the shards (in this case, query processing should be called remote, rather than distributed) or up to any number of shards. In each shard, you can specify from one to any number of replicas. You can specify a different number of replicas for each shard.
|
|
|
|
You can specify as many clusters as you wish in the configuration.
|
|
|
|
To view your clusters, use the 'system.clusters' table.
|
|
|
|
The Distributed engine allows working with a cluster like a local server. However, the cluster is inextensible: you must write its configuration in the server config file (even better, for all the cluster's servers).
|
|
|
|
There is no support for Distributed tables that look at other Distributed tables (except in cases when a Distributed table only has one shard). As an alternative, make the Distributed table look at the "final" tables.
|
|
|
|
The Distributed engine requires writing clusters to the config file. Clusters from config are updated on the fly, it does not require server restart. If you need to send a query to an unknown set of shards and replicas each time, you don't need to create a Distributed table - use the 'remote' table function instead. See the section "Table functions".
|
|
|
|
There are two methods for writing data to a cluster:
|
|
|
|
First, you can define which servers to write which data to, and perform the write directly on each shard. In other words, perform INSERT in the tables that the distributed table "looks at".
|
|
This is the most flexible solution - you can use any sharding scheme, which could be non-trivial due to the requirements of the subject area.
|
|
This is also the most optimal solution, since data can be written to different shards completely independently.
|
|
|
|
Second, you can perform INSERT in a Distributed table. In this case, the table will distribute the inserted data across servers itself.
|
|
In order to write to a Distributed table, it must have a sharding key set (the last parameter). In addition, if there is only one shard, the write operation works without specifying the sharding key, since it doesn't have any meaning in this case.
|
|
|
|
Each shard can have a weight defined in the config file. By default, the weight is equal to one. Data is distributed across shards in the amount proportional to the shard weight. For example, if there are two shards and the first has a weight of 9 while the second has a weight of 10, the first will be sent 9 / 19 parts of the rows, and the second will be sent 10 / 19.
|
|
|
|
Each shard can have the 'internal_replication' parameter defined in the config file.
|
|
|
|
If this parameter is set to 'true', the write operation selects the first healthy replica and writes data to it. Use this alternative if the Distributed table "looks at" replicated tables. In other words, if the table where data will be written is going to replicate them itself.
|
|
|
|
If it is set to 'false' (the default), data is written to all replicas. In essence, this means that the Distributed table replicates data itself. This is worse than using replicated tables, because the consistency of replicas is not checked, and over time they will contain slightly different data.
|
|
|
|
To select the shard that a row of data is sent to, the sharding expression is analyzed, and its remainder is taken from dividing it by the total weight of the shards. The row is sent to the shard that corresponds to the half-interval of the remainders from 'prev_weight' to 'prev_weights + weight', where 'prev_weights' is the total weight of the shards with the smallest number, and 'weight' is the weight of this shard. For example, if there are two shards, and the first has a weight of 9 while the second has a weight of 10, the row will be sent to the first shard for the remainders from the range [0, 9), and to the second for the remainders from the range [10, 19).
|
|
|
|
The sharding expression can be any expression from constants and table columns that returns an integer. For example, you can use the expression 'rand()' for random distribution of data, or 'UserID' for distribution by the remainder from dividing the user's ID (then the data of a single user will reside on a single shard, which simplifies running IN and JOIN by users). If one of the columns is not distributed evenly enough, you can wrap it in a hash function: intHash64(UserID).
|
|
|
|
A simple remainder from division is a limited solution for sharding and isn't always appropriate. It works for medium and large volumes of data (dozens of servers), but not for very large volumes of data (hundreds of servers or more). In the latter case, use the sharding scheme required by the subject area, rather than using entries in Distributed tables.
|
|
|
|
When using Replicated tables, it is possible to reshard data - look at "Resharding" section. But in many cases, better to do without it. SELECT queries are sent to all the shards, and work regardless of how data is distributed across the shards (they can be distributed completely randomly). When you add a new shard, you don't have to transfer the old data to it. You can write new data with a heavier weight - the data will be distributed slightly unevenly, but queries will work correctly and efficiently.
|
|
|
|
You should be concerned about the sharding scheme in the following cases:
|
|
- Queries are used that require joining data (IN or JOIN) by a specific key. If data is sharded by this key, you can use local IN or JOIN instead of GLOBAL IN or GLOBAL JOIN, which is much more efficient.
|
|
- A large number of servers is used (hundreds or more) with a large number of small queries (queries of individual clients - websites, advertisers, or partners). In order for the small queries to not affect the entire cluster, it makes sense to locate data for a single client on a single shard. Alternatively, as we've done in Yandex.Metrica, you can set up bi-level sharding: divide the entire cluster into "layers", where a layer may consist of multiple shards. Data for a single client is located on a single layer, but shards can be added to a layer as necessary, and data is randomly distributed within them. Distributed tables are created for each layer, and a single shared distributed table is created for global queries.
|
|
|
|
Data is written asynchronously. For an INSERT to a Distributed table, the data block is just written to the local file system. The data is sent to the remote servers in the background as soon as possible. You should check whether data is sent successfully by checking the list of files (data waiting to be sent) in the table directory:
|
|
/var/lib/clickhouse/data/database/table/.
|
|
|
|
If the server ceased to exist or had a rough restart (for example, after a device failure) after an INSERT to a Distributed table, the inserted data might be lost. If a damaged data part is detected in the table directory, it is transferred to the 'broken' subdirectory and no longer used.
|