mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-07 16:14:52 +00:00
cd14f9ebcb
* split up select.md * array-join.md basic refactoring * distinct.md basic refactoring * format.md basic refactoring * from.md basic refactoring * group-by.md basic refactoring * having.md basic refactoring * additional index.md refactoring * into-outfile.md basic refactoring * join.md basic refactoring * limit.md basic refactoring * limit-by.md basic refactoring * order-by.md basic refactoring * prewhere.md basic refactoring * adjust operators/index.md links * adjust sample.md links * adjust more links * adjust operatots links * fix some links * adjust aggregate function article titles * basic refactor of remaining select clauses * absolute paths in make_links.sh * run make_links.sh * remove old select.md locations * translate docs/es * translate docs/fr * translate docs/fa * remove old operators.md location * change operators.md links * adjust links in docs/es * adjust links in docs/es * minor texts adjustments * wip * update machine translations to use new links * fix changelog * es build fixes * get rid of some select.md links * temporary adjust ru links * temporary adjust more ru links * improve curly brace handling * adjust ru as well * fa build fix * ru link fixes * zh link fixes * temporary disable part of anchor checks
177 lines
7.8 KiB
Markdown
177 lines
7.8 KiB
Markdown
---
|
||
machine_translated: true
|
||
machine_translated_rev: 72537a2d527c63c07aa5d2361a8829f3895cf2bd
|
||
toc_priority: 59
|
||
toc_title: "\u062A\u0627\u062A\u0631-\u06A9\u067E\u06CC"
|
||
---
|
||
|
||
# تاتر-کپی {#clickhouse-copier}
|
||
|
||
کپی داده ها از جداول در یک خوشه به جداول در یکی دیگر از (یا همان) خوشه.
|
||
|
||
شما می توانید چند اجرا `clickhouse-copier` نمونه بر روی سرور های مختلف برای انجام همان کار. باغ وحش برای همگام سازی فرایندها استفاده می شود.
|
||
|
||
پس از شروع, `clickhouse-copier`:
|
||
|
||
- قابلیت اتصال به باغ وحش و دریافت:
|
||
|
||
- شغل کپی.
|
||
- دولت از مشاغل کپی.
|
||
|
||
- این کار را انجام می دهد.
|
||
|
||
هر فرایند در حال اجرا را انتخاب “closest” سفال از خوشه منبع و کپی داده ها را به خوشه مقصد, تغییر شکل داده ها در صورت لزوم.
|
||
|
||
`clickhouse-copier` تغییرات باغ وحش را دنبال می کند و در پرواز اعمال می شود.
|
||
|
||
برای کاهش ترافیک شبکه توصیه می کنیم در حال اجرا `clickhouse-copier` در همان سرور که داده های منبع واقع شده است.
|
||
|
||
## در حال اجرا تاتر-کپی {#running-clickhouse-copier}
|
||
|
||
ابزار باید به صورت دستی اجرا شود:
|
||
|
||
``` bash
|
||
$ clickhouse-copier copier --daemon --config zookeeper.xml --task-path /task/path --base-dir /path/to/dir
|
||
```
|
||
|
||
پارامترها:
|
||
|
||
- `daemon` — Starts `clickhouse-copier` در حالت شبح.
|
||
- `config` — The path to the `zookeeper.xml` فایل با پارامترهای اتصال به باغ وحش.
|
||
- `task-path` — The path to the ZooKeeper node. This node is used for syncing `clickhouse-copier` پردازش و ذخیره سازی وظایف. وظایف در ذخیره می شود `$task-path/description`.
|
||
- `task-file` — Optional path to file with task configuration for initial upload to ZooKeeper.
|
||
- `task-upload-force` — Force upload `task-file` حتی اگر گره در حال حاضر وجود دارد.
|
||
- `base-dir` — The path to logs and auxiliary files. When it starts, `clickhouse-copier` ایجاد `clickhouse-copier_YYYYMMHHSS_<PID>` زیرشاخه در `$base-dir`. اگر این پارامتر حذف شده است, دایرکتوری ها در دایرکتوری که ایجاد `clickhouse-copier` راه اندازی شد.
|
||
|
||
## قالب باغ وحش.شمع {#format-of-zookeeper-xml}
|
||
|
||
``` xml
|
||
<yandex>
|
||
<logger>
|
||
<level>trace</level>
|
||
<size>100M</size>
|
||
<count>3</count>
|
||
</logger>
|
||
|
||
<zookeeper>
|
||
<node index="1">
|
||
<host>127.0.0.1</host>
|
||
<port>2181</port>
|
||
</node>
|
||
</zookeeper>
|
||
</yandex>
|
||
```
|
||
|
||
## پیکربندی وظایف کپی کردن {#configuration-of-copying-tasks}
|
||
|
||
``` xml
|
||
<yandex>
|
||
<!-- Configuration of clusters as in an ordinary server config -->
|
||
<remote_servers>
|
||
<source_cluster>
|
||
<shard>
|
||
<internal_replication>false</internal_replication>
|
||
<replica>
|
||
<host>127.0.0.1</host>
|
||
<port>9000</port>
|
||
</replica>
|
||
</shard>
|
||
...
|
||
</source_cluster>
|
||
|
||
<destination_cluster>
|
||
...
|
||
</destination_cluster>
|
||
</remote_servers>
|
||
|
||
<!-- How many simultaneously active workers are possible. If you run more workers superfluous workers will sleep. -->
|
||
<max_workers>2</max_workers>
|
||
|
||
<!-- Setting used to fetch (pull) data from source cluster tables -->
|
||
<settings_pull>
|
||
<readonly>1</readonly>
|
||
</settings_pull>
|
||
|
||
<!-- Setting used to insert (push) data to destination cluster tables -->
|
||
<settings_push>
|
||
<readonly>0</readonly>
|
||
</settings_push>
|
||
|
||
<!-- Common setting for fetch (pull) and insert (push) operations. Also, copier process context uses it.
|
||
They are overlaid by <settings_pull/> and <settings_push/> respectively. -->
|
||
<settings>
|
||
<connect_timeout>3</connect_timeout>
|
||
<!-- Sync insert is set forcibly, leave it here just in case. -->
|
||
<insert_distributed_sync>1</insert_distributed_sync>
|
||
</settings>
|
||
|
||
<!-- Copying tasks description.
|
||
You could specify several table task in the same task description (in the same ZooKeeper node), they will be performed
|
||
sequentially.
|
||
-->
|
||
<tables>
|
||
<!-- A table task, copies one table. -->
|
||
<table_hits>
|
||
<!-- Source cluster name (from <remote_servers/> section) and tables in it that should be copied -->
|
||
<cluster_pull>source_cluster</cluster_pull>
|
||
<database_pull>test</database_pull>
|
||
<table_pull>hits</table_pull>
|
||
|
||
<!-- Destination cluster name and tables in which the data should be inserted -->
|
||
<cluster_push>destination_cluster</cluster_push>
|
||
<database_push>test</database_push>
|
||
<table_push>hits2</table_push>
|
||
|
||
<!-- Engine of destination tables.
|
||
If destination tables have not be created, workers create them using columns definition from source tables and engine
|
||
definition from here.
|
||
|
||
NOTE: If the first worker starts insert data and detects that destination partition is not empty then the partition will
|
||
be dropped and refilled, take it into account if you already have some data in destination tables. You could directly
|
||
specify partitions that should be copied in <enabled_partitions/>, they should be in quoted format like partition column of
|
||
system.parts table.
|
||
-->
|
||
<engine>
|
||
ENGINE=ReplicatedMergeTree('/clickhouse/tables/{cluster}/{shard}/hits2', '{replica}')
|
||
PARTITION BY toMonday(date)
|
||
ORDER BY (CounterID, EventDate)
|
||
</engine>
|
||
|
||
<!-- Sharding key used to insert data to destination cluster -->
|
||
<sharding_key>jumpConsistentHash(intHash64(UserID), 2)</sharding_key>
|
||
|
||
<!-- Optional expression that filter data while pull them from source servers -->
|
||
<where_condition>CounterID != 0</where_condition>
|
||
|
||
<!-- This section specifies partitions that should be copied, other partition will be ignored.
|
||
Partition names should have the same format as
|
||
partition column of system.parts table (i.e. a quoted text).
|
||
Since partition key of source and destination cluster could be different,
|
||
these partition names specify destination partitions.
|
||
|
||
NOTE: In spite of this section is optional (if it is not specified, all partitions will be copied),
|
||
it is strictly recommended to specify them explicitly.
|
||
If you already have some ready partitions on destination cluster they
|
||
will be removed at the start of the copying since they will be interpeted
|
||
as unfinished data from the previous copying!!!
|
||
-->
|
||
<enabled_partitions>
|
||
<partition>'2018-02-26'</partition>
|
||
<partition>'2018-03-05'</partition>
|
||
...
|
||
</enabled_partitions>
|
||
</table_hits>
|
||
|
||
<!-- Next table to copy. It is not copied until previous table is copying. -->
|
||
</table_visits>
|
||
...
|
||
</table_visits>
|
||
...
|
||
</tables>
|
||
</yandex>
|
||
```
|
||
|
||
`clickhouse-copier` پیگیری تغییرات در `/task/path/description` و اونا رو تو پرواز بکار میبره برای مثال, اگر شما ارزش تغییر `max_workers` تعداد فرایندهای در حال اجرا وظایف نیز تغییر خواهد کرد.
|
||
|
||
[مقاله اصلی](https://clickhouse.tech/docs/en/operations/utils/clickhouse-copier/) <!--hide-->
|