ClickHouse/tests/integration/test_cluster_copier/task_trivial.xml

64 lines
1.9 KiB
XML
Raw Normal View History

<clickhouse>
2020-03-16 21:05:38 +00:00
<!-- How many simualteneous workers are posssible -->
<max_workers>3</max_workers>
<!-- Common setting for pull and push operations -->
<settings>
<connect_timeout>1</connect_timeout>
</settings>
<!-- Setting used to fetch data -->
<settings_pull>
<max_rows_in_distinct>0</max_rows_in_distinct>
</settings_pull>
<!-- Setting used to insert data -->
<settings_push>
</settings_push>
<!-- Tasks -->
<tables>
<hits>
<cluster_pull>source_trivial_cluster</cluster_pull>
<database_pull>default</database_pull>
<table_pull>trivial</table_pull>
<cluster_push>destination_trivial_cluster</cluster_push>
<database_push>default</database_push>
<table_push>trivial</table_push>
<!-- Engine of destination tables -->
2020-04-21 17:37:40 +00:00
<engine>ENGINE=ReplicatedMergeTree('/clickhouse/tables/cluster{cluster}/{shard}/hits', '{replica}') PARTITION BY d % 5 ORDER BY (d, sipHash64(d)) SAMPLE BY sipHash64(d) SETTINGS index_granularity = 16</engine>
2020-03-16 21:05:38 +00:00
<!-- Which sarding key to use while copying -->
<sharding_key>d + 1</sharding_key>
<!-- Optional expression that filter copying data -->
<where_condition>d - d = 0</where_condition>
</hits>
</tables>
<!-- Configuration of clusters -->
<remote_servers>
<source_trivial_cluster>
<shard>
<replica>
2021-06-08 01:50:43 +00:00
<host>first_trivial</host>
2020-03-16 21:05:38 +00:00
<port>9000</port>
</replica>
</shard>
</source_trivial_cluster>
<destination_trivial_cluster>
<shard>
<replica>
2021-06-08 01:50:43 +00:00
<host>second_trivial</host>
2020-03-16 21:05:38 +00:00
<port>9000</port>
</replica>
</shard>
</destination_trivial_cluster>
</remote_servers>
</clickhouse>