# pylint: disable=line-too-long # pylint: disable=unused-argument # pylint: disable=redefined-outer-name: import time import pytest from helpers.cluster import ClickHouseCluster from helpers.network import PartitionManager from helpers.test_tools import assert_eq_with_retry cluster = ClickHouseCluster(__file__) def _fill_nodes(nodes, shard): for node in nodes: node.query( ''' CREATE DATABASE test; CREATE TABLE real_table(date Date, id UInt32, dummy UInt32) ENGINE = MergeTree(date, id, 8192); CREATE TABLE other_table(date Date, id UInt32, dummy UInt32) ENGINE = MergeTree(date, id, 8192); CREATE TABLE test_table(date Date, id UInt32, dummy UInt32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/test{shard}/replicated', '{replica}', date, id, 8192); '''.format(shard=shard, replica=node.name)) node1 = cluster.add_instance('node1', main_configs=['configs/remote_servers.xml'], with_zookeeper=True) node2 = cluster.add_instance('node2', main_configs=['configs/remote_servers.xml'], with_zookeeper=True) @pytest.fixture(scope="module") def normal_work(): try: cluster.start() _fill_nodes([node1, node2], 1) yield cluster finally: cluster.shutdown() def test_normal_work(normal_work): node1.query("insert into test_table values ('2017-06-16', 111, 0)") node1.query("insert into real_table values ('2017-06-16', 222, 0)") assert_eq_with_retry(node1, "SELECT id FROM test_table order by id", '111') assert_eq_with_retry(node1, "SELECT id FROM real_table order by id", '222') assert_eq_with_retry(node2, "SELECT id FROM test_table order by id", '111') node1.query("ALTER TABLE test_table REPLACE PARTITION 201706 FROM real_table") assert_eq_with_retry(node1, "SELECT id FROM test_table order by id", '222') assert_eq_with_retry(node2, "SELECT id FROM test_table order by id", '222') node3 = cluster.add_instance('node3', main_configs=['configs/remote_servers.xml'], with_zookeeper=True) node4 = cluster.add_instance('node4', main_configs=['configs/remote_servers.xml'], with_zookeeper=True) @pytest.fixture(scope="module") def drop_failover(): try: cluster.start() _fill_nodes([node3, node4], 2) yield cluster finally: cluster.shutdown() def test_drop_failover(drop_failover): node3.query("insert into test_table values ('2017-06-16', 111, 0)") node3.query("insert into real_table values ('2017-06-16', 222, 0)") assert_eq_with_retry(node3, "SELECT id FROM test_table order by id", '111') assert_eq_with_retry(node3, "SELECT id FROM real_table order by id", '222') assert_eq_with_retry(node4, "SELECT id FROM test_table order by id", '111') with PartitionManager() as pm: # Hinder replication between replicas pm.partition_instances(node3, node4, port=9009) # Disconnect Node4 from zookeper pm.drop_instance_zk_connections(node4) node3.query("ALTER TABLE test_table REPLACE PARTITION 201706 FROM real_table") # Node3 replace is ok assert_eq_with_retry(node3, "SELECT id FROM test_table order by id", '222') # Network interrupted -- replace is not ok, but it's ok assert_eq_with_retry(node4, "SELECT id FROM test_table order by id", '111') # Drop partition on source node node3.query("ALTER TABLE test_table DROP PARTITION 201706") # Wait few seconds for connection to zookeeper to be restored time.sleep(5) msg = node4.query_with_retry( "select last_exception from system.replication_queue where type = 'REPLACE_RANGE'", check_callback=lambda x: 'Not found part' not in x, sleep_time=1) assert 'Not found part' not in msg assert_eq_with_retry(node4, "SELECT id FROM test_table order by id", '') node5 = cluster.add_instance('node5', main_configs=['configs/remote_servers.xml'], with_zookeeper=True) node6 = cluster.add_instance('node6', main_configs=['configs/remote_servers.xml'], with_zookeeper=True) @pytest.fixture(scope="module") def replace_after_replace_failover(): try: cluster.start() _fill_nodes([node5, node6], 3) yield cluster finally: cluster.shutdown() def test_replace_after_replace_failover(replace_after_replace_failover): node5.query("insert into test_table values ('2017-06-16', 111, 0)") node5.query("insert into real_table values ('2017-06-16', 222, 0)") node5.query("insert into other_table values ('2017-06-16', 333, 0)") assert_eq_with_retry(node5, "SELECT id FROM test_table order by id", '111') assert_eq_with_retry(node5, "SELECT id FROM real_table order by id", '222') assert_eq_with_retry(node5, "SELECT id FROM other_table order by id", '333') assert_eq_with_retry(node6, "SELECT id FROM test_table order by id", '111') with PartitionManager() as pm: # Hinder replication between replicas pm.partition_instances(node5, node6, port=9009) # Disconnect Node6 from zookeper pm.drop_instance_zk_connections(node6) node5.query("ALTER TABLE test_table REPLACE PARTITION 201706 FROM real_table") # Node5 replace is ok assert_eq_with_retry(node5, "SELECT id FROM test_table order by id", '222') # Network interrupted -- replace is not ok, but it's ok assert_eq_with_retry(node6, "SELECT id FROM test_table order by id", '111') # Replace partition on source node node5.query("ALTER TABLE test_table REPLACE PARTITION 201706 FROM other_table") assert_eq_with_retry(node5, "SELECT id FROM test_table order by id", '333') # Wait few seconds for connection to zookeeper to be restored time.sleep(5) msg = node6.query_with_retry( "select last_exception from system.replication_queue where type = 'REPLACE_RANGE'", check_callback=lambda x: 'Not found part' not in x, sleep_time=1) assert 'Not found part' not in msg assert_eq_with_retry(node6, "SELECT id FROM test_table order by id", '333')