diff --git a/tests/integration/runner b/tests/integration/runner index c327898b675..34bfa75dc4b 100755 --- a/tests/integration/runner +++ b/tests/integration/runner @@ -413,11 +413,11 @@ if __name__ == "__main__": subprocess.check_call(f"docker kill {' '.join(containers)}", shell=True) print(f"Containers {containers} killed") - # Avoid overlaps with previous runs - subprocess.check_call("dmesg --clear", shell=True) + # # Avoid overlaps with previous runs + # subprocess.check_call("dmesg --clear", shell=True) print(("Running pytest container as: '" + cmd + "'.")) subprocess.check_call(cmd, shell=True) - # Dump dmesg (to capture possible OOMs) - subprocess.check_call("dmesg -T", shell=True) + # # Dump dmesg (to capture possible OOMs) + # subprocess.check_call("dmesg -T", shell=True) diff --git a/tests/integration/test_keeper_multinode_simple/test.py b/tests/integration/test_keeper_multinode_simple/test.py index 1dcbb290fa8..1999f361dd4 100644 --- a/tests/integration/test_keeper_multinode_simple/test.py +++ b/tests/integration/test_keeper_multinode_simple/test.py @@ -63,6 +63,10 @@ def test_read_write_multinode(started_cluster): node2_zk = get_fake_zk("node2") node3_zk = get_fake_zk("node3") + # Cleanup + if node1_zk.exists("/test_read_write_multinode_node1") != None: + node1_zk.delete("/test_read_write_multinode_node1") + node1_zk.create("/test_read_write_multinode_node1", b"somedata1") node2_zk.create("/test_read_write_multinode_node2", b"somedata2") node3_zk.create("/test_read_write_multinode_node3", b"somedata3") @@ -105,6 +109,10 @@ def test_watch_on_follower(started_cluster): node2_zk = get_fake_zk("node2") node3_zk = get_fake_zk("node3") + # Cleanup + if node1_zk.exists("/test_data_watches") != None: + node1_zk.delete("/test_data_watches") + node1_zk.create("/test_data_watches") node2_zk.set("/test_data_watches", b"hello") node3_zk.set("/test_data_watches", b"world") @@ -163,6 +171,10 @@ def test_session_expiration(started_cluster): node3_zk = get_fake_zk("node3", timeout=3.0) print("Node3 session id", node3_zk._session_id) + # Cleanup + if node3_zk.exists("/test_ephemeral_node") != None: + node3_zk.delete("/test_ephemeral_node") + node3_zk.create("/test_ephemeral_node", b"world", ephemeral=True) with PartitionManager() as pm: @@ -201,13 +213,18 @@ def test_follower_restart(started_cluster): try: wait_nodes() node1_zk = get_fake_zk("node1") - - node1_zk.create("/test_restart_node", b"hello") - - node3.restart_clickhouse(kill=True) - node3_zk = get_fake_zk("node3") + # Cleanup + if node1_zk.exists("/test_restart_node") != None: + node1_zk.delete("/test_restart_node") + + node1_zk.create("/test_restart_node", b"hello") + node3.restart_clickhouse(kill=True) + + wait_nodes() + + node3_zk = get_fake_zk("node3") # got data from log assert node3_zk.get("/test_restart_node")[0] == b"hello" @@ -225,11 +242,11 @@ def test_follower_restart(started_cluster): def test_simple_replicated_table(started_cluster): wait_nodes() + for i, node in enumerate([node1, node2, node3]): + node.query("DROP TABLE IF EXISTS t SYNC") node.query( - "CREATE TABLE t (value UInt64) ENGINE = ReplicatedMergeTree('/clickhouse/t', '{}') ORDER BY tuple()".format( - i + 1 - ) + f"CREATE TABLE t (value UInt64) ENGINE = ReplicatedMergeTree('/clickhouse/t', '{i + 1}') ORDER BY tuple()" ) node2.query("INSERT INTO t SELECT number FROM numbers(10)")