ClickHouse/tests/integration/test_polymorphic_parts/test.py

491 lines
15 KiB
Python
Raw Normal View History

import os
2020-02-13 14:59:15 +00:00
import random
import string
2020-04-27 13:37:05 +00:00
import struct
2020-02-13 14:59:15 +00:00
import pytest
2024-02-26 12:45:20 +00:00
from helpers.cluster import ClickHouseCluster, CLICKHOUSE_CI_MIN_TESTED_VERSION
from helpers.test_tools import TSV
2021-05-15 12:33:01 +00:00
from helpers.test_tools import assert_eq_with_retry, exec_query_with_retry
2020-02-13 14:59:15 +00:00
cluster = ClickHouseCluster(__file__)
2020-02-13 14:59:15 +00:00
def get_random_array():
return [random.randint(0, 1000) % 1000 for _ in range(random.randint(0, 1000))]
2020-02-13 14:59:15 +00:00
def get_random_string():
length = random.randint(0, 1000)
return "".join(
random.choice(string.ascii_uppercase + string.digits) for _ in range(length)
)
2020-02-13 14:59:15 +00:00
2020-02-13 14:59:15 +00:00
def insert_random_data(table, node, size):
data = [
"("
+ ",".join(
(
"'2019-10-11'",
str(i),
"'" + get_random_string() + "'",
str(get_random_array()),
)
)
+ ")"
for i in range(size)
2020-02-13 14:59:15 +00:00
]
2020-08-07 17:16:59 +00:00
node.query("INSERT INTO {} VALUES {}".format(table, ",".join(data)))
2020-02-13 14:59:15 +00:00
def create_tables(name, nodes, node_settings, shard):
2020-02-13 14:59:15 +00:00
for i, (node, settings) in enumerate(zip(nodes, node_settings)):
node.query(
"""
CREATE TABLE {name}(date Date, id UInt32, s String, arr Array(Int32))
ENGINE = ReplicatedMergeTree('/clickhouse/tables/test/{shard}/{name}', '{repl}')
PARTITION BY toYYYYMM(date)
ORDER BY id
SETTINGS index_granularity = 64, index_granularity_bytes = {index_granularity_bytes},
2023-05-02 22:31:09 +00:00
min_rows_for_wide_part = {min_rows_for_wide_part},
min_bytes_for_wide_part = 0
""".format(
name=name, shard=shard, repl=i, **settings
)
)
2020-02-13 14:59:15 +00:00
2020-02-19 14:55:03 +00:00
def create_tables_old_format(name, nodes, shard):
for i, node in enumerate(nodes):
node.query(
"""
CREATE TABLE {name}(date Date, id UInt32, s String, arr Array(Int32))
ENGINE = ReplicatedMergeTree('/clickhouse/tables/test/{shard}/{name}', '{repl}', date, id, 64)
""".format(
name=name, shard=shard, repl=i
2022-06-24 17:10:33 +00:00
),
settings={"allow_deprecated_syntax_for_merge_tree": 1},
)
2020-02-13 14:59:15 +00:00
node1 = cluster.add_instance(
"node1",
main_configs=[],
user_configs=["configs/users.d/not_optimize_count.xml"],
with_zookeeper=True,
)
node2 = cluster.add_instance(
"node2",
main_configs=[],
user_configs=["configs/users.d/not_optimize_count.xml"],
with_zookeeper=True,
)
settings_default = {
"index_granularity_bytes": 10485760,
"min_rows_for_wide_part": 512,
}
settings_compact_only = {
"index_granularity_bytes": 10485760,
"min_rows_for_wide_part": 1000000,
}
settings_not_adaptive = {
"index_granularity_bytes": 0,
"min_rows_for_wide_part": 512,
}
node3 = cluster.add_instance(
"node3",
main_configs=[],
user_configs=["configs/users.d/not_optimize_count.xml"],
with_zookeeper=True,
)
node4 = cluster.add_instance(
"node4",
user_configs=["configs/users.d/not_optimize_count.xml"],
main_configs=["configs/no_leader.xml"],
with_zookeeper=True,
)
2020-02-13 14:59:15 +00:00
settings_compact = {
"index_granularity_bytes": 10485760,
"min_rows_for_wide_part": 512,
}
settings_wide = {
"index_granularity_bytes": 10485760,
"min_rows_for_wide_part": 0,
}
node5 = cluster.add_instance(
"node5", main_configs=["configs/compact_parts.xml"], with_zookeeper=True
)
node6 = cluster.add_instance(
"node6", main_configs=["configs/compact_parts.xml"], with_zookeeper=True
)
2020-02-13 14:59:15 +00:00
node9 = cluster.add_instance("node9", with_zookeeper=True, stay_alive=True)
node10 = cluster.add_instance("node10", with_zookeeper=True)
node11 = cluster.add_instance(
"node11",
main_configs=["configs/do_not_merge.xml"],
with_zookeeper=True,
stay_alive=True,
)
node12 = cluster.add_instance(
"node12",
main_configs=["configs/do_not_merge.xml"],
with_zookeeper=True,
stay_alive=True,
)
2020-04-29 17:14:49 +00:00
2020-02-13 14:59:15 +00:00
@pytest.fixture(scope="module")
def start_cluster():
try:
cluster.start()
create_tables(
"polymorphic_table",
[node1, node2],
[settings_default, settings_default],
"shard1",
)
create_tables(
"compact_parts_only",
[node1, node2],
[settings_compact_only, settings_compact_only],
"shard1",
)
create_tables(
"non_adaptive_table",
[node1, node2],
[settings_not_adaptive, settings_not_adaptive],
"shard1",
)
create_tables(
"polymorphic_table_compact",
[node3, node4],
[settings_compact, settings_wide],
"shard2",
)
create_tables(
"polymorphic_table_wide",
[node3, node4],
[settings_wide, settings_compact],
"shard2",
)
create_tables_old_format("polymorphic_table", [node5, node6], "shard3")
2020-02-13 14:59:15 +00:00
yield cluster
finally:
cluster.shutdown()
2020-02-19 14:55:03 +00:00
@pytest.mark.parametrize(
("first_node", "second_node"),
2020-02-19 14:55:03 +00:00
[
(node1, node2), # compact parts
(node5, node6), # compact parts, old-format
],
2020-02-19 14:55:03 +00:00
)
def test_polymorphic_parts_basics(start_cluster, first_node, second_node):
first_node.query("SYSTEM STOP MERGES")
second_node.query("SYSTEM STOP MERGES")
2020-02-13 14:59:15 +00:00
for size in [300, 300, 600]:
insert_random_data("polymorphic_table", first_node, size)
2020-02-19 14:55:03 +00:00
second_node.query("SYSTEM SYNC REPLICA polymorphic_table", timeout=20)
2020-02-13 14:59:15 +00:00
2020-02-19 14:55:03 +00:00
assert first_node.query("SELECT count() FROM polymorphic_table") == "1200\n"
assert second_node.query("SELECT count() FROM polymorphic_table") == "1200\n"
2020-02-13 14:59:15 +00:00
expected = "Compact\t2\nWide\t1\n"
assert TSV(
first_node.query(
"SELECT part_type, count() FROM system.parts "
"WHERE table = 'polymorphic_table' AND active GROUP BY part_type ORDER BY part_type"
)
) == TSV(expected)
assert TSV(
second_node.query(
"SELECT part_type, count() FROM system.parts "
"WHERE table = 'polymorphic_table' AND active GROUP BY part_type ORDER BY part_type"
)
) == TSV(expected)
2020-02-13 14:59:15 +00:00
2020-02-19 14:55:03 +00:00
first_node.query("SYSTEM START MERGES")
second_node.query("SYSTEM START MERGES")
2020-02-13 14:59:15 +00:00
for _ in range(40):
insert_random_data("polymorphic_table", first_node, 10)
insert_random_data("polymorphic_table", second_node, 10)
2020-02-13 14:59:15 +00:00
2020-02-19 14:55:03 +00:00
first_node.query("SYSTEM SYNC REPLICA polymorphic_table", timeout=20)
second_node.query("SYSTEM SYNC REPLICA polymorphic_table", timeout=20)
2020-02-13 14:59:15 +00:00
2020-02-19 14:55:03 +00:00
assert first_node.query("SELECT count() FROM polymorphic_table") == "2000\n"
assert second_node.query("SELECT count() FROM polymorphic_table") == "2000\n"
2020-02-13 14:59:15 +00:00
2020-02-19 14:55:03 +00:00
first_node.query("OPTIMIZE TABLE polymorphic_table FINAL")
second_node.query("SYSTEM SYNC REPLICA polymorphic_table", timeout=20)
2020-02-13 14:59:15 +00:00
2020-02-19 14:55:03 +00:00
assert first_node.query("SELECT count() FROM polymorphic_table") == "2000\n"
assert second_node.query("SELECT count() FROM polymorphic_table") == "2000\n"
2020-02-13 14:59:15 +00:00
assert (
first_node.query(
"SELECT DISTINCT part_type FROM system.parts WHERE table = 'polymorphic_table' AND active"
)
== "Wide\n"
)
assert (
second_node.query(
"SELECT DISTINCT part_type FROM system.parts WHERE table = 'polymorphic_table' AND active"
)
== "Wide\n"
)
2020-02-13 14:59:15 +00:00
# Check alters and mutations also work
2020-02-19 14:55:03 +00:00
first_node.query("ALTER TABLE polymorphic_table ADD COLUMN ss String")
first_node.query("ALTER TABLE polymorphic_table UPDATE ss = toString(id) WHERE 1")
2020-02-19 14:55:03 +00:00
second_node.query("SYSTEM SYNC REPLICA polymorphic_table", timeout=20)
2020-02-19 14:55:03 +00:00
first_node.query("SELECT count(ss) FROM polymorphic_table") == "2000\n"
first_node.query("SELECT uniqExact(ss) FROM polymorphic_table") == "600\n"
2020-02-19 14:55:03 +00:00
second_node.query("SELECT count(ss) FROM polymorphic_table") == "2000\n"
second_node.query("SELECT uniqExact(ss) FROM polymorphic_table") == "600\n"
2020-04-28 23:07:11 +00:00
# Checks mostly that merge from compact part to compact part works.
def test_compact_parts_only(start_cluster):
for i in range(20):
insert_random_data("compact_parts_only", node1, 100)
insert_random_data("compact_parts_only", node2, 100)
2020-04-28 23:07:11 +00:00
node1.query("SYSTEM SYNC REPLICA compact_parts_only", timeout=20)
node2.query("SYSTEM SYNC REPLICA compact_parts_only", timeout=20)
assert node1.query("SELECT count() FROM compact_parts_only") == "4000\n"
assert node2.query("SELECT count() FROM compact_parts_only") == "4000\n"
assert (
node1.query(
"SELECT DISTINCT part_type FROM system.parts WHERE table = 'compact_parts_only' AND active"
)
== "Compact\n"
)
assert (
node2.query(
"SELECT DISTINCT part_type FROM system.parts WHERE table = 'compact_parts_only' AND active"
)
== "Compact\n"
)
2020-04-28 23:07:11 +00:00
node1.query("OPTIMIZE TABLE compact_parts_only FINAL")
node2.query("SYSTEM SYNC REPLICA compact_parts_only", timeout=20)
assert node2.query("SELECT count() FROM compact_parts_only") == "4000\n"
expected = "Compact\t1\n"
assert TSV(
node1.query(
"SELECT part_type, count() FROM system.parts "
"WHERE table = 'compact_parts_only' AND active GROUP BY part_type ORDER BY part_type"
)
) == TSV(expected)
assert TSV(
node2.query(
"SELECT part_type, count() FROM system.parts "
"WHERE table = 'compact_parts_only' AND active GROUP BY part_type ORDER BY part_type"
)
) == TSV(expected)
2020-04-28 23:07:11 +00:00
2020-02-13 14:59:15 +00:00
# Check that follower replicas create parts of the same type, which leader has chosen at merge.
@pytest.mark.parametrize(
("table", "part_type"),
[("polymorphic_table_compact", "Compact"), ("polymorphic_table_wide", "Wide")],
2020-02-13 14:59:15 +00:00
)
def test_different_part_types_on_replicas(start_cluster, table, part_type):
leader = node3
follower = node4
assert (
leader.query(
"SELECT is_leader FROM system.replicas WHERE table = '{}'".format(table)
)
== "1\n"
)
assert (
node4.query(
"SELECT is_leader FROM system.replicas WHERE table = '{}'".format(table)
)
== "0\n"
)
2020-02-13 14:59:15 +00:00
for _ in range(3):
insert_random_data(table, leader, 100)
2020-02-13 14:59:15 +00:00
2024-03-15 22:30:49 +00:00
exec_query_with_retry(
leader,
"OPTIMIZE TABLE {} FINAL".format(table),
settings={"optimize_throw_if_noop": 1},
silent=True,
)
follower.query("SYSTEM SYNC REPLICA {}".format(table), timeout=20)
2020-02-13 14:59:15 +00:00
expected = "{}\t1\n".format(part_type)
assert TSV(
leader.query(
"SELECT part_type, count() FROM system.parts "
"WHERE table = '{}' AND active GROUP BY part_type ORDER BY part_type".format(
table
)
)
) == TSV(expected)
assert TSV(
follower.query(
"SELECT part_type, count() FROM system.parts "
"WHERE table = '{}' AND active GROUP BY part_type ORDER BY part_type".format(
table
)
)
) == TSV(expected)
node7 = cluster.add_instance(
"node7",
user_configs=["configs_old/users.d/not_optimize_count.xml"],
with_zookeeper=True,
2024-02-26 12:45:20 +00:00
image="clickhouse/clickhouse-server",
tag=CLICKHOUSE_CI_MIN_TESTED_VERSION,
stay_alive=True,
with_installed_binary=True,
)
node8 = cluster.add_instance(
"node8",
user_configs=["configs/users.d/not_optimize_count.xml"],
with_zookeeper=True,
)
settings7 = {"index_granularity_bytes": 10485760}
settings8 = {
"index_granularity_bytes": 10485760,
"min_rows_for_wide_part": 512,
}
@pytest.fixture(scope="module")
def start_cluster_diff_versions():
try:
for name in ["polymorphic_table", "polymorphic_table_2"]:
cluster.start()
node7.query(
"""
CREATE TABLE {name}(date Date, id UInt32, s String, arr Array(Int32))
ENGINE = ReplicatedMergeTree('/clickhouse/tables/test/shard5/{name}', '1')
PARTITION BY toYYYYMM(date)
ORDER BY id
SETTINGS index_granularity = 64, index_granularity_bytes = {index_granularity_bytes}
""".format(
name=name, **settings7
)
)
node8.query(
"""
CREATE TABLE {name}(date Date, id UInt32, s String, arr Array(Int32))
ENGINE = ReplicatedMergeTree('/clickhouse/tables/test/shard5/{name}', '2')
PARTITION BY toYYYYMM(date)
ORDER BY id
SETTINGS index_granularity = 64, index_granularity_bytes = {index_granularity_bytes},
min_rows_for_wide_part = {min_rows_for_wide_part}, min_bytes_for_wide_part = {min_bytes_for_wide_part}
""".format(
name=name, **settings8
)
)
yield cluster
finally:
cluster.shutdown()
def test_polymorphic_parts_non_adaptive(start_cluster):
node1.query("SYSTEM STOP MERGES")
node2.query("SYSTEM STOP MERGES")
insert_random_data("non_adaptive_table", node1, 100)
node2.query("SYSTEM SYNC REPLICA non_adaptive_table", timeout=20)
insert_random_data("non_adaptive_table", node2, 100)
node1.query("SYSTEM SYNC REPLICA non_adaptive_table", timeout=20)
assert TSV(
node1.query(
"SELECT part_type, count() FROM system.parts "
"WHERE table = 'non_adaptive_table' AND active GROUP BY part_type ORDER BY part_type"
)
) == TSV("Wide\t2\n")
assert TSV(
node2.query(
"SELECT part_type, count() FROM system.parts "
"WHERE table = 'non_adaptive_table' AND active GROUP BY part_type ORDER BY part_type"
)
) == TSV("Wide\t2\n")
assert node1.contains_in_log(
"<Warning> default.non_adaptive_table ([0-9a-f-]*): Table can't create parts with adaptive granularity"
)
2020-04-29 17:14:49 +00:00
2020-04-27 13:37:05 +00:00
def test_polymorphic_parts_index(start_cluster):
node1.query(
2022-06-23 19:40:05 +00:00
"CREATE DATABASE test_index ENGINE=Ordinary",
settings={"allow_deprecated_database_ordinary": 1},
) # Different paths with Atomic
node1.query(
"""
2020-09-21 21:09:50 +00:00
CREATE TABLE test_index.index_compact(a UInt32, s String)
2020-08-07 17:16:59 +00:00
ENGINE = MergeTree ORDER BY a
2023-07-17 23:50:43 +00:00
SETTINGS min_rows_for_wide_part = 1000, index_granularity = 128, merge_max_block_size = 100, compress_marks=false, compress_primary_key=false, ratio_of_defaults_for_sparse_serialization=1"""
)
node1.query(
"INSERT INTO test_index.index_compact SELECT number, toString(number) FROM numbers(100)"
)
node1.query(
"INSERT INTO test_index.index_compact SELECT number, toString(number) FROM numbers(30)"
)
2020-09-21 21:09:50 +00:00
node1.query("OPTIMIZE TABLE test_index.index_compact FINAL")
2020-04-27 13:37:05 +00:00
assert (
node1.query(
"SELECT part_type FROM system.parts WHERE table = 'index_compact' AND active"
)
== "Compact\n"
)
assert (
node1.query(
"SELECT marks FROM system.parts WHERE table = 'index_compact' AND active"
)
== "2\n"
)
index_path = os.path.join(
node1.path, "database/data/test_index/index_compact/all_1_2_1/primary.idx"
)
f = open(index_path, "rb")
2020-04-27 13:37:05 +00:00
assert os.path.getsize(index_path) == 8
assert struct.unpack("I", f.read(4))[0] == 0
assert struct.unpack("I", f.read(4))[0] == 99