2021-07-06 10:48:13 +00:00
|
|
|
import logging
|
2018-06-29 13:46:28 +00:00
|
|
|
import random
|
2020-09-16 04:26:10 +00:00
|
|
|
import threading
|
|
|
|
import time
|
2018-06-29 13:46:28 +00:00
|
|
|
from collections import Counter
|
|
|
|
|
|
|
|
import pytest
|
|
|
|
from helpers.cluster import ClickHouseCluster
|
|
|
|
|
|
|
|
cluster = ClickHouseCluster(__file__)
|
|
|
|
|
2019-08-21 13:10:33 +00:00
|
|
|
node1 = cluster.add_instance("node1", macros={"cluster": "test1"}, with_zookeeper=True)
|
2019-06-17 19:41:48 +00:00
|
|
|
# Check, that limits on max part size for merges doesn`t affect mutations
|
2020-09-16 04:26:10 +00:00
|
|
|
node2 = cluster.add_instance(
|
|
|
|
"node2",
|
|
|
|
macros={"cluster": "test1"},
|
|
|
|
main_configs=["configs/merge_tree.xml"],
|
2023-07-18 20:13:28 +00:00
|
|
|
user_configs=["configs/users.xml"],
|
2020-09-16 04:26:10 +00:00
|
|
|
with_zookeeper=True,
|
|
|
|
)
|
2019-08-21 13:10:33 +00:00
|
|
|
|
2020-09-16 04:26:10 +00:00
|
|
|
node3 = cluster.add_instance(
|
|
|
|
"node3",
|
|
|
|
macros={"cluster": "test2"},
|
|
|
|
main_configs=["configs/merge_tree_max_parts.xml"],
|
2023-07-18 20:13:28 +00:00
|
|
|
user_configs=["configs/users.xml"],
|
2020-09-16 04:26:10 +00:00
|
|
|
with_zookeeper=True,
|
|
|
|
)
|
|
|
|
node4 = cluster.add_instance(
|
|
|
|
"node4",
|
|
|
|
macros={"cluster": "test2"},
|
|
|
|
main_configs=["configs/merge_tree_max_parts.xml"],
|
2023-07-18 20:13:28 +00:00
|
|
|
user_configs=["configs/users.xml"],
|
2020-09-16 04:26:10 +00:00
|
|
|
with_zookeeper=True,
|
|
|
|
)
|
2019-08-21 13:10:33 +00:00
|
|
|
|
2019-08-22 19:54:42 +00:00
|
|
|
node5 = cluster.add_instance(
|
|
|
|
"node5",
|
|
|
|
macros={"cluster": "test3"},
|
|
|
|
main_configs=["configs/merge_tree_max_parts.xml"],
|
2023-07-18 20:13:28 +00:00
|
|
|
user_configs=["configs/users.xml"],
|
2019-08-22 19:54:42 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
all_nodes = [node1, node2, node3, node4, node5]
|
2018-06-29 13:46:28 +00:00
|
|
|
|
2022-03-22 16:39:58 +00:00
|
|
|
|
2021-07-28 10:21:59 +00:00
|
|
|
def prepare_cluster():
|
|
|
|
for node in all_nodes:
|
|
|
|
node.query("DROP TABLE IF EXISTS test_mutations SYNC")
|
|
|
|
|
|
|
|
for node in [node1, node2, node3, node4]:
|
|
|
|
node.query(
|
|
|
|
"""
|
|
|
|
CREATE TABLE test_mutations(d Date, x UInt32, i UInt32)
|
|
|
|
ENGINE ReplicatedMergeTree('/clickhouse/{cluster}/tables/test/test_mutations', '{instance}')
|
|
|
|
ORDER BY x
|
|
|
|
PARTITION BY toYYYYMM(d)
|
|
|
|
SETTINGS number_of_free_entries_in_pool_to_execute_mutation=0
|
|
|
|
"""
|
|
|
|
)
|
|
|
|
|
|
|
|
node5.query(
|
|
|
|
"CREATE TABLE test_mutations(d Date, x UInt32, i UInt32) ENGINE MergeTree() ORDER BY x PARTITION BY toYYYYMM(d)"
|
|
|
|
)
|
|
|
|
|
2020-09-16 04:26:10 +00:00
|
|
|
|
2018-06-29 13:46:28 +00:00
|
|
|
@pytest.fixture(scope="module")
|
|
|
|
def started_cluster():
|
|
|
|
try:
|
|
|
|
cluster.start()
|
|
|
|
yield cluster
|
|
|
|
|
|
|
|
finally:
|
|
|
|
cluster.shutdown()
|
|
|
|
|
|
|
|
|
|
|
|
class Runner:
|
2019-08-21 13:10:33 +00:00
|
|
|
def __init__(self, nodes):
|
|
|
|
self.nodes = nodes
|
2018-06-29 13:46:28 +00:00
|
|
|
self.mtx = threading.Lock()
|
|
|
|
self.total_inserted_xs = 0
|
|
|
|
self.total_inserted_rows = 0
|
|
|
|
|
|
|
|
self.total_mutations = 0
|
|
|
|
self.total_deleted_xs = 0
|
|
|
|
self.total_deleted_rows = 0
|
|
|
|
|
|
|
|
self.current_xs = Counter()
|
|
|
|
|
|
|
|
self.currently_inserting_xs = Counter()
|
|
|
|
self.currently_deleting_xs = set()
|
|
|
|
|
|
|
|
self.stop_ev = threading.Event()
|
|
|
|
|
2019-08-21 13:10:33 +00:00
|
|
|
self.exceptions = []
|
|
|
|
|
2019-08-22 19:54:42 +00:00
|
|
|
def do_insert(self, thread_num, partitions_num):
|
2018-06-29 13:46:28 +00:00
|
|
|
self.stop_ev.wait(random.random())
|
|
|
|
|
|
|
|
# Each thread inserts a small random number of rows with random year, month 01 and day determined
|
|
|
|
# by the thread number. The idea is to avoid spurious duplicates and to insert into a
|
|
|
|
# nontrivial number of partitions.
|
|
|
|
month = "01"
|
|
|
|
day = str(thread_num + 1).zfill(2)
|
|
|
|
i = 1
|
|
|
|
while not self.stop_ev.is_set():
|
|
|
|
xs = [random.randint(1, 10) for _ in range(random.randint(1, 10))]
|
|
|
|
with self.mtx:
|
|
|
|
xs = [x for x in xs if x not in self.currently_deleting_xs]
|
|
|
|
if len(xs) == 0:
|
|
|
|
continue
|
|
|
|
for x in xs:
|
|
|
|
self.currently_inserting_xs[x] += 1
|
|
|
|
|
2019-08-22 19:54:42 +00:00
|
|
|
year = 2000 + random.randint(0, partitions_num)
|
2018-06-29 13:46:28 +00:00
|
|
|
date_str = "{year}-{month}-{day}".format(year=year, month=month, day=day)
|
|
|
|
payload = ""
|
|
|
|
for x in xs:
|
|
|
|
payload += "{date_str} {x} {i}\n".format(date_str=date_str, x=x, i=i)
|
|
|
|
i += 1
|
|
|
|
|
|
|
|
try:
|
2021-07-06 10:48:13 +00:00
|
|
|
logging.debug(f"thread {thread_num}: insert for {date_str}: {xs}")
|
2019-08-21 13:10:33 +00:00
|
|
|
random.choice(self.nodes).query(
|
|
|
|
"INSERT INTO test_mutations FORMAT TSV", payload
|
|
|
|
)
|
2018-06-29 13:46:28 +00:00
|
|
|
|
|
|
|
with self.mtx:
|
|
|
|
for x in xs:
|
|
|
|
self.current_xs[x] += 1
|
|
|
|
self.total_inserted_xs += sum(xs)
|
|
|
|
self.total_inserted_rows += len(xs)
|
|
|
|
|
2020-10-02 16:54:07 +00:00
|
|
|
except Exception as e:
|
2021-07-06 10:48:13 +00:00
|
|
|
logging.debug(f"Exception while inserting: {e}")
|
2019-08-21 13:10:33 +00:00
|
|
|
self.exceptions.append(e)
|
2018-06-29 13:46:28 +00:00
|
|
|
finally:
|
|
|
|
with self.mtx:
|
|
|
|
for x in xs:
|
|
|
|
self.currently_inserting_xs[x] -= 1
|
|
|
|
|
|
|
|
self.stop_ev.wait(0.2 + random.random() / 5)
|
|
|
|
|
|
|
|
def do_delete(self, thread_num):
|
|
|
|
self.stop_ev.wait(1.0 + random.random())
|
|
|
|
|
|
|
|
while not self.stop_ev.is_set():
|
|
|
|
chosen = False
|
|
|
|
with self.mtx:
|
|
|
|
if self.current_xs:
|
|
|
|
x = random.choice(list(self.current_xs.elements()))
|
|
|
|
|
|
|
|
if (
|
|
|
|
self.currently_inserting_xs[x] == 0
|
|
|
|
and x not in self.currently_deleting_xs
|
|
|
|
):
|
|
|
|
chosen = True
|
|
|
|
self.currently_deleting_xs.add(x)
|
|
|
|
to_delete_count = self.current_xs[x]
|
|
|
|
|
|
|
|
if not chosen:
|
|
|
|
self.stop_ev.wait(0.1 * random.random())
|
|
|
|
continue
|
|
|
|
|
|
|
|
try:
|
2021-07-06 10:48:13 +00:00
|
|
|
logging.debug(f"thread {thread_num}: delete {to_delete_count} * {x}")
|
2019-08-21 13:10:33 +00:00
|
|
|
random.choice(self.nodes).query(
|
|
|
|
"ALTER TABLE test_mutations DELETE WHERE x = {}".format(x)
|
2022-03-22 16:39:58 +00:00
|
|
|
)
|
2018-06-29 13:46:28 +00:00
|
|
|
|
|
|
|
with self.mtx:
|
|
|
|
self.total_mutations += 1
|
|
|
|
self.current_xs[x] -= to_delete_count
|
|
|
|
self.total_deleted_xs += to_delete_count * x
|
|
|
|
self.total_deleted_rows += to_delete_count
|
|
|
|
|
2020-10-02 16:54:07 +00:00
|
|
|
except Exception as e:
|
2021-07-06 10:48:13 +00:00
|
|
|
logging.debug(f"Exception while deleting: {e}")
|
2018-06-29 13:46:28 +00:00
|
|
|
finally:
|
|
|
|
with self.mtx:
|
|
|
|
self.currently_deleting_xs.remove(x)
|
|
|
|
|
|
|
|
self.stop_ev.wait(1.0 + random.random() * 2)
|
|
|
|
|
|
|
|
|
2019-08-21 13:10:33 +00:00
|
|
|
def wait_for_mutations(nodes, number_of_mutations):
|
|
|
|
for i in range(100): # wait for replication 80 seconds max
|
|
|
|
time.sleep(0.8)
|
|
|
|
|
|
|
|
def get_done_mutations(node):
|
|
|
|
return int(
|
|
|
|
node.query(
|
|
|
|
"SELECT sum(is_done) FROM system.mutations WHERE table = 'test_mutations'"
|
|
|
|
).rstrip()
|
2022-03-22 16:39:58 +00:00
|
|
|
)
|
2019-08-21 13:10:33 +00:00
|
|
|
|
|
|
|
if all([get_done_mutations(n) == number_of_mutations for n in nodes]):
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
|
2018-06-29 13:46:28 +00:00
|
|
|
def test_mutations(started_cluster):
|
2021-07-28 10:21:59 +00:00
|
|
|
prepare_cluster()
|
|
|
|
|
2018-09-07 11:51:51 +00:00
|
|
|
DURATION_SECONDS = 30
|
2019-08-21 13:10:33 +00:00
|
|
|
nodes = [node1, node2]
|
2018-06-29 13:46:28 +00:00
|
|
|
|
2019-08-21 13:10:33 +00:00
|
|
|
runner = Runner(nodes)
|
2018-06-29 13:46:28 +00:00
|
|
|
|
|
|
|
threads = []
|
|
|
|
for thread_num in range(5):
|
2019-08-22 19:54:42 +00:00
|
|
|
threads.append(threading.Thread(target=runner.do_insert, args=(thread_num, 10)))
|
2018-06-29 13:46:28 +00:00
|
|
|
|
|
|
|
for thread_num in (11, 12, 13):
|
|
|
|
threads.append(threading.Thread(target=runner.do_delete, args=(thread_num,)))
|
|
|
|
|
|
|
|
for t in threads:
|
|
|
|
t.start()
|
|
|
|
|
|
|
|
time.sleep(DURATION_SECONDS)
|
|
|
|
runner.stop_ev.set()
|
|
|
|
|
|
|
|
for t in threads:
|
|
|
|
t.join()
|
|
|
|
|
|
|
|
# Sanity check: at least something was inserted and something was deleted
|
|
|
|
assert runner.total_inserted_rows > 0
|
|
|
|
assert runner.total_mutations > 0
|
|
|
|
|
2019-08-21 13:10:33 +00:00
|
|
|
all_done = wait_for_mutations(nodes, runner.total_mutations)
|
2021-07-06 10:48:13 +00:00
|
|
|
logging.debug(f"Total mutations: {runner.total_mutations}")
|
2019-08-22 19:54:42 +00:00
|
|
|
for node in nodes:
|
2021-07-06 10:48:13 +00:00
|
|
|
logging.debug(
|
|
|
|
node.query(
|
2020-10-02 16:54:07 +00:00
|
|
|
"SELECT mutation_id, command, parts_to_do, is_done FROM system.mutations WHERE table = 'test_mutations' FORMAT TSVWithNames"
|
|
|
|
)
|
2022-03-22 16:39:58 +00:00
|
|
|
)
|
2018-06-29 13:46:28 +00:00
|
|
|
assert all_done
|
|
|
|
|
|
|
|
expected_sum = runner.total_inserted_xs - runner.total_deleted_xs
|
|
|
|
actual_sums = []
|
|
|
|
for i, node in enumerate(nodes):
|
|
|
|
actual_sums.append(
|
|
|
|
int(node.query("SELECT sum(x) FROM test_mutations").rstrip())
|
2022-03-22 16:39:58 +00:00
|
|
|
)
|
2018-06-29 13:46:28 +00:00
|
|
|
assert actual_sums[i] == expected_sum
|
2019-08-21 13:10:33 +00:00
|
|
|
|
|
|
|
|
2019-08-22 19:54:42 +00:00
|
|
|
@pytest.mark.parametrize(
|
2020-09-16 04:26:10 +00:00
|
|
|
("nodes",),
|
2019-08-22 19:54:42 +00:00
|
|
|
[
|
2020-09-16 04:26:10 +00:00
|
|
|
(
|
2022-03-22 16:39:58 +00:00
|
|
|
[
|
2020-09-16 04:26:10 +00:00
|
|
|
node5,
|
|
|
|
],
|
|
|
|
), # MergeTree
|
|
|
|
([node3, node4],), # ReplicatedMergeTree
|
2019-08-22 19:54:42 +00:00
|
|
|
],
|
|
|
|
)
|
|
|
|
def test_mutations_dont_prevent_merges(started_cluster, nodes):
|
2021-07-28 10:21:59 +00:00
|
|
|
prepare_cluster()
|
|
|
|
|
2019-08-22 19:54:42 +00:00
|
|
|
for year in range(2000, 2016):
|
2019-08-21 13:10:33 +00:00
|
|
|
rows = ""
|
|
|
|
date_str = "{}-01-{}".format(year, random.randint(1, 10))
|
|
|
|
for i in range(10):
|
|
|
|
rows += "{} {} {}\n".format(date_str, random.randint(1, 10), i)
|
2019-08-22 19:54:42 +00:00
|
|
|
nodes[0].query("INSERT INTO test_mutations FORMAT TSV", rows)
|
2019-08-21 13:10:33 +00:00
|
|
|
|
2019-08-22 19:54:42 +00:00
|
|
|
# will run mutations of 16 parts in parallel, mutations will sleep for about 20 seconds
|
|
|
|
nodes[0].query("ALTER TABLE test_mutations UPDATE i = sleepEachRow(2) WHERE 1")
|
2019-08-21 13:10:33 +00:00
|
|
|
|
|
|
|
runner = Runner(nodes)
|
|
|
|
threads = []
|
2019-08-22 19:54:42 +00:00
|
|
|
for thread_num in range(2):
|
|
|
|
threads.append(threading.Thread(target=runner.do_insert, args=(thread_num, 0)))
|
2019-08-21 13:10:33 +00:00
|
|
|
|
2019-08-22 19:54:42 +00:00
|
|
|
# will insert approx 8-10 new parts per 1 second into one partition
|
2019-08-21 13:10:33 +00:00
|
|
|
for t in threads:
|
|
|
|
t.start()
|
|
|
|
|
|
|
|
all_done = wait_for_mutations(nodes, 1)
|
|
|
|
|
|
|
|
runner.stop_ev.set()
|
|
|
|
for t in threads:
|
|
|
|
t.join()
|
|
|
|
|
2019-08-22 19:54:42 +00:00
|
|
|
for node in nodes:
|
2021-07-06 10:48:13 +00:00
|
|
|
logging.debug(
|
|
|
|
node.query(
|
2020-10-02 16:54:07 +00:00
|
|
|
"SELECT mutation_id, command, parts_to_do, is_done FROM system.mutations WHERE table = 'test_mutations' FORMAT TSVWithNames"
|
|
|
|
)
|
2022-03-22 16:39:58 +00:00
|
|
|
)
|
2021-07-06 10:48:13 +00:00
|
|
|
logging.debug(
|
|
|
|
node.query(
|
2020-10-02 16:54:07 +00:00
|
|
|
"SELECT partition, count(name), sum(active), sum(active*rows) FROM system.parts WHERE table ='test_mutations' GROUP BY partition FORMAT TSVWithNames"
|
|
|
|
)
|
2022-03-22 16:39:58 +00:00
|
|
|
)
|
2019-08-22 19:54:42 +00:00
|
|
|
|
2021-09-30 21:26:24 +00:00
|
|
|
assert all_done, "All done"
|
2019-08-21 13:10:33 +00:00
|
|
|
assert all([str(e).find("Too many parts") < 0 for e in runner.exceptions])
|