2019-12-17 13:45:53 +00:00
|
|
|
import logging
|
2020-12-10 09:19:42 +00:00
|
|
|
import sys
|
2019-12-17 13:45:53 +00:00
|
|
|
|
|
|
|
import pytest
|
|
|
|
from helpers.cluster import ClickHouseCluster
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.fixture(scope="module")
|
|
|
|
def cluster():
|
|
|
|
try:
|
|
|
|
cluster = ClickHouseCluster(__file__)
|
2022-03-22 16:39:58 +00:00
|
|
|
cluster.add_instance(
|
|
|
|
"node",
|
2024-05-13 12:46:42 +00:00
|
|
|
main_configs=["configs/storage_configuration.xml", "configs/ssl.xml"],
|
2022-03-22 16:39:58 +00:00
|
|
|
with_minio=True,
|
|
|
|
)
|
2019-12-17 13:45:53 +00:00
|
|
|
logging.info("Starting cluster...")
|
|
|
|
cluster.start()
|
|
|
|
logging.info("Cluster started")
|
|
|
|
|
|
|
|
yield cluster
|
|
|
|
finally:
|
|
|
|
cluster.shutdown()
|
|
|
|
|
|
|
|
|
2022-03-22 16:39:58 +00:00
|
|
|
def assert_objects_count(cluster, objects_count, path="data/"):
|
2020-12-08 18:31:57 +00:00
|
|
|
minio = cluster.minio_client
|
2022-09-16 12:06:26 +00:00
|
|
|
s3_objects = list(minio.list_objects(cluster.minio_bucket, path, recursive=True))
|
2020-12-08 18:31:57 +00:00
|
|
|
if objects_count != len(s3_objects):
|
|
|
|
for s3_object in s3_objects:
|
|
|
|
object_meta = minio.stat_object(cluster.minio_bucket, s3_object.object_name)
|
|
|
|
logging.info("Existing S3 object: %s", str(object_meta))
|
|
|
|
assert objects_count == len(s3_objects)
|
|
|
|
|
2022-03-22 16:39:58 +00:00
|
|
|
|
2021-10-09 12:14:23 +00:00
|
|
|
# TinyLog: files: id.bin, sizes.json
|
|
|
|
# INSERT overwrites 1 file (`sizes.json`) and appends 1 file (`id.bin`), so
|
|
|
|
# files_overhead=1, files_overhead_per_insert=1
|
|
|
|
#
|
|
|
|
# Log: files: id.bin, __marks.mrk, sizes.json
|
|
|
|
# INSERT overwrites 1 file (`sizes.json`), and appends 2 files (`id.bin`, `__marks.mrk`), so
|
|
|
|
# files_overhead=1, files_overhead_per_insert=2
|
|
|
|
#
|
|
|
|
# StripeLog: files: data.bin, index.mrk, sizes.json
|
|
|
|
# INSERT overwrites 1 file (`sizes.json`), and appends 2 files (`index.mrk`, `data.bin`), so
|
|
|
|
# files_overhead=1, files_overhead_per_insert=2
|
2020-02-14 14:28:33 +00:00
|
|
|
@pytest.mark.parametrize(
|
|
|
|
"log_engine,files_overhead,files_overhead_per_insert",
|
2022-06-03 16:46:06 +00:00
|
|
|
[
|
|
|
|
pytest.param("TinyLog", 1, 1, id="TinyLog"),
|
|
|
|
pytest.param("Log", 1, 2, id="Log"),
|
|
|
|
pytest.param("StripeLog", 1, 2, id="StripeLog"),
|
|
|
|
],
|
2022-03-22 16:39:58 +00:00
|
|
|
)
|
2020-02-14 14:28:33 +00:00
|
|
|
def test_log_family_s3(cluster, log_engine, files_overhead, files_overhead_per_insert):
|
2019-12-17 13:45:53 +00:00
|
|
|
node = cluster.instances["node"]
|
|
|
|
|
2022-03-22 16:39:58 +00:00
|
|
|
node.query(
|
|
|
|
"CREATE TABLE s3_test (id UInt64) ENGINE={} SETTINGS disk = 's3'".format(
|
|
|
|
log_engine
|
|
|
|
)
|
|
|
|
)
|
2020-01-27 18:44:30 +00:00
|
|
|
|
2022-06-03 16:45:06 +00:00
|
|
|
try:
|
|
|
|
node.query("INSERT INTO s3_test SELECT number FROM numbers(5)")
|
|
|
|
assert node.query("SELECT * FROM s3_test") == "0\n1\n2\n3\n4\n"
|
|
|
|
assert_objects_count(cluster, files_overhead_per_insert + files_overhead)
|
2020-01-27 18:44:30 +00:00
|
|
|
|
2022-06-03 16:45:06 +00:00
|
|
|
node.query("INSERT INTO s3_test SELECT number + 5 FROM numbers(3)")
|
2022-06-03 16:46:43 +00:00
|
|
|
assert (
|
|
|
|
node.query("SELECT * FROM s3_test order by id")
|
|
|
|
== "0\n1\n2\n3\n4\n5\n6\n7\n"
|
|
|
|
)
|
2022-06-03 16:45:06 +00:00
|
|
|
assert_objects_count(cluster, files_overhead_per_insert * 2 + files_overhead)
|
2020-01-27 18:44:30 +00:00
|
|
|
|
2022-06-03 16:45:06 +00:00
|
|
|
node.query("INSERT INTO s3_test SELECT number + 8 FROM numbers(1)")
|
|
|
|
assert (
|
2022-06-03 16:46:43 +00:00
|
|
|
node.query("SELECT * FROM s3_test order by id")
|
|
|
|
== "0\n1\n2\n3\n4\n5\n6\n7\n8\n"
|
2022-06-03 16:45:06 +00:00
|
|
|
)
|
|
|
|
assert_objects_count(cluster, files_overhead_per_insert * 3 + files_overhead)
|
2020-01-27 18:44:30 +00:00
|
|
|
|
2022-06-03 16:45:06 +00:00
|
|
|
node.query("TRUNCATE TABLE s3_test")
|
|
|
|
assert_objects_count(cluster, 0)
|
|
|
|
finally:
|
|
|
|
node.query("DROP TABLE s3_test")
|
2024-05-13 12:46:42 +00:00
|
|
|
|
|
|
|
|
|
|
|
# Imitate case when error occurs while inserting into table.
|
|
|
|
# For examle S3::TooManyRequests.
|
|
|
|
# In that case we can update data file, but not the size file.
|
|
|
|
# So due to exception we should do truncate of the data file to undo the insert query.
|
|
|
|
# See FileChecker::repair().
|
|
|
|
def test_stripe_log_truncate(cluster):
|
|
|
|
node = cluster.instances["node"]
|
|
|
|
|
|
|
|
node.query(
|
|
|
|
"""
|
|
|
|
CREATE TABLE stripe_table (
|
|
|
|
a int
|
|
|
|
) ENGINE = StripeLog()
|
|
|
|
SETTINGS storage_policy='s3_no_retries'
|
|
|
|
"""
|
|
|
|
)
|
|
|
|
|
|
|
|
node.query("SYSTEM ENABLE FAILPOINT stripe_log_sink_write_fallpoint")
|
|
|
|
node.query(
|
|
|
|
"""
|
|
|
|
INSERT INTO stripe_table SELECT number FROM numbers(10)
|
|
|
|
""",
|
|
|
|
ignore_error=True,
|
|
|
|
)
|
|
|
|
node.query("SYSTEM DISABLE FAILPOINT stripe_log_sink_write_fallpoint")
|
|
|
|
node.query("SELECT count(*) FROM stripe_table") == "0\n"
|
|
|
|
node.query("INSERT INTO stripe_table SELECT number FROM numbers(10)")
|
|
|
|
node.query("SELECT count(*) FROM stripe_table") == "10\n"
|
|
|
|
|
|
|
|
# Make sure that everything is okey with the table after restart.
|
|
|
|
node.query("DETACH TABLE stripe_table")
|
|
|
|
node.query("ATTACH TABLE stripe_table")
|
|
|
|
|
|
|
|
assert node.query("DROP TABLE stripe_table") == ""
|