ClickHouse/tests/integration/test_disks_app_func/test.py
2024-07-29 01:15:15 +03:00

339 lines
7.5 KiB
Python

from helpers.cluster import ClickHouseCluster
import pytest
@pytest.fixture(scope="module")
def started_cluster():
global cluster
try:
cluster = ClickHouseCluster(__file__)
cluster.add_instance(
"disks_app_test",
main_configs=["config.xml"],
with_minio=True,
)
cluster.start()
yield cluster
finally:
cluster.shutdown()
def write(source, disk, path):
source.exec_in_container(
[
"bash",
"-c",
"echo 'tester' |"
+ " ".join(
[
"/usr/bin/clickhouse",
"disks",
"--save-logs",
"--disk",
f"{disk}",
"--query",
f"'write {path}'",
]
),
]
)
def mkdir(source, disk, path):
source.exec_in_container(
[
"/usr/bin/clickhouse",
"disks",
"--save-logs",
"--disk",
f"{disk}",
"--query",
f"mkdir {path}",
]
)
def ls(source, disk, path):
return source.exec_in_container(
[
"/usr/bin/clickhouse",
"disks",
"--save-logs",
"--disk",
f"{disk}",
"--query",
f"list {path}",
]
)
def remove(source, disk, path):
return source.exec_in_container(
[
"/usr/bin/clickhouse",
"disks",
"--save-logs",
"--disk",
f"{disk}",
"--query",
f"remove {path}",
]
)
def init_data(source):
source.query("DROP TABLE IF EXISTS test_table")
source.query(
"CREATE TABLE test_table(word String, value UInt64) "
"ENGINE=MergeTree() "
"ORDER BY word SETTINGS storage_policy = 'test1'"
)
source.query("INSERT INTO test_table(*) VALUES ('test1', 2)")
def init_data_s3(source):
source.query("DROP TABLE IF EXISTS test_table_s3")
source.query(
"CREATE TABLE test_table_s3(word String, value UInt64) "
"ENGINE=MergeTree() "
"ORDER BY word SETTINGS storage_policy = 'test3'"
)
source.query("INSERT INTO test_table_s3(*) VALUES ('test1', 2)")
def init_data_s3_rm_rec(source):
# a / a / a, b, c
# / b / a, b, c, d, e
# / c /
# / d / a
mkdir(source, "test3", "a")
mkdir(source, "test3", "a/a")
mkdir(source, "test3", "a/b")
mkdir(source, "test3", "a/c")
mkdir(source, "test3", "a/d")
write(source, "test3", "a/a/a")
write(source, "test3", "a/a/b")
write(source, "test3", "a/a/c")
write(source, "test3", "a/b/a")
write(source, "test3", "a/b/b")
write(source, "test3", "a/b/c")
write(source, "test3", "a/b/d")
write(source, "test3", "a/b/e")
write(source, "test3", "d/a")
def test_disks_app_func_ld(started_cluster):
source = cluster.instances["disks_app_test"]
out = source.exec_in_container(
["/usr/bin/clickhouse", "disks", "--save-logs", "--query", "list-disks"]
)
disks = list(
sorted(
map(
lambda x: x.split(":")[0], filter(lambda x: len(x) > 1, out.split("\n"))
)
)
)
assert disks[:4] == ["default", "local", "test1", "test2"]
def test_disks_app_func_ls(started_cluster):
source = cluster.instances["disks_app_test"]
init_data(source)
out = ls(source, "test1", ".")
files = out.split("\n")
assert files[0] == "store"
out = ls(source, "test1", ". --recursive")
assert ".:\nstore\n" in out
assert "\n./store:\n" in out
def test_disks_app_func_cp(started_cluster):
source = cluster.instances["disks_app_test"]
init_data(source)
write(source, "test1", "path1")
source.exec_in_container(
[
"/usr/bin/clickhouse",
"disks",
"--query",
"copy --recursive --disk-from test1 --disk-to test2 . .",
]
)
out = ls(source, "test2", ".")
assert "path1" in out
remove(source, "test2", "path1")
remove(source, "test1", "path1")
# alesapin: Why we need list one more time?
# kssenii: it is an assertion that the file is indeed deleted
out = ls(source, "test2", ".")
assert "path1" not in out
out = ls(source, "test1", ".")
assert "path1" not in out
def test_disks_app_func_ln(started_cluster):
source = cluster.instances["disks_app_test"]
init_data(source)
source.exec_in_container(
[
"/usr/bin/clickhouse",
"disks",
"--query",
"link data/default/test_table data/default/z_tester",
]
)
out = source.exec_in_container(
["/usr/bin/clickhouse", "disks", "--save-logs", "--query", "list data/default/"]
)
files = out.split("\n")
assert "z_tester" in files
def test_disks_app_func_rm(started_cluster):
source = cluster.instances["disks_app_test"]
init_data(source)
write(source, "test2", "path3")
out = ls(source, "test2", ".")
assert "path3" in out
remove(source, "test2", "path3")
out = ls(source, "test2", ".")
assert "path3" not in out
def test_disks_app_func_rm_shared_recursive(started_cluster):
source = cluster.instances["disks_app_test"]
init_data_s3_rm_rec(source)
out = ls(source, "test3", ". --recursive")
assert (
out
== ".:\na\n\n./a:\na\nb\nc\nd\n\n./a/a:\na\nb\nc\n\n./a/b:\na\nb\nc\nd\ne\n\n./a/c:\n\n./a/d:\n\n"
)
remove(source, "test3", "a/a --recursive")
out = ls(source, "test3", ". --recursive")
assert (
out == ".:\na\n\n./a:\nb\nc\nd\n\n./a/b:\na\nb\nc\nd\ne\n\n./a/c:\n\n./a/d:\n\n"
)
remove(source, "test3", "a/b --recursive")
out = ls(source, "test3", ". --recursive")
assert out == ".:\na\n\n./a:\nc\nd\n\n./a/c:\n\n./a/d:\n\n"
remove(source, "test3", "a/c --recursive")
out = ls(source, "test3", ". --recursive")
assert out == ".:\na\n\n./a:\nd\n\n./a/d:\n\n"
remove(source, "test3", "a --recursive")
out = ls(source, "test3", ". --recursive")
assert out == ".:\n\n"
def test_disks_app_func_mv(started_cluster):
source = cluster.instances["disks_app_test"]
init_data(source)
out = ls(source, "test1", ".")
files = out.split("\n")
assert "old_store" not in files
assert "store" in files
source.exec_in_container(
[
"/usr/bin/clickhouse",
"disks",
"--disk",
"test1",
"--query",
"move store old_store",
]
)
out = ls(source, "test1", ".")
files = out.split("\n")
assert "old_store" in files
assert "store" not in files
def test_disks_app_func_read_write(started_cluster):
source = cluster.instances["disks_app_test"]
write(source, "test1", "5.txt")
out = source.exec_in_container(
[
"/usr/bin/clickhouse",
"disks",
"--save-logs",
"--disk",
"test1",
"--query",
"read 5.txt",
]
)
files = out.split("\n")
assert files[0] == "tester"
def test_remote_disk_list(started_cluster):
source = cluster.instances["disks_app_test"]
init_data_s3(source)
out = ls(source, "test3", ".")
files = out.split("\n")
assert files[0] == "store"
out = ls(source, "test3", ". --recursive")
assert ".:\nstore\n" in out
assert "\n./store:\n" in out