ClickHouse/tests/integration/test_attach_backup_from_s3_plain/test.py

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

109 lines
3.0 KiB
Python
Raw Normal View History

# pylint: disable=global-statement
# pylint: disable=line-too-long
import pytest
2024-09-27 10:19:39 +00:00
from helpers.cluster import ClickHouseCluster
cluster = ClickHouseCluster(__file__)
node = cluster.add_instance(
"node",
main_configs=["configs/disk_s3.xml"],
with_minio=True,
)
@pytest.fixture(scope="module", autouse=True)
def start_cluster():
try:
cluster.start()
yield
finally:
cluster.shutdown()
2024-02-26 09:37:17 +00:00
s3_disk_def = """disk(type=s3_plain,
endpoint='http://minio1:9001/root/data/disks/disk_s3_plain/{}/',
access_key_id='minio',
secret_access_key='minio123');"""
2024-02-27 10:07:19 +00:00
local_disk_def = "disk(type=object_storage, object_storage_type = 'local_blob_storage', metadata_type = 'plain', path = '/local_plain/{}/');"
@pytest.mark.parametrize(
"table_name,backup_name,storage_policy,disk_def,min_bytes_for_wide_part",
[
pytest.param(
2024-02-26 09:37:17 +00:00
"compact",
"backup_compact_s3",
"backup_disk_s3_plain",
s3_disk_def,
int(1e9),
id="compact",
),
pytest.param(
"wide",
"backup_wide_s3",
"backup_disk_s3_plain",
s3_disk_def,
int(0),
id="wide",
),
pytest.param(
2024-02-26 09:37:17 +00:00
"compact",
"backup_compact_local",
"backup_disk_local_plain",
local_disk_def,
int(1e9),
id="compact",
),
pytest.param(
2024-02-26 09:37:17 +00:00
"wide",
"backup_wide_local",
"backup_disk_local_plain",
local_disk_def,
int(0),
id="wide",
),
],
)
2024-02-26 09:37:17 +00:00
def test_attach_part(
table_name, backup_name, storage_policy, disk_def, min_bytes_for_wide_part
):
disk_definition = disk_def.format(backup_name)
node.query(
f"""
-- Catch any errors (NOTE: warnings are ok)
set send_logs_level='error';
-- BACKUP writes Ordinary like structure
-- but what is more important that you cannot ATTACH directly from Atomic
-- so Memory engine will work (to avoid using allow_deprecated_database_ordinary).
create database ordinary_db engine=Memory;
create table ordinary_db.{table_name} engine=MergeTree() order by key partition by part
settings min_bytes_for_wide_part={min_bytes_for_wide_part}
as select number%5 part, number key from numbers(100);
backup table ordinary_db.{table_name} TO Disk('{storage_policy}', '{backup_name}') settings deduplicate_files=0;
drop table ordinary_db.{table_name};
attach table ordinary_db.{table_name} (part UInt8, key UInt64)
engine=MergeTree()
order by key partition by part
settings
max_suspicious_broken_parts=0,
disk={disk_definition}
"""
)
assert int(node.query(f"select count() from ordinary_db.{table_name}")) == 100
node.query(
f"""
-- NOTE: DROP DATABASE cannot be done w/o this due to metadata leftovers
set force_remove_data_recursively_on_drop=1;
drop database ordinary_db sync;
"""
)