ClickHouse/tests/integration/test_system_logs/test_system_logs.py

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

169 lines
5.1 KiB
Python
Raw Normal View History

# pylint: disable=line-too-long
# pylint: disable=unused-argument
# pylint: disable=redefined-outer-name
import pytest
from helpers.cluster import ClickHouseCluster
cluster = ClickHouseCluster(__file__)
node1 = cluster.add_instance(
"node1",
2023-05-27 14:56:00 +00:00
base_config_dir="configs",
main_configs=["configs/config.d/system_logs_order_by.xml"],
stay_alive=True,
)
node2 = cluster.add_instance(
"node2",
2023-05-27 14:56:00 +00:00
base_config_dir="configs",
main_configs=[
"configs/config.d/system_logs_engine.xml",
"configs/config.d/disks.xml",
],
stay_alive=True,
)
node3 = cluster.add_instance(
"node3",
2023-05-27 14:56:00 +00:00
base_config_dir="configs",
main_configs=[
"configs/config.d/system_logs_settings.xml",
"configs/config.d/disks.xml",
],
stay_alive=True,
)
2023-05-24 01:50:29 +00:00
2024-04-07 06:07:12 +00:00
node4 = cluster.add_instance(
"node4",
base_config_dir="configs",
main_configs=[
"configs/config.d/system_logs_engine_s3_plain_rewritable_policy.xml",
"configs/config.d/disks.xml",
],
with_minio=True,
stay_alive=True,
)
@pytest.fixture(scope="module", autouse=True)
def start_cluster():
try:
cluster.start()
yield cluster
finally:
cluster.shutdown()
2023-05-24 01:50:29 +00:00
def test_system_logs_order_by_expr(start_cluster):
node1.query("SET log_query_threads = 1")
node1.query("SELECT count() FROM system.tables")
node1.query("SYSTEM FLUSH LOGS")
# Check 'sorting_key' of system.query_log.
2023-05-24 01:50:29 +00:00
assert (
node1.query(
"SELECT sorting_key FROM system.tables WHERE database='system' and name='query_log'"
)
== "event_date, event_time, initial_query_id\n"
2023-05-24 01:50:29 +00:00
)
# Check 'sorting_key' of system.query_thread_log.
2023-05-24 01:50:29 +00:00
assert (
node1.query(
"SELECT sorting_key FROM system.tables WHERE database='system' and name='query_thread_log'"
)
== "event_date, event_time, query_id\n"
2023-05-24 01:50:29 +00:00
)
def test_system_logs_engine_expr(start_cluster):
node2.query("SET log_query_threads = 1")
node2.query("SELECT count() FROM system.tables")
node2.query("SYSTEM FLUSH LOGS")
# Check 'engine_full' of system.query_log.
expected = "MergeTree PARTITION BY event_date ORDER BY event_time TTL event_date + toIntervalDay(30) SETTINGS storage_policy = \\'policy2\\', ttl_only_drop_parts = 1"
assert expected in node2.query(
2023-05-24 01:50:29 +00:00
"SELECT engine_full FROM system.tables WHERE database='system' and name='query_log'"
2024-04-07 06:07:12 +00:00
)
def test_system_logs_engine_s3_plain_rw_expr(start_cluster):
node4.query("SET log_query_threads = 1")
node4.query("SELECT count() FROM system.tables")
node4.query("SYSTEM FLUSH LOGS")
# Check 'engine_full' of system.query_log.
expected = "MergeTree PARTITION BY event_date ORDER BY event_time TTL event_date + toIntervalDay(30) SETTINGS storage_policy = \\'s3_plain_rewritable\\', ttl_only_drop_parts = 1"
assert expected in node4.query(
"SELECT engine_full FROM system.tables WHERE database='system' and name='query_log'"
)
node4.restart_clickhouse()
assert expected in node4.query(
"SELECT engine_full FROM system.tables WHERE database='system' and name='query_log'"
2023-05-24 01:50:29 +00:00
)
def test_system_logs_settings_expr(start_cluster):
node3.query("SET log_query_threads = 1")
node3.query("SELECT count() FROM system.tables")
node3.query("SYSTEM FLUSH LOGS")
# Check 'engine_full' of system.query_log.
expected = "MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time, initial_query_id) TTL event_date + toIntervalDay(30) SETTINGS storage_policy = \\'policy1\\', storage_policy = \\'policy2\\', ttl_only_drop_parts = 1"
assert expected in node3.query(
2023-05-24 01:50:29 +00:00
"SELECT engine_full FROM system.tables WHERE database='system' and name='query_log'"
)
2023-07-28 07:23:34 +00:00
def test_max_size_0(start_cluster):
node1.exec_in_container(
[
"bash",
"-c",
f"""echo "
<clickhouse>
<query_log>
<max_size_rows replace=\\"replace\\">0</max_size_rows>
<reserved_size_rows replace=\\"replace\\">0</reserved_size_rows>
</query_log>
</clickhouse>
" > /etc/clickhouse-server/config.d/yyy-override-query_log.xml
""",
]
)
with pytest.raises(Exception):
node1.restart_clickhouse()
node1.exec_in_container(
["rm", f"/etc/clickhouse-server/config.d/yyy-override-query_log.xml"]
)
node1.restart_clickhouse()
def test_reserved_size_greater_max_size(start_cluster):
node1.exec_in_container(
[
"bash",
"-c",
f"""echo "
<clickhouse>
<query_log>
<max_size_rows replace=\\"replace\\">10</max_size_rows>
<reserved_size_rows replace=\\"replace\\">11</reserved_size_rows>
</query_log>
</clickhouse>
" > /etc/clickhouse-server/config.d/yyy-override-query_log.xml
""",
]
)
with pytest.raises(Exception):
node1.restart_clickhouse()
node1.exec_in_container(
["rm", f"/etc/clickhouse-server/config.d/yyy-override-query_log.xml"]
)
node1.restart_clickhouse()