mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-23 16:12:01 +00:00
217 lines
5.9 KiB
Python
217 lines
5.9 KiB
Python
import time
|
|
|
|
import pytest
|
|
from helpers.cluster import ClickHouseCluster
|
|
from helpers.test_tools import assert_eq_with_retry
|
|
from helpers.network import PartitionManager
|
|
|
|
|
|
def fill_nodes(nodes, shard):
|
|
for node in nodes:
|
|
node.query(
|
|
"""
|
|
CREATE DATABASE test;
|
|
|
|
CREATE TABLE test.test_table(date Date, id UInt32)
|
|
ENGINE = ReplicatedMergeTree('/clickhouse/tables/test{shard}/replicated', '{replica}') ORDER BY id PARTITION BY toYYYYMM(date)
|
|
SETTINGS min_replicated_logs_to_keep=3, max_replicated_logs_to_keep=5,
|
|
cleanup_delay_period=0, cleanup_delay_period_random_add=0, cleanup_thread_preferred_points_per_iteration=0;
|
|
""".format(
|
|
shard=shard, replica=node.name
|
|
)
|
|
)
|
|
|
|
|
|
cluster = ClickHouseCluster(__file__)
|
|
node1 = cluster.add_instance(
|
|
"node1", main_configs=["configs/remote_servers.xml"], with_zookeeper=True
|
|
)
|
|
node2 = cluster.add_instance(
|
|
"node2", main_configs=["configs/remote_servers.xml"], with_zookeeper=True
|
|
)
|
|
|
|
|
|
@pytest.fixture(scope="module")
|
|
def start_cluster():
|
|
try:
|
|
cluster.start()
|
|
|
|
fill_nodes([node1, node2], 1)
|
|
|
|
yield cluster
|
|
|
|
except Exception as ex:
|
|
print(ex)
|
|
|
|
finally:
|
|
cluster.shutdown()
|
|
|
|
|
|
def test_readonly_metrics(start_cluster):
|
|
assert (
|
|
node1.query("SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'")
|
|
== "0\n"
|
|
)
|
|
|
|
with PartitionManager() as pm:
|
|
## make node1 readonly -> heal -> readonly -> heal -> detach table -> heal -> attach table
|
|
pm.drop_instance_zk_connections(node1)
|
|
assert_eq_with_retry(
|
|
node1,
|
|
"SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'",
|
|
"1\n",
|
|
retry_count=300,
|
|
sleep_time=1,
|
|
)
|
|
|
|
pm.heal_all()
|
|
assert_eq_with_retry(
|
|
node1,
|
|
"SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'",
|
|
"0\n",
|
|
retry_count=300,
|
|
sleep_time=1,
|
|
)
|
|
|
|
pm.drop_instance_zk_connections(node1)
|
|
assert_eq_with_retry(
|
|
node1,
|
|
"SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'",
|
|
"1\n",
|
|
retry_count=300,
|
|
sleep_time=1,
|
|
)
|
|
|
|
node1.query("DETACH TABLE test.test_table")
|
|
assert "0\n" == node1.query(
|
|
"SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'"
|
|
)
|
|
|
|
pm.heal_all()
|
|
node1.query("ATTACH TABLE test.test_table")
|
|
assert_eq_with_retry(
|
|
node1,
|
|
"SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'",
|
|
"0\n",
|
|
retry_count=300,
|
|
sleep_time=1,
|
|
)
|
|
|
|
|
|
# For LowCardinality-columns, the bytes for N rows is not N*size of 1 row.
|
|
def test_metrics_storage_buffer_size(start_cluster):
|
|
node1.query(
|
|
"""
|
|
CREATE TABLE test.test_mem_table
|
|
(
|
|
`str` LowCardinality(String)
|
|
)
|
|
ENGINE = Memory;
|
|
|
|
CREATE TABLE test.buffer_table
|
|
(
|
|
`str` LowCardinality(String)
|
|
)
|
|
ENGINE = Buffer('test', 'test_mem_table', 1, 600, 600, 1000, 100000, 100000, 10000000);
|
|
"""
|
|
)
|
|
|
|
# before flush
|
|
node1.query("INSERT INTO test.buffer_table VALUES('hello');")
|
|
assert (
|
|
node1.query(
|
|
"SELECT value FROM system.metrics WHERE metric = 'StorageBufferRows'"
|
|
)
|
|
== "1\n"
|
|
)
|
|
# By the way, this metric does not count the LowCardinality's dictionary size.
|
|
assert (
|
|
node1.query(
|
|
"SELECT value FROM system.metrics WHERE metric = 'StorageBufferBytes'"
|
|
)
|
|
== "1\n"
|
|
)
|
|
|
|
node1.query("INSERT INTO test.buffer_table VALUES('hello');")
|
|
assert (
|
|
node1.query(
|
|
"SELECT value FROM system.metrics WHERE metric = 'StorageBufferRows'"
|
|
)
|
|
== "2\n"
|
|
)
|
|
assert (
|
|
node1.query(
|
|
"SELECT value FROM system.metrics WHERE metric = 'StorageBufferBytes'"
|
|
)
|
|
== "2\n"
|
|
)
|
|
|
|
# flush
|
|
node1.query("OPTIMIZE TABLE test.buffer_table")
|
|
assert (
|
|
node1.query(
|
|
"SELECT value FROM system.metrics WHERE metric = 'StorageBufferRows'"
|
|
)
|
|
== "0\n"
|
|
)
|
|
assert (
|
|
node1.query(
|
|
"SELECT value FROM system.metrics WHERE metric = 'StorageBufferBytes'"
|
|
)
|
|
== "0\n"
|
|
)
|
|
|
|
|
|
def test_attach_without_zk_incr_readonly_metric(start_cluster):
|
|
assert (
|
|
node1.query("SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'")
|
|
== "0\n"
|
|
)
|
|
|
|
node1.query(
|
|
"ATTACH TABLE test.test_no_zk UUID 'a50b7933-59b2-49ce-8db6-59da3c9b4413' (i Int8, d Date) ENGINE = ReplicatedMergeTree('no_zk', 'replica') ORDER BY tuple()"
|
|
)
|
|
assert_eq_with_retry(
|
|
node1,
|
|
"SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'",
|
|
"1\n",
|
|
retry_count=300,
|
|
sleep_time=1,
|
|
)
|
|
|
|
node1.query("DETACH TABLE test.test_no_zk")
|
|
assert_eq_with_retry(
|
|
node1,
|
|
"SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'",
|
|
"0\n",
|
|
retry_count=300,
|
|
sleep_time=1,
|
|
)
|
|
|
|
node1.query("ATTACH TABLE test.test_no_zk")
|
|
assert_eq_with_retry(
|
|
node1,
|
|
"SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'",
|
|
"1\n",
|
|
retry_count=300,
|
|
sleep_time=1,
|
|
)
|
|
|
|
node1.query("SYSTEM RESTORE REPLICA test.test_no_zk")
|
|
assert_eq_with_retry(
|
|
node1,
|
|
"SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'",
|
|
"0\n",
|
|
retry_count=300,
|
|
sleep_time=1,
|
|
)
|
|
|
|
node1.query("DROP TABLE test.test_no_zk")
|
|
assert_eq_with_retry(
|
|
node1,
|
|
"SELECT value FROM system.metrics WHERE metric = 'ReadonlyReplica'",
|
|
"0\n",
|
|
retry_count=300,
|
|
sleep_time=1,
|
|
)
|