mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-25 09:02:00 +00:00
180 lines
5.6 KiB
Python
180 lines
5.6 KiB
Python
import time
|
|
from http import HTTPStatus
|
|
|
|
import pytest
|
|
import requests
|
|
|
|
from helpers.cluster import ClickHouseCluster
|
|
|
|
cluster = ClickHouseCluster(__file__)
|
|
|
|
node = cluster.add_instance(
|
|
"node",
|
|
main_configs=["configs/prometheus.xml"],
|
|
user_configs=["configs/allow_experimental_time_series_table.xml"],
|
|
with_prometheus=True,
|
|
handle_prometheus_remote_write=True,
|
|
handle_prometheus_remote_read=True,
|
|
)
|
|
|
|
|
|
def execute_query_on_prometheus_writer(query, timestamp):
|
|
return execute_query_impl(
|
|
cluster.get_instance_ip(cluster.prometheus_writer_host),
|
|
cluster.prometheus_writer_port,
|
|
"/api/v1/query",
|
|
query,
|
|
timestamp,
|
|
)
|
|
|
|
|
|
def execute_query_on_prometheus_reader(query, timestamp):
|
|
return execute_query_impl(
|
|
cluster.get_instance_ip(cluster.prometheus_reader_host),
|
|
cluster.prometheus_reader_port,
|
|
"/api/v1/query",
|
|
query,
|
|
timestamp,
|
|
)
|
|
|
|
|
|
def execute_query_impl(host, port, path, query, timestamp):
|
|
if not path.startswith("/"):
|
|
path += "/"
|
|
url = f"http://{host}:{port}/{path.strip('/')}?query={query}&time={timestamp}"
|
|
print(f"Requesting {url}")
|
|
r = requests.get(url)
|
|
print(f"Status code: {r.status_code} {HTTPStatus(r.status_code).phrase}")
|
|
if r.status_code != requests.codes.ok:
|
|
print(f"Response: {r.text}")
|
|
raise Exception(f"Got unexpected status code {r.status_code}")
|
|
return r.json()
|
|
|
|
|
|
def show_query_result(query):
|
|
evaluation_time = time.time()
|
|
print(f"Evaluating query: {query}")
|
|
print(f"Evaluation time: {evaluation_time}")
|
|
result_from_writer = execute_query_on_prometheus_writer(query, evaluation_time)
|
|
print(f"Result from prometheus_writer: {result_from_writer}")
|
|
result_from_reader = execute_query_on_prometheus_reader(query, evaluation_time)
|
|
print(f"Result from prometheus_reader: {result_from_reader}")
|
|
|
|
|
|
def compare_query(query):
|
|
timeout = 60
|
|
start_time = time.time()
|
|
evaluation_time = start_time
|
|
print(f"Evaluating query: {query}")
|
|
print(f"Evaluation time: {evaluation_time}")
|
|
while time.time() < start_time + timeout:
|
|
result_from_writer = execute_query_on_prometheus_writer(query, evaluation_time)
|
|
time.sleep(1)
|
|
result_from_reader = execute_query_on_prometheus_reader(query, evaluation_time)
|
|
print(f"Result from prometheus_writer: {result_from_writer}")
|
|
print(f"Result from prometheus_reader: {result_from_reader}")
|
|
if result_from_writer == result_from_reader:
|
|
return
|
|
raise Exception(
|
|
f"Got different results from prometheus_writer and prometheus_reader"
|
|
)
|
|
|
|
|
|
def compare_queries():
|
|
compare_query("up")
|
|
|
|
|
|
@pytest.fixture(scope="module", autouse=True)
|
|
def start_cluster():
|
|
try:
|
|
cluster.start()
|
|
yield cluster
|
|
finally:
|
|
cluster.shutdown()
|
|
|
|
|
|
@pytest.fixture(autouse=True)
|
|
def cleanup_after_test():
|
|
try:
|
|
yield
|
|
finally:
|
|
node.query("DROP TABLE IF EXISTS prometheus SYNC")
|
|
node.query("DROP TABLE IF EXISTS original SYNC")
|
|
node.query("DROP TABLE IF EXISTS mydata SYNC")
|
|
node.query("DROP TABLE IF EXISTS mytable SYNC")
|
|
node.query("DROP TABLE IF EXISTS mymetrics SYNC")
|
|
|
|
|
|
def test_default():
|
|
node.query("CREATE TABLE prometheus ENGINE=TimeSeries")
|
|
compare_queries()
|
|
|
|
|
|
def test_tags_to_columns():
|
|
node.query(
|
|
"CREATE TABLE prometheus ENGINE=TimeSeries SETTINGS tags_to_columns = {'job': 'job', 'instance': 'instance'}"
|
|
)
|
|
compare_queries()
|
|
|
|
|
|
def test_64bit_id():
|
|
node.query("CREATE TABLE prometheus (id UInt64) ENGINE=TimeSeries")
|
|
compare_queries()
|
|
|
|
|
|
def test_custom_id_algorithm():
|
|
node.query(
|
|
"CREATE TABLE prometheus (id FixedString(16) DEFAULT murmurHash3_128(metric_name, all_tags)) ENGINE=TimeSeries"
|
|
)
|
|
compare_queries()
|
|
|
|
|
|
def test_create_as_table():
|
|
node.query("CREATE TABLE original ENGINE=TimeSeries")
|
|
node.query("CREATE TABLE prometheus AS original")
|
|
compare_queries()
|
|
|
|
|
|
def test_inner_engines():
|
|
node.query(
|
|
"CREATE TABLE prometheus ENGINE=TimeSeries "
|
|
"DATA ENGINE=MergeTree ORDER BY (id, timestamp) "
|
|
"TAGS ENGINE=AggregatingMergeTree ORDER BY (metric_name, id) "
|
|
"METRICS ENGINE=ReplacingMergeTree ORDER BY metric_family_name"
|
|
)
|
|
compare_queries()
|
|
|
|
|
|
def test_external_tables():
|
|
node.query("DROP TABLE IF EXISTS mydata")
|
|
node.query("DROP TABLE IF EXISTS mytags")
|
|
node.query("DROP TABLE IF EXISTS mymetrics")
|
|
node.query("DROP TABLE IF EXISTS prometheus")
|
|
|
|
node.query(
|
|
"CREATE TABLE mydata (id UUID, timestamp DateTime64(3), value Float64) "
|
|
"ENGINE=MergeTree ORDER BY (id, timestamp)"
|
|
)
|
|
node.query(
|
|
"CREATE TABLE mytags ("
|
|
"id UUID, "
|
|
"metric_name LowCardinality(String), "
|
|
"tags Map(LowCardinality(String), String), "
|
|
"min_time SimpleAggregateFunction(min, Nullable(DateTime64(3))), "
|
|
"max_time SimpleAggregateFunction(max, Nullable(DateTime64(3)))) "
|
|
"ENGINE=AggregatingMergeTree ORDER BY (metric_name, id)"
|
|
)
|
|
|
|
# FIXME: The table structure should be:
|
|
# "CREATE TABLE mymetrics (metric_family_name String, type LowCardinality(String), unit LowCardinality(String), help String)"
|
|
# Renamed it because of the bug and potential type mismatch.
|
|
node.query(
|
|
"CREATE TABLE mymetrics (metric_family_name String, type String, unit String, help String) "
|
|
"ENGINE=ReplacingMergeTree ORDER BY metric_family_name"
|
|
)
|
|
node.query(
|
|
"CREATE TABLE prometheus ENGINE=TimeSeries "
|
|
"DATA mydata TAGS mytags METRICS mymetrics"
|
|
)
|
|
compare_queries()
|