mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-22 07:31:57 +00:00
apply black formatter
This commit is contained in:
parent
21f5d20b9e
commit
0ee8dfad53
@ -10,31 +10,38 @@ import requests
|
|||||||
import tempfile
|
import tempfile
|
||||||
|
|
||||||
|
|
||||||
DEFAULT_URL = 'https://clickhouse-datasets.s3.amazonaws.com'
|
DEFAULT_URL = "https://clickhouse-datasets.s3.amazonaws.com"
|
||||||
|
|
||||||
AVAILABLE_DATASETS = {
|
AVAILABLE_DATASETS = {
|
||||||
'hits': 'hits_v1.tar',
|
"hits": "hits_v1.tar",
|
||||||
'visits': 'visits_v1.tar',
|
"visits": "visits_v1.tar",
|
||||||
}
|
}
|
||||||
|
|
||||||
RETRIES_COUNT = 5
|
RETRIES_COUNT = 5
|
||||||
|
|
||||||
|
|
||||||
def _get_temp_file_name():
|
def _get_temp_file_name():
|
||||||
return os.path.join(tempfile._get_default_tempdir(), next(tempfile._get_candidate_names()))
|
return os.path.join(
|
||||||
|
tempfile._get_default_tempdir(), next(tempfile._get_candidate_names())
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def build_url(base_url, dataset):
|
def build_url(base_url, dataset):
|
||||||
return os.path.join(base_url, dataset, 'partitions', AVAILABLE_DATASETS[dataset])
|
return os.path.join(base_url, dataset, "partitions", AVAILABLE_DATASETS[dataset])
|
||||||
|
|
||||||
|
|
||||||
def dowload_with_progress(url, path):
|
def dowload_with_progress(url, path):
|
||||||
logging.info("Downloading from %s to temp path %s", url, path)
|
logging.info("Downloading from %s to temp path %s", url, path)
|
||||||
for i in range(RETRIES_COUNT):
|
for i in range(RETRIES_COUNT):
|
||||||
try:
|
try:
|
||||||
with open(path, 'wb') as f:
|
with open(path, "wb") as f:
|
||||||
response = requests.get(url, stream=True)
|
response = requests.get(url, stream=True)
|
||||||
response.raise_for_status()
|
response.raise_for_status()
|
||||||
total_length = response.headers.get('content-length')
|
total_length = response.headers.get("content-length")
|
||||||
if total_length is None or int(total_length) == 0:
|
if total_length is None or int(total_length) == 0:
|
||||||
logging.info("No content-length, will download file without progress")
|
logging.info(
|
||||||
|
"No content-length, will download file without progress"
|
||||||
|
)
|
||||||
f.write(response.content)
|
f.write(response.content)
|
||||||
else:
|
else:
|
||||||
dl = 0
|
dl = 0
|
||||||
@ -46,7 +53,11 @@ def dowload_with_progress(url, path):
|
|||||||
if sys.stdout.isatty():
|
if sys.stdout.isatty():
|
||||||
done = int(50 * dl / total_length)
|
done = int(50 * dl / total_length)
|
||||||
percent = int(100 * float(dl) / total_length)
|
percent = int(100 * float(dl) / total_length)
|
||||||
sys.stdout.write("\r[{}{}] {}%".format('=' * done, ' ' * (50-done), percent))
|
sys.stdout.write(
|
||||||
|
"\r[{}{}] {}%".format(
|
||||||
|
"=" * done, " " * (50 - done), percent
|
||||||
|
)
|
||||||
|
)
|
||||||
sys.stdout.flush()
|
sys.stdout.flush()
|
||||||
break
|
break
|
||||||
except Exception as ex:
|
except Exception as ex:
|
||||||
@ -56,14 +67,21 @@ def dowload_with_progress(url, path):
|
|||||||
if os.path.exists(path):
|
if os.path.exists(path):
|
||||||
os.remove(path)
|
os.remove(path)
|
||||||
else:
|
else:
|
||||||
raise Exception("Cannot download dataset from {}, all retries exceeded".format(url))
|
raise Exception(
|
||||||
|
"Cannot download dataset from {}, all retries exceeded".format(url)
|
||||||
|
)
|
||||||
|
|
||||||
sys.stdout.write("\n")
|
sys.stdout.write("\n")
|
||||||
logging.info("Downloading finished")
|
logging.info("Downloading finished")
|
||||||
|
|
||||||
|
|
||||||
def unpack_to_clickhouse_directory(tar_path, clickhouse_path):
|
def unpack_to_clickhouse_directory(tar_path, clickhouse_path):
|
||||||
logging.info("Will unpack data from temp path %s to clickhouse db %s", tar_path, clickhouse_path)
|
logging.info(
|
||||||
with tarfile.open(tar_path, 'r') as comp_file:
|
"Will unpack data from temp path %s to clickhouse db %s",
|
||||||
|
tar_path,
|
||||||
|
clickhouse_path,
|
||||||
|
)
|
||||||
|
with tarfile.open(tar_path, "r") as comp_file:
|
||||||
comp_file.extractall(path=clickhouse_path)
|
comp_file.extractall(path=clickhouse_path)
|
||||||
logging.info("Unpack finished")
|
logging.info("Unpack finished")
|
||||||
|
|
||||||
@ -72,15 +90,21 @@ if __name__ == "__main__":
|
|||||||
logging.basicConfig(level=logging.INFO)
|
logging.basicConfig(level=logging.INFO)
|
||||||
|
|
||||||
parser = argparse.ArgumentParser(
|
parser = argparse.ArgumentParser(
|
||||||
description="Simple tool for dowloading datasets for clickhouse from S3")
|
description="Simple tool for dowloading datasets for clickhouse from S3"
|
||||||
|
)
|
||||||
|
|
||||||
parser.add_argument('--dataset-names', required=True, nargs='+', choices=list(AVAILABLE_DATASETS.keys()))
|
parser.add_argument(
|
||||||
parser.add_argument('--url-prefix', default=DEFAULT_URL)
|
"--dataset-names",
|
||||||
parser.add_argument('--clickhouse-data-path', default='/var/lib/clickhouse/')
|
required=True,
|
||||||
|
nargs="+",
|
||||||
|
choices=list(AVAILABLE_DATASETS.keys()),
|
||||||
|
)
|
||||||
|
parser.add_argument("--url-prefix", default=DEFAULT_URL)
|
||||||
|
parser.add_argument("--clickhouse-data-path", default="/var/lib/clickhouse/")
|
||||||
|
|
||||||
args = parser.parse_args()
|
args = parser.parse_args()
|
||||||
datasets = args.dataset_names
|
datasets = args.dataset_names
|
||||||
logging.info("Will fetch following datasets: %s", ', '.join(datasets))
|
logging.info("Will fetch following datasets: %s", ", ".join(datasets))
|
||||||
for dataset in datasets:
|
for dataset in datasets:
|
||||||
logging.info("Processing %s", dataset)
|
logging.info("Processing %s", dataset)
|
||||||
temp_archive_path = _get_temp_file_name()
|
temp_archive_path = _get_temp_file_name()
|
||||||
@ -92,10 +116,11 @@ if __name__ == "__main__":
|
|||||||
logging.info("Some exception occured %s", str(ex))
|
logging.info("Some exception occured %s", str(ex))
|
||||||
raise
|
raise
|
||||||
finally:
|
finally:
|
||||||
logging.info("Will remove downloaded file %s from filesystem if it exists", temp_archive_path)
|
logging.info(
|
||||||
|
"Will remove downloaded file %s from filesystem if it exists",
|
||||||
|
temp_archive_path,
|
||||||
|
)
|
||||||
if os.path.exists(temp_archive_path):
|
if os.path.exists(temp_archive_path):
|
||||||
os.remove(temp_archive_path)
|
os.remove(temp_archive_path)
|
||||||
logging.info("Processing of %s finished", dataset)
|
logging.info("Processing of %s finished", dataset)
|
||||||
logging.info("Fetch finished, enjoy your tables!")
|
logging.info("Fetch finished, enjoy your tables!")
|
||||||
|
|
||||||
|
|
||||||
|
@ -95,7 +95,13 @@ class HTTPError(Exception):
|
|||||||
|
|
||||||
# Helpers to execute queries via HTTP interface.
|
# Helpers to execute queries via HTTP interface.
|
||||||
def clickhouse_execute_http(
|
def clickhouse_execute_http(
|
||||||
base_args, query, timeout=30, settings=None, default_format=None, max_http_retries=5, retry_error_codes=False
|
base_args,
|
||||||
|
query,
|
||||||
|
timeout=30,
|
||||||
|
settings=None,
|
||||||
|
default_format=None,
|
||||||
|
max_http_retries=5,
|
||||||
|
retry_error_codes=False,
|
||||||
):
|
):
|
||||||
if args.secure:
|
if args.secure:
|
||||||
client = http.client.HTTPSConnection(
|
client = http.client.HTTPSConnection(
|
||||||
@ -146,12 +152,36 @@ def clickhouse_execute_http(
|
|||||||
|
|
||||||
return data
|
return data
|
||||||
|
|
||||||
def clickhouse_execute(base_args, query, timeout=30, settings=None, max_http_retries=5, retry_error_codes=False):
|
|
||||||
return clickhouse_execute_http(base_args, query, timeout, settings, max_http_retries=max_http_retries, retry_error_codes=retry_error_codes).strip()
|
def clickhouse_execute(
|
||||||
|
base_args,
|
||||||
|
query,
|
||||||
|
timeout=30,
|
||||||
|
settings=None,
|
||||||
|
max_http_retries=5,
|
||||||
|
retry_error_codes=False,
|
||||||
|
):
|
||||||
|
return clickhouse_execute_http(
|
||||||
|
base_args,
|
||||||
|
query,
|
||||||
|
timeout,
|
||||||
|
settings,
|
||||||
|
max_http_retries=max_http_retries,
|
||||||
|
retry_error_codes=retry_error_codes,
|
||||||
|
).strip()
|
||||||
|
|
||||||
|
|
||||||
def clickhouse_execute_json(base_args, query, timeout=60, settings=None, max_http_retries=5):
|
def clickhouse_execute_json(
|
||||||
data = clickhouse_execute_http(base_args, query, timeout, settings, "JSONEachRow", max_http_retries=max_http_retries)
|
base_args, query, timeout=60, settings=None, max_http_retries=5
|
||||||
|
):
|
||||||
|
data = clickhouse_execute_http(
|
||||||
|
base_args,
|
||||||
|
query,
|
||||||
|
timeout,
|
||||||
|
settings,
|
||||||
|
"JSONEachRow",
|
||||||
|
max_http_retries=max_http_retries,
|
||||||
|
)
|
||||||
if not data:
|
if not data:
|
||||||
return None
|
return None
|
||||||
rows = []
|
rows = []
|
||||||
@ -648,7 +678,9 @@ class TestCase:
|
|||||||
|
|
||||||
clickhouse_execute(
|
clickhouse_execute(
|
||||||
args,
|
args,
|
||||||
"CREATE DATABASE IF NOT EXISTS " + database + get_db_engine(testcase_args, database),
|
"CREATE DATABASE IF NOT EXISTS "
|
||||||
|
+ database
|
||||||
|
+ get_db_engine(testcase_args, database),
|
||||||
settings=get_create_database_settings(args, testcase_args),
|
settings=get_create_database_settings(args, testcase_args),
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -831,7 +863,8 @@ class TestCase:
|
|||||||
|
|
||||||
# TODO: remove checking "no-upgrade-check" after 23.1
|
# TODO: remove checking "no-upgrade-check" after 23.1
|
||||||
elif args.upgrade_check and (
|
elif args.upgrade_check and (
|
||||||
"no-upgrade-check" in tags or "no-upgrade-check" in tags):
|
"no-upgrade-check" in tags or "no-upgrade-check" in tags
|
||||||
|
):
|
||||||
return FailureReason.NO_UPGRADE_CHECK
|
return FailureReason.NO_UPGRADE_CHECK
|
||||||
|
|
||||||
elif tags and ("no-s3-storage" in tags) and args.s3_storage:
|
elif tags and ("no-s3-storage" in tags) and args.s3_storage:
|
||||||
@ -1051,7 +1084,11 @@ class TestCase:
|
|||||||
@staticmethod
|
@staticmethod
|
||||||
def send_test_name_failed(suite: str, case: str):
|
def send_test_name_failed(suite: str, case: str):
|
||||||
pid = os.getpid()
|
pid = os.getpid()
|
||||||
clickhouse_execute(args, f"SELECT 'Running test {suite}/{case} from pid={pid}'", retry_error_codes=True)
|
clickhouse_execute(
|
||||||
|
args,
|
||||||
|
f"SELECT 'Running test {suite}/{case} from pid={pid}'",
|
||||||
|
retry_error_codes=True,
|
||||||
|
)
|
||||||
|
|
||||||
def run_single_test(
|
def run_single_test(
|
||||||
self, server_logs_level, client_options
|
self, server_logs_level, client_options
|
||||||
@ -2220,6 +2257,7 @@ def find_binary(name):
|
|||||||
|
|
||||||
raise Exception(f"{name} was not found in PATH")
|
raise Exception(f"{name} was not found in PATH")
|
||||||
|
|
||||||
|
|
||||||
def find_clickhouse_command(binary, command):
|
def find_clickhouse_command(binary, command):
|
||||||
symlink = binary + "-" + command
|
symlink = binary + "-" + command
|
||||||
if os.access(symlink, os.X_OK):
|
if os.access(symlink, os.X_OK):
|
||||||
@ -2228,6 +2266,7 @@ def find_clickhouse_command(binary, command):
|
|||||||
# To avoid requiring symlinks (in case you download binary from CI)
|
# To avoid requiring symlinks (in case you download binary from CI)
|
||||||
return binary + " " + command
|
return binary + " " + command
|
||||||
|
|
||||||
|
|
||||||
def get_additional_client_options(args):
|
def get_additional_client_options(args):
|
||||||
if args.client_option:
|
if args.client_option:
|
||||||
return " ".join("--" + option for option in args.client_option)
|
return " ".join("--" + option for option in args.client_option)
|
||||||
@ -2569,7 +2608,9 @@ if __name__ == "__main__":
|
|||||||
"WARNING: --extract_from_config option is deprecated and will be removed the the future",
|
"WARNING: --extract_from_config option is deprecated and will be removed the the future",
|
||||||
file=sys.stderr,
|
file=sys.stderr,
|
||||||
)
|
)
|
||||||
args.extract_from_config = find_clickhouse_command(args.binary, "extract-from-config")
|
args.extract_from_config = find_clickhouse_command(
|
||||||
|
args.binary, "extract-from-config"
|
||||||
|
)
|
||||||
|
|
||||||
if args.configclient:
|
if args.configclient:
|
||||||
args.client += " --config-file=" + args.configclient
|
args.client += " --config-file=" + args.configclient
|
||||||
|
@ -243,11 +243,18 @@ if __name__ == "__main__":
|
|||||||
)
|
)
|
||||||
|
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--no-random", action="store", dest="no_random", help="Disable tests order randomization"
|
"--no-random",
|
||||||
|
action="store",
|
||||||
|
dest="no_random",
|
||||||
|
help="Disable tests order randomization",
|
||||||
)
|
)
|
||||||
|
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--pre-pull", action="store_true", default=False, dest="pre_pull", help="Pull images for docker_compose before all other actions"
|
"--pre-pull",
|
||||||
|
action="store_true",
|
||||||
|
default=False,
|
||||||
|
dest="pre_pull",
|
||||||
|
help="Pull images for docker_compose before all other actions",
|
||||||
)
|
)
|
||||||
|
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
@ -306,7 +313,6 @@ if __name__ == "__main__":
|
|||||||
# if not args.no_random:
|
# if not args.no_random:
|
||||||
# rand_args += f"--random-seed={os.getpid()}"
|
# rand_args += f"--random-seed={os.getpid()}"
|
||||||
|
|
||||||
|
|
||||||
net = ""
|
net = ""
|
||||||
if args.network:
|
if args.network:
|
||||||
net = "--net={}".format(args.network)
|
net = "--net={}".format(args.network)
|
||||||
@ -417,7 +423,10 @@ if __name__ == "__main__":
|
|||||||
)
|
)
|
||||||
|
|
||||||
cmd = cmd_base + " " + args.command
|
cmd = cmd_base + " " + args.command
|
||||||
cmd_pre_pull = cmd_base + " find /compose -name docker_compose_*.yml -exec docker-compose -f '{}' pull \;"
|
cmd_pre_pull = (
|
||||||
|
cmd_base
|
||||||
|
+ " find /compose -name docker_compose_*.yml -exec docker-compose -f '{}' pull \;"
|
||||||
|
)
|
||||||
|
|
||||||
containers = subprocess.check_output(
|
containers = subprocess.check_output(
|
||||||
f"docker ps --all --quiet --filter name={CONTAINER_NAME} --format={{{{.ID}}}}",
|
f"docker ps --all --quiet --filter name={CONTAINER_NAME} --format={{{{.ID}}}}",
|
||||||
|
@ -1,57 +1,72 @@
|
|||||||
#!/usr/bin/env python3
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
|
||||||
def gen_queries():
|
def gen_queries():
|
||||||
create_template = 'create table tab_00386 (a Int8, b String, c Tuple(Int8), d Tuple(Tuple(Int8)), e Tuple(Int8, String), f Tuple(Tuple(Int8, String))) engine = MergeTree order by ({}) partition by {}'
|
create_template = "create table tab_00386 (a Int8, b String, c Tuple(Int8), d Tuple(Tuple(Int8)), e Tuple(Int8, String), f Tuple(Tuple(Int8, String))) engine = MergeTree order by ({}) partition by {}"
|
||||||
drop_query = 'drop table if exists tab_00386'
|
drop_query = "drop table if exists tab_00386"
|
||||||
values = ('1', "'a'", 'tuple(1)', 'tuple(tuple(1))', "(1, 'a')", "tuple((1, 'a'))")
|
values = ("1", "'a'", "tuple(1)", "tuple(tuple(1))", "(1, 'a')", "tuple((1, 'a'))")
|
||||||
insert_query = "insert into tab_00386 values (1, 'a', tuple(1), tuple(tuple(1)), (1, 'a'), tuple((1, 'a')))"
|
insert_query = "insert into tab_00386 values (1, 'a', tuple(1), tuple(tuple(1)), (1, 'a'), tuple((1, 'a')))"
|
||||||
columns = tuple('a b c d'.split())
|
columns = tuple("a b c d".split())
|
||||||
order_by_columns = tuple('a b c'.split())
|
order_by_columns = tuple("a b c".split())
|
||||||
partition_by_columns = tuple(' tuple() a'.split())
|
partition_by_columns = tuple(" tuple() a".split())
|
||||||
|
|
||||||
for partition in partition_by_columns:
|
for partition in partition_by_columns:
|
||||||
for key_mask in range(1, 1 << len(order_by_columns)):
|
for key_mask in range(1, 1 << len(order_by_columns)):
|
||||||
key = ','.join(order_by_columns[i] for i in range(len(order_by_columns)) if (1 << i) & key_mask != 0)
|
key = ",".join(
|
||||||
|
order_by_columns[i]
|
||||||
|
for i in range(len(order_by_columns))
|
||||||
|
if (1 << i) & key_mask != 0
|
||||||
|
)
|
||||||
create_query = create_template.format(key, partition)
|
create_query = create_template.format(key, partition)
|
||||||
for q in (drop_query, create_query, insert_query):
|
for q in (drop_query, create_query, insert_query):
|
||||||
yield q
|
yield q
|
||||||
|
|
||||||
for column, value in zip(columns, values):
|
for column, value in zip(columns, values):
|
||||||
yield 'select {} in {} from tab_00386'.format(column, value)
|
yield "select {} in {} from tab_00386".format(column, value)
|
||||||
yield 'select {} in tuple({}) from tab_00386'.format(column, value)
|
yield "select {} in tuple({}) from tab_00386".format(column, value)
|
||||||
yield 'select {} in (select {} from tab_00386) from tab_00386'.format(column, column)
|
yield "select {} in (select {} from tab_00386) from tab_00386".format(
|
||||||
|
column, column
|
||||||
|
)
|
||||||
|
|
||||||
for i in range(len(columns)):
|
for i in range(len(columns)):
|
||||||
for j in range(i, len(columns)):
|
for j in range(i, len(columns)):
|
||||||
yield 'select ({}, {}) in tuple({}, {}) from tab_00386'.format(columns[i], columns[j], values[i], values[j])
|
yield "select ({}, {}) in tuple({}, {}) from tab_00386".format(
|
||||||
yield 'select ({}, {}) in (select {}, {} from tab_00386) from tab_00386'.format(columns[i], columns[j], columns[i], columns[j])
|
columns[i], columns[j], values[i], values[j]
|
||||||
yield 'select ({}, {}) in (select ({}, {}) from tab_00386) from tab_00386'.format(columns[i], columns[j], columns[i], columns[j])
|
)
|
||||||
|
yield "select ({}, {}) in (select {}, {} from tab_00386) from tab_00386".format(
|
||||||
|
columns[i], columns[j], columns[i], columns[j]
|
||||||
|
)
|
||||||
|
yield "select ({}, {}) in (select ({}, {}) from tab_00386) from tab_00386".format(
|
||||||
|
columns[i], columns[j], columns[i], columns[j]
|
||||||
|
)
|
||||||
|
|
||||||
yield "select e in (1, 'a') from tab_00386"
|
yield "select e in (1, 'a') from tab_00386"
|
||||||
yield "select f in tuple((1, 'a')) from tab_00386"
|
yield "select f in tuple((1, 'a')) from tab_00386"
|
||||||
yield "select f in tuple(tuple((1, 'a'))) from tab_00386"
|
yield "select f in tuple(tuple((1, 'a'))) from tab_00386"
|
||||||
|
|
||||||
yield 'select e in (select a, b from tab_00386) from tab_00386'
|
yield "select e in (select a, b from tab_00386) from tab_00386"
|
||||||
yield 'select e in (select (a, b) from tab_00386) from tab_00386'
|
yield "select e in (select (a, b) from tab_00386) from tab_00386"
|
||||||
yield 'select f in (select tuple((a, b)) from tab_00386) from tab_00386'
|
yield "select f in (select tuple((a, b)) from tab_00386) from tab_00386"
|
||||||
yield 'select tuple(f) in (select tuple(tuple((a, b))) from tab_00386) from tab_00386'
|
yield "select tuple(f) in (select tuple(tuple((a, b))) from tab_00386) from tab_00386"
|
||||||
|
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
import os
|
import os
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
url = os.environ['CLICKHOUSE_URL']
|
url = os.environ["CLICKHOUSE_URL"]
|
||||||
|
|
||||||
for q in gen_queries():
|
for q in gen_queries():
|
||||||
resp = requests.post(url, data=q)
|
resp = requests.post(url, data=q)
|
||||||
if resp.status_code != 200 or resp.text.strip() not in ('1', ''):
|
if resp.status_code != 200 or resp.text.strip() not in ("1", ""):
|
||||||
print('Query:', q)
|
print("Query:", q)
|
||||||
print('Code:', resp.status_code)
|
print("Code:", resp.status_code)
|
||||||
print(resp.text)
|
print(resp.text)
|
||||||
break
|
break
|
||||||
|
|
||||||
requests.post(url, data='drop table tab_00386')
|
requests.post(url, data="drop table tab_00386")
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
main()
|
main()
|
||||||
|
|
||||||
|
@ -2,8 +2,20 @@
|
|||||||
|
|
||||||
import os, itertools, urllib.request, urllib.parse, urllib.error, urllib.request, urllib.error, urllib.parse, sys
|
import os, itertools, urllib.request, urllib.parse, urllib.error, urllib.request, urllib.error, urllib.parse, sys
|
||||||
|
|
||||||
|
|
||||||
def get_ch_answer(query):
|
def get_ch_answer(query):
|
||||||
return urllib.request.urlopen(os.environ.get('CLICKHOUSE_URL', 'http://localhost:' + os.environ.get('CLICKHOUSE_PORT_HTTP', '8123') ), data=query.encode()).read().decode()
|
return (
|
||||||
|
urllib.request.urlopen(
|
||||||
|
os.environ.get(
|
||||||
|
"CLICKHOUSE_URL",
|
||||||
|
"http://localhost:" + os.environ.get("CLICKHOUSE_PORT_HTTP", "8123"),
|
||||||
|
),
|
||||||
|
data=query.encode(),
|
||||||
|
)
|
||||||
|
.read()
|
||||||
|
.decode()
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def check_answers(query, answer):
|
def check_answers(query, answer):
|
||||||
ch_answer = get_ch_answer(query)
|
ch_answer = get_ch_answer(query)
|
||||||
@ -13,16 +25,18 @@ def check_answers(query, answer):
|
|||||||
print("Fetched answer :", ch_answer)
|
print("Fetched answer :", ch_answer)
|
||||||
exit(-1)
|
exit(-1)
|
||||||
|
|
||||||
|
|
||||||
def get_values():
|
def get_values():
|
||||||
values = [0, 1, -1]
|
values = [0, 1, -1]
|
||||||
for bits in [8, 16, 32, 64]:
|
for bits in [8, 16, 32, 64]:
|
||||||
values += [2**bits, 2**bits - 1]
|
values += [2**bits, 2**bits - 1]
|
||||||
values += [2 ** (bits - 1) - 1, 2 ** (bits - 1), 2 ** (bits - 1) + 1]
|
values += [2 ** (bits - 1) - 1, 2 ** (bits - 1), 2 ** (bits - 1) + 1]
|
||||||
values += [-2**(bits-1) - 1, -2**(bits-1), -2**(bits-1) + 1]
|
values += [-(2 ** (bits - 1)) - 1, -(2 ** (bits - 1)), -(2 ** (bits - 1)) + 1]
|
||||||
return values
|
return values
|
||||||
|
|
||||||
|
|
||||||
def is_valid_integer(x):
|
def is_valid_integer(x):
|
||||||
return -2**63 <= x and x <= 2**64-1
|
return -(2**63) <= x and x <= 2**64 - 1
|
||||||
|
|
||||||
|
|
||||||
TEST_WITH_CASTING = True
|
TEST_WITH_CASTING = True
|
||||||
@ -31,16 +45,12 @@ GENERATE_TEST_FILES=False
|
|||||||
TYPES = {
|
TYPES = {
|
||||||
"UInt8": {"bits": 8, "sign": False, "float": False},
|
"UInt8": {"bits": 8, "sign": False, "float": False},
|
||||||
"Int8": {"bits": 8, "sign": True, "float": False},
|
"Int8": {"bits": 8, "sign": True, "float": False},
|
||||||
|
|
||||||
"UInt16": {"bits": 16, "sign": False, "float": False},
|
"UInt16": {"bits": 16, "sign": False, "float": False},
|
||||||
"Int16": {"bits": 16, "sign": True, "float": False},
|
"Int16": {"bits": 16, "sign": True, "float": False},
|
||||||
|
|
||||||
"UInt32": {"bits": 32, "sign": False, "float": False},
|
"UInt32": {"bits": 32, "sign": False, "float": False},
|
||||||
"Int32": {"bits": 32, "sign": True, "float": False},
|
"Int32": {"bits": 32, "sign": True, "float": False},
|
||||||
|
|
||||||
"UInt64": {"bits": 64, "sign": False, "float": False},
|
"UInt64": {"bits": 64, "sign": False, "float": False},
|
||||||
"Int64": {"bits": 64, "sign": True, "float": False}
|
"Int64": {"bits": 64, "sign": True, "float": False}
|
||||||
|
|
||||||
# "Float32" : { "bits" : 32, "sign" : True, "float" : True },
|
# "Float32" : { "bits" : 32, "sign" : True, "float" : True },
|
||||||
# "Float64" : { "bits" : 64, "sign" : True, "float" : True }
|
# "Float64" : { "bits" : 64, "sign" : True, "float" : True }
|
||||||
}
|
}
|
||||||
@ -55,14 +65,18 @@ def inside_range(value, type_name):
|
|||||||
return True
|
return True
|
||||||
|
|
||||||
if signed:
|
if signed:
|
||||||
return -2**(bits-1) <= value and value <= 2**(bits-1) - 1
|
return -(2 ** (bits - 1)) <= value and value <= 2 ** (bits - 1) - 1
|
||||||
else:
|
else:
|
||||||
return 0 <= value and value <= 2**bits - 1
|
return 0 <= value and value <= 2**bits - 1
|
||||||
|
|
||||||
|
|
||||||
def test_operators(v1, v2, v1_passed, v2_passed):
|
def test_operators(v1, v2, v1_passed, v2_passed):
|
||||||
query_str = "{v1} = {v2}, {v1} != {v2}, {v1} < {v2}, {v1} <= {v2}, {v1} > {v2}, {v1} >= {v2},\t".format(v1=v1_passed, v2=v2_passed)
|
query_str = "{v1} = {v2}, {v1} != {v2}, {v1} < {v2}, {v1} <= {v2}, {v1} > {v2}, {v1} >= {v2},\t".format(
|
||||||
query_str += "{v1} = {v2}, {v1} != {v2}, {v1} < {v2}, {v1} <= {v2}, {v1} > {v2}, {v1} >= {v2} ".format(v1=v2_passed, v2=v1_passed)
|
v1=v1_passed, v2=v2_passed
|
||||||
|
)
|
||||||
|
query_str += "{v1} = {v2}, {v1} != {v2}, {v1} < {v2}, {v1} <= {v2}, {v1} > {v2}, {v1} >= {v2} ".format(
|
||||||
|
v1=v2_passed, v2=v1_passed
|
||||||
|
)
|
||||||
|
|
||||||
answers = [v1 == v2, v1 != v2, v1 < v2, v1 <= v2, v1 > v2, v1 >= v2]
|
answers = [v1 == v2, v1 != v2, v1 < v2, v1 <= v2, v1 > v2, v1 >= v2]
|
||||||
answers += [v2 == v1, v2 != v1, v2 < v1, v2 <= v1, v2 > v1, v2 >= v1]
|
answers += [v2 == v1, v2 != v1, v2 < v1, v2 <= v1, v2 > v1, v2 >= v1]
|
||||||
@ -74,6 +88,7 @@ def test_operators(v1, v2, v1_passed, v2_passed):
|
|||||||
|
|
||||||
VALUES = [x for x in get_values() if is_valid_integer(x)]
|
VALUES = [x for x in get_values() if is_valid_integer(x)]
|
||||||
|
|
||||||
|
|
||||||
def test_pair(v1, v2):
|
def test_pair(v1, v2):
|
||||||
query = "SELECT {}, {}, ".format(v1, v2)
|
query = "SELECT {}, {}, ".format(v1, v2)
|
||||||
answers = "{}\t{}\t".format(v1, v2)
|
answers = "{}\t{}\t".format(v1, v2)
|
||||||
@ -87,19 +102,58 @@ def test_pair(v1, v2):
|
|||||||
if inside_range(v1, t1):
|
if inside_range(v1, t1):
|
||||||
for t2 in TYPES.keys():
|
for t2 in TYPES.keys():
|
||||||
if inside_range(v2, t2):
|
if inside_range(v2, t2):
|
||||||
q, a = test_operators(v1, v2, 'to{}({})'.format(t1, v1), 'to{}({})'.format(t2, v2))
|
q, a = test_operators(
|
||||||
query += ', ' + q
|
v1, v2, "to{}({})".format(t1, v1), "to{}({})".format(t2, v2)
|
||||||
|
)
|
||||||
|
query += ", " + q
|
||||||
answers += "\t" + a
|
answers += "\t" + a
|
||||||
|
|
||||||
check_answers(query, answers)
|
check_answers(query, answers)
|
||||||
return query, answers
|
return query, answers
|
||||||
|
|
||||||
|
|
||||||
VALUES_INT = [0, -1, 1, 2**64-1, 2**63, -2**63, 2**63-1, 2**51, 2**52, 2**53-1, 2**53, 2**53+1, 2**53+2, -2**53+1, -2**53, -2**53-1, -2**53-2, 2*52, -2**52]
|
VALUES_INT = [
|
||||||
VALUES_FLOAT = [float(x) for x in VALUES_INT + [-0.5, 0.5, -1.5, 1.5, 2**53, 2**51 - 0.5, 2**51 + 0.5, 2**60, -2**60, -2**63 - 10000, 2**63 + 10000]]
|
0,
|
||||||
|
-1,
|
||||||
|
1,
|
||||||
|
2**64 - 1,
|
||||||
|
2**63,
|
||||||
|
-(2**63),
|
||||||
|
2**63 - 1,
|
||||||
|
2**51,
|
||||||
|
2**52,
|
||||||
|
2**53 - 1,
|
||||||
|
2**53,
|
||||||
|
2**53 + 1,
|
||||||
|
2**53 + 2,
|
||||||
|
-(2**53) + 1,
|
||||||
|
-(2**53),
|
||||||
|
-(2**53) - 1,
|
||||||
|
-(2**53) - 2,
|
||||||
|
2 * 52,
|
||||||
|
-(2**52),
|
||||||
|
]
|
||||||
|
VALUES_FLOAT = [
|
||||||
|
float(x)
|
||||||
|
for x in VALUES_INT
|
||||||
|
+ [
|
||||||
|
-0.5,
|
||||||
|
0.5,
|
||||||
|
-1.5,
|
||||||
|
1.5,
|
||||||
|
2**53,
|
||||||
|
2**51 - 0.5,
|
||||||
|
2**51 + 0.5,
|
||||||
|
2**60,
|
||||||
|
-(2**60),
|
||||||
|
-(2**63) - 10000,
|
||||||
|
2**63 + 10000,
|
||||||
|
]
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
def test_float_pair(i, f):
|
def test_float_pair(i, f):
|
||||||
f_str = ("%.9f" % f)
|
f_str = "%.9f" % f
|
||||||
query = "SELECT '{}', '{}', ".format(i, f_str)
|
query = "SELECT '{}', '{}', ".format(i, f_str)
|
||||||
answers = "{}\t{}\t".format(i, f_str)
|
answers = "{}\t{}\t".format(i, f_str)
|
||||||
|
|
||||||
@ -110,8 +164,8 @@ def test_float_pair(i, f):
|
|||||||
if TEST_WITH_CASTING:
|
if TEST_WITH_CASTING:
|
||||||
for t1 in TYPES.keys():
|
for t1 in TYPES.keys():
|
||||||
if inside_range(i, t1):
|
if inside_range(i, t1):
|
||||||
q, a = test_operators(i, f, 'to{}({})'.format(t1, i), f_str)
|
q, a = test_operators(i, f, "to{}({})".format(t1, i), f_str)
|
||||||
query += ', ' + q
|
query += ", " + q
|
||||||
answers += "\t" + a
|
answers += "\t" + a
|
||||||
|
|
||||||
check_answers(query, answers)
|
check_answers(query, answers)
|
||||||
@ -120,22 +174,26 @@ def test_float_pair(i, f):
|
|||||||
|
|
||||||
def main():
|
def main():
|
||||||
if GENERATE_TEST_FILES:
|
if GENERATE_TEST_FILES:
|
||||||
base_name = '00411_accurate_number_comparison'
|
base_name = "00411_accurate_number_comparison"
|
||||||
sql_file = open(base_name + '.sql', 'wt')
|
sql_file = open(base_name + ".sql", "wt")
|
||||||
ref_file = open(base_name + '.reference', 'wt')
|
ref_file = open(base_name + ".reference", "wt")
|
||||||
|
|
||||||
num_int_tests = len(list(itertools.combinations(VALUES, 2)))
|
num_int_tests = len(list(itertools.combinations(VALUES, 2)))
|
||||||
|
|
||||||
num_parts = 4
|
num_parts = 4
|
||||||
for part in range(0, num_parts):
|
for part in range(0, num_parts):
|
||||||
if 'int' + str(part + 1) in sys.argv[1:]:
|
if "int" + str(part + 1) in sys.argv[1:]:
|
||||||
for (v1, v2) in itertools.islice(itertools.combinations(VALUES, 2), part * num_int_tests // num_parts, (part + 1) * num_int_tests // num_parts):
|
for (v1, v2) in itertools.islice(
|
||||||
|
itertools.combinations(VALUES, 2),
|
||||||
|
part * num_int_tests // num_parts,
|
||||||
|
(part + 1) * num_int_tests // num_parts,
|
||||||
|
):
|
||||||
q, a = test_pair(v1, v2)
|
q, a = test_pair(v1, v2)
|
||||||
if GENERATE_TEST_FILES:
|
if GENERATE_TEST_FILES:
|
||||||
sql_file.write(q + ";\n")
|
sql_file.write(q + ";\n")
|
||||||
ref_file.write(a + "\n")
|
ref_file.write(a + "\n")
|
||||||
|
|
||||||
if 'float' in sys.argv[1:]:
|
if "float" in sys.argv[1:]:
|
||||||
for (i, f) in itertools.product(VALUES_INT, VALUES_FLOAT):
|
for (i, f) in itertools.product(VALUES_INT, VALUES_FLOAT):
|
||||||
q, a = test_float_pair(i, f)
|
q, a = test_float_pair(i, f)
|
||||||
if GENERATE_TEST_FILES:
|
if GENERATE_TEST_FILES:
|
||||||
|
@ -12,6 +12,7 @@ import subprocess
|
|||||||
from io import StringIO
|
from io import StringIO
|
||||||
from http.server import BaseHTTPRequestHandler, HTTPServer
|
from http.server import BaseHTTPRequestHandler, HTTPServer
|
||||||
|
|
||||||
|
|
||||||
def is_ipv6(host):
|
def is_ipv6(host):
|
||||||
try:
|
try:
|
||||||
socket.inet_aton(host)
|
socket.inet_aton(host)
|
||||||
@ -19,6 +20,7 @@ def is_ipv6(host):
|
|||||||
except:
|
except:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
def get_local_port(host, ipv6):
|
def get_local_port(host, ipv6):
|
||||||
if ipv6:
|
if ipv6:
|
||||||
family = socket.AF_INET6
|
family = socket.AF_INET6
|
||||||
@ -29,8 +31,9 @@ def get_local_port(host, ipv6):
|
|||||||
fd.bind((host, 0))
|
fd.bind((host, 0))
|
||||||
return fd.getsockname()[1]
|
return fd.getsockname()[1]
|
||||||
|
|
||||||
CLICKHOUSE_HOST = os.environ.get('CLICKHOUSE_HOST', '127.0.0.1')
|
|
||||||
CLICKHOUSE_PORT_HTTP = os.environ.get('CLICKHOUSE_PORT_HTTP', '8123')
|
CLICKHOUSE_HOST = os.environ.get("CLICKHOUSE_HOST", "127.0.0.1")
|
||||||
|
CLICKHOUSE_PORT_HTTP = os.environ.get("CLICKHOUSE_PORT_HTTP", "8123")
|
||||||
|
|
||||||
#####################################################################################
|
#####################################################################################
|
||||||
# This test starts an HTTP server and serves data to clickhouse url-engine based table.
|
# This test starts an HTTP server and serves data to clickhouse url-engine based table.
|
||||||
@ -39,27 +42,42 @@ CLICKHOUSE_PORT_HTTP = os.environ.get('CLICKHOUSE_PORT_HTTP', '8123')
|
|||||||
#####################################################################################
|
#####################################################################################
|
||||||
|
|
||||||
# IP-address of this host accessible from the outside world. Get the first one
|
# IP-address of this host accessible from the outside world. Get the first one
|
||||||
HTTP_SERVER_HOST = subprocess.check_output(['hostname', '-i']).decode('utf-8').strip().split()[0]
|
HTTP_SERVER_HOST = (
|
||||||
|
subprocess.check_output(["hostname", "-i"]).decode("utf-8").strip().split()[0]
|
||||||
|
)
|
||||||
IS_IPV6 = is_ipv6(HTTP_SERVER_HOST)
|
IS_IPV6 = is_ipv6(HTTP_SERVER_HOST)
|
||||||
HTTP_SERVER_PORT = get_local_port(HTTP_SERVER_HOST, IS_IPV6)
|
HTTP_SERVER_PORT = get_local_port(HTTP_SERVER_HOST, IS_IPV6)
|
||||||
|
|
||||||
# IP address and port of the HTTP server started from this script.
|
# IP address and port of the HTTP server started from this script.
|
||||||
HTTP_SERVER_ADDRESS = (HTTP_SERVER_HOST, HTTP_SERVER_PORT)
|
HTTP_SERVER_ADDRESS = (HTTP_SERVER_HOST, HTTP_SERVER_PORT)
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
HTTP_SERVER_URL_STR = 'http://' + f'[{str(HTTP_SERVER_ADDRESS[0])}]:{str(HTTP_SERVER_ADDRESS[1])}' + "/"
|
HTTP_SERVER_URL_STR = (
|
||||||
|
"http://"
|
||||||
|
+ f"[{str(HTTP_SERVER_ADDRESS[0])}]:{str(HTTP_SERVER_ADDRESS[1])}"
|
||||||
|
+ "/"
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
HTTP_SERVER_URL_STR = 'http://' + f'{str(HTTP_SERVER_ADDRESS[0])}:{str(HTTP_SERVER_ADDRESS[1])}' + "/"
|
HTTP_SERVER_URL_STR = (
|
||||||
|
"http://" + f"{str(HTTP_SERVER_ADDRESS[0])}:{str(HTTP_SERVER_ADDRESS[1])}" + "/"
|
||||||
|
)
|
||||||
|
|
||||||
|
CSV_DATA = os.path.join(
|
||||||
|
tempfile._get_default_tempdir(), next(tempfile._get_candidate_names())
|
||||||
|
)
|
||||||
|
|
||||||
CSV_DATA = os.path.join(tempfile._get_default_tempdir(), next(tempfile._get_candidate_names()))
|
|
||||||
|
|
||||||
def get_ch_answer(query):
|
def get_ch_answer(query):
|
||||||
host = CLICKHOUSE_HOST
|
host = CLICKHOUSE_HOST
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
host = f'[{host}]'
|
host = f"[{host}]"
|
||||||
|
|
||||||
url = os.environ.get('CLICKHOUSE_URL', 'http://{host}:{port}'.format(host=CLICKHOUSE_HOST, port=CLICKHOUSE_PORT_HTTP))
|
url = os.environ.get(
|
||||||
|
"CLICKHOUSE_URL",
|
||||||
|
"http://{host}:{port}".format(host=CLICKHOUSE_HOST, port=CLICKHOUSE_PORT_HTTP),
|
||||||
|
)
|
||||||
return urllib.request.urlopen(url, data=query.encode()).read().decode()
|
return urllib.request.urlopen(url, data=query.encode()).read().decode()
|
||||||
|
|
||||||
|
|
||||||
def check_answers(query, answer):
|
def check_answers(query, answer):
|
||||||
ch_answer = get_ch_answer(query)
|
ch_answer = get_ch_answer(query)
|
||||||
if ch_answer.strip() != answer.strip():
|
if ch_answer.strip() != answer.strip():
|
||||||
@ -68,18 +86,19 @@ def check_answers(query, answer):
|
|||||||
print("Fetched answer :", ch_answer, file=sys.stderr)
|
print("Fetched answer :", ch_answer, file=sys.stderr)
|
||||||
raise Exception("Fail on query")
|
raise Exception("Fail on query")
|
||||||
|
|
||||||
|
|
||||||
class CSVHTTPServer(BaseHTTPRequestHandler):
|
class CSVHTTPServer(BaseHTTPRequestHandler):
|
||||||
def _set_headers(self):
|
def _set_headers(self):
|
||||||
self.send_response(200)
|
self.send_response(200)
|
||||||
self.send_header('Content-type', 'text/csv')
|
self.send_header("Content-type", "text/csv")
|
||||||
self.end_headers()
|
self.end_headers()
|
||||||
|
|
||||||
def do_GET(self):
|
def do_GET(self):
|
||||||
self._set_headers()
|
self._set_headers()
|
||||||
with open(CSV_DATA, 'r') as fl:
|
with open(CSV_DATA, "r") as fl:
|
||||||
reader = csv.reader(fl, delimiter=',')
|
reader = csv.reader(fl, delimiter=",")
|
||||||
for row in reader:
|
for row in reader:
|
||||||
self.wfile.write((', '.join(row) + '\n').encode())
|
self.wfile.write((", ".join(row) + "\n").encode())
|
||||||
return
|
return
|
||||||
|
|
||||||
def do_HEAD(self):
|
def do_HEAD(self):
|
||||||
@ -87,33 +106,33 @@ class CSVHTTPServer(BaseHTTPRequestHandler):
|
|||||||
return
|
return
|
||||||
|
|
||||||
def read_chunk(self):
|
def read_chunk(self):
|
||||||
msg = ''
|
msg = ""
|
||||||
while True:
|
while True:
|
||||||
sym = self.rfile.read(1)
|
sym = self.rfile.read(1)
|
||||||
if sym == '':
|
if sym == "":
|
||||||
break
|
break
|
||||||
msg += sym.decode('utf-8')
|
msg += sym.decode("utf-8")
|
||||||
if msg.endswith('\r\n'):
|
if msg.endswith("\r\n"):
|
||||||
break
|
break
|
||||||
length = int(msg[:-2], 16)
|
length = int(msg[:-2], 16)
|
||||||
if length == 0:
|
if length == 0:
|
||||||
return ''
|
return ""
|
||||||
content = self.rfile.read(length)
|
content = self.rfile.read(length)
|
||||||
self.rfile.read(2) # read sep \r\n
|
self.rfile.read(2) # read sep \r\n
|
||||||
return content.decode('utf-8')
|
return content.decode("utf-8")
|
||||||
|
|
||||||
def do_POST(self):
|
def do_POST(self):
|
||||||
data = ''
|
data = ""
|
||||||
while True:
|
while True:
|
||||||
chunk = self.read_chunk()
|
chunk = self.read_chunk()
|
||||||
if not chunk:
|
if not chunk:
|
||||||
break
|
break
|
||||||
data += chunk
|
data += chunk
|
||||||
with StringIO(data) as fl:
|
with StringIO(data) as fl:
|
||||||
reader = csv.reader(fl, delimiter=',')
|
reader = csv.reader(fl, delimiter=",")
|
||||||
with open(CSV_DATA, 'a') as d:
|
with open(CSV_DATA, "a") as d:
|
||||||
for row in reader:
|
for row in reader:
|
||||||
d.write(','.join(row) + '\n')
|
d.write(",".join(row) + "\n")
|
||||||
self._set_headers()
|
self._set_headers()
|
||||||
self.wfile.write(b"ok")
|
self.wfile.write(b"ok")
|
||||||
|
|
||||||
@ -124,6 +143,7 @@ class CSVHTTPServer(BaseHTTPRequestHandler):
|
|||||||
class HTTPServerV6(HTTPServer):
|
class HTTPServerV6(HTTPServer):
|
||||||
address_family = socket.AF_INET6
|
address_family = socket.AF_INET6
|
||||||
|
|
||||||
|
|
||||||
def start_server():
|
def start_server():
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
httpd = HTTPServerV6(HTTP_SERVER_ADDRESS, CSVHTTPServer)
|
httpd = HTTPServerV6(HTTP_SERVER_ADDRESS, CSVHTTPServer)
|
||||||
@ -133,49 +153,76 @@ def start_server():
|
|||||||
t = threading.Thread(target=httpd.serve_forever)
|
t = threading.Thread(target=httpd.serve_forever)
|
||||||
return t, httpd
|
return t, httpd
|
||||||
|
|
||||||
|
|
||||||
# test section
|
# test section
|
||||||
|
|
||||||
def test_select(table_name="", schema="str String,numuint UInt32,numint Int32,double Float64", requests=[], answers=[], test_data=""):
|
|
||||||
with open(CSV_DATA, 'w') as f: # clear file
|
def test_select(
|
||||||
f.write('')
|
table_name="",
|
||||||
|
schema="str String,numuint UInt32,numint Int32,double Float64",
|
||||||
|
requests=[],
|
||||||
|
answers=[],
|
||||||
|
test_data="",
|
||||||
|
):
|
||||||
|
with open(CSV_DATA, "w") as f: # clear file
|
||||||
|
f.write("")
|
||||||
|
|
||||||
if test_data:
|
if test_data:
|
||||||
with open(CSV_DATA, 'w') as f:
|
with open(CSV_DATA, "w") as f:
|
||||||
f.write(test_data + "\n")
|
f.write(test_data + "\n")
|
||||||
|
|
||||||
if table_name:
|
if table_name:
|
||||||
get_ch_answer("drop table if exists {}".format(table_name))
|
get_ch_answer("drop table if exists {}".format(table_name))
|
||||||
get_ch_answer("create table {} ({}) engine=URL('{}', 'CSV')".format(table_name, schema, HTTP_SERVER_URL_STR))
|
get_ch_answer(
|
||||||
|
"create table {} ({}) engine=URL('{}', 'CSV')".format(
|
||||||
|
table_name, schema, HTTP_SERVER_URL_STR
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
for i in range(len(requests)):
|
for i in range(len(requests)):
|
||||||
tbl = table_name
|
tbl = table_name
|
||||||
if not tbl:
|
if not tbl:
|
||||||
tbl = "url('{addr}', 'CSV', '{schema}')".format(addr=HTTP_SERVER_URL_STR, schema=schema)
|
tbl = "url('{addr}', 'CSV', '{schema}')".format(
|
||||||
|
addr=HTTP_SERVER_URL_STR, schema=schema
|
||||||
|
)
|
||||||
check_answers(requests[i].format(tbl=tbl), answers[i])
|
check_answers(requests[i].format(tbl=tbl), answers[i])
|
||||||
|
|
||||||
if table_name:
|
if table_name:
|
||||||
get_ch_answer("drop table if exists {}".format(table_name))
|
get_ch_answer("drop table if exists {}".format(table_name))
|
||||||
|
|
||||||
|
|
||||||
def test_insert(table_name="", schema="str String,numuint UInt32,numint Int32,double Float64", requests_insert=[], requests_select=[], answers=[]):
|
def test_insert(
|
||||||
with open(CSV_DATA, 'w') as f: # flush test file
|
table_name="",
|
||||||
f.write('')
|
schema="str String,numuint UInt32,numint Int32,double Float64",
|
||||||
|
requests_insert=[],
|
||||||
|
requests_select=[],
|
||||||
|
answers=[],
|
||||||
|
):
|
||||||
|
with open(CSV_DATA, "w") as f: # flush test file
|
||||||
|
f.write("")
|
||||||
|
|
||||||
if table_name:
|
if table_name:
|
||||||
get_ch_answer("drop table if exists {}".format(table_name))
|
get_ch_answer("drop table if exists {}".format(table_name))
|
||||||
get_ch_answer("create table {} ({}) engine=URL('{}', 'CSV')".format(table_name, schema, HTTP_SERVER_URL_STR))
|
get_ch_answer(
|
||||||
|
"create table {} ({}) engine=URL('{}', 'CSV')".format(
|
||||||
|
table_name, schema, HTTP_SERVER_URL_STR
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
for req in requests_insert:
|
for req in requests_insert:
|
||||||
tbl = table_name
|
tbl = table_name
|
||||||
if not tbl:
|
if not tbl:
|
||||||
tbl = "table function url('{addr}', 'CSV', '{schema}')".format(addr=HTTP_SERVER_URL_STR, schema=schema)
|
tbl = "table function url('{addr}', 'CSV', '{schema}')".format(
|
||||||
|
addr=HTTP_SERVER_URL_STR, schema=schema
|
||||||
|
)
|
||||||
get_ch_answer(req.format(tbl=tbl))
|
get_ch_answer(req.format(tbl=tbl))
|
||||||
|
|
||||||
|
|
||||||
for i in range(len(requests_select)):
|
for i in range(len(requests_select)):
|
||||||
tbl = table_name
|
tbl = table_name
|
||||||
if not tbl:
|
if not tbl:
|
||||||
tbl = "url('{addr}', 'CSV', '{schema}')".format(addr=HTTP_SERVER_URL_STR, schema=schema)
|
tbl = "url('{addr}', 'CSV', '{schema}')".format(
|
||||||
|
addr=HTTP_SERVER_URL_STR, schema=schema
|
||||||
|
)
|
||||||
check_answers(requests_select[i].format(tbl=tbl), answers[i])
|
check_answers(requests_select[i].format(tbl=tbl), answers[i])
|
||||||
|
|
||||||
if table_name:
|
if table_name:
|
||||||
@ -185,9 +232,11 @@ def test_insert(table_name="", schema="str String,numuint UInt32,numint Int32,do
|
|||||||
def main():
|
def main():
|
||||||
test_data = "Hello,2,-2,7.7\nWorld,2,-5,8.8"
|
test_data = "Hello,2,-2,7.7\nWorld,2,-5,8.8"
|
||||||
select_only_requests = {
|
select_only_requests = {
|
||||||
"select str,numuint,numint,double from {tbl}" : test_data.replace(',', '\t'),
|
"select str,numuint,numint,double from {tbl}": test_data.replace(",", "\t"),
|
||||||
"select numuint, count(*) from {tbl} group by numuint": "2\t2",
|
"select numuint, count(*) from {tbl} group by numuint": "2\t2",
|
||||||
"select str,numuint,numint,double from {tbl} limit 1": test_data.split("\n")[0].replace(',', '\t'),
|
"select str,numuint,numint,double from {tbl} limit 1": test_data.split("\n")[
|
||||||
|
0
|
||||||
|
].replace(",", "\t"),
|
||||||
}
|
}
|
||||||
|
|
||||||
insert_requests = [
|
insert_requests = [
|
||||||
@ -196,21 +245,41 @@ def main():
|
|||||||
]
|
]
|
||||||
|
|
||||||
select_requests = {
|
select_requests = {
|
||||||
"select distinct numuint from {tbl} order by numuint": '\n'.join([str(i) for i in range(11)]),
|
"select distinct numuint from {tbl} order by numuint": "\n".join(
|
||||||
"select count(*) from {tbl}": '12',
|
[str(i) for i in range(11)]
|
||||||
'select double, count(*) from {tbl} group by double order by double': "7.7\t2\n9.9\t10"
|
),
|
||||||
|
"select count(*) from {tbl}": "12",
|
||||||
|
"select double, count(*) from {tbl} group by double order by double": "7.7\t2\n9.9\t10",
|
||||||
}
|
}
|
||||||
|
|
||||||
t, httpd = start_server()
|
t, httpd = start_server()
|
||||||
t.start()
|
t.start()
|
||||||
# test table with url engine
|
# test table with url engine
|
||||||
test_select(table_name="test_table_select", requests=list(select_only_requests.keys()), answers=list(select_only_requests.values()), test_data=test_data)
|
test_select(
|
||||||
|
table_name="test_table_select",
|
||||||
|
requests=list(select_only_requests.keys()),
|
||||||
|
answers=list(select_only_requests.values()),
|
||||||
|
test_data=test_data,
|
||||||
|
)
|
||||||
# test table function url
|
# test table function url
|
||||||
test_select(requests=list(select_only_requests.keys()), answers=list(select_only_requests.values()), test_data=test_data)
|
test_select(
|
||||||
|
requests=list(select_only_requests.keys()),
|
||||||
|
answers=list(select_only_requests.values()),
|
||||||
|
test_data=test_data,
|
||||||
|
)
|
||||||
# test insert into table with url engine
|
# test insert into table with url engine
|
||||||
test_insert(table_name="test_table_insert", requests_insert=insert_requests, requests_select=list(select_requests.keys()), answers=list(select_requests.values()))
|
test_insert(
|
||||||
|
table_name="test_table_insert",
|
||||||
|
requests_insert=insert_requests,
|
||||||
|
requests_select=list(select_requests.keys()),
|
||||||
|
answers=list(select_requests.values()),
|
||||||
|
)
|
||||||
# test insert into table function url
|
# test insert into table function url
|
||||||
test_insert(requests_insert=insert_requests, requests_select=list(select_requests.keys()), answers=list(select_requests.values()))
|
test_insert(
|
||||||
|
requests_insert=insert_requests,
|
||||||
|
requests_select=list(select_requests.keys()),
|
||||||
|
answers=list(select_requests.values()),
|
||||||
|
)
|
||||||
|
|
||||||
httpd.shutdown()
|
httpd.shutdown()
|
||||||
t.join()
|
t.join()
|
||||||
|
@ -12,35 +12,46 @@ HAYSTACKS = [
|
|||||||
|
|
||||||
NEEDLE = "needle"
|
NEEDLE = "needle"
|
||||||
|
|
||||||
HAY_RE = re.compile(r'\bhay\b', re.IGNORECASE)
|
HAY_RE = re.compile(r"\bhay\b", re.IGNORECASE)
|
||||||
NEEDLE_RE = re.compile(r'\bneedle\b', re.IGNORECASE)
|
NEEDLE_RE = re.compile(r"\bneedle\b", re.IGNORECASE)
|
||||||
|
|
||||||
|
|
||||||
def replace_follow_case(replacement):
|
def replace_follow_case(replacement):
|
||||||
def func(match):
|
def func(match):
|
||||||
g = match.group()
|
g = match.group()
|
||||||
if g.islower(): return replacement.lower()
|
if g.islower():
|
||||||
if g.istitle(): return replacement.title()
|
return replacement.lower()
|
||||||
if g.isupper(): return replacement.upper()
|
if g.istitle():
|
||||||
|
return replacement.title()
|
||||||
|
if g.isupper():
|
||||||
|
return replacement.upper()
|
||||||
return replacement
|
return replacement
|
||||||
|
|
||||||
return func
|
return func
|
||||||
|
|
||||||
|
|
||||||
def replace_separators(query, new_sep):
|
def replace_separators(query, new_sep):
|
||||||
SEP_RE = re.compile('\\s+')
|
SEP_RE = re.compile("\\s+")
|
||||||
result = SEP_RE.sub(new_sep, query)
|
result = SEP_RE.sub(new_sep, query)
|
||||||
return result
|
return result
|
||||||
|
|
||||||
def enlarge_haystack(query, times, separator=''):
|
|
||||||
return HAY_RE.sub(replace_follow_case(('hay' + separator) * times), query)
|
def enlarge_haystack(query, times, separator=""):
|
||||||
|
return HAY_RE.sub(replace_follow_case(("hay" + separator) * times), query)
|
||||||
|
|
||||||
|
|
||||||
def small_needle(query):
|
def small_needle(query):
|
||||||
return NEEDLE_RE.sub(replace_follow_case('n'), query)
|
return NEEDLE_RE.sub(replace_follow_case("n"), query)
|
||||||
|
|
||||||
|
|
||||||
def remove_needle(query):
|
def remove_needle(query):
|
||||||
return NEEDLE_RE.sub('', query)
|
return NEEDLE_RE.sub("", query)
|
||||||
|
|
||||||
|
|
||||||
def replace_needle(query, new_needle):
|
def replace_needle(query, new_needle):
|
||||||
return NEEDLE_RE.sub(new_needle, query)
|
return NEEDLE_RE.sub(new_needle, query)
|
||||||
|
|
||||||
|
|
||||||
# with str.lower, str.uppert, str.title and such
|
# with str.lower, str.uppert, str.title and such
|
||||||
def transform_needle(query, string_transformation_func):
|
def transform_needle(query, string_transformation_func):
|
||||||
def replace_with_transformation(match):
|
def replace_with_transformation(match):
|
||||||
@ -49,19 +60,21 @@ def transform_needle(query, string_transformation_func):
|
|||||||
|
|
||||||
return NEEDLE_RE.sub(replace_with_transformation, query)
|
return NEEDLE_RE.sub(replace_with_transformation, query)
|
||||||
|
|
||||||
def create_cases(case_sensitive_func, case_insensitive_func, table_row_template, table_query_template, const_query_template):
|
|
||||||
|
def create_cases(
|
||||||
|
case_sensitive_func,
|
||||||
|
case_insensitive_func,
|
||||||
|
table_row_template,
|
||||||
|
table_query_template,
|
||||||
|
const_query_template,
|
||||||
|
):
|
||||||
const_queries = []
|
const_queries = []
|
||||||
table_rows = []
|
table_rows = []
|
||||||
table_queries = set()
|
table_queries = set()
|
||||||
|
|
||||||
def add_case(func, haystack, needle, match):
|
def add_case(func, haystack, needle, match):
|
||||||
match = int(match)
|
match = int(match)
|
||||||
args = dict(
|
args = dict(func=func, haystack=haystack, needle=needle, match=match)
|
||||||
func = func,
|
|
||||||
haystack = haystack,
|
|
||||||
needle = needle,
|
|
||||||
match = match
|
|
||||||
)
|
|
||||||
const_queries.append(const_query_template.substitute(args))
|
const_queries.append(const_query_template.substitute(args))
|
||||||
table_queries.add(table_query_template.substitute(args))
|
table_queries.add(table_query_template.substitute(args))
|
||||||
table_rows.append(table_row_template.substitute(args))
|
table_rows.append(table_row_template.substitute(args))
|
||||||
@ -69,14 +82,28 @@ def create_cases(case_sensitive_func, case_insensitive_func, table_row_template,
|
|||||||
def add_case_sensitive(haystack, needle, match):
|
def add_case_sensitive(haystack, needle, match):
|
||||||
add_case(case_sensitive_func, haystack, needle, match)
|
add_case(case_sensitive_func, haystack, needle, match)
|
||||||
if match:
|
if match:
|
||||||
add_case(case_sensitive_func, transform_needle(haystack, str.swapcase), transform_needle(needle, str.swapcase), match)
|
add_case(
|
||||||
|
case_sensitive_func,
|
||||||
|
transform_needle(haystack, str.swapcase),
|
||||||
|
transform_needle(needle, str.swapcase),
|
||||||
|
match,
|
||||||
|
)
|
||||||
|
|
||||||
def add_case_insensitive(haystack, needle, match):
|
def add_case_insensitive(haystack, needle, match):
|
||||||
add_case(case_insensitive_func, haystack, needle, match)
|
add_case(case_insensitive_func, haystack, needle, match)
|
||||||
if match:
|
if match:
|
||||||
add_case(case_insensitive_func, transform_needle(haystack, str.swapcase), needle, match)
|
add_case(
|
||||||
add_case(case_insensitive_func, haystack, transform_needle(needle, str.swapcase), match)
|
case_insensitive_func,
|
||||||
|
transform_needle(haystack, str.swapcase),
|
||||||
|
needle,
|
||||||
|
match,
|
||||||
|
)
|
||||||
|
add_case(
|
||||||
|
case_insensitive_func,
|
||||||
|
haystack,
|
||||||
|
transform_needle(needle, str.swapcase),
|
||||||
|
match,
|
||||||
|
)
|
||||||
|
|
||||||
# Negative cases
|
# Negative cases
|
||||||
add_case_sensitive(remove_needle(HAYSTACKS[0]), NEEDLE, False)
|
add_case_sensitive(remove_needle(HAYSTACKS[0]), NEEDLE, False)
|
||||||
@ -85,7 +112,7 @@ def create_cases(case_sensitive_func, case_insensitive_func, table_row_template,
|
|||||||
for haystack in HAYSTACKS:
|
for haystack in HAYSTACKS:
|
||||||
add_case_sensitive(transform_needle(haystack, str.swapcase), NEEDLE, False)
|
add_case_sensitive(transform_needle(haystack, str.swapcase), NEEDLE, False)
|
||||||
|
|
||||||
sep = ''
|
sep = ""
|
||||||
h = replace_separators(haystack, sep)
|
h = replace_separators(haystack, sep)
|
||||||
|
|
||||||
add_case_sensitive(h, NEEDLE, False)
|
add_case_sensitive(h, NEEDLE, False)
|
||||||
@ -102,8 +129,7 @@ def create_cases(case_sensitive_func, case_insensitive_func, table_row_template,
|
|||||||
add_case_sensitive(haystack, NEEDLE, True)
|
add_case_sensitive(haystack, NEEDLE, True)
|
||||||
add_case_insensitive(haystack, NEEDLE, True)
|
add_case_insensitive(haystack, NEEDLE, True)
|
||||||
|
|
||||||
|
for sep in list(""" ,"""):
|
||||||
for sep in list(''' ,'''):
|
|
||||||
h = replace_separators(haystack, sep)
|
h = replace_separators(haystack, sep)
|
||||||
add_case_sensitive(h, NEEDLE, True)
|
add_case_sensitive(h, NEEDLE, True)
|
||||||
add_case_sensitive(small_needle(h), small_needle(NEEDLE), True)
|
add_case_sensitive(small_needle(h), small_needle(NEEDLE), True)
|
||||||
@ -114,32 +140,43 @@ def create_cases(case_sensitive_func, case_insensitive_func, table_row_template,
|
|||||||
add_case_insensitive(enlarge_haystack(h, 200, sep), NEEDLE, True)
|
add_case_insensitive(enlarge_haystack(h, 200, sep), NEEDLE, True)
|
||||||
|
|
||||||
# case insesitivity works only on ASCII strings
|
# case insesitivity works only on ASCII strings
|
||||||
add_case_sensitive(replace_needle(h, 'иголка'), replace_needle(NEEDLE, 'иголка'), True)
|
add_case_sensitive(
|
||||||
add_case_sensitive(replace_needle(h, '指针'), replace_needle(NEEDLE, '指针'), True)
|
replace_needle(h, "иголка"), replace_needle(NEEDLE, "иголка"), True
|
||||||
|
)
|
||||||
|
add_case_sensitive(
|
||||||
|
replace_needle(h, "指针"), replace_needle(NEEDLE, "指针"), True
|
||||||
|
)
|
||||||
|
|
||||||
for sep in list('''~!@$%^&*()-=+|]}[{";:/?.><\t''') + [r'\\\\']:
|
for sep in list("""~!@$%^&*()-=+|]}[{";:/?.><\t""") + [r"\\\\"]:
|
||||||
h = replace_separators(HAYSTACKS[0], sep)
|
h = replace_separators(HAYSTACKS[0], sep)
|
||||||
add_case(case_sensitive_func, h, NEEDLE, True)
|
add_case(case_sensitive_func, h, NEEDLE, True)
|
||||||
|
|
||||||
return table_rows, table_queries, const_queries
|
return table_rows, table_queries, const_queries
|
||||||
|
|
||||||
def main():
|
|
||||||
|
|
||||||
|
def main():
|
||||||
def query(x):
|
def query(x):
|
||||||
print(x)
|
print(x)
|
||||||
|
|
||||||
CONST_QUERY = Template("""SELECT ${func}('${haystack}', '${needle}'), ' expecting ', ${match};""")
|
CONST_QUERY = Template(
|
||||||
TABLE_QUERY = Template("""WITH '${needle}' as n
|
"""SELECT ${func}('${haystack}', '${needle}'), ' expecting ', ${match};"""
|
||||||
|
)
|
||||||
|
TABLE_QUERY = Template(
|
||||||
|
"""WITH '${needle}' as n
|
||||||
SELECT haystack, needle, ${func}(haystack, n) as result
|
SELECT haystack, needle, ${func}(haystack, n) as result
|
||||||
FROM ht
|
FROM ht
|
||||||
WHERE func = '${func}' AND needle = n AND result != match;""")
|
WHERE func = '${func}' AND needle = n AND result != match;"""
|
||||||
|
)
|
||||||
TABLE_ROW = Template("""('${haystack}', '${needle}', ${match}, '${func}')""")
|
TABLE_ROW = Template("""('${haystack}', '${needle}', ${match}, '${func}')""")
|
||||||
|
|
||||||
rows, table_queries, const_queries = create_cases('hasToken', 'hasTokenCaseInsensitive', TABLE_ROW, TABLE_QUERY, CONST_QUERY)
|
rows, table_queries, const_queries = create_cases(
|
||||||
|
"hasToken", "hasTokenCaseInsensitive", TABLE_ROW, TABLE_QUERY, CONST_QUERY
|
||||||
|
)
|
||||||
for q in const_queries:
|
for q in const_queries:
|
||||||
query(q)
|
query(q)
|
||||||
|
|
||||||
query("""DROP TABLE IF EXISTS ht;
|
query(
|
||||||
|
"""DROP TABLE IF EXISTS ht;
|
||||||
CREATE TABLE IF NOT EXISTS
|
CREATE TABLE IF NOT EXISTS
|
||||||
ht
|
ht
|
||||||
(
|
(
|
||||||
@ -150,11 +187,15 @@ def main():
|
|||||||
)
|
)
|
||||||
ENGINE MergeTree()
|
ENGINE MergeTree()
|
||||||
ORDER BY haystack;
|
ORDER BY haystack;
|
||||||
INSERT INTO ht VALUES {values};""".format(values=", ".join(rows)))
|
INSERT INTO ht VALUES {values};""".format(
|
||||||
|
values=", ".join(rows)
|
||||||
|
)
|
||||||
|
)
|
||||||
for q in sorted(table_queries):
|
for q in sorted(table_queries):
|
||||||
query(q)
|
query(q)
|
||||||
|
|
||||||
query("""DROP TABLE ht""")
|
query("""DROP TABLE ht""")
|
||||||
|
|
||||||
if __name__ == '__main__':
|
|
||||||
|
if __name__ == "__main__":
|
||||||
main()
|
main()
|
||||||
|
@ -8,28 +8,32 @@ import sys
|
|||||||
import signal
|
import signal
|
||||||
|
|
||||||
|
|
||||||
CLICKHOUSE_CLIENT = os.environ.get('CLICKHOUSE_CLIENT')
|
CLICKHOUSE_CLIENT = os.environ.get("CLICKHOUSE_CLIENT")
|
||||||
CLICKHOUSE_CURL = os.environ.get('CLICKHOUSE_CURL')
|
CLICKHOUSE_CURL = os.environ.get("CLICKHOUSE_CURL")
|
||||||
CLICKHOUSE_URL = os.environ.get('CLICKHOUSE_URL')
|
CLICKHOUSE_URL = os.environ.get("CLICKHOUSE_URL")
|
||||||
|
|
||||||
|
|
||||||
def send_query(query):
|
def send_query(query):
|
||||||
cmd = list(CLICKHOUSE_CLIENT.split())
|
cmd = list(CLICKHOUSE_CLIENT.split())
|
||||||
cmd += ['--query', query]
|
cmd += ["--query", query]
|
||||||
# print(cmd)
|
# print(cmd)
|
||||||
return subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT).stdout
|
return subprocess.Popen(
|
||||||
|
cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT
|
||||||
|
).stdout
|
||||||
|
|
||||||
|
|
||||||
def send_query_in_process_group(query):
|
def send_query_in_process_group(query):
|
||||||
cmd = list(CLICKHOUSE_CLIENT.split())
|
cmd = list(CLICKHOUSE_CLIENT.split())
|
||||||
cmd += ['--query', query]
|
cmd += ["--query", query]
|
||||||
# print(cmd)
|
# print(cmd)
|
||||||
return subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, preexec_fn=os.setsid)
|
return subprocess.Popen(
|
||||||
|
cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, preexec_fn=os.setsid
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def read_lines_and_push_to_queue(pipe, queue):
|
def read_lines_and_push_to_queue(pipe, queue):
|
||||||
try:
|
try:
|
||||||
for line in iter(pipe.readline, ''):
|
for line in iter(pipe.readline, ""):
|
||||||
line = line.strip()
|
line = line.strip()
|
||||||
print(line)
|
print(line)
|
||||||
sys.stdout.flush()
|
sys.stdout.flush()
|
||||||
@ -41,41 +45,44 @@ def read_lines_and_push_to_queue(pipe, queue):
|
|||||||
|
|
||||||
|
|
||||||
def test():
|
def test():
|
||||||
send_query('DROP TABLE IF EXISTS test.lv').read()
|
send_query("DROP TABLE IF EXISTS test.lv").read()
|
||||||
send_query('DROP TABLE IF EXISTS test.mt').read()
|
send_query("DROP TABLE IF EXISTS test.mt").read()
|
||||||
send_query('CREATE TABLE test.mt (a Int32) Engine=MergeTree order by tuple()').read()
|
send_query(
|
||||||
send_query('CREATE LIVE VIEW test.lv AS SELECT sum(a) FROM test.mt').read()
|
"CREATE TABLE test.mt (a Int32) Engine=MergeTree order by tuple()"
|
||||||
|
).read()
|
||||||
|
send_query("CREATE LIVE VIEW test.lv AS SELECT sum(a) FROM test.mt").read()
|
||||||
|
|
||||||
q = queue.Queue()
|
q = queue.Queue()
|
||||||
p = send_query_in_process_group('WATCH test.lv')
|
p = send_query_in_process_group("WATCH test.lv")
|
||||||
thread = threading.Thread(target=read_lines_and_push_to_queue, args=(p.stdout, q))
|
thread = threading.Thread(target=read_lines_and_push_to_queue, args=(p.stdout, q))
|
||||||
thread.start()
|
thread.start()
|
||||||
|
|
||||||
line = q.get()
|
line = q.get()
|
||||||
print(line)
|
print(line)
|
||||||
assert (line == '0\t1')
|
assert line == "0\t1"
|
||||||
|
|
||||||
send_query('INSERT INTO test.mt VALUES (1),(2),(3)').read()
|
send_query("INSERT INTO test.mt VALUES (1),(2),(3)").read()
|
||||||
line = q.get()
|
line = q.get()
|
||||||
print(line)
|
print(line)
|
||||||
assert (line == '6\t2')
|
assert line == "6\t2"
|
||||||
|
|
||||||
send_query('INSERT INTO test.mt VALUES (4),(5),(6)').read()
|
send_query("INSERT INTO test.mt VALUES (4),(5),(6)").read()
|
||||||
line = q.get()
|
line = q.get()
|
||||||
print(line)
|
print(line)
|
||||||
assert (line == '21\t3')
|
assert line == "21\t3"
|
||||||
|
|
||||||
# Send Ctrl+C to client.
|
# Send Ctrl+C to client.
|
||||||
os.killpg(os.getpgid(p.pid), signal.SIGINT)
|
os.killpg(os.getpgid(p.pid), signal.SIGINT)
|
||||||
# This insert shouldn't affect lv.
|
# This insert shouldn't affect lv.
|
||||||
send_query('INSERT INTO test.mt VALUES (7),(8),(9)').read()
|
send_query("INSERT INTO test.mt VALUES (7),(8),(9)").read()
|
||||||
line = q.get()
|
line = q.get()
|
||||||
print(line)
|
print(line)
|
||||||
assert (line is None)
|
assert line is None
|
||||||
|
|
||||||
send_query('DROP TABLE if exists test.lv').read()
|
send_query("DROP TABLE if exists test.lv").read()
|
||||||
send_query('DROP TABLE if exists test.lv').read()
|
send_query("DROP TABLE if exists test.lv").read()
|
||||||
|
|
||||||
thread.join()
|
thread.join()
|
||||||
|
|
||||||
|
|
||||||
test()
|
test()
|
||||||
|
@ -7,26 +7,30 @@ import os
|
|||||||
import sys
|
import sys
|
||||||
|
|
||||||
|
|
||||||
CLICKHOUSE_CLIENT = os.environ.get('CLICKHOUSE_CLIENT')
|
CLICKHOUSE_CLIENT = os.environ.get("CLICKHOUSE_CLIENT")
|
||||||
CLICKHOUSE_CURL = os.environ.get('CLICKHOUSE_CURL')
|
CLICKHOUSE_CURL = os.environ.get("CLICKHOUSE_CURL")
|
||||||
CLICKHOUSE_URL = os.environ.get('CLICKHOUSE_URL')
|
CLICKHOUSE_URL = os.environ.get("CLICKHOUSE_URL")
|
||||||
|
|
||||||
|
|
||||||
def send_query(query):
|
def send_query(query):
|
||||||
cmd = list(CLICKHOUSE_CLIENT.split())
|
cmd = list(CLICKHOUSE_CLIENT.split())
|
||||||
cmd += ['--query', query]
|
cmd += ["--query", query]
|
||||||
# print(cmd)
|
# print(cmd)
|
||||||
return subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT).stdout
|
return subprocess.Popen(
|
||||||
|
cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT
|
||||||
|
).stdout
|
||||||
|
|
||||||
|
|
||||||
def send_http_query(query):
|
def send_http_query(query):
|
||||||
cmd = list(CLICKHOUSE_CURL.split()) # list(['curl', '-sSN', '--max-time', '10'])
|
cmd = list(CLICKHOUSE_CURL.split()) # list(['curl', '-sSN', '--max-time', '10'])
|
||||||
cmd += ['-sSN', CLICKHOUSE_URL, '-d', query]
|
cmd += ["-sSN", CLICKHOUSE_URL, "-d", query]
|
||||||
return subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT).stdout
|
return subprocess.Popen(
|
||||||
|
cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT
|
||||||
|
).stdout
|
||||||
|
|
||||||
|
|
||||||
def read_lines_and_push_to_queue(pipe, queue):
|
def read_lines_and_push_to_queue(pipe, queue):
|
||||||
for line in iter(pipe.readline, ''):
|
for line in iter(pipe.readline, ""):
|
||||||
line = line.strip()
|
line = line.strip()
|
||||||
print(line)
|
print(line)
|
||||||
sys.stdout.flush()
|
sys.stdout.flush()
|
||||||
@ -36,28 +40,31 @@ def read_lines_and_push_to_queue(pipe, queue):
|
|||||||
|
|
||||||
|
|
||||||
def test():
|
def test():
|
||||||
send_query('DROP TABLE IF EXISTS test.lv').read()
|
send_query("DROP TABLE IF EXISTS test.lv").read()
|
||||||
send_query('DROP TABLE IF EXISTS test.mt').read()
|
send_query("DROP TABLE IF EXISTS test.mt").read()
|
||||||
send_query('CREATE TABLE test.mt (a Int32) Engine=MergeTree order by tuple()').read()
|
send_query(
|
||||||
send_query('CREATE LIVE VIEW test.lv AS SELECT sum(a) FROM test.mt').read()
|
"CREATE TABLE test.mt (a Int32) Engine=MergeTree order by tuple()"
|
||||||
|
).read()
|
||||||
|
send_query("CREATE LIVE VIEW test.lv AS SELECT sum(a) FROM test.mt").read()
|
||||||
|
|
||||||
q = queue.Queue()
|
q = queue.Queue()
|
||||||
pipe = send_http_query('WATCH test.lv')
|
pipe = send_http_query("WATCH test.lv")
|
||||||
thread = threading.Thread(target=read_lines_and_push_to_queue, args=(pipe, q))
|
thread = threading.Thread(target=read_lines_and_push_to_queue, args=(pipe, q))
|
||||||
thread.start()
|
thread.start()
|
||||||
|
|
||||||
line = q.get()
|
line = q.get()
|
||||||
print(line)
|
print(line)
|
||||||
assert (line == '0\t1')
|
assert line == "0\t1"
|
||||||
|
|
||||||
send_query('INSERT INTO test.mt VALUES (1),(2),(3)').read()
|
send_query("INSERT INTO test.mt VALUES (1),(2),(3)").read()
|
||||||
line = q.get()
|
line = q.get()
|
||||||
print(line)
|
print(line)
|
||||||
assert (line == '6\t2')
|
assert line == "6\t2"
|
||||||
|
|
||||||
send_query('DROP TABLE if exists test.lv').read()
|
send_query("DROP TABLE if exists test.lv").read()
|
||||||
send_query('DROP TABLE if exists test.lv').read()
|
send_query("DROP TABLE if exists test.lv").read()
|
||||||
|
|
||||||
thread.join()
|
thread.join()
|
||||||
|
|
||||||
|
|
||||||
test()
|
test()
|
||||||
|
@ -6,24 +6,38 @@ import pandas as pd
|
|||||||
import numpy as np
|
import numpy as np
|
||||||
|
|
||||||
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
||||||
sys.path.insert(0, os.path.join(CURDIR, 'helpers'))
|
sys.path.insert(0, os.path.join(CURDIR, "helpers"))
|
||||||
|
|
||||||
from pure_http_client import ClickHouseClient
|
from pure_http_client import ClickHouseClient
|
||||||
|
|
||||||
|
|
||||||
def test_and_check(name, a, b, t_stat, p_value, precision=1e-2):
|
def test_and_check(name, a, b, t_stat, p_value, precision=1e-2):
|
||||||
client = ClickHouseClient()
|
client = ClickHouseClient()
|
||||||
client.query("DROP TABLE IF EXISTS ttest;")
|
client.query("DROP TABLE IF EXISTS ttest;")
|
||||||
client.query("CREATE TABLE ttest (left Float64, right UInt8) ENGINE = Memory;");
|
client.query("CREATE TABLE ttest (left Float64, right UInt8) ENGINE = Memory;")
|
||||||
client.query("INSERT INTO ttest VALUES {};".format(", ".join(['({},{})'.format(i, 0) for i in a])))
|
client.query(
|
||||||
client.query("INSERT INTO ttest VALUES {};".format(", ".join(['({},{})'.format(j, 1) for j in b])))
|
"INSERT INTO ttest VALUES {};".format(
|
||||||
|
", ".join(["({},{})".format(i, 0) for i in a])
|
||||||
|
)
|
||||||
|
)
|
||||||
|
client.query(
|
||||||
|
"INSERT INTO ttest VALUES {};".format(
|
||||||
|
", ".join(["({},{})".format(j, 1) for j in b])
|
||||||
|
)
|
||||||
|
)
|
||||||
real = client.query_return_df(
|
real = client.query_return_df(
|
||||||
"SELECT roundBankers({}(left, right).1, 16) as t_stat, ".format(name) +
|
"SELECT roundBankers({}(left, right).1, 16) as t_stat, ".format(name)
|
||||||
"roundBankers({}(left, right).2, 16) as p_value ".format(name) +
|
+ "roundBankers({}(left, right).2, 16) as p_value ".format(name)
|
||||||
"FROM ttest FORMAT TabSeparatedWithNames;")
|
+ "FROM ttest FORMAT TabSeparatedWithNames;"
|
||||||
real_t_stat = real['t_stat'][0]
|
)
|
||||||
real_p_value = real['p_value'][0]
|
real_t_stat = real["t_stat"][0]
|
||||||
assert(abs(real_t_stat - np.float64(t_stat)) < precision), "clickhouse_t_stat {}, scipy_t_stat {}".format(real_t_stat, t_stat)
|
real_p_value = real["p_value"][0]
|
||||||
assert(abs(real_p_value - np.float64(p_value)) < precision), "clickhouse_p_value {}, scipy_p_value {}".format(real_p_value, p_value)
|
assert (
|
||||||
|
abs(real_t_stat - np.float64(t_stat)) < precision
|
||||||
|
), "clickhouse_t_stat {}, scipy_t_stat {}".format(real_t_stat, t_stat)
|
||||||
|
assert (
|
||||||
|
abs(real_p_value - np.float64(p_value)) < precision
|
||||||
|
), "clickhouse_p_value {}, scipy_p_value {}".format(real_p_value, p_value)
|
||||||
client.query("DROP TABLE IF EXISTS ttest;")
|
client.query("DROP TABLE IF EXISTS ttest;")
|
||||||
|
|
||||||
|
|
||||||
@ -48,6 +62,7 @@ def test_student():
|
|||||||
s, p = stats.ttest_ind(rvs1, rvs2, equal_var=True)
|
s, p = stats.ttest_ind(rvs1, rvs2, equal_var=True)
|
||||||
test_and_check("studentTTest", rvs1, rvs2, s, p)
|
test_and_check("studentTTest", rvs1, rvs2, s, p)
|
||||||
|
|
||||||
|
|
||||||
def test_welch():
|
def test_welch():
|
||||||
rvs1 = np.round(stats.norm.rvs(loc=1, scale=15, size=500), 2)
|
rvs1 = np.round(stats.norm.rvs(loc=1, scale=15, size=500), 2)
|
||||||
rvs2 = np.round(stats.norm.rvs(loc=10, scale=5, size=500), 2)
|
rvs2 = np.round(stats.norm.rvs(loc=10, scale=5, size=500), 2)
|
||||||
@ -69,6 +84,7 @@ def test_welch():
|
|||||||
s, p = stats.ttest_ind(rvs1, rvs2, equal_var=False)
|
s, p = stats.ttest_ind(rvs1, rvs2, equal_var=False)
|
||||||
test_and_check("welchTTest", rvs1, rvs2, s, p)
|
test_and_check("welchTTest", rvs1, rvs2, s, p)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
test_student()
|
test_student()
|
||||||
test_welch()
|
test_welch()
|
||||||
|
@ -6,7 +6,7 @@ import pandas as pd
|
|||||||
import numpy as np
|
import numpy as np
|
||||||
|
|
||||||
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
||||||
sys.path.insert(0, os.path.join(CURDIR, 'helpers'))
|
sys.path.insert(0, os.path.join(CURDIR, "helpers"))
|
||||||
|
|
||||||
from pure_http_client import ClickHouseClient
|
from pure_http_client import ClickHouseClient
|
||||||
|
|
||||||
@ -14,40 +14,51 @@ from pure_http_client import ClickHouseClient
|
|||||||
def test_and_check(name, a, b, t_stat, p_value):
|
def test_and_check(name, a, b, t_stat, p_value):
|
||||||
client = ClickHouseClient()
|
client = ClickHouseClient()
|
||||||
client.query("DROP TABLE IF EXISTS mann_whitney;")
|
client.query("DROP TABLE IF EXISTS mann_whitney;")
|
||||||
client.query("CREATE TABLE mann_whitney (left Float64, right UInt8) ENGINE = Memory;");
|
client.query(
|
||||||
client.query("INSERT INTO mann_whitney VALUES {};".format(", ".join(['({},{}), ({},{})'.format(i, 0, j, 1) for i,j in zip(a, b)])))
|
"CREATE TABLE mann_whitney (left Float64, right UInt8) ENGINE = Memory;"
|
||||||
|
)
|
||||||
|
client.query(
|
||||||
|
"INSERT INTO mann_whitney VALUES {};".format(
|
||||||
|
", ".join(["({},{}), ({},{})".format(i, 0, j, 1) for i, j in zip(a, b)])
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
real = client.query_return_df(
|
real = client.query_return_df(
|
||||||
"SELECT roundBankers({}(left, right).1, 16) as t_stat, ".format(name) +
|
"SELECT roundBankers({}(left, right).1, 16) as t_stat, ".format(name)
|
||||||
"roundBankers({}(left, right).2, 16) as p_value ".format(name) +
|
+ "roundBankers({}(left, right).2, 16) as p_value ".format(name)
|
||||||
"FROM mann_whitney FORMAT TabSeparatedWithNames;")
|
+ "FROM mann_whitney FORMAT TabSeparatedWithNames;"
|
||||||
real_t_stat = real['t_stat'][0]
|
)
|
||||||
real_p_value = real['p_value'][0]
|
real_t_stat = real["t_stat"][0]
|
||||||
assert(abs(real_t_stat - np.float64(t_stat) < 1e-2)), "clickhouse_t_stat {}, scipy_t_stat {}".format(real_t_stat, t_stat)
|
real_p_value = real["p_value"][0]
|
||||||
assert(abs(real_p_value - np.float64(p_value)) < 1e-2), "clickhouse_p_value {}, scipy_p_value {}".format(real_p_value, p_value)
|
assert abs(
|
||||||
|
real_t_stat - np.float64(t_stat) < 1e-2
|
||||||
|
), "clickhouse_t_stat {}, scipy_t_stat {}".format(real_t_stat, t_stat)
|
||||||
|
assert (
|
||||||
|
abs(real_p_value - np.float64(p_value)) < 1e-2
|
||||||
|
), "clickhouse_p_value {}, scipy_p_value {}".format(real_p_value, p_value)
|
||||||
client.query("DROP TABLE IF EXISTS mann_whitney;")
|
client.query("DROP TABLE IF EXISTS mann_whitney;")
|
||||||
|
|
||||||
|
|
||||||
def test_mann_whitney():
|
def test_mann_whitney():
|
||||||
rvs1 = np.round(stats.norm.rvs(loc=1, scale=5, size=500), 5)
|
rvs1 = np.round(stats.norm.rvs(loc=1, scale=5, size=500), 5)
|
||||||
rvs2 = np.round(stats.expon.rvs(scale=0.2, size=500), 5)
|
rvs2 = np.round(stats.expon.rvs(scale=0.2, size=500), 5)
|
||||||
s, p = stats.mannwhitneyu(rvs1, rvs2, alternative='two-sided')
|
s, p = stats.mannwhitneyu(rvs1, rvs2, alternative="two-sided")
|
||||||
test_and_check("mannWhitneyUTest", rvs1, rvs2, s, p)
|
test_and_check("mannWhitneyUTest", rvs1, rvs2, s, p)
|
||||||
test_and_check("mannWhitneyUTest('two-sided')", rvs1, rvs2, s, p)
|
test_and_check("mannWhitneyUTest('two-sided')", rvs1, rvs2, s, p)
|
||||||
|
|
||||||
equal = np.round(stats.cauchy.rvs(scale=5, size=500), 5)
|
equal = np.round(stats.cauchy.rvs(scale=5, size=500), 5)
|
||||||
s, p = stats.mannwhitneyu(equal, equal, alternative='two-sided')
|
s, p = stats.mannwhitneyu(equal, equal, alternative="two-sided")
|
||||||
test_and_check("mannWhitneyUTest('two-sided')", equal, equal, s, p)
|
test_and_check("mannWhitneyUTest('two-sided')", equal, equal, s, p)
|
||||||
|
|
||||||
s, p = stats.mannwhitneyu(equal, equal, alternative='less', use_continuity=False)
|
s, p = stats.mannwhitneyu(equal, equal, alternative="less", use_continuity=False)
|
||||||
test_and_check("mannWhitneyUTest('less', 0)", equal, equal, s, p)
|
test_and_check("mannWhitneyUTest('less', 0)", equal, equal, s, p)
|
||||||
|
|
||||||
|
|
||||||
rvs1 = np.round(stats.cauchy.rvs(scale=10, size=65536), 5)
|
rvs1 = np.round(stats.cauchy.rvs(scale=10, size=65536), 5)
|
||||||
rvs2 = np.round(stats.norm.rvs(loc=0, scale=10, size=65536), 5)
|
rvs2 = np.round(stats.norm.rvs(loc=0, scale=10, size=65536), 5)
|
||||||
s, p = stats.mannwhitneyu(rvs1, rvs2, alternative='greater')
|
s, p = stats.mannwhitneyu(rvs1, rvs2, alternative="greater")
|
||||||
test_and_check("mannWhitneyUTest('greater')", rvs1, rvs2, s, p)
|
test_and_check("mannWhitneyUTest('greater')", rvs1, rvs2, s, p)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
test_mann_whitney()
|
test_mann_whitney()
|
||||||
print("Ok.")
|
print("Ok.")
|
@ -4,14 +4,18 @@ from random import randint, choices
|
|||||||
import sys
|
import sys
|
||||||
|
|
||||||
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
||||||
sys.path.insert(0, os.path.join(CURDIR, 'helpers'))
|
sys.path.insert(0, os.path.join(CURDIR, "helpers"))
|
||||||
|
|
||||||
from pure_http_client import ClickHouseClient
|
from pure_http_client import ClickHouseClient
|
||||||
|
|
||||||
client = ClickHouseClient()
|
client = ClickHouseClient()
|
||||||
|
|
||||||
N = 10
|
N = 10
|
||||||
create_query = "CREATE TABLE t_cnf_fuzz(" + ", ".join([f"c{i} UInt8" for i in range(N)]) + ") ENGINE = Memory"
|
create_query = (
|
||||||
|
"CREATE TABLE t_cnf_fuzz("
|
||||||
|
+ ", ".join([f"c{i} UInt8" for i in range(N)])
|
||||||
|
+ ") ENGINE = Memory"
|
||||||
|
)
|
||||||
|
|
||||||
client.query("DROP TABLE IF EXISTS t_cnf_fuzz")
|
client.query("DROP TABLE IF EXISTS t_cnf_fuzz")
|
||||||
client.query(create_query)
|
client.query(create_query)
|
||||||
@ -35,6 +39,7 @@ client.query(insert_query)
|
|||||||
MAX_CLAUSES = 10
|
MAX_CLAUSES = 10
|
||||||
MAX_ATOMS = 5
|
MAX_ATOMS = 5
|
||||||
|
|
||||||
|
|
||||||
def generate_dnf():
|
def generate_dnf():
|
||||||
clauses = []
|
clauses = []
|
||||||
num_clauses = randint(1, MAX_CLAUSES)
|
num_clauses = randint(1, MAX_CLAUSES)
|
||||||
@ -42,12 +47,17 @@ def generate_dnf():
|
|||||||
num_atoms = randint(1, MAX_ATOMS)
|
num_atoms = randint(1, MAX_ATOMS)
|
||||||
atom_ids = choices(range(N), k=num_atoms)
|
atom_ids = choices(range(N), k=num_atoms)
|
||||||
negates = choices([0, 1], k=num_atoms)
|
negates = choices([0, 1], k=num_atoms)
|
||||||
atoms = [f"(NOT c{i})" if neg else f"c{i}" for (i, neg) in zip(atom_ids, negates)]
|
atoms = [
|
||||||
|
f"(NOT c{i})" if neg else f"c{i}" for (i, neg) in zip(atom_ids, negates)
|
||||||
|
]
|
||||||
clauses.append("(" + " AND ".join(atoms) + ")")
|
clauses.append("(" + " AND ".join(atoms) + ")")
|
||||||
|
|
||||||
return " OR ".join(clauses)
|
return " OR ".join(clauses)
|
||||||
|
|
||||||
select_query = "SELECT count() FROM t_cnf_fuzz WHERE {} SETTINGS convert_query_to_cnf = {}"
|
|
||||||
|
select_query = (
|
||||||
|
"SELECT count() FROM t_cnf_fuzz WHERE {} SETTINGS convert_query_to_cnf = {}"
|
||||||
|
)
|
||||||
|
|
||||||
fail_report = """
|
fail_report = """
|
||||||
Failed query: '{}'.
|
Failed query: '{}'.
|
||||||
|
@ -5,15 +5,20 @@ import random
|
|||||||
import string
|
import string
|
||||||
|
|
||||||
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
||||||
sys.path.insert(0, os.path.join(CURDIR, 'helpers'))
|
sys.path.insert(0, os.path.join(CURDIR, "helpers"))
|
||||||
|
|
||||||
from pure_http_client import ClickHouseClient
|
from pure_http_client import ClickHouseClient
|
||||||
|
|
||||||
|
|
||||||
def get_random_string(length):
|
def get_random_string(length):
|
||||||
return ''.join(random.choice(string.ascii_uppercase + string.digits) for _ in range(length))
|
return "".join(
|
||||||
|
random.choice(string.ascii_uppercase + string.digits) for _ in range(length)
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
client = ClickHouseClient()
|
client = ClickHouseClient()
|
||||||
|
|
||||||
|
|
||||||
def insert_block(table_name, block_granularity_rows, block_rows):
|
def insert_block(table_name, block_granularity_rows, block_rows):
|
||||||
global client
|
global client
|
||||||
block_data = []
|
block_data = []
|
||||||
@ -25,9 +30,12 @@ def insert_block(table_name, block_granularity_rows, block_rows):
|
|||||||
values_row = ", ".join("(1, '" + row + "')" for row in block_data)
|
values_row = ", ".join("(1, '" + row + "')" for row in block_data)
|
||||||
client.query("INSERT INTO {} VALUES {}".format(table_name, values_row))
|
client.query("INSERT INTO {} VALUES {}".format(table_name, values_row))
|
||||||
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
client.query("DROP TABLE IF EXISTS t")
|
client.query("DROP TABLE IF EXISTS t")
|
||||||
client.query("CREATE TABLE t (v UInt8, data String) ENGINE = MergeTree() ORDER BY tuple() SETTINGS min_bytes_for_wide_part = 0")
|
client.query(
|
||||||
|
"CREATE TABLE t (v UInt8, data String) ENGINE = MergeTree() ORDER BY tuple() SETTINGS min_bytes_for_wide_part = 0"
|
||||||
|
)
|
||||||
|
|
||||||
client.query("SYSTEM STOP MERGES t")
|
client.query("SYSTEM STOP MERGES t")
|
||||||
|
|
||||||
@ -53,6 +61,10 @@ try:
|
|||||||
client.query("SYSTEM START MERGES t")
|
client.query("SYSTEM START MERGES t")
|
||||||
client.query("OPTIMIZE TABLE t FINAL")
|
client.query("OPTIMIZE TABLE t FINAL")
|
||||||
|
|
||||||
print(client.query_return_df("SELECT COUNT() as C FROM t FORMAT TabSeparatedWithNames")['C'][0])
|
print(
|
||||||
|
client.query_return_df(
|
||||||
|
"SELECT COUNT() as C FROM t FORMAT TabSeparatedWithNames"
|
||||||
|
)["C"][0]
|
||||||
|
)
|
||||||
finally:
|
finally:
|
||||||
client.query("DROP TABLE IF EXISTS t")
|
client.query("DROP TABLE IF EXISTS t")
|
||||||
|
@ -21,6 +21,7 @@ def is_ipv6(host):
|
|||||||
except:
|
except:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
def get_local_port(host, ipv6):
|
def get_local_port(host, ipv6):
|
||||||
if ipv6:
|
if ipv6:
|
||||||
family = socket.AF_INET6
|
family = socket.AF_INET6
|
||||||
@ -31,8 +32,9 @@ def get_local_port(host, ipv6):
|
|||||||
fd.bind((host, 0))
|
fd.bind((host, 0))
|
||||||
return fd.getsockname()[1]
|
return fd.getsockname()[1]
|
||||||
|
|
||||||
CLICKHOUSE_HOST = os.environ.get('CLICKHOUSE_HOST', 'localhost')
|
|
||||||
CLICKHOUSE_PORT_HTTP = os.environ.get('CLICKHOUSE_PORT_HTTP', '8123')
|
CLICKHOUSE_HOST = os.environ.get("CLICKHOUSE_HOST", "localhost")
|
||||||
|
CLICKHOUSE_PORT_HTTP = os.environ.get("CLICKHOUSE_PORT_HTTP", "8123")
|
||||||
|
|
||||||
#####################################################################################
|
#####################################################################################
|
||||||
# This test starts an HTTP server and serves data to clickhouse url-engine based table.
|
# This test starts an HTTP server and serves data to clickhouse url-engine based table.
|
||||||
@ -42,16 +44,24 @@ CLICKHOUSE_PORT_HTTP = os.environ.get('CLICKHOUSE_PORT_HTTP', '8123')
|
|||||||
#####################################################################################
|
#####################################################################################
|
||||||
|
|
||||||
# IP-address of this host accessible from the outside world. Get the first one
|
# IP-address of this host accessible from the outside world. Get the first one
|
||||||
HTTP_SERVER_HOST = subprocess.check_output(['hostname', '-i']).decode('utf-8').strip().split()[0]
|
HTTP_SERVER_HOST = (
|
||||||
|
subprocess.check_output(["hostname", "-i"]).decode("utf-8").strip().split()[0]
|
||||||
|
)
|
||||||
IS_IPV6 = is_ipv6(HTTP_SERVER_HOST)
|
IS_IPV6 = is_ipv6(HTTP_SERVER_HOST)
|
||||||
HTTP_SERVER_PORT = get_local_port(HTTP_SERVER_HOST, IS_IPV6)
|
HTTP_SERVER_PORT = get_local_port(HTTP_SERVER_HOST, IS_IPV6)
|
||||||
|
|
||||||
# IP address and port of the HTTP server started from this script.
|
# IP address and port of the HTTP server started from this script.
|
||||||
HTTP_SERVER_ADDRESS = (HTTP_SERVER_HOST, HTTP_SERVER_PORT)
|
HTTP_SERVER_ADDRESS = (HTTP_SERVER_HOST, HTTP_SERVER_PORT)
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
HTTP_SERVER_URL_STR = 'http://' + f'[{str(HTTP_SERVER_ADDRESS[0])}]:{str(HTTP_SERVER_ADDRESS[1])}' + "/"
|
HTTP_SERVER_URL_STR = (
|
||||||
|
"http://"
|
||||||
|
+ f"[{str(HTTP_SERVER_ADDRESS[0])}]:{str(HTTP_SERVER_ADDRESS[1])}"
|
||||||
|
+ "/"
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
HTTP_SERVER_URL_STR = 'http://' + f'{str(HTTP_SERVER_ADDRESS[0])}:{str(HTTP_SERVER_ADDRESS[1])}' + "/"
|
HTTP_SERVER_URL_STR = (
|
||||||
|
"http://" + f"{str(HTTP_SERVER_ADDRESS[0])}:{str(HTTP_SERVER_ADDRESS[1])}" + "/"
|
||||||
|
)
|
||||||
|
|
||||||
# Because we need to check the content of file.csv we can create this content and avoid reading csv
|
# Because we need to check the content of file.csv we can create this content and avoid reading csv
|
||||||
CSV_DATA = "Hello, 1\nWorld, 2\nThis, 152\nis, 9283\ntesting, 2313213\ndata, 555\n"
|
CSV_DATA = "Hello, 1\nWorld, 2\nThis, 152\nis, 9283\ntesting, 2313213\ndata, 555\n"
|
||||||
@ -59,19 +69,24 @@ CSV_DATA = "Hello, 1\nWorld, 2\nThis, 152\nis, 9283\ntesting, 2313213\ndata, 555
|
|||||||
|
|
||||||
# Choose compression method
|
# Choose compression method
|
||||||
# (Will change during test, need to check standard data sending, to make sure that nothing broke)
|
# (Will change during test, need to check standard data sending, to make sure that nothing broke)
|
||||||
COMPRESS_METHOD = 'none'
|
COMPRESS_METHOD = "none"
|
||||||
ADDING_ENDING = ''
|
ADDING_ENDING = ""
|
||||||
ENDINGS = ['.gz', '.xz']
|
ENDINGS = [".gz", ".xz"]
|
||||||
SEND_ENCODING = True
|
SEND_ENCODING = True
|
||||||
|
|
||||||
|
|
||||||
def get_ch_answer(query):
|
def get_ch_answer(query):
|
||||||
host = CLICKHOUSE_HOST
|
host = CLICKHOUSE_HOST
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
host = f'[{host}]'
|
host = f"[{host}]"
|
||||||
|
|
||||||
url = os.environ.get('CLICKHOUSE_URL', 'http://{host}:{port}'.format(host=CLICKHOUSE_HOST, port=CLICKHOUSE_PORT_HTTP))
|
url = os.environ.get(
|
||||||
|
"CLICKHOUSE_URL",
|
||||||
|
"http://{host}:{port}".format(host=CLICKHOUSE_HOST, port=CLICKHOUSE_PORT_HTTP),
|
||||||
|
)
|
||||||
return urllib.request.urlopen(url, data=query.encode()).read().decode()
|
return urllib.request.urlopen(url, data=query.encode()).read().decode()
|
||||||
|
|
||||||
|
|
||||||
def check_answers(query, answer):
|
def check_answers(query, answer):
|
||||||
ch_answer = get_ch_answer(query)
|
ch_answer = get_ch_answer(query)
|
||||||
if ch_answer.strip() != answer.strip():
|
if ch_answer.strip() != answer.strip():
|
||||||
@ -80,18 +95,19 @@ def check_answers(query, answer):
|
|||||||
print("Fetched answer :", ch_answer, file=sys.stderr)
|
print("Fetched answer :", ch_answer, file=sys.stderr)
|
||||||
raise Exception("Fail on query")
|
raise Exception("Fail on query")
|
||||||
|
|
||||||
|
|
||||||
# Server with head method which is useful for debuging by hands
|
# Server with head method which is useful for debuging by hands
|
||||||
class HttpProcessor(SimpleHTTPRequestHandler):
|
class HttpProcessor(SimpleHTTPRequestHandler):
|
||||||
def _set_headers(self):
|
def _set_headers(self):
|
||||||
self.send_response(200)
|
self.send_response(200)
|
||||||
if SEND_ENCODING:
|
if SEND_ENCODING:
|
||||||
self.send_header('Content-Encoding', COMPRESS_METHOD)
|
self.send_header("Content-Encoding", COMPRESS_METHOD)
|
||||||
if COMPRESS_METHOD == 'none':
|
if COMPRESS_METHOD == "none":
|
||||||
self.send_header('Content-Length', len(CSV_DATA.encode()))
|
self.send_header("Content-Length", len(CSV_DATA.encode()))
|
||||||
else:
|
else:
|
||||||
self.compress_data()
|
self.compress_data()
|
||||||
self.send_header('Content-Length', len(self.data))
|
self.send_header("Content-Length", len(self.data))
|
||||||
self.send_header('Content-Type', 'text/csv')
|
self.send_header("Content-Type", "text/csv")
|
||||||
self.end_headers()
|
self.end_headers()
|
||||||
|
|
||||||
def do_HEAD(self):
|
def do_HEAD(self):
|
||||||
@ -99,18 +115,17 @@ class HttpProcessor(SimpleHTTPRequestHandler):
|
|||||||
return
|
return
|
||||||
|
|
||||||
def compress_data(self):
|
def compress_data(self):
|
||||||
if COMPRESS_METHOD == 'gzip':
|
if COMPRESS_METHOD == "gzip":
|
||||||
self.data = gzip.compress((CSV_DATA).encode())
|
self.data = gzip.compress((CSV_DATA).encode())
|
||||||
elif COMPRESS_METHOD == 'lzma':
|
elif COMPRESS_METHOD == "lzma":
|
||||||
self.data = lzma.compress((CSV_DATA).encode())
|
self.data = lzma.compress((CSV_DATA).encode())
|
||||||
else:
|
else:
|
||||||
self.data = 'WRONG CONVERSATION'.encode()
|
self.data = "WRONG CONVERSATION".encode()
|
||||||
|
|
||||||
|
|
||||||
def do_GET(self):
|
def do_GET(self):
|
||||||
self._set_headers()
|
self._set_headers()
|
||||||
|
|
||||||
if COMPRESS_METHOD == 'none':
|
if COMPRESS_METHOD == "none":
|
||||||
self.wfile.write(CSV_DATA.encode())
|
self.wfile.write(CSV_DATA.encode())
|
||||||
else:
|
else:
|
||||||
self.wfile.write(self.data)
|
self.wfile.write(self.data)
|
||||||
@ -119,9 +134,11 @@ class HttpProcessor(SimpleHTTPRequestHandler):
|
|||||||
def log_message(self, format, *args):
|
def log_message(self, format, *args):
|
||||||
return
|
return
|
||||||
|
|
||||||
|
|
||||||
class HTTPServerV6(HTTPServer):
|
class HTTPServerV6(HTTPServer):
|
||||||
address_family = socket.AF_INET6
|
address_family = socket.AF_INET6
|
||||||
|
|
||||||
|
|
||||||
def start_server(requests_amount):
|
def start_server(requests_amount):
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
httpd = HTTPServerV6(HTTP_SERVER_ADDRESS, HttpProcessor)
|
httpd = HTTPServerV6(HTTP_SERVER_ADDRESS, HttpProcessor)
|
||||||
@ -135,11 +152,19 @@ def start_server(requests_amount):
|
|||||||
t = threading.Thread(target=real_func)
|
t = threading.Thread(target=real_func)
|
||||||
return t
|
return t
|
||||||
|
|
||||||
|
|
||||||
#####################################################################
|
#####################################################################
|
||||||
# Testing area.
|
# Testing area.
|
||||||
#####################################################################
|
#####################################################################
|
||||||
|
|
||||||
def test_select(dict_name="", schema="word String, counter UInt32", requests=[], answers=[], test_data=""):
|
|
||||||
|
def test_select(
|
||||||
|
dict_name="",
|
||||||
|
schema="word String, counter UInt32",
|
||||||
|
requests=[],
|
||||||
|
answers=[],
|
||||||
|
test_data="",
|
||||||
|
):
|
||||||
global ADDING_ENDING
|
global ADDING_ENDING
|
||||||
global SEND_ENCODING
|
global SEND_ENCODING
|
||||||
global COMPRESS_METHOD
|
global COMPRESS_METHOD
|
||||||
@ -150,37 +175,37 @@ def test_select(dict_name="", schema="word String, counter UInt32", requests=[],
|
|||||||
|
|
||||||
if dict_name:
|
if dict_name:
|
||||||
get_ch_answer("drop dictionary if exists {}".format(dict_name))
|
get_ch_answer("drop dictionary if exists {}".format(dict_name))
|
||||||
get_ch_answer('''CREATE DICTIONARY {} ({})
|
get_ch_answer(
|
||||||
|
"""CREATE DICTIONARY {} ({})
|
||||||
PRIMARY KEY word
|
PRIMARY KEY word
|
||||||
SOURCE(HTTP(url '{}' format 'CSV'))
|
SOURCE(HTTP(url '{}' format 'CSV'))
|
||||||
LAYOUT(complex_key_hashed())
|
LAYOUT(complex_key_hashed())
|
||||||
LIFETIME(0)'''.format(dict_name, schema, HTTP_SERVER_URL_STR + '/test.csv' + ADDING_ENDING))
|
LIFETIME(0)""".format(
|
||||||
|
dict_name, schema, HTTP_SERVER_URL_STR + "/test.csv" + ADDING_ENDING
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
COMPRESS_METHOD = requests[i]
|
COMPRESS_METHOD = requests[i]
|
||||||
print(i, COMPRESS_METHOD, ADDING_ENDING, SEND_ENCODING)
|
print(i, COMPRESS_METHOD, ADDING_ENDING, SEND_ENCODING)
|
||||||
check_answers("SELECT * FROM {} ORDER BY word".format(dict_name), answers[i])
|
check_answers("SELECT * FROM {} ORDER BY word".format(dict_name), answers[i])
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
# first three for encoding, second three for url
|
# first three for encoding, second three for url
|
||||||
insert_requests = [
|
insert_requests = ["none", "gzip", "lzma", "gzip", "lzma"]
|
||||||
'none',
|
|
||||||
'gzip',
|
|
||||||
'lzma',
|
|
||||||
'gzip',
|
|
||||||
'lzma'
|
|
||||||
]
|
|
||||||
|
|
||||||
# This answers got experemently in non compressed mode and they are correct
|
# This answers got experemently in non compressed mode and they are correct
|
||||||
answers = ['''Hello 1\nThis 152\nWorld 2\ndata 555\nis 9283\ntesting 2313213'''] * 5
|
answers = ["""Hello 1\nThis 152\nWorld 2\ndata 555\nis 9283\ntesting 2313213"""] * 5
|
||||||
|
|
||||||
t = start_server(len(insert_requests))
|
t = start_server(len(insert_requests))
|
||||||
t.start()
|
t.start()
|
||||||
test_select(dict_name="test_table_select", requests=insert_requests, answers=answers)
|
test_select(
|
||||||
|
dict_name="test_table_select", requests=insert_requests, answers=answers
|
||||||
|
)
|
||||||
t.join()
|
t.join()
|
||||||
print("PASSED")
|
print("PASSED")
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
try:
|
try:
|
||||||
main()
|
main()
|
||||||
@ -191,5 +216,3 @@ if __name__ == "__main__":
|
|||||||
sys.stderr.flush()
|
sys.stderr.flush()
|
||||||
|
|
||||||
os._exit(1)
|
os._exit(1)
|
||||||
|
|
||||||
|
|
||||||
|
@ -5,9 +5,10 @@ import socket
|
|||||||
import os
|
import os
|
||||||
import uuid
|
import uuid
|
||||||
|
|
||||||
CLICKHOUSE_HOST = os.environ.get('CLICKHOUSE_HOST', '127.0.0.1')
|
CLICKHOUSE_HOST = os.environ.get("CLICKHOUSE_HOST", "127.0.0.1")
|
||||||
CLICKHOUSE_PORT = int(os.environ.get('CLICKHOUSE_PORT_TCP', '900000'))
|
CLICKHOUSE_PORT = int(os.environ.get("CLICKHOUSE_PORT_TCP", "900000"))
|
||||||
CLICKHOUSE_DATABASE = os.environ.get('CLICKHOUSE_DATABASE', 'default')
|
CLICKHOUSE_DATABASE = os.environ.get("CLICKHOUSE_DATABASE", "default")
|
||||||
|
|
||||||
|
|
||||||
def writeVarUInt(x, ba):
|
def writeVarUInt(x, ba):
|
||||||
for _ in range(0, 9):
|
for _ in range(0, 9):
|
||||||
@ -24,7 +25,7 @@ def writeVarUInt(x, ba):
|
|||||||
|
|
||||||
|
|
||||||
def writeStringBinary(s, ba):
|
def writeStringBinary(s, ba):
|
||||||
b = bytes(s, 'utf-8')
|
b = bytes(s, "utf-8")
|
||||||
writeVarUInt(len(s), ba)
|
writeVarUInt(len(s), ba)
|
||||||
ba.extend(b)
|
ba.extend(b)
|
||||||
|
|
||||||
@ -48,18 +49,23 @@ def readUInt(s, size=1):
|
|||||||
val += res[i] << (i * 8)
|
val += res[i] << (i * 8)
|
||||||
return val
|
return val
|
||||||
|
|
||||||
|
|
||||||
def readUInt8(s):
|
def readUInt8(s):
|
||||||
return readUInt(s)
|
return readUInt(s)
|
||||||
|
|
||||||
|
|
||||||
def readUInt16(s):
|
def readUInt16(s):
|
||||||
return readUInt(s, 2)
|
return readUInt(s, 2)
|
||||||
|
|
||||||
|
|
||||||
def readUInt32(s):
|
def readUInt32(s):
|
||||||
return readUInt(s, 4)
|
return readUInt(s, 4)
|
||||||
|
|
||||||
|
|
||||||
def readUInt64(s):
|
def readUInt64(s):
|
||||||
return readUInt(s, 8)
|
return readUInt(s, 8)
|
||||||
|
|
||||||
|
|
||||||
def readVarUInt(s):
|
def readVarUInt(s):
|
||||||
x = 0
|
x = 0
|
||||||
for i in range(9):
|
for i in range(9):
|
||||||
@ -75,25 +81,25 @@ def readVarUInt(s):
|
|||||||
def readStringBinary(s):
|
def readStringBinary(s):
|
||||||
size = readVarUInt(s)
|
size = readVarUInt(s)
|
||||||
s = readStrict(s, size)
|
s = readStrict(s, size)
|
||||||
return s.decode('utf-8')
|
return s.decode("utf-8")
|
||||||
|
|
||||||
|
|
||||||
def sendHello(s):
|
def sendHello(s):
|
||||||
ba = bytearray()
|
ba = bytearray()
|
||||||
writeVarUInt(0, ba) # Hello
|
writeVarUInt(0, ba) # Hello
|
||||||
writeStringBinary('simple native protocol', ba)
|
writeStringBinary("simple native protocol", ba)
|
||||||
writeVarUInt(21, ba)
|
writeVarUInt(21, ba)
|
||||||
writeVarUInt(9, ba)
|
writeVarUInt(9, ba)
|
||||||
writeVarUInt(54449, ba)
|
writeVarUInt(54449, ba)
|
||||||
writeStringBinary('default', ba) # database
|
writeStringBinary("default", ba) # database
|
||||||
writeStringBinary('default', ba) # user
|
writeStringBinary("default", ba) # user
|
||||||
writeStringBinary('', ba) # pwd
|
writeStringBinary("", ba) # pwd
|
||||||
s.sendall(ba)
|
s.sendall(ba)
|
||||||
|
|
||||||
|
|
||||||
def receiveHello(s):
|
def receiveHello(s):
|
||||||
p_type = readVarUInt(s)
|
p_type = readVarUInt(s)
|
||||||
assert (p_type == 0) # Hello
|
assert p_type == 0 # Hello
|
||||||
server_name = readStringBinary(s)
|
server_name = readStringBinary(s)
|
||||||
# print("Server name: ", server_name)
|
# print("Server name: ", server_name)
|
||||||
server_version_major = readVarUInt(s)
|
server_version_major = readVarUInt(s)
|
||||||
@ -111,18 +117,18 @@ def receiveHello(s):
|
|||||||
|
|
||||||
|
|
||||||
def serializeClientInfo(ba, query_id):
|
def serializeClientInfo(ba, query_id):
|
||||||
writeStringBinary('default', ba) # initial_user
|
writeStringBinary("default", ba) # initial_user
|
||||||
writeStringBinary(query_id, ba) # initial_query_id
|
writeStringBinary(query_id, ba) # initial_query_id
|
||||||
writeStringBinary('127.0.0.1:9000', ba) # initial_address
|
writeStringBinary("127.0.0.1:9000", ba) # initial_address
|
||||||
ba.extend([0] * 8) # initial_query_start_time_microseconds
|
ba.extend([0] * 8) # initial_query_start_time_microseconds
|
||||||
ba.append(1) # TCP
|
ba.append(1) # TCP
|
||||||
writeStringBinary('os_user', ba) # os_user
|
writeStringBinary("os_user", ba) # os_user
|
||||||
writeStringBinary('client_hostname', ba) # client_hostname
|
writeStringBinary("client_hostname", ba) # client_hostname
|
||||||
writeStringBinary('client_name', ba) # client_name
|
writeStringBinary("client_name", ba) # client_name
|
||||||
writeVarUInt(21, ba)
|
writeVarUInt(21, ba)
|
||||||
writeVarUInt(9, ba)
|
writeVarUInt(9, ba)
|
||||||
writeVarUInt(54449, ba)
|
writeVarUInt(54449, ba)
|
||||||
writeStringBinary('', ba) # quota_key
|
writeStringBinary("", ba) # quota_key
|
||||||
writeVarUInt(0, ba) # distributed_depth
|
writeVarUInt(0, ba) # distributed_depth
|
||||||
writeVarUInt(1, ba) # client_version_patch
|
writeVarUInt(1, ba) # client_version_patch
|
||||||
ba.append(0) # No telemetry
|
ba.append(0) # No telemetry
|
||||||
@ -139,8 +145,8 @@ def sendQuery(s, query):
|
|||||||
# client info
|
# client info
|
||||||
serializeClientInfo(ba, query_id)
|
serializeClientInfo(ba, query_id)
|
||||||
|
|
||||||
writeStringBinary('', ba) # No settings
|
writeStringBinary("", ba) # No settings
|
||||||
writeStringBinary('', ba) # No interserver secret
|
writeStringBinary("", ba) # No interserver secret
|
||||||
writeVarUInt(2, ba) # Stage - Complete
|
writeVarUInt(2, ba) # Stage - Complete
|
||||||
ba.append(0) # No compression
|
ba.append(0) # No compression
|
||||||
writeStringBinary(query, ba) # query, finally
|
writeStringBinary(query, ba) # query, finally
|
||||||
@ -158,7 +164,7 @@ def serializeBlockInfo(ba):
|
|||||||
def sendEmptyBlock(s):
|
def sendEmptyBlock(s):
|
||||||
ba = bytearray()
|
ba = bytearray()
|
||||||
writeVarUInt(2, ba) # Data
|
writeVarUInt(2, ba) # Data
|
||||||
writeStringBinary('', ba)
|
writeStringBinary("", ba)
|
||||||
serializeBlockInfo(ba)
|
serializeBlockInfo(ba)
|
||||||
writeVarUInt(0, ba) # rows
|
writeVarUInt(0, ba) # rows
|
||||||
writeVarUInt(0, ba) # columns
|
writeVarUInt(0, ba) # columns
|
||||||
@ -166,7 +172,8 @@ def sendEmptyBlock(s):
|
|||||||
|
|
||||||
|
|
||||||
def assertPacket(packet, expected):
|
def assertPacket(packet, expected):
|
||||||
assert(packet == expected), packet
|
assert packet == expected, packet
|
||||||
|
|
||||||
|
|
||||||
def readHeader(s):
|
def readHeader(s):
|
||||||
packet_type = readVarUInt(s)
|
packet_type = readVarUInt(s)
|
||||||
@ -196,7 +203,7 @@ def readException(s):
|
|||||||
text = readStringBinary(s)
|
text = readStringBinary(s)
|
||||||
readStringBinary(s) # trace
|
readStringBinary(s) # trace
|
||||||
assertPacket(readUInt8(s), 0) # has_nested
|
assertPacket(readUInt8(s), 0) # has_nested
|
||||||
return "code {}: {}".format(code, text.replace('DB::Exception:', ''))
|
return "code {}: {}".format(code, text.replace("DB::Exception:", ""))
|
||||||
|
|
||||||
|
|
||||||
def insertValidLowCardinalityRow():
|
def insertValidLowCardinalityRow():
|
||||||
@ -205,7 +212,12 @@ def insertValidLowCardinalityRow():
|
|||||||
s.connect((CLICKHOUSE_HOST, CLICKHOUSE_PORT))
|
s.connect((CLICKHOUSE_HOST, CLICKHOUSE_PORT))
|
||||||
sendHello(s)
|
sendHello(s)
|
||||||
receiveHello(s)
|
receiveHello(s)
|
||||||
sendQuery(s, 'insert into {}.tab settings input_format_defaults_for_omitted_fields=0 format TSV'.format(CLICKHOUSE_DATABASE))
|
sendQuery(
|
||||||
|
s,
|
||||||
|
"insert into {}.tab settings input_format_defaults_for_omitted_fields=0 format TSV".format(
|
||||||
|
CLICKHOUSE_DATABASE
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
# external tables
|
# external tables
|
||||||
sendEmptyBlock(s)
|
sendEmptyBlock(s)
|
||||||
@ -214,16 +226,18 @@ def insertValidLowCardinalityRow():
|
|||||||
# Data
|
# Data
|
||||||
ba = bytearray()
|
ba = bytearray()
|
||||||
writeVarUInt(2, ba) # Data
|
writeVarUInt(2, ba) # Data
|
||||||
writeStringBinary('', ba)
|
writeStringBinary("", ba)
|
||||||
serializeBlockInfo(ba)
|
serializeBlockInfo(ba)
|
||||||
writeVarUInt(1, ba) # rows
|
writeVarUInt(1, ba) # rows
|
||||||
writeVarUInt(1, ba) # columns
|
writeVarUInt(1, ba) # columns
|
||||||
writeStringBinary('x', ba)
|
writeStringBinary("x", ba)
|
||||||
writeStringBinary('LowCardinality(String)', ba)
|
writeStringBinary("LowCardinality(String)", ba)
|
||||||
ba.extend([1] + [0] * 7) # SharedDictionariesWithAdditionalKeys
|
ba.extend([1] + [0] * 7) # SharedDictionariesWithAdditionalKeys
|
||||||
ba.extend([3, 2] + [0] * 6) # indexes type: UInt64 [3], with additional keys [2]
|
ba.extend(
|
||||||
|
[3, 2] + [0] * 6
|
||||||
|
) # indexes type: UInt64 [3], with additional keys [2]
|
||||||
ba.extend([1] + [0] * 7) # num_keys in dict
|
ba.extend([1] + [0] * 7) # num_keys in dict
|
||||||
writeStringBinary('hello', ba) # key
|
writeStringBinary("hello", ba) # key
|
||||||
ba.extend([1] + [0] * 7) # num_indexes
|
ba.extend([1] + [0] * 7) # num_indexes
|
||||||
ba.extend([0] * 8) # UInt64 index (0 for 'hello')
|
ba.extend([0] * 8) # UInt64 index (0 for 'hello')
|
||||||
s.sendall(ba)
|
s.sendall(ba)
|
||||||
@ -241,7 +255,12 @@ def insertLowCardinalityRowWithIndexOverflow():
|
|||||||
s.connect((CLICKHOUSE_HOST, CLICKHOUSE_PORT))
|
s.connect((CLICKHOUSE_HOST, CLICKHOUSE_PORT))
|
||||||
sendHello(s)
|
sendHello(s)
|
||||||
receiveHello(s)
|
receiveHello(s)
|
||||||
sendQuery(s, 'insert into {}.tab settings input_format_defaults_for_omitted_fields=0 format TSV'.format(CLICKHOUSE_DATABASE))
|
sendQuery(
|
||||||
|
s,
|
||||||
|
"insert into {}.tab settings input_format_defaults_for_omitted_fields=0 format TSV".format(
|
||||||
|
CLICKHOUSE_DATABASE
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
# external tables
|
# external tables
|
||||||
sendEmptyBlock(s)
|
sendEmptyBlock(s)
|
||||||
@ -250,16 +269,18 @@ def insertLowCardinalityRowWithIndexOverflow():
|
|||||||
# Data
|
# Data
|
||||||
ba = bytearray()
|
ba = bytearray()
|
||||||
writeVarUInt(2, ba) # Data
|
writeVarUInt(2, ba) # Data
|
||||||
writeStringBinary('', ba)
|
writeStringBinary("", ba)
|
||||||
serializeBlockInfo(ba)
|
serializeBlockInfo(ba)
|
||||||
writeVarUInt(1, ba) # rows
|
writeVarUInt(1, ba) # rows
|
||||||
writeVarUInt(1, ba) # columns
|
writeVarUInt(1, ba) # columns
|
||||||
writeStringBinary('x', ba)
|
writeStringBinary("x", ba)
|
||||||
writeStringBinary('LowCardinality(String)', ba)
|
writeStringBinary("LowCardinality(String)", ba)
|
||||||
ba.extend([1] + [0] * 7) # SharedDictionariesWithAdditionalKeys
|
ba.extend([1] + [0] * 7) # SharedDictionariesWithAdditionalKeys
|
||||||
ba.extend([3, 2] + [0] * 6) # indexes type: UInt64 [3], with additional keys [2]
|
ba.extend(
|
||||||
|
[3, 2] + [0] * 6
|
||||||
|
) # indexes type: UInt64 [3], with additional keys [2]
|
||||||
ba.extend([1] + [0] * 7) # num_keys in dict
|
ba.extend([1] + [0] * 7) # num_keys in dict
|
||||||
writeStringBinary('hello', ba) # key
|
writeStringBinary("hello", ba) # key
|
||||||
ba.extend([1] + [0] * 7) # num_indexes
|
ba.extend([1] + [0] * 7) # num_indexes
|
||||||
ba.extend([0] * 7 + [1]) # UInt64 index (overflow)
|
ba.extend([0] * 7 + [1]) # UInt64 index (overflow)
|
||||||
s.sendall(ba)
|
s.sendall(ba)
|
||||||
@ -275,7 +296,12 @@ def insertLowCardinalityRowWithIncorrectDictType():
|
|||||||
s.connect((CLICKHOUSE_HOST, CLICKHOUSE_PORT))
|
s.connect((CLICKHOUSE_HOST, CLICKHOUSE_PORT))
|
||||||
sendHello(s)
|
sendHello(s)
|
||||||
receiveHello(s)
|
receiveHello(s)
|
||||||
sendQuery(s, 'insert into {}.tab settings input_format_defaults_for_omitted_fields=0 format TSV'.format(CLICKHOUSE_DATABASE))
|
sendQuery(
|
||||||
|
s,
|
||||||
|
"insert into {}.tab settings input_format_defaults_for_omitted_fields=0 format TSV".format(
|
||||||
|
CLICKHOUSE_DATABASE
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
# external tables
|
# external tables
|
||||||
sendEmptyBlock(s)
|
sendEmptyBlock(s)
|
||||||
@ -284,16 +310,18 @@ def insertLowCardinalityRowWithIncorrectDictType():
|
|||||||
# Data
|
# Data
|
||||||
ba = bytearray()
|
ba = bytearray()
|
||||||
writeVarUInt(2, ba) # Data
|
writeVarUInt(2, ba) # Data
|
||||||
writeStringBinary('', ba)
|
writeStringBinary("", ba)
|
||||||
serializeBlockInfo(ba)
|
serializeBlockInfo(ba)
|
||||||
writeVarUInt(1, ba) # rows
|
writeVarUInt(1, ba) # rows
|
||||||
writeVarUInt(1, ba) # columns
|
writeVarUInt(1, ba) # columns
|
||||||
writeStringBinary('x', ba)
|
writeStringBinary("x", ba)
|
||||||
writeStringBinary('LowCardinality(String)', ba)
|
writeStringBinary("LowCardinality(String)", ba)
|
||||||
ba.extend([1] + [0] * 7) # SharedDictionariesWithAdditionalKeys
|
ba.extend([1] + [0] * 7) # SharedDictionariesWithAdditionalKeys
|
||||||
ba.extend([3, 3] + [0] * 6) # indexes type: UInt64 [3], with global dict and add keys [1 + 2]
|
ba.extend(
|
||||||
|
[3, 3] + [0] * 6
|
||||||
|
) # indexes type: UInt64 [3], with global dict and add keys [1 + 2]
|
||||||
ba.extend([1] + [0] * 7) # num_keys in dict
|
ba.extend([1] + [0] * 7) # num_keys in dict
|
||||||
writeStringBinary('hello', ba) # key
|
writeStringBinary("hello", ba) # key
|
||||||
ba.extend([1] + [0] * 7) # num_indexes
|
ba.extend([1] + [0] * 7) # num_indexes
|
||||||
ba.extend([0] * 8) # UInt64 index (overflow)
|
ba.extend([0] * 8) # UInt64 index (overflow)
|
||||||
s.sendall(ba)
|
s.sendall(ba)
|
||||||
@ -302,13 +330,19 @@ def insertLowCardinalityRowWithIncorrectDictType():
|
|||||||
print(readException(s))
|
print(readException(s))
|
||||||
s.close()
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
def insertLowCardinalityRowWithIncorrectAdditionalKeys():
|
def insertLowCardinalityRowWithIncorrectAdditionalKeys():
|
||||||
with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s:
|
with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s:
|
||||||
s.settimeout(30)
|
s.settimeout(30)
|
||||||
s.connect((CLICKHOUSE_HOST, CLICKHOUSE_PORT))
|
s.connect((CLICKHOUSE_HOST, CLICKHOUSE_PORT))
|
||||||
sendHello(s)
|
sendHello(s)
|
||||||
receiveHello(s)
|
receiveHello(s)
|
||||||
sendQuery(s, 'insert into {}.tab settings input_format_defaults_for_omitted_fields=0 format TSV'.format(CLICKHOUSE_DATABASE))
|
sendQuery(
|
||||||
|
s,
|
||||||
|
"insert into {}.tab settings input_format_defaults_for_omitted_fields=0 format TSV".format(
|
||||||
|
CLICKHOUSE_DATABASE
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
# external tables
|
# external tables
|
||||||
sendEmptyBlock(s)
|
sendEmptyBlock(s)
|
||||||
@ -317,16 +351,18 @@ def insertLowCardinalityRowWithIncorrectAdditionalKeys():
|
|||||||
# Data
|
# Data
|
||||||
ba = bytearray()
|
ba = bytearray()
|
||||||
writeVarUInt(2, ba) # Data
|
writeVarUInt(2, ba) # Data
|
||||||
writeStringBinary('', ba)
|
writeStringBinary("", ba)
|
||||||
serializeBlockInfo(ba)
|
serializeBlockInfo(ba)
|
||||||
writeVarUInt(1, ba) # rows
|
writeVarUInt(1, ba) # rows
|
||||||
writeVarUInt(1, ba) # columns
|
writeVarUInt(1, ba) # columns
|
||||||
writeStringBinary('x', ba)
|
writeStringBinary("x", ba)
|
||||||
writeStringBinary('LowCardinality(String)', ba)
|
writeStringBinary("LowCardinality(String)", ba)
|
||||||
ba.extend([1] + [0] * 7) # SharedDictionariesWithAdditionalKeys
|
ba.extend([1] + [0] * 7) # SharedDictionariesWithAdditionalKeys
|
||||||
ba.extend([3, 0] + [0] * 6) # indexes type: UInt64 [3], with NO additional keys [0]
|
ba.extend(
|
||||||
|
[3, 0] + [0] * 6
|
||||||
|
) # indexes type: UInt64 [3], with NO additional keys [0]
|
||||||
ba.extend([1] + [0] * 7) # num_keys in dict
|
ba.extend([1] + [0] * 7) # num_keys in dict
|
||||||
writeStringBinary('hello', ba) # key
|
writeStringBinary("hello", ba) # key
|
||||||
ba.extend([1] + [0] * 7) # num_indexes
|
ba.extend([1] + [0] * 7) # num_indexes
|
||||||
ba.extend([0] * 8) # UInt64 index (0 for 'hello')
|
ba.extend([0] * 8) # UInt64 index (0 for 'hello')
|
||||||
s.sendall(ba)
|
s.sendall(ba)
|
||||||
@ -335,11 +371,13 @@ def insertLowCardinalityRowWithIncorrectAdditionalKeys():
|
|||||||
print(readException(s))
|
print(readException(s))
|
||||||
s.close()
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
insertValidLowCardinalityRow()
|
insertValidLowCardinalityRow()
|
||||||
insertLowCardinalityRowWithIndexOverflow()
|
insertLowCardinalityRowWithIndexOverflow()
|
||||||
insertLowCardinalityRowWithIncorrectDictType()
|
insertLowCardinalityRowWithIncorrectDictType()
|
||||||
insertLowCardinalityRowWithIncorrectAdditionalKeys()
|
insertLowCardinalityRowWithIncorrectAdditionalKeys()
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
main()
|
main()
|
||||||
|
@ -12,6 +12,7 @@ import subprocess
|
|||||||
from io import StringIO
|
from io import StringIO
|
||||||
from http.server import BaseHTTPRequestHandler, HTTPServer
|
from http.server import BaseHTTPRequestHandler, HTTPServer
|
||||||
|
|
||||||
|
|
||||||
def is_ipv6(host):
|
def is_ipv6(host):
|
||||||
try:
|
try:
|
||||||
socket.inet_aton(host)
|
socket.inet_aton(host)
|
||||||
@ -19,6 +20,7 @@ def is_ipv6(host):
|
|||||||
except:
|
except:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
def get_local_port(host, ipv6):
|
def get_local_port(host, ipv6):
|
||||||
if ipv6:
|
if ipv6:
|
||||||
family = socket.AF_INET6
|
family = socket.AF_INET6
|
||||||
@ -29,8 +31,9 @@ def get_local_port(host, ipv6):
|
|||||||
fd.bind((host, 0))
|
fd.bind((host, 0))
|
||||||
return fd.getsockname()[1]
|
return fd.getsockname()[1]
|
||||||
|
|
||||||
CLICKHOUSE_HOST = os.environ.get('CLICKHOUSE_HOST', '127.0.0.1')
|
|
||||||
CLICKHOUSE_PORT_HTTP = os.environ.get('CLICKHOUSE_PORT_HTTP', '8123')
|
CLICKHOUSE_HOST = os.environ.get("CLICKHOUSE_HOST", "127.0.0.1")
|
||||||
|
CLICKHOUSE_PORT_HTTP = os.environ.get("CLICKHOUSE_PORT_HTTP", "8123")
|
||||||
|
|
||||||
#####################################################################################
|
#####################################################################################
|
||||||
# This test starts an HTTP server and serves data to clickhouse url-engine based table.
|
# This test starts an HTTP server and serves data to clickhouse url-engine based table.
|
||||||
@ -39,27 +42,42 @@ CLICKHOUSE_PORT_HTTP = os.environ.get('CLICKHOUSE_PORT_HTTP', '8123')
|
|||||||
#####################################################################################
|
#####################################################################################
|
||||||
|
|
||||||
# IP-address of this host accessible from the outside world. Get the first one
|
# IP-address of this host accessible from the outside world. Get the first one
|
||||||
HTTP_SERVER_HOST = subprocess.check_output(['hostname', '-i']).decode('utf-8').strip().split()[0]
|
HTTP_SERVER_HOST = (
|
||||||
|
subprocess.check_output(["hostname", "-i"]).decode("utf-8").strip().split()[0]
|
||||||
|
)
|
||||||
IS_IPV6 = is_ipv6(HTTP_SERVER_HOST)
|
IS_IPV6 = is_ipv6(HTTP_SERVER_HOST)
|
||||||
HTTP_SERVER_PORT = get_local_port(HTTP_SERVER_HOST, IS_IPV6)
|
HTTP_SERVER_PORT = get_local_port(HTTP_SERVER_HOST, IS_IPV6)
|
||||||
|
|
||||||
# IP address and port of the HTTP server started from this script.
|
# IP address and port of the HTTP server started from this script.
|
||||||
HTTP_SERVER_ADDRESS = (HTTP_SERVER_HOST, HTTP_SERVER_PORT)
|
HTTP_SERVER_ADDRESS = (HTTP_SERVER_HOST, HTTP_SERVER_PORT)
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
HTTP_SERVER_URL_STR = 'http://' + f'[{str(HTTP_SERVER_ADDRESS[0])}]:{str(HTTP_SERVER_ADDRESS[1])}' + "/"
|
HTTP_SERVER_URL_STR = (
|
||||||
|
"http://"
|
||||||
|
+ f"[{str(HTTP_SERVER_ADDRESS[0])}]:{str(HTTP_SERVER_ADDRESS[1])}"
|
||||||
|
+ "/"
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
HTTP_SERVER_URL_STR = 'http://' + f'{str(HTTP_SERVER_ADDRESS[0])}:{str(HTTP_SERVER_ADDRESS[1])}' + "/"
|
HTTP_SERVER_URL_STR = (
|
||||||
|
"http://" + f"{str(HTTP_SERVER_ADDRESS[0])}:{str(HTTP_SERVER_ADDRESS[1])}" + "/"
|
||||||
|
)
|
||||||
|
|
||||||
|
CSV_DATA = os.path.join(
|
||||||
|
tempfile._get_default_tempdir(), next(tempfile._get_candidate_names())
|
||||||
|
)
|
||||||
|
|
||||||
CSV_DATA = os.path.join(tempfile._get_default_tempdir(), next(tempfile._get_candidate_names()))
|
|
||||||
|
|
||||||
def get_ch_answer(query):
|
def get_ch_answer(query):
|
||||||
host = CLICKHOUSE_HOST
|
host = CLICKHOUSE_HOST
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
host = f'[{host}]'
|
host = f"[{host}]"
|
||||||
|
|
||||||
url = os.environ.get('CLICKHOUSE_URL', 'http://{host}:{port}'.format(host=CLICKHOUSE_HOST, port=CLICKHOUSE_PORT_HTTP))
|
url = os.environ.get(
|
||||||
|
"CLICKHOUSE_URL",
|
||||||
|
"http://{host}:{port}".format(host=CLICKHOUSE_HOST, port=CLICKHOUSE_PORT_HTTP),
|
||||||
|
)
|
||||||
return urllib.request.urlopen(url, data=query.encode()).read().decode()
|
return urllib.request.urlopen(url, data=query.encode()).read().decode()
|
||||||
|
|
||||||
|
|
||||||
def check_answers(query, answer):
|
def check_answers(query, answer):
|
||||||
ch_answer = get_ch_answer(query)
|
ch_answer = get_ch_answer(query)
|
||||||
if ch_answer.strip() != answer.strip():
|
if ch_answer.strip() != answer.strip():
|
||||||
@ -68,15 +86,16 @@ def check_answers(query, answer):
|
|||||||
print("Fetched answer :", ch_answer, file=sys.stderr)
|
print("Fetched answer :", ch_answer, file=sys.stderr)
|
||||||
raise Exception("Fail on query")
|
raise Exception("Fail on query")
|
||||||
|
|
||||||
|
|
||||||
class CSVHTTPServer(BaseHTTPRequestHandler):
|
class CSVHTTPServer(BaseHTTPRequestHandler):
|
||||||
def _set_headers(self):
|
def _set_headers(self):
|
||||||
self.send_response(200)
|
self.send_response(200)
|
||||||
self.send_header('Content-type', 'text/csv')
|
self.send_header("Content-type", "text/csv")
|
||||||
self.end_headers()
|
self.end_headers()
|
||||||
|
|
||||||
def do_GET(self):
|
def do_GET(self):
|
||||||
self._set_headers()
|
self._set_headers()
|
||||||
self.wfile.write(('hello, world').encode())
|
self.wfile.write(("hello, world").encode())
|
||||||
# with open(CSV_DATA, 'r') as fl:
|
# with open(CSV_DATA, 'r') as fl:
|
||||||
# reader = csv.reader(fl, delimiter=',')
|
# reader = csv.reader(fl, delimiter=',')
|
||||||
# for row in reader:
|
# for row in reader:
|
||||||
@ -84,33 +103,33 @@ class CSVHTTPServer(BaseHTTPRequestHandler):
|
|||||||
return
|
return
|
||||||
|
|
||||||
def read_chunk(self):
|
def read_chunk(self):
|
||||||
msg = ''
|
msg = ""
|
||||||
while True:
|
while True:
|
||||||
sym = self.rfile.read(1)
|
sym = self.rfile.read(1)
|
||||||
if sym == '':
|
if sym == "":
|
||||||
break
|
break
|
||||||
msg += sym.decode('utf-8')
|
msg += sym.decode("utf-8")
|
||||||
if msg.endswith('\r\n'):
|
if msg.endswith("\r\n"):
|
||||||
break
|
break
|
||||||
length = int(msg[:-2], 16)
|
length = int(msg[:-2], 16)
|
||||||
if length == 0:
|
if length == 0:
|
||||||
return ''
|
return ""
|
||||||
content = self.rfile.read(length)
|
content = self.rfile.read(length)
|
||||||
self.rfile.read(2) # read sep \r\n
|
self.rfile.read(2) # read sep \r\n
|
||||||
return content.decode('utf-8')
|
return content.decode("utf-8")
|
||||||
|
|
||||||
def do_POST(self):
|
def do_POST(self):
|
||||||
data = ''
|
data = ""
|
||||||
while True:
|
while True:
|
||||||
chunk = self.read_chunk()
|
chunk = self.read_chunk()
|
||||||
if not chunk:
|
if not chunk:
|
||||||
break
|
break
|
||||||
data += chunk
|
data += chunk
|
||||||
with StringIO(data) as fl:
|
with StringIO(data) as fl:
|
||||||
reader = csv.reader(fl, delimiter=',')
|
reader = csv.reader(fl, delimiter=",")
|
||||||
with open(CSV_DATA, 'a') as d:
|
with open(CSV_DATA, "a") as d:
|
||||||
for row in reader:
|
for row in reader:
|
||||||
d.write(','.join(row) + '\n')
|
d.write(",".join(row) + "\n")
|
||||||
self._set_headers()
|
self._set_headers()
|
||||||
self.wfile.write(b"ok")
|
self.wfile.write(b"ok")
|
||||||
|
|
||||||
@ -121,6 +140,7 @@ class CSVHTTPServer(BaseHTTPRequestHandler):
|
|||||||
class HTTPServerV6(HTTPServer):
|
class HTTPServerV6(HTTPServer):
|
||||||
address_family = socket.AF_INET6
|
address_family = socket.AF_INET6
|
||||||
|
|
||||||
|
|
||||||
def start_server():
|
def start_server():
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
httpd = HTTPServerV6(HTTP_SERVER_ADDRESS, CSVHTTPServer)
|
httpd = HTTPServerV6(HTTP_SERVER_ADDRESS, CSVHTTPServer)
|
||||||
@ -130,57 +150,87 @@ def start_server():
|
|||||||
t = threading.Thread(target=httpd.serve_forever)
|
t = threading.Thread(target=httpd.serve_forever)
|
||||||
return t, httpd
|
return t, httpd
|
||||||
|
|
||||||
|
|
||||||
# test section
|
# test section
|
||||||
|
|
||||||
def test_select(table_name="", schema="str String,numuint UInt32,numint Int32,double Float64", requests=[], answers=[], test_data=""):
|
|
||||||
with open(CSV_DATA, 'w') as f: # clear file
|
def test_select(
|
||||||
f.write('')
|
table_name="",
|
||||||
|
schema="str String,numuint UInt32,numint Int32,double Float64",
|
||||||
|
requests=[],
|
||||||
|
answers=[],
|
||||||
|
test_data="",
|
||||||
|
):
|
||||||
|
with open(CSV_DATA, "w") as f: # clear file
|
||||||
|
f.write("")
|
||||||
|
|
||||||
if test_data:
|
if test_data:
|
||||||
with open(CSV_DATA, 'w') as f:
|
with open(CSV_DATA, "w") as f:
|
||||||
f.write(test_data + "\n")
|
f.write(test_data + "\n")
|
||||||
|
|
||||||
if table_name:
|
if table_name:
|
||||||
get_ch_answer("drop table if exists {}".format(table_name))
|
get_ch_answer("drop table if exists {}".format(table_name))
|
||||||
get_ch_answer("create table {} ({}) engine=URL('{}', 'CSV')".format(table_name, schema, HTTP_SERVER_URL_STR))
|
get_ch_answer(
|
||||||
|
"create table {} ({}) engine=URL('{}', 'CSV')".format(
|
||||||
|
table_name, schema, HTTP_SERVER_URL_STR
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
for i in range(len(requests)):
|
for i in range(len(requests)):
|
||||||
tbl = table_name
|
tbl = table_name
|
||||||
if not tbl:
|
if not tbl:
|
||||||
tbl = "url('{addr}', 'CSV', '{schema}')".format(addr=HTTP_SERVER_URL_STR, schema=schema)
|
tbl = "url('{addr}', 'CSV', '{schema}')".format(
|
||||||
|
addr=HTTP_SERVER_URL_STR, schema=schema
|
||||||
|
)
|
||||||
check_answers(requests[i].format(tbl=tbl), answers[i])
|
check_answers(requests[i].format(tbl=tbl), answers[i])
|
||||||
|
|
||||||
if table_name:
|
if table_name:
|
||||||
get_ch_answer("drop table if exists {}".format(table_name))
|
get_ch_answer("drop table if exists {}".format(table_name))
|
||||||
|
|
||||||
def test_insert(table_name="", schema="str String,numuint UInt32,numint Int32,double Float64", requests_insert=[], requests_select=[], answers=[]):
|
|
||||||
with open(CSV_DATA, 'w') as f: # flush test file
|
def test_insert(
|
||||||
f.write('')
|
table_name="",
|
||||||
|
schema="str String,numuint UInt32,numint Int32,double Float64",
|
||||||
|
requests_insert=[],
|
||||||
|
requests_select=[],
|
||||||
|
answers=[],
|
||||||
|
):
|
||||||
|
with open(CSV_DATA, "w") as f: # flush test file
|
||||||
|
f.write("")
|
||||||
|
|
||||||
if table_name:
|
if table_name:
|
||||||
get_ch_answer("drop table if exists {}".format(table_name))
|
get_ch_answer("drop table if exists {}".format(table_name))
|
||||||
get_ch_answer("create table {} ({}) engine=URL('{}', 'CSV')".format(table_name, schema, HTTP_SERVER_URL_STR))
|
get_ch_answer(
|
||||||
|
"create table {} ({}) engine=URL('{}', 'CSV')".format(
|
||||||
|
table_name, schema, HTTP_SERVER_URL_STR
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
for req in requests_insert:
|
for req in requests_insert:
|
||||||
tbl = table_name
|
tbl = table_name
|
||||||
if not tbl:
|
if not tbl:
|
||||||
tbl = "table function url('{addr}', 'CSV', '{schema}')".format(addr=HTTP_SERVER_URL_STR, schema=schema)
|
tbl = "table function url('{addr}', 'CSV', '{schema}')".format(
|
||||||
|
addr=HTTP_SERVER_URL_STR, schema=schema
|
||||||
|
)
|
||||||
get_ch_answer(req.format(tbl=tbl))
|
get_ch_answer(req.format(tbl=tbl))
|
||||||
|
|
||||||
|
|
||||||
for i in range(len(requests_select)):
|
for i in range(len(requests_select)):
|
||||||
tbl = table_name
|
tbl = table_name
|
||||||
if not tbl:
|
if not tbl:
|
||||||
tbl = "url('{addr}', 'CSV', '{schema}')".format(addr=HTTP_SERVER_URL_STR, schema=schema)
|
tbl = "url('{addr}', 'CSV', '{schema}')".format(
|
||||||
|
addr=HTTP_SERVER_URL_STR, schema=schema
|
||||||
|
)
|
||||||
check_answers(requests_select[i].format(tbl=tbl), answers[i])
|
check_answers(requests_select[i].format(tbl=tbl), answers[i])
|
||||||
|
|
||||||
if table_name:
|
if table_name:
|
||||||
get_ch_answer("drop table if exists {}".format(table_name))
|
get_ch_answer("drop table if exists {}".format(table_name))
|
||||||
|
|
||||||
|
|
||||||
def test_select_url_engine(requests=[], answers=[], test_data=""):
|
def test_select_url_engine(requests=[], answers=[], test_data=""):
|
||||||
for i in range(len(requests)):
|
for i in range(len(requests)):
|
||||||
check_answers(requests[i], answers[i])
|
check_answers(requests[i], answers[i])
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
test_data = "Hello,2,-2,7.7\nWorld,2,-5,8.8"
|
test_data = "Hello,2,-2,7.7\nWorld,2,-5,8.8"
|
||||||
"""
|
"""
|
||||||
@ -203,19 +253,29 @@ def main():
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
query = "select * from url('http://guest:guest@" + f'[{str(HTTP_SERVER_ADDRESS[0])}]:{str(HTTP_SERVER_ADDRESS[1])}' + "/', 'RawBLOB', 'a String')"
|
query = (
|
||||||
|
"select * from url('http://guest:guest@"
|
||||||
|
+ f"[{str(HTTP_SERVER_ADDRESS[0])}]:{str(HTTP_SERVER_ADDRESS[1])}"
|
||||||
|
+ "/', 'RawBLOB', 'a String')"
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
query = "select * from url('http://guest:guest@" + f'{str(HTTP_SERVER_ADDRESS[0])}:{str(HTTP_SERVER_ADDRESS[1])}' + "/', 'RawBLOB', 'a String')"
|
query = (
|
||||||
|
"select * from url('http://guest:guest@"
|
||||||
|
+ f"{str(HTTP_SERVER_ADDRESS[0])}:{str(HTTP_SERVER_ADDRESS[1])}"
|
||||||
|
+ "/', 'RawBLOB', 'a String')"
|
||||||
|
)
|
||||||
|
|
||||||
select_requests_url_auth = {
|
select_requests_url_auth = {
|
||||||
query : 'hello, world',
|
query: "hello, world",
|
||||||
}
|
}
|
||||||
|
|
||||||
t, httpd = start_server()
|
t, httpd = start_server()
|
||||||
t.start()
|
t.start()
|
||||||
test_select(requests=list(select_requests_url_auth.keys()), answers=list(select_requests_url_auth.values()), test_data=test_data)
|
test_select(
|
||||||
|
requests=list(select_requests_url_auth.keys()),
|
||||||
|
answers=list(select_requests_url_auth.values()),
|
||||||
|
test_data=test_data,
|
||||||
|
)
|
||||||
httpd.shutdown()
|
httpd.shutdown()
|
||||||
t.join()
|
t.join()
|
||||||
print("PASSED")
|
print("PASSED")
|
||||||
|
@ -8,7 +8,7 @@ import pandas as pd
|
|||||||
import numpy as np
|
import numpy as np
|
||||||
|
|
||||||
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
||||||
sys.path.insert(0, os.path.join(CURDIR, 'helpers'))
|
sys.path.insert(0, os.path.join(CURDIR, "helpers"))
|
||||||
|
|
||||||
from pure_http_client import ClickHouseClient
|
from pure_http_client import ClickHouseClient
|
||||||
|
|
||||||
@ -38,71 +38,171 @@ def twosample_proportion_ztest(s1, s2, t1, t2, alpha):
|
|||||||
def test_and_check(name, z_stat, p_value, ci_lower, ci_upper, precision=1e-2):
|
def test_and_check(name, z_stat, p_value, ci_lower, ci_upper, precision=1e-2):
|
||||||
client = ClickHouseClient()
|
client = ClickHouseClient()
|
||||||
real = client.query_return_df(
|
real = client.query_return_df(
|
||||||
"SELECT roundBankers({}.1, 16) as z_stat, ".format(name) +
|
"SELECT roundBankers({}.1, 16) as z_stat, ".format(name)
|
||||||
"roundBankers({}.2, 16) as p_value, ".format(name) +
|
+ "roundBankers({}.2, 16) as p_value, ".format(name)
|
||||||
"roundBankers({}.3, 16) as ci_lower, ".format(name) +
|
+ "roundBankers({}.3, 16) as ci_lower, ".format(name)
|
||||||
"roundBankers({}.4, 16) as ci_upper ".format(name) +
|
+ "roundBankers({}.4, 16) as ci_upper ".format(name)
|
||||||
"FORMAT TabSeparatedWithNames;")
|
+ "FORMAT TabSeparatedWithNames;"
|
||||||
real_z_stat = real['z_stat'][0]
|
)
|
||||||
real_p_value = real['p_value'][0]
|
real_z_stat = real["z_stat"][0]
|
||||||
real_ci_lower = real['ci_lower'][0]
|
real_p_value = real["p_value"][0]
|
||||||
real_ci_upper = real['ci_upper'][0]
|
real_ci_lower = real["ci_lower"][0]
|
||||||
assert((np.isnan(real_z_stat) and np.isnan(z_stat)) or abs(real_z_stat - np.float64(z_stat)) < precision), "clickhouse_z_stat {}, py_z_stat {}".format(real_z_stat, z_stat)
|
real_ci_upper = real["ci_upper"][0]
|
||||||
assert((np.isnan(real_p_value) and np.isnan(p_value)) or abs(real_p_value - np.float64(p_value)) < precision), "clickhouse_p_value {}, py_p_value {}".format(real_p_value, p_value)
|
assert (np.isnan(real_z_stat) and np.isnan(z_stat)) or abs(
|
||||||
assert((np.isnan(real_ci_lower) and np.isnan(ci_lower)) or abs(real_ci_lower - np.float64(ci_lower)) < precision), "clickhouse_ci_lower {}, py_ci_lower {}".format(real_ci_lower, ci_lower)
|
real_z_stat - np.float64(z_stat)
|
||||||
assert((np.isnan(real_ci_upper) and np.isnan(ci_upper)) or abs(real_ci_upper - np.float64(ci_upper)) < precision), "clickhouse_ci_upper {}, py_ci_upper {}".format(real_ci_upper, ci_upper)
|
) < precision, "clickhouse_z_stat {}, py_z_stat {}".format(real_z_stat, z_stat)
|
||||||
|
assert (np.isnan(real_p_value) and np.isnan(p_value)) or abs(
|
||||||
|
real_p_value - np.float64(p_value)
|
||||||
|
) < precision, "clickhouse_p_value {}, py_p_value {}".format(real_p_value, p_value)
|
||||||
|
assert (np.isnan(real_ci_lower) and np.isnan(ci_lower)) or abs(
|
||||||
|
real_ci_lower - np.float64(ci_lower)
|
||||||
|
) < precision, "clickhouse_ci_lower {}, py_ci_lower {}".format(
|
||||||
|
real_ci_lower, ci_lower
|
||||||
|
)
|
||||||
|
assert (np.isnan(real_ci_upper) and np.isnan(ci_upper)) or abs(
|
||||||
|
real_ci_upper - np.float64(ci_upper)
|
||||||
|
) < precision, "clickhouse_ci_upper {}, py_ci_upper {}".format(
|
||||||
|
real_ci_upper, ci_upper
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def test_mean_ztest():
|
def test_mean_ztest():
|
||||||
counts = [0, 0]
|
counts = [0, 0]
|
||||||
nobs = [0, 0]
|
nobs = [0, 0]
|
||||||
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(counts[0], counts[1], nobs[0], nobs[1], 0.05)
|
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(
|
||||||
test_and_check("proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')" % (counts[0], counts[1], nobs[0], nobs[1]), z_stat, p_value, ci_lower, ci_upper)
|
counts[0], counts[1], nobs[0], nobs[1], 0.05
|
||||||
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(10, 10, 10, 10, 0.05)
|
)
|
||||||
|
test_and_check(
|
||||||
|
"proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')"
|
||||||
|
% (counts[0], counts[1], nobs[0], nobs[1]),
|
||||||
|
z_stat,
|
||||||
|
p_value,
|
||||||
|
ci_lower,
|
||||||
|
ci_upper,
|
||||||
|
)
|
||||||
|
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(
|
||||||
|
10, 10, 10, 10, 0.05
|
||||||
|
)
|
||||||
|
|
||||||
counts = [10, 10]
|
counts = [10, 10]
|
||||||
nobs = [10, 10]
|
nobs = [10, 10]
|
||||||
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(counts[0], counts[1], nobs[0], nobs[1], 0.05)
|
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(
|
||||||
test_and_check("proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')" % (counts[0], counts[1], nobs[0], nobs[1]), z_stat, p_value, ci_lower, ci_upper)
|
counts[0], counts[1], nobs[0], nobs[1], 0.05
|
||||||
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(10, 10, 10, 10, 0.05)
|
)
|
||||||
|
test_and_check(
|
||||||
|
"proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')"
|
||||||
|
% (counts[0], counts[1], nobs[0], nobs[1]),
|
||||||
|
z_stat,
|
||||||
|
p_value,
|
||||||
|
ci_lower,
|
||||||
|
ci_upper,
|
||||||
|
)
|
||||||
|
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(
|
||||||
|
10, 10, 10, 10, 0.05
|
||||||
|
)
|
||||||
|
|
||||||
counts = [16, 16]
|
counts = [16, 16]
|
||||||
nobs = [16, 18]
|
nobs = [16, 18]
|
||||||
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(counts[0], counts[1], nobs[0], nobs[1], 0.05)
|
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(
|
||||||
test_and_check("proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')" % (counts[0], counts[1], nobs[0], nobs[1]), z_stat, p_value, ci_lower, ci_upper)
|
counts[0], counts[1], nobs[0], nobs[1], 0.05
|
||||||
|
)
|
||||||
|
test_and_check(
|
||||||
|
"proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')"
|
||||||
|
% (counts[0], counts[1], nobs[0], nobs[1]),
|
||||||
|
z_stat,
|
||||||
|
p_value,
|
||||||
|
ci_lower,
|
||||||
|
ci_upper,
|
||||||
|
)
|
||||||
|
|
||||||
counts = [10, 20]
|
counts = [10, 20]
|
||||||
nobs = [30, 40]
|
nobs = [30, 40]
|
||||||
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(counts[0], counts[1], nobs[0], nobs[1], 0.05)
|
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(
|
||||||
test_and_check("proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')" % (counts[0], counts[1], nobs[0], nobs[1]), z_stat, p_value, ci_lower, ci_upper)
|
counts[0], counts[1], nobs[0], nobs[1], 0.05
|
||||||
|
)
|
||||||
|
test_and_check(
|
||||||
|
"proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')"
|
||||||
|
% (counts[0], counts[1], nobs[0], nobs[1]),
|
||||||
|
z_stat,
|
||||||
|
p_value,
|
||||||
|
ci_lower,
|
||||||
|
ci_upper,
|
||||||
|
)
|
||||||
|
|
||||||
counts = [20, 10]
|
counts = [20, 10]
|
||||||
nobs = [40, 30]
|
nobs = [40, 30]
|
||||||
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(counts[0], counts[1], nobs[0], nobs[1], 0.05)
|
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(
|
||||||
test_and_check("proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')" % (counts[0], counts[1], nobs[0], nobs[1]), z_stat, p_value, ci_lower, ci_upper)
|
counts[0], counts[1], nobs[0], nobs[1], 0.05
|
||||||
|
)
|
||||||
|
test_and_check(
|
||||||
|
"proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')"
|
||||||
|
% (counts[0], counts[1], nobs[0], nobs[1]),
|
||||||
|
z_stat,
|
||||||
|
p_value,
|
||||||
|
ci_lower,
|
||||||
|
ci_upper,
|
||||||
|
)
|
||||||
|
|
||||||
counts = [randrange(10, 20), randrange(10, 20)]
|
counts = [randrange(10, 20), randrange(10, 20)]
|
||||||
nobs = [randrange(counts[0] + 1, counts[0] * 2), randrange(counts[1], counts[1] * 2)]
|
nobs = [
|
||||||
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(counts[0], counts[1], nobs[0], nobs[1], 0.05)
|
randrange(counts[0] + 1, counts[0] * 2),
|
||||||
test_and_check("proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')" % (counts[0], counts[1], nobs[0], nobs[1]), z_stat, p_value, ci_lower, ci_upper)
|
randrange(counts[1], counts[1] * 2),
|
||||||
|
]
|
||||||
|
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(
|
||||||
|
counts[0], counts[1], nobs[0], nobs[1], 0.05
|
||||||
|
)
|
||||||
|
test_and_check(
|
||||||
|
"proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')"
|
||||||
|
% (counts[0], counts[1], nobs[0], nobs[1]),
|
||||||
|
z_stat,
|
||||||
|
p_value,
|
||||||
|
ci_lower,
|
||||||
|
ci_upper,
|
||||||
|
)
|
||||||
|
|
||||||
counts = [randrange(1, 100), randrange(1, 200)]
|
counts = [randrange(1, 100), randrange(1, 200)]
|
||||||
nobs = [randrange(counts[0], counts[0] * 2), randrange(counts[1], counts[1] * 3)]
|
nobs = [randrange(counts[0], counts[0] * 2), randrange(counts[1], counts[1] * 3)]
|
||||||
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(counts[0], counts[1], nobs[0], nobs[1], 0.05)
|
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(
|
||||||
test_and_check("proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')" % (counts[0], counts[1], nobs[0], nobs[1]), z_stat, p_value, ci_lower, ci_upper)
|
counts[0], counts[1], nobs[0], nobs[1], 0.05
|
||||||
|
)
|
||||||
|
test_and_check(
|
||||||
|
"proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')"
|
||||||
|
% (counts[0], counts[1], nobs[0], nobs[1]),
|
||||||
|
z_stat,
|
||||||
|
p_value,
|
||||||
|
ci_lower,
|
||||||
|
ci_upper,
|
||||||
|
)
|
||||||
|
|
||||||
counts = [randrange(1, 200), randrange(1, 100)]
|
counts = [randrange(1, 200), randrange(1, 100)]
|
||||||
nobs = [randrange(counts[0], counts[0] * 3), randrange(counts[1], counts[1] * 2)]
|
nobs = [randrange(counts[0], counts[0] * 3), randrange(counts[1], counts[1] * 2)]
|
||||||
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(counts[0], counts[1], nobs[0], nobs[1], 0.05)
|
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(
|
||||||
test_and_check("proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')" % (counts[0], counts[1], nobs[0], nobs[1]), z_stat, p_value, ci_lower, ci_upper)
|
counts[0], counts[1], nobs[0], nobs[1], 0.05
|
||||||
|
)
|
||||||
|
test_and_check(
|
||||||
|
"proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')"
|
||||||
|
% (counts[0], counts[1], nobs[0], nobs[1]),
|
||||||
|
z_stat,
|
||||||
|
p_value,
|
||||||
|
ci_lower,
|
||||||
|
ci_upper,
|
||||||
|
)
|
||||||
|
|
||||||
counts = [randrange(1, 1000), randrange(1, 1000)]
|
counts = [randrange(1, 1000), randrange(1, 1000)]
|
||||||
nobs = [randrange(counts[0], counts[0] * 2), randrange(counts[1], counts[1] * 2)]
|
nobs = [randrange(counts[0], counts[0] * 2), randrange(counts[1], counts[1] * 2)]
|
||||||
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(counts[0], counts[1], nobs[0], nobs[1], 0.05)
|
z_stat, p_value, ci_lower, ci_upper = twosample_proportion_ztest(
|
||||||
test_and_check("proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')" % (counts[0], counts[1], nobs[0], nobs[1]), z_stat, p_value, ci_lower, ci_upper)
|
counts[0], counts[1], nobs[0], nobs[1], 0.05
|
||||||
|
)
|
||||||
|
test_and_check(
|
||||||
|
"proportionsZTest(%d, %d, %d, %d, 0.95, 'unpooled')"
|
||||||
|
% (counts[0], counts[1], nobs[0], nobs[1]),
|
||||||
|
z_stat,
|
||||||
|
p_value,
|
||||||
|
ci_lower,
|
||||||
|
ci_upper,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
test_mean_ztest()
|
test_mean_ztest()
|
||||||
print("Ok.")
|
print("Ok.")
|
||||||
|
|
||||||
|
@ -7,7 +7,7 @@ import pandas as pd
|
|||||||
import numpy as np
|
import numpy as np
|
||||||
|
|
||||||
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
||||||
sys.path.insert(0, os.path.join(CURDIR, 'helpers'))
|
sys.path.insert(0, os.path.join(CURDIR, "helpers"))
|
||||||
|
|
||||||
from pure_http_client import ClickHouseClient
|
from pure_http_client import ClickHouseClient
|
||||||
|
|
||||||
@ -30,23 +30,40 @@ def twosample_mean_ztest(rvs1, rvs2, alpha=0.05):
|
|||||||
def test_and_check(name, a, b, t_stat, p_value, ci_low, ci_high, precision=1e-2):
|
def test_and_check(name, a, b, t_stat, p_value, ci_low, ci_high, precision=1e-2):
|
||||||
client = ClickHouseClient()
|
client = ClickHouseClient()
|
||||||
client.query("DROP TABLE IF EXISTS ztest;")
|
client.query("DROP TABLE IF EXISTS ztest;")
|
||||||
client.query("CREATE TABLE ztest (left Float64, right UInt8) ENGINE = Memory;");
|
client.query("CREATE TABLE ztest (left Float64, right UInt8) ENGINE = Memory;")
|
||||||
client.query("INSERT INTO ztest VALUES {};".format(", ".join(['({},{})'.format(i, 0) for i in a])))
|
client.query(
|
||||||
client.query("INSERT INTO ztest VALUES {};".format(", ".join(['({},{})'.format(j, 1) for j in b])))
|
"INSERT INTO ztest VALUES {};".format(
|
||||||
|
", ".join(["({},{})".format(i, 0) for i in a])
|
||||||
|
)
|
||||||
|
)
|
||||||
|
client.query(
|
||||||
|
"INSERT INTO ztest VALUES {};".format(
|
||||||
|
", ".join(["({},{})".format(j, 1) for j in b])
|
||||||
|
)
|
||||||
|
)
|
||||||
real = client.query_return_df(
|
real = client.query_return_df(
|
||||||
"SELECT roundBankers({}(left, right).1, 16) as t_stat, ".format(name) +
|
"SELECT roundBankers({}(left, right).1, 16) as t_stat, ".format(name)
|
||||||
"roundBankers({}(left, right).2, 16) as p_value, ".format(name) +
|
+ "roundBankers({}(left, right).2, 16) as p_value, ".format(name)
|
||||||
"roundBankers({}(left, right).3, 16) as ci_low, ".format(name) +
|
+ "roundBankers({}(left, right).3, 16) as ci_low, ".format(name)
|
||||||
"roundBankers({}(left, right).4, 16) as ci_high ".format(name) +
|
+ "roundBankers({}(left, right).4, 16) as ci_high ".format(name)
|
||||||
"FROM ztest FORMAT TabSeparatedWithNames;")
|
+ "FROM ztest FORMAT TabSeparatedWithNames;"
|
||||||
real_t_stat = real['t_stat'][0]
|
)
|
||||||
real_p_value = real['p_value'][0]
|
real_t_stat = real["t_stat"][0]
|
||||||
real_ci_low = real['ci_low'][0]
|
real_p_value = real["p_value"][0]
|
||||||
real_ci_high = real['ci_high'][0]
|
real_ci_low = real["ci_low"][0]
|
||||||
assert(abs(real_t_stat - np.float64(t_stat)) < precision), "clickhouse_t_stat {}, py_t_stat {}".format(real_t_stat, t_stat)
|
real_ci_high = real["ci_high"][0]
|
||||||
assert(abs(real_p_value - np.float64(p_value)) < precision), "clickhouse_p_value {}, py_p_value {}".format(real_p_value, p_value)
|
assert (
|
||||||
assert(abs(real_ci_low - np.float64(ci_low)) < precision), "clickhouse_ci_low {}, py_ci_low {}".format(real_ci_low, ci_low)
|
abs(real_t_stat - np.float64(t_stat)) < precision
|
||||||
assert(abs(real_ci_high - np.float64(ci_high)) < precision), "clickhouse_ci_high {}, py_ci_high {}".format(real_ci_high, ci_high)
|
), "clickhouse_t_stat {}, py_t_stat {}".format(real_t_stat, t_stat)
|
||||||
|
assert (
|
||||||
|
abs(real_p_value - np.float64(p_value)) < precision
|
||||||
|
), "clickhouse_p_value {}, py_p_value {}".format(real_p_value, p_value)
|
||||||
|
assert (
|
||||||
|
abs(real_ci_low - np.float64(ci_low)) < precision
|
||||||
|
), "clickhouse_ci_low {}, py_ci_low {}".format(real_ci_low, ci_low)
|
||||||
|
assert (
|
||||||
|
abs(real_ci_high - np.float64(ci_high)) < precision
|
||||||
|
), "clickhouse_ci_high {}, py_ci_high {}".format(real_ci_high, ci_high)
|
||||||
client.query("DROP TABLE IF EXISTS ztest;")
|
client.query("DROP TABLE IF EXISTS ztest;")
|
||||||
|
|
||||||
|
|
||||||
@ -54,22 +71,54 @@ def test_mean_ztest():
|
|||||||
rvs1 = np.round(stats.norm.rvs(loc=1, scale=5, size=500), 2)
|
rvs1 = np.round(stats.norm.rvs(loc=1, scale=5, size=500), 2)
|
||||||
rvs2 = np.round(stats.norm.rvs(loc=10, scale=5, size=500), 2)
|
rvs2 = np.round(stats.norm.rvs(loc=10, scale=5, size=500), 2)
|
||||||
s, p, cl, ch = twosample_mean_ztest(rvs1, rvs2)
|
s, p, cl, ch = twosample_mean_ztest(rvs1, rvs2)
|
||||||
test_and_check("meanZTest(%f, %f, 0.95)" % (variance(rvs1), variance(rvs2)), rvs1, rvs2, s, p, cl, ch)
|
test_and_check(
|
||||||
|
"meanZTest(%f, %f, 0.95)" % (variance(rvs1), variance(rvs2)),
|
||||||
|
rvs1,
|
||||||
|
rvs2,
|
||||||
|
s,
|
||||||
|
p,
|
||||||
|
cl,
|
||||||
|
ch,
|
||||||
|
)
|
||||||
|
|
||||||
rvs1 = np.round(stats.norm.rvs(loc=0, scale=5, size=500), 2)
|
rvs1 = np.round(stats.norm.rvs(loc=0, scale=5, size=500), 2)
|
||||||
rvs2 = np.round(stats.norm.rvs(loc=0, scale=5, size=500), 2)
|
rvs2 = np.round(stats.norm.rvs(loc=0, scale=5, size=500), 2)
|
||||||
s, p, cl, ch = twosample_mean_ztest(rvs1, rvs2)
|
s, p, cl, ch = twosample_mean_ztest(rvs1, rvs2)
|
||||||
test_and_check("meanZTest(%f, %f, 0.95)" % (variance(rvs1), variance(rvs2)), rvs1, rvs2, s, p, cl, ch)
|
test_and_check(
|
||||||
|
"meanZTest(%f, %f, 0.95)" % (variance(rvs1), variance(rvs2)),
|
||||||
|
rvs1,
|
||||||
|
rvs2,
|
||||||
|
s,
|
||||||
|
p,
|
||||||
|
cl,
|
||||||
|
ch,
|
||||||
|
)
|
||||||
|
|
||||||
rvs1 = np.round(stats.norm.rvs(loc=2, scale=10, size=512), 2)
|
rvs1 = np.round(stats.norm.rvs(loc=2, scale=10, size=512), 2)
|
||||||
rvs2 = np.round(stats.norm.rvs(loc=5, scale=20, size=1024), 2)
|
rvs2 = np.round(stats.norm.rvs(loc=5, scale=20, size=1024), 2)
|
||||||
s, p, cl, ch = twosample_mean_ztest(rvs1, rvs2)
|
s, p, cl, ch = twosample_mean_ztest(rvs1, rvs2)
|
||||||
test_and_check("meanZTest(%f, %f, 0.95)" % (variance(rvs1), variance(rvs2)), rvs1, rvs2, s, p, cl, ch)
|
test_and_check(
|
||||||
|
"meanZTest(%f, %f, 0.95)" % (variance(rvs1), variance(rvs2)),
|
||||||
|
rvs1,
|
||||||
|
rvs2,
|
||||||
|
s,
|
||||||
|
p,
|
||||||
|
cl,
|
||||||
|
ch,
|
||||||
|
)
|
||||||
|
|
||||||
rvs1 = np.round(stats.norm.rvs(loc=0, scale=10, size=1024), 2)
|
rvs1 = np.round(stats.norm.rvs(loc=0, scale=10, size=1024), 2)
|
||||||
rvs2 = np.round(stats.norm.rvs(loc=0, scale=10, size=512), 2)
|
rvs2 = np.round(stats.norm.rvs(loc=0, scale=10, size=512), 2)
|
||||||
s, p, cl, ch = twosample_mean_ztest(rvs1, rvs2)
|
s, p, cl, ch = twosample_mean_ztest(rvs1, rvs2)
|
||||||
test_and_check("meanZTest(%f, %f, 0.95)" % (variance(rvs1), variance(rvs2)), rvs1, rvs2, s, p, cl, ch)
|
test_and_check(
|
||||||
|
"meanZTest(%f, %f, 0.95)" % (variance(rvs1), variance(rvs2)),
|
||||||
|
rvs1,
|
||||||
|
rvs2,
|
||||||
|
s,
|
||||||
|
p,
|
||||||
|
cl,
|
||||||
|
ch,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
@ -3,47 +3,71 @@ import os
|
|||||||
import sys
|
import sys
|
||||||
|
|
||||||
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
||||||
sys.path.insert(0, os.path.join(CURDIR, 'helpers'))
|
sys.path.insert(0, os.path.join(CURDIR, "helpers"))
|
||||||
|
|
||||||
CLICKHOUSE_URL = os.environ.get('CLICKHOUSE_URL')
|
CLICKHOUSE_URL = os.environ.get("CLICKHOUSE_URL")
|
||||||
CLICKHOUSE_TMP = os.environ.get('CLICKHOUSE_TMP')
|
CLICKHOUSE_TMP = os.environ.get("CLICKHOUSE_TMP")
|
||||||
|
|
||||||
from pure_http_client import ClickHouseClient
|
from pure_http_client import ClickHouseClient
|
||||||
|
|
||||||
client = ClickHouseClient()
|
client = ClickHouseClient()
|
||||||
|
|
||||||
|
|
||||||
def run_test(data_format, gen_data_template, settings):
|
def run_test(data_format, gen_data_template, settings):
|
||||||
print(data_format)
|
print(data_format)
|
||||||
client.query("TRUNCATE TABLE t_async_insert")
|
client.query("TRUNCATE TABLE t_async_insert")
|
||||||
|
|
||||||
expected = client.query(gen_data_template.format("TSV")).strip()
|
expected = client.query(gen_data_template.format("TSV")).strip()
|
||||||
data = client.query(gen_data_template.format(data_format), settings=settings,binary_result=True)
|
data = client.query(
|
||||||
|
gen_data_template.format(data_format), settings=settings, binary_result=True
|
||||||
|
)
|
||||||
|
|
||||||
insert_query = "INSERT INTO t_async_insert FORMAT {}".format(data_format)
|
insert_query = "INSERT INTO t_async_insert FORMAT {}".format(data_format)
|
||||||
client.query_with_data(insert_query, data, settings=settings)
|
client.query_with_data(insert_query, data, settings=settings)
|
||||||
|
|
||||||
result = client.query("SELECT * FROM t_async_insert FORMAT TSV").strip()
|
result = client.query("SELECT * FROM t_async_insert FORMAT TSV").strip()
|
||||||
if result != expected:
|
if result != expected:
|
||||||
print("Failed for format {}.\nExpected:\n{}\nGot:\n{}\n".format(data_format, expected, result))
|
print(
|
||||||
|
"Failed for format {}.\nExpected:\n{}\nGot:\n{}\n".format(
|
||||||
|
data_format, expected, result
|
||||||
|
)
|
||||||
|
)
|
||||||
exit(1)
|
exit(1)
|
||||||
|
|
||||||
formats = client.query("SELECT name FROM system.formats WHERE is_input AND is_output \
|
|
||||||
AND name NOT IN ('CapnProto', 'RawBLOB', 'Template', 'ProtobufSingle', 'LineAsString', 'Protobuf', 'ProtobufList') ORDER BY name").strip().split('\n')
|
formats = (
|
||||||
|
client.query(
|
||||||
|
"SELECT name FROM system.formats WHERE is_input AND is_output \
|
||||||
|
AND name NOT IN ('CapnProto', 'RawBLOB', 'Template', 'ProtobufSingle', 'LineAsString', 'Protobuf', 'ProtobufList') ORDER BY name"
|
||||||
|
)
|
||||||
|
.strip()
|
||||||
|
.split("\n")
|
||||||
|
)
|
||||||
|
|
||||||
# Generic formats
|
# Generic formats
|
||||||
client.query("DROP TABLE IF EXISTS t_async_insert")
|
client.query("DROP TABLE IF EXISTS t_async_insert")
|
||||||
client.query("CREATE TABLE t_async_insert (id UInt64, s String, arr Array(UInt64)) ENGINE = Memory")
|
client.query(
|
||||||
|
"CREATE TABLE t_async_insert (id UInt64, s String, arr Array(UInt64)) ENGINE = Memory"
|
||||||
|
)
|
||||||
gen_data_query = "SELECT number AS id, toString(number) AS s, range(number) AS arr FROM numbers(10) FORMAT {}"
|
gen_data_query = "SELECT number AS id, toString(number) AS s, range(number) AS arr FROM numbers(10) FORMAT {}"
|
||||||
|
|
||||||
for data_format in formats:
|
for data_format in formats:
|
||||||
run_test(data_format, gen_data_query, settings={"async_insert": 1, "wait_for_async_insert": 1})
|
run_test(
|
||||||
|
data_format,
|
||||||
|
gen_data_query,
|
||||||
|
settings={"async_insert": 1, "wait_for_async_insert": 1},
|
||||||
|
)
|
||||||
|
|
||||||
# LineAsString
|
# LineAsString
|
||||||
client.query("DROP TABLE IF EXISTS t_async_insert")
|
client.query("DROP TABLE IF EXISTS t_async_insert")
|
||||||
client.query("CREATE TABLE t_async_insert (s String) ENGINE = Memory")
|
client.query("CREATE TABLE t_async_insert (s String) ENGINE = Memory")
|
||||||
gen_data_query = "SELECT toString(number) AS s FROM numbers(10) FORMAT {}"
|
gen_data_query = "SELECT toString(number) AS s FROM numbers(10) FORMAT {}"
|
||||||
|
|
||||||
run_test('LineAsString', gen_data_query, settings={"async_insert": 1, "wait_for_async_insert": 1})
|
run_test(
|
||||||
|
"LineAsString",
|
||||||
|
gen_data_query,
|
||||||
|
settings={"async_insert": 1, "wait_for_async_insert": 1},
|
||||||
|
)
|
||||||
|
|
||||||
# TODO: add CapnProto and Protobuf
|
# TODO: add CapnProto and Protobuf
|
||||||
|
|
||||||
|
@ -17,6 +17,7 @@ def is_ipv6(host):
|
|||||||
except:
|
except:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
def get_local_port(host, ipv6):
|
def get_local_port(host, ipv6):
|
||||||
if ipv6:
|
if ipv6:
|
||||||
family = socket.AF_INET6
|
family = socket.AF_INET6
|
||||||
@ -27,20 +28,19 @@ def get_local_port(host, ipv6):
|
|||||||
fd.bind((host, 0))
|
fd.bind((host, 0))
|
||||||
return fd.getsockname()[1]
|
return fd.getsockname()[1]
|
||||||
|
|
||||||
CLICKHOUSE_HOST = os.environ.get('CLICKHOUSE_HOST', 'localhost')
|
|
||||||
CLICKHOUSE_PORT_HTTP = os.environ.get('CLICKHOUSE_PORT_HTTP', '8123')
|
CLICKHOUSE_HOST = os.environ.get("CLICKHOUSE_HOST", "localhost")
|
||||||
|
CLICKHOUSE_PORT_HTTP = os.environ.get("CLICKHOUSE_PORT_HTTP", "8123")
|
||||||
|
|
||||||
# Server returns this JSON response.
|
# Server returns this JSON response.
|
||||||
SERVER_JSON_RESPONSE = \
|
SERVER_JSON_RESPONSE = """{
|
||||||
'''{
|
|
||||||
"login": "ClickHouse",
|
"login": "ClickHouse",
|
||||||
"id": 54801242,
|
"id": 54801242,
|
||||||
"name": "ClickHouse",
|
"name": "ClickHouse",
|
||||||
"company": null
|
"company": null
|
||||||
}'''
|
}"""
|
||||||
|
|
||||||
EXPECTED_ANSWER = \
|
EXPECTED_ANSWER = """{\\n\\t"login": "ClickHouse",\\n\\t"id": 54801242,\\n\\t"name": "ClickHouse",\\n\\t"company": null\\n}"""
|
||||||
'''{\\n\\t"login": "ClickHouse",\\n\\t"id": 54801242,\\n\\t"name": "ClickHouse",\\n\\t"company": null\\n}'''
|
|
||||||
|
|
||||||
#####################################################################################
|
#####################################################################################
|
||||||
# This test starts an HTTP server and serves data to clickhouse url-engine based table.
|
# This test starts an HTTP server and serves data to clickhouse url-engine based table.
|
||||||
@ -51,26 +51,38 @@ EXPECTED_ANSWER = \
|
|||||||
#####################################################################################
|
#####################################################################################
|
||||||
|
|
||||||
# IP-address of this host accessible from the outside world. Get the first one
|
# IP-address of this host accessible from the outside world. Get the first one
|
||||||
HTTP_SERVER_HOST = subprocess.check_output(['hostname', '-i']).decode('utf-8').strip().split()[0]
|
HTTP_SERVER_HOST = (
|
||||||
|
subprocess.check_output(["hostname", "-i"]).decode("utf-8").strip().split()[0]
|
||||||
|
)
|
||||||
IS_IPV6 = is_ipv6(HTTP_SERVER_HOST)
|
IS_IPV6 = is_ipv6(HTTP_SERVER_HOST)
|
||||||
HTTP_SERVER_PORT = get_local_port(HTTP_SERVER_HOST, IS_IPV6)
|
HTTP_SERVER_PORT = get_local_port(HTTP_SERVER_HOST, IS_IPV6)
|
||||||
|
|
||||||
# IP address and port of the HTTP server started from this script.
|
# IP address and port of the HTTP server started from this script.
|
||||||
HTTP_SERVER_ADDRESS = (HTTP_SERVER_HOST, HTTP_SERVER_PORT)
|
HTTP_SERVER_ADDRESS = (HTTP_SERVER_HOST, HTTP_SERVER_PORT)
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
HTTP_SERVER_URL_STR = 'http://' + f'[{str(HTTP_SERVER_ADDRESS[0])}]:{str(HTTP_SERVER_ADDRESS[1])}' + "/"
|
HTTP_SERVER_URL_STR = (
|
||||||
|
"http://"
|
||||||
|
+ f"[{str(HTTP_SERVER_ADDRESS[0])}]:{str(HTTP_SERVER_ADDRESS[1])}"
|
||||||
|
+ "/"
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
HTTP_SERVER_URL_STR = 'http://' + f'{str(HTTP_SERVER_ADDRESS[0])}:{str(HTTP_SERVER_ADDRESS[1])}' + "/"
|
HTTP_SERVER_URL_STR = (
|
||||||
|
"http://" + f"{str(HTTP_SERVER_ADDRESS[0])}:{str(HTTP_SERVER_ADDRESS[1])}" + "/"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def get_ch_answer(query):
|
def get_ch_answer(query):
|
||||||
host = CLICKHOUSE_HOST
|
host = CLICKHOUSE_HOST
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
host = f'[{host}]'
|
host = f"[{host}]"
|
||||||
|
|
||||||
url = os.environ.get('CLICKHOUSE_URL', 'http://{host}:{port}'.format(host=CLICKHOUSE_HOST, port=CLICKHOUSE_PORT_HTTP))
|
url = os.environ.get(
|
||||||
|
"CLICKHOUSE_URL",
|
||||||
|
"http://{host}:{port}".format(host=CLICKHOUSE_HOST, port=CLICKHOUSE_PORT_HTTP),
|
||||||
|
)
|
||||||
return urllib.request.urlopen(url, data=query.encode()).read().decode()
|
return urllib.request.urlopen(url, data=query.encode()).read().decode()
|
||||||
|
|
||||||
|
|
||||||
def check_answers(query, answer):
|
def check_answers(query, answer):
|
||||||
ch_answer = get_ch_answer(query)
|
ch_answer = get_ch_answer(query)
|
||||||
if ch_answer.strip() != answer.strip():
|
if ch_answer.strip() != answer.strip():
|
||||||
@ -79,16 +91,17 @@ def check_answers(query, answer):
|
|||||||
print("Fetched answer :", ch_answer, file=sys.stderr)
|
print("Fetched answer :", ch_answer, file=sys.stderr)
|
||||||
raise Exception("Fail on query")
|
raise Exception("Fail on query")
|
||||||
|
|
||||||
|
|
||||||
# Server with check for User-Agent headers.
|
# Server with check for User-Agent headers.
|
||||||
class HttpProcessor(SimpleHTTPRequestHandler):
|
class HttpProcessor(SimpleHTTPRequestHandler):
|
||||||
def _set_headers(self):
|
def _set_headers(self):
|
||||||
user_agent = self.headers.get('User-Agent')
|
user_agent = self.headers.get("User-Agent")
|
||||||
if user_agent and user_agent.startswith('ClickHouse/'):
|
if user_agent and user_agent.startswith("ClickHouse/"):
|
||||||
self.send_response(200)
|
self.send_response(200)
|
||||||
else:
|
else:
|
||||||
self.send_response(403)
|
self.send_response(403)
|
||||||
|
|
||||||
self.send_header('Content-Type', 'text/csv')
|
self.send_header("Content-Type", "text/csv")
|
||||||
self.end_headers()
|
self.end_headers()
|
||||||
|
|
||||||
def do_GET(self):
|
def do_GET(self):
|
||||||
@ -98,9 +111,11 @@ class HttpProcessor(SimpleHTTPRequestHandler):
|
|||||||
def log_message(self, format, *args):
|
def log_message(self, format, *args):
|
||||||
return
|
return
|
||||||
|
|
||||||
|
|
||||||
class HTTPServerV6(HTTPServer):
|
class HTTPServerV6(HTTPServer):
|
||||||
address_family = socket.AF_INET6
|
address_family = socket.AF_INET6
|
||||||
|
|
||||||
|
|
||||||
def start_server(requests_amount):
|
def start_server(requests_amount):
|
||||||
if IS_IPV6:
|
if IS_IPV6:
|
||||||
httpd = HTTPServerV6(HTTP_SERVER_ADDRESS, HttpProcessor)
|
httpd = HTTPServerV6(HTTP_SERVER_ADDRESS, HttpProcessor)
|
||||||
@ -114,15 +129,18 @@ def start_server(requests_amount):
|
|||||||
t = threading.Thread(target=real_func)
|
t = threading.Thread(target=real_func)
|
||||||
return t
|
return t
|
||||||
|
|
||||||
|
|
||||||
#####################################################################
|
#####################################################################
|
||||||
# Testing area.
|
# Testing area.
|
||||||
#####################################################################
|
#####################################################################
|
||||||
|
|
||||||
|
|
||||||
def test_select():
|
def test_select():
|
||||||
global HTTP_SERVER_URL_STR
|
global HTTP_SERVER_URL_STR
|
||||||
query = 'SELECT * FROM url(\'{}\',\'JSONAsString\');'.format(HTTP_SERVER_URL_STR)
|
query = "SELECT * FROM url('{}','JSONAsString');".format(HTTP_SERVER_URL_STR)
|
||||||
check_answers(query, EXPECTED_ANSWER)
|
check_answers(query, EXPECTED_ANSWER)
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
# HEAD + GET
|
# HEAD + GET
|
||||||
t = start_server(3)
|
t = start_server(3)
|
||||||
@ -131,6 +149,7 @@ def main():
|
|||||||
t.join()
|
t.join()
|
||||||
print("PASSED")
|
print("PASSED")
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
try:
|
try:
|
||||||
main()
|
main()
|
||||||
@ -141,4 +160,3 @@ if __name__ == "__main__":
|
|||||||
sys.stderr.flush()
|
sys.stderr.flush()
|
||||||
|
|
||||||
os._exit(1)
|
os._exit(1)
|
||||||
|
|
||||||
|
@ -146,7 +146,9 @@ class HttpProcessor(BaseHTTPRequestHandler):
|
|||||||
self.send_error(416, "Requested Range Not Satisfiable")
|
self.send_error(416, "Requested Range Not Satisfiable")
|
||||||
return None
|
return None
|
||||||
|
|
||||||
retry_range_request = first != 0 and from_get is True and len(HttpProcessor.responses_to_get) > 0
|
retry_range_request = (
|
||||||
|
first != 0 and from_get is True and len(HttpProcessor.responses_to_get) > 0
|
||||||
|
)
|
||||||
if retry_range_request:
|
if retry_range_request:
|
||||||
code = HttpProcessor.responses_to_get.pop()
|
code = HttpProcessor.responses_to_get.pop()
|
||||||
if code not in HttpProcessor.responses:
|
if code not in HttpProcessor.responses:
|
||||||
@ -244,7 +246,9 @@ def run_test(allow_range, settings, check_retries=False):
|
|||||||
raise Exception("HTTP Range was not used when supported")
|
raise Exception("HTTP Range was not used when supported")
|
||||||
|
|
||||||
if check_retries and len(HttpProcessor.responses_to_get) > 0:
|
if check_retries and len(HttpProcessor.responses_to_get) > 0:
|
||||||
raise Exception("Expected to get http response 500, which had to be retried, but 200 ok returned and then retried")
|
raise Exception(
|
||||||
|
"Expected to get http response 500, which had to be retried, but 200 ok returned and then retried"
|
||||||
|
)
|
||||||
|
|
||||||
if retries_num > 0:
|
if retries_num > 0:
|
||||||
expected_get_call_num += retries_num - 1
|
expected_get_call_num += retries_num - 1
|
||||||
|
@ -7,7 +7,7 @@ import pandas as pd
|
|||||||
import numpy as np
|
import numpy as np
|
||||||
|
|
||||||
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
||||||
sys.path.insert(0, os.path.join(CURDIR, 'helpers'))
|
sys.path.insert(0, os.path.join(CURDIR, "helpers"))
|
||||||
|
|
||||||
from pure_http_client import ClickHouseClient
|
from pure_http_client import ClickHouseClient
|
||||||
|
|
||||||
@ -22,15 +22,22 @@ def test_and_check(rvs, n_groups, f_stat, p_value, precision=1e-2):
|
|||||||
client.query("DROP TABLE IF EXISTS anova;")
|
client.query("DROP TABLE IF EXISTS anova;")
|
||||||
client.query("CREATE TABLE anova (left Float64, right UInt64) ENGINE = Memory;")
|
client.query("CREATE TABLE anova (left Float64, right UInt64) ENGINE = Memory;")
|
||||||
for group in range(n_groups):
|
for group in range(n_groups):
|
||||||
client.query(f'''INSERT INTO anova VALUES {", ".join([f'({i},{group})' for i in rvs[group]])};''')
|
client.query(
|
||||||
|
f"""INSERT INTO anova VALUES {", ".join([f'({i},{group})' for i in rvs[group]])};"""
|
||||||
|
)
|
||||||
|
|
||||||
real = client.query_return_df(
|
real = client.query_return_df(
|
||||||
'''SELECT roundBankers(a.1, 16) as f_stat, roundBankers(a.2, 16) as p_value FROM (SELECT anova(left, right) as a FROM anova) FORMAT TabSeparatedWithNames;''')
|
"""SELECT roundBankers(a.1, 16) as f_stat, roundBankers(a.2, 16) as p_value FROM (SELECT anova(left, right) as a FROM anova) FORMAT TabSeparatedWithNames;"""
|
||||||
|
)
|
||||||
|
|
||||||
real_f_stat = real['f_stat'][0]
|
real_f_stat = real["f_stat"][0]
|
||||||
real_p_value = real['p_value'][0]
|
real_p_value = real["p_value"][0]
|
||||||
assert(abs(real_f_stat - np.float64(f_stat)) < precision), f"clickhouse_f_stat {real_f_stat}, py_f_stat {f_stat}"
|
assert (
|
||||||
assert(abs(real_p_value - np.float64(p_value)) < precision), f"clickhouse_p_value {real_p_value}, py_p_value {p_value}"
|
abs(real_f_stat - np.float64(f_stat)) < precision
|
||||||
|
), f"clickhouse_f_stat {real_f_stat}, py_f_stat {f_stat}"
|
||||||
|
assert (
|
||||||
|
abs(real_p_value - np.float64(p_value)) < precision
|
||||||
|
), f"clickhouse_p_value {real_p_value}, py_p_value {p_value}"
|
||||||
client.query("DROP TABLE IF EXISTS anova;")
|
client.query("DROP TABLE IF EXISTS anova;")
|
||||||
|
|
||||||
|
|
||||||
|
@ -123,10 +123,14 @@ Uses FinishSortingTransform: {}
|
|||||||
|
|
||||||
for query in queries:
|
for query in queries:
|
||||||
check_query(query["where"], query["order_by"], query["optimize"], False)
|
check_query(query["where"], query["order_by"], query["optimize"], False)
|
||||||
check_query(query["where"], query["order_by"] + ["e"], query["optimize"], query["optimize"])
|
check_query(
|
||||||
|
query["where"], query["order_by"] + ["e"], query["optimize"], query["optimize"]
|
||||||
|
)
|
||||||
|
|
||||||
where_columns = [f"bitNot({col})" for col in query["where"]]
|
where_columns = [f"bitNot({col})" for col in query["where"]]
|
||||||
check_query(where_columns, query["order_by"], query["optimize"], False)
|
check_query(where_columns, query["order_by"], query["optimize"], False)
|
||||||
check_query(where_columns, query["order_by"] + ["e"], query["optimize"], query["optimize"])
|
check_query(
|
||||||
|
where_columns, query["order_by"] + ["e"], query["optimize"], query["optimize"]
|
||||||
|
)
|
||||||
|
|
||||||
print("OK")
|
print("OK")
|
||||||
|
@ -8,8 +8,8 @@ TRANSFER_ENCODING_HEADER = "Transfer-Encoding"
|
|||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
host = os.environ['CLICKHOUSE_HOST']
|
host = os.environ["CLICKHOUSE_HOST"]
|
||||||
port = int(os.environ['CLICKHOUSE_PORT_HTTP'])
|
port = int(os.environ["CLICKHOUSE_PORT_HTTP"])
|
||||||
|
|
||||||
sock = socket(AF_INET, SOCK_STREAM)
|
sock = socket(AF_INET, SOCK_STREAM)
|
||||||
sock.connect((host, port))
|
sock.connect((host, port))
|
||||||
@ -47,4 +47,3 @@ def main():
|
|||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
main()
|
main()
|
||||||
|
|
||||||
|
@ -5,9 +5,10 @@ import os
|
|||||||
import uuid
|
import uuid
|
||||||
import json
|
import json
|
||||||
|
|
||||||
CLICKHOUSE_HOST = os.environ.get('CLICKHOUSE_HOST', '127.0.0.1')
|
CLICKHOUSE_HOST = os.environ.get("CLICKHOUSE_HOST", "127.0.0.1")
|
||||||
CLICKHOUSE_PORT = int(os.environ.get('CLICKHOUSE_PORT_TCP', '900000'))
|
CLICKHOUSE_PORT = int(os.environ.get("CLICKHOUSE_PORT_TCP", "900000"))
|
||||||
CLICKHOUSE_DATABASE = os.environ.get('CLICKHOUSE_DATABASE', 'default')
|
CLICKHOUSE_DATABASE = os.environ.get("CLICKHOUSE_DATABASE", "default")
|
||||||
|
|
||||||
|
|
||||||
def writeVarUInt(x, ba):
|
def writeVarUInt(x, ba):
|
||||||
for _ in range(0, 9):
|
for _ in range(0, 9):
|
||||||
@ -24,7 +25,7 @@ def writeVarUInt(x, ba):
|
|||||||
|
|
||||||
|
|
||||||
def writeStringBinary(s, ba):
|
def writeStringBinary(s, ba):
|
||||||
b = bytes(s, 'utf-8')
|
b = bytes(s, "utf-8")
|
||||||
writeVarUInt(len(s), ba)
|
writeVarUInt(len(s), ba)
|
||||||
ba.extend(b)
|
ba.extend(b)
|
||||||
|
|
||||||
@ -48,18 +49,23 @@ def readUInt(s, size=1):
|
|||||||
val += res[i] << (i * 8)
|
val += res[i] << (i * 8)
|
||||||
return val
|
return val
|
||||||
|
|
||||||
|
|
||||||
def readUInt8(s):
|
def readUInt8(s):
|
||||||
return readUInt(s)
|
return readUInt(s)
|
||||||
|
|
||||||
|
|
||||||
def readUInt16(s):
|
def readUInt16(s):
|
||||||
return readUInt(s, 2)
|
return readUInt(s, 2)
|
||||||
|
|
||||||
|
|
||||||
def readUInt32(s):
|
def readUInt32(s):
|
||||||
return readUInt(s, 4)
|
return readUInt(s, 4)
|
||||||
|
|
||||||
|
|
||||||
def readUInt64(s):
|
def readUInt64(s):
|
||||||
return readUInt(s, 8)
|
return readUInt(s, 8)
|
||||||
|
|
||||||
|
|
||||||
def readVarUInt(s):
|
def readVarUInt(s):
|
||||||
x = 0
|
x = 0
|
||||||
for i in range(9):
|
for i in range(9):
|
||||||
@ -75,25 +81,25 @@ def readVarUInt(s):
|
|||||||
def readStringBinary(s):
|
def readStringBinary(s):
|
||||||
size = readVarUInt(s)
|
size = readVarUInt(s)
|
||||||
s = readStrict(s, size)
|
s = readStrict(s, size)
|
||||||
return s.decode('utf-8')
|
return s.decode("utf-8")
|
||||||
|
|
||||||
|
|
||||||
def sendHello(s):
|
def sendHello(s):
|
||||||
ba = bytearray()
|
ba = bytearray()
|
||||||
writeVarUInt(0, ba) # Hello
|
writeVarUInt(0, ba) # Hello
|
||||||
writeStringBinary('simple native protocol', ba)
|
writeStringBinary("simple native protocol", ba)
|
||||||
writeVarUInt(21, ba)
|
writeVarUInt(21, ba)
|
||||||
writeVarUInt(9, ba)
|
writeVarUInt(9, ba)
|
||||||
writeVarUInt(54449, ba)
|
writeVarUInt(54449, ba)
|
||||||
writeStringBinary(CLICKHOUSE_DATABASE, ba) # database
|
writeStringBinary(CLICKHOUSE_DATABASE, ba) # database
|
||||||
writeStringBinary('default', ba) # user
|
writeStringBinary("default", ba) # user
|
||||||
writeStringBinary('', ba) # pwd
|
writeStringBinary("", ba) # pwd
|
||||||
s.sendall(ba)
|
s.sendall(ba)
|
||||||
|
|
||||||
|
|
||||||
def receiveHello(s):
|
def receiveHello(s):
|
||||||
p_type = readVarUInt(s)
|
p_type = readVarUInt(s)
|
||||||
assert (p_type == 0) # Hello
|
assert p_type == 0 # Hello
|
||||||
server_name = readStringBinary(s)
|
server_name = readStringBinary(s)
|
||||||
# print("Server name: ", server_name)
|
# print("Server name: ", server_name)
|
||||||
server_version_major = readVarUInt(s)
|
server_version_major = readVarUInt(s)
|
||||||
@ -111,18 +117,18 @@ def receiveHello(s):
|
|||||||
|
|
||||||
|
|
||||||
def serializeClientInfo(ba, query_id):
|
def serializeClientInfo(ba, query_id):
|
||||||
writeStringBinary('default', ba) # initial_user
|
writeStringBinary("default", ba) # initial_user
|
||||||
writeStringBinary(query_id, ba) # initial_query_id
|
writeStringBinary(query_id, ba) # initial_query_id
|
||||||
writeStringBinary('127.0.0.1:9000', ba) # initial_address
|
writeStringBinary("127.0.0.1:9000", ba) # initial_address
|
||||||
ba.extend([0] * 8) # initial_query_start_time_microseconds
|
ba.extend([0] * 8) # initial_query_start_time_microseconds
|
||||||
ba.append(1) # TCP
|
ba.append(1) # TCP
|
||||||
writeStringBinary('os_user', ba) # os_user
|
writeStringBinary("os_user", ba) # os_user
|
||||||
writeStringBinary('client_hostname', ba) # client_hostname
|
writeStringBinary("client_hostname", ba) # client_hostname
|
||||||
writeStringBinary('client_name', ba) # client_name
|
writeStringBinary("client_name", ba) # client_name
|
||||||
writeVarUInt(21, ba)
|
writeVarUInt(21, ba)
|
||||||
writeVarUInt(9, ba)
|
writeVarUInt(9, ba)
|
||||||
writeVarUInt(54449, ba)
|
writeVarUInt(54449, ba)
|
||||||
writeStringBinary('', ba) # quota_key
|
writeStringBinary("", ba) # quota_key
|
||||||
writeVarUInt(0, ba) # distributed_depth
|
writeVarUInt(0, ba) # distributed_depth
|
||||||
writeVarUInt(1, ba) # client_version_patch
|
writeVarUInt(1, ba) # client_version_patch
|
||||||
ba.append(0) # No telemetry
|
ba.append(0) # No telemetry
|
||||||
@ -139,8 +145,8 @@ def sendQuery(s, query):
|
|||||||
# client info
|
# client info
|
||||||
serializeClientInfo(ba, query_id)
|
serializeClientInfo(ba, query_id)
|
||||||
|
|
||||||
writeStringBinary('', ba) # No settings
|
writeStringBinary("", ba) # No settings
|
||||||
writeStringBinary('', ba) # No interserver secret
|
writeStringBinary("", ba) # No interserver secret
|
||||||
writeVarUInt(2, ba) # Stage - Complete
|
writeVarUInt(2, ba) # Stage - Complete
|
||||||
ba.append(0) # No compression
|
ba.append(0) # No compression
|
||||||
writeStringBinary(query, ba) # query, finally
|
writeStringBinary(query, ba) # query, finally
|
||||||
@ -158,7 +164,7 @@ def serializeBlockInfo(ba):
|
|||||||
def sendEmptyBlock(s):
|
def sendEmptyBlock(s):
|
||||||
ba = bytearray()
|
ba = bytearray()
|
||||||
writeVarUInt(2, ba) # Data
|
writeVarUInt(2, ba) # Data
|
||||||
writeStringBinary('', ba)
|
writeStringBinary("", ba)
|
||||||
serializeBlockInfo(ba)
|
serializeBlockInfo(ba)
|
||||||
writeVarUInt(0, ba) # rows
|
writeVarUInt(0, ba) # rows
|
||||||
writeVarUInt(0, ba) # columns
|
writeVarUInt(0, ba) # columns
|
||||||
@ -166,10 +172,10 @@ def sendEmptyBlock(s):
|
|||||||
|
|
||||||
|
|
||||||
def assertPacket(packet, expected):
|
def assertPacket(packet, expected):
|
||||||
assert(packet == expected), packet
|
assert packet == expected, packet
|
||||||
|
|
||||||
|
|
||||||
class Progress():
|
class Progress:
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
# NOTE: this is done in ctor to initialize __dict__
|
# NOTE: this is done in ctor to initialize __dict__
|
||||||
self.read_rows = 0
|
self.read_rows = 0
|
||||||
@ -198,11 +204,12 @@ class Progress():
|
|||||||
|
|
||||||
def __bool__(self):
|
def __bool__(self):
|
||||||
return (
|
return (
|
||||||
self.read_rows > 0 or
|
self.read_rows > 0
|
||||||
self.read_bytes > 0 or
|
or self.read_bytes > 0
|
||||||
self.total_rows_to_read > 0 or
|
or self.total_rows_to_read > 0
|
||||||
self.written_rows > 0 or
|
or self.written_rows > 0
|
||||||
self.written_bytes > 0)
|
or self.written_bytes > 0
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def readProgress(s):
|
def readProgress(s):
|
||||||
@ -219,13 +226,14 @@ def readProgress(s):
|
|||||||
progress.readPacket(s)
|
progress.readPacket(s)
|
||||||
return progress
|
return progress
|
||||||
|
|
||||||
|
|
||||||
def readException(s):
|
def readException(s):
|
||||||
code = readUInt32(s)
|
code = readUInt32(s)
|
||||||
name = readStringBinary(s)
|
name = readStringBinary(s)
|
||||||
text = readStringBinary(s)
|
text = readStringBinary(s)
|
||||||
readStringBinary(s) # trace
|
readStringBinary(s) # trace
|
||||||
assertPacket(readUInt8(s), 0) # has_nested
|
assertPacket(readUInt8(s), 0) # has_nested
|
||||||
return "code {}: {}".format(code, text.replace('DB::Exception:', ''))
|
return "code {}: {}".format(code, text.replace("DB::Exception:", ""))
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
@ -236,7 +244,10 @@ def main():
|
|||||||
receiveHello(s)
|
receiveHello(s)
|
||||||
# For 1 second sleep and 1000ms of interactive_delay we definitelly should have non zero progress packet.
|
# For 1 second sleep and 1000ms of interactive_delay we definitelly should have non zero progress packet.
|
||||||
# NOTE: interactive_delay=0 cannot be used since in this case CompletedPipelineExecutor will not call cancelled callback.
|
# NOTE: interactive_delay=0 cannot be used since in this case CompletedPipelineExecutor will not call cancelled callback.
|
||||||
sendQuery(s, "insert into function null('_ Int') select sleep(1) from numbers(2) settings max_block_size=1, interactive_delay=1000")
|
sendQuery(
|
||||||
|
s,
|
||||||
|
"insert into function null('_ Int') select sleep(1) from numbers(2) settings max_block_size=1, interactive_delay=1000",
|
||||||
|
)
|
||||||
|
|
||||||
# external tables
|
# external tables
|
||||||
sendEmptyBlock(s)
|
sendEmptyBlock(s)
|
||||||
|
@ -4,18 +4,19 @@ import os
|
|||||||
import sys
|
import sys
|
||||||
|
|
||||||
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
||||||
sys.path.insert(0, os.path.join(CURDIR, 'helpers'))
|
sys.path.insert(0, os.path.join(CURDIR, "helpers"))
|
||||||
|
|
||||||
from pure_http_client import ClickHouseClient
|
from pure_http_client import ClickHouseClient
|
||||||
|
|
||||||
|
|
||||||
class Tester:
|
class Tester:
|
||||||
'''
|
"""
|
||||||
- Creates test table
|
- Creates test table
|
||||||
- Deletes the specified range of rows
|
- Deletes the specified range of rows
|
||||||
- Masks another range using row-level policy
|
- Masks another range using row-level policy
|
||||||
- Runs some read queries and checks that the results
|
- Runs some read queries and checks that the results
|
||||||
'''
|
"""
|
||||||
|
|
||||||
def __init__(self, session, url, index_granularity, total_rows):
|
def __init__(self, session, url, index_granularity, total_rows):
|
||||||
self.session = session
|
self.session = session
|
||||||
self.url = url
|
self.url = url
|
||||||
@ -25,7 +26,7 @@ class Tester:
|
|||||||
self.repro_queries = []
|
self.repro_queries = []
|
||||||
|
|
||||||
def report_error(self):
|
def report_error(self):
|
||||||
print('Repro steps:', '\n\n\t'.join(self.repro_queries))
|
print("Repro steps:", "\n\n\t".join(self.repro_queries))
|
||||||
exit(1)
|
exit(1)
|
||||||
|
|
||||||
def query(self, query_text, include_in_repro_steps=True, expected_data=None):
|
def query(self, query_text, include_in_repro_steps=True, expected_data=None):
|
||||||
@ -36,113 +37,187 @@ class Tester:
|
|||||||
error = resp.text[0:40]
|
error = resp.text[0:40]
|
||||||
if error not in self.reported_errors:
|
if error not in self.reported_errors:
|
||||||
self.reported_errors.add(error)
|
self.reported_errors.add(error)
|
||||||
print('Code:', resp.status_code)
|
print("Code:", resp.status_code)
|
||||||
print('Result:', resp.text)
|
print("Result:", resp.text)
|
||||||
self.report_error()
|
self.report_error()
|
||||||
|
|
||||||
result = resp.text
|
result = resp.text
|
||||||
# Check that the result is as expected
|
# Check that the result is as expected
|
||||||
if ((not expected_data is None) and (int(result) != len(expected_data))):
|
if (not expected_data is None) and (int(result) != len(expected_data)):
|
||||||
print('Expected {} rows, got {}'.format(len(expected_data), result))
|
print("Expected {} rows, got {}".format(len(expected_data), result))
|
||||||
print('Expected data:' + str(expected_data))
|
print("Expected data:" + str(expected_data))
|
||||||
self.report_error()
|
self.report_error()
|
||||||
|
|
||||||
if not include_in_repro_steps:
|
if not include_in_repro_steps:
|
||||||
self.repro_queries.pop()
|
self.repro_queries.pop()
|
||||||
|
|
||||||
|
def check_data(
|
||||||
def check_data(self, all_data, delete_range_start, delete_range_end, row_level_policy_range_start, row_level_policy_range_end):
|
self,
|
||||||
|
all_data,
|
||||||
|
delete_range_start,
|
||||||
|
delete_range_end,
|
||||||
|
row_level_policy_range_start,
|
||||||
|
row_level_policy_range_end,
|
||||||
|
):
|
||||||
all_data_after_delete = all_data[
|
all_data_after_delete = all_data[
|
||||||
~((all_data.a == 0) &
|
~(
|
||||||
(all_data.b > delete_range_start) &
|
(all_data.a == 0)
|
||||||
(all_data.b <= delete_range_end))]
|
& (all_data.b > delete_range_start)
|
||||||
|
& (all_data.b <= delete_range_end)
|
||||||
|
)
|
||||||
|
]
|
||||||
all_data_after_row_policy = all_data_after_delete[
|
all_data_after_row_policy = all_data_after_delete[
|
||||||
(all_data_after_delete.b <= row_level_policy_range_start) |
|
(all_data_after_delete.b <= row_level_policy_range_start)
|
||||||
(all_data_after_delete.b > row_level_policy_range_end)]
|
| (all_data_after_delete.b > row_level_policy_range_end)
|
||||||
|
]
|
||||||
|
|
||||||
for to_select in ['count()', 'sum(d)']: # Test reading with and without column with default value
|
for to_select in [
|
||||||
self.query('SELECT {} FROM tab_02473;'.format(to_select), False, all_data_after_row_policy)
|
"count()",
|
||||||
|
"sum(d)",
|
||||||
|
]: # Test reading with and without column with default value
|
||||||
|
self.query(
|
||||||
|
"SELECT {} FROM tab_02473;".format(to_select),
|
||||||
|
False,
|
||||||
|
all_data_after_row_policy,
|
||||||
|
)
|
||||||
|
|
||||||
delta = 10
|
delta = 10
|
||||||
for query_range_start in [0, delta]:
|
for query_range_start in [0, delta]:
|
||||||
for query_range_end in [self.total_rows - delta]: # , self.total_rows]:
|
for query_range_end in [self.total_rows - delta]: # , self.total_rows]:
|
||||||
expected = all_data_after_row_policy[
|
expected = all_data_after_row_policy[
|
||||||
(all_data_after_row_policy.a == 0) &
|
(all_data_after_row_policy.a == 0)
|
||||||
(all_data_after_row_policy.b > query_range_start) &
|
& (all_data_after_row_policy.b > query_range_start)
|
||||||
(all_data_after_row_policy.b <= query_range_end)]
|
& (all_data_after_row_policy.b <= query_range_end)
|
||||||
self.query('SELECT {} from tab_02473 PREWHERE b > {} AND b <= {} WHERE a == 0;'.format(
|
]
|
||||||
to_select, query_range_start, query_range_end), False, expected)
|
self.query(
|
||||||
|
"SELECT {} from tab_02473 PREWHERE b > {} AND b <= {} WHERE a == 0;".format(
|
||||||
|
to_select, query_range_start, query_range_end
|
||||||
|
),
|
||||||
|
False,
|
||||||
|
expected,
|
||||||
|
)
|
||||||
|
|
||||||
expected = all_data_after_row_policy[
|
expected = all_data_after_row_policy[
|
||||||
(all_data_after_row_policy.a == 0) &
|
(all_data_after_row_policy.a == 0)
|
||||||
(all_data_after_row_policy.c > query_range_start) &
|
& (all_data_after_row_policy.c > query_range_start)
|
||||||
(all_data_after_row_policy.c <= query_range_end)]
|
& (all_data_after_row_policy.c <= query_range_end)
|
||||||
self.query('SELECT {} from tab_02473 PREWHERE c > {} AND c <= {} WHERE a == 0;'.format(
|
]
|
||||||
to_select, query_range_start, query_range_end), False, expected)
|
self.query(
|
||||||
|
"SELECT {} from tab_02473 PREWHERE c > {} AND c <= {} WHERE a == 0;".format(
|
||||||
|
to_select, query_range_start, query_range_end
|
||||||
|
),
|
||||||
|
False,
|
||||||
|
expected,
|
||||||
|
)
|
||||||
|
|
||||||
expected = all_data_after_row_policy[
|
expected = all_data_after_row_policy[
|
||||||
(all_data_after_row_policy.a == 0) &
|
(all_data_after_row_policy.a == 0)
|
||||||
((all_data_after_row_policy.c <= query_range_start) |
|
& (
|
||||||
(all_data_after_row_policy.c > query_range_end))]
|
(all_data_after_row_policy.c <= query_range_start)
|
||||||
self.query('SELECT {} from tab_02473 PREWHERE c <= {} OR c > {} WHERE a == 0;'.format(
|
| (all_data_after_row_policy.c > query_range_end)
|
||||||
to_select, query_range_start, query_range_end), False, expected)
|
)
|
||||||
|
]
|
||||||
|
self.query(
|
||||||
|
"SELECT {} from tab_02473 PREWHERE c <= {} OR c > {} WHERE a == 0;".format(
|
||||||
|
to_select, query_range_start, query_range_end
|
||||||
|
),
|
||||||
|
False,
|
||||||
|
expected,
|
||||||
|
)
|
||||||
|
|
||||||
|
def run_test(
|
||||||
def run_test(self, delete_range_start, delete_range_end, row_level_policy_range_start, row_level_policy_range_end):
|
self,
|
||||||
|
delete_range_start,
|
||||||
|
delete_range_end,
|
||||||
|
row_level_policy_range_start,
|
||||||
|
row_level_policy_range_end,
|
||||||
|
):
|
||||||
self.repro_queries = []
|
self.repro_queries = []
|
||||||
|
|
||||||
self.query('''
|
self.query(
|
||||||
|
"""
|
||||||
CREATE TABLE tab_02473 (a Int8, b Int32, c Int32, PRIMARY KEY (a))
|
CREATE TABLE tab_02473 (a Int8, b Int32, c Int32, PRIMARY KEY (a))
|
||||||
ENGINE = MergeTree() ORDER BY (a, b)
|
ENGINE = MergeTree() ORDER BY (a, b)
|
||||||
SETTINGS min_bytes_for_wide_part = 0, index_granularity = {};'''.format(self.index_granularity))
|
SETTINGS min_bytes_for_wide_part = 0, index_granularity = {};""".format(
|
||||||
|
self.index_granularity
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
self.query('INSERT INTO tab_02473 select 0, number+1, number+1 FROM numbers({});'.format(self.total_rows))
|
self.query(
|
||||||
|
"INSERT INTO tab_02473 select 0, number+1, number+1 FROM numbers({});".format(
|
||||||
|
self.total_rows
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
client = ClickHouseClient()
|
client = ClickHouseClient()
|
||||||
all_data = client.query_return_df("SELECT a, b, c, 1 as d FROM tab_02473 FORMAT TabSeparatedWithNames;")
|
all_data = client.query_return_df(
|
||||||
|
"SELECT a, b, c, 1 as d FROM tab_02473 FORMAT TabSeparatedWithNames;"
|
||||||
|
)
|
||||||
|
|
||||||
self.query('OPTIMIZE TABLE tab_02473 FINAL SETTINGS mutations_sync=2;')
|
self.query("OPTIMIZE TABLE tab_02473 FINAL SETTINGS mutations_sync=2;")
|
||||||
|
|
||||||
# After all data has been written add a column with default value
|
# After all data has been written add a column with default value
|
||||||
self.query('ALTER TABLE tab_02473 ADD COLUMN d Int64 DEFAULT 1;')
|
self.query("ALTER TABLE tab_02473 ADD COLUMN d Int64 DEFAULT 1;")
|
||||||
|
|
||||||
self.check_data(all_data, -100, -100, -100, -100)
|
self.check_data(all_data, -100, -100, -100, -100)
|
||||||
|
|
||||||
self.query('DELETE FROM tab_02473 WHERE a = 0 AND b > {} AND b <= {};'.format(
|
self.query(
|
||||||
delete_range_start, delete_range_end))
|
"DELETE FROM tab_02473 WHERE a = 0 AND b > {} AND b <= {};".format(
|
||||||
|
delete_range_start, delete_range_end
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
self.check_data(all_data, delete_range_start, delete_range_end, -100, -100)
|
self.check_data(all_data, delete_range_start, delete_range_end, -100, -100)
|
||||||
|
|
||||||
self.query('CREATE ROW POLICY policy_tab_02473 ON tab_02473 FOR SELECT USING b <= {} OR b > {} TO default;'.format(
|
self.query(
|
||||||
row_level_policy_range_start, row_level_policy_range_end))
|
"CREATE ROW POLICY policy_tab_02473 ON tab_02473 FOR SELECT USING b <= {} OR b > {} TO default;".format(
|
||||||
|
row_level_policy_range_start, row_level_policy_range_end
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
self.check_data(all_data, delete_range_start, delete_range_end, row_level_policy_range_start, row_level_policy_range_end)
|
self.check_data(
|
||||||
|
all_data,
|
||||||
|
delete_range_start,
|
||||||
|
delete_range_end,
|
||||||
|
row_level_policy_range_start,
|
||||||
|
row_level_policy_range_end,
|
||||||
|
)
|
||||||
|
|
||||||
self.query('DROP POLICY policy_tab_02473 ON tab_02473;')
|
self.query("DROP POLICY policy_tab_02473 ON tab_02473;")
|
||||||
|
|
||||||
self.query('DROP TABLE tab_02473;')
|
|
||||||
|
|
||||||
|
self.query("DROP TABLE tab_02473;")
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
# Set mutations to synchronous mode and enable lightweight DELETE's
|
# Set mutations to synchronous mode and enable lightweight DELETE's
|
||||||
url = os.environ['CLICKHOUSE_URL'] + '&max_threads=1'
|
url = os.environ["CLICKHOUSE_URL"] + "&max_threads=1"
|
||||||
|
|
||||||
default_index_granularity = 10;
|
default_index_granularity = 10
|
||||||
total_rows = 8 * default_index_granularity
|
total_rows = 8 * default_index_granularity
|
||||||
step = default_index_granularity
|
step = default_index_granularity
|
||||||
session = requests.Session()
|
session = requests.Session()
|
||||||
for index_granularity in [default_index_granularity-1, default_index_granularity]: # [default_index_granularity-1, default_index_granularity+1, default_index_granularity]:
|
for index_granularity in [
|
||||||
|
default_index_granularity - 1,
|
||||||
|
default_index_granularity,
|
||||||
|
]: # [default_index_granularity-1, default_index_granularity+1, default_index_granularity]:
|
||||||
tester = Tester(session, url, index_granularity, total_rows)
|
tester = Tester(session, url, index_granularity, total_rows)
|
||||||
# Test combinations of ranges of various size masked by lightweight DELETES
|
# Test combinations of ranges of various size masked by lightweight DELETES
|
||||||
# along with ranges of various size masked by row-level policies
|
# along with ranges of various size masked by row-level policies
|
||||||
for delete_range_start in range(0, total_rows, 3 * step):
|
for delete_range_start in range(0, total_rows, 3 * step):
|
||||||
for delete_range_end in range(delete_range_start + 3 * step, total_rows, 2 * step):
|
for delete_range_end in range(
|
||||||
|
delete_range_start + 3 * step, total_rows, 2 * step
|
||||||
|
):
|
||||||
for row_level_policy_range_start in range(0, total_rows, 3 * step):
|
for row_level_policy_range_start in range(0, total_rows, 3 * step):
|
||||||
for row_level_policy_range_end in range(row_level_policy_range_start + 3 * step, total_rows, 2 * step):
|
for row_level_policy_range_end in range(
|
||||||
tester.run_test(delete_range_start, delete_range_end, row_level_policy_range_start, row_level_policy_range_end)
|
row_level_policy_range_start + 3 * step, total_rows, 2 * step
|
||||||
|
):
|
||||||
|
tester.run_test(
|
||||||
|
delete_range_start,
|
||||||
|
delete_range_end,
|
||||||
|
row_level_policy_range_start,
|
||||||
|
row_level_policy_range_end,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
main()
|
main()
|
||||||
|
|
||||||
|
@ -4,16 +4,17 @@ import os
|
|||||||
import sys
|
import sys
|
||||||
|
|
||||||
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
||||||
sys.path.insert(0, os.path.join(CURDIR, 'helpers'))
|
sys.path.insert(0, os.path.join(CURDIR, "helpers"))
|
||||||
|
|
||||||
from pure_http_client import ClickHouseClient
|
from pure_http_client import ClickHouseClient
|
||||||
|
|
||||||
|
|
||||||
class Tester:
|
class Tester:
|
||||||
'''
|
"""
|
||||||
- Creates test table with multiple integer columns
|
- Creates test table with multiple integer columns
|
||||||
- Runs read queries with multiple range conditions on different columns in PREWHERE and check that the result is correct
|
- Runs read queries with multiple range conditions on different columns in PREWHERE and check that the result is correct
|
||||||
'''
|
"""
|
||||||
|
|
||||||
def __init__(self, session, url, index_granularity, total_rows):
|
def __init__(self, session, url, index_granularity, total_rows):
|
||||||
self.session = session
|
self.session = session
|
||||||
self.url = url
|
self.url = url
|
||||||
@ -23,7 +24,7 @@ class Tester:
|
|||||||
self.repro_queries = []
|
self.repro_queries = []
|
||||||
|
|
||||||
def report_error(self):
|
def report_error(self):
|
||||||
print('Repro steps:', '\n\n\t'.join(self.repro_queries))
|
print("Repro steps:", "\n\n\t".join(self.repro_queries))
|
||||||
exit(1)
|
exit(1)
|
||||||
|
|
||||||
def query(self, query_text, include_in_repro_steps=True, expected_data=None):
|
def query(self, query_text, include_in_repro_steps=True, expected_data=None):
|
||||||
@ -34,85 +35,130 @@ class Tester:
|
|||||||
error = resp.text[0:40]
|
error = resp.text[0:40]
|
||||||
if error not in self.reported_errors:
|
if error not in self.reported_errors:
|
||||||
self.reported_errors.add(error)
|
self.reported_errors.add(error)
|
||||||
print('Code:', resp.status_code)
|
print("Code:", resp.status_code)
|
||||||
print('Result:', resp.text)
|
print("Result:", resp.text)
|
||||||
self.report_error()
|
self.report_error()
|
||||||
|
|
||||||
result = resp.text
|
result = resp.text
|
||||||
# Check that the result is as expected
|
# Check that the result is as expected
|
||||||
if ((not expected_data is None) and (int(result) != len(expected_data))):
|
if (not expected_data is None) and (int(result) != len(expected_data)):
|
||||||
print('Expected {} rows, got {}'.format(len(expected_data), result))
|
print("Expected {} rows, got {}".format(len(expected_data), result))
|
||||||
print('Expected data:' + str(expected_data))
|
print("Expected data:" + str(expected_data))
|
||||||
self.report_error()
|
self.report_error()
|
||||||
|
|
||||||
if not include_in_repro_steps:
|
if not include_in_repro_steps:
|
||||||
self.repro_queries.pop()
|
self.repro_queries.pop()
|
||||||
|
|
||||||
|
def check_data(
|
||||||
def check_data(self, all_data, c_range_start, c_range_end, d_range_start, d_range_end):
|
self, all_data, c_range_start, c_range_end, d_range_start, d_range_end
|
||||||
for to_select in ['count()', 'sum(e)']: # Test reading with and without column with default value
|
):
|
||||||
self.query('SELECT {} FROM tab_02473;'.format(to_select), False, all_data)
|
for to_select in [
|
||||||
|
"count()",
|
||||||
|
"sum(e)",
|
||||||
|
]: # Test reading with and without column with default value
|
||||||
|
self.query("SELECT {} FROM tab_02473;".format(to_select), False, all_data)
|
||||||
|
|
||||||
delta = 10
|
delta = 10
|
||||||
for b_range_start in [0, delta]:
|
for b_range_start in [0, delta]:
|
||||||
for b_range_end in [self.total_rows - delta]: # , self.total_rows]:
|
for b_range_end in [self.total_rows - delta]: # , self.total_rows]:
|
||||||
expected = all_data[
|
expected = all_data[
|
||||||
(all_data.a == 0) &
|
(all_data.a == 0)
|
||||||
(all_data.b > b_range_start) &
|
& (all_data.b > b_range_start)
|
||||||
(all_data.b <= b_range_end)]
|
& (all_data.b <= b_range_end)
|
||||||
self.query('SELECT {} from tab_02473 PREWHERE b > {} AND b <= {} WHERE a == 0;'.format(
|
]
|
||||||
to_select, b_range_start, b_range_end), False, expected)
|
self.query(
|
||||||
|
"SELECT {} from tab_02473 PREWHERE b > {} AND b <= {} WHERE a == 0;".format(
|
||||||
|
to_select, b_range_start, b_range_end
|
||||||
|
),
|
||||||
|
False,
|
||||||
|
expected,
|
||||||
|
)
|
||||||
|
|
||||||
expected = all_data[
|
expected = all_data[
|
||||||
(all_data.a == 0) &
|
(all_data.a == 0)
|
||||||
(all_data.b > b_range_start) &
|
& (all_data.b > b_range_start)
|
||||||
(all_data.b <= b_range_end) &
|
& (all_data.b <= b_range_end)
|
||||||
(all_data.c > c_range_start) &
|
& (all_data.c > c_range_start)
|
||||||
(all_data.c <= c_range_end)]
|
& (all_data.c <= c_range_end)
|
||||||
self.query('SELECT {} from tab_02473 PREWHERE b > {} AND b <= {} AND c > {} AND c <= {} WHERE a == 0;'.format(
|
]
|
||||||
to_select, b_range_start, b_range_end, c_range_start, c_range_end), False, expected)
|
self.query(
|
||||||
|
"SELECT {} from tab_02473 PREWHERE b > {} AND b <= {} AND c > {} AND c <= {} WHERE a == 0;".format(
|
||||||
|
to_select,
|
||||||
|
b_range_start,
|
||||||
|
b_range_end,
|
||||||
|
c_range_start,
|
||||||
|
c_range_end,
|
||||||
|
),
|
||||||
|
False,
|
||||||
|
expected,
|
||||||
|
)
|
||||||
|
|
||||||
expected = all_data[
|
expected = all_data[
|
||||||
(all_data.a == 0) &
|
(all_data.a == 0)
|
||||||
(all_data.b > b_range_start) &
|
& (all_data.b > b_range_start)
|
||||||
(all_data.b <= b_range_end) &
|
& (all_data.b <= b_range_end)
|
||||||
(all_data.c > c_range_start) &
|
& (all_data.c > c_range_start)
|
||||||
(all_data.c <= c_range_end) &
|
& (all_data.c <= c_range_end)
|
||||||
(all_data.d > d_range_start) &
|
& (all_data.d > d_range_start)
|
||||||
(all_data.d <= d_range_end)]
|
& (all_data.d <= d_range_end)
|
||||||
self.query('SELECT {} from tab_02473 PREWHERE b > {} AND b <= {} AND c > {} AND c <= {} AND d > {} AND d <= {} WHERE a == 0;'.format(
|
]
|
||||||
to_select, b_range_start, b_range_end, c_range_start, c_range_end, d_range_start, d_range_end), False, expected)
|
self.query(
|
||||||
|
"SELECT {} from tab_02473 PREWHERE b > {} AND b <= {} AND c > {} AND c <= {} AND d > {} AND d <= {} WHERE a == 0;".format(
|
||||||
|
to_select,
|
||||||
|
b_range_start,
|
||||||
|
b_range_end,
|
||||||
|
c_range_start,
|
||||||
|
c_range_end,
|
||||||
|
d_range_start,
|
||||||
|
d_range_end,
|
||||||
|
),
|
||||||
|
False,
|
||||||
|
expected,
|
||||||
|
)
|
||||||
|
|
||||||
def run_test(self, c_range_start, c_range_end, d_range_start, d_range_end):
|
def run_test(self, c_range_start, c_range_end, d_range_start, d_range_end):
|
||||||
self.repro_queries = []
|
self.repro_queries = []
|
||||||
|
|
||||||
self.query('''
|
self.query(
|
||||||
|
"""
|
||||||
CREATE TABLE tab_02473 (a Int8, b Int32, c Int32, d Int32, PRIMARY KEY (a))
|
CREATE TABLE tab_02473 (a Int8, b Int32, c Int32, d Int32, PRIMARY KEY (a))
|
||||||
ENGINE = MergeTree() ORDER BY (a, b)
|
ENGINE = MergeTree() ORDER BY (a, b)
|
||||||
SETTINGS min_bytes_for_wide_part = 0, index_granularity = {};'''.format(self.index_granularity))
|
SETTINGS min_bytes_for_wide_part = 0, index_granularity = {};""".format(
|
||||||
|
self.index_granularity
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
self.query('INSERT INTO tab_02473 select 0, number+1, number+1, number+1 FROM numbers({});'.format(self.total_rows))
|
self.query(
|
||||||
|
"INSERT INTO tab_02473 select 0, number+1, number+1, number+1 FROM numbers({});".format(
|
||||||
|
self.total_rows
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
client = ClickHouseClient()
|
client = ClickHouseClient()
|
||||||
all_data = client.query_return_df("SELECT a, b, c, d, 1 as e FROM tab_02473 FORMAT TabSeparatedWithNames;")
|
all_data = client.query_return_df(
|
||||||
|
"SELECT a, b, c, d, 1 as e FROM tab_02473 FORMAT TabSeparatedWithNames;"
|
||||||
|
)
|
||||||
|
|
||||||
self.query('OPTIMIZE TABLE tab_02473 FINAL SETTINGS mutations_sync=2;')
|
self.query("OPTIMIZE TABLE tab_02473 FINAL SETTINGS mutations_sync=2;")
|
||||||
|
|
||||||
# After all data has been written add a column with default value
|
# After all data has been written add a column with default value
|
||||||
self.query('ALTER TABLE tab_02473 ADD COLUMN e Int64 DEFAULT 1;')
|
self.query("ALTER TABLE tab_02473 ADD COLUMN e Int64 DEFAULT 1;")
|
||||||
|
|
||||||
self.check_data(all_data, c_range_start, c_range_end, d_range_start, d_range_end)
|
self.check_data(
|
||||||
|
all_data, c_range_start, c_range_end, d_range_start, d_range_end
|
||||||
self.query('DROP TABLE tab_02473;')
|
)
|
||||||
|
|
||||||
|
self.query("DROP TABLE tab_02473;")
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
# Enable multiple prewhere read steps
|
# Enable multiple prewhere read steps
|
||||||
url = os.environ['CLICKHOUSE_URL'] + '&enable_multiple_prewhere_read_steps=1&move_all_conditions_to_prewhere=0&max_threads=1'
|
url = (
|
||||||
|
os.environ["CLICKHOUSE_URL"]
|
||||||
|
+ "&enable_multiple_prewhere_read_steps=1&move_all_conditions_to_prewhere=0&max_threads=1"
|
||||||
|
)
|
||||||
|
|
||||||
default_index_granularity = 10;
|
default_index_granularity = 10
|
||||||
total_rows = 8 * default_index_granularity
|
total_rows = 8 * default_index_granularity
|
||||||
step = default_index_granularity
|
step = default_index_granularity
|
||||||
session = requests.Session()
|
session = requests.Session()
|
||||||
@ -120,12 +166,19 @@ def main():
|
|||||||
tester = Tester(session, url, index_granularity, total_rows)
|
tester = Tester(session, url, index_granularity, total_rows)
|
||||||
# Test combinations of ranges of columns c and d
|
# Test combinations of ranges of columns c and d
|
||||||
for c_range_start in range(0, total_rows, int(2.3 * step)):
|
for c_range_start in range(0, total_rows, int(2.3 * step)):
|
||||||
for c_range_end in range(c_range_start + 3 * step, total_rows, int(2.1 * step)):
|
for c_range_end in range(
|
||||||
for d_range_start in range(int(0.5 * step), total_rows, int(2.7 * step)):
|
c_range_start + 3 * step, total_rows, int(2.1 * step)
|
||||||
for d_range_end in range(d_range_start + 3 * step, total_rows, int(2.2 * step)):
|
):
|
||||||
tester.run_test(c_range_start, c_range_end, d_range_start, d_range_end)
|
for d_range_start in range(
|
||||||
|
int(0.5 * step), total_rows, int(2.7 * step)
|
||||||
|
):
|
||||||
|
for d_range_end in range(
|
||||||
|
d_range_start + 3 * step, total_rows, int(2.2 * step)
|
||||||
|
):
|
||||||
|
tester.run_test(
|
||||||
|
c_range_start, c_range_end, d_range_start, d_range_end
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
main()
|
main()
|
||||||
|
|
||||||
|
@ -8,7 +8,7 @@ import time
|
|||||||
from threading import Thread
|
from threading import Thread
|
||||||
|
|
||||||
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
CURDIR = os.path.dirname(os.path.realpath(__file__))
|
||||||
sys.path.insert(0, os.path.join(CURDIR, 'helpers'))
|
sys.path.insert(0, os.path.join(CURDIR, "helpers"))
|
||||||
|
|
||||||
from pure_http_client import ClickHouseClient
|
from pure_http_client import ClickHouseClient
|
||||||
|
|
||||||
@ -16,14 +16,23 @@ client = ClickHouseClient()
|
|||||||
|
|
||||||
# test table without partition
|
# test table without partition
|
||||||
client.query("DROP TABLE IF EXISTS t_async_insert_dedup_no_part NO DELAY")
|
client.query("DROP TABLE IF EXISTS t_async_insert_dedup_no_part NO DELAY")
|
||||||
client.query('''
|
client.query(
|
||||||
|
"""
|
||||||
CREATE TABLE t_async_insert_dedup_no_part (
|
CREATE TABLE t_async_insert_dedup_no_part (
|
||||||
KeyID UInt32
|
KeyID UInt32
|
||||||
) Engine = ReplicatedMergeTree('/clickhouse/tables/{shard}/{database}/t_async_insert_dedup', '{replica}')
|
) Engine = ReplicatedMergeTree('/clickhouse/tables/{shard}/{database}/t_async_insert_dedup', '{replica}')
|
||||||
ORDER BY (KeyID)
|
ORDER BY (KeyID)
|
||||||
''')
|
"""
|
||||||
|
)
|
||||||
|
|
||||||
client.query("insert into t_async_insert_dedup_no_part values (1), (2), (3), (4), (5)", settings = {"async_insert": 1, "wait_for_async_insert": 1, "insert_keeper_fault_injection_probability": 0})
|
client.query(
|
||||||
|
"insert into t_async_insert_dedup_no_part values (1), (2), (3), (4), (5)",
|
||||||
|
settings={
|
||||||
|
"async_insert": 1,
|
||||||
|
"wait_for_async_insert": 1,
|
||||||
|
"insert_keeper_fault_injection_probability": 0,
|
||||||
|
},
|
||||||
|
)
|
||||||
result = client.query("select count(*) from t_async_insert_dedup_no_part")
|
result = client.query("select count(*) from t_async_insert_dedup_no_part")
|
||||||
print(result, flush=True)
|
print(result, flush=True)
|
||||||
client.query("DROP TABLE IF EXISTS t_async_insert_dedup_no_part NO DELAY")
|
client.query("DROP TABLE IF EXISTS t_async_insert_dedup_no_part NO DELAY")
|
||||||
@ -32,7 +41,7 @@ client.query("DROP TABLE IF EXISTS t_async_insert_dedup_no_part NO DELAY")
|
|||||||
def generate_data(q, total_number):
|
def generate_data(q, total_number):
|
||||||
old_data = []
|
old_data = []
|
||||||
max_chunk_size = 30
|
max_chunk_size = 30
|
||||||
partitions = ['2022-11-11 10:10:10', '2022-12-12 10:10:10']
|
partitions = ["2022-11-11 10:10:10", "2022-12-12 10:10:10"]
|
||||||
last_number = 0
|
last_number = 0
|
||||||
while True:
|
while True:
|
||||||
dup_simulate = random.randint(0, 3)
|
dup_simulate = random.randint(0, 3)
|
||||||
@ -65,24 +74,37 @@ def generate_data(q, total_number):
|
|||||||
# wait all the tasks is done.
|
# wait all the tasks is done.
|
||||||
q.join()
|
q.join()
|
||||||
|
|
||||||
|
|
||||||
def fetch_and_insert_data(q, client):
|
def fetch_and_insert_data(q, client):
|
||||||
while True:
|
while True:
|
||||||
insert = q.get()
|
insert = q.get()
|
||||||
client.query(insert, settings = {"async_insert": 1, "async_insert_deduplicate": 1, "wait_for_async_insert": 0, "async_insert_busy_timeout_ms": 1500, "insert_keeper_fault_injection_probability": 0})
|
client.query(
|
||||||
|
insert,
|
||||||
|
settings={
|
||||||
|
"async_insert": 1,
|
||||||
|
"async_insert_deduplicate": 1,
|
||||||
|
"wait_for_async_insert": 0,
|
||||||
|
"async_insert_busy_timeout_ms": 1500,
|
||||||
|
"insert_keeper_fault_injection_probability": 0,
|
||||||
|
},
|
||||||
|
)
|
||||||
q.task_done()
|
q.task_done()
|
||||||
sleep_time = random.randint(50, 500)
|
sleep_time = random.randint(50, 500)
|
||||||
time.sleep(sleep_time / 1000.0)
|
time.sleep(sleep_time / 1000.0)
|
||||||
|
|
||||||
|
|
||||||
# main process
|
# main process
|
||||||
client.query("DROP TABLE IF EXISTS t_async_insert_dedup NO DELAY")
|
client.query("DROP TABLE IF EXISTS t_async_insert_dedup NO DELAY")
|
||||||
client.query('''
|
client.query(
|
||||||
|
"""
|
||||||
CREATE TABLE t_async_insert_dedup (
|
CREATE TABLE t_async_insert_dedup (
|
||||||
EventDate DateTime,
|
EventDate DateTime,
|
||||||
KeyID UInt32
|
KeyID UInt32
|
||||||
) Engine = ReplicatedMergeTree('/clickhouse/tables/{shard}/{database}/t_async_insert_dedup', '{replica}')
|
) Engine = ReplicatedMergeTree('/clickhouse/tables/{shard}/{database}/t_async_insert_dedup', '{replica}')
|
||||||
PARTITION BY toYYYYMM(EventDate)
|
PARTITION BY toYYYYMM(EventDate)
|
||||||
ORDER BY (KeyID, EventDate) SETTINGS use_async_block_ids_cache = 1
|
ORDER BY (KeyID, EventDate) SETTINGS use_async_block_ids_cache = 1
|
||||||
''')
|
"""
|
||||||
|
)
|
||||||
|
|
||||||
q = queue.Queue(100)
|
q = queue.Queue(100)
|
||||||
total_number = 10000
|
total_number = 10000
|
||||||
@ -125,11 +147,15 @@ while True:
|
|||||||
print(len(result), flush=True)
|
print(len(result), flush=True)
|
||||||
break
|
break
|
||||||
|
|
||||||
result = client.query("SELECT value FROM system.metrics where metric = 'AsyncInsertCacheSize'")
|
result = client.query(
|
||||||
|
"SELECT value FROM system.metrics where metric = 'AsyncInsertCacheSize'"
|
||||||
|
)
|
||||||
result = int(result.split()[0])
|
result = int(result.split()[0])
|
||||||
if result <= 0:
|
if result <= 0:
|
||||||
raise Exception(f"AsyncInsertCacheSize should > 0, but got {result}")
|
raise Exception(f"AsyncInsertCacheSize should > 0, but got {result}")
|
||||||
result = client.query("SELECT value FROM system.events where event = 'AsyncInsertCacheHits'")
|
result = client.query(
|
||||||
|
"SELECT value FROM system.events where event = 'AsyncInsertCacheHits'"
|
||||||
|
)
|
||||||
result = int(result.split()[0])
|
result = int(result.split()[0])
|
||||||
if result <= 0:
|
if result <= 0:
|
||||||
raise Exception(f"AsyncInsertCacheHits should > 0, but got {result}")
|
raise Exception(f"AsyncInsertCacheHits should > 0, but got {result}")
|
||||||
|
@ -19,9 +19,9 @@ import tenacity
|
|||||||
import xmltodict
|
import xmltodict
|
||||||
import yaml
|
import yaml
|
||||||
|
|
||||||
SELECT_VERSION = r'SELECT version()'
|
SELECT_VERSION = r"SELECT version()"
|
||||||
|
|
||||||
SELECT_UPTIME = r'''
|
SELECT_UPTIME = r"""
|
||||||
{% if version_ge('21.3') -%}
|
{% if version_ge('21.3') -%}
|
||||||
SELECT formatReadableTimeDelta(uptime())
|
SELECT formatReadableTimeDelta(uptime())
|
||||||
{% else -%}
|
{% else -%}
|
||||||
@ -29,18 +29,18 @@ SELECT
|
|||||||
toString(floor(uptime() / 3600 / 24)) || ' days ' ||
|
toString(floor(uptime() / 3600 / 24)) || ' days ' ||
|
||||||
toString(floor(uptime() % (24 * 3600) / 3600, 1)) || ' hours'
|
toString(floor(uptime() % (24 * 3600) / 3600, 1)) || ' hours'
|
||||||
{% endif -%}
|
{% endif -%}
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_SYSTEM_TABLES = "SELECT name FROM system.tables WHERE database = 'system'"
|
SELECT_SYSTEM_TABLES = "SELECT name FROM system.tables WHERE database = 'system'"
|
||||||
|
|
||||||
SELECT_DATABASE_ENGINES = r'''SELECT
|
SELECT_DATABASE_ENGINES = r"""SELECT
|
||||||
engine,
|
engine,
|
||||||
count() "count"
|
count() "count"
|
||||||
FROM system.databases
|
FROM system.databases
|
||||||
GROUP BY engine
|
GROUP BY engine
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_DATABASES = r'''SELECT
|
SELECT_DATABASES = r"""SELECT
|
||||||
name,
|
name,
|
||||||
engine,
|
engine,
|
||||||
tables,
|
tables,
|
||||||
@ -62,17 +62,17 @@ LEFT JOIN
|
|||||||
) AS db_stats ON db.name = db_stats.database
|
) AS db_stats ON db.name = db_stats.database
|
||||||
ORDER BY bytes_on_disk DESC
|
ORDER BY bytes_on_disk DESC
|
||||||
LIMIT 10
|
LIMIT 10
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_TABLE_ENGINES = r'''SELECT
|
SELECT_TABLE_ENGINES = r"""SELECT
|
||||||
engine,
|
engine,
|
||||||
count() "count"
|
count() "count"
|
||||||
FROM system.tables
|
FROM system.tables
|
||||||
WHERE database != 'system'
|
WHERE database != 'system'
|
||||||
GROUP BY engine
|
GROUP BY engine
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_DICTIONARIES = r'''SELECT
|
SELECT_DICTIONARIES = r"""SELECT
|
||||||
source,
|
source,
|
||||||
type,
|
type,
|
||||||
status,
|
status,
|
||||||
@ -80,13 +80,13 @@ SELECT_DICTIONARIES = r'''SELECT
|
|||||||
FROM system.dictionaries
|
FROM system.dictionaries
|
||||||
GROUP BY source, type, status
|
GROUP BY source, type, status
|
||||||
ORDER BY status DESC, source
|
ORDER BY status DESC, source
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_ACCESS = "SHOW ACCESS"
|
SELECT_ACCESS = "SHOW ACCESS"
|
||||||
|
|
||||||
SELECT_QUOTA_USAGE = "SHOW QUOTA"
|
SELECT_QUOTA_USAGE = "SHOW QUOTA"
|
||||||
|
|
||||||
SELECT_REPLICAS = r'''SELECT
|
SELECT_REPLICAS = r"""SELECT
|
||||||
database,
|
database,
|
||||||
table,
|
table,
|
||||||
is_leader,
|
is_leader,
|
||||||
@ -98,9 +98,9 @@ SELECT_REPLICAS = r'''SELECT
|
|||||||
FROM system.replicas
|
FROM system.replicas
|
||||||
ORDER BY absolute_delay DESC
|
ORDER BY absolute_delay DESC
|
||||||
LIMIT 10
|
LIMIT 10
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_REPLICATION_QUEUE = r'''SELECT
|
SELECT_REPLICATION_QUEUE = r"""SELECT
|
||||||
database,
|
database,
|
||||||
table,
|
table,
|
||||||
replica_name,
|
replica_name,
|
||||||
@ -121,9 +121,9 @@ SELECT_REPLICATION_QUEUE = r'''SELECT
|
|||||||
FROM system.replication_queue
|
FROM system.replication_queue
|
||||||
ORDER BY create_time ASC
|
ORDER BY create_time ASC
|
||||||
LIMIT 20
|
LIMIT 20
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_REPLICATED_FETCHES = r'''SELECT
|
SELECT_REPLICATED_FETCHES = r"""SELECT
|
||||||
database,
|
database,
|
||||||
table,
|
table,
|
||||||
round(elapsed, 1) "elapsed",
|
round(elapsed, 1) "elapsed",
|
||||||
@ -140,9 +140,9 @@ SELECT_REPLICATED_FETCHES = r'''SELECT
|
|||||||
to_detached,
|
to_detached,
|
||||||
thread_id
|
thread_id
|
||||||
FROM system.replicated_fetches
|
FROM system.replicated_fetches
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_PARTS_PER_TABLE = r'''SELECT
|
SELECT_PARTS_PER_TABLE = r"""SELECT
|
||||||
database,
|
database,
|
||||||
table,
|
table,
|
||||||
count() "partitions",
|
count() "partitions",
|
||||||
@ -162,9 +162,9 @@ FROM
|
|||||||
GROUP BY database, table
|
GROUP BY database, table
|
||||||
ORDER BY max_parts_per_partition DESC
|
ORDER BY max_parts_per_partition DESC
|
||||||
LIMIT 10
|
LIMIT 10
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_MERGES = r'''SELECT
|
SELECT_MERGES = r"""SELECT
|
||||||
database,
|
database,
|
||||||
table,
|
table,
|
||||||
round(elapsed, 1) "elapsed",
|
round(elapsed, 1) "elapsed",
|
||||||
@ -187,9 +187,9 @@ SELECT_MERGES = r'''SELECT
|
|||||||
formatReadableSize(memory_usage) "memory_usage"
|
formatReadableSize(memory_usage) "memory_usage"
|
||||||
{% endif -%}
|
{% endif -%}
|
||||||
FROM system.merges
|
FROM system.merges
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_MUTATIONS = r'''SELECT
|
SELECT_MUTATIONS = r"""SELECT
|
||||||
database,
|
database,
|
||||||
table,
|
table,
|
||||||
mutation_id,
|
mutation_id,
|
||||||
@ -206,9 +206,9 @@ SELECT_MUTATIONS = r'''SELECT
|
|||||||
FROM system.mutations
|
FROM system.mutations
|
||||||
WHERE NOT is_done
|
WHERE NOT is_done
|
||||||
ORDER BY create_time DESC
|
ORDER BY create_time DESC
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_RECENT_DATA_PARTS = r'''SELECT
|
SELECT_RECENT_DATA_PARTS = r"""SELECT
|
||||||
database,
|
database,
|
||||||
table,
|
table,
|
||||||
engine,
|
engine,
|
||||||
@ -242,9 +242,9 @@ SELECT_RECENT_DATA_PARTS = r'''SELECT
|
|||||||
FROM system.parts
|
FROM system.parts
|
||||||
WHERE modification_time > now() - INTERVAL 3 MINUTE
|
WHERE modification_time > now() - INTERVAL 3 MINUTE
|
||||||
ORDER BY modification_time DESC
|
ORDER BY modification_time DESC
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_DETACHED_DATA_PARTS = r'''SELECT
|
SELECT_DETACHED_DATA_PARTS = r"""SELECT
|
||||||
database,
|
database,
|
||||||
table,
|
table,
|
||||||
partition_id,
|
partition_id,
|
||||||
@ -255,9 +255,9 @@ SELECT_DETACHED_DATA_PARTS = r'''SELECT
|
|||||||
max_block_number,
|
max_block_number,
|
||||||
level
|
level
|
||||||
FROM system.detached_parts
|
FROM system.detached_parts
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_PROCESSES = r'''SELECT
|
SELECT_PROCESSES = r"""SELECT
|
||||||
elapsed,
|
elapsed,
|
||||||
query_id,
|
query_id,
|
||||||
{% if normalize_queries -%}
|
{% if normalize_queries -%}
|
||||||
@ -285,9 +285,9 @@ SELECT_PROCESSES = r'''SELECT
|
|||||||
{% endif -%}
|
{% endif -%}
|
||||||
FROM system.processes
|
FROM system.processes
|
||||||
ORDER BY elapsed DESC
|
ORDER BY elapsed DESC
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_TOP_QUERIES_BY_DURATION = r'''SELECT
|
SELECT_TOP_QUERIES_BY_DURATION = r"""SELECT
|
||||||
type,
|
type,
|
||||||
query_start_time,
|
query_start_time,
|
||||||
query_duration_ms,
|
query_duration_ms,
|
||||||
@ -339,9 +339,9 @@ WHERE type != 'QueryStart'
|
|||||||
AND event_time >= now() - INTERVAL 1 DAY
|
AND event_time >= now() - INTERVAL 1 DAY
|
||||||
ORDER BY query_duration_ms DESC
|
ORDER BY query_duration_ms DESC
|
||||||
LIMIT 10
|
LIMIT 10
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_TOP_QUERIES_BY_MEMORY_USAGE = r'''SELECT
|
SELECT_TOP_QUERIES_BY_MEMORY_USAGE = r"""SELECT
|
||||||
type,
|
type,
|
||||||
query_start_time,
|
query_start_time,
|
||||||
query_duration_ms,
|
query_duration_ms,
|
||||||
@ -393,9 +393,9 @@ WHERE type != 'QueryStart'
|
|||||||
AND event_time >= now() - INTERVAL 1 DAY
|
AND event_time >= now() - INTERVAL 1 DAY
|
||||||
ORDER BY memory_usage DESC
|
ORDER BY memory_usage DESC
|
||||||
LIMIT 10
|
LIMIT 10
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_FAILED_QUERIES = r'''SELECT
|
SELECT_FAILED_QUERIES = r"""SELECT
|
||||||
type,
|
type,
|
||||||
query_start_time,
|
query_start_time,
|
||||||
query_duration_ms,
|
query_duration_ms,
|
||||||
@ -448,9 +448,9 @@ WHERE type != 'QueryStart'
|
|||||||
AND exception != ''
|
AND exception != ''
|
||||||
ORDER BY query_start_time DESC
|
ORDER BY query_start_time DESC
|
||||||
LIMIT 10
|
LIMIT 10
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_STACK_TRACES = r'''SELECT
|
SELECT_STACK_TRACES = r"""SELECT
|
||||||
'\n' || arrayStringConcat(
|
'\n' || arrayStringConcat(
|
||||||
arrayMap(
|
arrayMap(
|
||||||
x,
|
x,
|
||||||
@ -459,9 +459,9 @@ SELECT_STACK_TRACES = r'''SELECT
|
|||||||
arrayMap(x -> demangle(addressToSymbol(x)), trace)),
|
arrayMap(x -> demangle(addressToSymbol(x)), trace)),
|
||||||
'\n') AS trace
|
'\n') AS trace
|
||||||
FROM system.stack_trace
|
FROM system.stack_trace
|
||||||
'''
|
"""
|
||||||
|
|
||||||
SELECT_CRASH_LOG = r'''SELECT
|
SELECT_CRASH_LOG = r"""SELECT
|
||||||
event_time,
|
event_time,
|
||||||
signal,
|
signal,
|
||||||
thread_id,
|
thread_id,
|
||||||
@ -470,7 +470,7 @@ SELECT_CRASH_LOG = r'''SELECT
|
|||||||
version
|
version
|
||||||
FROM system.crash_log
|
FROM system.crash_log
|
||||||
ORDER BY event_time DESC
|
ORDER BY event_time DESC
|
||||||
'''
|
"""
|
||||||
|
|
||||||
|
|
||||||
def retry(exception_types, max_attempts=5, max_interval=5):
|
def retry(exception_types, max_attempts=5, max_interval=5):
|
||||||
@ -481,7 +481,8 @@ def retry(exception_types, max_attempts=5, max_interval=5):
|
|||||||
retry=tenacity.retry_if_exception_type(exception_types),
|
retry=tenacity.retry_if_exception_type(exception_types),
|
||||||
wait=tenacity.wait_random_exponential(multiplier=0.5, max=max_interval),
|
wait=tenacity.wait_random_exponential(multiplier=0.5, max=max_interval),
|
||||||
stop=tenacity.stop_after_attempt(max_attempts),
|
stop=tenacity.stop_after_attempt(max_attempts),
|
||||||
reraise=True)
|
reraise=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
class ClickhouseError(Exception):
|
class ClickhouseError(Exception):
|
||||||
@ -502,9 +503,9 @@ class ClickhouseClient:
|
|||||||
def __init__(self, *, host="localhost", port=8123, user="default", password):
|
def __init__(self, *, host="localhost", port=8123, user="default", password):
|
||||||
self._session = requests.Session()
|
self._session = requests.Session()
|
||||||
if user:
|
if user:
|
||||||
self._session.headers['X-ClickHouse-User'] = user
|
self._session.headers["X-ClickHouse-User"] = user
|
||||||
self._session.headers['X-ClickHouse-Key'] = password
|
self._session.headers["X-ClickHouse-Key"] = password
|
||||||
self._url = f'http://{host}:{port}'
|
self._url = f"http://{host}:{port}"
|
||||||
self._timeout = 60
|
self._timeout = 60
|
||||||
self._ch_version = None
|
self._ch_version = None
|
||||||
|
|
||||||
@ -516,7 +517,16 @@ class ClickhouseClient:
|
|||||||
return self._ch_version
|
return self._ch_version
|
||||||
|
|
||||||
@retry(requests.exceptions.ConnectionError)
|
@retry(requests.exceptions.ConnectionError)
|
||||||
def query(self, query, query_args=None, format=None, post_data=None, timeout=None, echo=False, dry_run=False):
|
def query(
|
||||||
|
self,
|
||||||
|
query,
|
||||||
|
query_args=None,
|
||||||
|
format=None,
|
||||||
|
post_data=None,
|
||||||
|
timeout=None,
|
||||||
|
echo=False,
|
||||||
|
dry_run=False,
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Execute query.
|
Execute query.
|
||||||
"""
|
"""
|
||||||
@ -524,28 +534,30 @@ class ClickhouseClient:
|
|||||||
query = self.render_query(query, **query_args)
|
query = self.render_query(query, **query_args)
|
||||||
|
|
||||||
if format:
|
if format:
|
||||||
query += f' FORMAT {format}'
|
query += f" FORMAT {format}"
|
||||||
|
|
||||||
if timeout is None:
|
if timeout is None:
|
||||||
timeout = self._timeout
|
timeout = self._timeout
|
||||||
|
|
||||||
if echo:
|
if echo:
|
||||||
print(sqlparse.format(query, reindent=True), '\n')
|
print(sqlparse.format(query, reindent=True), "\n")
|
||||||
|
|
||||||
if dry_run:
|
if dry_run:
|
||||||
return None
|
return None
|
||||||
|
|
||||||
try:
|
try:
|
||||||
response = self._session.post(self._url,
|
response = self._session.post(
|
||||||
|
self._url,
|
||||||
params={
|
params={
|
||||||
'query': query,
|
"query": query,
|
||||||
},
|
},
|
||||||
json=post_data,
|
json=post_data,
|
||||||
timeout=timeout)
|
timeout=timeout,
|
||||||
|
)
|
||||||
|
|
||||||
response.raise_for_status()
|
response.raise_for_status()
|
||||||
|
|
||||||
if format in ('JSON', 'JSONCompact'):
|
if format in ("JSON", "JSONCompact"):
|
||||||
return response.json()
|
return response.json()
|
||||||
|
|
||||||
return response.text.strip()
|
return response.text.strip()
|
||||||
@ -555,7 +567,9 @@ class ClickhouseClient:
|
|||||||
def render_query(self, query, **kwargs):
|
def render_query(self, query, **kwargs):
|
||||||
env = jinja2.Environment()
|
env = jinja2.Environment()
|
||||||
|
|
||||||
env.globals['version_ge'] = lambda version: version_ge(self.clickhouse_version, version)
|
env.globals["version_ge"] = lambda version: version_ge(
|
||||||
|
self.clickhouse_version, version
|
||||||
|
)
|
||||||
|
|
||||||
template = env.from_string(query)
|
template = env.from_string(query)
|
||||||
return template.render(kwargs)
|
return template.render(kwargs)
|
||||||
@ -578,11 +592,13 @@ class ClickhouseConfig:
|
|||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def load(cls):
|
def load(cls):
|
||||||
return ClickhouseConfig(cls._load_config('/var/lib/clickhouse/preprocessed_configs/config.xml'))
|
return ClickhouseConfig(
|
||||||
|
cls._load_config("/var/lib/clickhouse/preprocessed_configs/config.xml")
|
||||||
|
)
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _load_config(config_path):
|
def _load_config(config_path):
|
||||||
with open(config_path, 'r') as file:
|
with open(config_path, "r") as file:
|
||||||
return xmltodict.parse(file.read())
|
return xmltodict.parse(file.read())
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
@ -591,8 +607,8 @@ class ClickhouseConfig:
|
|||||||
for key, value in list(config.items()):
|
for key, value in list(config.items()):
|
||||||
if isinstance(value, MutableMapping):
|
if isinstance(value, MutableMapping):
|
||||||
cls._mask_secrets(config[key])
|
cls._mask_secrets(config[key])
|
||||||
elif key in ('password', 'secret_access_key', 'header', 'identity'):
|
elif key in ("password", "secret_access_key", "header", "identity"):
|
||||||
config[key] = '*****'
|
config[key] = "*****"
|
||||||
|
|
||||||
|
|
||||||
class DiagnosticsData:
|
class DiagnosticsData:
|
||||||
@ -603,53 +619,53 @@ class DiagnosticsData:
|
|||||||
def __init__(self, args):
|
def __init__(self, args):
|
||||||
self.args = args
|
self.args = args
|
||||||
self.host = args.host
|
self.host = args.host
|
||||||
self._sections = [{'section': None, 'data': {}}]
|
self._sections = [{"section": None, "data": {}}]
|
||||||
|
|
||||||
def add_string(self, name, value, section=None):
|
def add_string(self, name, value, section=None):
|
||||||
self._section(section)[name] = {
|
self._section(section)[name] = {
|
||||||
'type': 'string',
|
"type": "string",
|
||||||
'value': value,
|
"value": value,
|
||||||
}
|
}
|
||||||
|
|
||||||
def add_xml_document(self, name, document, section=None):
|
def add_xml_document(self, name, document, section=None):
|
||||||
self._section(section)[name] = {
|
self._section(section)[name] = {
|
||||||
'type': 'xml',
|
"type": "xml",
|
||||||
'value': document,
|
"value": document,
|
||||||
}
|
}
|
||||||
|
|
||||||
def add_query(self, name, query, result, section=None):
|
def add_query(self, name, query, result, section=None):
|
||||||
self._section(section)[name] = {
|
self._section(section)[name] = {
|
||||||
'type': 'query',
|
"type": "query",
|
||||||
'query': query,
|
"query": query,
|
||||||
'result': result,
|
"result": result,
|
||||||
}
|
}
|
||||||
|
|
||||||
def add_command(self, name, command, result, section=None):
|
def add_command(self, name, command, result, section=None):
|
||||||
self._section(section)[name] = {
|
self._section(section)[name] = {
|
||||||
'type': 'command',
|
"type": "command",
|
||||||
'command': command,
|
"command": command,
|
||||||
'result': result,
|
"result": result,
|
||||||
}
|
}
|
||||||
|
|
||||||
def dump(self, format):
|
def dump(self, format):
|
||||||
if format.startswith('json'):
|
if format.startswith("json"):
|
||||||
result = self._dump_json()
|
result = self._dump_json()
|
||||||
elif format.startswith('yaml'):
|
elif format.startswith("yaml"):
|
||||||
result = self._dump_yaml()
|
result = self._dump_yaml()
|
||||||
else:
|
else:
|
||||||
result = self._dump_wiki()
|
result = self._dump_wiki()
|
||||||
|
|
||||||
if format.endswith('.gz'):
|
if format.endswith(".gz"):
|
||||||
compressor = gzip.GzipFile(mode='wb', fileobj=sys.stdout.buffer)
|
compressor = gzip.GzipFile(mode="wb", fileobj=sys.stdout.buffer)
|
||||||
compressor.write(result.encode())
|
compressor.write(result.encode())
|
||||||
else:
|
else:
|
||||||
print(result)
|
print(result)
|
||||||
|
|
||||||
def _section(self, name=None):
|
def _section(self, name=None):
|
||||||
if self._sections[-1]['section'] != name:
|
if self._sections[-1]["section"] != name:
|
||||||
self._sections.append({'section': name, 'data': {}})
|
self._sections.append({"section": name, "data": {}})
|
||||||
|
|
||||||
return self._sections[-1]['data']
|
return self._sections[-1]["data"]
|
||||||
|
|
||||||
def _dump_json(self):
|
def _dump_json(self):
|
||||||
"""
|
"""
|
||||||
@ -669,85 +685,85 @@ class DiagnosticsData:
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
def _write_title(buffer, value):
|
def _write_title(buffer, value):
|
||||||
buffer.write(f'### {value}\n')
|
buffer.write(f"### {value}\n")
|
||||||
|
|
||||||
def _write_subtitle(buffer, value):
|
def _write_subtitle(buffer, value):
|
||||||
buffer.write(f'#### {value}\n')
|
buffer.write(f"#### {value}\n")
|
||||||
|
|
||||||
def _write_string_item(buffer, name, item):
|
def _write_string_item(buffer, name, item):
|
||||||
value = item['value']
|
value = item["value"]
|
||||||
if value != '':
|
if value != "":
|
||||||
value = f'**{value}**'
|
value = f"**{value}**"
|
||||||
buffer.write(f'{name}: {value}\n')
|
buffer.write(f"{name}: {value}\n")
|
||||||
|
|
||||||
def _write_xml_item(buffer, section_name, name, item):
|
def _write_xml_item(buffer, section_name, name, item):
|
||||||
if section_name:
|
if section_name:
|
||||||
buffer.write(f'##### {name}\n')
|
buffer.write(f"##### {name}\n")
|
||||||
else:
|
else:
|
||||||
_write_subtitle(buffer, name)
|
_write_subtitle(buffer, name)
|
||||||
|
|
||||||
_write_result(buffer, item['value'], format='XML')
|
_write_result(buffer, item["value"], format="XML")
|
||||||
|
|
||||||
def _write_query_item(buffer, section_name, name, item):
|
def _write_query_item(buffer, section_name, name, item):
|
||||||
if section_name:
|
if section_name:
|
||||||
buffer.write(f'##### {name}\n')
|
buffer.write(f"##### {name}\n")
|
||||||
else:
|
else:
|
||||||
_write_subtitle(buffer, name)
|
_write_subtitle(buffer, name)
|
||||||
|
|
||||||
_write_query(buffer, item['query'])
|
_write_query(buffer, item["query"])
|
||||||
_write_result(buffer, item['result'])
|
_write_result(buffer, item["result"])
|
||||||
|
|
||||||
def _write_command_item(buffer, section_name, name, item):
|
def _write_command_item(buffer, section_name, name, item):
|
||||||
if section_name:
|
if section_name:
|
||||||
buffer.write(f'##### {name}\n')
|
buffer.write(f"##### {name}\n")
|
||||||
else:
|
else:
|
||||||
_write_subtitle(buffer, name)
|
_write_subtitle(buffer, name)
|
||||||
|
|
||||||
_write_command(buffer, item['command'])
|
_write_command(buffer, item["command"])
|
||||||
_write_result(buffer, item['result'])
|
_write_result(buffer, item["result"])
|
||||||
|
|
||||||
def _write_unknown_item(buffer, section_name, name, item):
|
def _write_unknown_item(buffer, section_name, name, item):
|
||||||
if section_name:
|
if section_name:
|
||||||
buffer.write(f'**{name}**\n')
|
buffer.write(f"**{name}**\n")
|
||||||
else:
|
else:
|
||||||
_write_subtitle(buffer, name)
|
_write_subtitle(buffer, name)
|
||||||
|
|
||||||
json.dump(item, buffer, indent=2)
|
json.dump(item, buffer, indent=2)
|
||||||
|
|
||||||
def _write_query(buffer, query):
|
def _write_query(buffer, query):
|
||||||
buffer.write('**query**\n')
|
buffer.write("**query**\n")
|
||||||
buffer.write('```sql\n')
|
buffer.write("```sql\n")
|
||||||
buffer.write(query)
|
buffer.write(query)
|
||||||
buffer.write('\n```\n')
|
buffer.write("\n```\n")
|
||||||
|
|
||||||
def _write_command(buffer, command):
|
def _write_command(buffer, command):
|
||||||
buffer.write('**command**\n')
|
buffer.write("**command**\n")
|
||||||
buffer.write('```\n')
|
buffer.write("```\n")
|
||||||
buffer.write(command)
|
buffer.write(command)
|
||||||
buffer.write('\n```\n')
|
buffer.write("\n```\n")
|
||||||
|
|
||||||
def _write_result(buffer, result, format=None):
|
def _write_result(buffer, result, format=None):
|
||||||
buffer.write('**result**\n')
|
buffer.write("**result**\n")
|
||||||
buffer.write(f'```{format}\n' if format else '```\n')
|
buffer.write(f"```{format}\n" if format else "```\n")
|
||||||
buffer.write(result)
|
buffer.write(result)
|
||||||
buffer.write('\n```\n')
|
buffer.write("\n```\n")
|
||||||
|
|
||||||
buffer = io.StringIO()
|
buffer = io.StringIO()
|
||||||
|
|
||||||
_write_title(buffer, f'Diagnostics data for host {self.host}')
|
_write_title(buffer, f"Diagnostics data for host {self.host}")
|
||||||
for section in self._sections:
|
for section in self._sections:
|
||||||
section_name = section['section']
|
section_name = section["section"]
|
||||||
if section_name:
|
if section_name:
|
||||||
_write_subtitle(buffer, section_name)
|
_write_subtitle(buffer, section_name)
|
||||||
|
|
||||||
for name, item in section['data'].items():
|
for name, item in section["data"].items():
|
||||||
if item['type'] == 'string':
|
if item["type"] == "string":
|
||||||
_write_string_item(buffer, name, item)
|
_write_string_item(buffer, name, item)
|
||||||
elif item['type'] == 'query':
|
elif item["type"] == "query":
|
||||||
_write_query_item(buffer, section_name, name, item)
|
_write_query_item(buffer, section_name, name, item)
|
||||||
elif item['type'] == 'command':
|
elif item["type"] == "command":
|
||||||
_write_command_item(buffer, section_name, name, item)
|
_write_command_item(buffer, section_name, name, item)
|
||||||
elif item['type'] == 'xml':
|
elif item["type"] == "xml":
|
||||||
_write_xml_item(buffer, section_name, name, item)
|
_write_xml_item(buffer, section_name, name, item)
|
||||||
else:
|
else:
|
||||||
_write_unknown_item(buffer, section_name, name, item)
|
_write_unknown_item(buffer, section_name, name, item)
|
||||||
@ -760,126 +776,196 @@ def main():
|
|||||||
Program entry point.
|
Program entry point.
|
||||||
"""
|
"""
|
||||||
args = parse_args()
|
args = parse_args()
|
||||||
timestamp = datetime.strftime(datetime.now(), '%Y-%m-%d %H:%M:%S')
|
timestamp = datetime.strftime(datetime.now(), "%Y-%m-%d %H:%M:%S")
|
||||||
client = ClickhouseClient(host=args.host, port=args.port, user=args.user, password=args.password)
|
client = ClickhouseClient(
|
||||||
|
host=args.host, port=args.port, user=args.user, password=args.password
|
||||||
|
)
|
||||||
ch_config = ClickhouseConfig.load()
|
ch_config = ClickhouseConfig.load()
|
||||||
version = client.clickhouse_version
|
version = client.clickhouse_version
|
||||||
system_tables = [row[0] for row in execute_query(client, SELECT_SYSTEM_TABLES, format='JSONCompact')['data']]
|
system_tables = [
|
||||||
|
row[0]
|
||||||
|
for row in execute_query(client, SELECT_SYSTEM_TABLES, format="JSONCompact")[
|
||||||
|
"data"
|
||||||
|
]
|
||||||
|
]
|
||||||
|
|
||||||
diagnostics = DiagnosticsData(args)
|
diagnostics = DiagnosticsData(args)
|
||||||
diagnostics.add_string('Version', version)
|
diagnostics.add_string("Version", version)
|
||||||
diagnostics.add_string('Timestamp', timestamp)
|
diagnostics.add_string("Timestamp", timestamp)
|
||||||
diagnostics.add_string('Uptime', execute_query(client, SELECT_UPTIME))
|
diagnostics.add_string("Uptime", execute_query(client, SELECT_UPTIME))
|
||||||
|
|
||||||
diagnostics.add_xml_document('ClickHouse configuration', ch_config.dump())
|
diagnostics.add_xml_document("ClickHouse configuration", ch_config.dump())
|
||||||
|
|
||||||
if version_ge(version, '20.8'):
|
if version_ge(version, "20.8"):
|
||||||
add_query(diagnostics, 'Access configuration',
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Access configuration",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_ACCESS,
|
query=SELECT_ACCESS,
|
||||||
format='TSVRaw')
|
format="TSVRaw",
|
||||||
add_query(diagnostics, 'Quotas',
|
)
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Quotas",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_QUOTA_USAGE,
|
query=SELECT_QUOTA_USAGE,
|
||||||
format='Vertical')
|
format="Vertical",
|
||||||
|
)
|
||||||
|
|
||||||
add_query(diagnostics, 'Database engines',
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Database engines",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_DATABASE_ENGINES,
|
query=SELECT_DATABASE_ENGINES,
|
||||||
format='PrettyCompactNoEscapes',
|
format="PrettyCompactNoEscapes",
|
||||||
section='Schema')
|
section="Schema",
|
||||||
add_query(diagnostics, 'Databases (top 10 by size)',
|
)
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Databases (top 10 by size)",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_DATABASES,
|
query=SELECT_DATABASES,
|
||||||
format='PrettyCompactNoEscapes',
|
format="PrettyCompactNoEscapes",
|
||||||
section='Schema')
|
section="Schema",
|
||||||
add_query(diagnostics, 'Table engines',
|
)
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Table engines",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_TABLE_ENGINES,
|
query=SELECT_TABLE_ENGINES,
|
||||||
format='PrettyCompactNoEscapes',
|
format="PrettyCompactNoEscapes",
|
||||||
section='Schema')
|
section="Schema",
|
||||||
add_query(diagnostics, 'Dictionaries',
|
)
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Dictionaries",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_DICTIONARIES,
|
query=SELECT_DICTIONARIES,
|
||||||
format='PrettyCompactNoEscapes',
|
format="PrettyCompactNoEscapes",
|
||||||
section='Schema')
|
section="Schema",
|
||||||
|
)
|
||||||
|
|
||||||
add_query(diagnostics, 'Replicated tables (top 10 by absolute delay)',
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Replicated tables (top 10 by absolute delay)",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_REPLICAS,
|
query=SELECT_REPLICAS,
|
||||||
format='PrettyCompactNoEscapes',
|
format="PrettyCompactNoEscapes",
|
||||||
section='Replication')
|
section="Replication",
|
||||||
add_query(diagnostics, 'Replication queue (top 20 oldest tasks)',
|
)
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Replication queue (top 20 oldest tasks)",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_REPLICATION_QUEUE,
|
query=SELECT_REPLICATION_QUEUE,
|
||||||
format='Vertical',
|
format="Vertical",
|
||||||
section='Replication')
|
section="Replication",
|
||||||
if version_ge(version, '21.3'):
|
)
|
||||||
add_query(diagnostics, 'Replicated fetches',
|
if version_ge(version, "21.3"):
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Replicated fetches",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_REPLICATED_FETCHES,
|
query=SELECT_REPLICATED_FETCHES,
|
||||||
format='Vertical',
|
format="Vertical",
|
||||||
section='Replication')
|
section="Replication",
|
||||||
|
)
|
||||||
|
|
||||||
add_query(diagnostics, 'Top 10 tables by max parts per partition',
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Top 10 tables by max parts per partition",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_PARTS_PER_TABLE,
|
query=SELECT_PARTS_PER_TABLE,
|
||||||
format='PrettyCompactNoEscapes')
|
format="PrettyCompactNoEscapes",
|
||||||
add_query(diagnostics, 'Merges in progress',
|
)
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Merges in progress",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_MERGES,
|
query=SELECT_MERGES,
|
||||||
format='Vertical')
|
format="Vertical",
|
||||||
add_query(diagnostics, 'Mutations in progress',
|
)
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Mutations in progress",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_MUTATIONS,
|
query=SELECT_MUTATIONS,
|
||||||
format='Vertical')
|
format="Vertical",
|
||||||
add_query(diagnostics, 'Recent data parts (modification time within last 3 minutes)',
|
)
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Recent data parts (modification time within last 3 minutes)",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_RECENT_DATA_PARTS,
|
query=SELECT_RECENT_DATA_PARTS,
|
||||||
format='Vertical')
|
format="Vertical",
|
||||||
|
)
|
||||||
|
|
||||||
add_query(diagnostics, 'system.detached_parts',
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"system.detached_parts",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_DETACHED_DATA_PARTS,
|
query=SELECT_DETACHED_DATA_PARTS,
|
||||||
format='PrettyCompactNoEscapes',
|
format="PrettyCompactNoEscapes",
|
||||||
section='Detached data')
|
section="Detached data",
|
||||||
add_command(diagnostics, 'Disk space usage',
|
)
|
||||||
command='du -sh -L -c /var/lib/clickhouse/data/*/*/detached/* | sort -rsh',
|
add_command(
|
||||||
section='Detached data')
|
diagnostics,
|
||||||
|
"Disk space usage",
|
||||||
|
command="du -sh -L -c /var/lib/clickhouse/data/*/*/detached/* | sort -rsh",
|
||||||
|
section="Detached data",
|
||||||
|
)
|
||||||
|
|
||||||
add_query(diagnostics, 'Queries in progress (process list)',
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Queries in progress (process list)",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_PROCESSES,
|
query=SELECT_PROCESSES,
|
||||||
format='Vertical',
|
format="Vertical",
|
||||||
section='Queries')
|
section="Queries",
|
||||||
add_query(diagnostics, 'Top 10 queries by duration',
|
)
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Top 10 queries by duration",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_TOP_QUERIES_BY_DURATION,
|
query=SELECT_TOP_QUERIES_BY_DURATION,
|
||||||
format='Vertical',
|
format="Vertical",
|
||||||
section='Queries')
|
section="Queries",
|
||||||
add_query(diagnostics, 'Top 10 queries by memory usage',
|
)
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Top 10 queries by memory usage",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_TOP_QUERIES_BY_MEMORY_USAGE,
|
query=SELECT_TOP_QUERIES_BY_MEMORY_USAGE,
|
||||||
format='Vertical',
|
format="Vertical",
|
||||||
section='Queries')
|
section="Queries",
|
||||||
add_query(diagnostics, 'Last 10 failed queries',
|
)
|
||||||
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Last 10 failed queries",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_FAILED_QUERIES,
|
query=SELECT_FAILED_QUERIES,
|
||||||
format='Vertical',
|
format="Vertical",
|
||||||
section='Queries')
|
section="Queries",
|
||||||
|
)
|
||||||
|
|
||||||
add_query(diagnostics, 'Stack traces',
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Stack traces",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_STACK_TRACES,
|
query=SELECT_STACK_TRACES,
|
||||||
format='Vertical')
|
format="Vertical",
|
||||||
|
)
|
||||||
|
|
||||||
if 'crash_log' in system_tables:
|
if "crash_log" in system_tables:
|
||||||
add_query(diagnostics, 'Crash log',
|
add_query(
|
||||||
|
diagnostics,
|
||||||
|
"Crash log",
|
||||||
client=client,
|
client=client,
|
||||||
query=SELECT_CRASH_LOG,
|
query=SELECT_CRASH_LOG,
|
||||||
format='Vertical')
|
format="Vertical",
|
||||||
|
)
|
||||||
|
|
||||||
add_command(diagnostics, 'uname', 'uname -a')
|
add_command(diagnostics, "uname", "uname -a")
|
||||||
|
|
||||||
diagnostics.dump(args.format)
|
diagnostics.dump(args.format)
|
||||||
|
|
||||||
@ -889,29 +975,34 @@ def parse_args():
|
|||||||
Parse command-line arguments.
|
Parse command-line arguments.
|
||||||
"""
|
"""
|
||||||
parser = argparse.ArgumentParser()
|
parser = argparse.ArgumentParser()
|
||||||
parser.add_argument('--format',
|
parser.add_argument(
|
||||||
choices=['json', 'yaml', 'json.gz', 'yaml.gz', 'wiki', 'wiki.gz'],
|
"--format",
|
||||||
default='wiki')
|
choices=["json", "yaml", "json.gz", "yaml.gz", "wiki", "wiki.gz"],
|
||||||
parser.add_argument('--normalize-queries',
|
default="wiki",
|
||||||
action='store_true',
|
)
|
||||||
default=False)
|
parser.add_argument("--normalize-queries", action="store_true", default=False)
|
||||||
parser.add_argument('--host', dest="host", help="clickhouse host")
|
parser.add_argument("--host", dest="host", help="clickhouse host")
|
||||||
parser.add_argument('--port', dest="port", default=8123, help="clickhouse http port")
|
parser.add_argument(
|
||||||
parser.add_argument('--user', dest="user", default="default", help="clickhouse user")
|
"--port", dest="port", default=8123, help="clickhouse http port"
|
||||||
parser.add_argument('--password', dest="password", help="clickhouse password")
|
)
|
||||||
|
parser.add_argument(
|
||||||
|
"--user", dest="user", default="default", help="clickhouse user"
|
||||||
|
)
|
||||||
|
parser.add_argument("--password", dest="password", help="clickhouse password")
|
||||||
return parser.parse_args()
|
return parser.parse_args()
|
||||||
|
|
||||||
|
|
||||||
def add_query(diagnostics, name, client, query, format, section=None):
|
def add_query(diagnostics, name, client, query, format, section=None):
|
||||||
query_args = {
|
query_args = {
|
||||||
'normalize_queries': diagnostics.args.normalize_queries,
|
"normalize_queries": diagnostics.args.normalize_queries,
|
||||||
}
|
}
|
||||||
query = client.render_query(query, **query_args)
|
query = client.render_query(query, **query_args)
|
||||||
diagnostics.add_query(
|
diagnostics.add_query(
|
||||||
name=name,
|
name=name,
|
||||||
query=query,
|
query=query,
|
||||||
result=execute_query(client, query, render_query=False, format=format),
|
result=execute_query(client, query, render_query=False, format=format),
|
||||||
section=section)
|
section=section,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def execute_query(client, query, render_query=True, format=None):
|
def execute_query(client, query, render_query=True, format=None):
|
||||||
@ -926,14 +1017,18 @@ def execute_query(client, query, render_query=True, format=None):
|
|||||||
|
|
||||||
def add_command(diagnostics, name, command, section=None):
|
def add_command(diagnostics, name, command, section=None):
|
||||||
diagnostics.add_command(
|
diagnostics.add_command(
|
||||||
name=name,
|
name=name, command=command, result=execute_command(command), section=section
|
||||||
command=command,
|
)
|
||||||
result=execute_command(command),
|
|
||||||
section=section)
|
|
||||||
|
|
||||||
|
|
||||||
def execute_command(command, input=None):
|
def execute_command(command, input=None):
|
||||||
proc = subprocess.Popen(command, shell=True, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
proc = subprocess.Popen(
|
||||||
|
command,
|
||||||
|
shell=True,
|
||||||
|
stdin=subprocess.PIPE,
|
||||||
|
stdout=subprocess.PIPE,
|
||||||
|
stderr=subprocess.PIPE,
|
||||||
|
)
|
||||||
|
|
||||||
if isinstance(input, str):
|
if isinstance(input, str):
|
||||||
input = input.encode()
|
input = input.encode()
|
||||||
@ -941,7 +1036,7 @@ def execute_command(command, input=None):
|
|||||||
stdout, stderr = proc.communicate(input=input)
|
stdout, stderr = proc.communicate(input=input)
|
||||||
|
|
||||||
if proc.returncode:
|
if proc.returncode:
|
||||||
return f'failed with exit code {proc.returncode}\n{stderr.decode()}'
|
return f"failed with exit code {proc.returncode}\n{stderr.decode()}"
|
||||||
|
|
||||||
return stdout.decode()
|
return stdout.decode()
|
||||||
|
|
||||||
@ -957,8 +1052,8 @@ def parse_version(version):
|
|||||||
"""
|
"""
|
||||||
Parse version string.
|
Parse version string.
|
||||||
"""
|
"""
|
||||||
return [int(x) for x in version.strip().split('.') if x.isnumeric()]
|
return [int(x) for x in version.strip().split(".") if x.isnumeric()]
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == "__main__":
|
||||||
main()
|
main()
|
||||||
|
@ -35,32 +35,41 @@ class S3API(object):
|
|||||||
|
|
||||||
def call_back(x, y):
|
def call_back(x, y):
|
||||||
print("Uploaded {}/{} bytes".format(x, y))
|
print("Uploaded {}/{} bytes".format(x, y))
|
||||||
|
|
||||||
try:
|
try:
|
||||||
for i in range(chunkcount + 1):
|
for i in range(chunkcount + 1):
|
||||||
logging.info("Uploading chunk %s of %s", i, chunkcount + 1)
|
logging.info("Uploading chunk %s of %s", i, chunkcount + 1)
|
||||||
offset = chunksize * i
|
offset = chunksize * i
|
||||||
bytes_size = min(chunksize, filesize - offset)
|
bytes_size = min(chunksize, filesize - offset)
|
||||||
|
|
||||||
with open(file_path, 'r') as fp:
|
with open(file_path, "r") as fp:
|
||||||
fp.seek(offset)
|
fp.seek(offset)
|
||||||
key.upload_part_from_file(fp=fp, part_num=i+1,
|
key.upload_part_from_file(
|
||||||
size=bytes_size, cb=call_back,
|
fp=fp, part_num=i + 1, size=bytes_size, cb=call_back, num_cb=100
|
||||||
num_cb=100)
|
)
|
||||||
key.complete_upload()
|
key.complete_upload()
|
||||||
except Exception as ex:
|
except Exception as ex:
|
||||||
key.cancel_upload()
|
key.cancel_upload()
|
||||||
raise ex
|
raise ex
|
||||||
logging.info("Contents were set")
|
logging.info("Contents were set")
|
||||||
return "https://{bucket}.{mds_url}/{path}".format(
|
return "https://{bucket}.{mds_url}/{path}".format(
|
||||||
bucket=bucket_name, mds_url=self.mds_url, path=s3_path)
|
bucket=bucket_name, mds_url=self.mds_url, path=s3_path
|
||||||
|
)
|
||||||
|
|
||||||
def set_file_contents(self, bucket, local_file_path, s3_file_path):
|
def set_file_contents(self, bucket, local_file_path, s3_file_path):
|
||||||
key = Key(bucket)
|
key = Key(bucket)
|
||||||
key.key = s3_file_path
|
key.key = s3_file_path
|
||||||
file_size = os.stat(local_file_path).st_size
|
file_size = os.stat(local_file_path).st_size
|
||||||
logging.info("Uploading file `%s` to `%s`. Size is %s", local_file_path, s3_file_path, file_size)
|
logging.info(
|
||||||
|
"Uploading file `%s` to `%s`. Size is %s",
|
||||||
|
local_file_path,
|
||||||
|
s3_file_path,
|
||||||
|
file_size,
|
||||||
|
)
|
||||||
|
|
||||||
def call_back(x, y):
|
def call_back(x, y):
|
||||||
print("Uploaded {}/{} bytes".format(x, y))
|
print("Uploaded {}/{} bytes".format(x, y))
|
||||||
|
|
||||||
key.set_contents_from_filename(local_file_path, cb=call_back)
|
key.set_contents_from_filename(local_file_path, cb=call_back)
|
||||||
|
|
||||||
def upload_data_for_static_files_disk(self, bucket_name, directory_path, s3_path):
|
def upload_data_for_static_files_disk(self, bucket_name, directory_path, s3_path):
|
||||||
@ -79,7 +88,9 @@ class S3API(object):
|
|||||||
self.set_file_contents(bucket, local_file_path, s3_file_path)
|
self.set_file_contents(bucket, local_file_path, s3_file_path)
|
||||||
|
|
||||||
logging.info("Uploading finished")
|
logging.info("Uploading finished")
|
||||||
return "https://{bucket}.{mds_url}/{path}".format(bucket=bucket_name, mds_url=self.mds_url, path=s3_path)
|
return "https://{bucket}.{mds_url}/{path}".format(
|
||||||
|
bucket=bucket_name, mds_url=self.mds_url, path=s3_path
|
||||||
|
)
|
||||||
|
|
||||||
def list_bucket_keys(self, bucket_name):
|
def list_bucket_keys(self, bucket_name):
|
||||||
bucket = self.connection.get_bucket(bucket_name)
|
bucket = self.connection.get_bucket(bucket_name)
|
||||||
@ -91,100 +102,121 @@ class S3API(object):
|
|||||||
bucket.get_all_keys()
|
bucket.get_all_keys()
|
||||||
for obj in bucket.get_all_keys():
|
for obj in bucket.get_all_keys():
|
||||||
if obj.key.startswith(folder_path):
|
if obj.key.startswith(folder_path):
|
||||||
print('Removing ' + obj.key)
|
print("Removing " + obj.key)
|
||||||
obj.delete()
|
obj.delete()
|
||||||
|
|
||||||
|
|
||||||
def make_tar_file_for_table(clickhouse_data_path, db_name, table_name,
|
def make_tar_file_for_table(clickhouse_data_path, db_name, table_name, tmp_prefix):
|
||||||
tmp_prefix):
|
|
||||||
|
|
||||||
relative_data_path = os.path.join('data', db_name, table_name)
|
relative_data_path = os.path.join("data", db_name, table_name)
|
||||||
relative_meta_path = os.path.join('metadata', db_name, table_name + '.sql')
|
relative_meta_path = os.path.join("metadata", db_name, table_name + ".sql")
|
||||||
path_to_data = os.path.join(clickhouse_data_path, relative_data_path)
|
path_to_data = os.path.join(clickhouse_data_path, relative_data_path)
|
||||||
path_to_metadata = os.path.join(clickhouse_data_path, relative_meta_path)
|
path_to_metadata = os.path.join(clickhouse_data_path, relative_meta_path)
|
||||||
temporary_file_name = tmp_prefix + '/{tname}.tar'.format(tname=table_name)
|
temporary_file_name = tmp_prefix + "/{tname}.tar".format(tname=table_name)
|
||||||
with tarfile.open(temporary_file_name, "w") as bundle:
|
with tarfile.open(temporary_file_name, "w") as bundle:
|
||||||
bundle.add(path_to_data, arcname=relative_data_path)
|
bundle.add(path_to_data, arcname=relative_data_path)
|
||||||
bundle.add(path_to_metadata, arcname=relative_meta_path)
|
bundle.add(path_to_metadata, arcname=relative_meta_path)
|
||||||
return temporary_file_name
|
return temporary_file_name
|
||||||
|
|
||||||
|
|
||||||
USAGE_EXAMPLES = '''
|
USAGE_EXAMPLES = """
|
||||||
examples:
|
examples:
|
||||||
\t./s3uploader --dataset-name some_ds --access-key-id XXX --secret-access-key YYY --clickhouse-data-path /opt/clickhouse/ --table-name default.some_tbl --bucket-name some-bucket
|
\t./s3uploader --dataset-name some_ds --access-key-id XXX --secret-access-key YYY --clickhouse-data-path /opt/clickhouse/ --table-name default.some_tbl --bucket-name some-bucket
|
||||||
\t./s3uploader --dataset-name some_ds --access-key-id XXX --secret-access-key YYY --file-path some_ds.tsv.xz --bucket-name some-bucket --s3-path /path/to/
|
\t./s3uploader --dataset-name some_ds --access-key-id XXX --secret-access-key YYY --file-path some_ds.tsv.xz --bucket-name some-bucket --s3-path /path/to/
|
||||||
'''
|
"""
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
logging.basicConfig(level=logging.INFO, format='%(asctime)s %(message)s')
|
logging.basicConfig(level=logging.INFO, format="%(asctime)s %(message)s")
|
||||||
|
|
||||||
parser = argparse.ArgumentParser(
|
parser = argparse.ArgumentParser(
|
||||||
description="Simple tool for uploading datasets to clickhouse S3",
|
description="Simple tool for uploading datasets to clickhouse S3",
|
||||||
usage='%(prog)s [options] {}'.format(USAGE_EXAMPLES))
|
usage="%(prog)s [options] {}".format(USAGE_EXAMPLES),
|
||||||
parser.add_argument('--s3-api-url', default='s3.amazonaws.com')
|
)
|
||||||
parser.add_argument('--s3-common-url', default='s3.amazonaws.com')
|
parser.add_argument("--s3-api-url", default="s3.amazonaws.com")
|
||||||
parser.add_argument('--bucket-name', default='clickhouse-datasets')
|
parser.add_argument("--s3-common-url", default="s3.amazonaws.com")
|
||||||
parser.add_argument('--dataset-name', required=True,
|
parser.add_argument("--bucket-name", default="clickhouse-datasets")
|
||||||
help='Name of dataset, will be used in uploaded path')
|
parser.add_argument(
|
||||||
parser.add_argument('--access-key-id', required=True)
|
"--dataset-name",
|
||||||
parser.add_argument('--secret-access-key', required=True)
|
required=True,
|
||||||
parser.add_argument('--clickhouse-data-path',
|
help="Name of dataset, will be used in uploaded path",
|
||||||
default='/var/lib/clickhouse/',
|
)
|
||||||
help='Path to clickhouse database on filesystem')
|
parser.add_argument("--access-key-id", required=True)
|
||||||
parser.add_argument('--s3-path', help='Path in s3, where to upload file')
|
parser.add_argument("--secret-access-key", required=True)
|
||||||
parser.add_argument('--tmp-prefix', default='/tmp',
|
parser.add_argument(
|
||||||
help='Prefix to store temporary downloaded file')
|
"--clickhouse-data-path",
|
||||||
|
default="/var/lib/clickhouse/",
|
||||||
|
help="Path to clickhouse database on filesystem",
|
||||||
|
)
|
||||||
|
parser.add_argument("--s3-path", help="Path in s3, where to upload file")
|
||||||
|
parser.add_argument(
|
||||||
|
"--tmp-prefix", default="/tmp", help="Prefix to store temporary downloaded file"
|
||||||
|
)
|
||||||
data_group = parser.add_mutually_exclusive_group(required=True)
|
data_group = parser.add_mutually_exclusive_group(required=True)
|
||||||
table_name_argument = data_group.add_argument('--table-name',
|
table_name_argument = data_group.add_argument(
|
||||||
help='Name of table with database, if you are uploading partitions')
|
"--table-name",
|
||||||
data_group.add_argument('--file-path',
|
help="Name of table with database, if you are uploading partitions",
|
||||||
help='Name of file, if you are uploading')
|
)
|
||||||
data_group.add_argument('--directory-path', help='Path to directory with files to upload')
|
data_group.add_argument("--file-path", help="Name of file, if you are uploading")
|
||||||
data_group.add_argument('--list-directory', help='List s3 directory by --directory-path')
|
data_group.add_argument(
|
||||||
data_group.add_argument('--remove-directory', help='Remove s3 directory by --directory-path')
|
"--directory-path", help="Path to directory with files to upload"
|
||||||
|
)
|
||||||
|
data_group.add_argument(
|
||||||
|
"--list-directory", help="List s3 directory by --directory-path"
|
||||||
|
)
|
||||||
|
data_group.add_argument(
|
||||||
|
"--remove-directory", help="Remove s3 directory by --directory-path"
|
||||||
|
)
|
||||||
args = parser.parse_args()
|
args = parser.parse_args()
|
||||||
|
|
||||||
if args.table_name is not None and args.clickhouse_data_path is None:
|
if args.table_name is not None and args.clickhouse_data_path is None:
|
||||||
raise argparse.ArgumentError(table_name_argument,
|
raise argparse.ArgumentError(
|
||||||
"You should specify --clickhouse-data-path to upload --table")
|
table_name_argument,
|
||||||
|
"You should specify --clickhouse-data-path to upload --table",
|
||||||
|
)
|
||||||
|
|
||||||
s3_conn = S3API(
|
s3_conn = S3API(
|
||||||
args.access_key_id, args.secret_access_key,
|
args.access_key_id, args.secret_access_key, args.s3_api_url, args.s3_common_url
|
||||||
args.s3_api_url, args.s3_common_url)
|
)
|
||||||
|
|
||||||
file_path = ''
|
file_path = ""
|
||||||
directory_path = args.directory_path
|
directory_path = args.directory_path
|
||||||
s3_path = args.s3_path
|
s3_path = args.s3_path
|
||||||
|
|
||||||
if args.list_directory:
|
if args.list_directory:
|
||||||
s3_conn.list_bucket_keys(args.bucket_name)
|
s3_conn.list_bucket_keys(args.bucket_name)
|
||||||
elif args.remove_directory:
|
elif args.remove_directory:
|
||||||
print('Removing s3 path: ' + args.remove_directory)
|
print("Removing s3 path: " + args.remove_directory)
|
||||||
s3_conn.remove_folder_from_bucket(args.bucket_name, args.remove_directory)
|
s3_conn.remove_folder_from_bucket(args.bucket_name, args.remove_directory)
|
||||||
elif args.directory_path is not None:
|
elif args.directory_path is not None:
|
||||||
url = s3_conn.upload_data_for_static_files_disk(args.bucket_name, directory_path, s3_path)
|
url = s3_conn.upload_data_for_static_files_disk(
|
||||||
|
args.bucket_name, directory_path, s3_path
|
||||||
|
)
|
||||||
logging.info("Data uploaded: %s", url)
|
logging.info("Data uploaded: %s", url)
|
||||||
else:
|
else:
|
||||||
|
|
||||||
if args.table_name is not None:
|
if args.table_name is not None:
|
||||||
if '.' not in args.table_name:
|
if "." not in args.table_name:
|
||||||
db_name = 'default'
|
db_name = "default"
|
||||||
else:
|
else:
|
||||||
db_name, table_name = args.table_name.split('.')
|
db_name, table_name = args.table_name.split(".")
|
||||||
file_path = make_tar_file_for_table(
|
file_path = make_tar_file_for_table(
|
||||||
args.clickhouse_data_path, db_name, table_name, args.tmp_prefix)
|
args.clickhouse_data_path, db_name, table_name, args.tmp_prefix
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
file_path = args.file_path
|
file_path = args.file_path
|
||||||
|
|
||||||
if 'tsv' in file_path:
|
if "tsv" in file_path:
|
||||||
s3_path = os.path.join(
|
s3_path = os.path.join(
|
||||||
args.dataset_name, 'tsv', os.path.basename(file_path))
|
args.dataset_name, "tsv", os.path.basename(file_path)
|
||||||
|
)
|
||||||
if args.table_name is not None:
|
if args.table_name is not None:
|
||||||
s3_path = os.path.join(
|
s3_path = os.path.join(
|
||||||
args.dataset_name, 'partitions', os.path.basename(file_path))
|
args.dataset_name, "partitions", os.path.basename(file_path)
|
||||||
|
)
|
||||||
elif args.s3_path is not None:
|
elif args.s3_path is not None:
|
||||||
s3_path = os.path.join(
|
s3_path = os.path.join(
|
||||||
args.dataset_name, args.s3_path, os.path.basename(file_path))
|
args.dataset_name, args.s3_path, os.path.basename(file_path)
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
raise Exception("Don't know s3-path to upload")
|
raise Exception("Don't know s3-path to upload")
|
||||||
|
|
||||||
|
@ -11,13 +11,14 @@ from termcolor import colored
|
|||||||
import sys
|
import sys
|
||||||
|
|
||||||
COLORMAP = {
|
COLORMAP = {
|
||||||
"success": colored("success", 'green'),
|
"success": colored("success", "green"),
|
||||||
"failure": colored("failure", 'red'),
|
"failure": colored("failure", "red"),
|
||||||
"error": colored("error", 'red'),
|
"error": colored("error", "red"),
|
||||||
"pending": colored("pending", 'yellow'),
|
"pending": colored("pending", "yellow"),
|
||||||
"not run": colored("not run", 'white'),
|
"not run": colored("not run", "white"),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
def _filter_statuses(statuses):
|
def _filter_statuses(statuses):
|
||||||
"""
|
"""
|
||||||
Squash statuses to latest state
|
Squash statuses to latest state
|
||||||
@ -69,7 +70,7 @@ if __name__ == "__main__":
|
|||||||
date_since = datetime.datetime.strptime(args.since, "%Y-%m-%d %H:%M:%S")
|
date_since = datetime.datetime.strptime(args.since, "%Y-%m-%d %H:%M:%S")
|
||||||
|
|
||||||
gh = Github(args.token)
|
gh = Github(args.token)
|
||||||
repo = gh.get_repo('ClickHouse/ClickHouse')
|
repo = gh.get_repo("ClickHouse/ClickHouse")
|
||||||
commits = get_commits(repo, date_since)
|
commits = get_commits(repo, date_since)
|
||||||
|
|
||||||
longest_header = []
|
longest_header = []
|
||||||
@ -101,6 +102,6 @@ if __name__ == "__main__":
|
|||||||
result_data.append(current_result)
|
result_data.append(current_result)
|
||||||
|
|
||||||
if sys.stdout.isatty():
|
if sys.stdout.isatty():
|
||||||
longest_header = [colored(h, 'white', attrs=['bold']) for h in longest_header]
|
longest_header = [colored(h, "white", attrs=["bold"]) for h in longest_header]
|
||||||
|
|
||||||
print(tabulate.tabulate(result_data, headers=longest_header, tablefmt="grid"))
|
print(tabulate.tabulate(result_data, headers=longest_header, tablefmt="grid"))
|
||||||
|
Loading…
Reference in New Issue
Block a user