mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-24 16:42:05 +00:00
353 lines
12 KiB
Python
353 lines
12 KiB
Python
#!/usr/bin/env python3
|
|
import fileinput
|
|
import json
|
|
import logging
|
|
import os
|
|
import time
|
|
from pathlib import Path
|
|
from typing import Any, Dict, List, Optional
|
|
|
|
import requests
|
|
|
|
from env_helper import GITHUB_REPOSITORY
|
|
from get_robot_token import get_parameter_from_ssm
|
|
from pr_info import PRInfo
|
|
from report import TestResults
|
|
|
|
|
|
class CHException(Exception):
|
|
pass
|
|
|
|
|
|
class InsertException(Exception):
|
|
pass
|
|
|
|
|
|
class ClickHouseHelper:
|
|
def __init__(
|
|
self, url: Optional[str] = None, auth: Optional[Dict[str, str]] = None
|
|
):
|
|
if url is None:
|
|
url = get_parameter_from_ssm("clickhouse-test-stat-url")
|
|
|
|
self.url = url
|
|
self.auth = auth or {
|
|
"X-ClickHouse-User": get_parameter_from_ssm("clickhouse-test-stat-login"),
|
|
"X-ClickHouse-Key": get_parameter_from_ssm("clickhouse-test-stat-password"),
|
|
}
|
|
|
|
@staticmethod
|
|
def insert_file(
|
|
url: str,
|
|
auth: Optional[Dict[str, str]],
|
|
query: str,
|
|
file: Path,
|
|
additional_options: Optional[Dict[str, str]] = None,
|
|
**kwargs: Any,
|
|
) -> None:
|
|
params = {
|
|
"query": query,
|
|
"date_time_input_format": "best_effort",
|
|
"send_logs_level": "warning",
|
|
}
|
|
if additional_options:
|
|
for k, v in additional_options.items():
|
|
params[k] = v
|
|
|
|
with open(file, "rb") as data_fd:
|
|
ClickHouseHelper._insert_post(
|
|
url, params=params, data=data_fd, headers=auth, **kwargs
|
|
)
|
|
|
|
@staticmethod
|
|
def insert_json_str(url, auth, db, table, json_str):
|
|
params = {
|
|
"database": db,
|
|
"query": f"INSERT INTO {table} FORMAT JSONEachRow",
|
|
"date_time_input_format": "best_effort",
|
|
"send_logs_level": "warning",
|
|
}
|
|
ClickHouseHelper._insert_post(url, params=params, data=json_str, headers=auth)
|
|
|
|
@staticmethod
|
|
def _insert_post(*args, **kwargs):
|
|
url = ""
|
|
if args:
|
|
url = args[0]
|
|
url = kwargs.get("url", url)
|
|
timeout = kwargs.pop("timeout", 100)
|
|
|
|
for i in range(5):
|
|
try:
|
|
response = requests.post(*args, timeout=timeout, **kwargs)
|
|
except Exception as e:
|
|
error = f"Received exception while sending data to {url} on {i} attempt: {e}"
|
|
logging.warning(error)
|
|
continue
|
|
|
|
logging.info("Response content '%s'", response.content)
|
|
|
|
if response.ok:
|
|
break
|
|
|
|
error = (
|
|
f"Cannot insert data into clickhouse at try {i}: HTTP code "
|
|
f"{response.status_code}: '{response.text}'"
|
|
)
|
|
|
|
if response.status_code >= 500:
|
|
# A retryable error
|
|
time.sleep(1)
|
|
continue
|
|
|
|
logging.info(
|
|
"Request headers '%s', body '%s'",
|
|
response.request.headers,
|
|
response.request.body,
|
|
)
|
|
|
|
raise InsertException(error)
|
|
else:
|
|
raise InsertException(error)
|
|
|
|
def _insert_json_str_info(self, db, table, json_str):
|
|
self.insert_json_str(self.url, self.auth, db, table, json_str)
|
|
|
|
def insert_event_into(self, db, table, event, safe=True):
|
|
event_str = json.dumps(event)
|
|
try:
|
|
self._insert_json_str_info(db, table, event_str)
|
|
except InsertException as e:
|
|
logging.error(
|
|
"Exception happened during inserting data into clickhouse: %s", e
|
|
)
|
|
if not safe:
|
|
raise
|
|
|
|
def insert_events_into(self, db, table, events, safe=True):
|
|
jsons = []
|
|
for event in events:
|
|
jsons.append(json.dumps(event))
|
|
|
|
try:
|
|
self._insert_json_str_info(db, table, ",".join(jsons))
|
|
except InsertException as e:
|
|
logging.error(
|
|
"Exception happened during inserting data into clickhouse: %s", e
|
|
)
|
|
if not safe:
|
|
raise
|
|
|
|
def _select_and_get_json_each_row(self, db, query, query_params):
|
|
params = {
|
|
"database": db,
|
|
"query": query,
|
|
"default_format": "JSONEachRow",
|
|
}
|
|
if query_params is not None:
|
|
for name, value in query_params.items():
|
|
params[f"param_{name}"] = str(value)
|
|
|
|
for i in range(5):
|
|
response = None
|
|
try:
|
|
response = requests.get(
|
|
self.url, params=params, headers=self.auth, timeout=100
|
|
)
|
|
response.raise_for_status()
|
|
return response.text
|
|
except Exception as ex:
|
|
logging.warning("Select query failed with exception %s", str(ex))
|
|
if response:
|
|
logging.warning("Response text %s", response.text)
|
|
time.sleep(0.1 * i)
|
|
|
|
raise CHException("Cannot fetch data from clickhouse")
|
|
|
|
def select_json_each_row(self, db, query, query_params=None):
|
|
text = self._select_and_get_json_each_row(db, query, query_params)
|
|
result = []
|
|
for line in text.split("\n"):
|
|
if line:
|
|
result.append(json.loads(line))
|
|
return result
|
|
|
|
|
|
def _query_imds(path):
|
|
url = f"http://169.254.169.254/{path}"
|
|
for i in range(5):
|
|
try:
|
|
response = requests.get(url, timeout=1)
|
|
if response.status_code == 200:
|
|
return response.text
|
|
except Exception as e:
|
|
error = (
|
|
f"Received exception while sending data to {url} on {i} attempt: {e}"
|
|
)
|
|
logging.warning(error)
|
|
continue
|
|
return ""
|
|
|
|
|
|
# Obtain the machine type from IMDS:
|
|
def get_instance_type():
|
|
return _query_imds("latest/meta-data/instance-type")
|
|
|
|
|
|
# Obtain the instance id from IMDS:
|
|
def get_instance_id():
|
|
return _query_imds("latest/meta-data/instance-id")
|
|
|
|
|
|
def get_instance_lifecycle():
|
|
return _query_imds("latest/meta-data/instance-life-cycle")
|
|
|
|
|
|
def prepare_tests_results_for_clickhouse(
|
|
pr_info: PRInfo,
|
|
test_results: TestResults,
|
|
check_status: str,
|
|
check_duration: float,
|
|
check_start_time: str,
|
|
report_url: str,
|
|
check_name: str,
|
|
) -> List[dict]:
|
|
base_ref = pr_info.base_ref
|
|
base_repo = pr_info.base_name
|
|
head_ref = pr_info.head_ref
|
|
head_repo = pr_info.head_name
|
|
pull_request_url = f"https://github.com/{GITHUB_REPOSITORY}/commits/{head_ref}"
|
|
if pr_info.number != 0:
|
|
pull_request_url = pr_info.pr_html_url
|
|
|
|
common_properties = {
|
|
"pull_request_number": pr_info.number,
|
|
"commit_sha": pr_info.sha,
|
|
"commit_url": pr_info.commit_html_url,
|
|
"check_name": check_name,
|
|
"check_status": check_status,
|
|
"check_duration_ms": int(float(check_duration) * 1000),
|
|
"check_start_time": check_start_time,
|
|
"report_url": report_url,
|
|
"pull_request_url": pull_request_url,
|
|
"base_ref": base_ref,
|
|
"base_repo": base_repo,
|
|
"head_ref": head_ref,
|
|
"head_repo": head_repo,
|
|
"task_url": pr_info.task_url,
|
|
"instance_type": ",".join([get_instance_type(), get_instance_lifecycle()]),
|
|
"instance_id": get_instance_id(),
|
|
}
|
|
|
|
# Always publish a total record for all checks. For checks with individual
|
|
# tests, also publish a record per test.
|
|
result = [common_properties]
|
|
for test_result in test_results:
|
|
current_row = common_properties.copy()
|
|
test_name = test_result.name
|
|
test_status = test_result.status
|
|
|
|
test_time = test_result.time or 0
|
|
current_row["test_duration_ms"] = int(test_time * 1000)
|
|
current_row["test_name"] = test_name
|
|
current_row["test_status"] = test_status
|
|
if test_result.raw_logs:
|
|
# Protect from too big blobs that contain garbage
|
|
current_row["test_context_raw"] = test_result.raw_logs[: 32 * 1024]
|
|
else:
|
|
current_row["test_context_raw"] = ""
|
|
result.append(current_row)
|
|
|
|
return result
|
|
|
|
|
|
class CiLogsCredentials:
|
|
def __init__(self, config_path: Path):
|
|
self.config_path = config_path
|
|
try:
|
|
self._host = get_parameter_from_ssm("clickhouse_ci_logs_host") # type: str
|
|
self._password = get_parameter_from_ssm(
|
|
"clickhouse_ci_logs_password"
|
|
) # type: str
|
|
except:
|
|
logging.warning(
|
|
"Unable to retreive host and/or password from smm, all other "
|
|
"methods will noop"
|
|
)
|
|
self._host = ""
|
|
self._password = ""
|
|
|
|
def create_ci_logs_credentials(self) -> None:
|
|
if not (self.host and self.password):
|
|
logging.info(
|
|
"Hostname or password for CI logs instance are unknown, "
|
|
"skipping creating of credentials file, removing existing"
|
|
)
|
|
self.config_path.unlink(missing_ok=True)
|
|
return
|
|
self.config_path.parent.mkdir(parents=True, exist_ok=True)
|
|
self.config_path.write_text(
|
|
f"CLICKHOUSE_CI_LOGS_HOST={self.host}\n"
|
|
"CLICKHOUSE_CI_LOGS_USER=ci\n"
|
|
f"CLICKHOUSE_CI_LOGS_PASSWORD={self.password}\n",
|
|
encoding="utf-8",
|
|
)
|
|
|
|
def get_docker_arguments(
|
|
self, pr_info: PRInfo, check_start_time: str, check_name: str
|
|
) -> str:
|
|
run_by_hash_total = int(os.getenv("RUN_BY_HASH_TOTAL", "0"))
|
|
if run_by_hash_total > 1:
|
|
run_by_hash_num = int(os.getenv("RUN_BY_HASH_NUM", "0"))
|
|
check_name = f"{check_name} [{run_by_hash_num + 1}/{run_by_hash_total}]"
|
|
|
|
self.create_ci_logs_credentials()
|
|
if not self.config_path.exists():
|
|
logging.info("Do not use external logs pushing")
|
|
return ""
|
|
extra_columns = (
|
|
f"CAST({pr_info.number} AS UInt32) AS pull_request_number, '{pr_info.sha}' AS commit_sha, "
|
|
f"toDateTime('{check_start_time}', 'UTC') AS check_start_time, toLowCardinality('{check_name}') AS check_name, "
|
|
f"toLowCardinality('{get_instance_type()}') AS instance_type, '{get_instance_id()}' AS instance_id"
|
|
)
|
|
return (
|
|
f'-e EXTRA_COLUMNS_EXPRESSION="{extra_columns}" '
|
|
f"-e CLICKHOUSE_CI_LOGS_CREDENTIALS=/tmp/export-logs-config.sh "
|
|
f"--volume={self.config_path.absolute()}:/tmp/export-logs-config.sh:ro "
|
|
)
|
|
|
|
def clean_ci_logs_from_credentials(self, log_path: Path) -> None:
|
|
if not (self.host or self.password):
|
|
logging.info(
|
|
"Hostname and password for CI logs instance are unknown, "
|
|
"skipping cleaning %s",
|
|
log_path,
|
|
)
|
|
return
|
|
|
|
def process_line(line: str) -> str:
|
|
if self.host and self.password:
|
|
return line.replace(self.host, "CLICKHOUSE_CI_LOGS_HOST").replace(
|
|
self.password, "CLICKHOUSE_CI_LOGS_PASSWORD"
|
|
)
|
|
if self.host:
|
|
return line.replace(self.host, "CLICKHOUSE_CI_LOGS_HOST")
|
|
# the remaining is self.password
|
|
return line.replace(self.password, "CLICKHOUSE_CI_LOGS_PASSWORD")
|
|
|
|
# errors="surrogateescape" require python 3.10.
|
|
# With ubuntu 22.04 we are safe
|
|
with fileinput.input(
|
|
log_path, inplace=True, errors="surrogateescape"
|
|
) as log_fd:
|
|
for line in log_fd:
|
|
print(process_line(line), end="")
|
|
|
|
@property
|
|
def host(self) -> str:
|
|
return self._host
|
|
|
|
@property
|
|
def password(self) -> str:
|
|
return self._password
|