ClickHouse/tests/ci/report.py

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

873 lines
27 KiB
Python
Raw Normal View History

2021-09-10 11:52:21 +00:00
# -*- coding: utf-8 -*-
2024-02-02 14:33:08 +00:00
import csv
import datetime
import json
import logging
import os
from ast import literal_eval
from dataclasses import asdict, dataclass
2024-02-02 14:33:08 +00:00
from html import escape
from pathlib import Path
from typing import (
Dict,
Final,
Iterable,
List,
Literal,
Optional,
Sequence,
Tuple,
Union,
)
from build_download_helper import get_gh_api
2024-02-02 14:33:08 +00:00
from ci_config import CI_CONFIG, BuildConfig
from env_helper import REPORT_PATH, TEMP_PATH
2024-02-04 19:12:37 +00:00
from ci_utils import normalize_string
logger = logging.getLogger(__name__)
2023-09-01 17:15:11 +00:00
ERROR: Final = "error"
FAILURE: Final = "failure"
PENDING: Final = "pending"
SUCCESS: Final = "success"
OK: Final = "OK"
FAIL: Final = "FAIL"
StatusType = Literal["error", "failure", "pending", "success"]
STATUSES = [ERROR, FAILURE, PENDING, SUCCESS] # type: List[StatusType]
2023-09-01 17:15:11 +00:00
# The order of statuses from the worst to the best
def _state_rank(status: str) -> int:
"return the index of status or index of SUCCESS in case of wrong status"
try:
return STATUSES.index(status) # type: ignore
except ValueError:
return 3
2023-09-01 17:15:11 +00:00
def get_worst_status(statuses: Iterable[str]) -> StatusType:
worst_status = SUCCESS # type: StatusType
2023-09-01 17:15:11 +00:00
for status in statuses:
ind = _state_rank(status)
if ind < _state_rank(worst_status):
worst_status = STATUSES[ind]
2023-09-01 17:15:11 +00:00
if worst_status == ERROR:
break
return worst_status
2023-01-06 03:19:10 +00:00
### BEST FRONTEND PRACTICES BELOW
2021-09-10 11:52:21 +00:00
2023-09-01 12:05:18 +00:00
HEAD_HTML_TEMPLATE = """
2021-09-10 11:52:21 +00:00
<!DOCTYPE html>
<html>
2023-09-01 12:05:18 +00:00
<head>
2021-09-10 11:52:21 +00:00
<style>
2023-01-07 02:40:18 +00:00
:root {{
--color: white;
--background: hsl(190deg, 90%, 5%) linear-gradient(180deg, hsl(190deg, 90%, 10%), hsl(190deg, 90%, 0%));
--td-background: hsl(190deg, 90%, 15%);
--th-background: hsl(180deg, 90%, 15%);
--link-color: #FF5;
2023-01-07 02:40:18 +00:00
--link-hover-color: #F40;
--menu-background: hsl(190deg, 90%, 20%);
--menu-hover-background: hsl(190deg, 100%, 50%);
--menu-hover-color: black;
--text-gradient: linear-gradient(90deg, #8F8, #F88);
--shadow-intensity: 1;
--tr-hover-filter: brightness(120%);
--table-border-color: black;
2023-01-07 22:23:20 +00:00
}}
2023-01-07 02:40:18 +00:00
[data-theme="light"] {{
--color: black;
--background: hsl(190deg, 90%, 90%) linear-gradient(180deg, #EEE, #DEE);
--td-background: white;
--th-background: #EEE;
--link-color: #08F;
--link-hover-color: #F40;
--menu-background: white;
--menu-hover-background: white;
--menu-hover-color: #F40;
--text-gradient: linear-gradient(90deg, black, black);
--shadow-intensity: 0.1;
--tr-hover-filter: brightness(95%);
--table-border-color: #DDD;
}}
2023-01-06 03:19:10 +00:00
.gradient {{
2023-01-07 02:40:18 +00:00
background-image: var(--text-gradient);
2023-01-06 03:19:10 +00:00
background-size: 100%;
background-repeat: repeat;
background-clip: text;
-webkit-text-fill-color: transparent;
-moz-text-fill-color: transparent;
-webkit-background-clip: text;
-moz-background-clip: text;
}}
2023-01-07 02:40:18 +00:00
html {{ min-height: 100%; font-family: "DejaVu Sans", "Noto Sans", Arial, sans-serif; background: var(--background); color: var(--color); }}
2021-09-10 11:52:21 +00:00
h1 {{ margin-left: 10px; }}
2023-01-07 02:40:18 +00:00
th, td {{ padding: 5px 10px 5px 10px; text-align: left; vertical-align: top; line-height: 1.5; border: 1px solid var(--table-border-color); }}
td {{ background: var(--td-background); }}
th {{ background: var(--th-background); }}
a {{ color: var(--link-color); text-decoration: none; }}
a:hover, a:active {{ color: var(--link-hover-color); text-decoration: none; }}
table {{ box-shadow: 0 8px 25px -5px rgba(0, 0, 0, var(--shadow-intensity)); border-collapse: collapse; border-spacing: 0; }}
p.links a {{ padding: 5px; margin: 3px; background: var(--menu-background); line-height: 2.5; white-space: nowrap; box-shadow: 0 8px 25px -5px rgba(0, 0, 0, var(--shadow-intensity)); }}
p.links a:hover {{ background: var(--menu-hover-background); color: var(--menu-hover-color); }}
2021-09-10 11:52:21 +00:00
th {{ cursor: pointer; }}
2023-01-07 02:40:18 +00:00
tr:hover {{ filter: var(--tr-hover-filter); }}
.expandable {{ cursor: pointer; }}
.expandable-content {{ display: none; }}
2023-01-07 02:40:18 +00:00
#fish {{ display: none; float: right; position: relative; top: -20em; right: 2vw; margin-bottom: -20em; width: 30vw; filter: brightness(7%); z-index: -1; }}
.themes {{
float: right;
font-size: 20pt;
margin-bottom: 1rem;
}}
#toggle-dark, #toggle-light {{
padding-right: 0.5rem;
user-select: none;
cursor: pointer;
}}
#toggle-dark:hover, #toggle-light:hover {{
display: inline-block;
transform: translate(1px, 1px);
filter: brightness(125%);
}}
2021-09-10 11:52:21 +00:00
</style>
<title>{title}</title>
</head>
<body>
<div class="main">
2023-01-07 02:40:18 +00:00
<span class="nowrap themes"><span id="toggle-dark">🌚</span><span id="toggle-light">🌞</span></span>
2023-01-06 03:19:10 +00:00
<h1><span class="gradient">{header}</span></h1>
2023-09-01 12:05:18 +00:00
"""
FOOTER_HTML_TEMPLATE = """<img id="fish" src="https://presentations.clickhouse.com/images/fish.png" />
2021-09-10 11:52:21 +00:00
<script type="text/javascript">
/// Straight from https://stackoverflow.com/questions/14267781/sorting-html-table-with-javascript
2021-10-22 12:23:41 +00:00
const getCellValue = (tr, idx) => {{
var classes = tr.classList;
var elem = tr;
2023-09-12 11:54:35 +00:00
if (classes.contains("expandable-content") || classes.contains("expandable-content.open"))
2021-10-22 12:23:41 +00:00
elem = tr.previousElementSibling;
return elem.children[idx].innerText || elem.children[idx].textContent;
}}
2021-09-10 11:52:21 +00:00
const comparer = (idx, asc) => (a, b) => ((v1, v2) =>
v1 !== '' && v2 !== '' && !isNaN(v1) && !isNaN(v2) ? v1 - v2 : v1.toString().localeCompare(v2)
)(getCellValue(asc ? a : b, idx), getCellValue(asc ? b : a, idx));
document.querySelectorAll('th').forEach(th => th.addEventListener('click', (() => {{
const table = th.closest('table');
Array.from(table.querySelectorAll('tr:nth-child(n+2)'))
.sort(comparer(Array.from(th.parentNode.children).indexOf(th), this.asc = !this.asc))
.forEach(tr => table.appendChild(tr) );
}})));
2021-10-22 12:23:41 +00:00
Array.from(document.getElementsByClassName("expandable")).forEach(tr => tr.addEventListener('click', function() {{
2021-10-22 12:23:41 +00:00
var content = this.nextElementSibling;
content.classList.toggle("expandable-content");
2021-10-22 12:23:41 +00:00
}}));
2023-01-06 03:19:10 +00:00
2023-01-07 02:40:18 +00:00
let theme = 'dark';
function setTheme(new_theme) {{
theme = new_theme;
document.documentElement.setAttribute('data-theme', theme);
window.localStorage.setItem('theme', theme);
drawFish();
}}
function drawFish() {{
document.getElementById('fish').style.display = (document.body.clientHeight > 3000 && theme == 'dark') ? 'block' : 'none';
2023-01-06 03:19:10 +00:00
}}
2023-01-07 02:40:18 +00:00
document.getElementById('toggle-light').addEventListener('click', e => setTheme('light'));
document.getElementById('toggle-dark').addEventListener('click', e => setTheme('dark'));
let new_theme = window.localStorage.getItem('theme');
if (new_theme && new_theme != theme) {{
setTheme(new_theme);
}}
drawFish();
2021-09-10 11:52:21 +00:00
</script>
2023-01-06 03:19:10 +00:00
</body>
2021-09-10 11:52:21 +00:00
</html>
"""
2023-09-01 12:05:18 +00:00
HTML_BASE_TEST_TEMPLATE = (
f"{HEAD_HTML_TEMPLATE}"
"""<p class="links">
<a href="{raw_log_url}">{raw_log_name}</a>
<a href="{commit_url}">Commit</a>
{additional_urls}
<a href="{task_url}">Task (github actions)</a>
<a href="{job_url}">Job (github actions)</a>
</p>
{test_part}
"""
f"{FOOTER_HTML_TEMPLATE}"
)
2021-09-10 11:52:21 +00:00
HTML_TEST_PART = """
<table>
<tr>
{headers}
</tr>
{rows}
</table>
"""
BASE_HEADERS = ["Test name", "Test status"]
JOB_REPORT_FILE = Path(TEMP_PATH) / "job_report.json"
2021-09-10 11:52:21 +00:00
@dataclass
class TestResult:
name: str
status: str
# the following fields are optional
time: Optional[float] = None
log_files: Optional[Union[Sequence[str], Sequence[Path]]] = None
raw_logs: Optional[str] = None
# the field for uploaded logs URLs
log_urls: Optional[Sequence[str]] = None
def set_raw_logs(self, raw_logs: str) -> None:
self.raw_logs = raw_logs
def set_log_files(self, log_files_literal: str) -> None:
self.log_files = [] # type: Optional[List[Path]]
log_paths = literal_eval(log_files_literal)
if not isinstance(log_paths, list):
raise ValueError(
f"Malformed input: must be a list literal: {log_files_literal}"
)
for log_path in log_paths:
assert Path(log_path).exists(), log_path
self.log_files.append(log_path)
2023-09-12 18:06:00 +00:00
@staticmethod
2023-09-12 18:31:34 +00:00
def create_check_timeout_expired(timeout: float) -> "TestResult":
2023-09-12 18:06:00 +00:00
return TestResult("Check timeout expired", "FAIL", timeout)
TestResults = List[TestResult]
@dataclass
class JobReport:
status: str
description: str
test_results: TestResults
start_time: str
duration: float
additional_files: Union[Sequence[str], Sequence[Path]]
# clcikhouse version, build job only
version: str = ""
# checkname to set in commit status, set if differs from jjob name
check_name: str = ""
# directory with artifacts to upload on s3
build_dir_for_upload: Union[Path, str] = ""
# if False no GH commit status will be created by CI
need_commit_status: bool = True
@classmethod
def exist(cls) -> bool:
return JOB_REPORT_FILE.is_file()
@classmethod
def load(cls): # type: ignore
res = {}
with open(JOB_REPORT_FILE, "r") as json_file:
res = json.load(json_file)
# Deserialize the nested lists of TestResult
test_results_data = res.get("test_results", [])
test_results = [TestResult(**result) for result in test_results_data]
del res["test_results"]
return JobReport(test_results=test_results, **res)
@classmethod
def cleanup(cls):
if JOB_REPORT_FILE.exists():
JOB_REPORT_FILE.unlink()
def dump(self):
def path_converter(obj):
if isinstance(obj, Path):
return str(obj)
raise TypeError("Type not serializable")
with open(JOB_REPORT_FILE, "w") as json_file:
json.dump(asdict(self), json_file, default=path_converter, indent=2)
def read_test_results(results_path: Path, with_raw_logs: bool = True) -> TestResults:
results = [] # type: TestResults
with open(results_path, "r", encoding="utf-8") as descriptor:
reader = csv.reader(descriptor, delimiter="\t")
for line in reader:
name = line[0]
status = line[1]
time = None
if len(line) >= 3 and line[2] and line[2] != "\\N":
# The value can be emtpy, but when it's not,
# it's the time spent on the test
try:
time = float(line[2])
except ValueError:
pass
result = TestResult(name, status, time)
if len(line) == 4 and line[3]:
# The value can be emtpy, but when it's not,
# the 4th value is a pythonic list, e.g. ['file1', 'file2']
if with_raw_logs:
# Python does not support TSV, so we unescape manually
2023-02-02 12:40:50 +00:00
result.set_raw_logs(
line[3].replace("\\t", "\t").replace("\\n", "\n")
2023-02-02 00:59:49 +00:00
)
else:
result.set_log_files(line[3])
results.append(result)
return results
2023-01-04 11:18:53 +00:00
@dataclass
class BuildResult:
build_name: str
log_url: str
build_urls: List[str]
version: str
status: str
2023-01-04 11:18:53 +00:00
elapsed_seconds: int
job_api_url: str
pr_number: int = 0
head_ref: str = "dummy_branch_name"
_job_name: Optional[str] = None
_job_html_url: Optional[str] = None
_job_html_link: Optional[str] = None
_grouped_urls: Optional[List[List[str]]] = None
@classmethod
def cleanup(cls):
if Path(REPORT_PATH).exists():
for file in Path(REPORT_PATH).iterdir():
if "build_report" in file.name and file.name.endswith(".json"):
file.unlink()
@classmethod
def load(cls, build_name: str, pr_number: int, head_ref: str): # type: ignore
"""
loads report from a report file matched with given @pr_number and/or a @head_ref
"""
report_path = Path(REPORT_PATH) / BuildResult.get_report_name(
build_name, pr_number or head_ref
)
return cls.load_from_file(report_path)
@classmethod
def load_any(cls, build_name: str, pr_number: int, head_ref: str): # type: ignore
"""
loads report from suitable report file with the following priority:
1. report from PR with the same @pr_number
2. report from branch with the same @head_ref
3. report from the master
4. any other report
"""
reports = []
for file in Path(REPORT_PATH).iterdir():
if f"{build_name}.json" in file.name:
reports.append(file)
if not reports:
return None
file_path = None
for file in reports:
if pr_number and f"_{pr_number}_" in file.name:
file_path = file
break
if f"_{head_ref}_" in file.name:
file_path = file
break
if "_master_" in file.name:
file_path = file
break
return cls.load_from_file(file_path or reports[-1])
@classmethod
def load_from_file(cls, file: Union[Path, str]): # type: ignore
if not Path(file).exists():
return None
with open(file, "r") as json_file:
res = json.load(json_file)
return BuildResult(**res)
def as_json(self) -> str:
return json.dumps(asdict(self), indent=2)
@property
def build_config(self) -> Optional[BuildConfig]:
return CI_CONFIG.build_config.get(self.build_name, None)
@property
def comment(self) -> str:
if self.build_config is None:
return self._wrong_config_message
return self.build_config.comment
@property
def compiler(self) -> str:
if self.build_config is None:
return self._wrong_config_message
return self.build_config.compiler
@property
def debug_build(self) -> bool:
if self.build_config is None:
return False
return self.build_config.debug_build
@property
def sanitizer(self) -> str:
if self.build_config is None:
return self._wrong_config_message
return self.build_config.sanitizer
2024-02-07 19:58:40 +00:00
@property
2024-02-07 23:23:04 +00:00
def coverage(self) -> str:
2024-02-07 19:58:40 +00:00
if self.build_config is None:
return self._wrong_config_message
2024-02-07 23:23:04 +00:00
return str(self.build_config.coverage)
2024-02-07 19:58:40 +00:00
@property
def grouped_urls(self) -> List[List[str]]:
"Combine and preserve build_urls by artifact types"
if self._grouped_urls is not None:
return self._grouped_urls
if not self.build_urls:
self._grouped_urls = [[]]
return self._grouped_urls
artifacts_groups = {
"apk": [],
"deb": [],
"binary": [],
"tgz": [],
"rpm": [],
"performance": [],
} # type: Dict[str, List[str]]
for url in self.build_urls:
if url.endswith("performance.tar.zst"):
artifacts_groups["performance"].append(url)
elif (
url.endswith(".deb")
or url.endswith(".buildinfo")
or url.endswith(".changes")
or url.endswith(".tar.gz")
):
artifacts_groups["deb"].append(url)
elif url.endswith(".apk"):
artifacts_groups["apk"].append(url)
elif url.endswith(".rpm"):
artifacts_groups["rpm"].append(url)
elif url.endswith(".tgz") or url.endswith(".tgz.sha512"):
artifacts_groups["tgz"].append(url)
else:
artifacts_groups["binary"].append(url)
self._grouped_urls = [urls for urls in artifacts_groups.values() if urls]
return self._grouped_urls
@property
def _wrong_config_message(self) -> str:
return "missing"
@property
def is_missing(self) -> bool:
"The report is created for missing json file"
return not (
self.log_url
or self.build_urls
or self.version != "missing"
or self.status != ERROR
)
@property
def job_link(self) -> str:
if self._job_html_link is not None:
return self._job_html_link
self._job_html_link = f'<a href="{self.job_html_url}">{self.job_name}</a>'
return self._job_html_link
@property
def job_html_url(self) -> str:
if self._job_html_url is not None:
return self._job_html_url
self._set_properties()
return self._job_html_url or ""
@property
def job_name(self) -> str:
if self._job_name is not None:
return self._job_name
self._set_properties()
return self._job_name or ""
@job_name.setter
def job_name(self, job_name: str) -> None:
self._job_name = job_name
def _set_properties(self) -> None:
if all(p is not None for p in (self._job_name, self._job_html_url)):
return
job_data = {}
# quick check @self.job_api_url is valid url before request. it's set to "missing" for dummy BuildResult
if "http" in self.job_api_url:
try:
job_data = get_gh_api(self.job_api_url).json()
except Exception:
pass
# job_name can be set manually
self._job_name = self._job_name or job_data.get("name", "unknown")
self._job_html_url = job_data.get("html_url", "")
@staticmethod
def get_report_name(name: str, suffix: Union[str, int]) -> Path:
assert "/" not in str(suffix)
return Path(f"build_report_{suffix}_{name}.json")
@staticmethod
def missing_result(build_name: str) -> "BuildResult":
return BuildResult(build_name, "", [], "missing", ERROR, 0, "missing")
def write_json(self, directory: Union[Path, str] = REPORT_PATH) -> Path:
path = Path(directory) / self.get_report_name(
2024-02-04 19:12:37 +00:00
self.build_name, self.pr_number or normalize_string(self.head_ref)
)
path.write_text(
json.dumps(
{
"build_name": self.build_name,
"log_url": self.log_url,
"build_urls": self.build_urls,
"version": self.version,
"status": self.status,
"elapsed_seconds": self.elapsed_seconds,
"job_api_url": self.job_api_url,
"pr_number": self.pr_number,
"head_ref": self.head_ref,
}
),
encoding="utf-8",
)
# TODO: remove after the artifacts are in S3 completely
env_path = Path(os.getenv("GITHUB_ENV", "/dev/null"))
with env_path.open("a", encoding="utf-8") as ef:
ef.write(f"BUILD_URLS={path.stem}")
return path
2023-01-04 11:18:53 +00:00
BuildResults = List[BuildResult]
class ReportColorTheme:
class ReportColor:
yellow = "#FFB400"
red = "#F00"
green = "#0A0"
blue = "#00B4FF"
default = (ReportColor.green, ReportColor.red, ReportColor.yellow)
bugfixcheck = (ReportColor.yellow, ReportColor.blue, ReportColor.blue)
ColorTheme = Tuple[str, str, str]
def _format_header(
header: str, branch_name: str, branch_url: Optional[str] = None
) -> str:
# Following line does not lower CI->Ci and SQLancer->Sqlancer. It only
# capitalizes the first letter and doesn't touch the rest of the word
result = " ".join([w[0].upper() + w[1:] for w in header.split(" ") if w])
2021-09-10 11:52:21 +00:00
result = result.replace("Clickhouse", "ClickHouse")
result = result.replace("clickhouse", "ClickHouse")
if "ClickHouse" not in result:
result = f"ClickHouse {result}"
2021-09-10 11:52:21 +00:00
if branch_url:
result = f'{result} for <a href="{branch_url}">{branch_name}</a>'
2021-09-10 11:52:21 +00:00
else:
result = f"{result} for {branch_name}"
2021-09-10 11:52:21 +00:00
return result
def _get_status_style(status: str, colortheme: Optional[ColorTheme] = None) -> str:
2023-09-01 17:15:11 +00:00
ok_statuses = (OK, SUCCESS, "PASSED")
fail_statuses = (FAIL, FAILURE, ERROR, "FAILED", "Timeout", "NOT_FAILED")
if colortheme is None:
colortheme = ReportColorTheme.default
2021-09-10 11:52:21 +00:00
style = "font-weight: bold;"
if status in ok_statuses:
style += f"color: {colortheme[0]};"
elif status in fail_statuses:
style += f"color: {colortheme[1]};"
2021-09-10 11:52:21 +00:00
else:
style += f"color: {colortheme[2]};"
2021-09-10 11:52:21 +00:00
return style
def _get_html_url_name(url):
base_name = ""
if isinstance(url, str):
base_name = os.path.basename(url)
if isinstance(url, tuple):
base_name = url[1]
if "?" in base_name:
base_name = base_name.split("?")[0]
if base_name is not None:
return base_name.replace("%2B", "+").replace("%20", " ")
return None
2021-09-10 11:52:21 +00:00
def _get_html_url(url):
href = None
name = None
2021-09-10 11:52:21 +00:00
if isinstance(url, str):
href, name = url, _get_html_url_name(url)
2021-09-10 11:52:21 +00:00
if isinstance(url, tuple):
2022-02-21 07:29:10 +00:00
href, name = url[0], _get_html_url_name(url)
if href and name:
2022-11-14 14:48:51 +00:00
return f'<a href="{href}">{_get_html_url_name(url)}</a>'
return ""
def create_test_html_report(
header: str,
test_results: TestResults,
raw_log_url: str,
task_url: str,
job_url: str,
branch_url: str,
branch_name: str,
commit_url: str,
additional_urls: Optional[List[str]] = None,
statuscolors: Optional[ColorTheme] = None,
) -> str:
if additional_urls is None:
additional_urls = []
if test_results:
rows_part = []
2021-09-10 11:52:21 +00:00
num_fails = 0
has_test_time = False
has_log_urls = False
# Display entires with logs at the top (they correspond to failed tests)
test_results.sort(
key=lambda result: result.raw_logs is None and result.log_files is None
)
for test_result in test_results:
colspan = 0
if test_result.log_files is not None:
has_log_urls = True
2021-09-10 11:52:21 +00:00
row = []
if test_result.raw_logs is not None:
row.append('<tr class="expandable">')
else:
row.append("<tr>")
row.append(f"<td>{test_result.name}</td>")
colspan += 1
style = _get_status_style(test_result.status, colortheme=statuscolors)
2021-09-10 11:52:21 +00:00
# Allow to quickly scroll to the first failure.
fail_id = ""
has_error = test_result.status in ("FAIL", "NOT_FAILED")
2023-05-05 17:53:15 +00:00
if has_error:
2021-09-10 11:52:21 +00:00
num_fails = num_fails + 1
fail_id = f'id="fail{num_fails}" '
2021-09-10 11:52:21 +00:00
row.append(f'<td {fail_id}style="{style}">{test_result.status}</td>')
colspan += 1
2021-09-10 11:52:21 +00:00
if test_result.time is not None:
has_test_time = True
row.append(f"<td>{test_result.time}</td>")
colspan += 1
2021-09-10 11:52:21 +00:00
if test_result.log_urls is not None:
has_log_urls = True
test_logs_html = "<br>".join(
[_get_html_url(url) for url in test_result.log_urls]
)
row.append(f"<td>{test_logs_html}</td>")
colspan += 1
2021-09-10 11:52:21 +00:00
row.append("</tr>")
2023-09-11 16:13:52 +00:00
rows_part.append("\n".join(row))
if test_result.raw_logs is not None:
raw_logs = escape(test_result.raw_logs)
row_raw_logs = (
'<tr class="expandable-content">'
f'<td colspan="{colspan}"><pre>{raw_logs}</pre></td>'
"</tr>"
)
rows_part.append(row_raw_logs)
2021-09-10 11:52:21 +00:00
headers = BASE_HEADERS.copy()
2021-09-10 11:52:21 +00:00
if has_test_time:
headers.append("Test time, sec.")
if has_log_urls:
headers.append("Logs")
2021-09-10 11:52:21 +00:00
2022-11-14 14:47:09 +00:00
headers_html = "".join(["<th>" + h + "</th>" for h in headers])
test_part = HTML_TEST_PART.format(headers=headers_html, rows="".join(rows_part))
2021-09-10 11:52:21 +00:00
else:
test_part = ""
additional_html_urls = " ".join(
[_get_html_url(url) for url in sorted(additional_urls, key=_get_html_url_name)]
)
2021-09-10 11:52:21 +00:00
2022-08-09 16:49:55 +00:00
raw_log_name = os.path.basename(raw_log_url)
2022-08-11 10:05:16 +00:00
if "?" in raw_log_name:
raw_log_name = raw_log_name.split("?")[0]
2022-08-09 16:49:55 +00:00
html = HTML_BASE_TEST_TEMPLATE.format(
2021-09-10 11:52:21 +00:00
title=_format_header(header, branch_name),
header=_format_header(header, branch_name, branch_url),
2022-08-09 16:49:55 +00:00
raw_log_name=raw_log_name,
2021-09-10 11:52:21 +00:00
raw_log_url=raw_log_url,
task_url=task_url,
2022-08-09 16:49:55 +00:00
job_url=job_url,
2021-09-10 11:52:21 +00:00
test_part=test_part,
branch_name=branch_name,
commit_url=commit_url,
additional_urls=additional_html_urls,
2021-09-10 11:52:21 +00:00
)
return html
2021-09-10 11:52:21 +00:00
2023-09-01 12:05:18 +00:00
HTML_BASE_BUILD_TEMPLATE = (
f"{HEAD_HTML_TEMPLATE}"
"""<p class="links">
<a href="{commit_url}">Commit</a>
<a href="{task_url}">Task (github actions)</a>
</p>
2021-09-10 11:52:21 +00:00
<table>
<tr>
<th>Config/job name</th>
2021-09-10 11:52:21 +00:00
<th>Compiler</th>
<th>Build type</th>
<th>Version</th>
2021-09-10 11:52:21 +00:00
<th>Sanitizer</th>
2024-02-07 19:58:40 +00:00
<th>Coverage</th>
2021-09-10 11:52:21 +00:00
<th>Status</th>
<th>Build log</th>
<th>Build time</th>
<th class="artifacts">Artifacts</th>
2023-05-24 20:29:36 +00:00
<th>Comment</th>
2021-09-10 11:52:21 +00:00
</tr>
{rows}
</table>
"""
2023-09-01 12:05:18 +00:00
f"{FOOTER_HTML_TEMPLATE}"
)
2021-09-10 11:52:21 +00:00
LINK_TEMPLATE = '<a href="{url}">{text}</a>'
def create_build_html_report(
2023-01-04 11:18:53 +00:00
header: str,
build_results: BuildResults,
task_url: str,
branch_url: str,
branch_name: str,
commit_url: str,
) -> str:
rows = []
for build_result in build_results:
for artifact_urls in build_result.grouped_urls:
row = ["<tr>"]
row.append(
f"<td>{build_result.build_name}<br/>{build_result.job_link}</td>"
)
row.append(f"<td>{build_result.compiler}</td>")
if build_result.debug_build:
row.append("<td>debug</td>")
else:
row.append("<td>relwithdebuginfo</td>")
row.append(f"<td>{build_result.version}</td>")
if build_result.sanitizer:
row.append(f"<td>{build_result.sanitizer}</td>")
else:
row.append("<td>none</td>")
2024-02-07 19:58:40 +00:00
row.append(f"<td>{build_result.coverage}</td>")
if build_result.status:
style = _get_status_style(build_result.status)
row.append(f'<td style="{style}">{build_result.status}</td>')
else:
style = _get_status_style(ERROR)
row.append(f'<td style="{style}">error</td>')
2021-09-10 11:52:21 +00:00
row.append(f'<td><a href="{build_result.log_url}">link</a></td>')
2023-05-24 20:29:36 +00:00
delta = "unknown"
if build_result.elapsed_seconds:
delta = str(datetime.timedelta(seconds=build_result.elapsed_seconds))
row.append(f"<td>{delta}</td>")
links = []
link_separator = "<br/>"
if artifact_urls:
for artifact_url in artifact_urls:
links.append(
LINK_TEMPLATE.format(
text=_get_html_url_name(artifact_url), url=artifact_url
)
)
row.append(f"<td>{link_separator.join(links)}</td>")
comment = build_result.comment
if (
build_result.build_config is not None
and build_result.build_config.sparse_checkout
):
2024-01-18 18:21:26 +00:00
comment += " (note: sparse checkout is used, see update-submodules.sh)"
row.append(f"<td>{comment}</td>")
row.append("</tr>")
rows.append("".join(row))
2021-09-10 11:52:21 +00:00
return HTML_BASE_BUILD_TEMPLATE.format(
title=_format_header(header, branch_name),
header=_format_header(header, branch_name, branch_url),
rows="".join(rows),
2021-09-10 11:52:21 +00:00
task_url=task_url,
branch_name=branch_name,
commit_url=commit_url,
)