import json import time from dataclasses import dataclass, asdict from enum import Enum from pathlib import Path from typing import Dict, Optional, Any, Union, Sequence, List, Set from ci_config import CI from ci_utils import Utils, GH from commit_status_helper import CommitStatusData from env_helper import ( TEMP_PATH, CI_CONFIG_PATH, S3_BUILDS_BUCKET, GITHUB_RUN_URL, REPORT_PATH, ) from report import BuildResult from s3_helper import S3Helper from digest_helper import JobDigester @dataclass class PendingState: updated_at: float run_url: str class CiCache: """ CI cache is a bunch of records. Record is a file stored under special location on s3. The file name has a format: _[]--___.ci RECORD_TYPE: SUCCESSFUL - for successful jobs PENDING - for pending jobs ATTRIBUTES: release - for jobs being executed on the release branch including master branch (not a PR branch) """ _REQUIRED_DIGESTS = [CI.JobNames.DOCS_CHECK, CI.BuildNames.PACKAGE_RELEASE] _S3_CACHE_PREFIX = "CI_cache_v1" _CACHE_BUILD_REPORT_PREFIX = "build_report" _RECORD_FILE_EXTENSION = ".ci" _LOCAL_CACHE_PATH = Path(TEMP_PATH) / "ci_cache" _ATTRIBUTE_RELEASE = "release" # divider symbol 1 _DIV1 = "--" # divider symbol 2 _DIV2 = "_" assert _DIV1 != _DIV2 class RecordType(Enum): SUCCESSFUL = "successful" PENDING = "pending" FAILED = "failed" @dataclass class Record: record_type: "CiCache.RecordType" job_name: str job_digest: str batch: int num_batches: int release_branch: bool file: str = "" def to_str_key(self): """other fields must not be included in the hash str""" return "_".join( [self.job_name, self.job_digest, str(self.batch), str(self.num_batches)] ) class JobType(Enum): DOCS = "DOCS" SRCS = "SRCS" @classmethod def is_docs_job(cls, job_name: str) -> bool: return job_name == CI.JobNames.DOCS_CHECK @classmethod def is_srcs_job(cls, job_name: str) -> bool: return not cls.is_docs_job(job_name) @classmethod def get_type_by_name(cls, job_name: str) -> "CiCache.JobType": res = cls.SRCS if cls.is_docs_job(job_name): res = cls.DOCS elif cls.is_srcs_job(job_name): res = cls.SRCS else: assert False return res def __init__( self, s3: S3Helper, job_digests: Dict[str, str], cache_enabled: bool = True, ): self.enabled = cache_enabled self.jobs_to_skip = [] # type: List[str] self.jobs_to_wait = {} # type: Dict[str, CI.JobConfig] self.jobs_to_do = {} # type: Dict[str, CI.JobConfig] self.s3 = s3 self.job_digests = job_digests self.cache_s3_paths = { job_type: f"{self._S3_CACHE_PREFIX}/{job_type.value}-{self._get_digest_for_job_type(self.job_digests, job_type)}/" for job_type in self.JobType } self.s3_record_prefixes = { record_type: record_type.value for record_type in self.RecordType } self.records: Dict["CiCache.RecordType", Dict[str, "CiCache.Record"]] = { record_type: {} for record_type in self.RecordType } self.updated = False self.cache_data_fetched = True if not self._LOCAL_CACHE_PATH.exists(): self._LOCAL_CACHE_PATH.mkdir(parents=True, exist_ok=True) @classmethod def calc_digests_and_create( cls, s3: S3Helper, job_configs: Dict[str, CI.JobConfig], cache_enabled: bool = True, dry_run: bool = False, ) -> "CiCache": job_digester = JobDigester(dry_run=dry_run) digests = {} print("::group::Job Digests") for job, job_config in job_configs.items(): digest = job_digester.get_job_digest(job_config.digest) digests[job] = digest print(f" job [{job.rjust(50)}] has digest [{digest}]") for job in cls._REQUIRED_DIGESTS: if job not in job_configs: digest = job_digester.get_job_digest(CI.get_job_config(job).digest) digests[job] = digest print( f" job [{job.rjust(50)}] required for CI Cache has digest [{digest}]" ) print("::endgroup::") return CiCache(s3, digests, cache_enabled=cache_enabled) def _get_digest_for_job_type( self, job_digests: Dict[str, str], job_type: JobType ) -> str: if job_type == self.JobType.DOCS: res = job_digests[CI.JobNames.DOCS_CHECK] elif job_type == self.JobType.SRCS: if CI.BuildNames.PACKAGE_RELEASE in job_digests: res = job_digests[CI.BuildNames.PACKAGE_RELEASE] else: assert False, "BUG, no build job in digest' list" else: assert False, "BUG, New JobType? - please update the function" return res def _get_record_file_name( self, record_type: RecordType, job_name: str, batch: int, num_batches: int, release_branch: bool, ) -> str: prefix = self.s3_record_prefixes[record_type] prefix_extended = ( self._DIV2.join([prefix, self._ATTRIBUTE_RELEASE]) if release_branch else prefix ) assert self._DIV1 not in job_name, f"Invalid job name {job_name}" job_name = self._DIV2.join( [job_name, self.job_digests[job_name], str(batch), str(num_batches)] ) file_name = self._DIV1.join([prefix_extended, job_name]) file_name += self._RECORD_FILE_EXTENSION return file_name def _get_record_s3_path(self, job_name: str) -> str: return self.cache_s3_paths[self.JobType.get_type_by_name(job_name)] def _parse_record_file_name( self, record_type: RecordType, file_name: str ) -> Optional["CiCache.Record"]: # validate filename if ( not file_name.endswith(self._RECORD_FILE_EXTENSION) or not len(file_name.split(self._DIV1)) == 2 ): print("ERROR: wrong file name format") return None file_name = file_name.removesuffix(self._RECORD_FILE_EXTENSION) release_branch = False prefix_extended, job_suffix = file_name.split(self._DIV1) record_type_and_attribute = prefix_extended.split(self._DIV2) # validate filename prefix failure = False if not 0 < len(record_type_and_attribute) <= 2: print("ERROR: wrong file name prefix") failure = True if ( len(record_type_and_attribute) > 1 and record_type_and_attribute[1] != self._ATTRIBUTE_RELEASE ): print("ERROR: wrong record attribute") failure = True if record_type_and_attribute[0] != self.s3_record_prefixes[record_type]: print("ERROR: wrong record type") failure = True if failure: return None if ( len(record_type_and_attribute) > 1 and record_type_and_attribute[1] == self._ATTRIBUTE_RELEASE ): release_branch = True job_properties = job_suffix.split(self._DIV2) job_name, job_digest, batch, num_batches = ( self._DIV2.join(job_properties[:-3]), job_properties[-3], int(job_properties[-2]), int(job_properties[-1]), ) if not Utils.is_hex(job_digest): print("ERROR: wrong record job digest") return None record = self.Record( record_type, job_name, job_digest, batch, num_batches, release_branch, file="", ) return record def print_status(self): print(f"Cache enabled: [{self.enabled}]") for record_type in self.RecordType: GH.print_in_group( f"Cache records: [{record_type}]", list(self.records[record_type]) ) GH.print_in_group( "Jobs to do:", list(self.jobs_to_do.items()), ) GH.print_in_group("Jobs to skip:", self.jobs_to_skip) GH.print_in_group( "Jobs to wait:", list(self.jobs_to_wait.items()), ) return self @staticmethod def dump_run_config(indata: Dict[str, Any]) -> None: assert indata assert CI_CONFIG_PATH with open(CI_CONFIG_PATH, "w", encoding="utf-8") as json_file: json.dump(indata, json_file, indent=2) def update(self): """ Pulls cache records from s3. Only records name w/o content. """ if not self.enabled: return self for record_type in self.RecordType: prefix = self.s3_record_prefixes[record_type] cache_list = self.records[record_type] for job_type in self.JobType: path = self.cache_s3_paths[job_type] records = self.s3.list_prefix(f"{path}{prefix}", S3_BUILDS_BUCKET) records = [record.split("/")[-1] for record in records] for file in records: record = self._parse_record_file_name( record_type=record_type, file_name=file ) if not record: print(f"ERROR: failed to parse cache record [{file}]") continue if ( record.job_name not in self.job_digests or self.job_digests[record.job_name] != record.job_digest ): # skip records we are not interested in continue if record.to_str_key() not in cache_list: cache_list[record.to_str_key()] = record self.cache_data_fetched = False elif ( not cache_list[record.to_str_key()].release_branch and record.release_branch ): # replace a non-release record with a release one cache_list[record.to_str_key()] = record self.cache_data_fetched = False self.updated = True return self def fetch_records_data(self): """ Pulls CommitStatusData for all cached jobs from s3 """ if not self.updated: self.update() if self.cache_data_fetched: # there are no records without fetched data - no need to fetch return self # clean up for file in self._LOCAL_CACHE_PATH.glob("*.ci"): file.unlink() # download all record files for job_type in self.JobType: path = self.cache_s3_paths[job_type] for record_type in self.RecordType: prefix = self.s3_record_prefixes[record_type] _ = self.s3.download_files( bucket=S3_BUILDS_BUCKET, s3_path=f"{path}{prefix}", file_suffix=self._RECORD_FILE_EXTENSION, local_directory=self._LOCAL_CACHE_PATH, ) # validate we have files for all records and save file names meanwhile for record_type in self.RecordType: record_list = self.records[record_type] for _, record in record_list.items(): record_file_name = self._get_record_file_name( record_type, record.job_name, record.batch, record.num_batches, record.release_branch, ) assert ( self._LOCAL_CACHE_PATH / record_file_name ).is_file(), f"BUG. Record file must be present: {self._LOCAL_CACHE_PATH / record_file_name}" record.file = record_file_name self.cache_data_fetched = True return self def exist( self, record_type: "CiCache.RecordType", job: str, batch: int, num_batches: int, release_branch: bool, ) -> bool: if not self.updated: self.update() record_key = self.Record( record_type, job, self.job_digests[job], batch, num_batches, release_branch, ).to_str_key() res = record_key in self.records[record_type] if release_branch: return res and self.records[record_type][record_key].release_branch else: return res def push( self, record_type: "CiCache.RecordType", job: str, batches: Union[int, Sequence[int]], num_batches: int, status: Union[CommitStatusData, PendingState], release_branch: bool = False, ) -> None: """ Pushes a cache record (CommitStatusData) @release_branch adds "release" attribute to a record """ if isinstance(batches, int): batches = [batches] for batch in batches: record_file = self._LOCAL_CACHE_PATH / self._get_record_file_name( record_type, job, batch, num_batches, release_branch ) record_s3_path = self._get_record_s3_path(job) if record_type == self.RecordType.SUCCESSFUL: assert isinstance(status, CommitStatusData) status.dump_to_file(record_file) elif record_type == self.RecordType.FAILED: assert isinstance(status, CommitStatusData) status.dump_to_file(record_file) elif record_type == self.RecordType.PENDING: assert isinstance(status, PendingState) with open(record_file, "w", encoding="utf-8") as json_file: json.dump(asdict(status), json_file) else: assert False _ = self.s3.upload_file( bucket=S3_BUILDS_BUCKET, file_path=record_file, s3_path=record_s3_path + record_file.name, ) record = self.Record( record_type, job, self.job_digests[job], batch, num_batches, release_branch, file=record_file.name, ) if ( record.release_branch or record.to_str_key() not in self.records[record_type] ): self.records[record_type][record.to_str_key()] = record def get( self, record_type: "CiCache.RecordType", job: str, batch: int, num_batches: int ) -> Optional[Union[CommitStatusData, PendingState]]: """ Gets a cache record data for a job, or None if a cache miss """ if not self.cache_data_fetched: self.fetch_records_data() record_key = self.Record( record_type, job, self.job_digests[job], batch, num_batches, release_branch=False, ).to_str_key() if record_key not in self.records[record_type]: return None record_file_name = self.records[record_type][record_key].file res = CommitStatusData.load_from_file( self._LOCAL_CACHE_PATH / record_file_name ) # type: CommitStatusData return res def delete( self, record_type: "CiCache.RecordType", job: str, batch: int, num_batches: int, release_branch: bool, ) -> None: """ deletes record from the cache """ raise NotImplementedError("Let's try make cache push-and-read-only") # assert ( # record_type == self.RecordType.PENDING # ), "FIXME: delete is supported for pending records only" # record_file_name = self._get_record_file_name( # self.RecordType.PENDING, # job, # batch, # num_batches, # release_branch=release_branch, # ) # record_s3_path = self._get_record_s3_path(job) # self.s3.delete_file_from_s3(S3_BUILDS_BUCKET, record_s3_path + record_file_name) # record_key = self.Record( # record_type, # job, # self.job_digests[job], # batch, # num_batches, # release_branch=False, # ).to_str_key() # if record_key in self.records[record_type]: # del self.records[record_type][record_key] def is_successful( self, job: str, batch: int, num_batches: int, release_branch: bool ) -> bool: """ checks if a given job have already been done successfully """ return self.exist( self.RecordType.SUCCESSFUL, job, batch, num_batches, release_branch ) def has_evidence(self, job: str, job_config: CI.JobConfig) -> bool: """ checks if the job has been seen in master/release CI function is to be used to check if change did not affect the job :param job_config: :param job: :return: """ return ( self.is_successful( job=job, batch=0, num_batches=job_config.num_batches, release_branch=True, ) or self.is_pending( job=job, batch=0, num_batches=job_config.num_batches, release_branch=True, ) or self.is_failed( job=job, batch=0, num_batches=job_config.num_batches, release_branch=True, ) ) def is_failed( self, job: str, batch: int, num_batches: int, release_branch: bool ) -> bool: """ checks if a given job have already been done with failure """ return self.exist( self.RecordType.FAILED, job, batch, num_batches, release_branch ) def is_pending( self, job: str, batch: int, num_batches: int, release_branch: bool ) -> bool: """ check pending record in the cache for a given job @release_branch - checks that "release" attribute is set for a record """ if self.is_successful( job, batch, num_batches, release_branch ) or self.is_failed(job, batch, num_batches, release_branch): return False return self.exist( self.RecordType.PENDING, job, batch, num_batches, release_branch ) def push_successful( self, job: str, batch: int, num_batches: int, job_status: CommitStatusData, release_branch: bool = False, ) -> None: """ Pushes a cache record (CommitStatusData) @release_branch adds "release" attribute to a record """ self.push( self.RecordType.SUCCESSFUL, job, [batch], num_batches, job_status, release_branch, ) def push_failed( self, job: str, batch: int, num_batches: int, job_status: CommitStatusData, release_branch: bool = False, ) -> None: """ Pushes a cache record of type Failed (CommitStatusData) @release_branch adds "release" attribute to a record """ self.push( self.RecordType.FAILED, job, [batch], num_batches, job_status, release_branch, ) def push_pending( self, job: str, batches: List[int], num_batches: int, release_branch: bool ) -> None: """ pushes pending record for a job to the cache """ pending_state = PendingState(time.time(), run_url=GITHUB_RUN_URL) self.push( self.RecordType.PENDING, job, batches, num_batches, pending_state, release_branch, ) def push_pending_all(self, release_branch: bool) -> None: """ pushes pending records for all jobs that supposed to be run """ for job, job_config in self.jobs_to_do.items(): if ( job in self.jobs_to_wait or not job_config.has_digest() or job_config.disable_await ): # 1. "job in self.jobs_to_wait" - this job already has a pending record in cache # 2. "not job_config.has_digest()" - cache is not used for these jobs # 3. "job_config.disable_await" - await is explicitly disabled continue pending_state = PendingState(time.time(), run_url=GITHUB_RUN_URL) assert job_config.batches self.push( self.RecordType.PENDING, job, job_config.batches, job_config.num_batches, pending_state, release_branch, ) def get_successful( self, job: str, batch: int, num_batches: int ) -> Optional[CommitStatusData]: """ Gets a cache record (CommitStatusData) for a job, or None if a cache miss """ res = self.get(self.RecordType.SUCCESSFUL, job, batch, num_batches) assert res is None or isinstance(res, CommitStatusData) return res def delete_pending( self, job: str, batch: int, num_batches: int, release_branch: bool ) -> None: """ deletes pending record from the cache """ self.delete(self.RecordType.PENDING, job, batch, num_batches, release_branch) def download_build_reports(self, file_prefix: str = "") -> List[str]: """ not an ideal class for this method, but let it be as we store build reports in CI cache directory on s3 and CiCache knows where exactly @file_prefix allows filtering out reports by git head_ref """ report_path = Path(REPORT_PATH) report_path.mkdir(exist_ok=True, parents=True) path = ( self._get_record_s3_path(CI.BuildNames.PACKAGE_RELEASE) + self._CACHE_BUILD_REPORT_PREFIX ) if file_prefix: path += "_" + file_prefix reports_files = self.s3.download_files( bucket=S3_BUILDS_BUCKET, s3_path=path, file_suffix=".json", local_directory=report_path, ) return reports_files def upload_build_report(self, build_result: BuildResult) -> str: result_json_path = build_result.write_json(Path(TEMP_PATH)) s3_path = ( self._get_record_s3_path(CI.BuildNames.PACKAGE_RELEASE) + result_json_path.name ) return self.s3.upload_file( bucket=S3_BUILDS_BUCKET, file_path=result_json_path, s3_path=s3_path ) def filter_out_not_affected_jobs(self): """ Filter is to be applied in PRs to remove jobs that are not affected by the change :return: """ remove_from_workflow = [] required_builds = [] has_test_jobs_to_skip = False for job_name, job_config in self.jobs_to_do.items(): if CI.is_test_job(job_name) and job_name != CI.JobNames.BUILD_CHECK: if job_config.reference_job_name: reference_name = job_config.reference_job_name reference_config = CI.JOB_CONFIGS[reference_name] else: reference_name = job_name reference_config = job_config if self.has_evidence( job=reference_name, job_config=reference_config, ): remove_from_workflow.append(job_name) if job_name != CI.JobNames.DOCS_CHECK: has_test_jobs_to_skip = True else: required_builds += ( job_config.required_builds if job_config.required_builds else [] ) if has_test_jobs_to_skip: # If there are tests to skip, it means builds are not affected as well. # No need to test builds. Let's keep all builds required for test jobs and skip the others for job_name, job_config in self.jobs_to_do.items(): if CI.is_build_job(job_name): if job_name not in required_builds: remove_from_workflow.append(job_name) for job in remove_from_workflow: print(f"Filter job [{job}] - not affected by the change") if job in self.jobs_to_do: del self.jobs_to_do[job] if job in self.jobs_to_wait: del self.jobs_to_wait[job] if job in self.jobs_to_skip: self.jobs_to_skip.remove(job) # special handling for the special job: BUILD_CHECK has_builds = False for job in list(self.jobs_to_do) + self.jobs_to_skip: if CI.is_build_job(job): has_builds = True break if not has_builds: if CI.JobNames.BUILD_CHECK in self.jobs_to_do: print( f"Filter job [{CI.JobNames.BUILD_CHECK}] - no builds are required in the workflow" ) del self.jobs_to_do[CI.JobNames.BUILD_CHECK] def await_pending_jobs(self, is_release: bool, dry_run: bool = False) -> None: """ await pending jobs to be finished @jobs_with_params - jobs to await. {JOB_NAME: {"batches": [BATCHES...], "num_batches": NUM_BATCHES}} returns successfully finished jobs: {JOB_NAME: [BATCHES...]} """ if not self.jobs_to_wait: print("CI cache: no pending jobs to wait - continue") return poll_interval_sec = 300 # TIMEOUT * MAX_ROUNDS_TO_WAIT must be less than 6h (GH job timeout) with a room for rest RunConfig work TIMEOUT = 3000 # 50 min MAX_ROUNDS_TO_WAIT = 6 round_cnt = 0 if not is_release: # in PRs we can wait only for builds, TIMEOUT*MAX_ROUNDS_TO_WAIT=100min is enough MAX_ROUNDS_TO_WAIT = 2 while round_cnt < MAX_ROUNDS_TO_WAIT: round_cnt += 1 GH.print_in_group( f"Wait pending jobs, round [{round_cnt}/{MAX_ROUNDS_TO_WAIT}]:", list(self.jobs_to_wait), ) # this is an initial approach to wait pending jobs: # start waiting for the next TIMEOUT seconds if there are more than X(=4) jobs to wait # wait TIMEOUT seconds in rounds. Y(=5) is the max number of rounds expired_sec = 0 start_at = int(time.time()) while expired_sec < TIMEOUT and self.jobs_to_wait: await_finished: Set[str] = set() if not dry_run: time.sleep(poll_interval_sec) self.update() for job_name, job_config in self.jobs_to_wait.items(): num_batches = job_config.num_batches job_config = CI.get_job_config(job_name) assert job_config.pending_batches assert job_config.batches pending_batches = list(job_config.pending_batches) for batch in pending_batches: if self.is_pending( job_name, batch, num_batches, release_branch=is_release and job_config.required_on_release_branch, ): continue if self.is_successful( job_name, batch, num_batches, release_branch=is_release and job_config.required_on_release_branch, ): print( f"Job [{job_name}_[{batch}/{num_batches}]] is not pending anymore" ) job_config.batches.remove(batch) if not job_config.batches: print(f"Remove job [{job_name}] from jobs_to_do") self.jobs_to_skip.append(job_name) del self.jobs_to_do[job_name] else: print( f"NOTE: Job [{job_name}:{batch}] finished failed - do not add to ready" ) job_config.pending_batches.remove(batch) if not job_config.pending_batches: await_finished.add(job_name) for job in await_finished: del self.jobs_to_wait[job] if not dry_run: expired_sec = int(time.time()) - start_at print( f"...awaiting continues... seconds left [{TIMEOUT - expired_sec}]" ) else: # make up for 2 iterations in dry_run expired_sec += int(TIMEOUT / 2) + 1 GH.print_in_group( "Remaining jobs:", [list(self.jobs_to_wait)], ) def apply( self, job_configs: Dict[str, CI.JobConfig], is_release: bool ) -> "CiCache": if not self.enabled: self.jobs_to_do = job_configs return self if not self.updated: self.update() for job, job_config in job_configs.items(): assert ( job_config.batches ), "Batches must be generated. check ci_settings.apply()" if job_config.run_always: self.jobs_to_do[job] = job_config continue ready_batches = [] for batch in job_config.batches: if self.is_successful( job, batch, job_config.num_batches, release_branch=is_release and job_config.required_on_release_branch, ): ready_batches.append(batch) elif self.is_pending( job, batch, job_config.num_batches, release_branch=is_release and job_config.required_on_release_branch, ): if job_config.pending_batches is None: job_config.pending_batches = [] job_config.pending_batches.append(batch) if ready_batches == job_config.batches: self.jobs_to_skip.append(job) else: for batch in ready_batches: job_config.batches.remove(batch) self.jobs_to_do[job] = job_config if job_config.pending_batches: self.jobs_to_wait[job] = job_config return self if __name__ == "__main__": # for testing job_digest = { "package_release": "bbbd3519d1", "package_aarch64": "bbbd3519d1", "package_asan": "bbbd3519d1", "package_ubsan": "bbbd3519d1", "package_tsan": "bbbd3519d1", "package_msan": "bbbd3519d1", "package_debug": "bbbd3519d1", "package_release_coverage": "bbbd3519d1", "binary_release": "bbbd3519d1", "binary_tidy": "bbbd3519d1", "binary_darwin": "bbbd3519d1", "binary_aarch64": "bbbd3519d1", "binary_aarch64_v80compat": "bbbd3519d1", "binary_freebsd": "bbbd3519d1", "binary_darwin_aarch64": "bbbd3519d1", "binary_ppc64le": "bbbd3519d1", "binary_amd64_compat": "bbbd3519d1", "binary_amd64_musl": "bbbd3519d1", "binary_riscv64": "bbbd3519d1", "binary_s390x": "bbbd3519d1", "binary_loongarch64": "bbbd3519d1", "Builds": "f5dffeecb8", "Install packages (release)": "ba0c89660e", "Install packages (aarch64)": "ba0c89660e", "Stateful tests (asan)": "32a9a1aba9", "Stateful tests (tsan)": "32a9a1aba9", "Stateful tests (msan)": "32a9a1aba9", "Stateful tests (ubsan)": "32a9a1aba9", "Stateful tests (debug)": "32a9a1aba9", "Stateful tests (release)": "32a9a1aba9", "Stateful tests (coverage)": "32a9a1aba9", "Stateful tests (aarch64)": "32a9a1aba9", "Stateful tests (release, ParallelReplicas)": "32a9a1aba9", "Stateful tests (debug, ParallelReplicas)": "32a9a1aba9", "Stateless tests (asan)": "deb6778b88", "Stateless tests (tsan)": "deb6778b88", "Stateless tests (msan)": "deb6778b88", "Stateless tests (ubsan)": "deb6778b88", "Stateless tests (debug)": "deb6778b88", "Stateless tests (release)": "deb6778b88", "Stateless tests (coverage)": "deb6778b88", "Stateless tests (aarch64)": "deb6778b88", "Stateless tests (release, old analyzer, s3, DatabaseReplicated)": "deb6778b88", "Stateless tests (debug, s3 storage)": "deb6778b88", "Stateless tests (tsan, s3 storage)": "deb6778b88", "Stress test (debug)": "aa298abf10", "Stress test (tsan)": "aa298abf10", "Upgrade check (debug)": "5ce4d3ee02", "Integration tests (asan, old analyzer)": "42e58be3aa", "Integration tests (tsan)": "42e58be3aa", "Integration tests (aarch64)": "42e58be3aa", "Integration tests flaky check (asan)": "42e58be3aa", "Compatibility check (release)": "ecb69d8c4b", "Compatibility check (aarch64)": "ecb69d8c4b", "Unit tests (release)": "09d00b702e", "Unit tests (asan)": "09d00b702e", "Unit tests (msan)": "09d00b702e", "Unit tests (tsan)": "09d00b702e", "Unit tests (ubsan)": "09d00b702e", "AST fuzzer (debug)": "c38ebf947f", "AST fuzzer (asan)": "c38ebf947f", "AST fuzzer (msan)": "c38ebf947f", "AST fuzzer (tsan)": "c38ebf947f", "AST fuzzer (ubsan)": "c38ebf947f", "Stateless tests flaky check (asan)": "deb6778b88", "Performance Comparison (release)": "a8a7179258", "ClickBench (release)": "45c07c4aa6", "ClickBench (aarch64)": "45c07c4aa6", "Docker server image": "6a24d5b187", "Docker keeper image": "6a24d5b187", "Docs check": "4764154c62", "Fast test": "cb269133f2", "Style check": "ffffffffff", "Stateful tests (ubsan, ParallelReplicas)": "32a9a1aba9", "Stress test (msan)": "aa298abf10", "Upgrade check (asan)": "5ce4d3ee02", } ci_cache = CiCache(job_digests=job_digest, cache_enabled=True, s3=S3Helper()) ci_cache.update()