ClickHouse/tests/ci/merge_pr.py
2023-03-21 11:29:16 +01:00

289 lines
9.8 KiB
Python

#!/usr/bin/env python
"""Script to check if PR is mergeable and merge it"""
import argparse
import logging
from datetime import datetime
from os import getenv
from pprint import pformat
from typing import Dict, List
from github.PaginatedList import PaginatedList
from github.PullRequestReview import PullRequestReview
from github.WorkflowRun import WorkflowRun
from commit_status_helper import get_commit_filtered_statuses
from get_robot_token import get_best_robot_token
from github_helper import GitHub, NamedUser, PullRequest, Repository
from pr_info import PRInfo
# The team name for accepted approvals
TEAM_NAME = getenv("GITHUB_TEAM_NAME", "core")
class Reviews:
STATES = ["CHANGES_REQUESTED", "APPROVED"]
def __init__(self, pr: PullRequest):
"""The reviews are proceed in the next logic:
- if review for an author does not exist, set it
- the review status can be changed from CHANGES_REQUESTED and APPROVED
only to either one
"""
logging.info("Checking the PR for approvals")
self.pr = pr
reviews = pr.get_reviews()
# self.reviews is a dict of latest CHANGES_REQUESTED or APPROVED review
# per user
# NamedUsed has proper __eq__ and __hash__, so it's safe to use it
self.reviews = {} # type: Dict[NamedUser, PullRequestReview]
for r in reviews:
user = r.user
if not self.reviews.get(user):
self.reviews[user] = r
continue
# Do not process other statuses than STATES for existing user keys
if r.state not in self.STATES:
continue
# If the user has a status other than STATES, we overwrite it by a
# review w/ a proper state w/o checking the date
if self.reviews[user].state not in self.STATES:
self.reviews[user] = r
continue
# Keep the latest review per user
if self.reviews[user].submitted_at < r.submitted_at:
self.reviews[user] = r
def is_approved(self, team: List[NamedUser]) -> bool:
"""Checks if the PR is approved, and no changes made after the last approval"""
if not self.reviews:
logging.info("There aren't reviews for PR #%s", self.pr.number)
return False
logging.info(
"The following users have reviewed the PR:\n %s",
"\n ".join(
f"{user.login}: {review.state}" for user, review in self.reviews.items()
),
)
filtered_reviews = {
user: review
for user, review in self.reviews.items()
if review.state in self.STATES and user in team
}
# We consider reviews only from the given list of users
changes_requested = {
user: review
for user, review in filtered_reviews.items()
if review.state == "CHANGES_REQUESTED"
}
if changes_requested:
logging.info(
"The following users requested changes for the PR: %s",
", ".join(user.login for user in changes_requested.keys()),
)
return False
approved = {
user: review
for user, review in filtered_reviews.items()
if review.state == "APPROVED"
}
if not approved:
logging.info(
"The PR #%s is not approved by any of %s team member",
self.pr.number,
TEAM_NAME,
)
return False
logging.info(
"The following users from %s team approved the PR: %s",
TEAM_NAME,
", ".join(user.login for user in approved.keys()),
)
# The only reliable place to get the 100% accurate last_modified
# info is when the commit was pushed to GitHub. The info is
# available as a header 'last-modified' of /{org}/{repo}/commits/{sha}.
# Unfortunately, it's formatted as 'Wed, 04 Jan 2023 11:05:13 GMT'
commit = self.pr.head.repo.get_commit(self.pr.head.sha)
if commit.stats.last_modified is None:
logging.warning("Unable to get info about the commit %s", self.pr.head.sha)
return False
last_changed = datetime.strptime(
commit.stats.last_modified, "%a, %d %b %Y %H:%M:%S GMT"
)
logging.info("The PR is changed at %s", last_changed.isoformat())
approved_at = max(review.submitted_at for review in approved.values())
if approved_at == datetime.fromtimestamp(0):
logging.info(
"Unable to get `datetime.fromtimestamp(0)`, "
"here's debug info about reviews: %s",
"\n".join(pformat(review) for review in self.reviews.values()),
)
else:
logging.info("The PR is approved at %s", approved_at.isoformat())
if approved_at < last_changed:
logging.info(
"There are changes done at %s after approval at %s",
last_changed.isoformat(),
approved_at.isoformat(),
)
return False
return True
def get_workflows_for_head(repo: Repository, head_sha: str) -> List[WorkflowRun]:
# The monkey-patch until the PR is merged:
# https://github.com/PyGithub/PyGithub/pull/2408
return list(
PaginatedList(
WorkflowRun,
repo._requester, # type:ignore # pylint:disable=protected-access
f"{repo.url}/actions/runs",
{"head_sha": head_sha},
list_item="workflow_runs",
)
)
def parse_args() -> argparse.Namespace:
pr_info = PRInfo()
parser = argparse.ArgumentParser(
formatter_class=argparse.ArgumentDefaultsHelpFormatter,
description="Script to merge the given PR. Additional checks for approved "
"status and green commit statuses could be done",
)
parser.add_argument(
"--dry-run",
action="store_true",
help="if set, the script won't merge the PR, just check the conditions",
)
parser.add_argument(
"--check-approved",
action="store_true",
help="if set, checks that the PR is approved and no changes required",
)
parser.add_argument(
"--check-running-workflows", default=True, help=argparse.SUPPRESS
)
parser.add_argument(
"--no-check-running-workflows",
dest="check_running_workflows",
action="store_false",
default=argparse.SUPPRESS,
help="(dangerous) if set, skip checking for running workflows for the PR head",
)
parser.add_argument("--check-green", default=True, help=argparse.SUPPRESS)
parser.add_argument(
"--no-check-green",
dest="check_green",
action="store_false",
default=argparse.SUPPRESS,
help="(dangerous) if set, skip check commit to having all green statuses",
)
parser.add_argument(
"--repo",
default=pr_info.repo_full_name,
help="PR number to check",
)
parser.add_argument(
"--pr",
type=int,
default=pr_info.number,
help="PR number to check",
)
parser.add_argument(
"--token",
type=str,
default="",
help="a token to use for GitHub API requests, will be received from SSM "
"if empty",
)
args = parser.parse_args()
args.pr_info = pr_info
return args
def main():
logging.basicConfig(level=logging.INFO, format="%(asctime)s %(message)s")
args = parse_args()
logging.info("Going to process PR #%s in repo %s", args.pr, args.repo)
token = args.token or get_best_robot_token()
gh = GitHub(token)
repo = gh.get_repo(args.repo)
# An ugly and not nice fix to patch the wrong organization URL,
# see https://github.com/PyGithub/PyGithub/issues/2395#issuecomment-1378629710
# pylint: disable=protected-access
repo.organization._url.value = repo.organization.url.replace( # type: ignore
"/users/", "/orgs/", 1
)
# pylint: enable=protected-access
pr = repo.get_pull(args.pr)
if pr.is_merged():
logging.info("The PR #%s is already merged", pr.number)
return
not_ready_to_merge = pr.draft or "WIP" in pr.title
if not_ready_to_merge:
logging.info("The PR #%s is not ready for merge, stopping", pr.number)
return
if args.check_running_workflows:
workflows = get_workflows_for_head(repo, pr.head.sha)
workflows_in_progress = [wf for wf in workflows if wf.status != "completed"]
# At most one workflow in progress is fine. We check that there no
# cases like, e.g. PullRequestCI and DocksCheck in progress at once
if len(workflows_in_progress) > 1:
logging.info(
"The PR #%s has more than one workflows in progress, check URLs:\n%s",
pr.number,
"\n".join(wf.html_url for wf in workflows_in_progress),
)
return
if args.check_green:
logging.info("Checking that all PR's statuses are green")
commit = repo.get_commit(pr.head.sha)
failed_statuses = [
status.context
for status in get_commit_filtered_statuses(commit)
if status.state != "success"
]
if failed_statuses:
logging.warning(
"Some statuses aren't success:\n %s", ",\n ".join(failed_statuses)
)
return
if args.check_approved:
reviews = Reviews(pr)
team = repo.organization.get_team_by_slug(TEAM_NAME)
members = list(team.get_members())
if not reviews.is_approved(members):
logging.warning("We don't merge the PR")
return
logging.info("Merging the PR")
if not args.dry_run:
pr.merge()
if __name__ == "__main__":
main()