2021-11-08 14:30:27 +00:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
|
|
|
|
try:
|
|
|
|
from clickhouse.utils.github.cherrypick import CherryPick
|
|
|
|
from clickhouse.utils.github.query import Query as RemoteRepo
|
|
|
|
from clickhouse.utils.github.local import Repository as LocalRepo
|
|
|
|
except:
|
2021-11-09 11:14:30 +00:00
|
|
|
from .cherrypick import CherryPick
|
|
|
|
from .query import Query as RemoteRepo
|
|
|
|
from .local import Repository as LocalRepo
|
2021-11-08 14:30:27 +00:00
|
|
|
|
|
|
|
import argparse
|
|
|
|
import logging
|
|
|
|
import re
|
|
|
|
import sys
|
|
|
|
|
|
|
|
|
|
|
|
class Backport:
|
|
|
|
def __init__(self, token, owner, name, team):
|
|
|
|
self._gh = RemoteRepo(token, owner=owner, name=name, team=team, max_page_size=30, min_page_size=7)
|
|
|
|
self._token = token
|
|
|
|
self.default_branch_name = self._gh.default_branch
|
|
|
|
self.ssh_url = self._gh.ssh_url
|
|
|
|
|
|
|
|
def getPullRequests(self, from_commit):
|
|
|
|
return self._gh.get_pull_requests(from_commit)
|
|
|
|
|
|
|
|
def getBranchesWithRelease(self):
|
|
|
|
branches = set()
|
|
|
|
for pull_request in self._gh.find_pull_requests("release"):
|
|
|
|
branches.add(pull_request['headRefName'])
|
|
|
|
return branches
|
|
|
|
|
|
|
|
def execute(self, repo, upstream, until_commit, run_cherrypick):
|
|
|
|
repo = LocalRepo(repo, upstream, self.default_branch_name)
|
|
|
|
all_branches = repo.get_release_branches() # [(branch_name, base_commit)]
|
|
|
|
|
|
|
|
release_branches = self.getBranchesWithRelease()
|
|
|
|
|
|
|
|
branches = []
|
|
|
|
# iterate over all branches to preserve their precedence.
|
|
|
|
for branch in all_branches:
|
|
|
|
if branch[0] in release_branches:
|
|
|
|
branches.append(branch)
|
|
|
|
|
|
|
|
if not branches:
|
|
|
|
logging.info('No release branches found!')
|
|
|
|
return
|
|
|
|
|
|
|
|
for branch in branches:
|
|
|
|
logging.info('Found release branch: %s', branch[0])
|
|
|
|
|
|
|
|
if not until_commit:
|
|
|
|
until_commit = branches[0][1]
|
|
|
|
pull_requests = self.getPullRequests(until_commit)
|
|
|
|
|
|
|
|
backport_map = {}
|
|
|
|
|
|
|
|
RE_MUST_BACKPORT = re.compile(r'^v(\d+\.\d+)-must-backport$')
|
|
|
|
RE_NO_BACKPORT = re.compile(r'^v(\d+\.\d+)-no-backport$')
|
|
|
|
RE_BACKPORTED = re.compile(r'^v(\d+\.\d+)-backported$')
|
|
|
|
|
|
|
|
# pull-requests are sorted by ancestry from the most recent.
|
|
|
|
for pr in pull_requests:
|
|
|
|
while repo.comparator(branches[-1][1]) >= repo.comparator(pr['mergeCommit']['oid']):
|
|
|
|
logging.info("PR #{} is already inside {}. Dropping this branch for further PRs".format(pr['number'], branches[-1][0]))
|
|
|
|
branches.pop()
|
|
|
|
|
|
|
|
logging.info("Processing PR #{}".format(pr['number']))
|
|
|
|
|
|
|
|
assert len(branches)
|
|
|
|
|
|
|
|
branch_set = set([branch[0] for branch in branches])
|
|
|
|
|
|
|
|
# First pass. Find all must-backports
|
|
|
|
for label in pr['labels']['nodes']:
|
2021-12-13 17:40:06 +00:00
|
|
|
if label['name'] == 'pr-must-backport':
|
2021-11-08 14:30:27 +00:00
|
|
|
backport_map[pr['number']] = branch_set.copy()
|
|
|
|
continue
|
|
|
|
matched = RE_MUST_BACKPORT.match(label['name'])
|
|
|
|
if matched:
|
|
|
|
if pr['number'] not in backport_map:
|
|
|
|
backport_map[pr['number']] = set()
|
|
|
|
backport_map[pr['number']].add(matched.group(1))
|
|
|
|
|
|
|
|
# Second pass. Find all no-backports
|
|
|
|
for label in pr['labels']['nodes']:
|
|
|
|
if label['name'] == 'pr-no-backport' and pr['number'] in backport_map:
|
|
|
|
del backport_map[pr['number']]
|
|
|
|
break
|
|
|
|
matched_no_backport = RE_NO_BACKPORT.match(label['name'])
|
|
|
|
matched_backported = RE_BACKPORTED.match(label['name'])
|
|
|
|
if matched_no_backport and pr['number'] in backport_map and matched_no_backport.group(1) in backport_map[pr['number']]:
|
|
|
|
backport_map[pr['number']].remove(matched_no_backport.group(1))
|
|
|
|
logging.info('\tskipping %s because of forced no-backport', matched_no_backport.group(1))
|
|
|
|
elif matched_backported and pr['number'] in backport_map and matched_backported.group(1) in backport_map[pr['number']]:
|
|
|
|
backport_map[pr['number']].remove(matched_backported.group(1))
|
|
|
|
logging.info('\tskipping %s because it\'s already backported manually', matched_backported.group(1))
|
|
|
|
|
|
|
|
for pr, branches in list(backport_map.items()):
|
|
|
|
logging.info('PR #%s needs to be backported to:', pr)
|
|
|
|
for branch in branches:
|
|
|
|
logging.info('\t%s, and the status is: %s', branch, run_cherrypick(self._token, pr, branch))
|
|
|
|
|
|
|
|
# print API costs
|
|
|
|
logging.info('\nGitHub API total costs per query:')
|
|
|
|
for name, value in list(self._gh.api_costs.items()):
|
|
|
|
logging.info('%s : %s', name, value)
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == "__main__":
|
|
|
|
parser = argparse.ArgumentParser()
|
|
|
|
parser.add_argument('--token', type=str, required=True, help='token for Github access')
|
|
|
|
parser.add_argument('--repo', type=str, required=True, help='path to full repository', metavar='PATH')
|
|
|
|
parser.add_argument('--til', type=str, help='check PRs from HEAD til this commit', metavar='COMMIT')
|
|
|
|
parser.add_argument('--dry-run', action='store_true', help='do not create or merge any PRs', default=False)
|
|
|
|
parser.add_argument('--verbose', '-v', action='store_true', help='more verbose output', default=False)
|
|
|
|
parser.add_argument('--upstream', '-u', type=str, help='remote name of upstream in repository', default='origin')
|
|
|
|
args = parser.parse_args()
|
|
|
|
|
|
|
|
if args.verbose:
|
|
|
|
logging.basicConfig(format='%(message)s', stream=sys.stdout, level=logging.DEBUG)
|
|
|
|
else:
|
|
|
|
logging.basicConfig(format='%(message)s', stream=sys.stdout, level=logging.INFO)
|
|
|
|
|
|
|
|
cherrypick_run = lambda token, pr, branch: CherryPick(token, 'ClickHouse', 'ClickHouse', 'core', pr, branch).execute(args.repo, args.dry_run)
|
|
|
|
bp = Backport(args.token, 'ClickHouse', 'ClickHouse', 'core')
|
|
|
|
bp.execute(args.repo, args.upstream, args.til, cherrypick_run)
|