ClickHouse/utils/github/query.py

528 lines
19 KiB
Python
Raw Normal View History

# -*- coding: utf-8 -*-
import requests
class Query:
2019-04-11 15:08:25 +00:00
'''
Implements queries to the Github API using GraphQL
'''
_PULL_REQUEST = '''
author {{
... on User {{
id
login
}}
}}
baseRepository {{
nameWithOwner
}}
mergeCommit {{
oid
parents(first: {min_page_size}) {{
totalCount
nodes {{
oid
}}
}}
}}
mergedBy {{
... on User {{
id
login
}}
}}
baseRefName
closed
id
mergeable
merged
number
title
url
'''
def __init__(self, token, owner, name, team, max_page_size=100, min_page_size=5):
self._PULL_REQUEST = Query._PULL_REQUEST.format(min_page_size=min_page_size)
2019-04-11 15:08:25 +00:00
self._token = token
self._owner = owner
self._name = name
self._team = team
self._max_page_size = max_page_size
self._min_page_size = min_page_size
2019-09-06 12:42:15 +00:00
self.api_costs = {}
repo = self.get_repository()
self._id = repo['id']
self.ssh_url = repo['sshUrl']
self.default_branch = repo['defaultBranchRef']['name']
self.members = set(self.get_members())
def get_repository(self):
_QUERY = '''
repository(owner: "{owner}" name: "{name}") {{
defaultBranchRef {{
name
}}
id
sshUrl
}}
'''
query = _QUERY.format(owner=self._owner, name=self._name)
return self._run(query)['repository']
def get_members(self):
'''Get all team members for organization
Returns:
members: a map of members' logins to ids
'''
_QUERY = '''
organization(login: "{organization}") {{
team(slug: "{team}") {{
members(first: {max_page_size} {next}) {{
pageInfo {{
hasNextPage
endCursor
}}
nodes {{
id
login
}}
}}
}}
}}
'''
members = {}
not_end = True
query = _QUERY.format(organization=self._owner, team=self._team,
max_page_size=self._max_page_size,
next='')
while not_end:
result = self._run(query)['organization']['team']
if result is None:
break
result = result['members']
not_end = result['pageInfo']['hasNextPage']
query = _QUERY.format(organization=self._owner, team=self._team,
max_page_size=self._max_page_size,
next='after: "{}"'.format(result["pageInfo"]["endCursor"]))
members += dict([(node['login'], node['id']) for node in result['nodes']])
return members
def get_pull_request(self, number):
_QUERY = '''
repository(owner: "{owner}" name: "{name}") {{
pullRequest(number: {number}) {{
{pull_request_data}
2019-04-11 15:08:25 +00:00
}}
}}
'''
2019-04-08 19:28:08 +00:00
query = _QUERY.format(owner=self._owner, name=self._name, number=number,
2020-07-22 01:05:46 +00:00
pull_request_data=self._PULL_REQUEST, min_page_size=self._min_page_size)
return self._run(query)['repository']['pullRequest']
def find_pull_request(self, base, head):
_QUERY = '''
repository(owner: "{owner}" name: "{name}") {{
pullRequests(first: {min_page_size} baseRefName: "{base}" headRefName: "{head}") {{
nodes {{
{pull_request_data}
}}
totalCount
}}
}}
'''
query = _QUERY.format(owner=self._owner, name=self._name, base=base, head=head,
2020-07-22 01:05:46 +00:00
pull_request_data=self._PULL_REQUEST, min_page_size=self._min_page_size)
result = self._run(query)['repository']['pullRequests']
if result['totalCount'] > 0:
return result['nodes'][0]
else:
return {}
def get_pull_requests(self, before_commit):
'''
Get all merged pull-requests from the HEAD of default branch to the last commit (excluding)
'''
_QUERY = '''
repository(owner: "{owner}" name: "{name}") {{
defaultBranchRef {{
target {{
... on Commit {{
history(first: {max_page_size} {next}) {{
pageInfo {{
hasNextPage
endCursor
}}
nodes {{
oid
associatedPullRequests(first: {min_page_size}) {{
totalCount
nodes {{
... on PullRequest {{
{pull_request_data}
labels(first: {min_page_size}) {{
totalCount
pageInfo {{
hasNextPage
endCursor
}}
nodes {{
name
color
}}
}}
}}
2019-04-11 15:08:25 +00:00
}}
}}
}}
}}
}}
}}
}}
}}
'''
2019-04-11 15:08:25 +00:00
pull_requests = []
not_end = True
query = _QUERY.format(owner=self._owner, name=self._name,
max_page_size=self._max_page_size,
min_page_size=self._min_page_size,
pull_request_data=self._PULL_REQUEST,
next='')
2019-04-11 15:08:25 +00:00
while not_end:
result = self._run(query)['repository']['defaultBranchRef']['target']['history']
2019-04-11 15:08:25 +00:00
not_end = result['pageInfo']['hasNextPage']
query = _QUERY.format(owner=self._owner, name=self._name,
max_page_size=self._max_page_size,
min_page_size=self._min_page_size,
pull_request_data=self._PULL_REQUEST,
next='after: "{}"'.format(result["pageInfo"]["endCursor"]))
2019-04-11 15:08:25 +00:00
for commit in result['nodes']:
# FIXME: maybe include `before_commit`?
2019-04-11 15:08:25 +00:00
if str(commit['oid']) == str(before_commit):
not_end = False
break
2019-04-11 16:16:57 +00:00
# TODO: fetch all pull-requests that were merged in a single commit.
assert commit['associatedPullRequests']['totalCount'] <= self._min_page_size
2019-04-11 15:08:25 +00:00
for pull_request in commit['associatedPullRequests']['nodes']:
if(pull_request['baseRepository']['nameWithOwner'] == '{}/{}'.format(self._owner, self._name) and
pull_request['baseRefName'] == self.default_branch and
pull_request['mergeCommit']['oid'] == commit['oid']):
pull_requests.append(pull_request)
return pull_requests
def create_pull_request(self, source, target, title, description="", draft=False, can_modify=True):
_QUERY = '''
createPullRequest(input: {{
baseRefName: "{target}",
headRefName: "{source}",
repositoryId: "{id}",
title: "{title}",
body: "{body}",
draft: {draft},
maintainerCanModify: {modify}
}}) {{
pullRequest {{
{pull_request_data}
}}
}}
'''
query = _QUERY.format(target=target, source=source, id=self._id, title=title, body=description,
draft="true" if draft else "false", modify="true" if can_modify else "false",
2020-07-22 01:05:46 +00:00
pull_request_data=self._PULL_REQUEST)
return self._run(query, is_mutation=True)['createPullRequest']['pullRequest']
def merge_pull_request(self, id):
_QUERY = '''
mergePullRequest(input: {{
pullRequestId: "{id}"
}}) {{
pullRequest {{
{pull_request_data}
}}
}}
'''
2020-07-22 01:05:46 +00:00
query = _QUERY.format(id=id, pull_request_data=self._PULL_REQUEST)
return self._run(query, is_mutation=True)['mergePullRequest']['pullRequest']
# FIXME: figure out how to add more assignees at once
def add_assignee(self, pr, assignee):
_QUERY = '''
addAssigneesToAssignable(input: {{
assignableId: "{id1}",
assigneeIds: "{id2}"
}}) {{
clientMutationId
}}
'''
query = _QUERY.format(id1=pr['id'], id2=assignee['id'])
self._run(query, is_mutation=True)
def set_label(self, pull_request, label_name):
'''
Set label by name to the pull request
Args:
pull_request: JSON object returned by `get_pull_requests()`
label_name (string): label name
'''
_GET_LABEL = '''
repository(owner: "{owner}" name: "{name}") {{
labels(first: {max_page_size} {next} query: "{label_name}") {{
pageInfo {{
hasNextPage
endCursor
}}
nodes {{
id
name
color
}}
}}
}}
'''
_SET_LABEL = '''
addLabelsToLabelable(input: {{
labelableId: "{pr_id}",
labelIds: "{label_id}"
}}) {{
clientMutationId
}}
'''
labels = []
not_end = True
query = _GET_LABEL.format(owner=self._owner, name=self._name, label_name=label_name,
max_page_size=self._max_page_size,
next='')
while not_end:
result = self._run(query)['repository']['labels']
not_end = result['pageInfo']['hasNextPage']
query = _GET_LABEL.format(owner=self._owner, name=self._name, label_name=label_name,
max_page_size=self._max_page_size,
next='after: "{}"'.format(result["pageInfo"]["endCursor"]))
labels += [label for label in result['nodes']]
if not labels:
return
2020-07-22 01:05:46 +00:00
query = _SET_LABEL.format(pr_id=pull_request['id'], label_id=labels[0]['id'])
self._run(query, is_mutation=True)
2020-07-22 01:05:46 +00:00
# OLD METHODS
# _LABELS = '''
# repository(owner: "ClickHouse" name: "ClickHouse") {{
# pullRequest(number: {number}) {{
# labels(first: {max_page_size} {next}) {{
# pageInfo {{
# hasNextPage
# endCursor
# }}
# nodes {{
# name
# color
# }}
# }}
# }}
# }}
# '''
# def get_labels(self, pull_request):
# '''Fetchs all labels for given pull-request
# Args:
# pull_request: JSON object returned by `get_pull_requests()`
# Returns:
# labels: a list of JSON nodes with the name and color fields
# '''
# labels = [label for label in pull_request['labels']['nodes']]
# not_end = pull_request['labels']['pageInfo']['hasNextPage']
# query = Query._LABELS.format(number = pull_request['number'],
# max_page_size = self._max_page_size,
# next=f'after: "{pull_request["labels"]["pageInfo"]["endCursor"]}"')
# while not_end:
# result = self._run(query)['repository']['pullRequest']['labels']
# not_end = result['pageInfo']['hasNextPage']
# query = Query._LABELS.format(number=pull_request['number'],
# max_page_size=self._max_page_size,
# next=f'after: "{result["pageInfo"]["endCursor"]}"')
# labels += [label for label in result['nodes']]
# return labels
# _TIMELINE = '''
# repository(owner: "ClickHouse" name: "ClickHouse") {{
# pullRequest(number: {number}) {{
# timeline(first: {max_page_size} {next}) {{
# pageInfo {{
# hasNextPage
# endCursor
# }}
# nodes {{
# ... on CrossReferencedEvent {{
# isCrossRepository
# source {{
# ... on PullRequest {{
# number
# baseRefName
# merged
# labels(first: {max_page_size}) {{
# pageInfo {{
# hasNextPage
# endCursor
# }}
# nodes {{
# name
# color
# }}
# }}
# }}
# }}
# target {{
# ... on PullRequest {{
# number
# }}
# }}
# }}
# }}
# }}
# }}
# }}
# '''
# def get_timeline(self, pull_request):
# '''Fetchs all cross-reference events from pull-request's timeline
# Args:
# pull_request: JSON object returned by `get_pull_requests()`
# Returns:
# events: a list of JSON nodes for CrossReferenceEvent
# '''
# events = [event for event in pull_request['timeline']['nodes'] if event and event['source']]
# not_end = pull_request['timeline']['pageInfo']['hasNextPage']
# query = Query._TIMELINE.format(number = pull_request['number'],
# max_page_size = self._max_page_size,
# next=f'after: "{pull_request["timeline"]["pageInfo"]["endCursor"]}"')
# while not_end:
# result = self._run(query)['repository']['pullRequest']['timeline']
# not_end = result['pageInfo']['hasNextPage']
# query = Query._TIMELINE.format(number=pull_request['number'],
# max_page_size=self._max_page_size,
# next=f'after: "{result["pageInfo"]["endCursor"]}"')
# events += [event for event in result['nodes'] if event and event['source']]
# return events
# _DEFAULT = '''
# repository(owner: "ClickHouse", name: "ClickHouse") {
# defaultBranchRef {
# name
# }
# }
# '''
# def get_default_branch(self):
# '''Get short name of the default branch
# Returns:
# name (string): branch name
# '''
# return self._run(Query._DEFAULT)['repository']['defaultBranchRef']['name']
def _run(self, query, is_mutation=False):
2019-04-11 15:08:25 +00:00
from requests.adapters import HTTPAdapter
from urllib3.util.retry import Retry
def requests_retry_session(
retries=3,
backoff_factor=0.3,
status_forcelist=(500, 502, 504),
session=None,
):
session = session or requests.Session()
retry = Retry(
total=retries,
read=retries,
connect=retries,
backoff_factor=backoff_factor,
status_forcelist=status_forcelist,
)
adapter = HTTPAdapter(max_retries=retry)
session.mount('http://', adapter)
session.mount('https://', adapter)
return session
headers = {'Authorization': 'bearer {}'.format(self._token)}
if is_mutation:
query = '''
mutation {{
{query}
2019-09-06 12:42:15 +00:00
}}
'''.format(query=query)
else:
query = '''
query {{
{query}
rateLimit {{
cost
remaining
}}
}}
'''.format(query=query)
2020-03-16 15:27:07 +00:00
while True:
request = requests_retry_session().post('https://api.github.com/graphql', json={'query': query}, headers=headers)
if request.status_code == 200:
result = request.json()
if 'errors' in result:
raise Exception('Errors occured: {}\nOriginal query: {}'.format(result["errors"], query))
2020-03-16 15:27:07 +00:00
if not is_mutation:
import inspect
caller = inspect.getouterframes(inspect.currentframe(), 2)[1][3]
if caller not in self.api_costs.keys():
self.api_costs[caller] = 0
self.api_costs[caller] += result['data']['rateLimit']['cost']
return result['data']
else:
import json
raise Exception('Query failed with code {code}:\n{json}'.format(code=request.status_code, json=json.dumps(request.json(), indent=4)))