This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/spark-kubernetes-operator.git
The following commit(s) were added to refs/heads/main by this push: new c1e4879 Add merge_spark_pr.py c1e4879 is described below commit c1e48798a839bb0f77783cd68c88cab23c3276f3 Author: Dongjoon Hyun <dongj...@apache.org> AuthorDate: Fri Mar 29 14:53:31 2024 -0700 Add merge_spark_pr.py --- dev/merge_spark_pr.py | 717 ++++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 717 insertions(+) diff --git a/dev/merge_spark_pr.py b/dev/merge_spark_pr.py new file mode 100755 index 0000000..459f67b --- /dev/null +++ b/dev/merge_spark_pr.py @@ -0,0 +1,717 @@ +#!/usr/bin/env python3 + +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# + +# Utility for creating well-formed pull request merges and pushing them to Apache +# Spark. +# usage: ./merge_spark_pr.py (see config env vars below) +# +# This utility assumes you already have a local Spark git folder and that you +# have added remotes corresponding to both (i) the github apache Spark +# mirror and (ii) the apache git repo. + +import json +import os +import re +import subprocess +import sys +import traceback +from urllib.request import urlopen +from urllib.request import Request +from urllib.error import HTTPError + +try: + import jira.client + + JIRA_IMPORTED = True +except ImportError: + JIRA_IMPORTED = False + +# Location of your Spark git development area +SPARK_HOME = os.environ.get("SPARK_HOME", os.getcwd()) +# Remote name which points to the Github site +PR_REMOTE_NAME = os.environ.get("PR_REMOTE_NAME", "apache-github") +# Remote name which points to Apache git +PUSH_REMOTE_NAME = os.environ.get("PUSH_REMOTE_NAME", "apache") +# ASF JIRA username +JIRA_USERNAME = os.environ.get("JIRA_USERNAME", "") +# ASF JIRA password +JIRA_PASSWORD = os.environ.get("JIRA_PASSWORD", "") +# ASF JIRA access token +# If it is configured, username and password are dismissed +# Go to https://issues.apache.org/jira/secure/ViewProfile.jspa -> Personal Access Tokens for +# your own token management. +JIRA_ACCESS_TOKEN = os.environ.get("JIRA_ACCESS_TOKEN") +# OAuth key used for issuing requests against the GitHub API. If this is not defined, then requests +# will be unauthenticated. You should only need to configure this if you find yourself regularly +# exceeding your IP's unauthenticated request rate limit. You can create an OAuth key at +# https://github.com/settings/tokens. This script only requires the "public_repo" scope. +GITHUB_OAUTH_KEY = os.environ.get("GITHUB_OAUTH_KEY") + + +GITHUB_BASE = "https://github.com/apache/spark-kubernetes-operator/pull" +GITHUB_API_BASE = "https://api.github.com/repos/spark-kubernetes-operator/spark" +JIRA_BASE = "https://issues.apache.org/jira/browse" +JIRA_API_BASE = "https://issues.apache.org/jira" +# Prefix added to temporary branches +BRANCH_PREFIX = "PR_TOOL" + + +def print_error(msg): + print("\033[91m%s\033[0m" % msg) + + +def bold_input(prompt) -> str: + return input("\033[1m%s\033[0m" % prompt) + + +def get_json(url): + try: + request = Request(url) + if GITHUB_OAUTH_KEY: + request.add_header("Authorization", "token %s" % GITHUB_OAUTH_KEY) + return json.load(urlopen(request)) + except HTTPError as e: + if "X-RateLimit-Remaining" in e.headers and e.headers["X-RateLimit-Remaining"] == "0": + print_error( + "Exceeded the GitHub API rate limit; see the instructions in " + + "dev/merge_spark_pr.py to configure an OAuth token for making authenticated " + + "GitHub requests." + ) + elif e.code == 401: + print_error( + "GITHUB_OAUTH_KEY is invalid or expired. Please regenerate a new one with " + + "at least the 'public_repo' scope on https://github.com/settings/tokens and " + + "update your local settings before you try again." + ) + else: + print_error("Unable to fetch URL, exiting: %s" % url) + sys.exit(-1) + + +def fail(msg): + print_error(msg) + clean_up() + sys.exit(-1) + + +def run_cmd(cmd): + print(cmd) + if isinstance(cmd, list): + return subprocess.check_output(cmd).decode("utf-8") + else: + return subprocess.check_output(cmd.split(" ")).decode("utf-8") + + +def continue_maybe(prompt): + result = bold_input("%s (y/N): " % prompt) + if result.lower() != "y": + fail("Okay, exiting") + + +def clean_up(): + if "original_head" in globals(): + print("Restoring head pointer to %s" % original_head) + run_cmd("git checkout %s" % original_head) + + branches = run_cmd("git branch").replace(" ", "").split("\n") + + for branch in list(filter(lambda x: x.startswith(BRANCH_PREFIX), branches)): + print("Deleting local branch %s" % branch) + run_cmd("git branch -D %s" % branch) + + +# merge the requested PR and return the merge hash +def merge_pr(pr_num, target_ref, title, body, pr_repo_desc): + pr_branch_name = "%s_MERGE_PR_%s" % (BRANCH_PREFIX, pr_num) + target_branch_name = "%s_MERGE_PR_%s_%s" % (BRANCH_PREFIX, pr_num, target_ref.upper()) + run_cmd("git fetch %s pull/%s/head:%s" % (PR_REMOTE_NAME, pr_num, pr_branch_name)) + run_cmd("git fetch %s %s:%s" % (PUSH_REMOTE_NAME, target_ref, target_branch_name)) + run_cmd("git checkout %s" % target_branch_name) + + had_conflicts = False + try: + run_cmd(["git", "merge", pr_branch_name, "--squash"]) + except Exception as e: + msg = "Error merging: %s\nWould you like to manually fix-up this merge?" % e + continue_maybe(msg) + msg = "Okay, please fix any conflicts and 'git add' conflicting files... Finished?" + continue_maybe(msg) + had_conflicts = True + + # First commit author should be considered as the primary author when the rank is the same + commit_authors = run_cmd( + ["git", "log", "HEAD..%s" % pr_branch_name, "--pretty=format:%an <%ae>", "--reverse"] + ).split("\n") + distinct_authors = sorted( + list(dict.fromkeys(commit_authors)), key=lambda x: commit_authors.count(x), reverse=True + ) + primary_author = bold_input( + 'Enter primary author in the format of "name <email>" [%s]: ' % distinct_authors[0] + ) + if primary_author == "": + primary_author = distinct_authors[0] + else: + # When primary author is specified manually, de-dup it from author list and + # put it at the head of author list. + distinct_authors = list(filter(lambda x: x != primary_author, distinct_authors)) + distinct_authors.insert(0, primary_author) + + merge_message_flags = [] + + merge_message_flags += ["-m", title] + if body is not None: + # We remove @ symbols from the body to avoid triggering e-mails + # to people every time someone creates a public fork of Spark. + merge_message_flags += ["-m", body.replace("@", "")] + + committer_name = run_cmd("git config --get user.name").strip() + committer_email = run_cmd("git config --get user.email").strip() + + if had_conflicts: + message = "This patch had conflicts when merged, resolved by\nCommitter: %s <%s>" % ( + committer_name, + committer_email, + ) + merge_message_flags += ["-m", message] + + # The string "Closes #%s" string is required for GitHub to correctly close the PR + merge_message_flags += ["-m", "Closes #%s from %s." % (pr_num, pr_repo_desc)] + + authors = "Authored-by:" if len(distinct_authors) == 1 else "Lead-authored-by:" + authors += " %s" % (distinct_authors.pop(0)) + if len(distinct_authors) > 0: + authors += "\n" + "\n".join(["Co-authored-by: %s" % a for a in distinct_authors]) + authors += "\n" + "Signed-off-by: %s <%s>" % (committer_name, committer_email) + + merge_message_flags += ["-m", authors] + + run_cmd(["git", "commit", '--author="%s"' % primary_author] + merge_message_flags) + + continue_maybe( + "Merge complete (local ref %s). Push to %s?" % (target_branch_name, PUSH_REMOTE_NAME) + ) + + try: + run_cmd("git push %s %s:%s" % (PUSH_REMOTE_NAME, target_branch_name, target_ref)) + except Exception as e: + clean_up() + print_error("Exception while pushing: %s" % e) + + merge_hash = run_cmd("git rev-parse %s" % target_branch_name)[:8] + clean_up() + print("Pull request #%s merged!" % pr_num) + print("Merge hash: %s" % merge_hash) + return merge_hash + + +def cherry_pick(pr_num, merge_hash, default_branch): + pick_ref = bold_input("Enter a branch name [%s]: " % default_branch) + if pick_ref == "": + pick_ref = default_branch + + pick_branch_name = "%s_PICK_PR_%s_%s" % (BRANCH_PREFIX, pr_num, pick_ref.upper()) + + run_cmd("git fetch %s %s:%s" % (PUSH_REMOTE_NAME, pick_ref, pick_branch_name)) + run_cmd("git checkout %s" % pick_branch_name) + + try: + run_cmd("git cherry-pick -sx %s" % merge_hash) + except Exception as e: + msg = "Error cherry-picking: %s\nWould you like to manually fix-up this merge?" % e + continue_maybe(msg) + msg = "Okay, please fix any conflicts and finish the cherry-pick. Finished?" + continue_maybe(msg) + + continue_maybe( + "Pick complete (local ref %s). Push to %s?" % (pick_branch_name, PUSH_REMOTE_NAME) + ) + + try: + run_cmd("git push %s %s:%s" % (PUSH_REMOTE_NAME, pick_branch_name, pick_ref)) + except Exception as e: + clean_up() + fail("Exception while pushing: %s" % e) + + pick_hash = run_cmd("git rev-parse %s" % pick_branch_name)[:8] + clean_up() + + print("Pull request #%s picked into %s!" % (pr_num, pick_ref)) + print("Pick hash: %s" % pick_hash) + return pick_ref + + +def print_jira_issue_summary(issue): + summary = issue.fields.summary + assignee = issue.fields.assignee + if assignee is not None: + assignee = assignee.displayName + status = issue.fields.status.name + print("=== JIRA %s ===" % issue.key) + print( + "summary\t\t%s\nassignee\t%s\nstatus\t\t%s\nurl\t\t%s/%s\n" + % (summary, assignee, status, JIRA_BASE, issue.key) + ) + + +def get_jira_issue(prompt, default_jira_id=""): + jira_id = bold_input("%s [%s]: " % (prompt, default_jira_id)) + if jira_id == "": + jira_id = default_jira_id + if jira_id == "": + print("JIRA ID not found, skipping.") + return None + try: + issue = asf_jira.issue(jira_id) + print_jira_issue_summary(issue) + status = issue.fields.status.name + if status == "Resolved" or status == "Closed": + print("JIRA issue %s already has status '%s'" % (jira_id, status)) + return None + if bold_input("Check if the JIRA information is as expected (y/N): ").lower() == "y": + return issue + else: + return get_jira_issue("Enter the revised JIRA ID again or leave blank to skip") + except Exception as e: + print_error("ASF JIRA could not find %s: %s" % (jira_id, e)) + return get_jira_issue("Enter the revised JIRA ID again or leave blank to skip") + + +def resolve_jira_issue(merge_branches, comment, default_jira_id=""): + issue = get_jira_issue("Enter a JIRA id", default_jira_id) + if issue is None: + return + + if issue.fields.assignee is None: + choose_jira_assignee(issue) + + versions = asf_jira.project_versions("SPARK") + # Consider only x.y.z, unreleased, unarchived versions + versions = [ + x + for x in versions + if not x.raw["released"] and not x.raw["archived"] and re.match(r"\d+\.\d+\.\d+", x.name) + ] + versions = sorted(versions, key=lambda x: x.name, reverse=True) + + default_fix_versions = [] + for b in merge_branches: + if b == "master": + default_fix_versions.append(versions[0].name) + else: + found = False + found_versions = [] + for v in versions: + if v.name.startswith(b.replace("branch-", "")): + found_versions.append(v.name) + found = True + if found: + # There might be several unreleased versions for specific branches + # For example, assuming + # versions = ['4.0.0', '3.5.1', '3.5.0', '3.4.2', '3.3.4', '3.3.3'] + # we've found two candidates for branch-3.5, we pick the last/smallest one + default_fix_versions.append(found_versions[-1]) + else: + print_error( + "Target version for %s is not found on JIRA, it may be archived or " + "not created. Skipping it." % b + ) + + for v in default_fix_versions: + # Handles the case where we have forked a release branch but not yet made the release. + # In this case, if the PR is committed to the master branch and the release branch, we + # only consider the release branch to be the fix version. E.g. it is not valid to have + # both 1.1.0 and 1.0.0 as fix versions. + (major, minor, patch) = v.split(".") + if patch == "0": + previous = "%s.%s.%s" % (major, int(minor) - 1, 0) + if previous in default_fix_versions: + default_fix_versions = list(filter(lambda x: x != v, default_fix_versions)) + default_fix_versions = ",".join(default_fix_versions) + + available_versions = set(list(map(lambda v: v.name, versions))) + while True: + try: + fix_versions = bold_input( + "Enter comma-separated fix version(s) [%s]: " % default_fix_versions + ) + if fix_versions == "": + fix_versions = default_fix_versions + fix_versions = fix_versions.replace(" ", "").split(",") + if set(fix_versions).issubset(available_versions): + break + else: + print( + "Specified version(s) [%s] not found in the available versions, try " + "again (or leave blank and fix manually)." % (", ".join(fix_versions)) + ) + except KeyboardInterrupt: + raise + except BaseException: + traceback.print_exc() + print("Error setting fix version(s), try again (or leave blank and fix manually)") + + def get_version_json(version_str): + return list(filter(lambda v: v.name == version_str, versions))[0].raw + + jira_fix_versions = list(map(lambda v: get_version_json(v), fix_versions)) + + resolve = list(filter(lambda a: a["name"] == "Resolve Issue", asf_jira.transitions(issue.key)))[ + 0 + ] + resolution = list(filter(lambda r: r.raw["name"] == "Fixed", asf_jira.resolutions()))[0] + asf_jira.transition_issue( + issue.key, + resolve["id"], + fixVersions=jira_fix_versions, + comment=comment, + resolution={"id": resolution.raw["id"]}, + ) + + try: + print_jira_issue_summary(asf_jira.issue(issue.key)) + except Exception: + print("Unable to fetch JIRA issue %s after resolving" % issue.key) + print("Successfully resolved %s with fixVersions=%s!" % (issue.key, fix_versions)) + + +def choose_jira_assignee(issue): + """ + Prompt the user to choose who to assign the issue to in jira, given a list of candidates, + including the original reporter and all commentators + """ + while True: + try: + reporter = issue.fields.reporter + commentators = list(map(lambda x: x.author, issue.fields.comment.comments)) + candidates = set(commentators) + candidates.add(reporter) + candidates = list(candidates) + print("JIRA is unassigned, choose assignee") + for idx, author in enumerate(candidates): + if author.key == "apachespark": + continue + annotations = ["Reporter"] if author == reporter else [] + if author in commentators: + annotations.append("Commentator") + print("[%d] %s (%s)" % (idx, author.displayName, ",".join(annotations))) + raw_assignee = bold_input( + "Enter number of user, or userid, to assign to (blank to leave unassigned):" + ) + if raw_assignee == "": + return None + else: + try: + id = int(raw_assignee) + assignee = candidates[id] + except BaseException: + # assume it's a user id, and try to assign (might fail, we just prompt again) + assignee = asf_jira.user(raw_assignee) + try: + assign_issue(issue.key, assignee.name) + except Exception as e: + if ( + e.__class__.__name__ == "JIRAError" + and ("'%s' cannot be assigned" % assignee.name) + in getattr(e, "response").text + ): + continue_maybe( + "User '%s' cannot be assigned, add to contributors role and try again?" + % assignee.name + ) + grant_contributor_role(assignee.name) + assign_issue(issue.key, assignee.name) + else: + raise e + return assignee + except KeyboardInterrupt: + raise + except BaseException: + traceback.print_exc() + print("Error assigning JIRA, try again (or leave blank and fix manually)") + + +def grant_contributor_role(user: str): + role = asf_jira.project_role("SPARK", 10010) + role.add_user(user) + print("Successfully added user '%s' to contributors role" % user) + + +def assign_issue(issue: int, assignee: str) -> bool: + """ + Assign an issue to a user, which is a shorthand for jira.client.JIRA.assign_issue. + The original one has an issue that it will search users again and only choose the assignee + from 20 candidates. If it's unmatched, it picks the head blindly. In our case, the assignee + is already resolved. + """ + url = getattr(asf_jira, "_get_latest_url")(f"issue/{issue}/assignee") + payload = {"name": assignee} + getattr(asf_jira, "_session").put(url, data=json.dumps(payload)) + return True + + +def resolve_jira_issues(title, merge_branches, comment): + jira_ids = re.findall("SPARK-[0-9]{4,5}", title) + + if len(jira_ids) == 0: + resolve_jira_issue(merge_branches, comment) + for jira_id in jira_ids: + resolve_jira_issue(merge_branches, comment, jira_id) + + +def standardize_jira_ref(text): + """ + Standardize the [SPARK-XXXXX] [MODULE] prefix + Converts "[SPARK-XXX][mllib] Issue", "[MLLib] SPARK-XXX. Issue" or "SPARK XXX [MLLIB]: Issue" to + "[SPARK-XXX][MLLIB] Issue" + + >>> standardize_jira_ref( + ... "[SPARK-5821] [SQL] ParquetRelation2 CTAS should check if delete is successful") + '[SPARK-5821][SQL] ParquetRelation2 CTAS should check if delete is successful' + >>> standardize_jira_ref( + ... "[SPARK-4123][Project Infra][WIP]: Show new dependencies added in pull requests") + '[SPARK-4123][PROJECT INFRA][WIP] Show new dependencies added in pull requests' + >>> standardize_jira_ref("[MLlib] Spark 5954: Top by key") + '[SPARK-5954][MLLIB] Top by key' + >>> standardize_jira_ref("[SPARK-979] a LRU scheduler for load balancing in TaskSchedulerImpl") + '[SPARK-979] a LRU scheduler for load balancing in TaskSchedulerImpl' + >>> standardize_jira_ref( + ... "SPARK-1094 Support MiMa for reporting binary compatibility across versions.") + '[SPARK-1094] Support MiMa for reporting binary compatibility across versions.' + >>> standardize_jira_ref("[WIP] [SPARK-1146] Vagrant support for Spark") + '[SPARK-1146][WIP] Vagrant support for Spark' + >>> standardize_jira_ref( + ... "SPARK-1032. If Yarn app fails before registering, app master stays aroun...") + '[SPARK-1032] If Yarn app fails before registering, app master stays aroun...' + >>> standardize_jira_ref( + ... "[SPARK-6250][SPARK-6146][SPARK-5911][SQL] Types are now reserved words in DDL parser.") + '[SPARK-6250][SPARK-6146][SPARK-5911][SQL] Types are now reserved words in DDL parser.' + >>> standardize_jira_ref("Additional information for users building from source code") + 'Additional information for users building from source code' + """ + jira_refs = [] + components = [] + + # If the string is compliant, no need to process any further + if re.search(r"^\[SPARK-[0-9]{3,6}\](\[[A-Z0-9_\s,]+\] )+\S+", text): + return text + + # Extract JIRA ref(s): + pattern = re.compile(r"(SPARK[-\s]*[0-9]{3,6})+", re.IGNORECASE) + for ref in pattern.findall(text): + # Add brackets, replace spaces with a dash, & convert to uppercase + jira_refs.append("[" + re.sub(r"\s+", "-", ref.upper()) + "]") + text = text.replace(ref, "") + + # Extract spark component(s): + # Look for alphanumeric chars, spaces, dashes, periods, and/or commas + pattern = re.compile(r"(\[[\w\s,.-]+\])", re.IGNORECASE) + for component in pattern.findall(text): + components.append(component.upper()) + text = text.replace(component, "") + + # Cleanup any remaining symbols: + pattern = re.compile(r"^\W+(.*)", re.IGNORECASE) + if pattern.search(text) is not None: + text = pattern.search(text).groups()[0] + + # Assemble full text (JIRA ref(s), module(s), remaining text) + clean_text = "".join(jira_refs).strip() + "".join(components).strip() + " " + text.strip() + + # Replace multiple spaces with a single space, e.g. if no jira refs and/or components were + # included + clean_text = re.sub(r"\s+", " ", clean_text.strip()) + + return clean_text + + +def get_current_ref(): + ref = run_cmd("git rev-parse --abbrev-ref HEAD").strip() + if ref == "HEAD": + # The current ref is a detached HEAD, so grab its SHA. + return run_cmd("git rev-parse HEAD").strip() + else: + return ref + + +def initialize_jira(): + global asf_jira + jira_server = {"server": JIRA_API_BASE} + + if not JIRA_IMPORTED: + print_error("ERROR finding jira library. Run 'pip3 install jira' to install.") + continue_maybe("Continue without jira?") + elif JIRA_ACCESS_TOKEN: + client = jira.client.JIRA(jira_server, token_auth=JIRA_ACCESS_TOKEN) + try: + # Eagerly check if the token is valid to align with the behavior of username/password + # authn + client.current_user() + asf_jira = client + except Exception as e: + if e.__class__.__name__ == "JIRAError" and getattr(e, "status_code", None) == 401: + msg = ( + "ASF JIRA could not authenticate with the invalid or expired token '%s'" + % JIRA_ACCESS_TOKEN + ) + fail(msg) + else: + raise e + elif JIRA_USERNAME and JIRA_PASSWORD: + print("You can use JIRA_ACCESS_TOKEN instead of JIRA_USERNAME/JIRA_PASSWORD.") + print("Visit https://issues.apache.org/jira/secure/ViewProfile.jspa ") + print("and click 'Personal Access Tokens' menu to manage your own tokens.") + asf_jira = jira.client.JIRA(jira_server, basic_auth=(JIRA_USERNAME, JIRA_PASSWORD)) + else: + print("Neither JIRA_ACCESS_TOKEN nor JIRA_USERNAME/JIRA_PASSWORD are set.") + continue_maybe("Continue without jira?") + + +def main(): + initialize_jira() + global original_head + + os.chdir(SPARK_HOME) + original_head = get_current_ref() + + branches = get_json("%s/branches" % GITHUB_API_BASE) + branch_names = list(filter(lambda x: x.startswith("branch-"), [x["name"] for x in branches])) + # Assumes branch names can be sorted lexicographically + branch_names = sorted(branch_names, reverse=True) + branch_iter = iter(branch_names) + + pr_num = bold_input("Which pull request would you like to merge? (e.g. 34): ") + pr = get_json("%s/pulls/%s" % (GITHUB_API_BASE, pr_num)) + pr_events = get_json("%s/issues/%s/events" % (GITHUB_API_BASE, pr_num)) + + url = pr["url"] + + # Warn if the PR is WIP + if "[WIP]" in pr["title"]: + msg = "The PR title has `[WIP]`:\n%s\nContinue?" % pr["title"] + continue_maybe(msg) + + # Decide whether to use the modified title or not + modified_title = standardize_jira_ref(pr["title"]).rstrip(".") + if modified_title != pr["title"]: + print("I've re-written the title as follows to match the standard format:") + print("Original: %s" % pr["title"]) + print("Modified: %s" % modified_title) + result = bold_input("Would you like to use the modified title? (y/N): ") + if result.lower() == "y": + title = modified_title + print("Using modified title:") + else: + title = pr["title"] + print("Using original title:") + print(title) + else: + title = pr["title"] + + body = pr["body"] + if body is None: + body = "" + modified_body = re.sub(re.compile(r"<!--[^>]*-->\n?", re.DOTALL), "", body).lstrip() + if modified_body != body: + print("=" * 80) + print(modified_body) + print("=" * 80) + print("I've removed the comments from PR template like the above:") + result = bold_input("Would you like to use the modified body? (y/N): ") + if result.lower() == "y": + body = modified_body + print("Using modified body:") + else: + print("Using original body:") + print("=" * 80) + print(body) + print("=" * 80) + target_ref = pr["base"]["ref"] + user_login = pr["user"]["login"] + base_ref = pr["head"]["ref"] + pr_repo_desc = "%s/%s" % (user_login, base_ref) + + # Merged pull requests don't appear as merged in the GitHub API; + # Instead, they're closed by committers. + merge_commits = [e for e in pr_events if e["event"] == "closed" and e["commit_id"] is not None] + + if merge_commits and pr["state"] == "closed": + # A PR might have multiple merge commits, if it's reopened and merged again. We shall + # cherry-pick PRs in closed state with the latest merge hash. + # If the PR is still open(reopened), we shall not cherry-pick it but perform the normal + # merge as it could have been reverted earlier. + merge_commits = sorted(merge_commits, key=lambda x: x["created_at"]) + merge_hash = merge_commits[-1]["commit_id"] + message = get_json("%s/commits/%s" % (GITHUB_API_BASE, merge_hash))["commit"]["message"] + + print("Pull request %s has already been merged, assuming you want to backport" % pr_num) + commit_is_downloaded = ( + run_cmd(["git", "rev-parse", "--quiet", "--verify", "%s^{commit}" % merge_hash]).strip() + != "" + ) + if not commit_is_downloaded: + fail("Couldn't find any merge commit for #%s, you may need to update HEAD." % pr_num) + + print("Found commit %s:\n%s" % (merge_hash, message)) + cherry_pick(pr_num, merge_hash, next(branch_iter, branch_names[0])) + sys.exit(0) + + if not bool(pr["mergeable"]): + msg = ( + "Pull request %s is not mergeable in its current form.\n" % pr_num + + "Continue? (experts only!)" + ) + continue_maybe(msg) + + print("\n=== Pull Request #%s ===" % pr_num) + print("title\t%s\nsource\t%s\ntarget\t%s\nurl\t%s" % (title, pr_repo_desc, target_ref, url)) + continue_maybe("Proceed with merging pull request #%s?" % pr_num) + + merged_refs = [target_ref] + + merge_hash = merge_pr(pr_num, target_ref, title, body, pr_repo_desc) + + pick_prompt = "Would you like to pick %s into another branch?" % merge_hash + while bold_input("\n%s (y/N): " % pick_prompt).lower() == "y": + merged_refs = merged_refs + [ + cherry_pick(pr_num, merge_hash, next(branch_iter, branch_names[0])) + ] + + if asf_jira is not None: + continue_maybe("Would you like to update an associated JIRA?") + jira_comment = "Issue resolved by pull request %s\n[%s/%s]" % ( + pr_num, + GITHUB_BASE, + pr_num, + ) + resolve_jira_issues(title, merged_refs, jira_comment) + else: + print("Exiting without trying to close the associated JIRA.") + + +if __name__ == "__main__": + import doctest + + (failure_count, test_count) = doctest.testmod() + if failure_count: + sys.exit(-1) + try: + main() + except BaseException: + clean_up() + raise --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org