diff options
author | Lee Garrett <lgarrett@rocketjump.eu> | 2023-06-16 15:40:52 +0200 |
---|---|---|
committer | Lee Garrett <lgarrett@rocketjump.eu> | 2023-06-16 15:40:52 +0200 |
commit | 8d9a6d9cdf440b0a9b254a8a4bf063c0cb6a6201 (patch) | |
tree | e25b4160deb15b08aaf0aea65fc8c7bbc01dea12 /packaging | |
parent | 3cda7ad4dd15b514ff660905294b5b6330ecfb6f (diff) | |
download | debian-ansible-core-8d9a6d9cdf440b0a9b254a8a4bf063c0cb6a6201.zip |
New upstream version 2.14.6
Diffstat (limited to 'packaging')
-rw-r--r-- | packaging/pep517_backend/__init__.py | 1 | ||||
-rw-r--r-- | packaging/pep517_backend/_backend.py | 173 | ||||
-rw-r--r-- | packaging/pep517_backend/hooks.py | 9 | ||||
-rwxr-xr-x | packaging/release.py | 1469 |
4 files changed, 1652 insertions, 0 deletions
diff --git a/packaging/pep517_backend/__init__.py b/packaging/pep517_backend/__init__.py new file mode 100644 index 00000000..1d3bc14c --- /dev/null +++ b/packaging/pep517_backend/__init__.py @@ -0,0 +1 @@ +"""PEP 517 build backend for optionally pre-building docs before setuptools.""" diff --git a/packaging/pep517_backend/_backend.py b/packaging/pep517_backend/_backend.py new file mode 100644 index 00000000..ce97cfab --- /dev/null +++ b/packaging/pep517_backend/_backend.py @@ -0,0 +1,173 @@ +"""PEP 517 build backend wrapper for optionally pre-building docs for sdist.""" + +from __future__ import annotations + +import os +import re +import subprocess +import sys +import typing as t +from configparser import ConfigParser +from contextlib import contextmanager, suppress +from importlib import import_module +from io import StringIO +from pathlib import Path +from shutil import copytree +from tempfile import TemporaryDirectory + +try: + from contextlib import chdir as _chdir_cm +except ImportError: + @contextmanager + def _chdir_cm(path: os.PathLike) -> t.Iterator[None]: + original_wd = Path.cwd() + os.chdir(path) + try: + yield + finally: + os.chdir(original_wd) + +from setuptools.build_meta import ( + build_sdist as _setuptools_build_sdist, + get_requires_for_build_sdist as _setuptools_get_requires_for_build_sdist, +) + +with suppress(ImportError): + # NOTE: Only available for sdist builds that bundle manpages. Declared by + # NOTE: `get_requires_for_build_sdist()` when `--build-manpages` is passed. + from docutils.core import publish_file + from docutils.writers import manpage + + +__all__ = ( # noqa: WPS317, WPS410 + 'build_sdist', 'get_requires_for_build_sdist', +) + + +BUILD_MANPAGES_CONFIG_SETTING = '--build-manpages' +"""Config setting name toggle that is used to request manpage in sdists.""" + + +@contextmanager +def _run_in_temporary_directory() -> t.Iterator[Path]: + with TemporaryDirectory(prefix='.tmp-ansible-pep517-') as tmp_dir: + with _chdir_cm(tmp_dir): + yield Path(tmp_dir) + + +def _make_in_tree_ansible_importable() -> None: + """Add the library directory to module lookup paths.""" + lib_path = str(Path.cwd() / 'lib/') + sys.path.insert(0, lib_path) # NOTE: for the current runtime session + + +def _get_package_distribution_version() -> str: + """Retrieve the current version number from setuptools config.""" + setup_cfg_path = Path.cwd() / 'setup.cfg' + setup_cfg = ConfigParser() + setup_cfg.read_string(setup_cfg_path.read_text()) + cfg_version = setup_cfg.get('metadata', 'version') + importable_version_str = cfg_version.removeprefix('attr: ') + version_mod_str, version_var_str = importable_version_str.rsplit('.', 1) + _make_in_tree_ansible_importable() + return getattr(import_module(version_mod_str), version_var_str) + + +def _generate_rst_in_templates() -> t.Iterable[Path]: + """Create ``*.1.rst.in`` files out of CLI Python modules.""" + generate_man_cmd = ( + sys.executable, + 'hacking/build-ansible.py', + 'generate-man', + '--template-file=docs/templates/man.j2', + '--output-dir=docs/man/man1/', + '--output-format=man', + *Path('lib/ansible/cli/').glob('*.py'), + ) + subprocess.check_call(tuple(map(str, generate_man_cmd))) + return Path('docs/man/man1/').glob('*.1.rst.in') + + +def _convert_rst_in_template_to_manpage( + rst_doc_template: str, + destination_path: os.PathLike, + version_number: str, +) -> None: + """Render pre-made ``*.1.rst.in`` templates into manpages. + + This includes pasting the hardcoded version into the resulting files. + The resulting ``in``-files are wiped in the process. + """ + templated_rst_doc = rst_doc_template.replace('%VERSION%', version_number) + + with StringIO(templated_rst_doc) as in_mem_rst_doc: + publish_file( + source=in_mem_rst_doc, + destination_path=destination_path, + writer=manpage.Writer(), + ) + + +def build_sdist( # noqa: WPS210, WPS430 + sdist_directory: os.PathLike, + config_settings: dict[str, str] | None = None, +) -> str: + build_manpages_requested = BUILD_MANPAGES_CONFIG_SETTING in ( + config_settings or {} + ) + original_src_dir = Path.cwd().resolve() + with _run_in_temporary_directory() as tmp_dir: + tmp_src_dir = Path(tmp_dir) / 'src' + copytree(original_src_dir, tmp_src_dir, symlinks=True) + os.chdir(tmp_src_dir) + + if build_manpages_requested: + Path('docs/man/man1/').mkdir(exist_ok=True, parents=True) + version_number = _get_package_distribution_version() + for rst_in in _generate_rst_in_templates(): + _convert_rst_in_template_to_manpage( + rst_doc_template=rst_in.read_text(), + destination_path=rst_in.with_suffix('').with_suffix(''), + version_number=version_number, + ) + rst_in.unlink() + + Path('pyproject.toml').write_text( + re.sub( + r"""(?x) + backend-path\s=\s\[ # value is a list of double-quoted strings + [^]]+ + ].*\n + build-backend\s=\s"[^"]+".*\n # value is double-quoted + """, + 'build-backend = "setuptools.build_meta"\n', + Path('pyproject.toml').read_text(), + ) + ) + + built_sdist_basename = _setuptools_build_sdist( + sdist_directory=sdist_directory, + config_settings=config_settings, + ) + + return built_sdist_basename + + +def get_requires_for_build_sdist( + config_settings: dict[str, str] | None = None, +) -> list[str]: + build_manpages_requested = BUILD_MANPAGES_CONFIG_SETTING in ( + config_settings or {} + ) + build_manpages_requested = True # FIXME: Once pypa/build#559 is addressed. + + manpage_build_deps = [ + 'docutils', # provides `rst2man` + 'jinja2', # used in `hacking/build-ansible.py generate-man` + 'straight.plugin', # used in `hacking/build-ansible.py` for subcommand + 'pyyaml', # needed for importing in-tree `ansible-core` from `lib/` + ] if build_manpages_requested else [] + + return _setuptools_get_requires_for_build_sdist( + config_settings=config_settings, + ) + manpage_build_deps diff --git a/packaging/pep517_backend/hooks.py b/packaging/pep517_backend/hooks.py new file mode 100644 index 00000000..b834338a --- /dev/null +++ b/packaging/pep517_backend/hooks.py @@ -0,0 +1,9 @@ +# -*- coding: utf-8 -*- + +"""PEP 517 build backend for optionally pre-building docs before setuptools.""" + +from setuptools.build_meta import * # Re-exporting PEP 517 hooks # pylint: disable=unused-wildcard-import,wildcard-import + +from ._backend import ( # noqa: WPS436 # Re-exporting PEP 517 hooks + build_sdist, get_requires_for_build_sdist, +) diff --git a/packaging/release.py b/packaging/release.py new file mode 100755 index 00000000..795e3336 --- /dev/null +++ b/packaging/release.py @@ -0,0 +1,1469 @@ +#!/usr/bin/env python +# PYTHON_ARGCOMPLETE_OK +"""Manage upstream ansible-core releases.""" + +from __future__ import annotations + +import argparse +import contextlib +import dataclasses +import datetime +import enum +import functools +import gzip +import hashlib +import http.client +import inspect +import json +import math +import os +import pathlib +import re +import secrets +import shlex +import shutil +import stat +import subprocess +import sys +import tarfile +import tempfile +import typing as t +import urllib.error +import urllib.parse +import urllib.request +import venv +import webbrowser +import zipfile + +import jinja2 + +from packaging.version import Version, InvalidVersion + +# region CLI Framework + + +C = t.TypeVar("C", bound=t.Callable[[...], None]) + + +def path_to_str(value: t.Any) -> str: + """Return the given value converted to a string suitable for use as a command line argument.""" + return f"{value}/" if isinstance(value, pathlib.Path) and value.is_dir() else str(value) + + +def run( + *args: t.Any, + env: dict[str, t.Any] | None, + cwd: pathlib.Path | str, + capture_output: bool = False, +) -> CompletedProcess: + """Run the specified command.""" + args = [arg.relative_to(cwd) if isinstance(arg, pathlib.Path) else arg for arg in args] + + str_args = tuple(path_to_str(arg) for arg in args) + str_env = {key: path_to_str(value) for key, value in env.items()} if env is not None else None + + display.show(f"--> {shlex.join(str_args)}", color=Display.CYAN) + + try: + p = subprocess.run(str_args, check=True, text=True, env=str_env, cwd=cwd, capture_output=capture_output) + except subprocess.CalledProcessError as ex: + # improve type hinting and include stdout/stderr (if any) in the message + raise CalledProcessError( + message=str(ex), + cmd=str_args, + status=ex.returncode, + stdout=ex.stdout, + stderr=ex.stderr, + ) from None + + # improve type hinting + return CompletedProcess( + args=str_args, + stdout=p.stdout, + stderr=p.stderr, + ) + + +@contextlib.contextmanager +def suppress_when(error_as_warning: bool) -> None: + """Conditionally convert an ApplicationError in the provided context to a warning.""" + if error_as_warning: + try: + yield + except ApplicationError as ex: + display.warning(ex) + else: + yield + + +class ApplicationError(Exception): + """A fatal application error which will be shown without a traceback.""" + + +class CalledProcessError(Exception): + """Results from a failed process.""" + + def __init__(self, message: str, cmd: tuple[str, ...], status: int, stdout: str | None, stderr: str | None) -> None: + if stdout and (stdout := stdout.strip()): + message += f"\n>>> Standard Output\n{stdout}" + + if stderr and (stderr := stderr.strip()): + message += f"\n>>> Standard Error\n{stderr}" + + super().__init__(message) + + self.cmd = cmd + self.status = status + self.stdout = stdout + self.stderr = stderr + + +@dataclasses.dataclass(frozen=True) +class CompletedProcess: + """Results from a completed process.""" + + args: tuple[str, ...] + stdout: str | None + stderr: str | None + + +class Display: + """Display interface for sending output to the console.""" + + CLEAR = "\033[0m" + RED = "\033[31m" + BLUE = "\033[34m" + PURPLE = "\033[35m" + CYAN = "\033[36m" + + def fatal(self, message: t.Any) -> None: + """Print a fatal message to the console.""" + self.show(f"FATAL: {message}", color=self.RED) + + def warning(self, message: t.Any) -> None: + """Print a warning message to the console.""" + self.show(f"WARNING: {message}", color=self.PURPLE) + + def show(self, message: t.Any, color: str | None = None) -> None: + """Print a message to the console.""" + print(f"{color or self.CLEAR}{message}{self.CLEAR}", flush=True) + + +class CommandFramework: + """ + Simple command line framework inspired by nox. + + Argument parsing is handled by argparse. Each function annotated with an instance of this class becomes a subcommand. + Options are shared across all commands, and are defined by providing kwargs when creating an instance of this class. + Options are only defined for commands which have a matching parameter. + + The name of each kwarg is the option name, which will be prefixed with `--` and with underscores converted to dashes. + The value of each kwarg is passed as kwargs to ArgumentParser.add_argument. Passing None results in an internal only parameter. + + The following custom kwargs are recognized and are not passed to add_argument: + + name - Override the positional argument (option) passed to add_argument. + exclusive - Put the argument in an exclusive group of the given name. + """ + + def __init__(self, **kwargs: dict[str, t.Any] | None) -> None: + self.commands: list[C] = [] + self.arguments = kwargs + self.parsed_arguments: argparse.Namespace | None = None + + def __call__(self, func: C) -> C: + """Register the decorated function as a CLI command.""" + self.commands.append(func) + return func + + def run(self, *args: C, **kwargs) -> None: + """Run the specified command(s), using any provided internal args.""" + for arg in args: + self._run(arg, **kwargs) + + def main(self) -> None: + """Main program entry point.""" + parser = argparse.ArgumentParser(description=__doc__) + subparsers = parser.add_subparsers(metavar="COMMAND", required=True) + + for func in self.commands: + func_parser = subparsers.add_parser(self._format_command_name(func), description=func.__doc__, help=func.__doc__) + func_parser.set_defaults(func=func) + + exclusive_groups = {} + signature = inspect.signature(func) + + for name in signature.parameters: + if name not in self.arguments: + raise RuntimeError(f"The '{name}' argument, used by '{func.__name__}', has not been defined.") + + if (arguments := self.arguments.get(name)) is None: + continue # internal use + + arguments = arguments.copy() + exclusive = arguments.pop("exclusive", None) + + if exclusive: + if exclusive not in exclusive_groups: + exclusive_groups[exclusive] = func_parser.add_mutually_exclusive_group() + + command_parser = exclusive_groups[exclusive] + else: + command_parser = func_parser + + if option_name := arguments.pop("name", None): + arguments.update(dest=name) + else: + option_name = f"--{name.replace('_', '-')}" + + command_parser.add_argument(option_name, **arguments) + + try: + # noinspection PyUnresolvedReferences + import argcomplete + except ImportError: + pass + else: + argcomplete.autocomplete(parser) + + self.parsed_arguments = parser.parse_args() + + try: + self.run(self.parsed_arguments.func) + except ApplicationError as ex: + display.fatal(ex) + sys.exit(1) + + def _run(self, func: C, **kwargs) -> None: + """Run the specified command, using any provided internal args.""" + signature = inspect.signature(func) + func_args = {name: getattr(self.parsed_arguments, name) for name in signature.parameters if hasattr(self.parsed_arguments, name)} + func_args.update({name: value for name, value in kwargs.items() if name in signature.parameters}) + printable_args = ", ".join(f"{name}={repr(value)}" for name, value in func_args.items()) + label = f"{self._format_command_name(func)}({printable_args})" + + display.show(f"==> {label}", color=Display.BLUE) + + try: + func(**func_args) + except BaseException: + display.show(f"!!! {label}", color=Display.RED) + raise + + display.show(f"<== {label}", color=Display.BLUE) + + @staticmethod + def _format_command_name(func: C) -> str: + """Return the friendly name of the given command.""" + return func.__name__.replace("_", "-") + + +display = Display() + + +# endregion +# region Data Classes + + +@dataclasses.dataclass(frozen=True) +class GitHubRelease: + """Details required to create a GitHub release.""" + + user: str + repo: str + tag: str + target: str + title: str + body: str + pre_release: bool + + +@dataclasses.dataclass(frozen=True) +class PullRequest: + """Details required to create a pull request.""" + + upstream_user: str + upstream_repo: str + upstream_branch: str + user: str + repo: str + branch: str + title: str + body: str + + +@dataclasses.dataclass(frozen=True) +class Remote: + """Details about a git remote.""" + + name: str + user: str + repo: str + + +@dataclasses.dataclass(frozen=True) +class Remotes: + """Details about git removes.""" + + fork: Remote + upstream: Remote + + +@dataclasses.dataclass(frozen=True) +class GitState: + """Details about the state of the git repository.""" + + remotes: Remotes + branch: str | None + commit: str + + +@dataclasses.dataclass(frozen=True) +class ReleaseArtifact: + """Information about a release artifact on PyPI.""" + + package_type: str + package_label: str + url: str + size: int + digest: str + digest_algorithm: str + + +@dataclasses.dataclass(frozen=True) +class ReleaseAnnouncement: + """Contents of a release announcement.""" + + subject: str + body: str + + +# endregion +# region Utilities + + +SCRIPT_DIR = pathlib.Path(__file__).parent.resolve() +CHECKOUT_DIR = SCRIPT_DIR.parent + +ANSIBLE_LIB_DIR = CHECKOUT_DIR / "lib" +ANSIBLE_DIR = ANSIBLE_LIB_DIR / "ansible" +ANSIBLE_BIN_DIR = CHECKOUT_DIR / "bin" +ANSIBLE_RELEASE_FILE = ANSIBLE_DIR / "release.py" +ANSIBLE_REQUIREMENTS_FILE = CHECKOUT_DIR / "requirements.txt" + +DIST_DIR = CHECKOUT_DIR / "dist" +VENV_DIR = DIST_DIR / ".venv" / "release" + +CHANGELOGS_DIR = CHECKOUT_DIR / "changelogs" +CHANGELOGS_FRAGMENTS_DIR = CHANGELOGS_DIR / "fragments" + +ANSIBLE_VERSION_PATTERN = re.compile("^__version__ = '(?P<version>.*)'$", re.MULTILINE) +ANSIBLE_VERSION_FORMAT = "__version__ = '{version}'" + +DIGEST_ALGORITHM = "sha256" + +# These endpoint names match those defined as defaults in twine. +# See: https://github.com/pypa/twine/blob/9c2c0a1c535155931c3d879359330cb836950c6a/twine/utils.py#L82-L85 +PYPI_ENDPOINTS = dict( + pypi="https://pypi.org/pypi", + testpypi="https://test.pypi.org/pypi", +) + +PIP_ENV = dict( + PIP_REQUIRE_VIRTUALENV="yes", + PIP_DISABLE_PIP_VERSION_CHECK="yes", +) + + +class VersionMode(enum.Enum): + """How to handle the ansible-core version.""" + + DEFAULT = enum.auto() + """Do not allow development versions. Do not allow post release versions.""" + STRIP_POST = enum.auto() + """Do not allow development versions. Strip the post release from the version if present.""" + REQUIRE_POST = enum.auto() + """Do not allow development versions. Require a post release version.""" + REQUIRE_DEV_POST = enum.auto() + """Require a development or post release version.""" + ALLOW_DEV_POST = enum.auto() + """Allow development and post release versions.""" + + def apply(self, version: Version) -> Version: + """Apply the mode to the given version and return the result.""" + original_version = version + + release_component_count = 3 + + if len(version.release) != release_component_count: + raise ApplicationError(f"Version {version} contains {version.release} release components instead of {release_component_count}.") + + if version.epoch: + raise ApplicationError(f"Version {version} contains an epoch component: {version.epoch}") + + if version.local is not None: + raise ApplicationError(f"Version {version} contains a local component: {version.local}") + + if version.is_devrelease and version.is_postrelease: + raise ApplicationError(f"Version {version} is a development and post release version.") + + if self == VersionMode.ALLOW_DEV_POST: + return version + + if self == VersionMode.REQUIRE_DEV_POST: + if not version.is_devrelease and not version.is_postrelease: + raise ApplicationError(f"Version {version} is not a development or post release version.") + + return version + + if version.is_devrelease: + raise ApplicationError(f"Version {version} is a development release: {version.dev}") + + if self == VersionMode.STRIP_POST: + if version.is_postrelease: + version = Version(str(version).removesuffix(f".post{version.post}")) + display.warning(f"Using version {version} by stripping the post release suffix from version {original_version}.") + + return version + + if self == VersionMode.REQUIRE_POST: + if not version.is_postrelease: + raise ApplicationError(f"Version {version} is not a post release version.") + + return version + + if version.is_postrelease: + raise ApplicationError(f"Version {version} is a post release.") + + if self == VersionMode.DEFAULT: + return version + + raise NotImplementedError(self) + + +def git(*args: t.Any, capture_output: bool = False) -> CompletedProcess: + """Run the specified git command.""" + return run("git", *args, env=None, cwd=CHECKOUT_DIR, capture_output=capture_output) + + +def get_commit(rev: str | None = None) -> str: + """Return the commit associated with the given rev, or HEAD if no rev is given.""" + try: + return git("rev-parse", "--quiet", "--verify", "--end-of-options", f"{rev or 'HEAD'}^{{commit}}", capture_output=True).stdout.strip() + except CalledProcessError as ex: + if ex.status == 1 and not ex.stdout and not ex.stderr: + raise ApplicationError(f"Could not find commit: {rev}") from None + + raise + + +def prepare_pull_request(version: Version, branch: str, title: str, add: t.Iterable[pathlib.Path | str], allow_stale: bool) -> PullRequest: + """Return pull request parameters using the provided details.""" + git_state = get_git_state(version, allow_stale) + + if not git("status", "--porcelain", "--untracked-files=no", capture_output=True).stdout.strip(): + raise ApplicationError("There are no changes to commit. Did you skip a step?") + + upstream_branch = get_upstream_branch(version) + body = create_pull_request_body(title) + + git("checkout", "-b", branch) + git("add", *add) + git("commit", "-m", title) + git("push", "--set-upstream", git_state.remotes.fork.name, branch) + git("checkout", git_state.branch or git_state.commit) + git("branch", "-d", branch) + + pr = PullRequest( + upstream_user=git_state.remotes.upstream.user, + upstream_repo=git_state.remotes.upstream.repo, + upstream_branch=upstream_branch, + user=git_state.remotes.fork.user, + repo=git_state.remotes.fork.repo, + branch=branch, + title=title, + body=body, + ) + + return pr + + +def create_github_release(release: GitHubRelease) -> None: + """Open a browser tab for creating the given GitHub release.""" + # See: https://docs.github.com/en/repositories/releasing-projects-on-github/automation-for-release-forms-with-query-parameters + + params = dict( + tag=release.tag, + target=release.target, + title=release.title, + body=release.body, + prerelease=1 if release.pre_release else 0, + ) + + query_string = urllib.parse.urlencode(params) + url = f"https://github.com/{release.user}/{release.repo}/releases/new?{query_string}" + + display.show("Opening release creation page in new tab using default browser ...") + webbrowser.open_new_tab(url) + + +def create_pull_request(pr: PullRequest) -> None: + """Open a browser tab for creating the given pull request.""" + # See: https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/proposing-changes-to-your-work-with-pull-requests/using-query-parameters-to-create-a-pull-request # noqa + + params = dict( + quick_pull=1, + title=pr.title, + body=pr.body, + ) + + query_string = urllib.parse.urlencode(params) + url = f"https://github.com/{pr.upstream_user}/{pr.upstream_repo}/compare/{pr.upstream_branch}...{pr.user}:{pr.repo}:{pr.branch}?{query_string}" + + display.show("Opening pull request in new tab using default browser ...") + webbrowser.open_new_tab(url) + + +def create_pull_request_body(title: str) -> str: + """Return a simple pull request body created from the given title.""" + body = f""" +##### SUMMARY + +{title} + +##### ISSUE TYPE + +Feature Pull Request + +##### COMPONENT NAME + +ansible +""" + + return body.lstrip() + + +def get_remote(name: str, push: bool) -> Remote: + """Return details about the specified remote.""" + remote_url = git("remote", "get-url", *(["--push"] if push else []), name, capture_output=True).stdout.strip() + remote_match = re.search(r"[@/]github[.]com[:/](?P<user>[^/]+)/(?P<repo>[^.]+)(?:[.]git)?$", remote_url) + + if not remote_match: + raise RuntimeError(f"Unable to identify the user and repo in the '{name}' remote: {remote_url}") + + remote = Remote( + name=name, + user=remote_match.group("user"), + repo=remote_match.group("repo"), + ) + + return remote + + +@functools.cache +def get_remotes() -> Remotes: + """Return details about the remotes we need to use.""" + # assume the devel branch has its upstream remote pointing to the user's fork + fork_remote_name = git("branch", "--list", "devel", "--format=%(upstream:remotename)", capture_output=True).stdout.strip() + + if not fork_remote_name: + raise ApplicationError("Could not determine the remote for your fork of Ansible.") + + display.show(f"Detected '{fork_remote_name}' as the remote for your fork of Ansible.") + + # assume there is only one ansible org remote, which would allow release testing using another repo in the same org without special configuration + all_remotes = git("remote", "-v", capture_output=True).stdout.strip().splitlines() + ansible_remote_names = set(line.split()[0] for line in all_remotes if re.search(r"[@/]github[.]com[:/]ansible/", line)) + + if not ansible_remote_names: + raise ApplicationError(f"Could not determine the remote which '{fork_remote_name}' was forked from.") + + if len(ansible_remote_names) > 1: + raise ApplicationError(f"Found multiple candidates for the remote from which '{fork_remote_name}' was forked from: {', '.join(ansible_remote_names)}") + + upstream_remote_name = ansible_remote_names.pop() + + display.show(f"Detected '{upstream_remote_name}' as the remote from which '{fork_remote_name}' was forked from.") + + if fork_remote_name == upstream_remote_name: + raise ApplicationError("The remote for your fork of Ansible cannot be the same as the remote from which it was forked.") + + remotes = Remotes( + fork=get_remote(fork_remote_name, push=True), + upstream=get_remote(upstream_remote_name, push=False), + ) + + return remotes + + +def get_upstream_branch(version: Version) -> str: + """Return the upstream branch name for the given version.""" + return f"stable-{version.major}.{version.minor}" + + +def get_git_state(version: Version, allow_stale: bool) -> GitState: + """Return information about the current state of the git repository.""" + remotes = get_remotes() + upstream_branch = get_upstream_branch(version) + + git("fetch", remotes.upstream.name, upstream_branch) + + upstream_ref = f"{remotes.upstream.name}/{upstream_branch}" + upstream_commit = get_commit(upstream_ref) + + commit = get_commit() + + if commit != upstream_commit: + with suppress_when(allow_stale): + raise ApplicationError(f"The current commit ({commit}) does not match {upstream_ref} ({upstream_commit}).") + + branch = git("branch", "--show-current", capture_output=True).stdout.strip() or None + + state = GitState( + remotes=remotes, + branch=branch, + commit=commit, + ) + + return state + + +@functools.cache +def ensure_venv() -> dict[str, str]: + """Ensure the release venv is ready and return the env vars needed to use it.""" + + # TODO: consider freezing the ansible and release requirements along with their dependencies + + ansible_requirements = ANSIBLE_REQUIREMENTS_FILE.read_text() + + release_requirements = """ +build +twine +""" + + requirements_file = CHECKOUT_DIR / "test/sanity/code-smell/package-data.requirements.txt" + requirements_content = requirements_file.read_text() + requirements_content += ansible_requirements + requirements_content += release_requirements + + requirements_hash = hashlib.sha256(requirements_content.encode()).hexdigest()[:8] + + python_version = ".".join(map(str, sys.version_info[:2])) + + venv_dir = VENV_DIR / python_version / requirements_hash + venv_bin_dir = venv_dir / "bin" + venv_requirements_file = venv_dir / "requirements.txt" + venv_marker_file = venv_dir / "marker.txt" + + env = os.environ.copy() + env.pop("PYTHONPATH", None) # avoid interference from ansible being injected into the environment + env.update( + PATH=os.pathsep.join((str(venv_bin_dir), env["PATH"])), + ) + + if not venv_marker_file.exists(): + display.show(f"Creating a Python {python_version} virtual environment ({requirements_hash}) ...") + + if venv_dir.exists(): + shutil.rmtree(venv_dir) + + venv.create(venv_dir, with_pip=True) + + venv_requirements_file.write_text(requirements_content) + + run("pip", "install", "-r", venv_requirements_file, env=env | PIP_ENV, cwd=CHECKOUT_DIR) + + venv_marker_file.touch() + + return env + + +def get_ansible_version(version: str | None = None, /, commit: str | None = None, mode: VersionMode = VersionMode.DEFAULT) -> Version: + """Parse and return the current ansible-core version, the provided version or the version from the provided commit.""" + if version and commit: + raise ValueError("Specify only one of: version, commit") + + if version: + source = "" + else: + if commit: + current = git("show", f"{commit}:{ANSIBLE_RELEASE_FILE.relative_to(CHECKOUT_DIR)}", capture_output=True).stdout + else: + current = ANSIBLE_RELEASE_FILE.read_text() + + if not (match := ANSIBLE_VERSION_PATTERN.search(current)): + raise RuntimeError("Failed to get the ansible-core version.") + + version = match.group("version") + source = f" in '{ANSIBLE_RELEASE_FILE}'" + + try: + parsed_version = Version(version) + except InvalidVersion: + raise ApplicationError(f"Invalid version{source}: {version}") from None + + parsed_version = mode.apply(parsed_version) + + return parsed_version + + +def get_next_version(version: Version, /, final: bool = False, pre: str | None = None, mode: VersionMode = VersionMode.DEFAULT) -> Version: + """Return the next version after the specified version.""" + + # TODO: consider using development versions instead of post versions after a release is published + + pre = pre or "" + micro = version.micro + + if version.is_devrelease: + # The next version of a development release is the same version without the development component. + if final: + pre = "" + elif not pre and version.pre is not None: + pre = f"{version.pre[0]}{version.pre[1]}" + elif version.is_postrelease: + # The next version of a post release is the next pre-release *or* micro release component. + if final: + pre = "" + elif not pre and version.pre is not None: + pre = f"{version.pre[0]}{version.pre[1] + 1}" + + if version.pre is None: + micro = version.micro + 1 + else: + raise ApplicationError(f"Version {version} is not a development or post release version.") + + version = f"{version.major}.{version.minor}.{micro}{pre}" + + return get_ansible_version(version, mode=mode) + + +def check_ansible_version(current_version: Version, requested_version: Version) -> None: + """Verify the requested version is valid for the current version.""" + if requested_version.release[:2] != current_version.release[:2]: + raise ApplicationError(f"Version {requested_version} does not match the major and minor portion of the current version: {current_version}") + + if requested_version < current_version: + raise ApplicationError(f"Version {requested_version} is older than the current version: {current_version}") + + # TODO: consider additional checks to avoid mistakes when incrementing the release version + + +def set_ansible_version(current_version: Version, requested_version: Version) -> None: + """Set the current ansible-core version.""" + check_ansible_version(current_version, requested_version) + + if requested_version == current_version: + return + + display.show(f"Updating version {current_version} to {requested_version} ...") + + current = ANSIBLE_RELEASE_FILE.read_text() + updated = ANSIBLE_VERSION_PATTERN.sub(ANSIBLE_VERSION_FORMAT.format(version=requested_version), current) + + if current == updated: + raise RuntimeError("Failed to set the ansible-core version.") + + ANSIBLE_RELEASE_FILE.write_text(updated) + + +def create_reproducible_sdist(original_path: pathlib.Path, output_path: pathlib.Path, mtime: int) -> None: + """Read the specified sdist and write out a new copy with uniform file metadata at the specified location.""" + with tarfile.open(original_path) as original_archive: + with tempfile.TemporaryDirectory() as temp_dir: + tar_file = pathlib.Path(temp_dir) / "sdist.tar" + + with tarfile.open(tar_file, mode="w") as tar_archive: + for original_info in original_archive.getmembers(): # type: tarfile.TarInfo + tar_archive.addfile(create_reproducible_tar_info(original_info, mtime), original_archive.extractfile(original_info)) + + with tar_file.open("rb") as tar_archive: + with gzip.GzipFile(output_path, "wb", mtime=mtime) as output_archive: + shutil.copyfileobj(tar_archive, output_archive) + + +def create_reproducible_tar_info(original: tarfile.TarInfo, mtime: int) -> tarfile.TarInfo: + """Return a copy of the given TarInfo with uniform file metadata.""" + sanitized = tarfile.TarInfo() + sanitized.name = original.name + sanitized.size = original.size + sanitized.mtime = mtime + sanitized.mode = (original.mode & ~(stat.S_IRWXU | stat.S_IRWXG | stat.S_IRWXO)) | stat.S_IRUSR | stat.S_IRGRP | stat.S_IROTH | stat.S_IWUSR + sanitized.type = original.type + sanitized.linkname = original.linkname + sanitized.uid = 0 + sanitized.gid = 0 + sanitized.uname = "root" + sanitized.gname = "root" + + if original.mode & (stat.S_IXUSR | stat.S_IXGRP | stat.S_IXOTH): + sanitized.mode |= stat.S_IXUSR | stat.S_IXGRP | stat.S_IXOTH + + return sanitized + + +def test_built_artifact(path: pathlib.Path) -> None: + """Test the specified built artifact by installing it in a venv and running some basic commands.""" + with tempfile.TemporaryDirectory() as temp_dir_name: + temp_dir = pathlib.Path(temp_dir_name) + + venv_dir = temp_dir / "venv" + venv_bin_dir = venv_dir / "bin" + + venv.create(venv_dir, with_pip=True) + + env = os.environ.copy() + env.pop("PYTHONPATH", None) # avoid interference from ansible being injected into the environment + env.update( + PATH=os.pathsep.join((str(venv_bin_dir), env["PATH"])), + ) + + run("pip", "install", path, env=env | PIP_ENV, cwd=CHECKOUT_DIR) + + run("ansible", "--version", env=env, cwd=CHECKOUT_DIR) + run("ansible-test", "--version", env=env, cwd=CHECKOUT_DIR) + + +def get_sdist_path(version: Version, dist_dir: pathlib.Path = DIST_DIR) -> pathlib.Path: + """Return the path to the sdist file.""" + return dist_dir / f"ansible-core-{version}.tar.gz" + + +def get_wheel_path(version: Version, dist_dir: pathlib.Path = DIST_DIR) -> pathlib.Path: + """Return the path to the wheel file.""" + return dist_dir / f"ansible_core-{version}-py3-none-any.whl" + + +def calculate_digest(path: pathlib.Path) -> str: + """Return the digest for the specified file.""" + # TODO: use hashlib.file_digest once Python 3.11 is the minimum supported version + return hashlib.new(DIGEST_ALGORITHM, path.read_bytes()).hexdigest() + + +@functools.cache +def get_release_artifact_details(repository: str, version: Version, validate: bool) -> list[ReleaseArtifact]: + """Return information about the release artifacts hosted on PyPI.""" + endpoint = PYPI_ENDPOINTS[repository] + url = f"{endpoint}/ansible-core/{version}/json" + + opener = urllib.request.build_opener() + response: http.client.HTTPResponse + + try: + with opener.open(url) as response: + data = json.load(response) + except urllib.error.HTTPError as ex: + if ex.status == http.HTTPStatus.NOT_FOUND: + raise ApplicationError(f"Version {version} not found on PyPI.") from None + + raise RuntimeError(f"Failed to get {version} from PyPI: {ex}") from ex + + artifacts = [describe_release_artifact(version, item, validate) for item in data["urls"]] + + expected_artifact_types = {"bdist_wheel", "sdist"} + found_artifact_types = set(artifact.package_type for artifact in artifacts) + + if found_artifact_types != expected_artifact_types: + raise RuntimeError(f"Expected {expected_artifact_types} artifact types, but found {found_artifact_types} instead.") + + return artifacts + + +def describe_release_artifact(version: Version, item: dict[str, t.Any], validate: bool) -> ReleaseArtifact: + """Return release artifact details extracted from the given PyPI data.""" + package_type = item["packagetype"] + + # The artifact URL is documented as stable, so is safe to put in release notes and announcements. + # See: https://github.com/pypi/warehouse/blame/c95be4a1055f4b36a8852715eb80318c81fc00ca/docs/api-reference/integration-guide.rst#L86-L90 + url = item["url"] + + pypi_size = item["size"] + pypi_digest = item["digests"][DIGEST_ALGORITHM] + + if package_type == "bdist_wheel": + local_artifact_file = get_wheel_path(version) + package_label = "Built Distribution" + elif package_type == "sdist": + local_artifact_file = get_sdist_path(version) + package_label = "Source Distribution" + else: + raise NotImplementedError(f"Package type '{package_type}' is not supported.") + + if validate: + try: + local_size = local_artifact_file.stat().st_size + local_digest = calculate_digest(local_artifact_file) + except FileNotFoundError: + raise ApplicationError(f"Missing local artifact: {local_artifact_file.relative_to(CHECKOUT_DIR)}") from None + + if local_size != pypi_size: + raise ApplicationError(f"The {version} local {package_type} size {local_size} does not match the PyPI size {pypi_size}.") + + if local_digest != pypi_digest: + raise ApplicationError(f"The {version} local {package_type} digest '{local_digest}' does not match the PyPI digest '{pypi_digest}'.") + + return ReleaseArtifact( + package_type=package_type, + package_label=package_label, + url=url, + size=pypi_size, + digest=pypi_digest, + digest_algorithm=DIGEST_ALGORITHM.upper(), + ) + + +def get_next_release_date(start: datetime.date, step: int, after: datetime.date) -> datetime.date: + """Return the next release date.""" + if start > after: + raise ValueError(f"{start=} is greater than {after=}") + + current_delta = after - start + release_delta = datetime.timedelta(days=(math.floor(current_delta.days / step) + 1) * step) + + release = start + release_delta + + return release + + +def create_template_environment() -> jinja2.Environment: + """Create and return a jinja2 environment.""" + env = jinja2.Environment() + env.filters.update( + basename=os.path.basename, + ) + + return env + + +def create_github_release_notes(upstream: Remote, repository: str, version: Version, validate: bool) -> str: + """Create and return GitHub release notes.""" + env = create_template_environment() + template = env.from_string(GITHUB_RELEASE_NOTES_TEMPLATE) + + variables = dict( + version=version, + releases=get_release_artifact_details(repository, version, validate), + changelog=f"https://github.com/{upstream.user}/{upstream.repo}/blob/v{ version }/changelogs/CHANGELOG-v{ version.major }.{ version.minor }.rst", + ) + + release_notes = template.render(**variables).strip() + + return release_notes + + +def create_release_announcement(upstream: Remote, repository: str, version: Version, validate: bool) -> ReleaseAnnouncement: + """Create and return a release announcement message.""" + env = create_template_environment() + subject_template = env.from_string(RELEASE_ANNOUNCEMENT_SUBJECT_TEMPLATE) + body_template = env.from_string(RELEASE_ANNOUNCEMENT_BODY_TEMPLATE) + + today = datetime.datetime.now(tz=datetime.timezone.utc).date() + + variables = dict( + version=version, + info=dict( + name="ansible-core", + short=f"{version.major}.{version.minor}", + releases=get_release_artifact_details(repository, version, validate), + ), + next_rc=get_next_release_date(datetime.date(2021, 8, 9), 28, today), + next_ga=get_next_release_date(datetime.date(2021, 8, 16), 28, today), + rc=version.pre and version.pre[0] == "rc", + beta=version.pre and version.pre[0] == "b", + alpha=version.pre and version.pre[0] == "a", + major=version.micro == 0, + upstream=upstream, + ) + + if version.pre and version.pre[0] in ("a", "b"): + display.warning("The release announcement template does not populate the date for the next release.") + + subject = subject_template.render(**variables).strip() + body = body_template.render(**variables).strip() + + message = ReleaseAnnouncement( + subject=subject, + body=body, + ) + + return message + + +# endregion +# region Templates + + +FINAL_RELEASE_ANNOUNCEMENT_RECIPIENTS = [ + "ansible-announce@googlegroups.com", + "ansible-project@googlegroups.com", + "ansible-devel@googlegroups.com", +] + +PRE_RELEASE_ANNOUNCEMENT_RECIPIENTS = [ + "ansible-devel@googlegroups.com", +] + +GITHUB_RELEASE_NOTES_TEMPLATE = """ +# Changelog + +See the [full changelog]({{ changelog }}) for the changes included in this release. + +# Release Artifacts + +{%- for release in releases %} +* {{ release.package_label }}: [{{ release.url|basename }}]({{ release.url }}) - {{ release.size }} bytes + * {{ release.digest }} ({{ release.digest_algorithm }}) +{%- endfor %} +""" + +# These release templates were adapted from sivel's release announcement script. +# See: https://gist.github.com/sivel/937bc2862a9677d8db875f3b10744d8c + +RELEASE_ANNOUNCEMENT_SUBJECT_TEMPLATE = """ +New release{% if rc %} candidate{% elif beta %} beta{% elif alpha %} alpha{% endif %}: {{ info.name }} {{ version }} +""" + +# NOTE: Gmail will automatically wrap the plain text version when sending. +# There's no need to perform wrapping ahead of time for normal sentences. +# However, lines with special formatting should be kept short to avoid unwanted wrapping. +RELEASE_ANNOUNCEMENT_BODY_TEMPLATE = """ +Hi all- we're happy to announce the{{ " " }} +{%- if rc -%} +following release candidate +{%- elif beta -%} +beta release of +{%- elif alpha -%} +alpha release of +{%- else -%} +general release of +{%- endif -%}: + +{{ info.name }} {{ version }} + + +How to get it +------------- + +$ python3 -m pip install --user {{ info.name }}=={{ version }} + +The release artifacts can be found here: +{% for release in info.releases %} +# {{ release.package_label }}: {{ release.size }} bytes +# {{ release.digest_algorithm }}: {{ release.digest }} +{{ release.url }} +{%- endfor %} + + +What's new +---------- + +{% if major %} +This release is a major release. +{%- else -%} +This release is a maintenance release containing numerous bugfixes. +{% endif %} +The full changelog can be found here: + +https://github.com/{{ upstream.user }}/{{ upstream.repo }}/blob/v{{ version }}/changelogs/CHANGELOG-v{{ info.short }}.rst + + +Schedule for future releases +---------------------------- +{% if rc %} +The release candidate will become a general availability release on {{ next_ga.strftime('%-d %B %Y') }}. +{% elif beta %} +Subject to the need for additional beta releases, the first release candidate is scheduled for X. +{% elif alpha %} +Subject to the need for additional alpha releases, the first release beta is scheduled for X. +{% else %} +The next release candidate is planned to be released on {{ next_rc.strftime('%-d %B %Y') }}. The next general availability release will be one week after. +{% endif %} + +Porting help +------------ + +If you discover any errors or if any of your working playbooks break when you upgrade, please use the following link to report the regression: + +https://github.com/{{ upstream.user }}/{{ upstream.repo }}/issues/new/choose + +In your issue, be sure to mention the version that works and the one that doesn't. + +Thanks! +""" + +# endregion +# region Commands + +command = CommandFramework( + repository=dict(metavar="REPO", choices=tuple(PYPI_ENDPOINTS), default="pypi", help="PyPI repository to use: %(choices)s [%(default)s]"), + version=dict(exclusive="version", help="version to set"), + pre=dict(exclusive="version", help="increment version to the specified pre-release (aN, bN, rcN)"), + final=dict(exclusive="version", action="store_true", help="increment version to the next final release"), + commit=dict(help="commit to tag"), + mailto=dict(name="--no-mailto", action="store_false", help="write announcement to console instead of using a mailto: link"), + validate=dict(name="--no-validate", action="store_false", help="disable validation of PyPI artifacts against local ones"), + prompt=dict(name="--no-prompt", action="store_false", help="disable interactive prompt before publishing with twine"), + allow_tag=dict(action="store_true", help="allow an existing release tag (for testing)"), + allow_stale=dict(action="store_true", help="allow a stale checkout (for testing)"), + allow_dirty=dict(action="store_true", help="allow untracked files and files with changes (for testing)"), +) + + +@command +def instructions() -> None: + """Show instructions for the release process.""" + message = """ +Releases must be performed using an up-to-date checkout of a fork of the Ansible repository. + +1. Make sure your checkout is up-to-date. +2. Run the `prepare` command [1], then: + a. Submit the PR opened in the browser. + b. Wait for CI to pass. + c. Merge the PR. +3. Update your checkout to include the commit from the PR which was just merged. +4. Run the `complete` command [2], then: + a. Submit the GitHub release opened in the browser. + b. Submit the PR opened in the browser. + c. Send the release announcement opened in your browser. + d. Wait for CI to pass. + e. Merge the PR. + +[1] Use the `--final`, `--pre` or `--version` option for control over the version. +[2] During the `publish` step, `twine` may prompt for credentials. +""" + + display.show(message.strip()) + + +@command +def show_version(final: bool = False, pre: str | None = None) -> None: + """Show the current and next ansible-core version.""" + current_version = get_ansible_version(mode=VersionMode.ALLOW_DEV_POST) + display.show(f"Current version: {current_version}") + + try: + next_version = get_next_version(current_version, final=final, pre=pre) + except ApplicationError as ex: + display.show(f" Next version: Unknown - {ex}") + else: + display.show(f" Next version: {next_version}") + + check_ansible_version(current_version, next_version) + + +@command +def check_state(allow_stale: bool = False) -> None: + """Verify the git repository is in a usable state for creating a pull request.""" + get_git_state(get_ansible_version(), allow_stale) + + +# noinspection PyUnusedLocal +@command +def prepare(final: bool = False, pre: str | None = None, version: str | None = None) -> None: + """Prepare a release.""" + command.run( + update_version, + check_state, + generate_summary, + generate_changelog, + create_release_pr, + ) + + +@command +def update_version(final: bool = False, pre: str | None = None, version: str | None = None) -> None: + """Update the version embedded in the source code.""" + current_version = get_ansible_version(mode=VersionMode.REQUIRE_DEV_POST) + + if version: + requested_version = get_ansible_version(version) + else: + requested_version = get_next_version(current_version, final=final, pre=pre) + + set_ansible_version(current_version, requested_version) + + +@command +def generate_summary() -> None: + """Generate a summary changelog fragment for this release.""" + version = get_ansible_version() + release_date = datetime.datetime.now(datetime.timezone.utc).strftime("%Y-%m-%d") + summary_path = CHANGELOGS_FRAGMENTS_DIR / f"{version}_summary.yaml" + major_minor = f"{version.major}.{version.minor}" + + content = f""" +release_summary: | + | Release Date: {release_date} + | `Porting Guide <https://docs.ansible.com/ansible-core/{major_minor}/porting_guides/porting_guide_core_{major_minor}.html>`__ +""" + + summary_path.write_text(content.lstrip()) + + +@command +def generate_changelog() -> None: + """Generate the changelog and validate the results.""" + env = ensure_venv() + env.update( + PATH=os.pathsep.join((str(ANSIBLE_BIN_DIR), env["PATH"])), + PYTHONPATH=ANSIBLE_LIB_DIR, + ) + + # TODO: consider switching back to the original changelog generator instead of using antsibull-changelog + + run("antsibull-changelog", "release", "-vv", "--use-ansible-doc", env=env, cwd=CHECKOUT_DIR) + run("antsibull-changelog", "generate", "-vv", "--use-ansible-doc", env=env, cwd=CHECKOUT_DIR) + + run("ansible-test", "sanity", CHANGELOGS_DIR, ANSIBLE_RELEASE_FILE, env=env, cwd=CHECKOUT_DIR) + + +@command +def create_release_pr(allow_stale: bool = False) -> None: + """Create a branch and open a browser tab for creating a release pull request.""" + version = get_ansible_version() + + pr = prepare_pull_request( + version=version, + branch=f"release-{version}-{secrets.token_hex(4)}", + title=f"New release v{version}", + add=( + CHANGELOGS_DIR, + ANSIBLE_RELEASE_FILE, + ), + allow_stale=allow_stale, + ) + + create_pull_request(pr) + + +# noinspection PyUnusedLocal +@command +def complete(repository: str, mailto: bool = True, allow_dirty: bool = False) -> None: + """Complete a release after the prepared changes have been merged.""" + command.run( + check_state, + build, + test, + publish, + tag_release, + post_version, + create_post_pr, + release_announcement, + ) + + +@command +def build(allow_dirty: bool = False) -> None: + """Build the sdist and wheel.""" + version = get_ansible_version(mode=VersionMode.ALLOW_DEV_POST) + env = ensure_venv() + + dirty = git("status", "--porcelain", "--untracked-files=all", capture_output=True).stdout.strip().splitlines() + + if dirty: + with suppress_when(allow_dirty): + raise ApplicationError(f"There are {len(dirty)} files which are untracked and/or have changes, which will be omitted from the build.") + + sdist_file = get_sdist_path(version) + wheel_file = get_wheel_path(version) + + with tempfile.TemporaryDirectory(dir=DIST_DIR, prefix=f"build-{version}-", suffix=".tmp") as temp_dir_name: + temp_dir = pathlib.Path(temp_dir_name) + dist_dir = temp_dir / "dist" + + commit_time = int(git("show", "-s", "--format=%ct", capture_output=True).stdout) + + env.update( + SOURCE_DATE_EPOCH=str(commit_time), + ) + + git("worktree", "add", "-d", temp_dir) + + try: + run("python", "-m", "build", "--config-setting=--build-manpages", env=env, cwd=temp_dir) + + create_reproducible_sdist(get_sdist_path(version, dist_dir), sdist_file, commit_time) + get_wheel_path(version, dist_dir).rename(wheel_file) + finally: + git("worktree", "remove", temp_dir) + + +@command +def test() -> None: + """Test the sdist and wheel.""" + command.run( + test_sdist, + test_wheel, + ) + + +@command +def test_sdist() -> None: + """Test the sdist.""" + version = get_ansible_version(mode=VersionMode.ALLOW_DEV_POST) + sdist_file = get_sdist_path(version) + + with tempfile.TemporaryDirectory() as temp_dir_name: + temp_dir = pathlib.Path(temp_dir_name) + + with contextlib.ExitStack() as stack: + try: + sdist = stack.enter_context(tarfile.open(sdist_file)) + except FileNotFoundError: + raise ApplicationError(f"Missing sdist: {sdist_file.relative_to(CHECKOUT_DIR)}") from None + + sdist.extractall(temp_dir) + + man1_dir = temp_dir / sdist_file.with_suffix("").with_suffix("").name / "docs" / "man" / "man1" + man1_pages = sorted(man1_dir.glob("*.1")) + + if not man1_pages: + raise ApplicationError(f"No man pages found in the sdist at: {man1_dir.relative_to(temp_dir)}") + + pyc_glob = "*.pyc*" + pyc_files = sorted(path.relative_to(temp_dir) for path in temp_dir.rglob(pyc_glob)) + + if pyc_files: + raise ApplicationError(f"Found {len(pyc_files)} '{pyc_glob}' file(s): {', '.join(map(str, pyc_files))}") + + display.show(f"Found man1 pages: {', '.join([path.name for path in man1_pages])}") + + test_built_artifact(sdist_file) + + +@command +def test_wheel() -> None: + """Test the wheel.""" + version = get_ansible_version(mode=VersionMode.ALLOW_DEV_POST) + wheel_file = get_wheel_path(version) + + with tempfile.TemporaryDirectory() as temp_dir_name: + temp_dir = pathlib.Path(temp_dir_name) + + with contextlib.ExitStack() as stack: + try: + wheel = stack.enter_context(zipfile.ZipFile(wheel_file)) + except FileNotFoundError: + raise ApplicationError(f"Missing wheel for version {version}: {wheel_file}") from None + + wheel.extractall(temp_dir) + + test_built_artifact(wheel_file) + + +@command +def publish(repository: str, prompt: bool = True) -> None: + """Publish to PyPI.""" + version = get_ansible_version() + sdist_file = get_sdist_path(version) + wheel_file = get_wheel_path(version) + env = ensure_venv() + + if prompt: + try: + while input(f"Do you want to publish {version} to the '{repository}' repository?\nEnter the repository name to confirm: ") != repository: + pass + except KeyboardInterrupt: + display.show("") + raise ApplicationError("Publishing was aborted by the user.") from None + + run("twine", "upload", "-r", repository, sdist_file, wheel_file, env=env, cwd=CHECKOUT_DIR) + + +@command +def tag_release(repository: str, commit: str | None = None, validate: bool = True, allow_tag: bool = False) -> None: + """Create a GitHub release using the current or specified commit.""" + upstream = get_remotes().upstream + + if commit: + git("fetch", upstream.name) # fetch upstream to make sure the commit can be found + + commit = get_commit(commit) + version = get_ansible_version(commit=commit) + tag = f"v{version}" + + if upstream_tag := git("ls-remote", "--tags", upstream.name, tag, capture_output=True).stdout.strip(): + with suppress_when(allow_tag): + raise ApplicationError(f"Version {version} has already been tagged: {upstream_tag}") + + upstream_branch = get_upstream_branch(version) + upstream_refs = git("branch", "-r", "--format=%(refname)", "--contains", commit, capture_output=True).stdout.strip().splitlines() + upstream_ref = f"refs/remotes/{upstream.name}/{upstream_branch}" + + if upstream_ref not in upstream_refs: + raise ApplicationError(f"Commit {upstream_ref} not found. Found {len(upstream_refs)} upstream ref(s): {', '.join(upstream_refs)}") + + body = create_github_release_notes(upstream, repository, version, validate) + + release = GitHubRelease( + user=upstream.user, + repo=upstream.repo, + target=commit, + tag=tag, + title=tag, + body=body, + pre_release=version.pre is not None, + ) + + create_github_release(release) + + +@command +def post_version() -> None: + """Set the post release version.""" + current_version = get_ansible_version() + requested_version = get_ansible_version(f"{current_version}.post0", mode=VersionMode.REQUIRE_POST) + + set_ansible_version(current_version, requested_version) + + +@command +def create_post_pr(allow_stale: bool = False) -> None: + """Create a branch and open a browser tab for creating a post release pull request.""" + version = get_ansible_version(mode=VersionMode.REQUIRE_POST) + + pr = prepare_pull_request( + version=version, + branch=f"release-{version}-{secrets.token_hex(4)}", + title=f"Update Ansible release version to v{version}.", + add=(ANSIBLE_RELEASE_FILE,), + allow_stale=allow_stale, + ) + + create_pull_request(pr) + + +@command +def release_announcement(repository: str, version: str | None = None, mailto: bool = True, validate: bool = True) -> None: + """Generate a release announcement for the current or specified version.""" + parsed_version = get_ansible_version(version, mode=VersionMode.STRIP_POST) + upstream = get_remotes().upstream + message = create_release_announcement(upstream, repository, parsed_version, validate) + recipient_list = PRE_RELEASE_ANNOUNCEMENT_RECIPIENTS if parsed_version.is_prerelease else FINAL_RELEASE_ANNOUNCEMENT_RECIPIENTS + recipients = ", ".join(recipient_list) + + if mailto: + to = urllib.parse.quote(recipients) + + params = dict( + subject=message.subject, + body=message.body, + ) + + query_string = urllib.parse.urlencode(params) + url = f"mailto:{to}?{query_string}" + + display.show("Opening email client through default web browser ...") + webbrowser.open(url) + else: + print(f"TO: {recipients}") + print(f"SUBJECT: {message.subject}") + print() + print(message.body) + + +# endregion + + +if __name__ == "__main__": + command.main() |