Skip to content

Latest commit

 

History

History
1675 lines (1235 loc) · 69.3 KB

File metadata and controls

1675 lines (1235 loc) · 69.3 KB

Table of contents


Intro

What the provider distributions are

The Provider distributions are separate packages (one package per provider) that implement integrations with external services for Airflow in the form of installable Python packages.

The Release Manager prepares packages separately from the main Airflow Release, using breeze commands and accompanying scripts. This document provides an overview of the command line tools needed to prepare the packages.

NOTE!! When you have problems with any of those commands that run inside breeze docker image, you can run the command with --debug flag that will drop you in the shell inside the image and will print the command that you should run.

Decide when to release

You can release Provider distributions separately from the main Airflow on an ad-hoc basis, whenever we find that a given provider needs to be released due to new features or due to bug fixes. You can release each provider package separately, but due to voting and release overhead we try to group releases of Provider distributions together.

Collect ambiguities during the release (for a follow-up doc PR)

These instructions are imperfect. Every release uncovers at least one command that has drifted, one step that is under-documented, or one automation that silently did the wrong thing. As you run through this document, jot down any such observations in a scratch file kept outside the repo (anywhere that is not tracked by git — a note in your home directory, a scratchpad, a gist). Once the release has landed, turn those notes into a follow-up PR against this document.

Keeping the scratch file out of the repo avoids accidentally committing release-manager notes along with the release-prep PR, and makes it obvious that the notes are input to the next doc PR rather than something to keep around long-term.

Special procedures (done very infrequently)

Note

Those processes are done very infrequently, when there is time to bump minimum versions of providers or when you remove a provider. Usually you should just skip this section and go straight to Prepare Regular providers RC.

Bump min Airflow version for providers

Note

This should only happen when it is time to bump the minimum version of providers as agreed in related provider policy

  1. Update PROVIDERS_COMPATIBILITY_TESTS_MATRIX in src/airflow_breeze/global_constants.py to remove the versions of Airflow that are not applicable anymore.

  2. Check if Breeze unit tests in dev/breeze/tests/test_packages.py need adjustments. This is done by simply searching and replacing old version occurrences with newer one. For example 2.10.0 to 2.11.0

  3. Update minimum airflow version for all packages, you should modify MIN_AIRFLOW_VERSION in src/airflow_breeze/utils/packages.py and run the breeze release-management prepare-provider-documentation --only-min-version-update This will only update the min version in the __init__.py files and package documentation without bumping the provider versions.

  4. Remove AIRFLOW_V_X_Y_PLUS in all tests (review and update skipif and other conditional behavior and test_compat.py, where X is the TARGET version we change to. For example when we update min Airflow version to 3.0.0, we should remove all references to AIRFLOW_V_3_0_PLUS simply because "everything" in our tests is already 3.0.0+ and there is no need to exclude or modify tests for earlier versions of Airflow.

Note: Sometimes we are releasing a subset of providers and would not want to add the list of these providers to every breeze command we run, specifically: prepare-provider-distributions, build-docs , publish-docs, and, add-back-references. In this case, we can instead export an environment variable: DISTRIBUTIONS_LIST, and it will work for every breeze command involved in the release process. The value can also be passed as the --distributions-list argument. Follow the steps below to set the environment variable:

 export DISTRIBUTIONS_LIST="PACKAGE1 PACKAGE2"
branch="update-min-airflow-version"
git checkout -b "${branch}"
breeze release-management prepare-provider-documentation --only-min-version-update
git add .
git commit -m "Bump minimum Airflow version in providers to Airflow 3.0.0"
git push --set-upstream origin "${branch}"

Note: that this command will only bump the min airflow versions for those providers that do not have it set to a higher version. You do not have to skip specific providers - run it for all providers it will handle everything automatically.

Note: this step is not part of the release cycle. It should be done independently when the time to update min airflow version has come.

Move provider into remove state

Note

This is only needed in case some providers have been removed since last release wave.

The removed state needs to be in a release wave before you actually plan to remove the source code for the provider. Set provider with removed state -> release provider -> remove source code of the provider. When setting the provider in removed state you need also to clarify in the change log that there will be no more releases for this provider.

To set provider as removed do the following:

  1. In provider yaml change state from to ready to removed
  2. Place entry in changelog.txt that notify users about provider being removed.
  3. Update test_get_removed_providers in /dev/breeze/tests/test_packages.py by adding the provider to the list

Prepare Regular Provider distributions (RC)

This is the process that happens regularly (every 2 weeks).

Perform review of security issues that are marked for the release

We are keeping track of security issues in the Security Issues repository currently. As a release manager, you should have access to the repository. Please review and ensure that all security issues marked for the release have been addressed and resolved. Ping security team (comment in the issues) if anything missing or the issue does not seem to be addressed.

Additionally, the dependabot alerts and code scanning alerts should be reviewed and security team should be pinged to review and resolve them.

Convert commits to changelog entries and bump provider versions

First thing that release manager has to do is to convert commits for each provider into changelog entries and update version of the provider to a target version - depending on type of changes implemented in the providers. This is happening by running breeze prepare-provider-documentation. The tool aids the release manager to classify each of the changes and will automatically increase version of the provider.

Before running the command, you need to set the environment variable RELEASE_DATE to the date of the release, usually current on next day - depending on when you plan to start the release. This can be updated later, when you rebase the PR with the latest main before merging it.

export RELEASE_DATE=$(date "+%Y-%m-%d")

The RELEASE_DATE can also optionally have _01, _02 suffixes appended to it to indicate that the release is the first, second or third release candidate for the provider that day. It's unlikely to happen that we have several release candidates for the same provider in a single day, but this makes it possible to release multiple providers in a single day.

Note that versions of the provider will be updated in two places (the tool does it automatically):

  • provider.yaml - where there is a list of all versions of providers (the first one is the latest)
  • changelog.rst - where changelogs are sorted according to provider version and group changes in the right sections
breeze release-management prepare-provider-documentation

In case you prepare provider documentation for just a few selected providers, you can run:

breeze release-management prepare-provider-documentation [packages]

This happens automatically if you set DISTRIBUTIONS_LIST variable.

In case you want to also release a pre-installed provider that is in not-ready state (i.e. when you want to release it before you switch their state to ready), you need to pass --include-not-ready-providers flag to the command above.

The tool determines the new version of provider as follows:

  • increased patch-level for bugfix-only and doc-only changes
  • increased minor version if new features are added
  • increased major version if breaking changes are added

Update versions of dependent providers to the next version

Sometimes when contributors want to use next version of a dependent provider, instead of doing it immediately in the code they can add a comment # use next version to the line of pyproject.toml file of the provider that refers to the provider, which next version should be used. This comment will be picked up by theupdate-providers-next-version command and the version of the dependent provider will be updated to the next version and comment will be removed.

breeze release-management update-providers-next-version

Create a PR with the changes

Make sure to set labels: allow provider dependency bump and skip common compat check to the PR, so that the PR is not blocked by selective checks.

You can do it for example this way:

gh pr create \
  --title "Prepare providers release ${RELEASE_DATE}" \
  --label "allow provider dependency bump" \
  --label "skip common compat check" \
  --body "Prepare providers release ${RELEASE_DATE}" \
  --web

Apply incremental changes and merge the PR

When those changes are generated, you should commit the changes, create a PR and get it reviewed. This usually takes some time, so before merging you need to rebase it to latest main and see if there are no new, incremental updates (one or two merged commit in the meantime). If there are - you still have a chance to incorporate the changes via incremental-update process for provider documentation:

  breeze release-management prepare-provider-documentation --incremental-update

In case you prepare provider documentation for just a few selected providers, you can run:

breeze release-management prepare-provider-documentation --incremental-update [packages]

Once you do it, the diff will be generated so you will see if there are any new changes added to changelogs. If there are, you need to add them to PR and classify the changes manually:

  • move the changes to the right sections in changelog
  • remove the "Please review" comments generated by the incremental update process
  • if needed adjust version of provider - in changelog and provider.yaml, in case the new change changes classification of the upgrade (patchlevel/minor/major)

Commit the changes and create the PR. You need to apply the following labels to the PR:

  • skip common compat check
  • allow provider dependency bump

Once approved, merge it, be careful to do it quickly so that no new PRs are merged for providers in the meantime - if they are, you will miss them in the changelog.

In case you want to also release a pre-installed provider that is in not-ready state (i.e. when you want to release it before you switch their state to ready), you need to pass --include-not-ready-providers flag to the command above.

Note

In case you prepare provider's documentation in a branch different than main, you need to manually specify the base branch via --base-branch parameter. For example if you try to build a cncf.kubernetes provider that is build from provider-cncf-kubernetes/v4-4 branch should be prepared like this:

breeze release-management prepare-provider-documentation --include-removed-providers \
 --base-branch provider-cncf-kubernetes/v4-4 cncf.kubernetes

(Optional) Apply template updates

This step should only be executed if we want to change template files for the providers, i.e. change security information, commit/index/README content that is automatically generated.

Regenerate the documentation templates by running the command with --reapply-templates flag to the command above. This refreshes the content of:

  • __init__.py in provider's package
  • Provider Commits
  • Provider index for the documentation
  • Provider README file used when publishing package in PyPI
breeze release-management prepare-provider-documentation --include-removed-providers --reapply-templates-only

Build Provider distributions for SVN apache upload

Those packages might get promoted to "final" packages by just renaming the files, so internally they should keep the final version number without the rc suffix, even if they are rc1/rc2/... candidates.

They also need to be signed and have checksum files. You can generate the checksum/signature files by running the "dev/sign.sh" script (assuming you have the right PGP key set-up for signing). The script generates corresponding .asc and .sha512 files for each file to sign. note: sign script uses libassuan and gnupg if you don't have them installed run:

MacOS:

brew install libassuan
brew install gnupg

Linux (Debian/Ubuntu):

sudo apt-get install libassuan-dev gnupg

Verify your GPG signing key is ready

Before you spend 10+ minutes building artifacts only to discover that signing fails, run these checks once:

# 1. The apache.org key has a secret signing subkey available locally.
gpg --list-secret-keys apache.org

# 2. Signing actually works (exits 0, writes a .asc, verifies cleanly).
echo test > /tmp/sign-check && \
    gpg --yes --armor --local-user apache.org \
        --output /tmp/sign-check.asc --detach-sig /tmp/sign-check && \
    gpg --verify /tmp/sign-check.asc /tmp/sign-check && \
    rm -f /tmp/sign-check /tmp/sign-check.asc && \
    echo "GPG signing OK"

# 3. The fingerprint of your signing (sub)key appears in the Airflow KEYS file.
#    Without this, PMC verifiers cannot validate the release.
FINGERPRINT=$(gpg --list-keys --with-colons apache.org | awk -F: '/^fpr:/ {print $10; exit}')
curl -fsS https://dist.apache.org/repos/dist/release/airflow/KEYS | \
    grep -q "${FINGERPRINT}" && echo "Key ${FINGERPRINT} is in KEYS" || \
    echo "MISSING: add your key to KEYS before releasing"

If any of these fail, fix them before the build step. For first-time release managers, adding your key to the KEYS file is a separate PR against https://dist.apache.org/repos/dist/release/airflow/ (SVN).

sign.sh defaults to SIGN_WITH=apache.org. If your apache.org uid resolves to multiple keys (rare), set SIGN_WITH explicitly to the fingerprint of the key you want to use.

Build and sign the source and convenience packages

  • Cleanup dist folder:
export RELEASE_DATE=$(date "+%Y-%m-%d%n")
export AIRFLOW_REPO_ROOT=$(pwd -P)
rm -rf ${AIRFLOW_REPO_ROOT}/dist/*
  • Release candidate packages:

These instructions assume the standard remote naming convention (upstreamapache/airflow, origin → your fork — see contributing-docs/10_working_with_git.rst). Set tags for the providers in the repo.

echo "Tagging with providers/${RELEASE_DATE}"
git tag -s providers/${RELEASE_DATE} -m "Tag providers for ${RELEASE_DATE}" --force
git push upstream providers/${RELEASE_DATE}
breeze release-management prepare-provider-distributions  --include-removed-providers --distribution-format both
breeze release-management prepare-tarball --tarball-type apache_airflow_providers --version "${RELEASE_DATE}"

The prepare-*-distributions commands should produce the reproducible .whl, .tar.gz packages in the dist folder. The prepare-tarball command should produce reproducible -source.tar.gz tarball of sources.

if you only build few packages, run:

echo "Tagging with providers/${RELEASE_DATE}"
git tag -s providers/${RELEASE_DATE} -m "Tag providers for ${RELEASE_DATE}" --force
git push upstream providers/${RELEASE_DATE}
breeze release-management prepare-provider-distributions --include-removed-providers --distribution-format both PACKAGE PACKAGE ....
breeze release-management prepare-tarball --tarball-type apache_airflow_providers --version "${RELEASE_DATE}"

In case you want to also release a pre-installed provider that is in not-ready state (i.e. when you want to release it before you switch their state to ready), you need to pass --include-not-ready-providers flag to the command above.

  • Sign all your packages
pushd dist
../dev/sign.sh *
popd

If you see Library not loaded error it means that you are missing libassuan and gnupg. check above steps to install them.

Commit the source packages to Apache SVN repo

  • Push the artifacts to ASF dev dist repo
# First clone the repo if you do not have it
cd ..
[ -d asf-dist ] || svn checkout --depth=immediates https://dist.apache.org/repos/dist asf-dist
svn update --set-depth=infinity asf-dist/dev/airflow

# Go to provider's folder in asf-dist
cd asf-dist/dev/airflow/providers

# Create a new folder for the release.
mkdir -p ${RELEASE_DATE}

# Move the artifacts to svn folder
mv ${AIRFLOW_REPO_ROOT}/dist/* "${RELEASE_DATE}"

# Add and commit
svn add ${RELEASE_DATE}
svn commit -m "Add artifacts for Airflow Providers ${RELEASE_DATE}"

cd "$AIRFLOW_REPO_ROOT"

Verify that the files are available in the ${RELEASE_DATE} folder under providers

You should see only providers that you are about to release. If you are seeing others there is an issue. You can remove the redundant provider files manually with:

cd asf-dist/dev/airflow/providers
cd ${RELEASE_DATE}
svn rm file_name  // repeat that for every file
svn commit -m "delete old providers"
cd "$AIRFLOW_REPO_ROOT"

Publish the Regular distributions to PyPI (release candidates)

In order to publish release candidate to PyPI you just need to build and release packages. The packages should however contain the rcN suffix in the version file name but not internally in the package, so you need to use --version-suffix switch to prepare those packages. Note that these are different packages than the ones used for SVN upload though they should be generated from the same sources.

  • Generate the packages with the rc1 version (specify the version suffix with PyPI switch). Note that you should clean up dist folder before generating the packages, so you will only have the right packages there.
cd "$AIRFLOW_REPO_ROOT"
rm -rf ${AIRFLOW_REPO_ROOT}/dist/*

breeze release-management prepare-provider-distributions  --include-removed-providers \
 --version-suffix rc1 --distribution-format both

If you only build few packages, run:

breeze release-management prepare-provider-distributions \
--version-suffix rc1 --distribution-format both PACKAGE PACKAGE ....

Alternatively, if you have set the environment variable: DISTRIBUTIONS_LIST above, just run the command:

breeze release-management prepare-provider-distributions --version-suffix rc1

Or using --distributions-list argument:

breeze release-management prepare-provider-distributions --distributions-list "PACKAGE1 PACKAGE2" --version-suffix rc1

In case some packages already had rc1 suffix prepared and released, and they still need to be released, they will have automatically appropriate rcN suffix added to them. The suffix will be increased for each release candidate and checked if tag has been already created for that release candidate. If yes, the suffix will be increased until the tag is not found.

  • Verify the artifacts that would be uploaded:
twine check ${AIRFLOW_REPO_ROOT}/dist/*
  • Configure a short-lived PyPI token for this upload only. Until Trusted Publishing is deployed for the Airflow provider distributions on PyPI, the recommended practice is:

    1. Log in to https://pypi.org and create an API token right before the upload step. Scope caveat: you would ideally create a project-scoped token, but PyPI only allows project-scoped tokens for projects you already own/maintain on that account. Most Airflow release managers do not have per-project owner rights on every provider being released, so in practice you will need to create an account-wide ("all projects") token. That is acceptable only if you treat it as single-use and delete it immediately after the upload (step 4 below). Never keep an all-projects token on disk longer than the upload itself.
    2. Put it in ~/.pypirc (or export as TWINE_USERNAME=__token__ TWINE_PASSWORD=pypi-...).
    3. Run the upload (below).
    4. Immediately delete the token from the PyPI web UI after the upload completes. Do not keep long-lived release-manager tokens on disk.

    This is a defence-in-depth practice: the RM machine becomes a one-time release vehicle, not a persistent point of compromise.

  • Upload the package to PyPI:

twine upload -r pypi ${AIRFLOW_REPO_ROOT}/dist/*
  • Confirm that the packages are available under the links printed and look good.

  • Save these links for later, you'll need to paste them in the email you'll send to dev@airflow.apache.org

Push the RC tags

Earlier, we pushed the date tag, now that the RC(s) are ready we can push the tags for them.

breeze release-management tag-providers

Prepare documentation in Staging

Documentation is an essential part of the product and should be made available to users. In our cases, documentation for the released versions is published in the staging S3 bucket, and the site is kept in a separate repository - apache/airflow-site, but the documentation source code and build tools are available in the apache/airflow repository, so you need to run several workflows to publish the documentation. More details about it can be found in Docs README showing the architecture and workflows including manual workflows for emergency cases.

You usually use the breeze command to publish the documentation. The command does the following:

  1. Triggers Publish Docs to S3.
  2. Triggers workflow in apache/airflow-site to refresh
  3. Triggers S3 to GitHub Sync
  breeze workflow-run publish-docs --ref providers/${RELEASE_DATE} --site-env staging all-providers

Or if you just want to publish a few selected providers, you can run:

  breeze workflow-run publish-docs --ref providers/${RELEASE_DATE} --site-env staging PACKAGE1 PACKAGE2 ..

Warning

Make sure to NOT close the terminal while the command is running and to keep your computer from sleep. This command will run several (3!) workflows from your terminal and this is important to keep it running until completion.

There is also a manual way of running the workflows (see at the end of the document, this should normally not be needed unless there is some problem with workflow automation above)

Prepare issue in GitHub to keep status of testing

Create a GitHub issue with the content generated via manual execution of the command below. You will use link to that issue in the next step.

cd "${AIRFLOW_REPO_ROOT}"

breeze release-management generate-issue-content-providers --only-available-in-dist

By default the command will attempt to retrieve the GitHub token used to authenticate with GH and tackle rate limiting from locally run gh auth token command output - but if you do not have gh installed, you can generate such token in GitHub Interface and pass it to the command manually. When you use breeze release-management generate-issue-content-providers --help - you will see the link that you will be able to click to generate such token.

cd "${AIRFLOW_REPO_ROOT}"

breeze release-management generate-issue-content-providers --only-available-in-dist --github-token TOKEN

Sometimes, when there are big PRs implemented across many providers, you want to filter them out from the issue content. When there are many of the same PRs/issues they create a noise in the issue and not add value, usually those PRs and issues are about package preparation mechanism so they are tested well outside regular package testing.

The command will exclude automatically PRs that are commented out, but sometimes there are issues you want to exclude additionally.

You can optionally pass list of such PR to be excluded from the issue with --excluded-pr-list. This might limit the scope of verification. Some providers might disappear from the list and list of authors that will be pinged in the generated issue.

You can repeat that and regenerate the issue content until you are happy with the generated issue

cd "${AIRFLOW_REPO_ROOT}"

breeze release-management generate-issue-content-providers --only-available-in-dist --github-token TOKEN \
    --excluded-pr-list PR_NUMBER1,PR_NUMBER2

It's also OK to manually modify the content of such generated issue before actually creating the issue. There is a comment generated with NOTE TO RELEASE MANAGER about this in the issue content. Hit Preview button on "create issue" screen before creating it to verify how it will look like for the contributors.

By default, the command will ask whether to create the issue. You can answer Yes and it will attempt to create the issue with gh tool. This might not always be possible because often our issues create too long URL to create an issue, in which case you will have to copy&paste the issue title and content manually to the issue created with "New Issue" in GitHub,

Prepare voting email for Providers release candidate

Make sure the packages are in https://dist.apache.org/repos/dist/dev/airflow/providers/

Send out a vote to the dev@airflow.apache.org mailing list. Here you can prepare text of the email.

export VOTE_DURATION_IN_HOURS=72
export IS_SHORTEN_VOTE=$([ $VOTE_DURATION_IN_HOURS -ge 72 ] && echo "false" || echo "true")
export SHORTEN_VOTE_TEXT="This is a shortened ($VOTE_DURATION_IN_HOURS hours vote) as agreed by policy set it https://lists.apache.org/thread/cv194w1fqqykrhswhmm54zy9gnnv6kgm"
if [[ "${OSTYPE}" == *darwin* ]]; then
  export VOTE_END_TIME=$(date -u -v "+${VOTE_DURATION_IN_HOURS}H" -v "+10M" +'%Y-%m-%d %H:%M')
else  # Linux
  export VOTE_END_TIME=$(date --utc -d "now + $VOTE_DURATION_IN_HOURS hours + 10 minutes" +'%Y-%m-%d %H:%M')
fi
export RELEASE_MANAGER_NAME="TODO:RELEASE_MANAGER_NAME"
export GITHUB_ISSUE_LINK="TODO:ISSUE_LINK"

subject:

cat <<EOF
$([ $VOTE_DURATION_IN_HOURS -ge 72 ] && echo "[VOTE]" || echo "[ACCELERATED VOTE]") Airflow Providers, release preparation date ${RELEASE_DATE}
EOF
cat <<EOF
Hey all,

I have just cut the new wave Airflow Providers packages with release preparation date ${RELEASE_DATE}. This email is calling a vote on the release,
which will last for $VOTE_DURATION_IN_HOURS hours - which means that it will end on $VOTE_END_TIME UTC and until 3 binding +1 votes have been received.
$([ "$IS_SHORTEN_VOTE" = "true" ] && echo "${SHORTEN_VOTE_TEXT}" || echo "")

Consider this my (binding) +1.

<ADD ANY HIGH-LEVEL DESCRIPTION OF THE CHANGES HERE!>

Airflow Providers are available at:
https://dist.apache.org/repos/dist/dev/airflow/providers/${RELEASE_DATE}

*apache-airflow-providers-${RELEASE_DATE}-source.tar.gz* is the full source tarball of airflow repo - snapshot taken at the moment of provider's release.

*apache-airflow-providers-<PROVIDER>-*.tar.gz* are the convenience python "sdist" distributions that we publish in PyPI

*apache_airflow_providers_<PROVIDER>-*.whl are the convenience Python "wheel" distributions that we publish in PyPI.

The test procedure for PMC members is described in
https://github.com/apache/airflow/blob/main/dev/README_RELEASE_PROVIDERS.md#verify-the-release-candidate-by-pmc-members

The test procedure for and Contributors who would like to test this RC is described in:
https://github.com/apache/airflow/blob/main/dev/README_RELEASE_PROVIDERS.md#verify-the-release-candidate-by-contributors


Public keys are available at:
https://dist.apache.org/repos/dist/release/airflow/KEYS

Please vote accordingly:

[ ] +1 approve
[ ] +0 no opinion
[ ] -1 disapprove with the reason

Only votes from PMC members are binding, but members of the community are
encouraged to test the release and vote with "(non-binding)".

Please note that the version number excludes the 'rcX' string.
This will allow us to rename the artifact without modifying
the artifact checksums when we actually release it.

The status of testing the providers by the community is kept here:
$GITHUB_ISSUE_LINK

The issue is also the easiest way to see important PRs included in the RC candidates.
Detailed changelog for the providers will be published in the documentation after the
RC candidates are released.

You can find the RC packages in PyPI following these links:

<PASTE TWINE UPLOAD LINKS HERE. SORT THEM BEFORE!>

Cheers,
$RELEASE_MANAGER_NAME

EOF

Due to the nature of packages, not all packages have to be released as convenience packages in the final release. During the voting process the voting PMC members might decide to exclude certain packages from the release if some critical problems have been found in some packages.

Please modify the message above accordingly to clearly exclude those packages.

Note, For RC2/3 you may refer to shorten vote period as agreed in mailing list thread.

Release verification

Verify the release candidate by PMC members

Set expected release tag (the same as announced in the vote email):

export RELEASE_DATE=2025-11-03

Go to the directory where you have airflow checked out and set AIRFLOW_REPO_ROOT variable

export AIRFLOW_REPO_ROOT=$(pwd -P)

SVN check

The files should be present in Airflow dist

The following files should be present (6 files):

  • .tar.gz + .asc + .sha512 (one set of files per provider)
  • -py3-none-any.whl + .asc + .sha512 (one set of files per provider)

As a PMC member, you should be able to clone the SVN repository:

cd ..
[ -d asf-dist ] || svn checkout --depth=immediates https://dist.apache.org/repos/dist asf-dist
svn update --set-depth=infinity asf-dist/dev/airflow

Or update it if you already checked it out:

svn update --set-depth=infinity asf-dist/dev/airflow

Set environment variables: PATH_TO_AIRFLOW_SVN to the root of folder where you have providers and RELEASE_DATE to the release date you are verifying.

cd asf-dist/dev/airflow
export PATH_TO_AIRFLOW_SVN=$(pwd -P)

Optionally you can use the breeze release-management check-release-files command to verify that all expected files are present in SVN. This command will produce a Dockerfile.pmc which may help with verifying installation of the packages.

Once you have cloned/updated the SVN repository, copy the PyPi URLs shared in the email to a file called packages.txt in the $AIRFLOW_REPO_ROOT/dev directory.

cd "$AIRFLOW_REPO_ROOT"
# Copy packages.txt extracted from the mail sent by the release manager here
breeze release-management check-release-files providers --release-date "${RELEASE_DATE}" --packages-file ./dev/packages.txt --path-to-airflow-svn "${PATH_TO_AIRFLOW_SVN}"

After the above command completes you can build Dockerfile.pmc to trigger an installation of each provider package and verify the correct versions are installed:

docker build -f Dockerfile.pmc --tag local/airflow .
docker run --rm --entrypoint "airflow" local/airflow info
docker image rm local/airflow

Reproducible package builds checks

For Provider distributions we introduced a reproducible build mechanism - which means that whoever wants to use sources of Airflow from the release tag, can reproducibly build the same "wheel" and "sdist" packages as the release manager and they will be byte-by-byte identical, which makes them easy to verify - if they came from the same sources. This build is only done using released dependencies from PyPI and source code in our repository - no other binary dependencies are used during the build process and if the packages produced are byte-by-byte identical with the one we create from tagged sources it means that the build has a verified provenance.

How to verify it:

  1. Change directory where your airflow sources are checked out
cd "$AIRFLOW_REPO_ROOT"
  1. Check out the providers/YYYY-MM-DD tag:
cd "$AIRFLOW_REPO_ROOT"
git fetch upstream --tags
git checkout providers/${RELEASE_DATE}
  1. Remove all the packages you have in dist folder
rm -rf dist/*
  1. Build the packages using checked out sources
breeze release-management prepare-provider-distributions --include-removed-providers --distribution-format both
breeze release-management prepare-tarball --tarball-type apache_airflow_providers --version "${RELEASE_DATE}"
  1. Switch to the folder where you checked out the SVN dev files
cd ${PATH_TO_AIRFLOW_SVN}/providers/${RELEASE_DATE}
  1. Compare the packages in SVN to the ones you just built
for i in *.tar.gz *.whl
do
  echo -n "$i:"; diff $i ${AIRFLOW_REPO_ROOT}/dist/$i && echo "No diff found"
done

You should see output similar to:

apache_airflow_providers_amazon-8.12.0.tar.gz:No diff found
apache_airflow_providers_apache_impala-1.2.1.tar.gz:No diff found
apache_airflow_providers_atlassian_jira-2.3.0.tar.gz:No diff found
apache_airflow_providers_cncf_kubernetes-7.10.0.tar.gz:No diff found
apache_airflow_providers_common_io-1.1.0.tar.gz:No diff found
apache_airflow_providers_common_sql-1.8.1.tar.gz:No diff found
apache_airflow_providers_databricks-5.0.1.tar.gz:No diff found
apache_airflow_providers_dbt_cloud-3.4.1.tar.gz:No diff found
apache_airflow_providers_docker-3.8.2.tar.gz:No diff found
apache_airflow_providers_elasticsearch-5.2.0.tar.gz:No diff found
apache_airflow_providers_google-10.12.0.tar.gz:No diff found
apache_airflow_providers_microsoft_azure-8.3.0.tar.gz:No diff found
apache_airflow_providers_odbc-4.2.0.tar.gz:No diff found
apache_airflow_providers_openai-1.0.1.tar.gz:No diff found
apache_airflow_providers_opsgenie-5.3.0.tar.gz:No diff found
apache_airflow_providers_papermill-3.5.0.tar.gz:No diff found
apache_airflow_providers_redis-3.4.1.tar.gz:No diff found
apache_airflow_providers_snowflake-5.1.2.tar.gz:No diff found
apache_airflow_providers_trino-5.4.1.tar.gz:No diff found
apache_airflow_providers_amazon-8.12.0-py3-none-any.whl:No diff found
apache_airflow_providers_apache_impala-1.2.1-py3-none-any.whl:No diff found
apache_airflow_providers_atlassian_jira-2.3.0-py3-none-any.whl:No diff found
apache_airflow_providers_cncf_kubernetes-7.10.0-py3-none-any.whl:No diff found
apache_airflow_providers_common_io-1.1.0-py3-none-any.whl:No diff found
apache_airflow_providers_common_sql-1.8.1-py3-none-any.whl:No diff found
apache_airflow_providers_databricks-5.0.1-py3-none-any.whl:No diff found
apache_airflow_providers_dbt_cloud-3.4.1-py3-none-any.whl:No diff found
apache_airflow_providers_docker-3.8.2-py3-none-any.whl:No diff found
apache_airflow_providers_elasticsearch-5.2.0-py3-none-any.whl:No diff found
apache_airflow_providers_google-10.12.0-py3-none-any.whl:No diff found
apache_airflow_providers_microsoft_azure-8.3.0-py3-none-any.whl:No diff found
apache_airflow_providers_odbc-4.2.0-py3-none-any.whl:No diff found
apache_airflow_providers_openai-1.0.1-py3-none-any.whl:No diff found
apache_airflow_providers_opsgenie-5.3.0-py3-none-any.whl:No diff found
apache_airflow_providers_papermill-3.5.0-py3-none-any.whl:No diff found
apache_airflow_providers_redis-3.4.1-py3-none-any.whl:No diff found
apache_airflow_providers_snowflake-5.1.2-py3-none-any.whl:No diff found
apache_airflow_providers_trino-5.4.1-py3-none-any.whl:No diff found

Licence check

This can be done with the Apache RAT tool.

Download the latest jar from https://creadur.apache.org/rat/download_rat.cgi (unpack the binary, the jar is inside)

You can run this command to do it for you (including checksum verification for your own security):

# Checksum value is taken from https://downloads.apache.org/creadur/apache-rat-0.18/apache-rat-0.18-bin.tar.gz.sha512
wget -q https://archive.apache.org/dist/creadur/apache-rat-0.18/apache-rat-0.18-bin.tar.gz -O /tmp/apache-rat-0.18-bin.tar.gz
echo "315b16536526838237c42b5e6b613d29adc77e25a6e44a866b2b7f8b162e03d3629d49c9faea86ceb864a36b2c42838b8ce43d6f2db544e961f2259e242748f4  /tmp/apache-rat-0.18-bin.tar.gz" | sha512sum -c -
tar -xzf /tmp/apache-rat-0.18-bin.tar.gz -C /tmp

Unpack the release source archive (the <package + version>-source.tar.gz file) to a folder

rm -rf /tmp/apache-airflow-providers-src && mkdir -p /tmp/apache-airflow-providers-src && tar -xzf ${PATH_TO_AIRFLOW_SVN}/providers/${RELEASE_DATE}/apache_airflow_providers-*-source.tar.gz --strip-components 1 -C /tmp/apache-airflow-providers-src

Run the check:

cp ${AIRFLOW_REPO_ROOT}/.rat-excludes /tmp/apache-airflow-providers-src/.rat-excludes
java -jar /tmp/apache-rat-0.18/apache-rat-0.18.jar --input-exclude-file /tmp/apache-airflow-providers-src/.rat-excludes /tmp/apache-airflow-providers-src/ | grep -E "! |INFO: "

You should see no files reported as Unknown or with wrong licence and summary of the check similar to:

INFO: Apache Creadur RAT 0.18 (Apache Software Foundation)
INFO: Excluding patterns: .git-blame-ignore-revs, .github/*, .git ...
INFO: Excluding MISC collection.
INFO: Excluding HIDDEN_DIR collection.
SLF4J(W): No SLF4J providers were found.
SLF4J(W): Defaulting to no-operation (NOP) logger implementation
SLF4J(W): See https://www.slf4j.org/codes.html#noProviders for further details.
INFO: RAT summary:
INFO:   Approved:  15615
INFO:   Archives:  2
INFO:   Binaries:  813
INFO:   Document types:  5
INFO:   Ignored:  2392
INFO:   License categories:  2
INFO:   License names:  2
INFO:   Notices:  216
INFO:   Standards:  15609
INFO:   Unapproved:  0
INFO:   Unknown:  0

There should be no files reported as Unknown or Unapproved. The files that are unknown or unapproved should be shown with a line starting with !.

For example:

! Unapproved:         1    A count of unapproved licenses.
! /CODE_OF_CONDUCT.md

Signature check

Make sure you have imported into your GPG the PGP key of the person signing the release. You can find the valid keys in KEYS.

Download the KEYS file from the above link and save it locally.

You can import the whole KEYS file into gpg by running the following command:

wget https://dist.apache.org/repos/dist/release/airflow/KEYS
gpg --import KEYS

You can also import the keys individually from a keyserver. The below one uses Kaxil's key and retrieves it from the default GPG keyserver OpenPGP.org:

gpg --keyserver keys.openpgp.org --receive-keys CDE15C6E4D3A8EC4ECF4BA4B6674E08AD7DE406F

You should choose to import the key when asked.

Note that by being default, the OpenPGP server tends to be overloaded often and might respond with errors or timeouts. Many of the release managers also uploaded their keys to the GNUPG.net keyserver, and you can retrieve it from there.

gpg --keyserver keys.gnupg.net --receive-keys CDE15C6E4D3A8EC4ECF4BA4B6674E08AD7DE406F

Once you have the keys, the signatures can be verified by running this:

cd ${PATH_TO_AIRFLOW_SVN}/providers/${RELEASE_DATE}
for i in *.asc
do
   echo -e "Checking $i\n"; gpg --verify $i
done

This should produce results similar to the below. The "Good signature from ..." is indication that the signatures are correct. Do not worry about the "not certified with a trusted signature" warning. Most of the certificates used by release managers are self-signed, and that's why you get this warning. By importing the key either from the server in the previous step or from the KEYS page, you know that this is a valid key already. To suppress the warning you may edit the key's trust level by running gpg --edit-key <key id> trust and entering 5 to assign trust level ultimate.

Checking apache_airflow_providers-2025-11-03-source.tar.gz.asc

gpg: assuming signed data in 'apache_airflow_providers-2025-11-03-source.tar.gz'
gpg: Signature made Mon 03 Nov 2025 09:30:25 CET
gpg:                using EDDSA key 0B63993D3B84EE3E2EBB984B2A454DAA1E6A437B
gpg:                issuer "eladkal@apache.org"
gpg: Good signature from "Elad Kalif <eladkal@apache.org>" [unknown]
gpg: WARNING: The key's User ID is not certified with a trusted signature!
gpg:          There is no indication that the signature belongs to the owner.
Primary key fingerprint: 0B63 993D 3B84 EE3E 2EBB  984B 2A45 4DAA 1E6A 437B
Checking apache_airflow_providers_edge3-1.4.1-py3-none-any.whl.asc

gpg: assuming signed data in 'apache_airflow_providers_edge3-1.4.1-py3-none-any.whl'
gpg: Signature made Mon 03 Nov 2025 09:30:24 CET
gpg:                using EDDSA key 0B63993D3B84EE3E2EBB984B2A454DAA1E6A437B
gpg:                issuer "eladkal@apache.org"
gpg: Good signature from "Elad Kalif <eladkal@apache.org>" [unknown]
gpg: WARNING: The key's User ID is not certified with a trusted signature!
gpg:          There is no indication that the signature belongs to the owner.
Primary key fingerprint: 0B63 993D 3B84 EE3E 2EBB  984B 2A45 4DAA 1E6A 437B
Checking apache_airflow_providers_edge3-1.4.1.tar.gz.asc

gpg: assuming signed data in 'apache_airflow_providers_edge3-1.4.1.tar.gz'
gpg: Signature made Mon 03 Nov 2025 09:30:24 CET
gpg:                using EDDSA key 0B63993D3B84EE3E2EBB984B2A454DAA1E6A437B
gpg:                issuer "eladkal@apache.org"
gpg: Good signature from "Elad Kalif <eladkal@apache.org>" [unknown]
gpg: WARNING: The key's User ID is not certified with a trusted signature!
gpg:          There is no indication that the signature belongs to the owner.
Primary key fingerprint: 0B63 993D 3B84 EE3E 2EBB  984B 2A45 4DAA 1E6A 437B
⌁0% [jarek:~/code/asf-dist/dev/airflow/providers] dist ‡ for i in *.sha512

SHA512 check

Run this:

for i in *.sha512
do
    echo "Checking $i"; shasum -a 512 `basename $i .sha512 ` | diff - $i
done

You should get output similar to:

Checking apache_airflow_providers-2025-11-03-source.tar.gz.sha512
Checking apache_airflow_providers_edge3-1.4.1-py3-none-any.whl.sha512
Checking apache_airflow_providers_edge3-1.4.1.tar.gz.sha512
...

Verify the release candidate by Contributors

This can be done (and we encourage to) by any of the Contributors. In fact, it's best if the actual users of Apache Airflow test it in their own staging/test installations. Each release candidate is available on PyPI apart from SVN packages, so everyone should be able to install the release candidate version.

Breeze allows you to easily install and run pre-release candidates by following simple instructions described in Manually testing release candidate packages

But you can use any of the installation methods you prefer (you can even install it via the binary wheels downloaded from the SVN).

For values for ==rc, refer to the PR/email from Release Manager. For example, in order to test keycloak provider, PR-60496 lists ""Provider keycloak: 0.5.0rc1". Hence use "apache-airflow-providers-keycloak==0.5.0rc1" in below steps.

Installing in your local virtualenv

You have to make sure you have Airflow 3* installed in your PIP virtualenv (the version you want to install providers with).

pip install apache-airflow-providers-<provider>==<VERSION>rc<X>

Installing with Breeze

You can use any Airflow 3.X.Y version, like 3.2.0, to install specific version for testing, using breeze.

breeze start-airflow --use-airflow-version 3.1.3 --python 3.10 --backend postgres \
    --load-example-dags --load-default-connections

After you are in Breeze:

pip install apache-airflow-providers-<provider>==<VERSION>rc<X>

NOTE! After installing the provider package, restart the Airflow components so the new provider is loaded. If you started Breeze with breeze start-airflow, in the terminal multiplexer (mprocs or tmux) use the keyboard shortcuts to stop and then start each managed component: scheduler, api_server, triggerer, and dag_processor.

Building your own docker image

If you prefer to build your own image, you can also use the official image and PyPI packages to test Provider distributions. This is especially helpful when you want to test integrations, but you need to install additional tools. Below is an example Dockerfile, which installs providers for Google. Please note, these version numbers are arbitrary. You'll need to substitute the proper version numbers when running this yourself.

FROM apache/airflow:3.1.3
RUN pip install  --user apache-airflow-providers-google==18.1.0.rc1

USER ${AIRFLOW_UID}

To build an image build and run a shell, run:

docker build . --tag my-image:0.0.1
docker run  -ti \
    --rm \
    -v "$PWD/data:/opt/airflow/" \
    -v "$PWD/keys/:/keys/" \
    -p 8080:8080 \
    -e AIRFLOW__CORE__LOAD_EXAMPLES=True \
    my-image:0.0.1 bash

Additional Verification

Once you install and run Airflow, you can perform any verification you see as necessary to check that the Airflow works as you expected.

Publish release

export RELEASE_DATE=2025-11-14
export RELEASE_MANAGER_NAME="RELEASE_MANAGER_NAME"
echo "prepare release date is ${RELEASE_DATE}"

Summarize the voting for the Apache Airflow release

Once the vote has been passed, you will need to send a result vote to dev@airflow.apache.org:

In both subject and message update DATE OF RELEASE, FIRST/LAST NAMES and numbers). In case some providers were excluded, explain why they were excluded and what is the plan for them (otherwise remove the optional part of the message). There are two options for releasing the next RC candidates:

  • They will be released as an ad-hoc release with accelerated vote period on their own (when there are not many changes to other providers in the meantime and when we have a small bugfix for the providers that we want to release quickly.

  • They will be included together with the next wave of releases (our tooling supports automated calculation of RC version for candidates for the next wave of releases that already had earlier RCs.

Email subject:

cat <<EOF
[RESULT][VOTE] Airflow Providers - release preparation date ${RELEASE_DATE}
EOF

Email content:

cat <<EOF
Hello,

Apache Airflow Providers prepared on ${RELEASE_DATE} have been accepted.

3 "+1" binding votes received:
- FIRST LAST NAME (binding)
- FIRST LAST NAME (binding)
- FIRST LAST NAME (binding)

2 "+1" non-binding votes received:
- FIRST LAST NAME
- FIRST LAST NAME

[optional] The providers PROVIDER, PROVIDER have been excluded from the release.
This is due to REASON HERE.
The next RC candidates for those providers will be released [in the next wave
of providers] or [as an ad-hoc release on their own with accelerated vote period].

Vote thread: https://lists.apache.org/thread/cs6mcvpn2lk9w2p4oz43t20z3fg5nl7l

I'll continue with the release process, and the release announcement will follow shortly.

Cheers,
${RELEASE_MANAGER_NAME}
EOF

Publish release to SVN

The best way of doing this is to svn cp between the two repos (this avoids having to upload the binaries again, and gives a clearer history in the svn commit logs.

We also need to archive older releases before copying the new ones Release policy

cd "<ROOT_OF_YOUR_AIRFLOW_REPO>"
# Set AIRFLOW_REPO_ROOT to the path of your git repo
export AIRFLOW_REPO_ROOT="$(pwd -P)"

# Go the folder where you have checked out the release repo from SVN
# Make sure this is direct directory and a symbolic link
# Otherwise 'svn mv' errors out if it is with "E200033: Another process is blocking the working copy database
# Assuming that your `asf-dist` is checked out in the same folder as airflow:
cd ..

export ASF_DIST_PARENT="$(pwd -P)"
# make sure physical path is used, in case original directory is symbolically linked
cd "${ASF_DIST_PARENT}"

# or clone it if it's not done yet
[ -d asf-dist ] || svn checkout --depth=immediates https://dist.apache.org/repos/dist asf-dist
# Update to latest version
svn update --set-depth=infinity asf-dist/dev/airflow asf-dist/release/airflow

SOURCE_DIR="${ASF_DIST_PARENT}/asf-dist/dev/airflow/providers/${RELEASE_DATE}"

# If some packages have been excluded, remove them now
# Check the packages are there (replace <provider> with the name of the provider that you remove)
ls ${SOURCE_DIR}/*<provider>*
# Remove them
svn rm ${SOURCE_DIR}/*<provider>*

# All latest releases are kept in this providers folder without version sub-folder
cd "${ASF_DIST_PARENT}/asf-dist/release/airflow/providers"

# Copy your providers with the target name to dist directory and to SVN
rm -rf "${AIRFLOW_REPO_ROOT}"/dist/*

for file in "${SOURCE_DIR}"/*
do
 base_file=$(basename ${file})
 cp -v "${file}" "${AIRFLOW_REPO_ROOT}/dist/${base_file//rc[0-9]/}"
 svn mv "${file}" "${base_file//rc[0-9]/}"
done

# Check which old packages will be removed using dry run
breeze release-management clean-old-provider-artifacts --directory $(pwd -P) --dry-run

# Remove those packages
breeze release-management clean-old-provider-artifacts --directory $(pwd -P)

# You need to do go to the asf-dist directory in order to commit both dev and release together
cd ${ASF_DIST_PARENT}/asf-dist
# Commit to SVN
svn commit -m "Release Airflow Providers on $(date "+%Y-%m-%d%n"): release preparation date ${RELEASE_DATE}"

Verify that the packages appear in providers and the (always first) apache_airflow_providers-RELEASE_DATE-source.tar.gz file should have the right RELEASE_DATE.

You are expected to see all latest versions of providers. The ones you are about to release (with new version) and the ones that are not part of the current release.

Troubleshoot: In case that while viewing the packages in dist/release you see that a provider has files from current version and release version it probably means that you wanted to exclude the new version of provider from release but didn't remove all providers files as expected in previous step. Since you already commit to SVN you need to recover files from previous version with svn copy (svn merge will not work since you don't have copy of the file locally) for example:

svn copy https://dist.apache.org/repos/dist/release/airflow/providers/apache_airflow_providers_docker-3.4.0-py3-none-any.whl@59404
https://dist.apache.org/repos/dist/release/airflow/providers/apache_airflow_providers_docker-3.4.0-py3-none-any.whl

Where 59404 is the revision we want to copy the file from. Then you can commit again. You can also add -m "undeleted file" to the svn copy to commit in 1 step.

Then remove from svn the files of the new provider version that you wanted to exclude from release. If you had this issue you will need also to make adjustments in the next step to remove the provider from listed in twine check. This is simply by removing the relevant files locally.

Publish the packages to PyPI

By that time the packages should be in your dist folder.

cd "$AIRFLOW_REPO_ROOT"
git checkout providers/${RELEASE_DATE}

example git checkout providers/2025-10-31

Note you probably will see message You are in 'detached HEAD' state. This is expected, the RC tag is most likely behind the main branch.

  • Remove source artifact:
rm dist/apache_airflow_providers-${RELEASE_DATE}-source.tar.gz
  • Verify the artifacts that would be uploaded:
twine check ${AIRFLOW_REPO_ROOT}/dist/*.whl ${AIRFLOW_REPO_ROOT}/dist/*.tar.gz
  • Configure a short-lived PyPI token for this upload only. Until Trusted Publishing is deployed for the Airflow provider distributions on PyPI, the recommended practice is:

    1. Log in to https://pypi.org and create an API token right before the upload step. Scope caveat: you would ideally create a project-scoped token, but PyPI only allows project-scoped tokens for projects you already own/maintain on that account. Most Airflow release managers do not have per-project owner rights on every provider being released, so in practice you will need to create an account-wide ("all projects") token. That is acceptable only if you treat it as single-use and delete it immediately after the upload (step 4 below). Never keep an all-projects token on disk longer than the upload itself.
    2. Put it in ~/.pypirc (or export as TWINE_USERNAME=__token__ TWINE_PASSWORD=pypi-...).
    3. Run the upload (below).
    4. Immediately delete the token from the PyPI web UI after the upload completes. Do not keep long-lived release-manager tokens on disk.

    This is a defence-in-depth practice: the RM machine becomes a one-time release vehicle, not a persistent point of compromise.

  • Upload the package to PyPI:

twine upload -r pypi ${AIRFLOW_REPO_ROOT}/dist/*.whl ${AIRFLOW_REPO_ROOT}/dist/*.tar.gz
  • Verify that the packages are available under the links printed.

Copy links to updated packages, sort it alphabetically and save it on the side. You will need it for the announcement message.

  • Again, confirm that the packages are available under the links printed.

Add the final release tag in git

These instructions assume the standard remote naming convention (upstreamapache/airflow, origin → your fork — see contributing-docs/10_working_with_git.rst). breeze release-management tag-providers auto-detects which of your remotes points at apache/airflow, so it also works for release-manager setups where the apache/airflow clone is configured as origin. Set tags for the providers in the repository.

Sometimes in cases when there is a connectivity issue to GitHub, it might be possible that local tags get created and lead to annoying errors. The default behavior would be to clean such local tags up.

If you want to disable this behavior, set the env CLEAN_LOCAL_TAGS to false.

breeze release-management tag-providers

Publish documentation

Documentation is an essential part of the product and should be made available to users. In our cases, documentation for the released versions is published in the live S3 bucket, and the site is kept in a separate repository - apache/airflow-site, but the documentation source code and build tools are available in the apache/airflow repository, so you need to run several workflows to publish the documentation. More details about it can be found in Docs README showing the architecture and workflows including manual workflows for emergency cases.

You usually use the breeze command to publish the documentation. The command does the following:

  1. Triggers Publish Docs to S3.
  2. Triggers workflow in apache/airflow-site to refresh
  3. Triggers S3 to GitHub Sync

First - unset GITHUB_TOKEN if you have it set, this workflows reads token from your github repository configuration if you login with gh, do it only once - you do not have repeat it afterwards.

unset GITHUB_TOKEN
brew install gh
gh auth login

Run workflows:

  breeze workflow-run publish-docs --ref providers/${RELEASE_DATE} --site-env live all-providers

If you need to exclude some providers from the documentation you need to add --exclude-providers flag with space separated list of excluded providers.

  breeze workflow-run publish-docs --ref providers/${RELEASE_DATE} --site-env live all-providers --exclude-docs "apprise slack"

Or if you just want to publish a few selected providers, you can run:

  breeze workflow-run publish-docs --ref providers/${RELEASE_DATE} --site-env live PACKAGE1 PACKAGE2 ..

Warning

Make sure to NOT close the terminal while the command is running and to keep your computer from sleep. This command will run several (3!) workflows from your terminal and this is important to keep it running until completion.

There is also a manual way of running the workflows (see at the end of the document, this should normally not be needed unless there is some problem with workflow automation above)

Note

The Provider Registry at airflow.apache.org/registry/ is rebuilt automatically as part of the publish-docs-to-s3.yml workflow (it calls registry-build.yml as a post-publish job). The registry extracts metadata from provider.yaml files and PyPI, so it picks up new/updated providers without manual intervention. If you need to rebuild the registry independently, trigger the registry-build.yml workflow manually. See registry/README.md for details.

Update providers metadata

Create PR and open it to be merged:

cd "$AIRFLOW_REPO_ROOT"
git checkout main
git pull upstream main
current_date=$(date '+%Y-%m-%d%n')
branch="update-providers-metadata-${current_date}"
git checkout -b "${branch}"
breeze release-management generate-providers-metadata --refresh-constraints-and-airflow-releases
git add -p .
git commit -m "Update providers metadata ${current_date}"
gh pr create --title "Update providers metadata ${current_date}" --web

Notify developers of release

Notify users@airflow.apache.org (cc'ing dev@airflow.apache.org) that the artifacts have been published.

Subject:

cat <<EOF
[ANNOUNCE] Apache Airflow Providers prepared on ${RELEASE_DATE} are released
EOF

Body:

cat <<EOF
Dear Airflow community,

I'm happy to announce that new versions of Airflow Providers packages prepared on ${RELEASE_DATE} were just released.

Full list of PyPI packages released is added at the end of the message.

The source release, as well as the binary releases, are available here:

https://airflow.apache.org/docs/apache-airflow-providers/installing-from-sources.html

You can install the providers via PyPI: https://airflow.apache.org/docs/apache-airflow-providers/installing-from-pypi.html

The documentation index is available at https://airflow.apache.org/docs/ and documentation for individual provider versions is linked directly from PyPI.

----

Full list of released PyPI packages:

TODO: Paste the list of packages here that you put on the side. Sort them alphabetically.

Cheers,
${RELEASE_MANAGER_NAME}
EOF

Send the same email to announce@apache.org, except change the opening line to Dear community,. It is more reliable to send it via the web ui at https://lists.apache.org/list.html?announce@apache.org (press "c" to compose a new thread)

Note If you choose sending it with your email client make sure the email is set to plain text mode. Trying to send HTML content will result in failure.

Send announcements about security issues fixed in the release

The release manager should review and mark as READY all the security issues fixed in the release. Such issues can be listed under the Next wave of providers milestone in security issues. Go through the list of these issues and check for each of them the fix has been released as part of this release. Then the release manager should announce the issues via the CVE management tool at https://cveprocess.apache.org/.

Once announced, each of the issue should be linked with a 'reference' with tag 'vendor advisory' with the URL to the announcement published automatically by the CVE management tool. Note that the announce@apache.org is moderated, and the link to the email thread will not be published immediately, that's why it is recommended to add the link to users@airflow.apache.org which takes usually few seconds to be published after the CVE tool sends them.

The ASF Security will be notified and will submit to the CVE project and will set the state to 'PUBLIC'.

Announce about the release in social media

NOTE!

As a rule we announce only new providers that were added. If you believe there is a reason to announce in social media for another case consult with PMC members about it.

Example for special cases:

  • an exciting new capability that the community waited for and should have big impact.
  • big number of providers released at once.
  • bumping min airflow version (which is a special case of the above)

Announcement is done from official Apache-Airflow accounts.

Make sure attach the release image generated with Figma to the post. If you don't have access to the account ask a PMC member to post.


Add release data to Apache Committee Report Helper

You should get email about it to your account that should urge you to add it, but in case you don't, you can add it manually: add the release data (version and date) at: https://reporter.apache.org/addrelease.html?airflow

Close the testing status issue

Don't forget to thank the folks who tested and close the issue tracking the testing status.

Thank you everyone. Providers are released.

I invite everyone to help improve providers for the next release, a list of open issues can be found [here](https://github.com/apache/airflow/issues?q=is%3Aopen+is%3Aissue+label%3Aarea%3Aproviders).

Remove Provider distributions scheduled for removal

If there are Provider distributions scheduler for removal, create PR and merge it to remove them.

The following places should be checked:

  • airflow/providers/PROVIDER
  • tests/providers/PROVIDER
  • tests/system/providers/PROVIDER
  • tests/integration/providers/PROVIDER
  • docs/apache-airflow-providers-PROVIDER
  • docs/integration-logos/PROVIDER
  • .github/boring-cyborg.yml
  • airflow/contrib/hooks/__init__.py
  • airflow/contrib/operators/__init__.py
  • airflow/utils/db.py (for default connections)
  • dev/breeze/tests/test_packages.py (remove the providers from removed lists)
  • generated/provider_metadata.json

Run breeze setup regenerate-command-images --force

Update test_get_removed_providers in /dev/breeze/tests/test_packages.py by removing the provider from the list

Misc / Post release Helpers

Those processes are related to the release of Airflow but should be run in exceptional situations.

Fixing released documentation

Sometimes we want to rebuild the documentation with some fixes that were merged in main, for example when there are html layout changes or typo fixes, or formatting issue fixes.

In this case the process is as follows:

  • When you want to re-publish providers-PROVIDER/X.Y.Z docs, create (or pull if already created) providers-PROVIDER/X.Y.Z-docs branch
  • Cherry-pick changes you want to add and push to the main apache/airflow repo
  • Run the publishing workflow.

In case you are releasing latest released version of Provider (which should be most of the cases), run this:

breeze workflow-run publish-docs --site-env live --ref providers-PROVIDER/X.Y.Z-docs \
   --skip-tag-validation \
   PROVIDER

In case you are releasing an older version of Provider, you should skip writing to stable folder

breeze workflow-run publish-docs --site-env live --ref providers-PROVIDER/X.Y.Z-docs \
   --skip-tag-validation \
   --skip-write-to-stable-folder \
   PROVIDER

Similarly you can rebuild all provider docs:

  • When you want to re-publish providers/YYYY-MM-DD docs, create (or pull if already created) providers/YYYY-MM-DD-docs branch
  • Cherry-pick changes you want to add and push to the main apache/airflow repo
  • Run the publishing workflow.

In case you are releasing latest released version of Provider (which should be most of the cases), run this:

breeze workflow-run publish-docs --site-env live --ref providers/YYYY-MM-DD-docs \
   --skip-tag-validation \
   all-providers

In case you are releasing an older version of Provider, you should skip writing to stable folder

breeze workflow-run publish-docs --site-env live --ref providers/YYYY-MM-DD-docs \
   --skip-tag-validation \
   --skip-write-to-stable-folder \
   all-providers