Table of contents
The Provider packages are separate packages (one package per provider) that implement integrations with external services for Airflow in the form of installable Python packages.
The Release Manager prepares packages separately from the main Airflow Release, using breeze
commands and accompanying scripts. This document provides an overview of the command line tools needed to prepare the packages.
NOTE!! When you have problems with any of those commands that run inside breeze
docker image, you can run the command with --debug
flag that will drop you in the shell inside the image and will print the command that you should run.
The prerequisites to release Apache Airflow are described in README.md.
You can read more about the command line tools used to generate the packages in the Provider packages.
Update BASE_PROVIDERS_COMPATIBILITY_CHECKS
in src/airflow_breeze/global_constants.py
to remove the versions of Airflow that are not applicable anymore.
Check if Breeze unit tests in dev/breeze/tests/test_packages.py
need adjustments. This is done by simply searching and replacing old version occurrences with newer one. For example 2.5.0 to 2.6.0
Update minimum airflow version for all packages, you should modify MIN_AIRFLOW_VERSION
in src/airflow_breeze/utils/packages.py
and run the prepare-provider-documentation
command with the --only-min-version-update
flag. This will only update the min version in the __init__.py
files and package documentation without bumping the provider versions.
Note: Sometimes we are releasing a subset of providers and would not want to add the list of these providers to every breeze command we run, specifically: prepare-provider-packages
, build-docs
, publish-docs
, and, add-back-references
. In this case, we can instead export an environment variable: PACKAGE_LIST
, and it will work for every breeze command involved in the release process. The value can also be passed as the --package-list
argument. Follow the steps below to set the environment variable:
export PACKAGE_LIST=PACKAGE1,PACKAGE2
branch="update-min-airflow-version" git checkout -b "${branch}" breeze release-management prepare-provider-documentation --only-min-version-update git add . git commit -m "Bump minimum Airflow version in providers to Airflow 2.6.0" git push --set-upstream origin "${branch}"
Note: that this command will only bump the min airflow versions for those providers that do not have it set to a higher version. You do not have to skip specific providers - run it for all providers it will handle everything automatically.
Note: this step is not part of the release cycle. It should be done independently when the time to update min airflow version has come.
You can release provider packages separately from the main Airflow on an ad-hoc basis, whenever we find that a given provider needs to be released - due to new features or due to bug fixes. You can release each provider package separately, but due to voting and release overhead we try to group releases of provider packages together.
We are using the SEMVER versioning scheme for the provider packages. This is in order to give the users confidence about maintaining backwards compatibility in the new releases of those packages.
Details about maintaining the SEMVER version are going to be discussed and implemented in the related issue
The provider packages can be in one of several states.
not-ready
state is used when the provider has some in-progress changes (usually API changes) that we do not want to release yet as part of the regular release cycle. Providers in this state are excluded from being released as part of the regular release cycle (including documentation building). You can build and prepare such provider when you explicitly specify it as argument of a release command or by passing --include-not-ready-providers
flag in corresponding command. The not-ready
providers are treated as regular providers when it comes to running tests and preparing and releasing packages in CI
- as we want to make sure they are properly releasable any time and we want them to contribute to dependencies and we want to test them. Also in case of preinstalled providers, the not-ready
providers are contributing their dependencies rather than the provider package to requirements of Airflow.ready
state is the usual state of the provider that is released in the regular release cycle (including the documentation, package building and publishing). This is the state most providers are in.suspended
state is used when we have a good reason to suspend such provider, following the devlist discussion and vote or “lazy consensus”. The process of suspension is described in Provider's docs. The suspended
providers are excluded from being released as part of the regular release cycle (including documentation building) but also they do not contribute dependencies to the CI image and their tests are not run in CI process. You can build and prepare such provider when you explicitly specify it as argument of a release command or by passing --include-suspended-providers
flag in corresponding command (but it might or might not work at any time as the provider release commands are not regularly run on CI for the suspended providers). The suspended
providers are not released as part of the regular release cycle.removed
state is a temporary state after the provider has been voted (or agreed in “lazy consensus”) to be removed and it is only used for exactly one release cycle - in order to produce the final version of the package - identical to the previous version with the exception of the removal notice. The process of removal is described in Provider's docs. The removed
providers are included in the regular release cycle (including documentation building) because the --include-removed-providers
flag is passed to commands that release manager runs (see below). The difference between suspended
and removed
providers is that additional information is added to their documentation about the provider not being maintained any more by the community.This graph shows the possible transitions between the states:
graph TD; new[/new/] new -- Add to the code -->ready; ready ready-- Mark as not ready -->not-ready; not-ready-- Mark as ready -->ready; ready-- Suspend -->suspended; suspended-- Resume -->ready; ready-- Mark as removed -->removed; suspended-- Mark as removed -->removed; gone[\gone\] removed -- Remove from the code --> gone;
Sometimes (rare) we release providers that have dependencies on future version of Airflow - which means that they are released long before they are actually usable and it also means that versions in PyPI should be released with apache-airflow >= x.y.z.dev0
version, such providers should have the .dev0 suffix included in the apache-airflow
dependency specification, only the final release, just before the final Airflow x.y.z release should get it changed to >= x.y.z
. This is a rare case and should be handled with care.
We call such case chicken-egg providers as it's not clear who should be released first - the provider or the Airflow.
First thing that release manager has to do is to change version of the provider to a target version. Each provider has a provider.yaml
file that, among others, stores information about provider versions. When you attempt to release a provider you should update that information based on the changes for the provider, and its CHANGELOG.rst
. It might be that CHANGELOG.rst
already contains the right target version. This will be especially true if some changes in the provider add new features (then minor version is increased) or when the changes introduce backwards-incompatible, breaking change in the provider (then major version is incremented). Committers, when approving and merging changes to the providers, should pay attention that the CHANGELOG.rst
is updated whenever anything other than bugfix is added.
If there are no new features or breaking changes, the release manager should simply increase the patch-level version for the provider.
The new version should be first on the list.
Each of the provider packages contains Release notes in the form of the CHANGELOG.rst
file that is automatically generated from history of the changes and code of the provider. They are stored in the documentation directory. The README.md
file generated during package preparation is not stored anywhere in the repository - it contains however link to the Changelog generated.
When the provider package version has not been updated since the latest version, the release notes are not generated. Release notes are only generated, when the latest version of the package does not yet have a corresponding TAG.
The tags for providers is of the form providers-<PROVIDER_ID>/<VERSION>
for example providers-amazon/1.0.0
. During releasing, the rc*
tags are created (for example providers-amazon/1.0.0rc1
).
Details about maintaining the SEMVER version are going to be discussed and implemented in the related issue
breeze release-management prepare-provider-documentation --include-removed-providers
In case you prepare provider documentation for just a few selected providers, you can run:
breeze release-management prepare-provider-documentation [packages]
In case you want to also release a pre-installed provider that is in not-ready
state (i.e. when you want to release it before you switch their state to ready
), you need to pass --include-not-ready-providers
flag to the command above.
This command will not only prepare documentation but will also help the release manager to review changes implemented in all providers, and determine which of the providers should be released. For each provider details will be printed on what changes were implemented since the last release including links to particular commits.
This should help to determine which version of provider should be released:
It also helps the release manager to update CHANGELOG.rst where high-level overview of the changes should be documented for the providers released. You should iterate and re-generate the same content after any change as many times as you want. The generated files should be added and committed to the repository.
When you want to regenerate the changes before the release and make sure all changelogs are updated, run it in non-interactive mode:
breeze release-management prepare-provider-documentation --include-removed-providers --answer yes
In case you prepare provider documentation for just a few selected providers, you can run:
breeze release-management prepare-provider-documentation --answer yes [packages]
In case you want to also release a pre-installed provider that is in not-ready
state (i.e. when you want to release it before you switch their state to ready
), you need to pass --include-not-ready-providers
flag to the command above.
NOTE!! In case you prepare provider's documentation in a branch different than main, you need to manually specify the base branch via --base-branch
parameter. For example if you try to build a cncf.kubernetes
provider that is build from provider-cncf-kubernetes/v4-4
branch should be prepared like this:
breeze release-management prepare-provider-documentation --include-removed-providers \ --base-branch provider-cncf-kubernetes/v4-4 cncf.kubernetes
This step should only be executed if we want to change template files for the providers - i.e. change security information, commit/index/README content that is automatically generated.
Regenerate the documentation templates by running the command with --reapply-templates
flag to the command above. This refreshes the content of:
__init__.py
in provider's packagebreeze release-management prepare-provider-documentation --include-removed-providers --reapply-templates-only
At this point you should have providers yaml files and changelog updated. You should go over the change log and place changes in their relevant section (breaking change, feature, bugs, etc...) Once finished you should raise a PR : Prepare docs for MM YYYY wave of Providers In the PR we will verify if we want to release a specific package or if the versions chosen are right. Only after PR is merged you should proceed to next steps.
Those packages might get promoted to “final” packages by just renaming the files, so internally they should keep the final version number without the rc suffix, even if they are rc1/rc2/... candidates.
They also need to be signed and have checksum files. You can generate the checksum/signature files by running the “dev/sign.sh” script (assuming you have the right PGP key set-up for signing). The script generates corresponding .asc and .sha512 files for each file to sign.
export AIRFLOW_REPO_ROOT=$(pwd -P) rm -rf ${AIRFLOW_REPO_ROOT}/dist/*
breeze release-management prepare-provider-packages --include-removed-providers --package-format both
if you only build few packages, run:
breeze release-management prepare-provider-packages --include-removed-providers \ --package-format both PACKAGE PACKAGE ....
In case you want to also release a pre-installed provider that is in not-ready
state (i.e. when you want to release it before you switch their state to ready
), you need to pass --include-not-ready-providers
flag to the command above.
pushd dist ../dev/sign.sh * popd
# First clone the repo if you do not have it cd .. [ -d asf-dist ] || svn checkout --depth=immediates https://dist.apache.org/repos/dist asf-dist svn update --set-depth=infinity asf-dist/dev/airflow # Create a new folder for the release. cd asf-dist/dev/airflow/providers # Remove previously released providers svn rm * # Move the artifacts to svn folder mv ${AIRFLOW_REPO_ROOT}/dist/* . # Add and commit svn add * svn commit -m "Add artifacts for Airflow Providers $(date "+%Y-%m-%d%n")" cd ${AIRFLOW_REPO_ROOT}
Verify that the files are available at providers
You should see only providers that you are about to release. If you are seeing others there is an issue. You can remove the redundant provider files manually with:
svn rm file_name // repeate that for every file svn commit -m "delete old providers"
In order to publish release candidate to PyPI you just need to build and release packages. The packages should however contain the rcN suffix in the version file name but not internally in the package, so you need to use --version-suffix-for-pypi
switch to prepare those packages. Note that these are different packages than the ones used for SVN upload though they should be generated from the same sources.
rm -rf ${AIRFLOW_REPO_ROOT}/dist/* breeze release-management prepare-provider-packages --include-removed-providers \ --version-suffix-for-pypi rc1 --package-format both
If you only build few packages, run:
breeze release-management prepare-provider-packages \ --version-suffix-for-pypi rc1 --package-format both PACKAGE PACKAGE ....
Alternatively, if you have set the environment variable: PACKAGE_LIST
above, just run the command:
breeze release-management prepare-provider-packages
Or using --package-list
argument:
breeze release-management prepare-provider-packages --package-list PACKAGE1,PACKAGE2
In case some packages already had rc1 suffix prepared and released, and they still need to be released, they will have automatically appropriate rcN suffix added to them. The suffix will be increased for each release candidate and checked if tag has been already created for that release candidate. If yes, the suffix will be increased until the tag is not found.
twine check ${AIRFLOW_REPO_ROOT}/dist/*
twine upload -r pypi ${AIRFLOW_REPO_ROOT}/dist/*
Assume that your remote for apache repository is called apache
you should now set tags for the providers in the repo.
Sometimes in cases when there is a connectivity issue to Github, it might be possible that local tags get created and lead to annoying errors. The default behaviour would be to clean such local tags up.
If you want to disable this behaviour, set the env CLEAN_LOCAL_TAGS to false.
breeze release-management tag-providers
Documentation is an essential part of the product and should be made available to users. In our cases, documentation for the released versions is published in a separate repository - apache/airflow-site
, but the documentation source code and build tools are available in the apache/airflow
repository, so you have to coordinate between the two repositories to be able to build the documentation.
Documentation for providers can be found in the /docs/apache-airflow-providers
directory and the /docs/apache-airflow-providers-*/
directory. The first directory contains the package contents lists and should be updated every time a new version of provider packages is released.
AIRFLOW_SITE_DIRECTORY
.git clone https://github.com/apache/airflow-site.git airflow-site cd airflow-site export AIRFLOW_SITE_DIRECTORY="$(pwd -P)"
Note if this is not the first time you clone the repo make sure main branch is rebased:
cd "${AIRFLOW_SITE_DIRECTORY}" git checkout main git pull --rebase
cd "${AIRFLOW_REPO_ROOT}" breeze build-docs --clean-build apache-airflow-providers all-providers --include-removed-providers
Usually when we release packages we also build documentation for the “documentation-only” packages. This means that unless we release just few selected packages or if we need to deliberately skip some packages we should release documentation for all provider packages and the above command is the one to use.
If we want to just release some providers you can release them using package names:
cd "${AIRFLOW_REPO_ROOT}" breeze build-docs apache-airflow-providers cncf.kubernetes sftp --clean-build
Alternatively, if you have set the environment variable: PACKAGE_LIST
above, just run the command:
cd "${AIRFLOW_REPO_ROOT}" breeze build-docs --clean-build
Or using --package-list
argument:
breeze build-docs --package-list PACKAGE1,PACKAGE2
./docs/start_doc_server.sh
If you encounter error like:
airflow git:(main) ./docs/start_doc_server.sh ./docs/start_doc_server.sh: line 22: cd: /Users/eladkal/PycharmProjects/airflow/docs/_build: No such file or directory
That probably means that the doc folder is empty thus it can not build the doc server. This indicates that previous step of building the docs did not work.
airflow-site
repositoryAll providers (including overriding documentation for doc-only changes) - note that publishing is way faster on multi-cpu machines when you are publishing multiple providers:
cd "${AIRFLOW_REPO_ROOT}" breeze release-management publish-docs apache-airflow-providers all-providers --include-removed-providers \ --override-versioned --run-in-parallel breeze release-management add-back-references all-providers
If you have providers as list of provider ids because you just released them you can build them with
cd "${AIRFLOW_REPO_ROOT}" breeze release-management publish-docs amazon apache.beam google .... breeze release-management add-back-references all-providers
Alternatively, if you have set the environment variable: PACKAGE_LIST
above, just run the command:
breeze release-management publish-docs breeze release-management add-back-references all-providers
Or using --package-list
argument:
breeze release-management publish-docs --package-list PACKAGE1,PACKAGE2 breeze release-management add-back-references all-providers
Review the state of removed, suspended, new packages in the docs index:
If you publish a new package, you must add it to the list of packages in the index.
If there are changes to suspension or removal status of a package you must move it appropriate section.
Create the commit and push changes.
cd "${AIRFLOW_SITE_DIRECTORY}" branch="add-documentation-$(date "+%Y-%m-%d%n")" git checkout -b "${branch}" git add . git commit -m "Add documentation for packages - $(date "+%Y-%m-%d%n")" git push --set-upstream origin "${branch}"
Create a GitHub issue with the content generated via manual execution of the command below. You will use link to that issue in the next step.
cd "${AIRFLOW_REPO_ROOT}" breeze release-management generate-issue-content-providers --only-available-in-dist
GitHub API uses rate limiting that is based on the public IP address of client if you do not authenticate with GitHub, so when you retrieve bigger number of PRs or when you are behind NAT and share your public IP address with many other Anonymous GitHub API users, issue retrieval will be halted and your API calls might slow down to a crawl, you will need then a GITHUB_TOKEN set as your environment variable or pass the token as --github-token
option in the script.
cd "${AIRFLOW_REPO_ROOT}" breeze release-management generate-issue-content-providers --only-available-in-dist --github-token TOKEN
or
cd "${AIRFLOW_REPO_ROOT}" export GITHUB_TOKEN=TOKEN breeze release-management generate-issue-content-providers --only-available-in-dist
You can generate the token by following this link. Since it is easy to generate such token, by following the link, it is recommended to generate a new token for each release and delete it once you've generated the issue.
If you see in the output that some of the PRs are just “noise” (i.e. there is no need to verify them as they are misc/documentation kind of changes that have no impact on the actual installation of the provider or the code of the provider, can optionally pass list of PR to be excluded from the issue with --excluded-pr-list
. This might limit the scope of verification. Some providers might disappear from the list and list of authors that will be pinged in the generated issue.
You can repeat that and regenerate the issue content until you are happy with the generated issue.
cd "${AIRFLOW_REPO_ROOT}" breeze release-management generate-issue-content-providers --only-available-in-dist --github-token TOKEN \ --excluded-pr-list PR_NUMBER1,PR_NUMBER2
It's also OK to manually modify the content of such generated issue before actually creating the issue. There is a comment generated with NOTE TO RELEASE MANAGER about this in the issue content. Hit Preview button on “create issue” screen before creating it to verify how it will look like for the contributors.
Make sure the packages are in https://dist.apache.org/repos/dist/dev/airflow/providers/
Send out a vote to the dev@airflow.apache.org mailing list. Here you can prepare text of the email.
subject:
cat <<EOF [VOTE] Airflow Providers prepared on $(date "+%B %d, %Y") EOF
cat <<EOF Hey all, I have just cut the new wave Airflow Providers packages. This email is calling a vote on the release, which will last for 72 hours - which means that it will end on $(TZ=UTC date -v+3d "+%B %d, %Y %H:%M %p" ) UTC and until 3 binding +1 votes have been received. Consider this my (binding) +1. <ADD ANY HIGH-LEVEL DESCRIPTION OF THE CHANGES HERE!> Airflow Providers are available at: https://dist.apache.org/repos/dist/dev/airflow/providers/ *apache-airflow-providers-<PROVIDER>-*.tar.gz* are the binary Python "sdist" release - they are also official "sources" for the provider packages. *apache_airflow_providers_<PROVIDER>-*.whl are the binary Python "wheel" release. The test procedure for PMC members is described in https://github.com/apache/airflow/blob/main/dev/README_RELEASE_PROVIDER_PACKAGES.md#verify-the-release-candidate-by-pmc-members The test procedure for and Contributors who would like to test this RC is described in: https://github.com/apache/airflow/blob/main/dev/README_RELEASE_PROVIDER_PACKAGES.md#verify-the-release-candidate-by-contributors Public keys are available at: https://dist.apache.org/repos/dist/release/airflow/KEYS Please vote accordingly: [ ] +1 approve [ ] +0 no opinion [ ] -1 disapprove with the reason Only votes from PMC members are binding, but members of the community are encouraged to test the release and vote with "(non-binding)". Please note that the version number excludes the 'rcX' string. This will allow us to rename the artifact without modifying the artifact checksums when we actually release. The status of testing the providers by the community is kept here: <TODO COPY LINK TO THE ISSUE CREATED> The issue is also the easiest way to see important PRs included in the RC candidates. Detailed changelog for the providers will be published in the documentation after the RC candidates are released. You can find the RC packages in PyPI following these links: <PASTE TWINE UPLOAD LINKS HERE. SORT THEM BEFORE!> Cheers, <TODO: Your Name> EOF
Due to the nature of packages, not all packages have to be released as convenience packages in the final release. During the voting process the voting PMCs might decide to exclude certain packages from the release if some critical problems have been found in some packages.
Please modify the message above accordingly to clearly exclude those packages.
Note, For RC2/3 you may refer to shorten vote period as agreed in mailing list thread.
The files should be present in Airflow dist
The following files should be present (6 files):
As a PMC you should be able to clone the SVN repository:
svn co https://dist.apache.org/repos/dist/dev/airflow/
Or update it if you already checked it out:
svn update .
Optionally you can use the check_files.py
script to verify that all expected files are present in SVN. This script will produce a Dockerfile.pmc
which may help with verifying installation of the packages.
# Copy the list of packages (pypi urls) into `packages.txt` then run: python check_files.py providers -p {PATH_TO_SVN}
After the above script completes you can build Dockerfile.pmc
to trigger an installation of each provider package and verify the correct versions are installed:
docker build -f Dockerfile.pmc --tag local/airflow . docker run --rm --entrypoint "airflow" local/airflow info docker image rm local/airflow
For provider packages we introduced a reproducible build mechanism - which means that whoever wants to use sources of Airflow from the release tag, can reproducibly build the same “wheel” and “sdist” packages as the release manager and they will be byte-by-byte identical, which makes them easy to verify - if they came from the same sources. This build is only done using released dependencies from PyPI and source code in our repository - no other binary dependencies are used during the build process and if the packages produced are byte-by-byte identical with the one we create from tagged sources it means that the build has a verified provenance.
How to verify it:
cd "${AIRFLOW_REPO_ROOT}"
git checkout tags/providers-amazon-1.0.0rc1
rm -rf dist/*
breeze release-management prepare-provider-packages --include-removed-providers --package-format both
cd {PATH_TO_SVN} cd airflow/providers
for i in *.tar.gz *.whl do echo -n "$i:"; diff $i ${AIRFLOW_REPO_ROOT}/dist/$i && echo "No diff found" done
You should see output similar to:
apache_airflow_providers_amazon-8.12.0.tar.gz:No diff found apache_airflow_providers_apache_impala-1.2.1.tar.gz:No diff found apache_airflow_providers_atlassian_jira-2.3.0.tar.gz:No diff found apache_airflow_providers_cncf_kubernetes-7.10.0.tar.gz:No diff found apache_airflow_providers_common_io-1.1.0.tar.gz:No diff found apache_airflow_providers_common_sql-1.8.1.tar.gz:No diff found apache_airflow_providers_databricks-5.0.1.tar.gz:No diff found apache_airflow_providers_dbt_cloud-3.4.1.tar.gz:No diff found apache_airflow_providers_docker-3.8.2.tar.gz:No diff found apache_airflow_providers_elasticsearch-5.2.0.tar.gz:No diff found apache_airflow_providers_google-10.12.0.tar.gz:No diff found apache_airflow_providers_microsoft_azure-8.3.0.tar.gz:No diff found apache_airflow_providers_odbc-4.2.0.tar.gz:No diff found apache_airflow_providers_openai-1.0.1.tar.gz:No diff found apache_airflow_providers_opsgenie-5.3.0.tar.gz:No diff found apache_airflow_providers_papermill-3.5.0.tar.gz:No diff found apache_airflow_providers_redis-3.4.1.tar.gz:No diff found apache_airflow_providers_snowflake-5.1.2.tar.gz:No diff found apache_airflow_providers_trino-5.4.1.tar.gz:No diff found apache_airflow_providers_amazon-8.12.0-py3-none-any.whl:No diff found apache_airflow_providers_apache_impala-1.2.1-py3-none-any.whl:No diff found apache_airflow_providers_atlassian_jira-2.3.0-py3-none-any.whl:No diff found apache_airflow_providers_cncf_kubernetes-7.10.0-py3-none-any.whl:No diff found apache_airflow_providers_common_io-1.1.0-py3-none-any.whl:No diff found apache_airflow_providers_common_sql-1.8.1-py3-none-any.whl:No diff found apache_airflow_providers_databricks-5.0.1-py3-none-any.whl:No diff found apache_airflow_providers_dbt_cloud-3.4.1-py3-none-any.whl:No diff found apache_airflow_providers_docker-3.8.2-py3-none-any.whl:No diff found apache_airflow_providers_elasticsearch-5.2.0-py3-none-any.whl:No diff found apache_airflow_providers_google-10.12.0-py3-none-any.whl:No diff found apache_airflow_providers_microsoft_azure-8.3.0-py3-none-any.whl:No diff found apache_airflow_providers_odbc-4.2.0-py3-none-any.whl:No diff found apache_airflow_providers_openai-1.0.1-py3-none-any.whl:No diff found apache_airflow_providers_opsgenie-5.3.0-py3-none-any.whl:No diff found apache_airflow_providers_papermill-3.5.0-py3-none-any.whl:No diff found apache_airflow_providers_redis-3.4.1-py3-none-any.whl:No diff found apache_airflow_providers_snowflake-5.1.2-py3-none-any.whl:No diff found apache_airflow_providers_trino-5.4.1-py3-none-any.whl:No diff found
This can be done with the Apache RAT tool.
<package + version>.tar.gz
file) to a folderjava -jar ../../apache-rat-0.13/apache-rat-0.13.jar -E .rat-excludes -d .
where .rat-excludes
is the file in the root of Airflow source code.
Make sure you have imported into your GPG the PGP key of the person signing the release. You can find the valid keys in KEYS.
You can import the whole KEYS file:
gpg --import KEYS
You can also import the keys individually from a keyserver. The below one uses Kaxil's key and retrieves it from the default GPG keyserver OpenPGP.org:
gpg --keyserver keys.openpgp.org --receive-keys CDE15C6E4D3A8EC4ECF4BA4B6674E08AD7DE406F
You should choose to import the key when asked.
Note that by being default, the OpenPGP server tends to be overloaded often and might respond with errors or timeouts. Many of the release managers also uploaded their keys to the GNUPG.net keyserver, and you can retrieve it from there.
gpg --keyserver keys.gnupg.net --receive-keys CDE15C6E4D3A8EC4ECF4BA4B6674E08AD7DE406F
Once you have the keys, the signatures can be verified by running this:
for i in *.asc do echo -e "Checking $i\n"; gpg --verify $i done
This should produce results similar to the below. The “Good signature from ...” is indication that the signatures are correct. Do not worry about the “not certified with a trusted signature” warning. Most of the certificates used by release managers are self-signed, and that‘s why you get this warning. By importing the key either from the server in the previous step or from the KEYS page, you know that this is a valid key already. To suppress the warning you may edit the key’s trust level by running gpg --edit-key <key id> trust
and entering 5
to assign trust level ultimate
.
Checking apache-airflow-2.0.2rc4.tar.gz.asc gpg: assuming signed data in 'apache-airflow-2.0.2rc4.tar.gz' gpg: Signature made sob, 22 sie 2020, 20:28:28 CEST gpg: using RSA key 12717556040EEF2EEAF1B9C275FCCD0A25FA0E4B gpg: Good signature from "Kaxil Naik <kaxilnaik@gmail.com>" [unknown] gpg: WARNING: This key is not certified with a trusted signature! gpg: There is no indication that the signature belongs to the owner. Primary key fingerprint: 1271 7556 040E EF2E EAF1 B9C2 75FC CD0A 25FA 0E4B Checking apache_airflow-2.0.2rc4-py2.py3-none-any.whl.asc gpg: assuming signed data in 'apache_airflow-2.0.2rc4-py2.py3-none-any.whl' gpg: Signature made sob, 22 sie 2020, 20:28:31 CEST gpg: using RSA key 12717556040EEF2EEAF1B9C275FCCD0A25FA0E4B gpg: Good signature from "Kaxil Naik <kaxilnaik@gmail.com>" [unknown] gpg: WARNING: This key is not certified with a trusted signature! gpg: There is no indication that the signature belongs to the owner. Primary key fingerprint: 1271 7556 040E EF2E EAF1 B9C2 75FC CD0A 25FA 0E4B Checking apache-airflow-2.0.2rc4-source.tar.gz.asc gpg: assuming signed data in 'apache-airflow-2.0.2rc4-source.tar.gz' gpg: Signature made sob, 22 sie 2020, 20:28:25 CEST gpg: using RSA key 12717556040EEF2EEAF1B9C275FCCD0A25FA0E4B gpg: Good signature from "Kaxil Naik <kaxilnaik@gmail.com>" [unknown] gpg: WARNING: This key is not certified with a trusted signature! gpg: There is no indication that the signature belongs to the owner. Primary key fingerprint: 1271 7556 040E EF2E EAF1 B9C2 75FC CD0A 25FA 0E4B
Run this:
for i in *.sha512 do echo "Checking $i"; shasum -a 512 `basename $i .sha512 ` | diff - $i done
You should get output similar to:
Checking apache-airflow-providers-google-1.0.0rc1.tar.gz.sha512 Checking apache_airflow-providers-google-1.0.0rc1-py3-none-any.whl.sha512
This can be done (and we encourage to) by any of the Contributors. In fact, it's best if the actual users of Apache Airflow test it in their own staging/test installations. Each release candidate is available on PyPI apart from SVN packages, so everyone should be able to install the release candidate version.
Breeze allows you to easily install and run pre-release candidates by following simple instructions described in Manually testing release candidate packages
But you can use any of the installation methods you prefer (you can even install it via the binary wheels downloaded from the SVN).
You have to make sure you have Airflow 2* installed in your PIP virtualenv (the version you want to install providers with).
pip install apache-airflow-providers-<provider>==<VERSION>rc<X>
breeze start-airflow --use-airflow-version 2.2.4 --python 3.8 --backend postgres \ --load-example-dags --load-default-connections
After you are in Breeze:
pip install apache-airflow-providers-<provider>==<VERSION>rc<X>
NOTE! You should Ctrl-C
and restart the connections to restart airflow components and make sure new provider packages is used.
If you prefer to build your own image, you can also use the official image and PyPI packages to test provider packages. This is especially helpful when you want to test integrations, but you need to install additional tools. Below is an example Dockerfile, which installs providers for Google/
FROM apache/airflow:2.2.3 RUN pip install --user apache-airflow-providers-google==2.2.2.rc1 USER ${AIRFLOW_UID}
To build an image build and run a shell, run:
docker build . --tag my-image:0.0.1 docker run -ti \ --rm \ -v "$PWD/data:/opt/airflow/" \ -v "$PWD/keys/:/keys/" \ -p 8080:8080 \ -e AIRFLOW__CORE__LOAD_EXAMPLES=True \ my-image:0.0.1 bash
Once you install and run Airflow, you can perform any verification you see as necessary to check that the Airflow works as you expected.
Once the vote has been passed, you will need to send a result vote to dev@airflow.apache.org:
In both subject and message update DATE OF RELEASE, FIRST/LAST NAMES and numbers). In case some providers were excluded, explain why they were excluded and what is the plan for them (otherwise remove the optional part of the message). There are two options for releasing the next RC candidates:
They will be released as an ad-hoc release with accelerated vote period on their own (when there are not many changes to other providers in the meantime and when we have a small bugfix for the providers that we want to release quickly.
They will be included together with the next wave of releases (our tooling supports automated calculation of RC version for candidates for the next wave of releases that already had earlier RCs.
Email subject:
[RESULT][VOTE] Airflow Providers - release of DATE OF RELEASE
Email content:
Hello, Apache Airflow Providers prepared on DATE OF RELEASE have been accepted. 3 "+1" binding votes received: - FIRST LAST NAME (binding) - FIRST LAST NAME (binding) - FIRST LAST NAME (binding) 2 "+1" non-binding votes received: - FIRST LAST NAME - FIRST LAST NAME [optional] The providers PROVIDER, PROVIDER have been excluded from the release. This is due to REASON HERE. The next RC candidates for those providers will be released [in the next wave of providers] or [as an ad-hoc release on their own with accelerated vote period]. Vote thread: https://lists.apache.org/thread/cs6mcvpn2lk9w2p4oz43t20z3fg5nl7l I'll continue with the release process, and the release announcement will follow shortly. Cheers, <your name>
The best way of doing this is to svn cp between the two repos (this avoids having to upload the binaries again, and gives a clearer history in the svn commit logs.
We also need to archive older releases before copying the new ones Release policy
cd "<ROOT_OF_YOUR_AIRFLOW_REPO>" # Set AIRFLOW_REPO_ROOT to the path of your git repo export AIRFLOW_REPO_ROOT="$(pwd -P)" # Go the folder where you have checked out the release repo from SVN # Make sure this is direct directory and a symbolic link # Otherwise 'svn mv' errors out if it is with "E200033: Another process is blocking the working copy database cd "<ROOT_WHERE_YOUR_ASF_DIST_IS_CREATED>" export ASF_DIST_PARENT="$(pwd -P)" # make sure physical path is used, in case original directory is symbolically linked cd "${ASF_DIST_PARENT}" # or clone it if it's not done yet [ -d asf-dist ] || svn checkout --depth=immediates https://dist.apache.org/repos/dist asf-dist # Update to latest version svn update --set-depth=infinity asf-dist/dev/airflow asf-dist/release/airflow SOURCE_DIR="${ASF_DIST_PARENT}/asf-dist/dev/airflow/providers" # If some packages have been excluded, remove them now # Check the packages are there (replace <provider> with the name of the provider that you remove) ls ${SOURCE_DIR}/*<provider>* # Remove them svn rm ${SOURCE_DIR}/*<provider>* # Create providers folder if it does not exist # All latest releases are kept in this one folder without version sub-folder cd "${ASF_DIST_PARENT}/asf-dist/release/airflow" mkdir -pv providers cd providers # Copy your providers with the target name to dist directory and to SVN rm -rf "${AIRFLOW_REPO_ROOT}"/dist/* for file in "${SOURCE_DIR}"/* do base_file=$(basename ${file}) cp -v "${file}" "${AIRFLOW_REPO_ROOT}/dist/${base_file//rc[0-9]/}" svn mv "${file}" "${base_file//rc[0-9]/}" done # Check which old packages will be removed using dry run breeze release-management clean-old-provider-artifacts --directory $(pwd -P) --dry-run # Remove those packages breeze release-management clean-old-provider-artifacts --directory $(pwd -P) # You need to do go to the asf-dist directory in order to commit both dev and release together cd ${ASF_DIST_PARENT}/asf-dist # Commit to SVN svn commit -m "Release Airflow Providers on $(date "+%Y-%m-%d%n")"
Verify that the packages appear in providers
You are expected to see all latest versions of providers. The ones you are about to release (with new version) and the ones that are not part of the current release.
Troubleshoot: In case that while viewing the packages in dist/release you see that a provider has files from current version and release version it probably means that you wanted to exclude the new version of provider from release but didn‘t remove all providers files as expected in previous step. Since you already commit to SVN you need to recover files from previous version with svn copy (svn merge will not work since you don’t have copy of the file locally) for example:
svn copy https://dist.apache.org/repos/dist/release/airflow/providers/apache_airflow_providers_docker-3.4.0-py3-none-any.whl@59404 https://dist.apache.org/repos/dist/release/airflow/providers/apache_airflow_providers_docker-3.4.0-py3-none-any.whl
Where 59404
is the revision we want to copy the file from. Then you can commit again. You can also add -m "undeleted file"
to the svn copy
to commit in 1 step.
Then remove from svn the files of the new provider version that you wanted to exclude from release. If you had this issue you will need also to make adjustments in the next step to remove the provider from listed in twine check. This is simply by removing the relevant files locally.
By that time the packages should be in your dist folder.
cd ${AIRFLOW_REPO_ROOT} git checkout <ONE_OF_THE_RC_TAGS_FOR_ONE_OF_THE_RELEASED_PROVIDERS>
example git checkout providers-amazon/7.0.0rc2
Note you probably will see message You are in 'detached HEAD' state.
This is expected, the RC tag is most likely behind the main branch.
twine check ${AIRFLOW_REPO_ROOT}/dist/*.whl ${AIRFLOW_REPO_ROOT}/dist/*.tar.gz
twine upload -r pypi ${AIRFLOW_REPO_ROOT}/dist/*.whl ${AIRFLOW_REPO_ROOT}/dist/*.tar.gz
Copy links to updated packages, sort it alphabetically and save it on the side. You will need it for the announcement message.
Merge the PR that you prepared before with the documentation.
If you decided to remove some packages from the release make sure to do amend the commit in this way:
docs-archive/apache-airflow-providers-<PROVIDER>
stable.txt
to the previous versiondocs-archive/apache-airflow-providers-<PROVIDER>
folderdocs-archive/apache-airflow-providers/core-extensions/index.html
(2 places)docs-archive/apache-airflow-providers/core-extensions/connections.html
(2 places)docs-archive/apache-airflow-providers/core-extensions/extra-links.html
(2 places)docs-archive/apache-airflow-providers/core-extensions/packages-ref.html
(5 places)Assume that your remote for apache repository is called apache
you should now set tags for the providers in the repo.
Sometimes in cases when there is a connectivity issue to Github, it might be possible that local tags get created and lead to annoying errors. The default behaviour would be to clean such local tags up.
If you want to disable this behaviour, set the env CLEAN_LOCAL_TAGS to false.
breeze release-management tag-providers
cd ${AIRFLOW_REPO_ROOT} git checkout main git pull current_date=$(date '+%Y-%m-%d%n') branch="update-providers-metadata-${current_date}" git checkout -b "${branch}" breeze release-management generate-providers-metadata --refresh-constraints git add -p . git commit -m "Update providers metadata ${current_date}" git push --set-upstream origin "${branch}"
Create PR and get it merged
Notify users@airflow.apache.org (cc'ing dev@airflow.apache.org) that the artifacts have been published.
Subject:
[ANNOUNCE] Apache Airflow Providers prepared on DATE OF RELEASE are released
Body:
Dear Airflow community, I'm happy to announce that new versions of Airflow Providers packages prepared on DATE OF RELEASE were just released. Full list of PyPI packages released is added at the end of the message. The source release, as well as the binary releases, are available here: https://airflow.apache.org/docs/apache-airflow-providers/installing-from-sources You can install the providers via PyPI: https://airflow.apache.org/docs/apache-airflow-providers/installing-from-pypi The documentation is available at https://airflow.apache.org/docs/ and linked from the PyPI packages. ---- Full list of released PyPI packages: TODO: Paste the list of packages here that you put on the side. Sort them alphabetically. Cheers, <your name>
Send the same email to announce@apache.org, except change the opening line to Dear community,
. It is more reliable to send it via the web ui at https://lists.apache.org/list.html?announce@apache.org (press “c” to compose a new thread)
Note If you choose sending it with your email client make sure the email is set to plain text mode. Trying to send HTML content will result in failure.
The release manager should review and mark as READY all the security issues fixed in the release. Such issues are marked as affecting < <JUST_RELEASED_VERSION>
in the CVE management tool at https://cveprocess.apache.org/. Then the release manager should announced the issues via the tool.
Once announced, each of the issue should be linked with a ‘reference’ with tag ‘vendor advisory’ with the URL to the announcement published automatically by the CVE management tool. Note that the announce@apache.org is moderated, and the link to the email thread will not be published immediately, that's why it is recommended to add the link to users@airflow.apache.org which takes usually few seconds to be published after the CVE tool sends them.
The ASF Security will be notified and will submit to the CVE project and will set the state to ‘PUBLIC’.
NOTE!
As a rule we announce only new providers that were added. If you believe there is a reason to announce in social media for another case consult with PMCs about it.
Example for special cases:
Announcement is done from official Apache-Airflow accounts.
Make sure attach the release image generated with Figma to the post. If you don't have access to the account ask PMC to post.
Add the release data (version and date) at: https://reporter.apache.org/addrelease.html?airflow
Don't forget to thank the folks who tested and close the issue tracking the testing status.
Thank you everyone. Providers are released. I invite everyone to help improve providers for the next release, a list of open issues can be found [here](https://github.com/apache/airflow/issues?q=is%3Aopen+is%3Aissue+label%3Aarea%3Aproviders).
If there are provider packages scheduler for removal, create PR and merge it to remove them.
The following places should be checked:
airflow/providers/PROVIDER
tests/providers/PROVIDER
tests/system/providers/PROVIDER
tests/integration/providers/PROVIDER
docs/apache-airflow-providers-PROVIDER
docs/integration-logos/PROVIDER
.github/boring-cyborg.yml
airflow/contrib/hooks/__init__.py
airflow/contrib/operators/__init__.py
airflow/utils/db.py
(for default connections)dev/breeze/tests/test_packages.py
(remove the providers from removed
lists)generated/provider_metadata.json
Run breeze setup regenerate-command-images --force