Table of contents

You can find the prerequisites to release Apache Airflow in README.md.

Prepare the Apache Airflow Package RC

Build RC artifacts

The Release Candidate artifacts we vote upon should be the exact ones we vote against, without any modification other than renaming – i.e. the contents of the files must be the same between voted release candidate and final release. Because of this the version in the built artifacts that will become the official Apache releases must not include the rcN suffix.

  • Set environment variables

    # Set Version
    export VERSION=2.1.2rc3
    export VERSION_SUFFIX=rc3
    export VERSION_CONSTRAINT_BRANCH=2-1
    export VERSION_WITHOUT_RC=${VERSION/rc?/}
    
    # Set AIRFLOW_REPO_ROOT to the path of your git repo
    export AIRFLOW_REPO_ROOT=$(pwd)
    
    
    # Example after cloning
    git clone https://github.com/apache/airflow.git airflow
    cd airflow
    export AIRFLOW_REPO_ROOT=$(pwd)
    
  • Set your version to 2.0.N in setup.py (without the RC tag)

  • Commit the version change.

  • Tag your release

    git tag -s ${VERSION}
    
  • Clean the checkout: the sdist step below will

    git clean -fxd
    
  • Tarball the repo

    mkdir dist
    git archive --format=tar.gz ${VERSION} \
        --prefix=apache-airflow-${VERSION_WITHOUT_RC}/ \
        -o dist/apache-airflow-${VERSION_WITHOUT_RC}-source.tar.gz
    
  • Generate SHA512/ASC (If you have not generated a key yet, generate it by following instructions on http://www.apache.org/dev/openpgp.html#key-gen-generate-key)

    ./breeze prepare-airflow-packages --package-format both
    ${AIRFLOW_REPO_ROOT}/dev/sign.sh dist/*
    
  • Tag & Push the latest constraints files. This pushes constraints with rc suffix (this is expected)!

    git checkout constraints-${VERSION_CONSTRAINT_BRANCH}
    git tag -s "constraints-${VERSION}"
    git push origin "constraints-${VERSION}"
    
  • Push the artifacts to ASF dev dist repo

    # First clone the repo
    svn checkout https://dist.apache.org/repos/dist/dev/airflow airflow-dev
    
    # Create new folder for the release
    cd airflow-dev
    svn mkdir ${VERSION}
    
    # Move the artifacts to svn folder & commit
    mv ${AIRFLOW_REPO_ROOT}/dist/* ${VERSION}/
    cd ${VERSION}
    svn add *
    svn commit -m "Add artifacts for Airflow ${VERSION}"
    

Manually prepare production Docker Image

./scripts/ci/tools/prepare_prod_docker_images.sh ${VERSION}

This will wipe Breeze cache and docker-context-files in order to make sure the build is “clean”. It also performs image verification before pushing the images.

[\Optional] Create new release branch

When you just released the X.Y.0 version (first release of new minor version) you need to create release branches: vX-Y-test and vX-Y-stable (for example with 2.1.0rc1 release you need to create v2-1-test and v2-1-stable branches):

# First clone the repo
BRANCH_PREFIX=v2-1
git branch ${BRANCH_PREFIX}-test
git branch ${BRANCH_PREFIX}-stable
git push origin ${BRANCH_PREFIX}-test ${BRANCH_PREFIX}-stable

Search and replace all the vX-Y for previous branches (TODO: we should likely automate this a bit more)

Run script to re-tag images from the main branch to the vX-Y-test branch:

./dev/retag_docker_images.py --source-branch main --target-branch ${BRANCH_PREFIX}-test

Prepare PyPI convenience “snapshot” packages

At this point we have the artefact that we vote on, but as a convenience to developers we also want to publish “snapshots” of the RC builds to PyPI for installing via pip:

To do this we need to

  • Build the package:

    ./breeze prepare-airflow-packages --version-suffix-for-pypi "${VERSION_SUFFIX}" --package-format both
    
  • Verify the artifacts that would be uploaded:

    twine check dist/*
    
  • Upload the package to PyPI's test environment:

    twine upload -r pypitest dist/*
    
  • Verify that the test package looks good by downloading it and installing it into a virtual environment. The package download link is available at: https://test.pypi.org/project/apache-airflow/#files

  • Upload the package to PyPI's production environment: twine upload -r pypi dist/*

  • Again, confirm that the package is available here: https://pypi.python.org/pypi/apache-airflow

It is important to stress that this snapshot should not be named “release”, and it is not supposed to be used by and advertised to the end-users who do not read the devlist.

  • Push Tag for the release candidate

    This step should only be done now and not before, because it triggers an automated build of the production docker image, using the packages that are currently released in PyPI (both airflow and latest provider packages).

    git push origin ${VERSION}
    

Prepare production Docker Image

Production Docker images should be manually prepared and pushed by the release manager.

./scripts/ci/tools/prepare_prod_docker_images.sh ${VERSION}

This will wipe Breeze cache and docker-context-files in order to make sure the build is “clean”. It also performs image verification before pushing the images.

Prepare Vote email on the Apache Airflow release candidate

  • Use the dev/airflow-jira script to generate a list of Airflow JIRAs that were closed in the release.

  • Send out a vote to the dev@airflow.apache.org mailing list:

Subject:

[VOTE] Airflow 2.0.2rc3

Body:

Hey all,

I have cut Airflow 2.0.2 RC3. This email is calling a vote on the release,
which will last for 72 hours. Consider this my (binding) +1.

Airflow 2.0.2 RC3 is available at:
https://dist.apache.org/repos/dist/dev/airflow/2.0.2rc3/

*apache-airflow-2.0.2rc3-source.tar.gz* is a source release that comes
with INSTALL instructions.
*apache-airflow-2.0.2rc3-bin.tar.gz* is the binary Python "sdist" release.

Public keys are available at:
https://dist.apache.org/repos/dist/release/airflow/KEYS

Only votes from PMC members are binding, but the release manager should encourage members of the community
to test the release and vote with "(non-binding)".

The test procedure for PMCs and Contributors who would like to test this RC are described in
https://github.com/apache/airflow/blob/main/dev/README.md#vote-and-verify-the-apache-airflow-release-candidate

Please note that the version number excludes the `rcX` string, so it's now
simply 2.0.2. This will allow us to rename the artifact without modifying
the artifact checksums when we actually release.


Changes since 2.0.2rc2:
*Bugs*:
[AIRFLOW-3732] Fix issue when trying to edit connection in RBAC UI
[AIRFLOW-2866] Fix missing CSRF token head when using RBAC UI (#3804)
...


*Improvements*:
[AIRFLOW-3302] Small CSS fixes (#4140)
[Airflow-2766] Respect shared datetime across tabs
...


*New features*:
[AIRFLOW-2874] Enables FAB's theme support (#3719)
[AIRFLOW-3336] Add new TriggerRule for 0 upstream failures (#4182)
...


*Doc-only Change*:
[AIRFLOW-XXX] Fix BashOperator Docstring (#4052)
[AIRFLOW-3018] Fix Minor issues in Documentation
...

Cheers,
<your name>

Verify the release candidate by PMCs

The PMCs should verify the releases in order to make sure the release is following the Apache Legal Release Policy.

At least 3 (+1) votes should be recorded in accordance to Votes on Package Releases

The legal checks include:

  • checking if the packages are present in the right dist folder on svn
  • verifying if all the sources have correct licences
  • verifying if release manager signed the releases with the right key
  • verifying if all the checksums are valid for the release

SVN check

The files should be present in the sub-folder of Airflow dist

The following files should be present (9 files):

  • -bin-tar.gz + .asc + .sha512
  • -source.tar.gz + .asc + .sha512
  • -.whl + .asc + .sha512

As a PMC you should be able to clone the SVN repository:

svn co https://dist.apache.org/repos/dist/dev/airflow

Or update it if you already checked it out:

svn update .

Optionally you can use check_files.py script to verify that all expected files are present in SVN. This script may help also with verifying installation of the packages.

python check_files.py -v {VERSION} -t airflow -p {PATH_TO_SVN}

Licence check

This can be done with the Apache RAT tool.

  • Download the latest jar from https://creadur.apache.org/rat/download_rat.cgi (unpack the binary, the jar is inside)
  • Unpack the binary (-bin.tar.gz) to a folder
  • Enter the folder and run the check (point to the place where you extracted the .jar)
java -jar ../../apache-rat-0.13/apache-rat-0.13.jar -E .rat-excludes -d .

where .rat-excludes is the file in the root of Airflow source code.

Signature check

Make sure you have the key of person signed imported in your GPG. You can find the valid keys in KEYS.

You can import the whole KEYS file:

gpg --import KEYS

You can also import the keys individually from a keyserver. The below one uses Kaxil's key and retrieves it from the default GPG keyserver OpenPGP.org:

gpg --receive-keys 12717556040EEF2EEAF1B9C275FCCD0A25FA0E4B

You should choose to import the key when asked.

Note that by being default, the OpenPGP server tends to be overloaded often and might respond with errors or timeouts. Many of the release managers also uploaded their keys to the GNUPG.net keyserver, and you can retrieve it from there.

gpg --keyserver keys.gnupg.net --receive-keys 12717556040EEF2EEAF1B9C275FCCD0A25FA0E4B

Once you have the keys, the signatures can be verified by running this:

for i in *.asc
do
   echo "Checking $i"; gpg --verify $i
done

This should produce results similar to the below. The “Good signature from ...” is indication that the signatures are correct. Do not worry about the “not certified with a trusted signature” warning. Most of the certificates used by release managers are self signed, that's why you get this warning. By importing the server in the previous step and importing it via ID from KEYS page, you know that this is a valid Key already.

Checking apache-airflow-2.0.2rc4-bin.tar.gz.asc
gpg: assuming signed data in 'apache-airflow-2.0.2rc4-bin.tar.gz'
gpg: Signature made sob, 22 sie 2020, 20:28:28 CEST
gpg:                using RSA key 12717556040EEF2EEAF1B9C275FCCD0A25FA0E4B
gpg: Good signature from "Kaxil Naik <kaxilnaik@gmail.com>" [unknown]
gpg: WARNING: This key is not certified with a trusted signature!
gpg:          There is no indication that the signature belongs to the owner.
Primary key fingerprint: 1271 7556 040E EF2E EAF1  B9C2 75FC CD0A 25FA 0E4B
Checking apache_airflow-2.0.2rc4-py2.py3-none-any.whl.asc
gpg: assuming signed data in 'apache_airflow-2.0.2rc4-py2.py3-none-any.whl'
gpg: Signature made sob, 22 sie 2020, 20:28:31 CEST
gpg:                using RSA key 12717556040EEF2EEAF1B9C275FCCD0A25FA0E4B
gpg: Good signature from "Kaxil Naik <kaxilnaik@gmail.com>" [unknown]
gpg: WARNING: This key is not certified with a trusted signature!
gpg:          There is no indication that the signature belongs to the owner.
Primary key fingerprint: 1271 7556 040E EF2E EAF1  B9C2 75FC CD0A 25FA 0E4B
Checking apache-airflow-2.0.2rc4-source.tar.gz.asc
gpg: assuming signed data in 'apache-airflow-2.0.2rc4-source.tar.gz'
gpg: Signature made sob, 22 sie 2020, 20:28:25 CEST
gpg:                using RSA key 12717556040EEF2EEAF1B9C275FCCD0A25FA0E4B
gpg: Good signature from "Kaxil Naik <kaxilnaik@gmail.com>" [unknown]
gpg: WARNING: This key is not certified with a trusted signature!
gpg:          There is no indication that the signature belongs to the owner.
Primary key fingerprint: 1271 7556 040E EF2E EAF1  B9C2 75FC CD0A 25FA 0E4B

SHA512 sum check

Run this:

for i in *.sha512
do
    echo "Checking $i"; shasum -a 512 `basename $i .sha512 ` | diff - $i
done

You should get output similar to:

Checking apache-airflow-2.0.2rc4-bin.tar.gz.sha512
Checking apache_airflow-2.0.2rc4-py2.py3-none-any.whl.sha512
Checking apache-airflow-2.0.2rc4-source.tar.gz.sha512

Verify release candidates by Contributors

This can be done (and we encourage to) by any of the Contributors. In fact, it's best if the actual users of Apache Airflow test it in their own staging/test installations. Each release candidate is available on PyPI apart from SVN packages, so everyone should be able to install the release candidate version of Airflow via simply ( is 2.0.2 for example, and is release candidate number 1,2,3,....).

pip install apache-airflow==<VERSION>rc<X>

Optionally it can be followed with constraints

pip install apache-airflow==<VERSION>rc<X> \
  --constraint "https://raw.githubusercontent.com/apache/airflow/constraints-<VERSION>/constraints-3.6.txt"`

Note that the constraints contain python version that you are installing it with.

You can use any of the installation methods you prefer (you can even install it via the binary wheel downloaded from the SVN).

There is also an easy way of installation with Breeze if you have the latest sources of Apache Airflow. Running the following command will use tmux inside breeze, create admin user and run Webserver & Scheduler:

./breeze start-airflow --use-airflow-version <VERSION>rc<X> --python 3.7 --backend postgres

Once you install and run Airflow, you should perform any verification you see as necessary to check that the Airflow works as you expected.

Publish the final Apache Airflow release

Summarize the voting for the Apache Airflow release

Once the vote has been passed, you will need to send a result vote to dev@airflow.apache.org:

Subject:

[RESULT][VOTE] Airflow 2.0.2rc3

Message:

Hello,

Apache Airflow 2.0.2 (based on RC3) has been accepted.

4 “+1” binding votes received:
- Kaxil Naik
- Bolke de Bruin
- Ash Berlin-Taylor
- Tao Feng


4 "+1" non-binding votes received:

- Deng Xiaodong
- Stefan Seelmann
- Joshua Patchus
- Felix Uellendall

Vote thread:
https://lists.apache.org/thread.html/736404ca3d2b2143b296d0910630b9bd0f8b56a0c54e3a05f4c8b5fe@%3Cdev.airflow.apache.org%3E

I'll continue with the release process, and the release announcement will follow shortly.

Cheers,
<your name>

Publish release to SVN

You need to migrate the RC artifacts that passed to this repository: https://dist.apache.org/repos/dist/release/airflow/ (The migration should include renaming the files so that they no longer have the RC number in their filenames.)

The best way of doing this is to svn cp between the two repos (this avoids having to upload the binaries again, and gives a clearer history in the svn commit logs):

# GO to Airflow Sources first
cd <YOUR_AIRFLOW_SOURCES>
export AIRFLOW_SOURCES=$(pwd)

# GO to Checked out DEV repo. Should be checked out before via:
# svn checkout https://dist.apache.org/repos/dist/dev/airflow airflow-dev
cd <YOUR_AIFLOW_DEV_SVN>
svn update
export AIRFLOW_DEV_SVN=$(pwd)

# GO to Checked out RELEASE repo. Should be checked out before via:
# svn checkout https://dist.apache.org/repos/dist/release/airflow airflow-release
cd <YOUR_AIFLOW_RELEASE_SVN>
svn update

export RC=2.0.2rc5
export VERSION=${RC/rc?/}

# Create new folder for the release
cd airflow-release
svn mkdir "${VERSION}"
cd "${VERSION}"

# Move the artifacts to svn folder & commit
for f in ${AIRFLOW_DEV_SVN}/$RC/*; do
    svn cp "$f" "${$(basename $f)/}"
    # Those will be used to upload to PyPI
    cp "$f" "${AIRFLOW_SOURCES}/dist/${$(basename $f)/}"
done
svn commit -m "Release Airflow ${VERSION} from ${RC}"

# Remove old release
# See http://www.apache.org/legal/release-policy.html#when-to-archive
cd ..
export PREVIOUS_VERSION=2.0.2
svn rm "${PREVIOUS_VERSION}"
svn commit -m "Remove old release: ${PREVIOUS_VERSION}"

Verify that the packages appear in airflow

Prepare PyPI “release” packages

At this point we release an official package (they should be copied and renamed from the previously released RC candidates in “${AIRFLOW_SOURCES}/dist”:

  • Verify the artifacts that would be uploaded:

    cd "${AIRFLOW_SOURCES}"
    twine check dist/*
    
  • Upload the package to PyPI's test environment:

    twine upload -r pypitest dist/*
    
  • Verify that the test package looks good by downloading it and installing it into a virtual environment. The package download link is available at: https://test.pypi.org/project/apache-airflow/#files

  • Upload the package to PyPI's production environment:

    twine upload -r pypi dist/*
    
  • Again, confirm that the package is available here: https://pypi.python.org/pypi/apache-airflow

Update CHANGELOG.md

  • Get a diff between the last version and the current version:

    git log 1.8.0..1.9.0 --pretty=oneline
    
  • Update CHANGELOG.md with the details, and commit it.

  • Re-Tag & Push the constraints files with the final release version.

    git checkout constraints-${RC}
    git tag -s "constraints-${VERSION}"
    git push origin "constraints-${VERSION}"
    
  • Push Tag for the final version

    This step should only be done now and not before, because it triggers an automated build of the production docker image, using the packages that are currently released in PyPI (both airflow and latest provider packages).

    git push origin ${VERSION}
    

Manually prepare production Docker Image

./scripts/ci/tools/prepare_prod_docker_images.sh ${VERSION}

This will wipe Breeze cache and docker-context-files in order to make sure the build is “clean”. It also performs image verification before pushing the images.

If this is the newest image released, push the latest image as well.

docker tag "apache/airflow:${VERSION}" "apache/airflow:latest"
docker push "apache/airflow:latest"

Publish documentation

Documentation is an essential part of the product and should be made available to users. In our cases, documentation for the released versions is published in a separate repository - apache/airflow-site, but the documentation source code and build tools are available in the apache/airflow repository, so you have to coordinate between the two repositories to be able to build the documentation.

Documentation for providers can be found in the /docs/apache-airflow directory.

  • First, copy the airflow-site repository and set the environment variable AIRFLOW_SITE_DIRECTORY.

    git clone https://github.com/apache/airflow-site.git airflow-site
    cd airflow-site
    export AIRFLOW_SITE_DIRECTORY="$(pwd)"
    
  • Then you can go to the directory and build the necessary documentation packages

    cd "${AIRFLOW_REPO_ROOT}"
    ./breeze build-docs -- --package-filter apache-airflow --package-filter docker-stack --for-production
    
  • Now you can preview the documentation.

    ./docs/start_doc_server.sh
    
  • Copy the documentation to the airflow-site repository, create commit and push changes.

    ./docs/publish_docs.py --package-filter apache-airflow --package-filter docker-stack
    cd "${AIRFLOW_SITE_DIRECTORY}"
    git commit -m "Add documentation for Apache Airflow ${VERSION}"
    git push
    

Notify developers of release

Subject:

cat <<EOF
Airflow ${VERSION} is released
EOF

Body:

cat <<EOF
Dear Airflow community,

I'm happy to announce that Airflow ${VERSION} was just released.

The source release, as well as the binary "sdist" release, are available
here:

https://dist.apache.org/repos/dist/release/airflow/${VERSION}/

We also made this version available on PyPI for convenience (`pip install apache-airflow`):

https://pypi.python.org/pypi/apache-airflow

The documentation is available on:
https://airflow.apache.org/
https://airflow.apache.org/docs/apache-airflow/${VERSION}/

Find the CHANGELOG here for more details:

https://airflow.apache.org/changelog.html#airflow-1-10-2-2019-01-19

Cheers,
<your name>
EOF

Update Announcements page

Update “Announcements” page at the Official Airflow website