Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Expand
titleFinalize the release

Once the release candidate has been reviewed and approved by the community, the release should be finalized. This involves the final deployment of the release candidate to the release repositories, merging of the website changes, etc.

Deploy Python artifacts to PyPI (Since 1.9)

Release manager should create a PyPI account and ask the PMC add this account to pyflink collaborator list with Maintainer role (The PyPI admin account info can be found here. NOTE, only visible to PMC members) to deploy the Python artifacts to PyPI. The artifacts could be uploaded using twine(https://pypi.org/project/twine/). To install twine, just run:

Code Block
languagetext
pip install --upgrade twine==1.12.0

Download the python artifacts from dist.apache.org and upload it to pypi.org:

Code Block
languagetext
svn checkout https://dist.apache.org/repos/dist/dev/flink/flink-${RELEASE_VERSION}-rc${RC_NUM}
cd flink-${RELEASE_VERSION}-rc${RC_NUM}

# since 1.11
cd pythoncd python

#uploads wheels
for f in *.whl; do twine upload --repository-url https://upload.pypi.org/legacy/ $f $f.asc; done

##upload sincesource 1.13packages
twine upload --repository-url https://upload.pypi.org/legacy/ apache-flink-libraries-${RELEASE_VERSION}.tar.gz apache-flink-libraries-${RELEASE_VERSION}.tar.gz.asc

twine upload --repository-url https://upload.pypi.org/legacy/ apache-flink-${RELEASE_VERSION}.tar.gz apache-flink-${RELEASE_VERSION}.tar.gz.asc

If upload failed or incorrect for some reason(e.g. network transmission problem), you need to delete the uploaded release package of the same version(if exists) and rename the artifact to apache-flink-${RELEASE_VERSION}.post0.tar.gz, then re-upload.

Note: re-uploading to pypi.org must be avoided as much as possible because it will cause some irreparable problems. If that happens, users cannot install the apache-flink package by explicitly specifying the package version, i.e. the following command "pip install apache-flink==${RELEASE_VERSION}" will fail. Instead they have to run "pip install apache-flink" or "pip install apache-flink==${RELEASE_VERSION}.post0" to install the apache-flink package.

Deploy artifacts to Maven Central Repository

Use the Apache Nexus repository to release the staged binary artifacts to the Maven Central repository. In the Staging Repositories section, find the relevant release candidate orgapacheflink-XXX entry and click Release. Drop all other release candidates that are not being released.

Deploy source and binary releases to dist.apache.org

Copy the source and binary releases from the dev repository to the release repository at dist.apache.org using Subversion.

Code Block
svn move -m "Release Flink ${RELEASE_VERSION}" https://dist.apache.org/repos/dist/dev/flink/flink-${RELEASE_VERSION}-rc${RC_NUM} https://dist.apache.org/repos/dist/release/flink/flink-${RELEASE_VERSION}

(Note: Only PMC members have access to the release repository. If you do not have access, ask on the mailing list for assistance.)

Remove old release candidates from dist.apache.org

Remove the old release candidates from https://dist.apache.org/repos/dist/dev/flink using Subversion.

Code Block
languagebash
titleRemove old release candidates from dist.apache.org
svn checkout https://dist.apache.org/repos/dist/dev/flink --depth=immediates
cd flink
svn remove flink-${RELEASE_VERSION}-rc*
svn commit -m "Remove old release candidates for Apache Flink ${RELEASE_VERSION}"

Git tag

Create and push a new Git tag for the released version by copying the tag for the final release candidate, as follows:

Code Block
git tag -s "release-${RELEASE_VERSION}" refs/tags/${TAG}^{} -m "Release Flink ${RELEASE_VERSION}"
git push <remote> refs/tags/release-${RELEASE_VERSION}

Mark the version as released in JIRA

In JIRA, inside version management, hover over the current release and a settings menu will appear. Click Release, and select today’s date.

(Note: Only PMC members have access to the project administration. If you do not have access, ask on the mailing list for assistance.)

If PRs have been merged to the release branch after the the last release candidate was tagged, make sure that the corresponding Jira tickets have the correct Fix Version set.

Publish the Dockerfiles for the new release

Note: the official Dockerfiles fetch the binary distribution of the target Flink version from an Apache mirror. After publishing the binary release artifacts, mirrors can take some hours to start serving the new artifacts, so you may want to wait to do this step until you are ready to continue with the "Promote the release" steps below.

Follow the instructions in the flink-docker repo to build the new Dockerfiles and send an updated manifest to Docker Hub so the new images are built and published.

Checklist to proceed to the next step

  • Python artifacts released and indexed in the PyPI Repository
  • Maven artifacts released and indexed in the Maven Central Repository (usually takes about a day to show up)
  • Source & binary distributions available in the release repository of https://dist.apache.org/repos/dist/release/flink/
  • Dev repository https://dist.apache.org/repos/dist/dev/flink/ is empty
  • Release tagged in the source code repository
  • Release version finalized in JIRA. (Note: Not all committers have administrator access to JIRA. If you end up getting permissions errors ask on the mailing list for assistance)
  • Website contains links to new release binaries and sources in download page
  • For major releases, the front page references the correct new major release version and directs to the correct link
  • Dockerfiles in flink-docker updated for the new Flink release and pull request opened on the Docker official-images with an updated manifest

...

Expand
titlePromote the release

Once the release has been finalized, the last step of the process is to promote the release within the project and beyond. Please wait for 24h after finalizing the release in accordance with the ASF release policy.

Update japicmp configuration

Update the japicmp reference version and enable API compatibility checks for @PublicEvolving  APIs on the corresponding SNAPSHOT branch with the update_japicmp_configuration.sh script (see below).

For a new major release (x.y.0), run the same command also on the master branch for updating the japicmp reference version, and remove all out-dated exclusions in the japicmp configurations (which will usually mean removing all exclusions that aren't workarounds for Scala magic).

Code Block
languagebash
tools $ NEW_VERSION=$RELEASE_VERSION releasing/update_japicmp_configuration.sh
tools $ cd ..
$ git add *
$ git commit -m "Update japicmp configuration for $RELEASE_VERSION"

Merge website pull request

Merge the website pull request to list the release. Make sure to regenerate the website as well, as it isn't build automatically.

Remove outdated versions

dist.apache.org

For a new major release remove all release files older than 2 versions, e.g., when releasing 1.7, remove all releases <= 1.5.

For a new bugfix version remove all release files for previous bugfix releases in the same series, e.g., when releasing 1.7.1, remove the 1.7.0 release.

  1. If you have not already, check out the Flink section of the release repository on dist.apache.org via Subversion. In a fresh directory:

    Code Block
    languagebash
     svn checkout https://dist.apache.org/repos/dist/release/flink --depth=immediates
     cd flink


  2. Remove files for outdated releases and commit the changes.

    Code Block
    languagebash
     svn remove flink-<version_to_remove>
     svn commit


  3. Verify that files  are removed

    Warning
    Remember to remove the corresponding download links from the website.


CI

Disable the cron job for the now-unsupported version from (tools/azure-pipelines/build-apache-repo.yml) in the respective branch.

Apache mailing lists

Announce on the dev@ mailing list that the release has been finished.

Announce on the release on the user@ mailing list, listing major improvements and contributions.

Announce the release on the announce@apache.org mailing list.

Code Block
languagetext
From: Release Manager
To: dev@flink.apache.org, user@flink.apache.org, user-zh@flink.apache.org, announce@apache.org
Subject: [ANNOUNCE] Apache Flink 1.2.3 released

The Apache Flink community is very happy to announce the release of Apache Flink 1.2.3, which is the third bugfix release for the Apache Flink 1.2 series.

Apache Flink® is an open-source stream processing framework for distributed, high-performing, always-available, and accurate data streaming applications.

The release is available for download at:
https://flink.apache.org/downloads.html

Please check out the release blog post for an overview of the improvements for this bugfix release:
<blob post link>

The full release notes are available in Jira:
<jira release notes link>

We would like to thank all contributors of the Apache Flink community who made this release possible!

Regards,
Release Manager

Recordkeeping

Use reporter.apache.org to seed the information about the release into future project reports.

(Note: Only PMC members have access report releases. If you do not have access, ask on the mailing list for assistance.)

Flink blog

Major or otherwise important releases should have a blog post. Write one if needed for this particular release. Minor releases that don’t introduce new major functionality don’t necessarily need to be blogged.

Please make sure that the release notes of the documentation (see section "Review and update documentation") are linked from the blog post of a major release.
We usually include the names of all contributors in the announcement blog post. Use the following command to get the list of contributors:


Code Block
git log --pretty="%an%n%cn" <fromCommit>..<toCommit> | sort | uniq | tr "\n" "," | sed 's/,/, /g'


Social media

Tweet, post on Facebook, LinkedIn, and other platforms. Ask other contributors to do the same.

Checklist to declare the process completed

  1. Website pull request to list the release merged
  2. Release announced on the user@ mailing list.
  3. Blog post published, if applicable.
  4. Release recorded in reporter.apache.org.
  5. Release announced on social media.
  6. Completion declared on the dev@ mailing list.
  7. Update Homebrew: https://docs.brew.sh/How-To-Open-a-Homebrew-Pull-Request (seems to be done automatically - at least for minor releases)
  8. Update quickstart scripts in flink-web, under the q/ directory
  9. Updated the japicmp configuration
    • corresponding SNAPSHOT branch japicmp reference version set to the just released version, and API compatibiltity checks for @PublicEvolving  was enabled
    • (major only) master branch japicmp reference version set to the just released version
    • (major only) master branch japicmp exclusions have been cleared
  10. Update the list of previous version in docs/_config.yml on toml on the master branch.
  11. Set show_outdated_warning: true in docs/_config.yml in toml in the branch of the previous Flink version
  12. Update stable and master alias in flink.py found in https://github.com/apache/infrastructure-bb2/blob/master/flink.py
    1. Update flink_branches_nightly, e.g.

      Code Block
      flink_branches_nightly = [
          ('master', 'release-1.13'),
          ('release-1.12', 'stable'),
          'release-1.11',
      ]
      ->
      flink_branches_nightly = [
          ('master', 'release-1.14'),
          ('release-1.13', 'stable'),
          'release-1.12',
      ]


    2. Add the oldest release to flink_branches_manual


  13. Update docs to "stable" in docs/_config.yml in toml in the branch of the just-released version:
  14. Update CURRENT_VERSION in TypeSerializerUpgradeTestBase  with the new version. This will likely fail some tests because snapshots are missing for that version. Generate them, for example in TypeSerializerUpgradeTestBase. 
  15. (major/minor only) Update migration tests in master to cover migration from new version: (search for usages of FlinkVersion)
    • AbstractOperatorRestoreTestBase
    • CEPMigrationTest
    • BucketingSinkMigrationTest
    • FlinkKafkaConsumerBaseMigrationTest
    • ContinuousFileProcessingMigrationTest
    • WindowOperatorMigrationTest
    • StatefulJobSavepointMigrationITCase
    • StatefulJobWBroadcastStateMigrationITCase


...