The new major release of the Apache Airflow workflow management platform will soon be one year old. If you haven't considered upgrading before, now is a good time to do so. Moving to the new major release brings new features, faster workflow execution, and improved security. Nevertheless, upgrading involves effort and carries the risk of data loss and temporary system downtime.
To ensure that nothing stands in the way of a problem-free changeover, in this articel we’ll provide you with practical tips on what to bear in mind during the upgrade.
To increase the motivation to upgrade, it's worth taking a look at the new features. As an open-source tool, Apache Airflow is further developed by a wide-ranging community, which develops feature ideas from its own use cases. The new features thus promise user-centric workflow management at the pulse of current needs. What does this mean in concrete terms in version 2.1?
As soon as you want to follow a workflow live, the auto-refresh functionality comes in handy: The status of the tasks are automatically updated in the graph view and the page does not have to be reloaded.
Of course, further functionality will be added as part of the lifecycle of the new major release. For example, in the current development pipeline is a more extensive possibility to define the execution times of a workflow. Here, the last working day of the month will be a valid selection in the future.
Some aspects remain the same in the new major version of Airflow. Accordingly, it is still possible to start a workflow via the web interface and the workflows are still available in a Python file. Typical use cases remain and Apache Airflow best practices are still valid.
What is new, however, is the revised, modern web user interface. This is also the focus in the further development of Apache Airflow.While initially only the design was adapted, in the future Airflow will be transformed into a modern web application with an appealing technology stack.
Also relevant for users is the separation of Apache Airflow into a main application and provider packages. This measure is intended to keep the installation more lightweight. If Airflow is loaded locally on a single server, the Kubernetes package can be left out. The provider packages are also detached from the official Airflow releases and bug fixes are implemented faster. For users, the separation manifests itself in the customization of operator paths in new and old workflows.
For a successful upgrade, you can refer to the detailed documentation provided by Airflow. The necessary steps and background information are summarized there. Nevertheless, you should take the following tips to heart:
We wish you the best of success with the upgrade to the new major version of Apache Airflow. If you need assistance with the transition or implementing an integration scenario with the new API, feel free to contact us. Our consultants are ready to assist you with practical knowhow.