#SGVirtual 20.04 Introduction to Apache Airflow

preview_player
Показать описание
Apache Airflow is a tool created by community to programmatically author, schedule and monitor workflows. The biggest advantage of Airflow is the fact that it does not limit the scope of pipelines. Airflow can be used for building Machine Learning models, transferring data or managing the infrastructure. Because Airflow workflows are written in pure Python, there is no restriction on what a single task can do. It can execute a python callable, bash script or create a new Kubernetes pod. Thanks to that, a single Airflow workflow can include tasks for both setup and teardown of external infrastructure, as well as operations that will be performed using the new resources.

During this talk Tomek Urbaszek and Jarek Potiuk, both Apache Airflow official committers, provide an extensive introduction to Airflow. This talk was recorded at SG Virtual 20.04.
Рекомендации по теме