Using Apache Airflow to Create Data Workflows on Google Cloud.

preview_player
Показать описание
Using Apache Airflow to create Dynamic, Extensible, Elegant, Scalable Data Workflows on Google Cloud at SoulCycle.

In this webinar we are going to explore using Apache Airflow to create dynamic, extensible, elegant, scalable data workflows on Google Cloud at SoulCycle.

Caserta consultants will share, in technical detail, how to define and configure Airflow pipelines to create workflows that are maintainable, versionable, testable, and collaborative.

In this webinar you will learn best practices to create Airflow directed acyclic graphs (DAGs) of tasks, enforcing relationships and dependencies. The talk covers constructing Airflow pipelines with Spark based ETL, a Google Storage Data Lake, and a Big Query Data Warehouse.

The stage will be shared with Dallas S. Simpson, Director of Data Engineering at SoulCycle Inc. Dallas will co-present with Caserta to share his experience and lessons learned using Airflow on their internal GCP Data Analytics Platform project.

Data architects, data engineers and data scientists alike will enjoy the contents of this talk.
Рекомендации по теме
Комментарии
Автор

This is fantastic! Have you moved to Composer from VMs?

EthanLyon