Modern data stack encourages modular and flexible data architecture. It’s crucial to be able to integrate with the best-of-breed technologies on the market to provide the best experience to the users. SAS Studio is a modern and powerful IDE for building robust and reliable data pipelines – Studio Flows. But once data pipelines are ready, how do you bind them together, define dependencies and execute them regularly? Apache Airflow (with 30,000+ stars on GitHub) is a dominating open source solution for orchestration, scheduling and monitoring of data pipelines and other assets. Airflow supports seamless integration with various third-party systems with its mechanism of external providers. SAS Provider for Airflow is one such provider. It allows SAS users to easily create Airflow tasks to execute SAS Studio Flows and other SAS assets. The provider is a project maintained by SAS and available on GitHub at https://github.com/sassoftware/sas-airflow-provider under Apache 2.0 license. This demonstration shows you how to create Airflow DAGs with Studio Flows in SAS Studio and then monitor the execution of DAGs in Airflow UI. You'll also learn how to contribute to the project with your ideas and features.
Good news: We've extended SAS Hackathon registration until Sept. 12, so you still have time to be part of our biggest event yet – our five-year anniversary!
SAS Explore 2023 presentations are now available! (Also indexed for search at lexjansen.com!)
View all available SAS Explore content by category: