About Apache Airflow
Where data teams schedule, run and monitor their pipelines.
Apache Airflow is the open-source platform data teams use to schedule, run and monitor their pipelines. Engineers write each pipeline as a Python DAG (a graph of tasks with dependencies, retries and timing rules), Airflow schedules it, runs it, and keeps a record of every run, every task and every retry. The community maintains a long list of providers for the warehouses, SaaS systems and cloud services those tasks touch, so the same scheduler can load Snowflake, fire a dbt run, and ping Slack on failure.
What teams build on Airflow runs from nightly ELT into the data warehouse to ML training jobs, data-quality checks, report generation and customer-facing exports. With Airflow's run history pulled into your warehouse, the question of which DAG is fragile, which team owns it, and what each run is costing you becomes a dashboard, not a tail of the web UI.