Airflow

Adding Airflow to your decube connections helps your team to discover, document and monitor the quality of your pipelines.

Supported Capabilities

General

  • Metadata — metadata extraction and display of asset information (tables, columns, schemas). Types collected: Data Job, Data Run, Data Task

Data Quality Monitors

  • Job Failure

Connection Requirements

  • Airflow Username

  • Airflow User Password

  • Airflow API host address

  • Airflow API enabled and set to Basic Auth. See Airflow Documentationarrow-up-right for this.

  • Airflow Version 2.3.0 and above (Version >= 2.0.0 and < 2.3.0 may not work fully)

Airflow

Creating an Airflow User for Decube

  1. Go to Security > List Users

  1. Click on "+" to Add User

  1. Insert information for new decube user

    1. Username - suggested value: decube

    2. Role - Minimum required Op (which is a default role from Airflow)

    3. Password - Use a strong password

Airflow API is not Publicly Accessible

  • For decube to monitor your Airflow service, we will require that the Airflow API be publicly accessible or privately accessible to a SSH bastion host. Instruction on setting up a bastion host can be found here SSH Tunneling

Last updated