Datahub Airflow plugin to capture executions and send to Datahub
Project description
Datahub Airflow Plugin
Capabilities
DataHub supports integration of
- Airflow Pipeline (DAG) metadata
- DAG and Task run information
- Lineage information when present
Installation
- You need to install the required dependency in your airflow.
pip install acryl-datahub-airflow-plugin
- Disable lazy plugin load in your airflow.cfg
core.lazy_load_plugins : False
-
You must configure an Airflow hook for Datahub. We support both a Datahub REST hook and a Kafka-based hook, but you only need one.
# For REST-based: airflow connections add --conn-type 'datahub_rest' 'datahub_rest_default' --conn-host 'http://localhost:8080' # For Kafka-based (standard Kafka sink config can be passed via extras): airflow connections add --conn-type 'datahub_kafka' 'datahub_kafka_default' --conn-host 'broker:9092' --conn-extra '{}'
-
Add your
datahub_conn_id
and/orcluster
to yourairflow.cfg
file if it is not align with the default values. See configuration parameters belowConfiguration options:
Name Default value Description datahub.datahub_conn_id datahub_rest_deafault The name of the datahub connection you set in step 1. datahub.cluster prod name of the airflow cluster datahub.capture_ownership_info true If true, the owners field of the DAG will be capture as a DataHub corpuser. datahub.capture_tags_info true If true, the tags field of the DAG will be captured as DataHub tags. datahub.graceful_exceptions true If set to true, most runtime errors in the lineage backend will be suppressed and will not cause the overall task to fail. Note that configuration issues will still throw exceptions. -
Configure
inlets
andoutlets
for your Airflow operators. For reference, look at the sample DAG inlineage_backend_demo.py
, or referencelineage_backend_taskflow_demo.py
if you're using the TaskFlow API. -
[optional] Learn more about Airflow lineage, including shorthand notation and some automation.
How to validate installation
- Go and check in Airflow at Admin -> Plugins menu if you can see the Datahub plugin
- Run an Airflow DAG and you should see in the task logs Datahub releated log messages like:
Emitting Datahub ...
Additional references
Related Datahub videos: Airflow Lineage Airflow Run History in DataHub
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file treff7es-datahub-airflow-plugin-0.0.0.3.tar.gz
.
File metadata
- Download URL: treff7es-datahub-airflow-plugin-0.0.0.3.tar.gz
- Upload date:
- Size: 9.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.0 CPython/3.9.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | e134adb92244840e617763e2ba4bbf870704c60195942a6f6b4595f28aa23314 |
|
MD5 | afd21993c1f83219ef99bfbb79340d7b |
|
BLAKE2b-256 | 30096d1a1bc0be5c80f029dd721edfba938bae56be3c5bf971ea8c63f5621899 |
File details
Details for the file treff7es_datahub_airflow_plugin-0.0.0.3-py3-none-any.whl
.
File metadata
- Download URL: treff7es_datahub_airflow_plugin-0.0.0.3-py3-none-any.whl
- Upload date:
- Size: 7.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.0 CPython/3.9.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 8683ffdfb14f84e9b76d29e42a116df94459bbd4d5b83b2a0adde32ed16d8c8d |
|
MD5 | 0ac3e3d1bbdefb655f234b5356675f09 |
|
BLAKE2b-256 | fdcf9d3e6d5d963b292ddafbb886771c640273a4d633cc7f6a55ca99ac4f658d |