Skip to main content

Back-ported airflow.providers.databricks.* package for Airflow 1.10.*

Project description

Package apache-airflow-backport-providers-databricks

Release: 2020.6.24

Backport package

This is a backport providers package for databricks provider. All classes for this provider package are in airflow.providers.databricks python package.

Only Python 3.6+ is supported for this backport package.

While Airflow 1.10.* continues to support Python 2.7+ - you need to upgrade python to 3.6+ if you want to use this backport package.

Installation

You can install this package on top of an existing airflow 1.10.* installation via pip install apache-airflow-backport-providers-databricks

Compatibility

For full compatibility and test status of the backport packages check Airflow Backport Package Compatibility

PIP requirements

PIP package Version required
requests >=2.20.0, <3

Provider classes summary

In Airflow 2.0, all operators, transfers, hooks, sensors, secrets for the databricks provider are in the airflow.providers.databricks package. You can read more about the naming conventions used in Naming conventions for provider packages

Operators

Moved operators

Airflow 2.0 operators: airflow.providers.databricks package Airflow 1.10.* previous location (usually airflow.contrib)
operators.databricks.DatabricksRunNowOperator contrib.operators.databricks_operator.DatabricksRunNowOperator
operators.databricks.DatabricksSubmitRunOperator contrib.operators.databricks_operator.DatabricksSubmitRunOperator

Hooks

Moved hooks

Airflow 2.0 hooks: airflow.providers.databricks package Airflow 1.10.* previous location (usually airflow.contrib)
hooks.databricks.DatabricksHook contrib.hooks.databricks_hook.DatabricksHook

Releases

Release 2020.6.24

Commit Committed Subject
12af6a080 2020-06-19 Final cleanup for 2020.6.23rc1 release preparation (#9404)
c7e5bce57 2020-06-19 Prepare backport release candidate for 2020.6.23rc1 (#9370)
f6bd817a3 2020-06-16 Introduce 'transfers' packages (#9320)
0b0e4f7a4 2020-05-26 Preparing for RC3 relase of backports (#9026)
00642a46d 2020-05-26 Fixed name of 20 remaining wrongly named operators. (#8994)
f1073381e 2020-05-22 Add support for spark python and submit tasks in Databricks operator(#8846)
375d1ca22 2020-05-19 Release candidate 2 for backport packages 2020.05.20 (#8898)
12c5e5d8a 2020-05-17 Prepare release candidate for backport packages (#8891)
f3521fb0e 2020-05-16 Regenerate readme files for backport package release (#8886)
92585ca4c 2020-05-15 Added automated release notes generation for backport operators (#8807)
649935e8c 2020-04-27 [AIRFLOW-8472]: PATCH for Databricks hook _do_api_call (#8473)
16903ba3a 2020-04-24 [AIRFLOW-8474]: Adding possibility to get job_id from Databricks run (#8475)
5648dfbc3 2020-03-23 Add missing call to Super class in 'amazon', 'cloudant & 'databricks' providers (#7827)
3320e432a 2020-02-24 [AIRFLOW-6817] Lazy-load airflow.DAG to keep user-facing API untouched (#7517)
4d03e33c1 2020-02-22 [AIRFLOW-6817] remove imports from airflow/__init__.py, replaced implicit imports with explicit imports, added entry to UPDATING.MD - squashed/rebased (#7456)
97a429f9d 2020-02-02 [AIRFLOW-6714] Remove magic comments about UTF-8 (#7338)
83c037873 2020-01-30 [AIRFLOW-6674] Move example_dags in accordance with AIP-21 (#7287)
c42a375e7 2020-01-27 [AIRFLOW-6644][AIP-21] Move service classes to providers package (#7265)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

Built Distribution

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page