Skip to main content

Back-ported airflow.providers.databricks.* package for Airflow 1.10.*

Project description

Package apache-airflow-backport-providers-databricks

Release: 2020.6.24

Backport package

This is a backport providers package for databricks provider. All classes for this provider package are in airflow.providers.databricks python package.

Only Python 3.6+ is supported for this backport package.

While Airflow 1.10.* continues to support Python 2.7+ - you need to upgrade python to 3.6+ if you want to use this backport package.

Installation

You can install this package on top of an existing airflow 1.10.* installation via pip install apache-airflow-backport-providers-databricks

Compatibility

For full compatibility and test status of the backport packages check Airflow Backport Package Compatibility

PIP requirements

PIP package Version required
requests >=2.20.0, <3

Provider classes summary

In Airflow 2.0, all operators, transfers, hooks, sensors, secrets for the databricks provider are in the airflow.providers.databricks package. You can read more about the naming conventions used in Naming conventions for provider packages

Operators

Moved operators

Airflow 2.0 operators: airflow.providers.databricks package Airflow 1.10.* previous location (usually airflow.contrib)
operators.databricks.DatabricksRunNowOperator contrib.operators.databricks_operator.DatabricksRunNowOperator
operators.databricks.DatabricksSubmitRunOperator contrib.operators.databricks_operator.DatabricksSubmitRunOperator

Hooks

Moved hooks

Airflow 2.0 hooks: airflow.providers.databricks package Airflow 1.10.* previous location (usually airflow.contrib)
hooks.databricks.DatabricksHook contrib.hooks.databricks_hook.DatabricksHook

Releases

Release 2020.6.24

Commit Committed Subject
12af6a080 2020-06-19 Final cleanup for 2020.6.23rc1 release preparation (#9404)
c7e5bce57 2020-06-19 Prepare backport release candidate for 2020.6.23rc1 (#9370)
f6bd817a3 2020-06-16 Introduce 'transfers' packages (#9320)
0b0e4f7a4 2020-05-26 Preparing for RC3 relase of backports (#9026)
00642a46d 2020-05-26 Fixed name of 20 remaining wrongly named operators. (#8994)
f1073381e 2020-05-22 Add support for spark python and submit tasks in Databricks operator(#8846)
375d1ca22 2020-05-19 Release candidate 2 for backport packages 2020.05.20 (#8898)
12c5e5d8a 2020-05-17 Prepare release candidate for backport packages (#8891)
f3521fb0e 2020-05-16 Regenerate readme files for backport package release (#8886)
92585ca4c 2020-05-15 Added automated release notes generation for backport operators (#8807)
649935e8c 2020-04-27 [AIRFLOW-8472]: PATCH for Databricks hook _do_api_call (#8473)
16903ba3a 2020-04-24 [AIRFLOW-8474]: Adding possibility to get job_id from Databricks run (#8475)
5648dfbc3 2020-03-23 Add missing call to Super class in 'amazon', 'cloudant & 'databricks' providers (#7827)
3320e432a 2020-02-24 [AIRFLOW-6817] Lazy-load airflow.DAG to keep user-facing API untouched (#7517)
4d03e33c1 2020-02-22 [AIRFLOW-6817] remove imports from airflow/__init__.py, replaced implicit imports with explicit imports, added entry to UPDATING.MD - squashed/rebased (#7456)
97a429f9d 2020-02-02 [AIRFLOW-6714] Remove magic comments about UTF-8 (#7338)
83c037873 2020-01-30 [AIRFLOW-6674] Move example_dags in accordance with AIP-21 (#7287)
c42a375e7 2020-01-27 [AIRFLOW-6644][AIP-21] Move service classes to providers package (#7265)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

File details

Details for the file apache-airflow-backport-providers-databricks-2020.6.24.tar.gz.

File metadata

File hashes

Hashes for apache-airflow-backport-providers-databricks-2020.6.24.tar.gz
Algorithm Hash digest
SHA256 c36769b258b399806f37298dcb547e11cb11a8aff50a0edfae196aa59246e3c5
MD5 2689831572c81be31c4e28a11819c11f
BLAKE2b-256 b1e8d6cfafe0e13d1847c5df6e651f3dbe2f68c96a11830d4913a582cde51cde

See more details on using hashes here.

File details

Details for the file apache_airflow_backport_providers_databricks-2020.6.24-py3-none-any.whl.

File metadata

File hashes

Hashes for apache_airflow_backport_providers_databricks-2020.6.24-py3-none-any.whl
Algorithm Hash digest
SHA256 09f9e6e982bc760c489260220fd2a03e1875d68b851b9690fba0e7f6da93cbfb
MD5 53d031591a0933a49df3e66365782430
BLAKE2b-256 ba47b60458f5df919389cbe5a221b21123d7dc2bf2a1278982458e9df8d7fe92

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page