Skip to main content

Provider for Apache Airflow. Implements apache-airflow-providers-apache-hive package

Project description

Package apache-airflow-providers-apache-hive

Release: 6.1.0

Apache Hive

Provider package

This is a provider package for apache.hive provider. All classes for this provider package are in airflow.providers.apache.hive python package.

You can find package information and changelog for the provider in the documentation.

Installation

You can install this package on top of an existing Airflow 2 installation (see Requirements below for the minimum Airflow version supported) via pip install apache-airflow-providers-apache-hive

The package supports the following python versions: 3.7,3.8,3.9,3.10

Requirements

PIP package

Version required

apache-airflow

>=2.4.0

apache-airflow-providers-common-sql

>=1.3.1

hmsclient

>=0.1.0

pandas

>=0.17.1

pyhive[hive]

>=0.6.0

sasl

>=0.3.1; python_version>="3.9"

thrift

>=0.9.2

Cross provider package dependencies

Those are dependencies that might be needed in order to use all the features of the package. You need to install the specified provider packages in order to use them.

You can install such cross-provider dependencies when installing from PyPI. For example:

pip install apache-airflow-providers-apache-hive[amazon]

Dependent package

Extra

apache-airflow-providers-amazon

amazon

apache-airflow-providers-common-sql

common.sql

apache-airflow-providers-microsoft-mssql

microsoft.mssql

apache-airflow-providers-mysql

mysql

apache-airflow-providers-presto

presto

apache-airflow-providers-samba

samba

apache-airflow-providers-vertica

vertica

Changelog

6.1.0

Misc

  • Bump minimum Airflow version in providers (#30917)

  • Update return types of 'get_key' methods on 'S3Hook' (#30923)

6.0.0

Breaking changes

The auth option is moved from the extra field to the auth parameter in the Hook. If you have extra parameters defined in your connections as auth, you should move them to the DAG where your HiveOperator or other Hive related operators are used.

  • Move auth parameter from extra to Hook parameter (#30212)

5.1.3

Bug Fixes

  • Validate Hive Beeline parameters (#29502)

5.1.2

Misc

  • Fixed MyPy errors introduced by new mysql-connector-python (#28995)

5.1.1

Bug Fixes

  • Move local_infile option from extra to hook parameter (#28811)

5.1.0

Features

The apache.hive provider provides now hive macros that used to be provided by Airflow. As of 5.1.0 version of apache.hive the hive macros are provided by the Provider.

  • Move Hive macros to the provider (#28538)

  • Make pandas dependency optional for Amazon Provider (#28505)

5.0.0

Breaking changes

The hive_cli_params from connection were moved to the Hook. If you have extra parameters defined in your connections as hive_cli_params extra, you should move them to the DAG where your HiveOperator is used.

  • Move hive_cli_params to hook parameters (#28101)

Features

  • Improve filtering for invalid schemas in Hive hook (#27808)

4.1.1

Bug Fixes

  • Bump common.sql provider to 1.3.1 (#27888)

4.1.0

Misc

  • Move min airflow version to 2.3.0 for all providers (#27196)

Bug Fixes

  • Filter out invalid schemas in Hive hook (#27647)

4.0.1

Misc

  • Add common-sql lower bound for common-sql (#25789)

4.0.0

Breaking Changes

  • The hql parameter in get_records of HiveServer2Hook has been renamed to sql to match the get_records DbApiHook signature. If you used it as a positional parameter, this is no change for you, but if you used it as keyword one, you need to rename it.

  • hive_conf parameter has been renamed to parameters and it is now second parameter, to match get_records signature from the DbApiHook. You need to rename it if you used it.

  • schema parameter in get_records is an optional kwargs extra parameter that you can add, to match the schema of get_records from DbApiHook.

  • Deprecate hql parameters and synchronize DBApiHook method APIs (#25299)

  • Remove Smart Sensors (#25507)

3.1.0

Features

  • Move all SQL classes to common-sql provider (#24836)

Bug Fixes

  • fix connection extra parameter 'auth_mechanism' in 'HiveMetastoreHook' and 'HiveServer2Hook' (#24713)

3.0.0

Breaking changes

Misc

  • chore: Refactoring and Cleaning Apache Providers (#24219)

  • AIP-47 - Migrate hive DAGs to new design #22439 (#24204)

2.3.3

Bug Fixes

  • Fix HiveToMySqlOperator's wrong docstring (#23316)

2.3.2

Bug Fixes

  • Fix mistakenly added install_requires for all providers (#22382)

2.3.1

Misc

  • Add Trove classifiers in PyPI (Framework :: Apache Airflow :: Provider)

2.3.0

Features

  • Set larger limit get_partitions_by_filter in HiveMetastoreHook (#21504)

Bug Fixes

  • Fix Python 3.9 support in Hive (#21893)

  • Fix key typo in 'template_fields_renderers' for 'HiveOperator' (#21525)

Misc

  • Support for Python 3.10

  • Add how-to guide for hive operator (#21590)

2.2.0

Features

  • Add more SQL template fields renderers (#21237)

  • Add conditional 'template_fields_renderers' check for new SQL lexers (#21403)

2.1.0

Features

  • hive provider: restore HA support for metastore (#19777)

Bug Fixes

2.0.3

Bug Fixes

  • fix get_connections deprecation warn in hivemetastore hook (#18854)

2.0.2

Bug fixes

  • HiveHook fix get_pandas_df() failure when it tries to read an empty table (#17777)

Misc

  • Optimise connection importing for Airflow 2.2.0

2.0.1

Features

  • Add Python 3.9 support (#15515)

2.0.0

Breaking changes

  • Auto-apply apply_default decorator (#15667)

1.0.3

Bug fixes

  • Fix mistake and typos in doc/docstrings (#15180)

  • Fix grammar and remove duplicate words (#14647)

  • Resolve issue related to HiveCliHook kill (#14542)

1.0.2

Bug fixes

  • Corrections in docs and tools after releasing provider RCs (#14082)

1.0.1

Updated documentation and readme files.

Bug fixes

  • Remove password if in LDAP or CUSTOM mode HiveServer2Hook (#11767)

1.0.0

Initial version of the provider.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

Built Distribution

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page