Skip to main content

Provider for Apache Airflow. Implements apache-airflow-providers-apache-hdfs package

Project description

Package apache-airflow-providers-apache-hdfs

Release: 4.1.0

Hadoop Distributed File System (HDFS) and WebHDFS.

Provider package

This is a provider package for apache.hdfs provider. All classes for this provider package are in airflow.providers.apache.hdfs python package.

You can find package information and changelog for the provider in the documentation.

Installation

You can install this package on top of an existing Airflow 2 installation (see Requirements below for the minimum Airflow version supported) via pip install apache-airflow-providers-apache-hdfs

The package supports the following python versions: 3.8,3.9,3.10,3.11

Requirements

PIP package

Version required

apache-airflow

>=2.4.0

hdfs[avro,dataframe,kerberos]

>=2.0.4

Changelog

4.1.0

Features

  • Add ability to read/write task instance logs from HDFS (#31512)

4.0.0

Breaking changes

The original HDFS Hook and sensor has been removed. It used the old HDFS snakebite-py3 library that had no update in years and the protobuf they are using reached end of life.

The 3.* version of the provider is still available and can be used if you need to use the old hooks and sensors.

The HDFSHook, HDFSSensor, HdfsRegexSensor, HdfsRegexSensor that have been removed from this provider and they are not available any more. If you want to continue using them, you can use 3.* version of the provider, but the recommendation is to switch to the new WebHDFSHook and WebHDFSSensor that use the WebHDFS API.

  • Remove snakebite-py3 based HDFS hooks and sensors (#31262)

Misc

  • Bump minimum Airflow version in providers (#30917)

3.2.1

Bug Fixes

  • Fix HDFSHook HAClient is invalid (#30164)

3.2.0

Misc

  • Move min airflow version to 2.3.0 for all providers (#27196)

3.1.0

Features

  • Adding Authentication to webhdfs sensor (#25110)

3.0.1

Bug Fixes

  • 'WebHDFSHook' Bugfix/optional port (#24550)

3.0.0

Breaking changes

Misc

  • chore: Refactoring and Cleaning Apache Providers (#24219)

2.2.3

Bug Fixes

  • Fix mistakenly added install_requires for all providers (#22382)

2.2.2

Misc

  • Add Trove classifiers in PyPI (Framework :: Apache Airflow :: Provider)

2.2.1

Misc

  • Support for Python 3.10

  • Add how-to guide for WebHDFS operators (#21393)

2.2.0

Features

  • hdfs provider: restore HA support for webhdfs (#19711)

2.1.1

Bug Fixes

  • fix get_connections deprecation warning in webhdfs hook (#18331)

2.1.0

Features

  • hdfs provider: allow SSL webhdfs connections (#17637)

Misc

  • Optimise connection importing for Airflow 2.2.0

2.0.0

Breaking changes

  • Auto-apply apply_default decorator (#15667)

1.0.1

Updated documentation and readme files.

1.0.0

Initial version of the provider.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

apache-airflow-providers-apache-hdfs-4.1.0.tar.gz (21.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

File details

Details for the file apache-airflow-providers-apache-hdfs-4.1.0.tar.gz.

File metadata

File hashes

Hashes for apache-airflow-providers-apache-hdfs-4.1.0.tar.gz
Algorithm Hash digest
SHA256 754a42fcde6a24cd07634594898dada7a445fa6d1710b2eef1974e16bf30ef0c
MD5 7e7b031cb42c501016fa6fb1ddecd988
BLAKE2b-256 a491d38f246d5b27ed4e4d5a6cdb71f42f8289a1ea4d69b8e40ca9234b56bb4c

See more details on using hashes here.

File details

Details for the file apache_airflow_providers_apache_hdfs-4.1.0-py3-none-any.whl.

File metadata

File hashes

Hashes for apache_airflow_providers_apache_hdfs-4.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 912c0ce69bdccaea8f2a5e7a9b02890e3e0efc5681aa6db9e1c3a16c832ee0c8
MD5 6f88e8ca6f5c7fc116208c185aab4893
BLAKE2b-256 af4d12cb56c79bc1749775101c588c3b4373a750735d1082ef87cf8de403eda7

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page