Skip to main content

The missing Link between AWS services and the most popular Python data libraries

Project description

AWS Data Wrangler (BETA)

Code style: black

The missing link between AWS services and the most popular Python data libraries.

CAUTION: This project is in BETA version. And was not tested in battle yet.

Read the docs!

AWS Data Wrangler aims to fill a gap between AWS Analytics Services (Glue, Athena, EMR, Redshift) and the most popular Python libraries for lightweight workloads.

The rationale behind AWS Data Wrangler is to use the right tool for each job. And this project was developed with the lightweight jobs in mind. That is never so clear and depends of a lot of different factors, but a good rule of thumb that we discoverd during the tests is that if your workload is something around 5 GB in plan text or less, so you should go with AWS Data Wrangler instead of the consagrated big data tools.

Usually there are two different types of use cases when dealing with data, heavy workloads which are dealt better using distributed tools services like EMR/AWS Glue Spark Job and lightweight workloads that can be treated most efficient using simpler tools, and this is when aws data wrangler comes into action.

For example, in AWS Glue you can choose between two different types of Job, distributed with Apache Spark or single node with Python Shell. In this case data wrangler would use the single node with Python Shell job option (Or even AWS Lambda), resulting in less cost and less warm-up time.

Rationale Image


Contents: Installation | Usage | Known Limitations | Contributing | Dependencies | License


Installation

pip install awswrangler

AWS Data Wrangler runs on Python 2 and 3. And runs on AWS Lambda, AWS Glue, EC2, on-premises and local.

P.S. The Lambda Layer bundle and the Glue egg are available to download. It's just upload to your account and run! :rocket:

Usage

Writing Pandas Dataframe to Data Lake:

awswrangler.s3.write(
        df=df,
        database="database",
        path="s3://...",
        file_format="parquet",
        preserve_index=True,
        mode="overwrite",
        partition_cols=["col"],
    )

If a Glue Database name is passed, all the metadata will be created in the Glue Catalog. If not, only the s3 data write will be done.

Reading from Data Lake to Pandas Dataframe:

df = awswrangler.athena.read("database", "select * from table")

Reading from "infinite" S3 source to Pandas Dataframe through generators. That can set a maximum chunk size in bytes to fit in any memory size:

for df in awswrangler.s3.read(path="s3://...", max_size=500):
    print(df)

Typical ETL:

import pandas
import awswrangler

df = pandas.read_csv("s3//your_bucket/your_object.csv")  # Read from anywhere

# Typical Pandas, Numpy or Pyarrow transformation HERE!

awswrangler.s3.write(  # Storing the data and metadata to Data Lake
        df=df,
        database="database",
        path="s3://...",
        file_format="parquet",
        preserve_index=True,
        mode="overwrite",
        partition_cols=["col"],
    )

Dependencies

AWS Data Wrangler project relies on others great initiatives:

Known Limitations

  • By now only writes in Parquet and CSV file formats
  • By now there are not compression support
  • By now there are not nested type support

Contributing

For almost all features we need rely on AWS Services that didn't have mock tools in the community yet (AWS Glue, AWS Athena). So we are focusing on integration tests instead unit tests.

So, you will need provide a S3 bucket and a Glue/Athena database through environment variables.

export AWSWRANGLER_TEST_BUCKET=...

export AWSWRANGLER_TEST_DATABASE=...

CAUTION: This may this may incur costs in your AWS Account

make init

Make your changes...

make format

make lint

make test

License

This library is licensed under the Apache 2.0 License.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

awswrangler-0.0b2.tar.gz (15.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

awswrangler-0.0b2-py27,py36,py37-none-any.whl (19.2 kB view details)

Uploaded Python 2.7,py36,py37

File details

Details for the file awswrangler-0.0b2.tar.gz.

File metadata

  • Download URL: awswrangler-0.0b2.tar.gz
  • Upload date:
  • Size: 15.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.21.0 setuptools/40.8.0 requests-toolbelt/0.9.1 tqdm/4.31.1 CPython/3.7.1

File hashes

Hashes for awswrangler-0.0b2.tar.gz
Algorithm Hash digest
SHA256 0fee0197b42e91886732ceebbb9ed142bac41df5c564fcbd00f53fa956b7170f
MD5 3de4d7dbfc8c357ab86abf76d8ec085f
BLAKE2b-256 03e951ffc8f4e29a16d94062bbea995378dc7a50bb61b97d401171bfffdef662

See more details on using hashes here.

File details

Details for the file awswrangler-0.0b2-py27,py36,py37-none-any.whl.

File metadata

  • Download URL: awswrangler-0.0b2-py27,py36,py37-none-any.whl
  • Upload date:
  • Size: 19.2 kB
  • Tags: Python 2.7,py36,py37
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.21.0 setuptools/40.8.0 requests-toolbelt/0.9.1 tqdm/4.31.1 CPython/3.7.1

File hashes

Hashes for awswrangler-0.0b2-py27,py36,py37-none-any.whl
Algorithm Hash digest
SHA256 47ed59a0881f6c8c899954de1ce97c5ec43f6f735ea9177b96e8be6ab13bf642
MD5 f2271becdb4f1f6f1d8ddb36ae663798
BLAKE2b-256 597335e38cfb9ef9c9833ea7ce2e5042b3200cff9dc888e01dbfa63552de1f7c

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page