Skip to main content

Python library for working with Amazon Web Services such as Redshift and S3

Project description

Amazonian

Amazonian is a Python library for interacting easily with Amazon S3 and Redshift.

Installation

pip install amazonian

Usage

S3

from amazonian import S3

s3 = S3(key=None, secret=None, iam_role=None, root='s3://', spark=spark)

# get list of files:
s3.ls(path='s3://bucket/directory/subdirectory')

# get a tree representation of folder structure
s3.tree(path='s3://bucket/directory/subdirectory')

# get file size:
s3.get_size(path='some_file')

# save a Spark DataFrame as a Parquet
s3.save_parquet(data=my_data, path='s3://bucket/directory/subdirectory/name.parquet')

# load a Parquet into a Spark DataFrame
my_data = s3.load_parquet(path='s3://bucket/directory/subdirectory/name.parquet')

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

amazonian-2022.5.27.tar.gz (14.7 kB view hashes)

Uploaded Source

Built Distribution

amazonian-2022.5.27-py3-none-any.whl (18.6 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page