Python library for working with Amazon Web Services such as Redshift and S3
Project description
Amazonian
Amazonian is a Python library for interacting easily with Amazon S3 and Redshift.
Installation
pip install amazonian
Usage
S3
from amazonian import S3
s3 = S3(key=None, secret=None, iam_role=None, root='s3://', spark=spark)
# get list of files:
s3.ls(path='s3://bucket/directory/subdirectory')
# get a tree representation of folder structure
s3.tree(path='s3://bucket/directory/subdirectory')
# get file size:
s3.get_size(path='some_file')
# save a Spark DataFrame as a Parquet
s3.save_parquet(data=my_data, path='s3://bucket/directory/subdirectory/name.parquet')
# load a Parquet into a Spark DataFrame
my_data = s3.load_parquet(path='s3://bucket/directory/subdirectory/name.parquet')
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
amazonian-2021.8.20.1.tar.gz
(14.1 kB
view hashes)
Built Distribution
Close
Hashes for amazonian-2021.8.20.1-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | d382d67fc82df461dbec09b10b80946a85081ac6b81996f4e7562fa7145e178c |
|
MD5 | 51b58a70efeee4a96b1a437d73135c62 |
|
BLAKE2b-256 | 8d32da4f4b1b1642c4110d48785205a69fbfa3f31eee5d19d34bebb65b9a498a |