Add S3 support to dtool
Project description
Free software: MIT License
Features
Copy datasets to and from S3 object storage
List all the datasets in a S3 bucket
Create datasets directly in S3
Installation
To install the dtool-S3 package:
pip install dtool-s3
Configuration
Install the aws client, for details see https://docs.aws.amazon.com/cli/latest/userguide/installing.html. In short:
pip install awscli --upgrade --user
Configure the credentials using:
aws configure
These are needed for the boto3 library, for more details see https://boto3.readthedocs.io/en/latest/guide/quickstart.html.
Usage
To copy a dataset from local disk (my-dataset) to a S3 bucket (/data_raw) one can use the command below:
dtool copy ./my-dataset s3://data_raw
To list all the datasets in a S3 bucket one can use the command below:
dtool ls s3://data_raw
See the dtool documentation for more detail.
Configuring the local dtool S3 cache
When fetching items from a dataset, for example using the dtool item fetch command, the content of the item is cached in a file on local disk. The default cache directory is ~/.cache/dtool/s3.
One may want to change this directory. For example, if working on a HPC cluster to set it to a directory that lives on fast solid state disk. This can be achieved by setting the DTOOL_S3_CACHE_DIRECTORY environment variable. For example:
mkdir -p /tmp/dtool/s3 export DTOOL_S3_CACHE_DIRECTORY=/tmp/dtool/s3
Alternatively, when using the dtool command line interface one can add the DTOOL_S3_CACHE_DIRECTORY key to the ~/.config/dtool/dtool.json file. For example:
{ "DTOOL_S3_CACHE_DIRECTORY": "/tmp/dtool/s3" }
If the file does not exist one may need to create it.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
File details
Details for the file dtool-s3-0.4.1.tar.gz
.
File metadata
- Download URL: dtool-s3-0.4.1.tar.gz
- Upload date:
- Size: 7.3 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: Python-urllib/3.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 7cbb7398ff4e255979c9c82b726e6009703baa34739e87518052695f120639cc |
|
MD5 | 63ebe013d2b1b4b9eea827d593914c93 |
|
BLAKE2b-256 | b20b23aace0edc2c96b7dfe36f3ce535aa06f364aae7f80b47a1d65aef17adf0 |