Skip to main content

AWS CDK custom resource that handles large files deployment to S3 bucket.

Project description

Pipeline

B.CfnS3LargeDeployment

b-cfn-s3-large-deployment - AWS CDK custom resource that handles large files deployment to S3 bucket.

Description

This custom resource deploys local files or S3 bucket objects to a destination bucket retaining their file-system hierarchy.

Two types of deployment sources are available:

  • BucketDeploymentSource - uses another S3 bucket object(-s) as source for the deployment to a destination bucket. Only files up to 5TB are supported due to S3 bucket limitations;
  • AssetDeploymentSource - uses aws-cdk.aws-s3-assets lib to deploy local files as .zip files to assets bucket from which extracted contents are moved to the destination bucket. Asset files more than 2GB in size are not supported.

See "Known limits" sections below for more information on this resource limitations.

This resource implementation is based on GitHub pull-request https://github.com/aws/aws-cdk/pull/15220.

Remarks

Biomapas aims to modernise life-science industry by sharing its IT knowledge with other companies and the community.

Related technology

  • Python >= 3.8
  • Amazon Web Services (AWS)

Assumptions

The project assumes that the person working with it have basic knowledge in python programming.

Useful sources

See code documentation for any additional sources and references. Also see aws-cdk.s3-deployment library for more information as this implementation is based on work done there.

Install

Use the package manager pip to install this package. This project is not in the PyPi repository yet. Install directly from source or PyPI.

pip install .

Or

pip install b-cfn-s3-large-deployment

Usage & Examples

This AWS CloudFormation custom resource is used pretty much the same way as any other resource. Simply initialize it within any valid CDK scope giving it unique name/id, providing source(-s) and the destination for the deployment.

Optionally, if there's a need for larger files deployment than what AWS Lambda's /tmp directory supports, setting the DeploymentPops.use_efs and DeploymentPops.efs_props fields, AWS Elastic File Storage (EFS) can be enabled to allow such files handling.

A simple example of S3LargeDeploymentResource usage is shown below:

from aws_cdk.core import App, Stack, Construct
from aws_cdk.aws_s3 import Bucket

from b_cfn_s3_large_deployment.resource import S3LargeDeploymentResource
from b_cfn_s3_large_deployment.deployment_props import DeploymentProps
from b_cfn_s3_large_deployment.deployment_source import AssetDeploymentSource, BucketDeploymentSource


class ExampleStack(Stack):
    def __init__(self, scope: Construct):
        super().__init__(...)

        S3LargeDeploymentResource(
            scope=self,
            name='ExampleLargeDeployment',
            sources=[
                AssetDeploymentSource(path='/path/to/your/local/directory'),
                AssetDeploymentSource(path='/path/to/your/local/zip/file.zip'),
                BucketDeploymentSource(
                  bucket=..., 
                  zip_object_key='your-source-bucket-object-key'
                ),
                ...
            ],
            destination_bucket=Bucket(...),
            props=DeploymentProps(...)
        )
        ...


app = App()
ExampleStack(app, 'ExampleStack')

app.synth()

Here, three types of supported sources were used:

  1. whole, local directory given as a path, which is then deployed to the assets bucket as a .zip object:

    AssetDeploymentSource(path='/path/to/your/local/directory')
    
  2. single .zip file given as a path, which is then deployed to the assets bucket:

    AssetDeploymentSource(path='/path/to/your/local/zip/file.zip')
    
  3. Single .zip S3 object found in the source bucket, given as an object key. No further pre-processing is applied in this case:

    BucketDeploymentSource(
       bucket=...,
       zip_object_key='your-source-bucket-object-key'
    )
    

In all of these cases, final, source .zip objects are extracted inside S3LargeDeploymentResource's handler function storage and the available contents are then deployed to the configured destination. This is all done, while maintaining original file structure of source contents.

Known limits

  • aws_cdk.aws_s3_assets.Asset supports up to 2GB/asset (limited by NodeJS implementation).
  • S3 bucket supports up to 5TB objects.

Testing

No throughout testing is yet implemented - only a single integration test, checking the success of cdk synth command.

Contribution

Found a bug? Want to add or suggest a new feature? Contributions of any kind are gladly welcome. Contact your direct supervisor, create a pull-request or an issue in Jira platform.

Release history

1.1.3

  • Fixed dependencies bug between S3LargeDeploymentResource and its handler function.
  • Upgraded testing pipeline to use B.AwsCdkParallel.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

b-cfn-s3-large-deployment-1.1.3.tar.gz (18.4 kB view details)

Uploaded Source

Built Distribution

b_cfn_s3_large_deployment-1.1.3-py3-none-any.whl (19.6 kB view details)

Uploaded Python 3

File details

Details for the file b-cfn-s3-large-deployment-1.1.3.tar.gz.

File metadata

  • Download URL: b-cfn-s3-large-deployment-1.1.3.tar.gz
  • Upload date:
  • Size: 18.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.7.1 importlib_metadata/4.9.0 pkginfo/1.8.2 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.8.12

File hashes

Hashes for b-cfn-s3-large-deployment-1.1.3.tar.gz
Algorithm Hash digest
SHA256 6cb35b7cff4aca577c134668bb4048f9b09efe4c05db26054e9ba8a5662416ab
MD5 949652510750327769dbbccac7636e35
BLAKE2b-256 1458961251e54c2646b06782856c3b7767af088b802f7521ceda1b35753ddd43

See more details on using hashes here.

File details

Details for the file b_cfn_s3_large_deployment-1.1.3-py3-none-any.whl.

File metadata

  • Download URL: b_cfn_s3_large_deployment-1.1.3-py3-none-any.whl
  • Upload date:
  • Size: 19.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.7.1 importlib_metadata/4.9.0 pkginfo/1.8.2 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.8.12

File hashes

Hashes for b_cfn_s3_large_deployment-1.1.3-py3-none-any.whl
Algorithm Hash digest
SHA256 49c91a401cefdd9357af6a6a49411d5798e9e048b09b26ad410de68b1bcdffa6
MD5 3325bc0d14c36300b4c136d4a4ee6af2
BLAKE2b-256 af53e343f1dbe8f12ae56e611aa37d60c719abce99b11da1753d446e71939aa2

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page