Skip to main content

Infrastructure for AI applications and machine learning pipelines

Project description

PackYak image

PyPI version

[!NOTE] Still in active development.

PackYak

PackYak is an open source platform for building versioned Data Lakehouses in AWS with Python and the AWS CDK.

With one CLI command and an AWS account, you can spin up a world class Data Platform in your own AWS account with Git versioning of Data, SageMaker Domains, Spark Clusters, Streamlit Sites and more.

Maintain your Data Lake just like you do your code - with git! Leverage branches, tags and commits to version your table schemas and data. Provide a consistent view of the data for consumers, enable rapid experimentation and roll back mistakes with ease.

Development of ETL or ML training jobs is made easy with the yak CLI. Easily set up remote sessions on your Spark, Ray or Dask clusters to enjoy the power of cloud computing without giving up the experience of your local IDE.

How it Works

PackYak combines modern Software Development, Cloud Engineering and Data Engineering practices into one Python framework:

  1. Git-like versioning of Data Tables with Project Nessie - no more worrying about the version of data, simply use branches, tags and commits to freeze data or roll back mistakes.
  2. Software-defined Assets (as seen in Dagster) - think of your data pipelines in terms of the data it produces. Greatly simplify how data is produced, modified over time and backfilled in the event of errors.
  3. Infrastructure-as-Code (AWS CDK and Pulumi) - deploy in minutes and manage it all yourself with minimal effort.
  4. Apache Spark - write your ETL as simple python processes that are then scaled automatically over a managed AWS EMR Spark Cluster.
  5. Streamlit - build Streamlit applications that integrate the Data Lakehouse and Apache Spark to provide interactive reports and exploratory tools over the versioned data lake.

Get Started

Install Docker

If you haven't already, install Docker.

Install Python Poetry & Plugins

# Install the Python Poetry CLI
curl -sSL https://install.python-poetry.org | python3 -

# Add the export plugin to generate narrow requirements.txt
poetry self add poetry-plugin-export

Install the packyak CLI:

pip install packyak

Create a new Project

packyak new my-project
cd ./my-project

Deploy to AWS

poetry run cdk deploy

Git-like Data Catalog (Project Nessie)

PackYak comes with a Construct for hosting a Project Nessie catalog that supports Git-like versioning of the tables in a Data Lakehouse.

It deploys with an AWS DynamoDB Versioned store and an API hosted in AWS Lambda or AWS ECS. The Nessie Server is stateless and can be scaled easily with minimal-to-zero operational overhead.

Create a NessieDynamoDBVersionStore

from packyak.aws_cdk import DynamoDBNessieVersionStore

versionStore = DynamoDBNessieVersionStore(
  scope=stack,
  id="VersionStore",
  versionStoreName="my-version-store",
)

Create a Bucket to store Data Tables (e.g. Parquet files). This will store the "Repository"'s data.

myRepoBucket = Bucket(
  scope=stack,
  id="MyCatalogBucket",
)

Create the Nessie Catalog Service

# hosted on AWS ECS
myCatalog = NessieECSCatalog(
  scope=stack,
  id="MyCatalog",
  vpc=vpc,
  warehouseBucket=myRepoBucket,
  catalogName=lakeHouseName,
  versionStore=versionStore,
)

Create a Branch

Branch off the main branch of data into a dev branch to "freeze" the data as of a particular commit

CREATE BRANCH dev FROM main

Deploy a Spark Cluster

Create an EMR Cluster for processing data

spark = Cluster(
  scope=stack,
  id="Spark",
  clusterName="my-cluster",
  vpc=vpc,
  catalogs={
    # use the Nessie Catalog as the default data catalog for Spark SQL queries
    "spark_catalog": myCatalog,
  },
  installSSMAgent=true,
)

SSH into the Spark Cluster

yak ssh makes it easy to develop on AWS EMR, SageMaker and EC2 instances using your local VS Code IDE by facilitating SSH connections to the host over AWS SSM without complicated networking rules or bastion hosts. Everything is secured by AWS IAM.

yak ssh {ec2-instance-id}

Initialize a SparkSession

To create a SparkSession using your code reposistory's .venv file instead of the system one, use the init_session helper:

from packyak.spark import init_session

spark = init_session()

[!TIP] This is usually found in the first cell of a Jupyter notebook.

If you want to customize the SparkSession further, use session_builder instead:

spark = init_session().getOrCreate()

Remote VS Code over SSH

Once connected to a remote host, you can use VS Code's Remote SSH to start editing code and running commands on the remote host with the comfort of your local VS Code IDE.

  1. SSH in and forward port 22 to a local port of your choice:
yak ssh {ec2-instance-id} -L 9001:localhost:22
  1. Configure the remote host in your .ssh/config:
Host emr
  HostName localhost
  Port 9001
  User root
  IdentityFile ~/.ssh/id_rsa

Configure SparkSQL to be served over JDBC

sparkSQL = spark.jdbc(port=10001)

Deploy Streamlit Site

Stand up a Streamlit Site to serve interactive reports and applications over your data.

site = StreamlitSite(
  scope=stack,
  # Point it at the Streamlit site entrypoint
  home="app/home.py",
  # Where the Streamlit pages/tabs are, defaults to `dirname(home)/pages/*.py`
  # pages="app/pages"
)

Deploy to AWS

packyak deploy

Or via the AWS CDK CLI:

poetry run cdk deploy

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

packyak-0.4.22.tar.gz (26.5 kB view hashes)

Uploaded Source

Built Distribution

packyak-0.4.22-py3-none-any.whl (37.6 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page