Skip to main content

Convert JupyterNotebooks to Kubeflow Pipelines deployments

Project description

Kale Logo


Kale is a Python package that aims at automatically deploy a general purpose Jupyter Notebook as a running Kubeflow Pipelines instance, without requiring the use the specific KFP DSL.

The general idea of kale is to automatically arrange the cells included in a notebook, and transform them into a unified KFP-compliant pipeline. To do so, the user is only required to decide which cells correspond to which pipeline step, by the use of tags. In this way, a researcher can better focus on building and testing its code locally, and then scale it in a simple, organized and controlled way.

Tagging language

Jupyter provides a tagging feature out-of-the-box, that lets you associate each cells with custom defined tags. The feature is available also in JupyterLab via the jupyterlab-celltags extension.

The tags are used to tell Kale how to convert the notebook's code cells into an execution graph, by specifying the execution dependencies between the pipeline steps and which code cells to merge together.

This is a list of tags recognized by Kale:

Tag Description Example
^block:<block_name>(;<block_name>)*$ Assign the current cell to a (multiple) pipeline step block:train-model
block:processing-A;processing-B
^prev:<block_name>(;<block_name>)*$ Define an execution dependency of the current cell to n other pipeline steps `prev:load-dataset
imports|functions Tell Kale to add this code block at the beginning of every pipeline code block. Useful to add imports/function to every pipeline step -
skip 'Hide' the current cell from Kale. -

Where <block_name> is matched against the regex: [a-z0-9]. So any string containing only digits and lowercase characters.

Installation

Kale is provided as a Python package. Just clone the repository to your local machine and install the package in a virtual environment.

# Clone the repo to your local environment
git clone https://github.com/kubeflow-kale/kale
cd kale
# Install the package in your virtualenv
python setup.py install

Getting Started

First you need to have a running Kubeflow instance (Kubeflow getting started guide).

Kale provides a CLI command. Run kale --help for a detailed description of the execution parameters.

Example:

kale --nb examples/base_example_numpy.ipynb \
	--pipeline_name numpy_example \
	--pipeline_descr "Numpy Example" \
	--docker_image stefanofioravanzo/kale-kfp-examples:0.1

This will produce a python script examples/kfp_numpy_examples.kfp.py containing all the definitions of the KFP stand-alone functions and the necessary code to define and deploy a pipeline. If run with the --deploy flag, Kale will try to automatically deploy the generated pipeline (default KFP url is localhost:8080, see kale CLI parameters for customization).

Some example Notebooks can be found in the examples repository.

Architecture

core.py contains the Kale class, main entry-point of the application. The function run does the conversion from jupyter notebook to kfp pipeline by calling all the other sub-modules.

Kale was developed with a modular design. There are 4 main modules, each with a specific functionality.

1. nbparser

This module both defines the Kale's Jupyter tagging language and provides the machinery to convert a tagged notebook to a NetworkX graph representing the resulting KFP pipeline. The main function is parse_notebook(), which takes as input a tagged notebook and returns the graph. The pipeline steps Python code is contained inside a special attributed of the graph's nodes.

2. static_analysis

The purpose of this module is to detect the data dependencies between the code snippets in the execution graph. This is achieved by running PyFlakes that is able to detect all the potential missing variables errors, plus some custom routines that iterate over the AST. dep_analysis.variables_dependencies_detection() is the main function called by Kale to start the static analysis process.

3. marshal

Marshal is a module that implements two dispater classes. PatternDispatcher can register functions based a regex pattern, and then call one of these functions based on a user provided string. TypeDispatcher can register functions based on a pattern as well, but dispatches a function based on the type of the input objects instead of a string.

Example:

from kale.marshal import resource_save, resource_load

# register functions to load and save numpy objects
@resource_load.register('.*\.npy')  # match anything ending in '.npy'
def resource_numpy_load(uri, **kwargs):
    return np.load(uri)

@resource_save.register('numpy\..*')  # match anything starting with 'numpy'
def resource_numpy_save(obj, path, **kwargs):
    return np.save(path+".npy", obj)

a = np.random.random((10,10))
resource_save(a, 'marshal_dir/')
b = resource_load('marshal_dir/a.npy')  # a == b

Common backends to support multiple data types are registered in marshal/backends.py. The idea is to provide backends for the most used data types and have a common fallback to standard dill serialization in case the data type is not recognized.

NB: The code in this module is not used directly by Kale during the conversion process from notebook to pipeline, but is instead injected at the beginning and at the end of each generated pipeline step to serialize and de-serialize objects that need to be passed between pipeline steps. Have a look at templates/function_template.txt and at the generated python script (e.g. kfp_numpy_example.kfp.py).

4. codegen

This module provides a single function gen_kfp_code that exploits the Jinja2 templating engine to generate a fully functional python script. The templates used are under the templates/ folder.

Flask Server

The package provides the kale_server CLI command that runs a Flask server. The Flask server accepts requests at localhost:5000/kale. The API accepts POST requests containing a Jupyter Notebook in raw format (JSON) and will call the main Kale module to deploy the notebook to a KFP instance.

kale/ API available parameters:

  • nb: Raw JSON Jupyter Notebook
  • deploy: Boolean Flag. True to deploy the Notebook automatically
  • kfp_port: Port of the running KFP instance
  • pipeline_name: Name of the resulting KFP pipeline
  • pipeline_descr: Description of the resulting KFP pipeline
  • docker_image: Docker image to use for the pipeline steps

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

kubeflow-kale-0.3.4.tar.gz (239.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

kubeflow_kale-0.3.4-py3-none-any.whl (30.9 kB view details)

Uploaded Python 3

File details

Details for the file kubeflow-kale-0.3.4.tar.gz.

File metadata

  • Download URL: kubeflow-kale-0.3.4.tar.gz
  • Upload date:
  • Size: 239.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.15.0 pkginfo/1.5.0.1 requests/2.22.0 setuptools/40.6.2 requests-toolbelt/0.9.1 tqdm/4.36.1 CPython/3.6.8

File hashes

Hashes for kubeflow-kale-0.3.4.tar.gz
Algorithm Hash digest
SHA256 bffc6681f2184827dbca9542463fe6ffd29b39d242c2829f7913658f6b4ed506
MD5 ae72a8558992eeb1a9886d6805825ea0
BLAKE2b-256 a10b911246e44e7940ae700f7bc54781e6926e3954e1c23657f7580b7b042336

See more details on using hashes here.

File details

Details for the file kubeflow_kale-0.3.4-py3-none-any.whl.

File metadata

  • Download URL: kubeflow_kale-0.3.4-py3-none-any.whl
  • Upload date:
  • Size: 30.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.15.0 pkginfo/1.5.0.1 requests/2.22.0 setuptools/40.6.2 requests-toolbelt/0.9.1 tqdm/4.36.1 CPython/3.6.8

File hashes

Hashes for kubeflow_kale-0.3.4-py3-none-any.whl
Algorithm Hash digest
SHA256 122f5fe9acf623895a0be2f8c3de3158d0d46a592e69b7b3c7766f311e08062d
MD5 70e6d01de3710bfb8883a95d83dfe1bf
BLAKE2b-256 5cb30d23579993ca812cb76ffac364e62ecf28192910b848bc1045263f38791f

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page