Skip to main content

Machine Learning Operations Toolkit

Project description

Version Python version License Documentation Status

Tempo: The MLOps Software Development Kit


An open source framework to enable data scientists to productionise, test and deploy models with simple workflows that abstract the underlying complexity of scalable MLOps platforms.


Tempo provides a unified interface to multiple MLOps projects that enable data scientists to deploy and productionise machine learning systems.

  • Package your trained model artifacts to optimized server runtimes (Tensorflow, PyTorch, Sklearn, XGBoost etc)
  • Package custom business logic to production servers.
  • Build an inference pipeline of models and orchestration steps.
  • Include any custom python components as needed. Examples:
    • Outlier detectors with Alibi-Detect.
    • Explainers with Alibi-Explain.
  • Test Locally - Deploy to Production
    • Run with local unit tests.
    • Deploy locally to Docker to test with Docker runtimes.
    • Deploy to production on Kubernetes
    • Extract declarative Kubernetes yaml to follow GitOps workflows.
  • Supporting Seldon production runtimes
    • Seldon Core open source
    • Seldon Deploy enterprise
  • Create stateful services. Examples:
    • Multi-Armed Bandits.


  1. Develop locally.
  2. Test locally on Docker with production artifacts.
  3. Push artifacts to remote bucket store and launch remotely (on Kubernetes).


Motivating Synopsis

Data scientists can easily test their models and orchestrate them with pipelines.

Below we see two Models (sklearn and xgboost) with a function decorated pipeline to call both.

def get_tempo_artifacts(artifacts_folder: str) -> Tuple[Pipeline, Model, Model]:

    sklearn_model = Model(

    xgboost_model = Model(

        models=PipelineModels(sklearn=sklearn_model, xgboost=xgboost_model),
    def classifier(payload: np.ndarray) -> Tuple[np.ndarray, str]:
        res1 = classifier.models.sklearn(input=payload)

        if res1[0] == 1:
            return res1, SKLearnTag
            return classifier.models.xgboost(input=payload), XGBoostTag

    return classifier, sklearn_model, xgboost_model

Save the pipeline code.

from tempo.serve.loader import save

Deploy locally to docker.

from tempo import deploy_local
remote_model = deploy_local(classifier)

Make predictions on containerized servers that would be used in production.

remote_model.predict(np.array([[1, 2, 3, 4]]))

Deploy to Kubernetes for production.

from tempo.serve.metadata import SeldonCoreOptions
from tempo import deploy_remote

runtime_options = SeldonCoreOptions(**{
    "remote_options": {
        "namespace": "production",
        "authSecretName": "minio-secret"
remote_model = deploy_remote(classifier, options=runtime_options)

This is an extract from the multi-model introduction demo.

Project details

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

mlops-tempo-0.5.2.tar.gz (45.0 kB view hashes)

Uploaded source

Built Distribution

mlops_tempo-0.5.2-py3-none-any.whl (75.0 kB view hashes)

Uploaded py3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page