Skip to main content

Deploy your model in one line of code

Project description

Deploy your ML pipelines effortlessly, scalably and reliably

Datarmada aims at removing all the friction that comes with Machine Learning in production. We understand that Data Scientists are not trained to do that, and sometimes they are not even attracted by this Software Engineering / DevOps aspect.

This package aims at deploying your machine learning pipeline on a server in one line.

Your pipeline is deployed on an OVH server so that you own your data and it is compliant with European regulations.

Installation

Install the package python using pip

pip install auto-mlops

Deploy your pipeline

Import the Deployer class from the package.

from auto_mlops import Deployer
deployer = Deployer()

Now, deploy your pipeline by passing to the deploy method a list containing all of its elements. The pipeline elements (except for the last one) must be either :

  • A function returning transformed data if your pipeline element doesn't need to be fitted
  • An instance of a class implementing a transform method

The last element of the pipeline must be an instance of a class a predict methods, such as a scikit-learn or a Keras model.

from sklearn.linear_model import LogisticRegression

def preprocess(raw_data):
    # preprocess the data
    return preprocessed_data

class Featurizer:
    def transform(self, preprocessed_data):
        # transform the data
        return featurized_data

log_reg = LogisticRegression()
log_reg.fit(featurized_data, y)

deployer.deploy([preprocess, featurizer, log_reg])

Remember your elements must be fitted if they need to !

You will be asked for your email address so that we can keep track of the ownership of the pipelines deployed, and give you access to monitoring functions in the future.

deployer.deploy([preprocess, featurizer, log_reg])

>> Please enter your email address so that we can keep track of your pipelines:
you@example.com

>> Your pipeline has been deployed to https://cloud.datarmada.com/id

You can access your route whenever you want through deployer.route

Make predictions

You can now send data to the route by making a POST request as following

import requests

res = requests.post(
  "https://cloud.datarmada.com/id",
  json = {
    "data": your_raw_data
  }
)

print(res.json())

>> { "prediction" : prediction }

It may be possible that one of the package you are using is not available in the environment we are deploying your model. If you receive an error saying so, please email us at contact@datarmada.com so that we can fix it.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

auto-mlops-0.1.12.tar.gz (3.5 kB view details)

Uploaded Source

Built Distribution

auto_mlops-0.1.12-py3-none-any.whl (4.0 kB view details)

Uploaded Python 3

File details

Details for the file auto-mlops-0.1.12.tar.gz.

File metadata

  • Download URL: auto-mlops-0.1.12.tar.gz
  • Upload date:
  • Size: 3.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.0.3 CPython/3.7.4 Linux/5.3.0-40-generic

File hashes

Hashes for auto-mlops-0.1.12.tar.gz
Algorithm Hash digest
SHA256 67e67c8ba673d3002add5eea05fca9ab8bf8f2077c1d6ffeea475d05d94fb256
MD5 1bc25fb2341904ce063bcd718384b5a8
BLAKE2b-256 0d25e43817dbe1cb76fb427f997036918bd950bc5d22b360ccf0c1ae65ec7a6d

See more details on using hashes here.

File details

Details for the file auto_mlops-0.1.12-py3-none-any.whl.

File metadata

  • Download URL: auto_mlops-0.1.12-py3-none-any.whl
  • Upload date:
  • Size: 4.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.0.3 CPython/3.7.4 Linux/5.3.0-40-generic

File hashes

Hashes for auto_mlops-0.1.12-py3-none-any.whl
Algorithm Hash digest
SHA256 8ae761c8f1e646fe0b5a252113c0cb5e20086abd746eeb6ecc1f105ef0a7ebde
MD5 8e69a5d5ca07777d5c86170e92fe9ef7
BLAKE2b-256 5e703c8277d75730a95e33e8e7ab78cd719b5d998a894877dc69e6299b9f78fe

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page