Skip to main content

Aims to be the Simplest Machine Learning Model Inference Server

Project description

Pinferencia

Simple, but Powerful.

Language grade: Python PyPI PyPI - Python Version

Pinferencia


English Doc | Seriously, Doc | 中文文档 | 正襟危坐版文档

Help wanted. Translation, rap lyrics, all wanted. Feel free to create an issue.


Pinferencia tries to be the simplest AI model inference server ever!

Serving a model with REST API has never been so easy.

If you want to

  • find a simple but robust way to serve your model
  • write minimal codes while maintain controls over you service
  • avoid any heavy-weight solutions
  • easily to integrate with your CICD
  • make your model and service portable and runnable across machines

You're at the right place.

Features

Pinferencia features include:

  • Fast to code, fast to go alive. Minimal codes needed, minimal transformation needed. Just based on what you have.
  • 100% Test Coverage: Both statement and branch coverages, no kidding.
  • Easy to use, easy to understand.
  • Automatic API documentation page. All API explained in details with online try-out feature.
  • Serve any model, even a single function can be served.

Install

pip install "pinferencia[uvicorn]"

Quick Start

Serve Any Model

from pinferencia import Server


class MyModel:
    def predict(self, data):
        return sum(data)


model = MyModel()

service = Server()
service.register(
    model_name="mymodel",
    model=model,
    entrypoint="predict",
)

Just run:

uvicorn app:service --reload

Hooray, your service is alive. Go to http://127.0.0.1/ and have fun.

Any Deep Learning Models? Just as easy. Simple train or load your model, and register it with the service. Go alive immediately.

Pytorch

import torch

from pinferencia import Server


# train your models
model = "..."

# or load your models (1)
# from state_dict
model = TheModelClass(*args, **kwargs)
model.load_state_dict(torch.load(PATH))

# entire model
model = torch.load(PATH)

# torchscript
model = torch.jit.load('model_scripted.pt')

model.eval()

service = Server()
service.register(
    model_name="mymodel",
    model=model,
)

Tensorflow

import tensorflow as tf

from pinferencia import Server


# train your models
model = "..."

# or load your models (1)
# saved_model
model = tf.keras.models.load_model('saved_model/model')

# HDF5
model = tf.keras.models.load_model('model.h5')

# from weights
model = create_model()
model.load_weights('./checkpoints/my_checkpoint')
loss, acc = model.evaluate(test_images, test_labels, verbose=2)

service = Server()
service.register(
    model_name="mymodel",
    model=model,
    entrypoint="predict",
)

Any model of any framework will just work the same way. Now run uvicorn app:service --reload and enjoy!

Contributing

If you'd like to contribute, details are here

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pinferencia-0.1.0rc1.tar.gz (16.5 kB view details)

Uploaded Source

Built Distribution

pinferencia-0.1.0rc1-py3-none-any.whl (23.6 kB view details)

Uploaded Python 3

File details

Details for the file pinferencia-0.1.0rc1.tar.gz.

File metadata

  • Download URL: pinferencia-0.1.0rc1.tar.gz
  • Upload date:
  • Size: 16.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.1.13 CPython/3.8.9 Darwin/21.1.0

File hashes

Hashes for pinferencia-0.1.0rc1.tar.gz
Algorithm Hash digest
SHA256 42908615fd953beee4dd6dea2c2d6931d0dc773e62c1bdc7f1c8b40f402173c5
MD5 38eb977344945bf898968fcb4b668499
BLAKE2b-256 5dc4ccec5208bb56c403a9a369ff2b2ec73e00bcb9a6bdd8a234c11ee6b19ad9

See more details on using hashes here.

File details

Details for the file pinferencia-0.1.0rc1-py3-none-any.whl.

File metadata

  • Download URL: pinferencia-0.1.0rc1-py3-none-any.whl
  • Upload date:
  • Size: 23.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.1.13 CPython/3.8.9 Darwin/21.1.0

File hashes

Hashes for pinferencia-0.1.0rc1-py3-none-any.whl
Algorithm Hash digest
SHA256 dd92081b03a9ab527c730b991e9237ddb82bd40a1a23b4e1020d8c6adce42ccf
MD5 2e38fb81ca6d24313fec480aeddde8c1
BLAKE2b-256 6154d36708a86faae286d2ae6d8664f3fec84d9f2edab8f77be6386d08bdade3

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page