AsyncIO serving for data science models
Project description
Foxcross
AsyncIO serving for data science models built on Starlette
Documentation: https://www.foxcross.dev/
Requirements: Python 3.6+
Quick Start
Installation using pip
:
pip install foxcross
Create some test data and a simple model in the same directory to be served:
data.json
[1,2,3,4,5]
models.py
from foxcross.serving import ModelServing, run_model_serving
class AddOneModel(ModelServing):
test_data_path = "data.json"
def predict(self, data):
return [x + 1 for x in data]
if __name__ == "__main__":
run_model_serving()
Run the model locally:
python models.py
Navigate to localhost:8000/predict-test/
, and you should see the list incremented by 1.
You can visit localhost:8000/
to see all the available routes for your model.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
foxcross-0.4.0.tar.gz
(7.7 kB
view hashes)