The banana package is a python client to interact with your machine learning models hosted on Banana
Project description
server
An HTTP server designed for AI
Quickstart
Install the Banana server package
pip3 install banana-server
Create a python file called main.py
with this:
from banana_server import Banana
app = Banana("server")
@app.init
def init():
model = "my pytorch model"
app.optimize(model)
return app.set_cache({
"model": model,
"hello": "world"
})
@app.handler
def handler(cache, json_in) -> dict:
print("cache:", cache)
print("json_in:", json_in)
return {"json": "out"}
app.serve()
Run it with
python3 main.py
Test the running server with
curl -X POST -H "Content-Type: application/json" -d '{"name":"test"}' http://localhost:8000
Documentation
banana_server.Banana
from banana_server import Banana
app = Banana("server")
This instantiates your HTTP app, similar to popular frameworks like Flask
This HTTP server is production-ready out of the box.
@app.init
@app.init
def init():
model = "my pytorch model"
app.optimize(model)
return app.set_cache({
"model": model,
"hello": "world"
})
The @app.init
decorated function runs once on server startup, and is used to load any reuseable, heavy objects such as:
- Your AI model, loaded to GPU
- Tokenizers
- Precalculated embeddings
Once initialized, you must save those variables to the cache with app.set_cache({})
so they can be referenced later.
There may only be one @app.init
function.
@app.handler
@app.handler
def handler(cache, json_in) -> dict:
print("cache:", cache)
print("json_in:", json_in)
return {"json": "out"}
The @app.handler
decorated function runs for every http call, and is used to run inference or training workloads against your model(s).
Arg | Type | Description |
---|---|---|
cache | dict | The app's cache, set with set_cache() |
json_in | dict | The json body of the input call. If using the Banana client SDK, this is the same as model_inputs |
Return Val | Type | Description |
---|---|---|
json_out | dict | The json body to return to the client. If using the Banana client SDK, this is the same as model_outputs |
There may only be one @app.handler
function.
app.set_cache()
app.set_cache({})
app.set_cache
saves the input dictionary to the app's cache, for reuse in future calls. It may be used in both the @app.init
and @app.handler
functions.
app.set_cache
overwrites any preexisting cache.
app.get_cache()
cache = app.get_cache()
app.get_cache
fetches the dictionary to the app's cache. This value is automatically provided for you as the cache
argument in the @app.handler
function.
app.optimize(model)
model # some pytorch model
app.optimize(model)
app.optimize
is a feature specific to users hosting on Banana's serverless GPU infrastructure. It is run during buildtime rather than runtime, and is used to locate the model(s) to be targeted for Banana's Fastboot optimization.
Multiple models may be optimized. Only Pytorch models are currently supported.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file banana_server-0.0.1.tar.gz
.
File metadata
- Download URL: banana_server-0.0.1.tar.gz
- Upload date:
- Size: 7.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.9.15
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 1d62dc83ccbe890a54714d8adec49b2523d36dc43e60e1f60690328134baef5b |
|
MD5 | 51850890424f6b39d900373ea0e417d9 |
|
BLAKE2b-256 | dda84d1f5b2301b4c5af841af2afc54a8abc52e5f5dacfea64cbe3b30224a587 |
File details
Details for the file banana_server-0.0.1-py3-none-any.whl
.
File metadata
- Download URL: banana_server-0.0.1-py3-none-any.whl
- Upload date:
- Size: 7.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.9.15
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 9188af5d1f56c4f7caaebc40a9a7ea90b8ea7ea052aec9b977d7468fe0a2b12f |
|
MD5 | b78558f1ad7ce811db4673580842933d |
|
BLAKE2b-256 | b7dc145db7bf4d680e65b6abd395403609af5cf948a150582e48c0ccdf1a1aff |