Skip to main content

The banana package is a python client to interact with your machine learning models hosted on Banana

Project description

server

An HTTP server designed for AI

Quickstart

Install the Banana server package

pip3 install banana-server

Create a python file called main.py with this:

from banana_server import Banana

app = Banana("server")

@app.init
def init():
    model = "my pytorch model"
    app.optimize(model)

    return app.set_cache({
        "model": model,
        "hello": "world"
    })

@app.handler
def handler(cache, json_in) -> dict:
    print("cache:", cache)
    print("json_in:", json_in)
    return {"json": "out"}

app.serve()

Run it with

python3 main.py

Test the running server with

curl -X POST -H "Content-Type: application/json" -d '{"name":"test"}' http://localhost:8000

Documentation

banana_server.Banana

from banana_server import Banana

app = Banana("server")

This instantiates your HTTP app, similar to popular frameworks like Flask

This HTTP server is production-ready out of the box.

@app.init

@app.init
def init():
    model = "my pytorch model"
    app.optimize(model)

    return app.set_cache({
        "model": model,
        "hello": "world"
    })

The @app.init decorated function runs once on server startup, and is used to load any reuseable, heavy objects such as:

  • Your AI model, loaded to GPU
  • Tokenizers
  • Precalculated embeddings

Once initialized, you must save those variables to the cache with app.set_cache({}) so they can be referenced later.

There may only be one @app.init function.

@app.handler

@app.handler
def handler(cache, json_in) -> dict:
    print("cache:", cache)
    print("json_in:", json_in)
    return {"json": "out"}

The @app.handler decorated function runs for every http call, and is used to run inference or training workloads against your model(s).

Arg Type Description
cache dict The app's cache, set with set_cache()
json_in dict The json body of the input call. If using the Banana client SDK, this is the same as model_inputs
Return Val Type Description
json_out dict The json body to return to the client. If using the Banana client SDK, this is the same as model_outputs

There may only be one @app.handler function.

app.set_cache()

app.set_cache({})

app.set_cache saves the input dictionary to the app's cache, for reuse in future calls. It may be used in both the @app.init and @app.handler functions.

app.set_cache overwrites any preexisting cache.

app.get_cache()

cache = app.get_cache()

app.get_cache fetches the dictionary to the app's cache. This value is automatically provided for you as the cache argument in the @app.handler function.

app.optimize(model)

model # some pytorch model
app.optimize(model)

app.optimize is a feature specific to users hosting on Banana's serverless GPU infrastructure. It is run during buildtime rather than runtime, and is used to locate the model(s) to be targeted for Banana's Fastboot optimization.

Multiple models may be optimized. Only Pytorch models are currently supported.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

banana_server-0.0.1.tar.gz (7.5 kB view details)

Uploaded Source

Built Distribution

banana_server-0.0.1-py3-none-any.whl (7.7 kB view details)

Uploaded Python 3

File details

Details for the file banana_server-0.0.1.tar.gz.

File metadata

  • Download URL: banana_server-0.0.1.tar.gz
  • Upload date:
  • Size: 7.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.15

File hashes

Hashes for banana_server-0.0.1.tar.gz
Algorithm Hash digest
SHA256 1d62dc83ccbe890a54714d8adec49b2523d36dc43e60e1f60690328134baef5b
MD5 51850890424f6b39d900373ea0e417d9
BLAKE2b-256 dda84d1f5b2301b4c5af841af2afc54a8abc52e5f5dacfea64cbe3b30224a587

See more details on using hashes here.

File details

Details for the file banana_server-0.0.1-py3-none-any.whl.

File metadata

File hashes

Hashes for banana_server-0.0.1-py3-none-any.whl
Algorithm Hash digest
SHA256 9188af5d1f56c4f7caaebc40a9a7ea90b8ea7ea052aec9b977d7468fe0a2b12f
MD5 b78558f1ad7ce811db4673580842933d
BLAKE2b-256 b7dc145db7bf4d680e65b6abd395403609af5cf948a150582e48c0ccdf1a1aff

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page