Deploy FastAPI applications on MLFlow
Project description
mlrpc [Experimental]
Host REST APIs via FastAPI in Databricks Serverless Model serving via this rpc abstraction.
MLRPC does not modify your FastAPI code! It acts as a build step/proxy layer and invokes your FastAPI code in a databricks serverless model serving environment and routes to the right endpoints properly without you having to know about custom python models.
This is currently experimental in nature. If interest increases, will work on hardening and ensure it has a stable build and tests.
"If you know fastapi, you can deploy models on databricks model serving!"
Why FastAPI?
It is a modern, fast (high-performance), web framework for building APIs with Python 3.7+ based on standard Python type hints and pydantic. It is a great way to build APIs and has a lot of features that make it easy to use.
https://pypistats.org/packages/fastapi
Key features
- FastAPI: Host your FastAPI code in databricks
- Generate OpenAPI client objects
- Swagger UI to explore your endpoint
- Hot reload your FastAPI code to the remote endpoint
- Hot reloads of code are encrypted
- No need to manage your own pyfunc model
- Full support for environments (dev, test, prod) and interactive development
Limitations
- Only supports FastAPI
- Requires mlrpc client or follow spec to query the endpoint
- Does not yet support FastAPI lifecycle events
- Does not support websockets (model serving does not support websockets)
- Does not support SSE (server side events)
Installation
Use the cli to install the package
pip install -U 'mlrpc[cli]'
Instructions
Keep in mind that the cli has a lot of options but all of them are also managable via the config file.
1. Make a config file
mlrpc init
2. Edit the config file
## The following lines are minimum required config
## You need to have the app section
# [app]
# name = <app name>
# uc_catalog = <catalog name>
# uc_schema = <schema name>
# endpoint_name = <endpoint name>
# data_dir=data # this is optional if you want to upload larger binaries like chroma, sqlite, faiss, lancedb, etc
# Cost controls
# size = Small # Small, Medium, Large
# scale_to_zero_enabled = true # true, false
## The following lines are optional and not required
# latest_alias_name = <alternative alias> uses "current" by default
# experiment_name = <experiment name> # optional specify your experiment name
## Define where your code is
# app_root_dir = <root of your directory>
## The following lines are optional if you need secrets
# secret_scope = <secret scope in databricks>
# secret_key = <secret key>
# env_file = <location of your env file relative to this file>
## use this setting if you want to delete old versions
# only_last_n_versions=10
3. Deploy the artifacts
mlrpc deploy -p <databricks profile>
4. [OPTIONAL] verify in valid model deployment server locally
This may not work for you if your models require gpus or other resources that wont exist on your laptop
mlrpc local -p <databricks profile>
5. Deploy to model serving infra
mlrpc serve -p <databricks profile>
6. Explore the deployed endpoint via swagger proxy and hot reloading
mlrpc swagger -p <databricks profile>
Disclaimer
mlrpc is not developed, endorsed not supported by Databricks. It is provided as-is; no warranty is derived from using this package. For more details, please refer to the license.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.