bentoml: A framework for machine learning model serving
Project description
Model Serving Made Easy
BentoML let you create machine learning powered prediction service in minutes and bridges the gap between data science and DevOps.
Why BentoML
- The easiest way to get your ML models into production.
- High performance model serving, all in Python.
- Package your model once and deploy it anywhere.
- Support all major ML model training frameworks.
Getting Started
- Quickstart guide will show you a simple example of using BentoML in action. In under 10 minutes, you'll be able to serve your ML model over an HTTP API endpoint, and build a docker image that is ready to be deployed in production.
- Main concepts will give a comprehensive tour of BentoML's components and introduce you to its philosophy. After reading, you will see what drives BentoML's design, and know what
bento
andrunner
stands for. - Playground notebook gets your hands dirty in a notebook environment, for you to try out all the core features in BentoML.
- ML Frameworks lays out best practices and example usages by the ML framework used for training models.
- Advanced Guides show cases advanced features in BentoML, including GPU support, inference graph, monitoring, and customizing docker environment etc.
Community
- To report a bug or suggest a feature request, use GitHub Issues.
- For other discussions, use Github Discussions.
- To receive release announcements, please subscribe to our mailing list or join us on Slack.
Contributing
There are many ways to contribute to the project:
- If you have any feedback on the project, share it with the community in Github Discussions of this project.
- Report issues you're facing and "Thumbs up" on issues and feature requests that are relevant to you.
- Investigate bugs and reviewing other developer's pull requests.
- Contributing code or documentation to the project by submitting a Github pull request. See the development guide.
- See more in the contributing guide.
Usage Reporting
BentoML by default collects anonymous usage data using Amplitude. It only collects BentoML library's own actions and parameters, no user or model data will be collected. Here is the code that does it.
This helps the BentoML team to understand how the community is using this tool and what to build next.
You can easily opt-out of usage tracking by running the BentoML commands with the --do-not-track
option.
> bentoml [command] --do-not-track
You can also opt-out via setting environment variable BENTOML_DO_NOT_TRACK=True
> export BENTOML_DO_NOT_TRACK=True
License
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file bentoml-1.0.0.dev0.tar.gz
.
File metadata
- Download URL: bentoml-1.0.0.dev0.tar.gz
- Upload date:
- Size: 493.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.7.0 importlib_metadata/4.8.2 pkginfo/1.8.2 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.8.12
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 3f3d35028985c738f44e1f13ac21c63aee2a109319f11a7e696fe3e51b4a09cd |
|
MD5 | 126db9b9d025339de94a265183339dae |
|
BLAKE2b-256 | 32c711006e7f6c004b7a6db0e21437882b282d0b9ddcdd98239cb9ec638b02f4 |
File details
Details for the file bentoml-1.0.0.dev0-py3-none-any.whl
.
File metadata
- Download URL: bentoml-1.0.0.dev0-py3-none-any.whl
- Upload date:
- Size: 567.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.7.0 importlib_metadata/4.8.2 pkginfo/1.8.2 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.8.12
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | b493e3ec76a4eee29fda0a04eedebb12fcc27f31f0ffb5046eed519918f722f8 |
|
MD5 | 6a42b912223278b3a6f6fe04c3394058 |
|
BLAKE2b-256 | 2ee05c8aec43ea9e82dcbf217484040fbbaff982b317a66f38b7ca5d750042f9 |