OnnxSlim: A Toolkit to Help Optimize Large Onnx Model
Project description
OnnxSlim
OnnxSlim can help you slim your onnx model, with less operators, but same accuracy, better inference speed.
- 🚀 OnnxSlim is merged to mnn-llm, performance increased by 5%
Installation
Using Prebuilt
pip install onnxslim
Build From Source
pip install .
How to use
onnxslim your_onnx_model slimmed_onnx_model
For more usage, see onnxslim -h or refer to our examples
References
Contact
Discord: https://discord.gg/nRw2Fd3VUS
QQ Group: 873569894
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
onnxslim-0.1.22.tar.gz
(65.2 kB
view hashes)
Built Distribution
onnxslim-0.1.22-py3-none-any.whl
(75.4 kB
view hashes)
Close
Hashes for onnxslim-0.1.22-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 179a3f1d8d999e63379ea24917758865601d0e845fd07933a9f8e1b0fe74d587 |
|
MD5 | d77ce553c156fca910488237d6b101da |
|
BLAKE2b-256 | e99885b36fb61a0ffa5eee0ab586599453bb1faef7c4f7a479018ebebf479b02 |