OnnxSlim: A Toolkit to Help Optimize Large Onnx Model
Project description
OnnxSlim
OnnxSlim can help you slim your onnx model, with less operators, but same accuracy, better inference speed.
- 🚀 OnnxSlim is merged to mnn-llm, performance increased by 5%
- 🚀 Rank 1st in the AICAS 2024 LLM inference optimization challenge held by Arm and T-head
- 🚀 OnnxSlim is merged into ultralytics ❤️❤️❤️
- 🚀 OnnxSlim is merged into transformers.js 🤗🤗🤗
Installation
Using Prebuilt
pip install onnxslim
Install From Source
pip install git+https://github.com/inisis/OnnxSlim@main
Install From Local
git clone https://github.com/inisis/OnnxSlim && cd OnnxSlim/
pip install .
How to use
onnxslim your_onnx_model slimmed_onnx_model
For more usage, see onnxslim -h or refer to our examples
References
Contact
Discord: https://discord.gg/nRw2Fd3VUS QQ Group: 873569894
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
onnxslim-0.1.33.tar.gz
(119.1 kB
view hashes)
Built Distribution
onnxslim-0.1.33-py3-none-any.whl
(140.5 kB
view hashes)
Close
Hashes for onnxslim-0.1.33-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | c3b5874524a9de3df38bd8a68adde6ade8d33156930dcf3d9ef02f3ed7fad406 |
|
MD5 | c5fb951217dfec56011270d2ef4b759c |
|
BLAKE2b-256 | 0efc3ad1127d2fbe152fb53c2fc28ba94547fead44ae78d0075e92203852793d |