An Easy-to-Use and High-Performance AI deployment framework
Project description
nndeploy: An Easy-to-Use, and High-Performance AI Deployment Framework
Introduction
nndeploy is an easy-to-use, and high-performance AI deployment framework. Based on the design concepts of visual workflows and multi-backend inference, developers can quickly develop SDKs for specified platforms and hardware from algorithm repositories, significantly saving development time. Furthermore, the framework has already deployed numerous AI models including LLM, AIGC generation, face swap, object detection, image segmentation, etc., ready to use out-of-the-box.
Simple and Easy to Use
- Visual Workflow: Deploy AI algorithms through drag-and-drop operations. Visually adjust all node parameters of the AI algorithm in the frontend and quickly preview the effect after parameter tuning.
- Custom Nodes: Support Python/C++ custom nodes, seamlessly integrated into the visual interface without frontend code.
- Algorithm Combination: Flexibly combine different algorithms to quickly build innovative AI applications.
- One-Click Deployment: The completed workflow can be exported as a JSON configuration file with one click, supporting direct calls via Python/C++ API, achieving seamless transition from development to production environments, and fully supporting platforms like Linux, Windows, macOS, Android, iOS, etc.
High Performance
-
Parallel Optimization: Supports execution modes like serial, pipeline parallel, task parallel, etc.
-
Memory Optimization: Optimization strategies like zero-copy, memory pool, memory reuse, etc.
-
High-Performance Optimization: Built-in nodes optimized with C++/CUDA/Ascend C/SIMD, etc.
-
Multi-Backend Inference: One workflow, multiple backend inference. Integrates 13 mainstream inference frameworks with zero abstraction cost, covering all platforms including cloud, desktop, mobile, edge, etc.
Inference Framework Application Scenario Status ONNXRuntime Cross-platform inference ✅ TensorRT NVIDIA GPU high-performance inference ✅ OpenVINO Intel CPU/GPU optimization ✅ MNN Mobile inference engine by Alibaba ✅ TNN Mobile inference engine by Tencent ✅ ncnn Mobile inference engine by Tencent ✅ CoreML iOS/macOS native acceleration ✅ AscendCL Huawei Ascend AI chip inference framework ✅ RKNN Rockchip NPU inference framework ✅ SNPE Qualcomm Snapdragon NPU inference framework ✅ TVM Deep learning compiler stack ✅ PyTorch Rapid prototyping / Cloud deployment ✅ Self-developed Inference Framework Default inference framework ✅
Out-of-the-Box Algorithms
List of deployed models, with 100+ nodes created. We will continue to deploy more high-value AI algorithms. If you have algorithms you need deployed, please let us know via issue.
| Application Scenario | Available Models | Remarks |
|---|---|---|
| Large Language Model | QWen-2.5, QWen-3 | |
| Image Generation | Stable Diffusion 1.5, Stable Diffusion XL, Stable Diffusion 3, HunyuanDiT, etc. | Supports text-to-image, image-to-image, inpainting; based on diffusers |
| Face Swap | deep-live-cam | |
| OCR | Paddle OCR | |
| Object Detection | YOLOv5, YOLOv6, YOLOv7, YOLOv8, YOLOv11, YOLOx | |
| Object Tracking | FairMot | |
| Image Segmentation | RBMGv1.4, PPMatting, Segment Anything | |
| Classification | ResNet, MobileNet, EfficientNet, PPLcNet, GhostNet, ShuffleNet, SqueezeNet | |
| API Service | OPENAI, DeepSeek, Moonshot | Supports LLM and AIGC services |
See more details in the Deployed Model List Details
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distributions
Built Distributions
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file nndeploy-3.0.9-cp312-cp312-manylinux_2_28_x86_64.whl.
File metadata
- Download URL: nndeploy-3.0.9-cp312-cp312-manylinux_2_28_x86_64.whl
- Upload date:
- Size: 98.6 MB
- Tags: CPython 3.12, manylinux: glibc 2.28+ x86-64
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
4f62ea7f0177a07ba9be07d43c207b19798ade20f09d6f179f47a5b599d2654f
|
|
| MD5 |
42b035418a40162aa5f75b45c83047dd
|
|
| BLAKE2b-256 |
3b2ca0d1847de437334411c9070fb90cda7b0d9ebb4c8c9bb0a38ed1a2ff2b34
|
File details
Details for the file nndeploy-3.0.9-cp311-cp311-manylinux_2_28_x86_64.whl.
File metadata
- Download URL: nndeploy-3.0.9-cp311-cp311-manylinux_2_28_x86_64.whl
- Upload date:
- Size: 97.1 MB
- Tags: CPython 3.11, manylinux: glibc 2.28+ x86-64
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
2aad87b4e2d9a7bd26e4af5b10bd2697273316308b601c3aaa83e79fd1253003
|
|
| MD5 |
7d2a2786bfa76950de42ebe330a86489
|
|
| BLAKE2b-256 |
5619b4879ccb2718e958eda5ba5116cc362754f41af805583838cbe97ea8eef5
|
File details
Details for the file nndeploy-3.0.9-cp310-cp310-manylinux_2_28_x86_64.whl.
File metadata
- Download URL: nndeploy-3.0.9-cp310-cp310-manylinux_2_28_x86_64.whl
- Upload date:
- Size: 95.6 MB
- Tags: CPython 3.10, manylinux: glibc 2.28+ x86-64
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
7b90230e2dd99f599bb91d5f8c846cc75ff81faff068105440c8340e99e1d7da
|
|
| MD5 |
b6dcc35b90447534b14cfce04791ae2b
|
|
| BLAKE2b-256 |
5bc97c2464b571ceec8bf8bed1c0d7cda3eb6876b3ca2269ebe12ebb1445da0b
|