Skip to main content

An Easy-to-Use and High-Performance AI deployment framework

Project description

nndeploy: An Easy-to-Use, and High-Performance AI Deployment Framework

Introduction

nndeploy is an easy-to-use, and high-performance AI deployment framework. Based on the design concepts of visual workflows and multi-backend inference, developers can quickly develop SDKs for specified platforms and hardware from algorithm repositories, significantly saving development time. Furthermore, the framework has already deployed numerous AI models including LLM, AIGC generation, face swap, object detection, image segmentation, etc., ready to use out-of-the-box.

Simple and Easy to Use

  • Visual Workflow: Deploy AI algorithms through drag-and-drop operations. Visually adjust all node parameters of the AI algorithm in the frontend and quickly preview the effect after parameter tuning.
  • Custom Nodes: Support Python/C++ custom nodes, seamlessly integrated into the visual interface without frontend code.
  • Algorithm Combination: Flexibly combine different algorithms to quickly build innovative AI applications.
  • One-Click Deployment: The completed workflow can be exported as a JSON configuration file with one click, supporting direct calls via Python/C++ API, achieving seamless transition from development to production environments, and fully supporting platforms like Linux, Windows, macOS, Android, iOS, etc.

High Performance

  • Parallel Optimization: Supports execution modes like serial, pipeline parallel, task parallel, etc.

  • Memory Optimization: Optimization strategies like zero-copy, memory pool, memory reuse, etc.

  • High-Performance Optimization: Built-in nodes optimized with C++/CUDA/Ascend C/SIMD, etc.

  • Multi-Backend Inference: One workflow, multiple backend inference. Integrates 13 mainstream inference frameworks with zero abstraction cost, covering all platforms including cloud, desktop, mobile, edge, etc.

    Inference Framework Application Scenario Status
    ONNXRuntime Cross-platform inference
    TensorRT NVIDIA GPU high-performance inference
    OpenVINO Intel CPU/GPU optimization
    MNN Mobile inference engine by Alibaba
    TNN Mobile inference engine by Tencent
    ncnn Mobile inference engine by Tencent
    CoreML iOS/macOS native acceleration
    AscendCL Huawei Ascend AI chip inference framework
    RKNN Rockchip NPU inference framework
    SNPE Qualcomm Snapdragon NPU inference framework
    TVM Deep learning compiler stack
    PyTorch Rapid prototyping / Cloud deployment
    Self-developed Inference Framework Default inference framework

Out-of-the-Box Algorithms

List of deployed models, with 100+ nodes created. We will continue to deploy more high-value AI algorithms. If you have algorithms you need deployed, please let us know via issue.

Application Scenario Available Models Remarks
Large Language Model QWen-2.5, QWen-3
Image Generation Stable Diffusion 1.5, Stable Diffusion XL, Stable Diffusion 3, HunyuanDiT, etc. Supports text-to-image, image-to-image, inpainting; based on diffusers
Face Swap deep-live-cam
OCR Paddle OCR
Object Detection YOLOv5, YOLOv6, YOLOv7, YOLOv8, YOLOv11, YOLOx
Object Tracking FairMot
Image Segmentation RBMGv1.4, PPMatting, Segment Anything
Classification ResNet, MobileNet, EfficientNet, PPLcNet, GhostNet, ShuffleNet, SqueezeNet
API Service OPENAI, DeepSeek, Moonshot Supports LLM and AIGC services

See more details in the Deployed Model List Details

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distributions

If you're not sure about the file name format, learn more about wheel file names.

nndeploy-3.0.9-cp312-cp312-manylinux_2_28_x86_64.whl (98.6 MB view details)

Uploaded CPython 3.12manylinux: glibc 2.28+ x86-64

nndeploy-3.0.9-cp311-cp311-manylinux_2_28_x86_64.whl (97.1 MB view details)

Uploaded CPython 3.11manylinux: glibc 2.28+ x86-64

nndeploy-3.0.9-cp310-cp310-manylinux_2_28_x86_64.whl (95.6 MB view details)

Uploaded CPython 3.10manylinux: glibc 2.28+ x86-64

File details

Details for the file nndeploy-3.0.9-cp312-cp312-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for nndeploy-3.0.9-cp312-cp312-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 4f62ea7f0177a07ba9be07d43c207b19798ade20f09d6f179f47a5b599d2654f
MD5 42b035418a40162aa5f75b45c83047dd
BLAKE2b-256 3b2ca0d1847de437334411c9070fb90cda7b0d9ebb4c8c9bb0a38ed1a2ff2b34

See more details on using hashes here.

File details

Details for the file nndeploy-3.0.9-cp311-cp311-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for nndeploy-3.0.9-cp311-cp311-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 2aad87b4e2d9a7bd26e4af5b10bd2697273316308b601c3aaa83e79fd1253003
MD5 7d2a2786bfa76950de42ebe330a86489
BLAKE2b-256 5619b4879ccb2718e958eda5ba5116cc362754f41af805583838cbe97ea8eef5

See more details on using hashes here.

File details

Details for the file nndeploy-3.0.9-cp310-cp310-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for nndeploy-3.0.9-cp310-cp310-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 7b90230e2dd99f599bb91d5f8c846cc75ff81faff068105440c8340e99e1d7da
MD5 b6dcc35b90447534b14cfce04791ae2b
BLAKE2b-256 5bc97c2464b571ceec8bf8bed1c0d7cda3eb6876b3ca2269ebe12ebb1445da0b

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page