Skip to main content

Popular Machine Learning models optimized for Qualcomm chipsets.

Project description

Qualcomm® AI Hub Models

Release Tag PyPi Python 3.10, 3.11, 3.12, 3.13

The Qualcomm® AI Hub Models are a collection of state-of-the-art machine learning models optimized for deployment on Qualcomm® devices.

See supported: On-Device Runtimes, Hardware Targets & Precision, Chipsets, Devices

 

Setup

1. Install Python Package

The package is available via pip:

# NOTE for Snapdragon X Elite users:
# Only AMDx64 (64-bit) Python is supported on Windows.
# Installation will fail when using Windows ARM64 Python.

pip install qai_hub_models

Some models (e.g. YOLOv7) require additional dependencies. View the model README (at qai_hub_models/models/model_id) for installation instructions.

 

2. Configure AI Hub Workbench Access

Many features of AI Hub Models (such as model compilation, on-device profiling, etc.) require access to Qualcomm® AI Hub Workbench:

 

Getting Started

Export and Run A Model on a Physical Device

All models in our directory can be compiled and profiled on a hosted Qualcomm® device:

pip install "qai_hub_models[yolov7]"

python -m qai_hub_models.models.yolov7.export [--target-runtime ...] [--device ...] [--help]

Using Qualcomm® AI Hub Workbench, the export script will:

  1. Compile the model for the chosen device and target runtime (see: Compiling Models on AI Hub Workbench).
  2. If applicable, Quantize the model (see: Quantization on AI Hub Workbench)
  3. Profile the compiled model on a real device in the cloud (see: Profiling Models on AI Hub Workbench).
  4. Run inference with a sample input data on a real device in the cloud, and compare on-device model output with PyTorch output (see: Running Inference on AI Hub Workbench)
  5. Download the compiled model to disk.

 

End-To-End Model Demos

Most models in our directory contain CLI demos that run the model end-to-end:

pip install "qai_hub_models[yolov7]"
# Predict and draw bounding boxes on the provided image
python -m qai_hub_models.models.yolov7.demo [--image ...] [--eval-mode {fp,on-device}] [--help]

End-to-end demos:

  1. Preprocess human-readable input into model input
  2. Run model inference
  3. Postprocess model output to a human-readable format

Many end-to-end demos use AI Hub Workbench to run inference on a real cloud-hosted device (with --eval-mode on-device). All end-to-end demos can also run locally via PyTorch (with --eval-mode fp).

 

Sample Applications

Native applications that can run our models (with pre- and post-processing) on physical devices are published in the AI Hub Apps repository.

Python applications are defined for all models (from qai_hub_models.models.<model_name> import App). These apps wrap model inference with pre- and post-processing steps written using torch & numpy. These apps are optimized to be an easy-to-follow example, rather than to minimize prediction time.

 

Model Support Data

On-Device Runtimes

Runtime Supported OS
Qualcomm AI Engine Direct Android, Linux, Windows
LiteRT (TensorFlow Lite) Android, Linux
ONNX Android, Linux, Windows

Device Hardware & Precision

Device Compute Unit Supported Precision
CPU FP32, INT16, INT8
GPU FP32, FP16
NPU (includes Hexagon DSP, HTP) FP16*, INT16, INT8

*Some older chipsets do not support fp16 inference on their NPU.

Chipsets

and many more.

Devices

  • Samsung Galaxy S21, S22, S23, and S24 Series
  • Xiaomi 12 and 13
  • Snapdragon X Elite CRD (Compute Reference Device)
  • Qualcomm RB3 Gen 2, RB5

and many more.

 

Model Directory

Computer Vision

Model README
Image Classification
Beit qai_hub_models.models.beit
ConvNext-Base qai_hub_models.models.convnext_base
ConvNext-Tiny qai_hub_models.models.convnext_tiny
DLA-102-X qai_hub_models.models.dla102x
DenseNet-121 qai_hub_models.models.densenet121
EfficientFormer qai_hub_models.models.efficientformer
EfficientNet-B0 qai_hub_models.models.efficientnet_b0
EfficientNet-B4 qai_hub_models.models.efficientnet_b4
EfficientNet-V2-s qai_hub_models.models.efficientnet_v2_s
EfficientViT-b2-cls qai_hub_models.models.efficientvit_b2_cls
EfficientViT-l2-cls qai_hub_models.models.efficientvit_l2_cls
GPUNet qai_hub_models.models.gpunet
GoogLeNet qai_hub_models.models.googlenet
Inception-v3 qai_hub_models.models.inception_v3
InternImage qai_hub_models.models.internimage
LeViT qai_hub_models.models.levit
MNASNet05 qai_hub_models.models.mnasnet05
Mobile-VIT qai_hub_models.models.mobile_vit
MobileNet-v2 qai_hub_models.models.mobilenet_v2
MobileNet-v3-Large qai_hub_models.models.mobilenet_v3_large
MobileNet-v3-Small qai_hub_models.models.mobilenet_v3_small
NASNet qai_hub_models.models.nasnet
RegNet qai_hub_models.models.regnet
RegNet-Y-800MF qai_hub_models.models.regnet_y_800mf
ResNeXt101 qai_hub_models.models.resnext101
ResNeXt50 qai_hub_models.models.resnext50
ResNet101 qai_hub_models.models.resnet101
ResNet18 qai_hub_models.models.resnet18
ResNet50 qai_hub_models.models.resnet50
Sequencer2D qai_hub_models.models.sequencer2d
Shufflenet-v2 qai_hub_models.models.shufflenet_v2
SqueezeNet-1.1 qai_hub_models.models.squeezenet1_1
VIT qai_hub_models.models.vit
WideResNet50 qai_hub_models.models.wideresnet50
Image Editing
AOT-GAN qai_hub_models.models.aotgan
DDColor qai_hub_models.models.ddcolor
LaMa-Dilated qai_hub_models.models.lama_dilated
Super Resolution
Real-ESRGAN-x4plus qai_hub_models.models.real_esrgan_x4plus
XLSR qai_hub_models.models.xlsr
Semantic Segmentation
BGNet qai_hub_models.models.bgnet
BiseNet qai_hub_models.models.bisenet
DDRNet23-Slim qai_hub_models.models.ddrnet23_slim
DeepLabV3-Plus-MobileNet qai_hub_models.models.deeplabv3_plus_mobilenet
DeepLabXception qai_hub_models.models.deeplab_xception
EdgeTAM qai_hub_models.models.edgetam
FCN-ResNet50 qai_hub_models.models.fcn_resnet50
FFNet-122NS-LowRes qai_hub_models.models.ffnet_122ns_lowres
FFNet-40S qai_hub_models.models.ffnet_40s
FFNet-54S qai_hub_models.models.ffnet_54s
FFNet-78S qai_hub_models.models.ffnet_78s
FFNet-78S-LowRes qai_hub_models.models.ffnet_78s_lowres
FastSam-S qai_hub_models.models.fastsam_s
FastSam-X qai_hub_models.models.fastsam_x
HRNet-W48-OCR qai_hub_models.models.hrnet_w48_ocr
MaskRCNN qai_hub_models.models.maskrcnn
MediaPipe-Selfie-Segmentation qai_hub_models.models.mediapipe_selfie
MobileSam qai_hub_models.models.mobilesam
PSPNet qai_hub_models.models.pspnet
PidNet qai_hub_models.models.pidnet
SINet qai_hub_models.models.sinet
SalsaNext qai_hub_models.models.salsanext
Segformer-Base qai_hub_models.models.segformer_base
Segment-Anything-Model-2 qai_hub_models.models.sam2
Unet-Segmentation qai_hub_models.models.unet_segmentation
YOLOv11-Segmentation qai_hub_models.models.yolov11_seg
YOLOv8-Segmentation qai_hub_models.models.yolov8_seg
Video Classification
ResNet-2Plus1D qai_hub_models.models.resnet_2plus1d
ResNet-3D qai_hub_models.models.resnet_3d
ResNet-Mixed-Convolution qai_hub_models.models.resnet_mixed
Video-MAE qai_hub_models.models.video_mae
Video Generation
First-Order-Motion-Model qai_hub_models.models.fomm
Video Object Tracking
Track-Anything qai_hub_models.models.track_anything
Object Detection
3D-Deep-BOX qai_hub_models.models.deepbox
CavaFace qai_hub_models.models.cavaface
CenterNet-2D qai_hub_models.models.centernet_2d
Conditional-DETR-ResNet50 qai_hub_models.models.conditional_detr_resnet50
DETR-ResNet101 qai_hub_models.models.detr_resnet101
DETR-ResNet101-DC5 qai_hub_models.models.detr_resnet101_dc5
DETR-ResNet50 qai_hub_models.models.detr_resnet50
DETR-ResNet50-DC5 qai_hub_models.models.detr_resnet50_dc5
DeformableDETR qai_hub_models.models.deformable_detr
Detectron2-Detection qai_hub_models.models.detectron2_detection
Facial-Attribute-Detection qai_hub_models.models.face_attrib_net
HRNetFace qai_hub_models.models.hrnet_face
Lightweight-Face-Detection qai_hub_models.models.face_det_lite
MediaPipe-Face-Detection qai_hub_models.models.mediapipe_face
MediaPipe-Hand-Detection qai_hub_models.models.mediapipe_hand
MediaPipe-Hand-Gesture-Recognition qai_hub_models.models.mediapipe_hand_gesture
PPE-Detection qai_hub_models.models.gear_guard_net
Person-Foot-Detection qai_hub_models.models.foot_track_net
RTMDet qai_hub_models.models.rtmdet
YOLOv10-Detection qai_hub_models.models.yolov10_det
YOLOv11-Detection qai_hub_models.models.yolov11_det
YOLOv8-Detection qai_hub_models.models.yolov8_det
Yolo-R qai_hub_models.models.yolor
Yolo-X qai_hub_models.models.yolox
Yolo-v3 qai_hub_models.models.yolov3
Yolo-v5 qai_hub_models.models.yolov5
Yolo-v6 qai_hub_models.models.yolov6
Yolo-v7 qai_hub_models.models.yolov7
Pose Estimation
CenterNet-Pose qai_hub_models.models.centernet_pose
Facial-Landmark-Detection qai_hub_models.models.facemap_3dmm
HRNetPose qai_hub_models.models.hrnet_pose
LiteHRNet qai_hub_models.models.litehrnet
MediaPipe-Pose-Estimation qai_hub_models.models.mediapipe_pose
Posenet-Mobilenet qai_hub_models.models.posenet_mobilenet
RTMPose-Body2d qai_hub_models.models.rtmpose_body2d
Gaze Estimation
EyeGaze qai_hub_models.models.eyegaze
Depth Estimation
Depth-Anything qai_hub_models.models.depth_anything
Depth-Anything-V2 qai_hub_models.models.depth_anything_v2
Depth-Anything-V3 qai_hub_models.models.depth_anything_v3
Midas-V2 qai_hub_models.models.midas
Driver Assistance
BEVDet qai_hub_models.models.bevdet
BEVFusion qai_hub_models.models.bevfusion_det
CVT qai_hub_models.models.cvt
CenterNet-3D qai_hub_models.models.centernet_3d
CenterPoint qai_hub_models.models.centerpoint
GKT qai_hub_models.models.gkt
StateTransformer qai_hub_models.models.statetransformer
Robotics
ACT qai_hub_models.models.act

Multimodal

Model README
EasyOCR qai_hub_models.models.easyocr
Nomic-Embed-Text qai_hub_models.models.nomic_embed_text
OpenAI-Clip qai_hub_models.models.openai_clip
OpusMT-En-Es qai_hub_models.models.opus_mt_en_es
OpusMT-En-Zh qai_hub_models.models.opus_mt_en_zh
OpusMT-Es-En qai_hub_models.models.opus_mt_es_en
OpusMT-Zh-En qai_hub_models.models.opus_mt_zh_en
TrOCR qai_hub_models.models.trocr

Audio

Model README
Speech Recognition
HuggingFace-WavLM-Base-Plus qai_hub_models.models.huggingface_wavlm_base_plus
Whisper-Base qai_hub_models.models.whisper_base
Whisper-Large-V3-Turbo qai_hub_models.models.whisper_large_v3_turbo
Whisper-Small qai_hub_models.models.whisper_small
Whisper-Small-Quantized qai_hub_models.models.whisper_small_quantized
Whisper-Tiny qai_hub_models.models.whisper_tiny
Zipformer qai_hub_models.models.zipformer
Audio Classification
YamNet qai_hub_models.models.yamnet
Audio Generation
MeloTTS-EN qai_hub_models.models.melotts_en

Generative AI

Model README
Image Generation
ControlNet-Canny qai_hub_models.models.controlnet_canny
Stable-Diffusion-v1.5 qai_hub_models.models.stable_diffusion_v1_5
Stable-Diffusion-v2.1 qai_hub_models.models.stable_diffusion_v2_1
Text Generation
Albert-Base-V2-Hf qai_hub_models.models.albert_base_v2_hf
Baichuan2-7B qai_hub_models.models.baichuan2_7b
Bert-Base-Uncased-Hf qai_hub_models.models.bert_base_uncased_hf
Distil-Bert-Base-Uncased-Hf qai_hub_models.models.distil_bert_base_uncased_hf
Electra-Bert-Base-Discrim-Google qai_hub_models.models.electra_bert_base_discrim_google
Falcon3-7B-Instruct qai_hub_models.models.falcon_v3_7b_instruct
IBM-Granite-v3.1-8B-Instruct qai_hub_models.models.ibm_granite_v3_1_8b_instruct
IndusQ-1.1B qai_hub_models.models.indus_1b
JAIS-6p7b-Chat qai_hub_models.models.jais_6p7b_chat
Llama-SEA-LION-v3.5-8B-R qai_hub_models.models.llama_v3_1_sea_lion_3_5_8b_r
Llama-v2-7B-Chat qai_hub_models.models.llama_v2_7b_chat
Llama-v3-8B-Instruct qai_hub_models.models.llama_v3_8b_instruct
Llama-v3-ELYZA-JP-8B qai_hub_models.models.llama_v3_elyza_jp_8b
Llama-v3.1-8B-Instruct qai_hub_models.models.llama_v3_1_8b_instruct
Llama-v3.2-1B-Instruct qai_hub_models.models.llama_v3_2_1b_instruct
Llama-v3.2-3B-Instruct qai_hub_models.models.llama_v3_2_3b_instruct
Llama3-TAIDE-LX-8B-Chat-Alpha1 qai_hub_models.models.llama_v3_taide_8b_chat
Mistral-7B-Instruct-v0.3 qai_hub_models.models.mistral_7b_instruct_v0_3
Mobile-Bert-Uncased-Google qai_hub_models.models.mobile_bert_uncased_google
PLaMo-1B qai_hub_models.models.plamo_1b
Phi-3.5-Mini-Instruct qai_hub_models.models.phi_3_5_mini_instruct
Qwen2-7B-Instruct qai_hub_models.models.qwen2_7b_instruct
Qwen2.5-7B-Instruct qai_hub_models.models.qwen2_5_7b_instruct

Need help?

Slack: https://aihub.qualcomm.com/community/slack

GitHub Issues: https://github.com/quic/ai-hub-models/issues

Email: ai-hub-support@qti.qualcomm.com.

LICENSE

Qualcomm® AI Hub Models is licensed under BSD-3. See the LICENSE file.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

qai_hub_models-0.47.0-py3-none-any.whl (3.6 MB view details)

Uploaded Python 3

File details

Details for the file qai_hub_models-0.47.0-py3-none-any.whl.

File metadata

File hashes

Hashes for qai_hub_models-0.47.0-py3-none-any.whl
Algorithm Hash digest
SHA256 ee15a13c5b6cd154da2182d33b6b96bc663a187726e88749ea1d71e63008d0d5
MD5 aa2973520cd59127cdc91a93cfb49dae
BLAKE2b-256 c5e841dab1820de45918a7d8c89e11c611a490daf98bf210544779625d2304c1

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page