Nexa AI SDK
Project description
Nexa SDK
The Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), and text-to-speech (TTS) capabilities. Additionally, it offers an OpenAI-compatible API server with JSON schema mode for function calling and streaming support, and a user-friendly Streamlit UI.
Features
-
Model Support:
- ONNX & GGML models
- Conversion Engine
- Inference Engine:
- Text Generation
- Image Generation
- Vision-Language Models (VLM)
- Text-to-Speech (TTS)
Detailed API documentation is available here.
- Server:
- OpenAI-compatible API
- JSON schema mode for function calling
- Streaming support
- Streamlit UI for interactive model deployment and testing
Installation
GPU version(optional)
check if you have GPU acceleration (torch required)
CUDA:
import torch
torch.cuda.is_available()
if True
CMAKE_ARGS="-DGGML_CUDA=on -DSD_CUBLAS=ON" pip install nexaai-gpu
Apple M Chip:
Apple icon -> about this mac -> Graphicsif True:
CMAKE_ARGS="-DGGML_METAL=on -DSD_METAL=ON" pip install nexaai-gpu
AMD graphics card:
CMAKE_ARGS="-DGGML_HIPBLAS=on" pip install nexaai-gpu
CPU version
Mac with Intel chips
CMAKE_ARGS="-DCMAKE_CXX_FLAGS=-fopenmp" pip install nexaai
Mac with M chips or other Operating systems:
pip install nexaai
Nexa CLI commands
Model Commands
NLP Models
Model | Type | Format | Command |
---|---|---|---|
octopus-v2 | NLP | GGUF | nexa gen-text octopus-v2 |
octopus-v4 | NLP | GGUF | nexa gen-text octopus-v4 |
tinyllama | NLP | GGUF | nexa gen-text tinyllama |
llama2 | NLP | GGUF/ONNX | nexa gen-text llama2 |
llama3 | NLP | GGUF/ONNX | nexa gen-text llama3 |
llama3.1 | NLP | GGUF/ONNX | nexa gen-text llama3.1 |
gemma | NLP | GGUF/ONNX | nexa gen-text gemma |
gemma2 | NLP | GGUF | nexa gen-text gemma2 |
qwen1.5 | NLP | GGUF | nexa gen-text qwen1.5 |
qwen2 | NLP | GGUF/ONNX | nexa gen-text qwen2 |
mistral | NLP | GGUF/ONNX | nexa gen-text mistral |
codegemma | NLP | GGUF | nexa gen-text codegemma |
codellama | NLP | GGUF | nexa gen-text codellama |
codeqwen | NLP | GGUF | nexa gen-text codeqwen |
deepseek-coder | NLP | GGUF | nexa gen-text deepseek-coder |
dolphin-mistral | NLP | GGUF | nexa gen-text dolphin-mistral |
nomic-embed-text | NLP | GGUF | nexa gen-text nomic-embed-text |
phi2 | NLP | GGUF | nexa gen-text phi2 |
phi3 | NLP | GGUF/ONNX | nexa gen-text phi3 |
Multimodal Models
Model | Type | Format | Command |
---|---|---|---|
nanollava | Multimodal | GGUF | nexa vlm nanollava |
llava-phi3 | Multimodal | GGUF | nexa vlm llava-phi3 |
llava-llama3 | Multimodal | GGUF | nexa vlm llava-llama3 |
llava1.6-mistral | Multimodal | GGUF | nexa vlm llava1.6-mistral |
llava1.6-vicuna | Multimodal | GGUF | nexa vlm llava1.6-vicuna |
Computer Vision Models
Model | Type | Format | Command |
---|---|---|---|
stable-diffusion-v1-4 | Computer Vision | GGUF | nexa gen-image sd1-4 |
stable-diffusion-v1-5 | Computer Vision | GGUF/ONNX | nexa gen-image sd1-5 |
lcm-dreamshaper | Computer Vision | GGUF/ONNX | nexa gen-image lcm-dreamshaper |
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
nexaai_metal-0.0.3.tar.gz
(38.1 MB
view details)
File details
Details for the file nexaai_metal-0.0.3.tar.gz
.
File metadata
- Download URL: nexaai_metal-0.0.3.tar.gz
- Upload date:
- Size: 38.1 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.10.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | d38347496ec5e3904a9b3e7ef054d913afba442ecf374fb03451aa0de53a7b26 |
|
MD5 | edd5ac47486991a6258a485ef9ab5092 |
|
BLAKE2b-256 | 2433c0b4cd421bc77e56bc0b3e17f22df61a23cb53f6d313fdebce6f5cd700ab |