Skip to main content

Nexa AI SDK

Project description

Nexa SDK

The Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), and text-to-speech (TTS) capabilities. Additionally, it offers an OpenAI-compatible API server with JSON schema mode for function calling and streaming support, and a user-friendly Streamlit UI.

Features

  • Model Support:
    • ONNX & GGML models
    • Conversion Engine
    • Inference Engine:
      • Text Generation
      • Image Generation
      • Vision-Language Models (VLM)
      • Text-to-Speech (TTS)
  • Server:
    • OpenAI-compatible API
    • JSON schema mode for function calling
    • Streaming support
  • Streamlit UI for interactive model deployment and testing

Installation

For CPU version

pip install nexaai
pip install nexaai[onnx] # if you want to use ONNX models

For GPU version

pip install nexaai-gpu
pip install nexaai-gpu[onnx] # if you want to use ONNX models

Nexa CLI commands

Model Commands

NLP Models

Model Type Format Command
octopus-v2 NLP GGUF nexa-cli gen-text octopus-v2
octopus-v4 NLP GGUF nexa-cli gen-text octopus-v4
tinyllama NLP GGUF nexa-cli gen-text tinyllama
llama2 NLP GGUF/ONNX nexa-cli gen-text llama2
llama3 NLP GGUF/ONNX nexa-cli gen-text llama3
llama3.1 NLP GGUF/ONNX nexa-cli gen-text llama3.1
gemma NLP GGUF/ONNX nexa-cli gen-text gemma
gemma2 NLP GGUF nexa-cli gen-text gemma2
qwen1.5 NLP GGUF nexa-cli gen-text qwen1.5
qwen2 NLP GGUF/ONNX nexa-cli gen-text qwen2
mistral NLP GGUF/ONNX nexa-cli gen-text mistral
codegemma NLP GGUF nexa-cli gen-text codegemma
codellama NLP GGUF nexa-cli gen-text codellama
codeqwen NLP GGUF nexa-cli gen-text codeqwen
deepseek-coder NLP GGUF nexa-cli gen-text deepseek-coder
dolphin-mistral NLP GGUF nexa-cli gen-text dolphin-mistral
nomic-embed-text NLP GGUF nexa-cli gen-text nomic-embed-text
phi2 NLP GGUF nexa-cli gen-text phi2
phi3 NLP GGUF/ONNX nexa-cli gen-text phi3

Multimodal Models

Model Type Format Command
nanollava Multimodal GGUF nexa-cli vlm nanollava
llava-phi3 Multimodal GGUF nexa-cli vlm llava-phi3
llava-llama3 Multimodal GGUF nexa-cli vlm llava-llama3
llava1.6-mistral Multimodal GGUF nexa-cli vlm llava1.6-mistral
llava1.6-vicuna Multimodal GGUF nexa-cli vlm llava1.6-vicuna

Computer Vision Models

Model Type Format Command
stable-diffusion-v1-4 Computer Vision GGUF nexa-cli gen-image sd1-4
stable-diffusion-v1-5 Computer Vision GGUF/ONNX nexa-cli gen-image sd1-5
lcm-dreamshaper Computer Vision GGUF/ONNX nexa-cli gen-image lcm-dreamshaper

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

nexaai_gpu-0.0.1.dev0-py3-none-any.whl (159.5 kB view details)

Uploaded Python 3

File details

Details for the file nexaai_gpu-0.0.1.dev0-py3-none-any.whl.

File metadata

File hashes

Hashes for nexaai_gpu-0.0.1.dev0-py3-none-any.whl
Algorithm Hash digest
SHA256 ad9dc8c00486145acbe438c7c14b5c20a15905b506191602796992f55cffa43e
MD5 4eb51710229e9df4c6a10a2f0085681c
BLAKE2b-256 d911017ee444c36f4a8c056c022575bd9e4c5760eee6a205ea74fb7e7fc44b5c

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page