Skip to main content

A production-ready, zero-dependency ONNX runtime for OpenAI CLIP and OpenCLIP models.

Project description

Modern ONNX CLIP

A modern, lightweight, and robust ONNX runtime for CLIP models.

This library allows you to run OpenAI CLIP and OpenCLIP models in production environments without installing PyTorch. It provides a simple CLI to convert models from the massive OpenCLIP model zoo and a pure-Python inference engine powered by onnxruntime, numpy, and pillow.

🚀 Features

  • Zero PyTorch Dependency in Production: Run inference with just numpy and onnxruntime. Drastically reduces Docker image size and memory usage.
  • Easy Conversion: Convert any model from OpenCLIP (ViT-B-32, ViT-L-14, SigLIP, etc.) with a single command.
  • Modern Tooling: Built with uv, ruff, and strictly typed with pyright.
  • Fast: Leverages ONNX Runtime (CPU or CUDA) for high-performance inference.
  • Drop-in Replacement: Designed to replace the unmaintained onnx_clip package with better model support.

📦 Installation

For Production (Inference Only)

If you only need to run models, install the base package. This does not install PyTorch.

pip install modern-onnx-clip
# or with uv
uv add modern-onnx-clip

For Development & Exporting

To convert models, you need the export dependencies (PyTorch, OpenCLIP).

pip install "modern-onnx-clip[export]"

🛠️ Usage

1. Convert a Model

First, convert a model from the OpenCLIP registry. You need the [export] extras installed for this step.

# Syntax: onnx-clip convert --model <ARCH> --pretrained <TAG> --output <DIR>

# Example: Standard ViT-B-32
onnx-clip convert --model ViT-B-32 --pretrained laion2b_s34b_b79k --output ./models/vit-b-32

# Example: ViT-L-14 (Higher accuracy)
onnx-clip convert --model ViT-L-14 --pretrained openai --output ./models/vit-l-14

This will create a folder containing visual.onnx, textual.onnx, and configuration files.

2. Run Inference (Python)

Now you can use the model in your application. This step works without PyTorch.

from onnx_clip import OnnxClip
from PIL import Image

# 1. Load the model (Provide the directory where you exported the model)
model = OnnxClip(model_dir="./models/vit-b-32", device="cpu") # use 'cuda' for GPU

# 2. Get Image Embeddings
image = Image.open("cat.jpg")
image_features = model.get_image_embedding(image)
# shape: (1, 512)

# 3. Get Text Embeddings
text_features = model.get_text_embedding(["a photo of a cat", "a photo of a dog"])
# shape: (2, 512)

# 4. Calculate Similarity
# (The embeddings are already normalized)
similarity = image_features @ text_features.T
print(similarity)
# [[0.28, 0.15]]

3. CLI Inference (Testing)

You can also test a model directly from the CLI:

onnx-clip run --model-dir ./models/vit-b-32 --image cat.jpg --text "a cute cat"

⚙️ GPU Support

To run on NVIDIA GPUs, install onnxruntime-gpu instead of the standard onnxruntime.

pip uninstall onnxruntime
pip install onnxruntime-gpu

Then initialize the model with device="cuda".

🏗️ Project Structure

  • exporter.py: Handles loading PyTorch models and exporting them to ONNX graphs.
  • model.py: The lightweight inference engine. Abstraction over ONNX Runtime sessions.
  • preprocessor.py: Reimplementation of CLIP's image preprocessing using only NumPy and Pillow.
  • tokenizer.py: Handles text tokenization (BPE) without heavy external dependencies.

License

MIT License.

Acknowledgements

Built on top of the incredible work by OpenAI and OpenCLIP. Inspired by the original onnx_clip package.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

modern_onnx_clip-0.1.0.tar.gz (123.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

modern_onnx_clip-0.1.0-py3-none-any.whl (13.2 kB view details)

Uploaded Python 3

File details

Details for the file modern_onnx_clip-0.1.0.tar.gz.

File metadata

  • Download URL: modern_onnx_clip-0.1.0.tar.gz
  • Upload date:
  • Size: 123.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for modern_onnx_clip-0.1.0.tar.gz
Algorithm Hash digest
SHA256 58ce20688841532d89cb97fb1244a8afea5fca65c16a8451923e57d56b1adbde
MD5 d4af99bb0ba5f2b8012676a93cbea837
BLAKE2b-256 3bd61477df5e6121472b197d8e81dfda8670cfdbbe13a3ae8035a5c45abadda4

See more details on using hashes here.

Provenance

The following attestation bundles were made for modern_onnx_clip-0.1.0.tar.gz:

Publisher: publish.yml on Neizvestnyj/modern-onnx-clip

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file modern_onnx_clip-0.1.0-py3-none-any.whl.

File metadata

File hashes

Hashes for modern_onnx_clip-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 21853d12f478263f8a7c38cb73ac495558d7315b55fe6581c99f7226c2b1fcbc
MD5 b870e00a4aa52e799167c6f94ec7dc46
BLAKE2b-256 c871a6ce1cb165754c6a29b13d16bdcb028d40656c681c49ad680ac1428d7850

See more details on using hashes here.

Provenance

The following attestation bundles were made for modern_onnx_clip-0.1.0-py3-none-any.whl:

Publisher: publish.yml on Neizvestnyj/modern-onnx-clip

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page