Skip to main content

A production-ready, zero-dependency ONNX runtime for OpenAI CLIP and OpenCLIP models.

Project description

Modern ONNX CLIP

A modern, lightweight, and robust ONNX runtime for CLIP models.

This library allows you to run OpenAI CLIP and OpenCLIP models in production environments without installing PyTorch. It provides a simple CLI to convert models from the massive OpenCLIP model zoo and a pure-Python inference engine powered by onnxruntime, numpy, and pillow.

🚀 Features

  • Zero PyTorch Dependency in Production: Run inference with just numpy and onnxruntime. Drastically reduces Docker image size and memory usage.
  • Easy Conversion: Convert any model from OpenCLIP (ViT-B-32, ViT-L-14, SigLIP, etc.) with a single command.
  • Modern Tooling: Built with uv, ruff, and strictly typed with pyright.
  • Fast: Leverages ONNX Runtime (CPU or CUDA) for high-performance inference.
  • Drop-in Replacement: Designed to replace the unmaintained onnx_clip package with better model support.

📦 Installation

For Production (Inference Only)

If you only need to run models, install the base package. This does not install PyTorch.

pip install modern-onnx-clip
# or with uv
uv add modern-onnx-clip

For Development & Exporting

To convert models, you need the export dependencies (PyTorch, OpenCLIP).

pip install "modern-onnx-clip[export]"

🛠️ Usage

1. Convert a Model

First, convert a model from the OpenCLIP registry. You need the [export] extras installed for this step.

# Syntax: onnx-clip convert --model <ARCH> --pretrained <TAG> --output <DIR>

# Example: Standard ViT-B-32
onnx-clip convert --model ViT-B-32 --pretrained laion2b_s34b_b79k --output ./models/vit-b-32

# Example: ViT-L-14 (Higher accuracy)
onnx-clip convert --model ViT-L-14 --pretrained openai --output ./models/vit-l-14

This will create a folder containing visual.onnx, textual.onnx, and configuration files.

2. Run Inference (Python)

Now you can use the model in your application. This step works without PyTorch.

from onnx_clip import OnnxClip
from PIL import Image

# 1. Load the model (Provide the directory where you exported the model)
model = OnnxClip(model_dir="./models/vit-b-32", device="cpu") # use 'cuda' for GPU

# 2. Get Image Embeddings
image = Image.open("cat.jpg")
image_features = model.get_image_embedding(image)
# shape: (1, 512)

# 3. Get Text Embeddings
text_features = model.get_text_embedding(["a photo of a cat", "a photo of a dog"])
# shape: (2, 512)

# 4. Calculate Similarity
# (The embeddings are already normalized)
similarity = image_features @ text_features.T
print(similarity)
# [[0.28, 0.15]]

3. CLI Inference (Testing)

You can also test a model directly from the CLI:

onnx-clip run --model-dir ./models/vit-b-32 --image cat.jpg --text "a cute cat"

⚙️ GPU Support

To run on NVIDIA GPUs, install onnxruntime-gpu instead of the standard onnxruntime.

pip uninstall onnxruntime
pip install onnxruntime-gpu

Then initialize the model with device="cuda".

🏗️ Project Structure

  • exporter.py: Handles loading PyTorch models and exporting them to ONNX graphs.
  • model.py: The lightweight inference engine. Abstraction over ONNX Runtime sessions.
  • preprocessor.py: Reimplementation of CLIP's image preprocessing using only NumPy and Pillow.
  • tokenizer.py: Handles text tokenization (BPE) without heavy external dependencies.

License

MIT License.

Acknowledgements

Built on top of the incredible work by OpenAI and OpenCLIP. Inspired by the original onnx_clip package.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

modern_onnx_clip-0.1.2.tar.gz (125.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

modern_onnx_clip-0.1.2-py3-none-any.whl (14.2 kB view details)

Uploaded Python 3

File details

Details for the file modern_onnx_clip-0.1.2.tar.gz.

File metadata

  • Download URL: modern_onnx_clip-0.1.2.tar.gz
  • Upload date:
  • Size: 125.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for modern_onnx_clip-0.1.2.tar.gz
Algorithm Hash digest
SHA256 f69e4f3278c95e9bdf4fe4c7f657a8b2ca9ed3d150958c7e769419ed5df0af9a
MD5 e30527032e63fc79ac93ce4a1752d396
BLAKE2b-256 442c2b8ca7f6e2ffb81b662db21c9c0b8b3095c06539d5295b488946aad2e7f1

See more details on using hashes here.

Provenance

The following attestation bundles were made for modern_onnx_clip-0.1.2.tar.gz:

Publisher: publish.yml on Neizvestnyj/modern-onnx-clip

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file modern_onnx_clip-0.1.2-py3-none-any.whl.

File metadata

File hashes

Hashes for modern_onnx_clip-0.1.2-py3-none-any.whl
Algorithm Hash digest
SHA256 f8f68cff6a1637c000c53ff43c614576096e3eb528204b068cb52578b4ec4cf6
MD5 9792b1ef36d0ebdf1d940557cc2214e7
BLAKE2b-256 6ea1b76cef8522dd1a88496383a88abc93dd06f8e7efe43565cc38bd1781ca48

See more details on using hashes here.

Provenance

The following attestation bundles were made for modern_onnx_clip-0.1.2-py3-none-any.whl:

Publisher: publish.yml on Neizvestnyj/modern-onnx-clip

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page