Skip to main content

A production-ready, zero-dependency ONNX runtime for OpenAI CLIP and OpenCLIP models.

Project description

Modern ONNX CLIP

A modern, lightweight, and robust ONNX runtime for CLIP models.

This library allows you to run OpenAI CLIP and OpenCLIP models in production environments without installing PyTorch. It provides a simple CLI to convert models from the massive OpenCLIP model zoo and a pure-Python inference engine powered by onnxruntime, numpy, and pillow.

🚀 Features

  • Zero PyTorch Dependency in Production: Run inference with just numpy and onnxruntime. Drastically reduces Docker image size and memory usage.
  • Easy Conversion: Convert any model from OpenCLIP (ViT-B-32, ViT-L-14, SigLIP, etc.) with a single command.
  • Modern Tooling: Built with uv, ruff, and strictly typed with pyright.
  • Fast: Leverages ONNX Runtime (CPU or CUDA) for high-performance inference.
  • Drop-in Replacement: Designed to replace the unmaintained onnx_clip package with better model support.

📦 Installation

For Production (Inference Only)

If you only need to run models, install the base package. This does not install PyTorch.

pip install modern-onnx-clip
# or with uv
uv add modern-onnx-clip

For Development & Exporting

To convert models, you need the export dependencies (PyTorch, OpenCLIP).

pip install "modern-onnx-clip[export]"

🛠️ Usage

1. Convert a Model

First, convert a model from the OpenCLIP registry. You need the [export] extras installed for this step.

# Syntax: onnx-clip convert --model <ARCH> --pretrained <TAG> --output <DIR>

# Example: Standard ViT-B-32
onnx-clip convert --model ViT-B-32 --pretrained laion2b_s34b_b79k --output ./models/vit-b-32

# Example: ViT-L-14 (Higher accuracy)
onnx-clip convert --model ViT-L-14 --pretrained openai --output ./models/vit-l-14

This will create a folder containing visual.onnx, textual.onnx, and configuration files.

2. Run Inference (Python)

Now you can use the model in your application. This step works without PyTorch.

from onnx_clip import OnnxClip
from PIL import Image

# 1. Load the model (Provide the directory where you exported the model)
model = OnnxClip(model_dir="./models/vit-b-32", device="cpu")  # use 'cuda' for GPU

# 2. Get Image Embeddings
image = Image.open("cat.jpg")
image_features = model.get_image_embedding(image)
# shape: (1, 512)

# 3. Get Text Embeddings
text_features = model.get_text_embedding(["a photo of a cat", "a photo of a dog"])
# shape: (2, 512)

# 4. Calculate Similarity
# (The embeddings are already normalized)
similarity = image_features @ text_features.T
print(similarity)
# [[0.28, 0.15]]

3. CLI Inference (Testing)

You can also test a model directly from the CLI:

onnx-clip run --model-dir ./models/vit-b-32 --image cat.jpg --text "a cute cat"

⚙️ GPU Support

To run on NVIDIA GPUs, install onnxruntime-gpu instead of the standard onnxruntime.

pip uninstall onnxruntime
pip install onnxruntime-gpu

Then initialize the model with device="cuda".

🏗️ Project Structure

  • exporter.py: Handles loading PyTorch models and exporting them to ONNX graphs.
  • model.py: The lightweight inference engine. Abstraction over ONNX Runtime sessions.
  • preprocessor.py: Reimplementation of CLIP's image preprocessing using only NumPy and Pillow.
  • tokenizer.py: Handles text tokenization (BPE) without heavy external dependencies.

🧪 Development & Testing

We use pytest for testing.

Standard Tests

Run the standard test suite (does not require PyTorch/CLIP):

pytest

Manual Verification Tests

The tests/manual/ directory contains scripts to verify numerical consistency between this library (ONNX) and the original PyTorch CLIP. These tests are skipped by default if dependencies are missing. To run them:

  1. Install the clip library manually (it cannot be a package dependency due to PyPI restrictions):

    pip install git+https://github.com/openai/CLIP.git
    
  2. Export a model to a local directory (e.g., ../models/ViT-B-32):

    onnx-clip convert --model ViT-B-32 --pretrained laion2b_s34b_b79k --output ../models/ViT-B-32
    
  3. Set the environment variable and run:

    # Linux/Mac
    export ONNX_CLIP_MODEL_DIR="../models/ViT-B-32"
    pytest tests/manual/
    
    # Windows (PowerShell)
    $env:ONNX_CLIP_MODEL_DIR="../models/ViT-B-32"
    pytest tests/manual/
    

License

MIT License.

Acknowledgements

Built on top of the incredible work by OpenAI and OpenCLIP. Inspired by the original onnx_clip package.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

modern_onnx_clip-0.1.3.tar.gz (127.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

modern_onnx_clip-0.1.3-py3-none-any.whl (14.5 kB view details)

Uploaded Python 3

File details

Details for the file modern_onnx_clip-0.1.3.tar.gz.

File metadata

  • Download URL: modern_onnx_clip-0.1.3.tar.gz
  • Upload date:
  • Size: 127.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for modern_onnx_clip-0.1.3.tar.gz
Algorithm Hash digest
SHA256 4618abbb63c852307c5cf504e0637494d60cb1fc65f5683ed351eabd7ce48619
MD5 879d9ec82e9eb13a5f9b085b12f20b11
BLAKE2b-256 41dd3c43eb8156c30540898e341cca8eb7cf8ef531abc2fae1ab46262bc008e7

See more details on using hashes here.

Provenance

The following attestation bundles were made for modern_onnx_clip-0.1.3.tar.gz:

Publisher: publish.yml on Neizvestnyj/modern-onnx-clip

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file modern_onnx_clip-0.1.3-py3-none-any.whl.

File metadata

File hashes

Hashes for modern_onnx_clip-0.1.3-py3-none-any.whl
Algorithm Hash digest
SHA256 e04e2d643d356153768c3b31e63689246035cc3006cf4660eb33d6f9f84ef887
MD5 db605c1b76a20ca1f93ef175728aa448
BLAKE2b-256 0acf663d0da166f075f4f6564c774d9191e7ae4b8fa3c49afc2b8c4b5a5be837

See more details on using hashes here.

Provenance

The following attestation bundles were made for modern_onnx_clip-0.1.3-py3-none-any.whl:

Publisher: publish.yml on Neizvestnyj/modern-onnx-clip

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page