Convert GGUF models to OomLlama's compact OOM format - 2x smaller
Project description
gguf2oom
Convert GGUF models to OomLlama's compact OOM format - 2x smaller
Quick Start
pip install gguf2oom
# Convert any GGUF to OOM Q2
gguf2oom model.gguf model.oom
# Show GGUF file info
gguf2oom --info model.gguf
Why Convert to OOM?
| Format | 32B Model | 70B Model |
|---|---|---|
| GGUF Q4_K | ~20 GB | ~40 GB |
| OOM Q2 | ~10 GB | ~20 GB |
The OOM format uses Q2 quantization (2-bit weights) with per-block scale/min values, achieving ~2x compression vs GGUF Q4.
Usage
# Basic conversion
gguf2oom input.gguf output.oom
# Show model info without converting
gguf2oom --info input.gguf
# Help
gguf2oom --help
How It Works
- Reads GGUF file (any quantization: Q4_K, Q8_0, F16, etc.)
- Dequantizes each tensor to FP32
- Requantizes to OOM Q2 format (2 bits per weight)
- Writes compact .oom file with OOML magic header
Use with OomLlama
# Install both
pip install gguf2oom oomllama
# Convert
gguf2oom humotica-32b.gguf humotica-32b.oom
# Run inference
oomllama generate --model humotica-32b.oom "Hello!"
Platform Support
The converter automatically downloads the right binary for your platform:
- Linux x86_64
- Linux aarch64 (coming soon)
- macOS x86_64 (coming soon)
- macOS arm64 (coming soon)
Binaries are cached in ~/.cache/gguf2oom/
Links
- OomLlama - Run OOM models
- GitHub
- HuggingFace Models
Credits
- Converter: Humotica AI Lab
- OOM Format: Gemini IDD & Root AI
- GGUF Reader: Inspired by llama.cpp
One Love, One fAmIly 🦙
Built by Humotica AI Lab
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file gguf2oom-0.1.0.tar.gz.
File metadata
- Download URL: gguf2oom-0.1.0.tar.gz
- Upload date:
- Size: 3.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.5
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
711a9eff9750fb4b02208d3e2456db0c39cc3133dc252154cd70bda87173380b
|
|
| MD5 |
a01696cc872991873806d727c357e7e3
|
|
| BLAKE2b-256 |
68382ea57cc6d5797f685ae32821fce4332a83cb71e80cd5555391b16fbffe31
|
File details
Details for the file gguf2oom-0.1.0-py3-none-any.whl.
File metadata
- Download URL: gguf2oom-0.1.0-py3-none-any.whl
- Upload date:
- Size: 3.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.5
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
64b980dab05f7cd4233e169602fda57bb3d41c0995b0eab2750ee5d1ca53b114
|
|
| MD5 |
8166162a016b38dfc71b93e4e6207d70
|
|
| BLAKE2b-256 |
43902ec2c4a7d288ca03f6afc4c6545941007f5965d241a2c29815657491e3df
|