Skip to main content

Convert GGUF models to OomLlama's compact OOM format - 2x smaller

Project description

gguf2oom

Convert GGUF models to OomLlama's compact OOM format - 2x smaller

PyPI License: MIT

Quick Start

pip install gguf2oom

# Convert any GGUF to OOM Q2
gguf2oom model.gguf model.oom

# Show GGUF file info
gguf2oom --info model.gguf

Why Convert to OOM?

Format 32B Model 70B Model
GGUF Q4_K ~20 GB ~40 GB
OOM Q2 ~10 GB ~20 GB

The OOM format uses Q2 quantization (2-bit weights) with per-block scale/min values, achieving ~2x compression vs GGUF Q4.

Usage

# Basic conversion
gguf2oom input.gguf output.oom

# Show model info without converting
gguf2oom --info input.gguf

# Help
gguf2oom --help

How It Works

  1. Reads GGUF file (any quantization: Q4_K, Q8_0, F16, etc.)
  2. Dequantizes each tensor to FP32
  3. Requantizes to OOM Q2 format (2 bits per weight)
  4. Writes compact .oom file with OOML magic header

Use with OomLlama

# Install both
pip install gguf2oom oomllama

# Convert
gguf2oom humotica-32b.gguf humotica-32b.oom

# Run inference
oomllama generate --model humotica-32b.oom "Hello!"

Platform Support

The converter automatically downloads the right binary for your platform:

  • Linux x86_64
  • Linux aarch64 (coming soon)
  • macOS x86_64 (coming soon)
  • macOS arm64 (coming soon)

Binaries are cached in ~/.cache/gguf2oom/

Links

Credits

  • Converter: Humotica AI Lab
  • OOM Format: Gemini IDD & Root AI
  • GGUF Reader: Inspired by llama.cpp

One Love, One fAmIly 🦙

Built by Humotica AI Lab

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

gguf2oom-0.1.0.tar.gz (3.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

gguf2oom-0.1.0-py3-none-any.whl (3.8 kB view details)

Uploaded Python 3

File details

Details for the file gguf2oom-0.1.0.tar.gz.

File metadata

  • Download URL: gguf2oom-0.1.0.tar.gz
  • Upload date:
  • Size: 3.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.5

File hashes

Hashes for gguf2oom-0.1.0.tar.gz
Algorithm Hash digest
SHA256 711a9eff9750fb4b02208d3e2456db0c39cc3133dc252154cd70bda87173380b
MD5 a01696cc872991873806d727c357e7e3
BLAKE2b-256 68382ea57cc6d5797f685ae32821fce4332a83cb71e80cd5555391b16fbffe31

See more details on using hashes here.

File details

Details for the file gguf2oom-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: gguf2oom-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 3.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.5

File hashes

Hashes for gguf2oom-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 64b980dab05f7cd4233e169602fda57bb3d41c0995b0eab2750ee5d1ca53b114
MD5 8166162a016b38dfc71b93e4e6207d70
BLAKE2b-256 43902ec2c4a7d288ca03f6afc4c6545941007f5965d241a2c29815657491e3df

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page