Skip to main content

The Context Optimization Layer for LLM Applications - Cut costs by 50-90%

Project description

Headroom

Compress everything your AI agent reads. Same answers, fraction of the tokens.

CI codecov PyPI npm Model: Kompress-base Tokens saved: 60B+ License: Apache 2.0 Docs

Headroom in action

Every tool call, log line, DB read, RAG chunk, and file your agent injects into a prompt is mostly boilerplate. Headroom strips the noise and keeps the signal — losslessly, locally, and without touching accuracy.

100 logs. One FATAL error buried at position 67. Both runs found it. Baseline 10,144 tokens → Headroom 1,260 tokens87% fewer, identical answer. python examples/needle_in_haystack_test.py


Quick start

Works with Anthropic, OpenAI, Google, Bedrock, Vertex, Azure, OpenRouter, and 100+ models via LiteLLM.

Wrap your coding agent — one command:

pip install "headroom-ai[all]"

headroom wrap claude      # Claude Code
headroom wrap codex       # Codex
headroom wrap cursor      # Cursor
headroom wrap aider       # Aider
headroom wrap copilot     # GitHub Copilot CLI

Using pipx? Current release wheels are built for Python 3.10 through 3.13, so choose a supported interpreter explicitly:

pipx install --python python3.13 "headroom-ai[all]"

Drop it into your own code — Python or TypeScript:

from headroom import compress

result = compress(messages, model="claude-sonnet-4-5")
response = client.messages.create(model="claude-sonnet-4-5", messages=result.messages)
print(f"Saved {result.tokens_saved} tokens ({result.compression_ratio:.0%})")
import { compress } from 'headroom-ai';
const result = await compress(messages, { model: 'gpt-4o' });

Or run it as a proxy — zero code changes, any language:

headroom proxy --port 8787
ANTHROPIC_BASE_URL=http://localhost:8787 your-app
OPENAI_BASE_URL=http://localhost:8787/v1 your-app

Why Headroom

  • Accuracy-preserving. GSM8K 0.870 → 0.870 (±0.000). TruthfulQA +0.030. SQuAD v2 and BFCL both 97% accuracy after compression. Validated on public OSS benchmarks you can rerun yourself.
  • Runs on your machine. No cloud API, no data egress. Compression latency is milliseconds — faster end-to-end for Sonnet / Opus / GPT-4 class models than a hosted service round-trip.
  • Kompress-base on HuggingFace. Our open-source text compressor, fine-tuned on real agentic traces — tool outputs, logs, RAG chunks, code. Install with pip install "headroom-ai[ml]".
  • Cross-agent memory and learning. Claude Code saves a fact, Codex reads it back. headroom learn mines failed sessions and writes corrections straight to CLAUDE.md / AGENTS.md / GEMINI.md — reliability compounds over time.
  • Reversible (CCR). Compression is not deletion. The model can always call headroom_retrieve to pull the original bytes. Nothing is thrown away.

Bundles managed RTK and lean-ctx binaries for local CLI context filtering — full attribution below.


How it fits

 Your agent / app
   (Claude Code, Cursor, Codex, LangChain, Agno, Strands, your own code…)
        │   prompts · tool outputs · logs · RAG results · files
        ▼
    ┌────────────────────────────────────────────────────┐
    │  Headroom   (runs locally — your data stays here)  │
    │  ───────────────────────────────────────────────   │
    │  CacheAligner  →  ContentRouter  →  CCR             │
    │                    ├─ SmartCrusher   (JSON)         │
    │                    ├─ CodeCompressor (AST)          │
    │                    └─ Kompress-base  (text, HF)     │
    │                                                     │
    │  Cross-agent memory  ·  headroom learn  ·  MCP      │
    └────────────────────────────────────────────────────┘
        │   compressed prompt  +  retrieval tool
        ▼
 LLM provider  (Anthropic · OpenAI · Bedrock · …)

Architecture · CCR reversible compression · Kompress-base model card

Canonical pipeline lifecycle

Headroom now exposes one stable request lifecycle across compress(), the SDK, and the proxy:

SetupPre-StartPost-StartInput ReceivedInput CachedInput RoutedInput CompressedInput RememberedPre-SendPost-SendResponse Received

  • Transforms still do the work: CacheAligner, ContentRouter, SmartCrusher, CodeCompressor, Kompress-base, IntelligentContext / RollingWindow.
  • Pipeline extensions observe or customize those lifecycle stages via on_pipeline_event(...).
  • Compression hooks still work and now sit alongside the canonical lifecycle instead of being the only extension seam.
  • Proxy extensions remain the server/app integration seam for ASGI middleware, routes, and startup policy.

Provider slices

Provider and tool-specific behavior is being moved behind dedicated modules under headroom/providers/ so core orchestration stays focused on lifecycle, sequencing, and policy.

  • CLI/tool slices: headroom/providers/claude, copilot, codex, openclaw
  • Provider runtime slices: headroom/providers/claude, gemini, plus shared backend/runtime dispatch in headroom/providers/registry.py
  • Core files stay orchestration-first: wrap.py, client.py, cli/proxy.py, and proxy/server.py now delegate provider-specific env shaping, API target normalization, backend selection, and transport dispatch instead of inlining those rules.

Proof

Savings on real agent workloads:

Workload Before After Savings
Code search (100 results) 17,765 1,408 92%
SRE incident debugging 65,694 5,118 92%
GitHub issue triage 54,174 14,761 73%
Codebase exploration 78,502 41,254 47%

Accuracy preserved on standard benchmarks:

Benchmark Category N Baseline Headroom Delta
GSM8K Math 100 0.870 0.870 ±0.000
TruthfulQA Factual 100 0.530 0.560 +0.030
SQuAD v2 QA 100 97% 19% compression
BFCL Tools 100 97% 32% compression

Reproduce:

python -m headroom.evals suite --tier 1

Community, live:

Full benchmarks & methodology


Built for coding agents

Agent One-command wrap Notes
Claude Code headroom wrap claude --memory for cross-agent memory, --code-graph for codebase intel
Codex headroom wrap codex --memory Shares the same memory store as Claude
Cursor headroom wrap cursor Prints Cursor config — paste once, done
Aider headroom wrap aider Starts proxy, launches Aider
Copilot CLI headroom wrap copilot Starts proxy, launches Copilot
OpenClaw headroom wrap openclaw Installs Headroom as ContextEngine plugin

MCP-native too — headroom mcp install exposes headroom_compress, headroom_retrieve, and headroom_stats to any MCP client.

headroom learn in action

Integrations

Drop Headroom into any stack
Your setup Hook in with
Any Python app compress(messages, model=…)
Any TypeScript app await compress(messages, { model })
Anthropic / OpenAI SDK withHeadroom(new Anthropic()) · withHeadroom(new OpenAI())
Vercel AI SDK wrapLanguageModel({ model, middleware: headroomMiddleware() })
LiteLLM litellm.callbacks = [HeadroomCallback()]
LangChain HeadroomChatModel(your_llm)
Agno HeadroomAgnoModel(your_model)
Strands Strands guide
ASGI apps app.add_middleware(CompressionMiddleware)
Multi-agent SharedContext().put / .get
MCP clients headroom mcp install
What's inside
  • SmartCrusher — universal JSON: arrays of dicts, nested objects, mixed types.
  • CodeCompressor — AST-aware for Python, JS, Go, Rust, Java, C++.
  • Kompress-base — our HuggingFace model, trained on agentic traces.
  • Image compression — 40–90% reduction via trained ML router.
  • CacheAligner — stabilizes prefixes so Anthropic/OpenAI KV caches actually hit.
  • IntelligentContext — score-based context fitting with learned importance.
  • CCR — reversible compression; LLM retrieves originals on demand.
  • Cross-agent memory — shared store, agent provenance, auto-dedup.
  • SharedContext — compressed context passing across multi-agent workflows.
  • headroom learn — plugin-based failure mining for Claude, Codex, Gemini.

Install

pip install "headroom-ai[all]"          # Python, everything
npm  install headroom-ai                # TypeScript / Node
docker pull ghcr.io/chopratejas/headroom:latest

Granular extras: [proxy], [mcp], [ml] (Kompress-base), [agno], [langchain], [evals]. Requires Python 3.10+.

Installation guide — Docker tags, persistent service, PowerShell, devcontainers.


Documentation

Start here Go deeper
Quickstart Architecture
Proxy How compression works
MCP tools CCR — reversible compression
Memory Cache optimization
Failure learning Benchmarks
Configuration Limitations

Compared to

Headroom runs locally, covers every content type (not just CLI or text), works with every major framework, and is reversible.

Scope Deploy Local Reversible
Headroom All context — tools, RAG, logs, files, history Proxy · library · middleware · MCP Yes Yes
RTK CLI command outputs CLI wrapper Yes No
lean-ctx CLI commands, MCP tools, editor rules CLI wrapper · MCP Yes No
Compresr, Token Co. Text sent to their API Hosted API call No No
OpenAI Compaction Conversation history Provider-native No No

Attribution. Headroom ships with the excellent RTK binary for shell-output rewriting — git showgit show --short, noisy ls → scoped, chatty installers → summarized. Huge thanks to the RTK team; their tool is a first-class part of our stack, and Headroom compresses everything downstream of it. Headroom can also use lean-ctx as the selected CLI context tool; set HEADROOM_CONTEXT_TOOL=lean-ctx before running headroom wrap ....


Contributing

git clone https://github.com/chopratejas/headroom.git && cd headroom
pip install -e ".[dev]" && pytest

Devcontainers in .devcontainer/ (default + memory-stack with Qdrant & Neo4j). See CONTRIBUTING.md.


Community

License

Apache 2.0 — see LICENSE.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

headroom_ai-0.21.27.tar.gz (1.6 MB view details)

Uploaded Source

Built Distributions

If you're not sure about the file name format, learn more about wheel file names.

headroom_ai-0.21.27-cp313-cp313-manylinux_2_28_x86_64.whl (18.5 MB view details)

Uploaded CPython 3.13manylinux: glibc 2.28+ x86-64

headroom_ai-0.21.27-cp313-cp313-manylinux_2_28_aarch64.whl (19.6 MB view details)

Uploaded CPython 3.13manylinux: glibc 2.28+ ARM64

headroom_ai-0.21.27-cp313-cp313-macosx_11_0_arm64.whl (17.4 MB view details)

Uploaded CPython 3.13macOS 11.0+ ARM64

headroom_ai-0.21.27-cp312-cp312-manylinux_2_28_x86_64.whl (18.5 MB view details)

Uploaded CPython 3.12manylinux: glibc 2.28+ x86-64

headroom_ai-0.21.27-cp312-cp312-manylinux_2_28_aarch64.whl (19.6 MB view details)

Uploaded CPython 3.12manylinux: glibc 2.28+ ARM64

headroom_ai-0.21.27-cp312-cp312-macosx_11_0_arm64.whl (17.4 MB view details)

Uploaded CPython 3.12macOS 11.0+ ARM64

headroom_ai-0.21.27-cp311-cp311-manylinux_2_28_x86_64.whl (18.5 MB view details)

Uploaded CPython 3.11manylinux: glibc 2.28+ x86-64

headroom_ai-0.21.27-cp311-cp311-manylinux_2_28_aarch64.whl (19.6 MB view details)

Uploaded CPython 3.11manylinux: glibc 2.28+ ARM64

headroom_ai-0.21.27-cp311-cp311-macosx_11_0_arm64.whl (17.4 MB view details)

Uploaded CPython 3.11macOS 11.0+ ARM64

headroom_ai-0.21.27-cp310-cp310-manylinux_2_28_x86_64.whl (18.5 MB view details)

Uploaded CPython 3.10manylinux: glibc 2.28+ x86-64

headroom_ai-0.21.27-cp310-cp310-manylinux_2_28_aarch64.whl (19.6 MB view details)

Uploaded CPython 3.10manylinux: glibc 2.28+ ARM64

headroom_ai-0.21.27-cp310-cp310-macosx_11_0_arm64.whl (17.4 MB view details)

Uploaded CPython 3.10macOS 11.0+ ARM64

File details

Details for the file headroom_ai-0.21.27.tar.gz.

File metadata

  • Download URL: headroom_ai-0.21.27.tar.gz
  • Upload date:
  • Size: 1.6 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for headroom_ai-0.21.27.tar.gz
Algorithm Hash digest
SHA256 55bc7a080b5a86e637ce7019f8155fe236cfde6c3591be388c0c992fbefcad16
MD5 566aa169740125c3632bd6a32aed1ce4
BLAKE2b-256 4719edff343e6e25baf6106ace8284b363c81c5dd604f2ace0e31d33c24d40c9

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27.tar.gz:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp313-cp313-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp313-cp313-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 c085b472b637cdcce7b852349bb6899ce264f88b03d803de65d43178fdb28a77
MD5 9bf8580095be4fb19e95c01d7c6fb6c7
BLAKE2b-256 3bcb543e79c6fc06af0b09057078264b28b297b0dac2004ccdc8c6153aa5c610

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp313-cp313-manylinux_2_28_x86_64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp313-cp313-manylinux_2_28_aarch64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp313-cp313-manylinux_2_28_aarch64.whl
Algorithm Hash digest
SHA256 5fecb424ab5816a74b3ec8326b3ea0b3f5397e85fb5f9bb26e0a03b5d27c319e
MD5 90a3964881ebe0cd5cefa07d60ecb0ac
BLAKE2b-256 c7c782ce0fe7df398ba2d737fbaabcddaefb8b8eb38eaa138e5f6bab0870705e

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp313-cp313-manylinux_2_28_aarch64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp313-cp313-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp313-cp313-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 9b1dc66a7f8135b2262c7901768ab760fdb196a77067e9b450e8aa71feeba050
MD5 d993624270829675c63cc20714725e48
BLAKE2b-256 65f8c85ae02439525ca1e3afda0fafb287821319a400a5f8e956a712f131cb89

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp313-cp313-macosx_11_0_arm64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp312-cp312-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp312-cp312-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 fc45aa8fd7100bfd67f91a5ae734059f189a77e4c2d84a00914bc473c8370adb
MD5 441814571b8e50186dcb0913db05f759
BLAKE2b-256 1e4a0e0617c741b51a74eea508824384a62e7e29e7f30ae8be3b5ddc35e081af

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp312-cp312-manylinux_2_28_x86_64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp312-cp312-manylinux_2_28_aarch64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp312-cp312-manylinux_2_28_aarch64.whl
Algorithm Hash digest
SHA256 3c2f8946029bdd2321ed962e8c30ee85698403d2262259df3eb9085edd2a3644
MD5 3a022942931c4fa82f66dee35f4fff1a
BLAKE2b-256 79072c98d30a526034cb5093445d6bbd614ca0db95321f20005d60273e2fa59f

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp312-cp312-manylinux_2_28_aarch64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp312-cp312-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp312-cp312-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 e50280f71bd6f4992dc9443add1a03c1c90dfaeae14d24e3cd75d84a1a0a3017
MD5 46e46ba946210b553aac29ad60f69743
BLAKE2b-256 17def24e6bef6bfce85dd9056b6d8253c72c9725a6750797dba4780c1bf7b55d

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp312-cp312-macosx_11_0_arm64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp311-cp311-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp311-cp311-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 6616b9380e5eeb265f78ec59e4cecd9021e3618516b68c3f4bac1d1ef76df69b
MD5 29aa29e89abb5741bd0fdf2a42b7ad5e
BLAKE2b-256 db000bd7021c333b3e8ca73ca50684a06219069ddc1808f025fe5c399a40a8ff

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp311-cp311-manylinux_2_28_x86_64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp311-cp311-manylinux_2_28_aarch64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp311-cp311-manylinux_2_28_aarch64.whl
Algorithm Hash digest
SHA256 3af51a416a8e5175b637f65fa0ccb94ff19e5c26ccbec36aaa8e9d3c6037aa8b
MD5 70efae6fd90b6e2bcef807756f924b08
BLAKE2b-256 0d38277a12e6d48312f3286dbb1de3f0ab9a27dab786fe426f6bfc271c774665

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp311-cp311-manylinux_2_28_aarch64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp311-cp311-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp311-cp311-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 871b30599fcad4b7d7cf2fdf05df16129ac96319adbaca2850748cfcaa50d15e
MD5 f20179d3a6d75144bd0498f2a4233b27
BLAKE2b-256 50aeb465ab23ca7f94d8aa522d1b1b4de66b35d647f08211f0adbb92002d1eb4

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp311-cp311-macosx_11_0_arm64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp310-cp310-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp310-cp310-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 407b679fa64907e1dac43fe84cd79bc8cce1f6c03750e883b76da7ecd26d3e42
MD5 ce61550d20a8c7ee6bf0fa27994428cf
BLAKE2b-256 62f9807b315b20635f5491cfb129535ba715d92cd154ccf032ebb76ea1ad69cd

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp310-cp310-manylinux_2_28_x86_64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp310-cp310-manylinux_2_28_aarch64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp310-cp310-manylinux_2_28_aarch64.whl
Algorithm Hash digest
SHA256 153b7825145ab9047ef20b93212b2b5052c808382911b4ed88c2d2158552800d
MD5 4cfc1ee33e0c5e508a702b2f2d95bd61
BLAKE2b-256 bffbd21538a0719962adcd48fea86c217d102b739e1be2a18031b782a6ff81d7

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp310-cp310-manylinux_2_28_aarch64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file headroom_ai-0.21.27-cp310-cp310-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for headroom_ai-0.21.27-cp310-cp310-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 794a5782c040d6f2eb2b496cba411a8d21d4e1628843ae3b88c89b3928b91e5a
MD5 cf530528add1ba139485eb2fb475f95e
BLAKE2b-256 e0e420093367886acfd0930ace00f56653c805d62120452c70c1701695d5ebb0

See more details on using hashes here.

Provenance

The following attestation bundles were made for headroom_ai-0.21.27-cp310-cp310-macosx_11_0_arm64.whl:

Publisher: release.yml on chopratejas/headroom

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page