Skip to main content

Deterministic conversational state engine for LLM applications.

Project description

Context Compiler

PyPI version Python versions License

A deterministic directive engine that converts explicit user instructions into structured conversational state for LLM applications.

LLMs are good at reasoning but unreliable at maintaining consistent state. Constraints drift, corrections compete, and long conversations accumulate contradictions.

The Context Compiler introduces a deterministic state layer that governs authoritative conversational state independently of the model.

The model performs reasoning and generation while the compiler manages premise and policies. Once accepted, directives remain authoritative until explicitly corrected or reset.

Quickstart

pip install context-compiler
context-compiler

Or in code:

from context_compiler import create_engine

engine = create_engine()

user_input = "prohibit peanuts"
decision = engine.step(user_input)

if decision["kind"] == "clarify":
    show_to_user(decision["prompt_to_user"])
elif decision["kind"] == "update":
    messages = build_messages(engine.state, user_input)
    render(call_llm(messages))
else:
    render(call_llm(user_input))

Installation

Requirements:

  • Python 3.11+

Install:

pip install context-compiler

Development

uv sync --group dev
uv run pytest

Why “Compiler”?

Context Compiler treats explicit user directives as inputs to a deterministic process.

Instead of relying on the LLM to remember constraints across a conversation, user instructions are compiled into structured state before the model runs.

The idea is similar to a traditional compiler: user directives are translated into a structured representation that the rest of the system can rely on.


10-Second Example

User sets a constraint once:

User: prohibit peanuts

Outcome: policy state includes "peanuts": "prohibit".

Later in the conversation:

User: how should I make this curry?

The host supplies the authoritative state to the model so the constraint persists across turns.


Evidence (cross-model runs)

Behavior was evaluated using a fixed set of deterministic demo scenarios.

A run is considered a "pass" if the model output satisfies the scenario’s expected behavior.

  • Models tested: llama3.1:8b, gpt-4o-mini, gpt-4.1, gpt-5, claude-sonnet-4, claude-opus-4
  • Demo scenarios (all pass with compiler) cover ambiguity handling, constraint persistence, correction replacement, and tool governance.
  • Pass-rate summary: baseline (LLM only) 2–4 / 6; with compiler 6 / 6; with compiler + compaction 6 / 6.
  • Context reduction in long conversations: up to 99%
  • Prompt size reduction: about 50%
  • SWE curated results (compiler vs baseline) — cross-model evaluation on 6 tasks showing mostly positive deltas

Architecture

User Input
     │
     ▼
Context Compiler
     │
     ▼
Decision
     │
     ▼
Host Application
 ├─ clarify → ask user
 ├─ passthrough → call LLM
 └─ update → call LLM with compiled state

The compiler governs authoritative state and never calls the LLM. The host decides whether to call the model based on the returned Decision.


Decision API

Each user message produces a Decision.

class Decision(TypedDict):
    kind: Literal["passthrough", "update", "clarify"]
    state: dict | None
    prompt_to_user: str | None

Meaning:

kind host behavior
passthrough forward user input to LLM
update forward input with updated state
clarify show prompt_to_user and do not call the LLM

API Reference

API Description
create_engine(state=None) Create a new compiler engine; optional state provides initial authoritative state (validated/canonicalized).
step(user_input) Parse one user turn and return a deterministic Decision.
compile_transcript(messages) Replay a transcript from a fresh engine and return either final state or a confirmation prompt.
engine.apply_transcript(messages) Replay a transcript onto the current engine state and return either final state or a confirmation prompt.
engine.state Read current authoritative in-memory state snapshot.
get_premise_value(state) Read the current premise value from a state snapshot.
get_policy_items(state, value=None) Read policy items from a state snapshot (all, use, or prohibit).
engine.export_json() Export current state as JSON for persistence/transport.
engine.import_json(payload) Load/restore state from exported JSON payload.

State Model

The compiler maintains an authoritative state snapshot.

  • Premise is a single value that can be set or replaced
  • Policies are per-item (use or prohibit)
  • State changes only through explicit directives
  • No inference or semantic reasoning

Identical input sequences always produce identical state.

The internal structure of the state is intentionally opaque to host applications.


Directive Examples

Set and change premise:

User: set premise concise replies
User: change premise to concise bullet points

Per-item policies:

User: use docker
User: prohibit peanuts

Replacement:

User: use podman instead of docker

Removal and reset:

User: remove policy peanuts
User: reset policies
User: clear state

Conflicting directives trigger clarification instead of changing state.

For full directive grammar and edge-case behavior, see DirectiveGrammarSpec.md.


Examples


Conformance Fixtures

Cross-language conformance tests are defined in tests/fixtures/.


Advanced topics


Guarantees

  • State changes only through explicit user directives or confirmation.
  • Identical input sequences produce identical compiler state.
  • Model responses never modify compiler state.
  • Ambiguous directives trigger clarification instead of changing state.

These invariants are verified through behavioral tests and Hypothesis-based property tests.


Design Notes

More detailed design and milestone documents are available in:


License

Apache-2.0.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

context_compiler-0.5.2.tar.gz (238.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

context_compiler-0.5.2-py3-none-any.whl (15.0 kB view details)

Uploaded Python 3

File details

Details for the file context_compiler-0.5.2.tar.gz.

File metadata

  • Download URL: context_compiler-0.5.2.tar.gz
  • Upload date:
  • Size: 238.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for context_compiler-0.5.2.tar.gz
Algorithm Hash digest
SHA256 356a699b323ac567c4b8e4d2174d2be6526d076b9d96e7caf297179175a1573f
MD5 98c1d80dd560d0a2544ad1631b3608e0
BLAKE2b-256 a08662f56c7c2346d1a150a18a4457dee6d22fb4fe5471ce8e443c71d51c3136

See more details on using hashes here.

Provenance

The following attestation bundles were made for context_compiler-0.5.2.tar.gz:

Publisher: publish-pypi.yml on rlippmann/context-compiler

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file context_compiler-0.5.2-py3-none-any.whl.

File metadata

File hashes

Hashes for context_compiler-0.5.2-py3-none-any.whl
Algorithm Hash digest
SHA256 6ff71f9db4c711fbd514fb21066901d5c356ef24d0560cc39c668e1ad4e8cccc
MD5 dd1ba91ca7f5355d4bebd0c640227439
BLAKE2b-256 7e26f2545d99442b83e41c7cc66e196e20be73d7763e9a714ca7e3e3d2af55e2

See more details on using hashes here.

Provenance

The following attestation bundles were made for context_compiler-0.5.2-py3-none-any.whl:

Publisher: publish-pypi.yml on rlippmann/context-compiler

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page