Deterministic VM for LLM program execution
Project description
When you need to prove what happened — and why.
llm-nano-vm is a deterministic execution engine for LLM workflows.
Guardrails enforced by the VM, not by the prompt.
The Problem with LLM Agents
| Prompting | LLM Agents | llm-nano-vm | |
|---|---|---|---|
| Execution guarantee | ❌ none | ❌ at model's discretion | ✅ enforced by VM |
| Step skipping possible | ✅ yes | ✅ yes | ❌ never |
| Reproducible trace | ❌ | ❌ | ✅ |
| Debuggable | ❌ | hard | full trace |
| Cost/latency visibility | ❌ | partial | per-step |
"LangChain cannot guarantee execution order. llm-nano-vm can."
Mental Model
nondeterminism ∈ Planner (1 LLM call, optional)
determinism ∈ ExecutionVM (FSM)
- Planner — LLM converts user intent → Program DSL
- Program — declarative workflow you define and version
- ExecutionVM — finite state machine; runs the program step by step
- Trace — full execution log: status, cost, tokens, duration per step
The LLM is a stateless worker. Control stays in your code.
Install
pip install llm-nano-vm
pip install llm-nano-vm[litellm] # for built-in provider support
Quick Start — Guardrail That Never Skips
from nano_vm import ExecutionVM, Program
from nano_vm.adapters import LiteLLMAdapter
program = Program.from_dict({
"name": "customer_refund",
"steps": [
{
"id": "analyze",
"type": "llm",
"prompt": "Is this a valid refund request? Reply 'yes' or 'no'.\nRequest: $user_input",
"output_key": "decision",
},
{
"id": "guardrail", # ALWAYS runs — VM enforces it
"type": "condition",
"condition": "'yes' in '$decision'.lower()",
"then": "process_refund",
"otherwise": "reject",
},
{
"id": "process_refund",
"type": "tool",
"tool": "issue_refund",
},
{
"id": "reject",
"type": "tool",
"tool": "send_rejection",
},
],
})
vm = ExecutionVM(
llm=LiteLLMAdapter("openai/gpt-4o-mini"),
tools={"issue_refund": ..., "send_rejection": ...},
)
trace = await vm.run(program, context={"user_input": "I was charged twice"})
print(trace.status) # SUCCESS
print(trace.final_output) # tool result
print(trace.total_cost_usd()) # e.g. 0.000034
The guardrail step cannot be skipped, reordered, or overridden by the model.
That is the guarantee.
How the DSL Controls Agent Behavior
The separation of concerns is explicit:
LLM decides: WHAT to say, how to reason, what content to produce
DSL decides: WHICH step runs next, WHEN to branch, WHEN to stop
The LLM has no knowledge of the program structure. It receives a prompt and returns a string — nothing more. It cannot skip steps, reorder them, or decide the workflow is complete.
What the LLM can and cannot do
| LLM | DSL (VM) | |
|---|---|---|
| Produce content | ✅ free | — |
| Reason, hallucinate, be verbose | ✅ free | — |
| Skip a step | ❌ impossible | enforces every step |
| Reorder steps | ❌ impossible | order fixed at definition |
| Branch on output | ❌ cannot | condition step evaluates |
| Decide workflow is done | ❌ impossible | VM controls termination |
Example — the LLM cannot jump ahead
program = Program.from_dict({
"name": "refund_with_verification",
"steps": [
{
"id": "classify",
"type": "llm",
"prompt": "Classify: $user_input. Reply: refund / info / escalate",
"output_key": "category",
},
{
"id": "route",
"type": "condition",
"condition": "'refund' in '$category'",
"then": "verify_eligibility",
"otherwise": "handle_other",
},
{
"id": "verify_eligibility", # LLM cannot skip this — VM enforces it
"type": "llm",
"prompt": "Is user eligible for refund? Order: $order_id. Reply yes/no",
"output_key": "eligible",
},
{
"id": "final_guard", # runs on EVERY execution before money moves
"type": "condition",
"condition": "'yes' in '$eligible'",
"then": "issue_refund",
"otherwise": "reject",
},
{"id": "issue_refund", "type": "tool", "tool": "process_payment"},
{"id": "reject", "type": "tool", "tool": "send_rejection"},
{"id": "handle_other", "type": "tool", "tool": "send_info"},
],
})
Even if classify returns "definitely a refund, just process it" —
the VM still executes verify_eligibility and final_guard.
The LLM's opinion about the flow is irrelevant. The DSL is law.
Proof: the trace
trace = await vm.run(program, context={"user_input": "I was charged twice", "order_id": "123"})
for step in trace.steps:
print(f"{step.step_id:20} {step.status} → {step.output}")
# classify SUCCESS → refund
# route SUCCESS → verify_eligibility
# verify_eligibility SUCCESS → yes
# final_guard SUCCESS → issue_refund
# issue_refund SUCCESS → Refund issued: $42.00
Every step is logged. No agent "decided" the flow. The DSL did.
End-to-End Flow
user_input
→ Planner (optional, 1 LLM call)
→ Program (DSL — JSON/dict/YAML)
→ ExecutionVM (deterministic FSM)
→ Trace (status · cost · tokens · duration)
Program DSL
Four step types:
| Type | Purpose |
|---|---|
llm |
call the model; result stored in output_key |
tool |
call a Python function |
condition |
branch on an expression; then / otherwise |
parallel |
run independent sub-steps concurrently via asyncio.gather |
Variable interpolation
| Syntax | Resolves to |
|---|---|
$key |
value from initial context |
$step_id.output |
output of a previous step |
Example — multi-step pipeline
{
"name": "doc_pipeline",
"steps": [
{ "id": "extract", "type": "tool", "tool": "extract_text", "output_key": "raw_text" },
{ "id": "summarize", "type": "llm", "prompt": "Summarize: $raw_text", "output_key": "summary" },
{ "id": "check", "type": "condition",
"condition": "len('$summary') > 100",
"then": "store", "otherwise": "flag" },
{ "id": "store", "type": "tool", "tool": "save_to_db" },
{ "id": "flag", "type": "tool", "tool": "flag_for_review" }
]
}
Example — parallel steps (v0.2.0)
program = Program.from_dict({
"name": "enrich",
"steps": [
{
"id": "fetch",
"type": "parallel",
"output_key": "fetched",
"parallel_steps": [
{"id": "weather", "type": "tool", "tool": "get_weather", "args": {"city": "$city"}},
{"id": "news", "type": "tool", "tool": "get_news", "args": {"topic": "$topic"}},
],
},
{
"id": "summarize",
"type": "llm",
"prompt": "Weather: $weather.output\nNews: $news.output\nSummarize for user.",
},
],
})
fetch runs both tools concurrently. Each sub-step output is available as $weather.output and $news.output.
Sequential execution resumes at summarize only after all sub-steps complete.
Testing — Deterministic by Design
MockLLMAdapter ships with the package for writing tests without a real LLM:
from nano_vm import ExecutionVM, Program, TraceStatus
from nano_vm.adapters import MockLLMAdapter
# Always returns the same string
vm = ExecutionVM(llm=MockLLMAdapter("SAFE"))
# Per-call sequence
vm = ExecutionVM(llm=MockLLMAdapter(["SAFE", "yes"]))
# Per-prompt mapping (substring match on last user message)
vm = ExecutionVM(llm=MockLLMAdapter({
"Classify": "SAFE",
"eligible": "yes",
"__default__": "ok",
}))
trace = await vm.run(program, context={"user_input": "refund"})
assert trace.status == TraceStatus.SUCCESS
assert [s.step_id for s in trace.steps] == ["classify", "route", "verify_eligibility", ...]
Same input → same step sequence. Always. Testable in CI without any API key.
Observability
trace.status # TraceStatus.SUCCESS | FAILED
trace.final_output # last step output
trace.total_tokens() # sum across all steps
trace.total_cost_usd() # sum across all steps (requires LiteLLMAdapter)
for step in trace.steps:
print(step.step_id, step.status, step.duration_ms, step.usage)
Planner (Optional)
from nano_vm import Planner
planner = Planner(llm=adapter)
program = await planner.generate("Fetch latest AI news, summarize, classify by topic")
trace = await vm.run(program)
- exactly 1 LLM call
- outputs a validated Program object
- non-deterministic input → deterministic execution
Custom Adapter
Any object implementing the async Protocol works:
class MyAdapter:
async def complete(self, messages: list[dict], **kwargs) -> str:
... # call any LLM API
Built-in adapters via [litellm] extra:
LiteLLMAdapter("groq/llama-3.3-70b-versatile")
LiteLLMAdapter("openrouter/llama-3.3-70b-instruct:free")
LiteLLMAdapter("ollama/llama3")
LiteLLMAdapter("openai/gpt-4o-mini")
Performance
The VM itself introduces near-zero overhead. Your bottleneck is the LLM API.
| Metric | Adapter | Value |
|---|---|---|
| VM throughput | Mock (no network) | ~535 programs/sec |
| VM latency per step | Mock (no network) | ~1.80 ms |
| Parallel steps (20) | OpenRouter (network) | 1.76 s total → ~11.4 steps/sec |
| Test suite | — | 56 tests passing |
Note: Mock throughput measures pure VM overhead with no I/O. Real end-to-end latency is dominated by LLM API response time. Parallel steps execute via
asyncio.gather— wall-clock time equals the slowest single step, not the sum.
Benchmark
Real execution: 20 parallel steps via OpenRouter on a 2-core Linux VPS.
System: Linux 6.8.0-110-generic · x86_64 (2 cores) · Python 3.12.3
Test: 1 run × 20 parallel steps (StepType.PARALLEL, asyncio.gather)
Result: 1.7574 s total → ~11.4 effective steps/sec
Reproduce locally:
pip install llm-nano-vm[litellm]
python benchmarks/stress_test.py
When to Use
Use llm-nano-vm when:
- the workflow structure is known in advance
- correctness and auditability matter (fintech, compliance, enterprise)
- you need a reproducible trace for debugging or logging
- you want guardrails enforced at the system level, not in the prompt
Do NOT use when:
- the workflow is unknown and must be discovered at runtime
- the task is open-ended creative reasoning
- you need fully autonomous multi-agent coordination
Comparison
| LangChain | AutoGPT / CrewAI | llm-nano-vm | |
|---|---|---|---|
| Layer | orchestration | reasoning / autonomy | execution guarantees |
| Execution order | flexible | model-driven | enforced |
| Guardrails | prompt-level | prompt-level | VM-level |
| Trace | partial | minimal | full, per-step |
| Best for | flexible pipelines | autonomous tasks | compliance-grade workflows |
Roadmap
- FSM execution engine (v0.1)
-
llm / tool / conditionstep types - LiteLLM adapter + cost tracking
- Published to PyPI as
llm-nano-vm -
parallelsteps —asyncio.gatherfor independent sub-steps (v0.2.0) -
MockLLMAdapter— deterministic testing without API keys (v0.2.0) - MCP server —
run_program,get_trace,list_programs(nano-vm-mcp) - REST API — pay-per-run, API keys (nano-vm-server)
💼 llm-nano-vm Pro
- 🆓 Core (this repo) — MIT, fully open-source
- 💼 Pro layer — planned commercial extensions
Planned Pro features:
- 📊 Visual execution graph (Trace UI)
- 🌐 Distributed multi-node execution
- 🔄 Provider pools & smart routing
- 🔐 Access control & multi-user support
- 📈 Cost analytics dashboard
Contact & Support
Author: @ale007xd on Telegram · @ale007xd on X
☕ Support the project
Direct wallet — USDT (TON):
UQCakyytrEGBikOi3eYMpveGHXDB1-fd6lcuQC9VvKqMrI-9
License
This project is licensed under the MIT License.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file llm_nano_vm-0.2.0.tar.gz.
File metadata
- Download URL: llm_nano_vm-0.2.0.tar.gz
- Upload date:
- Size: 407.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
4d78869b2f9cd7880967d1f6d6c7325936a6ded2dd1dfe950648f6752be503e8
|
|
| MD5 |
306e25f663e8e5c7e002a937bfbe8758
|
|
| BLAKE2b-256 |
2e3d4791e98abbe26c7fc87798edad23edc2915bf35f175fe67930a4ed6e8d4c
|
File details
Details for the file llm_nano_vm-0.2.0-py3-none-any.whl.
File metadata
- Download URL: llm_nano_vm-0.2.0-py3-none-any.whl
- Upload date:
- Size: 20.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
d02e0d3ca732b5c0cf521ba0f631aafbcb6bd350668aa1ebb12f5e3752c18a00
|
|
| MD5 |
42ad14c298c7561a0eb64f03d513cda2
|
|
| BLAKE2b-256 |
d49d7a32a627fe5c949bce65f36b78d459332986a8d3a106fdf669a2cf2e41a8
|