OverClaw — autonomous agent optimization through structured experimentation
Project description
OverClaw
Automatically optimize your AI agent's prompts, tool definitions, model selection, and pipeline logic through structured experimentation.
What it does
OverClaw runs your agent against a test dataset, traces every LLM call and tool invocation, scores the outputs, and uses a strong reasoning model to generate concrete improvements. Changes that raise the score are kept; the rest are reverted. After several rounds you get a measurably better agent — without manual prompt tweaking.
What makes OverClaw different is policy-driven optimization. You define the decision rules, constraints, and expectations your agent must follow, and those policies guide every stage: evaluation criteria, test data synthesis, optimization diagnosis, and scoring.
What gets optimized
- System prompts — more precise instructions, output format enforcement
- Tool descriptions — clearer parameters, better usage guidance
- Model selection — find the right quality/cost tradeoff
- Agent logic — tool-call ordering, iteration limits, output parsing
- Policy compliance — alignment with your domain rules and constraints
Quick start
Requirements: Python 3.10+, uv, and API keys for at least one LLM provider (OpenAI, Anthropic).
# 1. Install OverClaw as a CLI tool
uv tool install -e .
# 2. Set API keys and model defaults
overclaw init
# 3. Register your agent with a name and entrypoint
overclaw agent register lead-qualification agents.agent1.sample_agent:run
# 4. Analyze your agent, define policies, and build evaluation criteria
overclaw setup lead-qualification
# 5. Run the optimizer
overclaw optimize lead-qualification
Have an existing policy document? Pass it directly to setup:
overclaw setup lead-qualification --policy docs/my_policy.md
Using
uv runinstead? If you prefer not to install globally, all commands work asuv run overclaw <command>afteruv sync.
How it works
1. Initialize (overclaw init)
Configure API keys and default models. Writes .overclaw/.env in the current
directory. Safe to re-run.
2. Register your agent (overclaw agent register)
Point OverClaw at the Python function it should call for each test case:
overclaw agent register <name> <module:function>
The module path is resolved relative to the project root. Your function receives an input dict and must return a dict.
Other registry commands:
| Command | Description |
|---|---|
overclaw agent list |
List all registered agents |
overclaw agent show <name> |
Show registration details and pipeline status |
overclaw agent update <name> <mod:fn> |
Update the entrypoint (e.g. after renaming a file) |
overclaw agent remove <name> |
Remove from registry (does not delete files) |
3. Setup (overclaw setup)
An interactive flow that prepares everything the optimizer needs:
| Phase | What happens |
|---|---|
| Agent analysis | An LLM reads your agent code to detect the input/output schema, tools, and decision logic. |
| Policy generation | If you pass --policy, your document is analyzed against the code and improvements are suggested. Otherwise, a policy is inferred from the code automatically. You can refine either version in a conversational loop until you approve it. |
| Dataset | OverClaw either uses your existing test data or generates diverse synthetic cases based on the policy and agent description. |
| Evaluation criteria | Scoring rules are proposed for each output field. Policy constraints inform stricter scoring where relevant. You can accept, refine, or edit manually. |
Setup produces two artifacts in .overclaw/agents/<name>/setup_spec/:
eval_spec.json— machine-readable evaluation spec (used at runtime)policies.md— human-readable policy document you maintain
Both are editable after generation.
| Flag | Description |
|---|---|
--fast |
Skip all prompts. Requires ANALYZER_MODEL and SYNTHETIC_DATAGEN_MODEL in .env. |
--policy PATH |
Provide an existing policy document. OverClaw analyzes it against agent code and suggests edits. |
4. Optimize (overclaw optimize)
The iterative optimization loop. You configure a few settings interactively
(or use --fast for defaults):
| Setting | Description |
|---|---|
| Analyzer model | The strong model that diagnoses failures and generates code fixes. |
| LLM-as-Judge | Optional semantic scoring alongside mechanical matching (adds ~10% eval cost). |
| Iterations | Number of optimize → evaluate → accept/revert rounds (default: 5). |
| Candidates per iteration | How many variant fixes to generate per round (best-of-N). Each biases edits toward a different area — tool descriptions, core logic, input handling, system prompt. Higher N improves odds but costs more. |
| Parallel execution | Run agent evaluations across multiple workers. |
What happens each iteration
- Run the agent on every test case and collect traces + outputs.
- Score outputs against the eval spec (0–100 across dynamic dimensions).
- Diagnose — the analyzer receives traces, scores, policy, and code. It identifies failure patterns and root causes.
- Generate N candidate fixes, each targeting a different area of the code. If N≥3, the last candidate uses a separate diagnosis for diversity.
- Validate — syntax checks, interface checks, and a smoke test on a small case subset.
- Evaluate — surviving candidates are scored on the full dataset.
- Accept or revert — the best candidate is kept only if it improves the score without regressing too many individual cases.
Advanced settings (available during interactive config) include regression thresholds, train/holdout splits to detect overfitting, early stopping patience, and diagnosis visibility controls.
| Flag | Description |
|---|---|
--fast |
Skip all prompts. Requires ANALYZER_MODEL in .env. Uses defaults. |
Multi-file agents
By default OverClaw optimizes the single registered entry file. For agents split across multiple modules, it automatically resolves local imports, extracts individual functions and classes, and applies targeted edits back to the original files — so your project structure stays intact.
Agent policies
Policies are the foundation of meaningful optimization. They tell the optimizer what the agent should do, not just how it currently scores — preventing improvements that raise numbers but violate business rules.
A policies.md looks like this:
# Agent Policy: Lead Qualification
## Purpose
Qualifies inbound sales leads by analyzing company data and inquiry content.
## Decision Rules
1. If the inquiry mentions "enterprise" or "custom pricing", classify as hot
2. Companies with 500+ employees get a minimum lead score of 60
## Constraints
- Never disqualify without checking company size
- Score and category must be consistent (hot = 70+, warm = 40-69, cold = <40)
## Priority Order
1. Accuracy of category classification
2. Score calibration
3. Reasoning quality
## Edge Cases
| Scenario | Expected Behaviour |
|----------------------|---------------------------------------|
| Missing company name | Default to cold, note in reasoning |
| Competitor inquiry | Classify as cold, recommend nurture |
## Quality Expectations
- Reasoning should reference specific data points from the input
- Scores should be calibrated: hot leads 70-100, warm 40-69, cold 0-39
Policies feed into diagnosis prompts, code generation constraints, synthetic data generation, and LLM-as-Judge scoring — so every stage of the pipeline respects your domain rules.
Using your own data
Data files are JSON arrays where each element has an input and
expected_output:
[
{
"input": { "company_name": "Acme Corp", "inquiry": "Need enterprise pricing" },
"expected_output": { "category": "hot", "lead_score": 85 }
}
]
Place data files in your agent directory under data/ and OverClaw will
detect them during setup. If you don't have data, OverClaw generates realistic
synthetic test cases using the policy and agent description.
Output
After optimization, results are saved under .overclaw/agents/<name>/:
| Path | Description |
|---|---|
setup_spec/policies.md |
Agent policy document |
setup_spec/eval_spec.json |
Evaluation criteria with embedded policy |
setup_spec/dataset.json |
Test dataset used for optimization |
experiments/best_agent.py |
The highest-scoring agent version |
experiments/best_agent/ |
All optimized files (multi-file agents only) |
experiments/results.tsv |
Score history for every iteration |
experiments/traces/ |
Detailed JSON traces of every agent run |
experiments/report.md |
Summary report with scores and diffs |
CLI reference
overclaw init Configure API keys and models
overclaw agent register <name> <mod:fn> Register an agent
overclaw agent list List registered agents
overclaw agent show <name> Show agent status
overclaw agent update <name> <mod:fn> Update entrypoint
overclaw agent remove <name> Remove from registry
overclaw setup <name> [--fast] [--policy] Analyze agent, build eval spec
overclaw optimize <name> [--fast] Run optimization loop
Run overclaw <command> --help for full documentation on any command.
License
MIT
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file overclaw-0.2.0.tar.gz.
File metadata
- Download URL: overclaw-0.2.0.tar.gz
- Upload date:
- Size: 542.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.11.0 {"installer":{"name":"uv","version":"0.11.0","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
afda312c591b7af8bc4a0a1ce2dc4ba2253913cd9a9f8d1dddacd21480ba03f7
|
|
| MD5 |
7492a2933cfa38a68c3d4bc548164760
|
|
| BLAKE2b-256 |
f15bc19dcec353dd793a10c6ae661bf8605191bcb81432393505cbc07caaa00e
|
File details
Details for the file overclaw-0.2.0-py3-none-any.whl.
File metadata
- Download URL: overclaw-0.2.0-py3-none-any.whl
- Upload date:
- Size: 274.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.11.0 {"installer":{"name":"uv","version":"0.11.0","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
bcc26d20a9fa868be93443f7e9a531aa6d5d67029312f2395016c8a770c3380c
|
|
| MD5 |
4e0918c34a97c63d8f9a9a7c790f8ad0
|
|
| BLAKE2b-256 |
ed1792363e64408c87dd2553fbc3cb3d3a2157c83266e99e3645665f4a0fa89e
|