Sage - Simplified AI agent definition and deployment via configuration
Project description
Sage Agent
Yes, I shamelessly named it after me ;)
Inspired by the recent sprawl of OpenClaw, PicoBot, ZeroClaw, and whatever else popped up last Tuesday — I decided to write my own. Written from the ground up in Python.
Sage doesn't aspire to be the next Claude Code. Instead, it's intentionally designed to be a clean slate out of the box, so that you can make it more intelligent. No opinions. No bloat. Just a solid foundation you can build on top of.
Built-in evaluation and CI/headless execution are included. See .docs/eval.md and .docs/ci-headless.md.
Key Features
Agents
The core unit. Define an agent in a Markdown file with YAML frontmatter — name, model, system prompt — and you're running. No boilerplate classes, no framework ceremony. Just config and go.
---
name: assistant
model: gpt-4o
---
You are a helpful AI assistant.
Subagents & Delegation
Agents can have subagents. When they do, they automatically get a delegate tool — the LLM decides when and how to hand off work. It's orchestration without the orchestration code.
Tools via @tool Decorator
Write a Python function. Decorate it with @tool. Sage auto-generates the JSON schema from your type hints. That's it. No manual schema wrangling.
@tool
def word_count(text: str) -> str:
"""Count the number of words in the given text."""
return str(len(text.split()))
Built-in tools included — or load them all at once with sage.tools.builtins:
| Category | Tools |
|---|---|
| Core | shell, file_read, file_write, file_edit, http_request |
| Memory | memory_store, memory_recall |
| Web | web_fetch, web_search |
Skills
Reusable capabilities defined as Markdown files. Drop them in a skills/ directory and all agents share them automatically. Sage resolves the global skill pool via a waterfall (skills_dir in config.toml → ./skills/ → ~/.agents/skills/ → ~/.claude/skills/). Each agent can optionally limit its skills to a named subset via an allowlist in config.toml. Flat files or directory-per-skill — both work.
Orchestration
Four flavors:
- Pipeline (
>>) — chain agents sequentially. Output of one feeds the next. - Parallel — run multiple agents concurrently via
Orchestrator.run_parallel(). - Race — first agent to complete wins via
Orchestrator.run_race(). - Autonomous delegation — an orchestrator agent with subagents decides who does what, on its own.
100+ LLM Providers
Powered by litellm. OpenAI, Azure, Anthropic, Ollama, and basically everything else. One model string, any provider.
| Provider | Model String |
|---|---|
| OpenAI | gpt-4o, gpt-4o-mini |
| Azure | azure/gpt-4o |
| Anthropic | anthropic/claude-sonnet-4-20250514 |
| Ollama | ollama/llama3 |
MCP Support
Connect to MCP servers (stdio or SSE) or expose your tools as an MCP server. Both directions work.
Semantic Memory
SQLite-backed with litellm embeddings. Zero-config persistent recall across sessions. Compaction built in so context doesn't bloat forever.
Permissions
Control what tools can do via a single permission: block in YAML frontmatter. Each permission category (read, edit, shell, web, memory) maps to a set of built-in tools. Set a category to allow, deny, or ask, or use pattern matching for fine-grained shell control. When set to deny, tools are invisible to the LLM. Interactive prompts in the TUI when policy is ask.
Hook System
A lifecycle event bus for intercepting and extending agent behavior without modifying core code. Register async handlers against named HookEvent values (PRE_LLM_CALL, POST_LLM_CALL, POST_TOOL_EXECUTE, ON_DELEGATION, ON_COMPACTION, …). Built-in hooks cover credential scrubbing, query-based model routing, bail-out retry (follow-through), and automatic memory injection. Hooks that raise never crash the agent — errors are logged and swallowed.
from sage.hooks.registry import HookRegistry
from sage.hooks.base import HookEvent
hr = HookRegistry()
async def log_calls(event, data):
print(f"{event}: {data.get('model')}")
hr.register(HookEvent.PRE_LLM_CALL, log_calls)
agent = Agent(name="a", model="gpt-4o", hook_registry=hr)
Coordination
Agent-to-agent messaging and lifecycle primitives for multi-agent systems:
- MessageBus — in-memory per-agent inboxes with TTL expiry, idempotency, overflow protection, and broadcast delivery
- CancellationScope — propagate cancel signals across async tasks; child scopes inherit parent cancellation
- SessionManager — create, track, and destroy concurrent agent sessions with typed metadata
Context Management
Token-aware context window management. Automatic compaction when approaching the model's limit — tries LLM summarization first, then emergency drop, then deterministic trim as a guaranteed last resort. Configurable reserve tokens and optional pruning of large tool outputs.
TUI
A full interactive terminal UI built with Textual. Split-screen layout — chat on the left, live tool-call feed on the right, status bar at the bottom. It's actually nice to use.
sage tui --agent-config AGENTS.md
Protocol-Based Architecture
ProviderProtocol, MemoryProtocol, EmbeddingProtocol — swap out any layer. Don't like the SQLite memory backend? Write your own. Don't want litellm? Implement the protocol. Everything is async-first.
Quick Start
pip install sage-agent
# or
uv tool install sage-agent
export OPENAI_API_KEY=sk-...
sage agent run AGENTS.md --input "What is the capital of France?"
Code API
import asyncio
from sage import Agent
agent = Agent(
name="assistant",
model="gpt-4o",
body="You are a helpful assistant.",
)
result = asyncio.run(agent.run("What is 2 + 2?"))
print(result)
Or load from config:
agent = Agent.from_config("AGENTS.md")
result = asyncio.run(agent.run("Hello"))
Pipelines
pipeline = researcher >> summarizer
result = asyncio.run(pipeline.run("Explain quantum computing"))
Parallel Execution
from sage import Orchestrator
results = asyncio.run(Orchestrator.run_parallel(agents, "Analyze this topic"))
Race Execution
winner = asyncio.run(Orchestrator.run_race(agents, "Solve this problem"))
Autonomous Orchestration
---
name: orchestrator
model: gpt-4o
subagents:
- research_agent
- summarize_agent
---
You are an orchestrator. Use the delegate tool to assign tasks to your subagents.
sage agent run orchestrator/AGENTS.md --input "Research and summarize quantum computing"
CLI
sage agent run AGENTS.md --input "Hello" [--stream] # Run an agent
sage agent validate AGENTS.md # Validate config
sage agent list [directory] # List agent configs
sage agent orchestrate AGENTS.md --input "text" # Run subagents in parallel
sage tool list AGENTS.md # List available tools
sage init [--name my-agent] [--model gpt-4o] # Scaffold a new project
sage tui --agent-config AGENTS.md # Launch interactive TUI
sage exec AGENTS.md -i "Hello" [-o text|jsonl|quiet] [--timeout N] [--yes] # Run headless (CI/scripting)
sage eval run suite.yaml [--min-pass-rate 0.9] [--runs N] # Run evaluation suite
sage eval validate suite.yaml # Validate suite file
sage eval history [--suite NAME] [--last N] # Show run history
sage eval compare <run-id-1> <run-id-2> # Compare two runs
sage eval list [directory] # Find suite files
Configuration Reference
Agent Config (Markdown Frontmatter)
---
name: my-agent
model: gpt-4o
description: "A helpful assistant" # Display only, NOT sent to model
max_turns: 10
# Tool access: permission categories drive tool registration
# Categories: read, edit, shell, web, memory, task
# Values: "allow" | "deny" | "ask" | {pattern: action, ...}
permission:
read: allow
edit: allow
shell:
"*": ask
"git log*": allow
"git diff*": allow
web: allow
# Custom tool modules (in addition to permission-derived built-ins)
extensions:
- myapp.tools # Your own tools (module path)
memory:
backend: sqlite # "sqlite" (default) or "file"
path: memory.db
embedding: text-embedding-3-large
compaction_threshold: 50
auto_load: false # Auto-inject recalled memories pre-LLM-call
auto_load_top_k: 5 # How many memories to inject
subagents:
- research_agent # Directory containing AGENTS.md
- config: helper.md # Reference another .md file
- name: inline-helper # Or define inline
model: gpt-4o-mini
mcp_servers:
- transport: stdio
command: npx
args: ["-y", "@modelcontextprotocol/server-filesystem", "/tmp"]
- transport: sse
url: http://localhost:8080/sse
context:
compaction_threshold: 0.75 # Compact at 75% of context window
reserve_tokens: 4096
prune_tool_outputs: true
tool_output_max_chars: 5000
model_params:
temperature: 0.7
max_tokens: 2048
# Hook-driven features (all optional)
credential_scrubbing:
enabled: true
patterns: ["sk-.*", "Bearer .*"]
allowlist: ["sk-test"]
query_classification:
rules:
- keywords: ["python", "code"]
patterns: []
priority: 1
target_model: gpt-4o
follow_through:
enabled: true
patterns: ["I cannot", "I'm unable"]
research:
enabled: true
max_sources: 5
timeout: 15.0
session:
enabled: true
---
You are a helpful AI assistant.
Main Config (TOML)
Sage supports a global TOML config file for defaults and per-agent overrides. It's auto-discovered at ./config.toml or ~/.config/sage/config.toml, or set via SAGE_CONFIG_PATH.
# Optional: global skills directory (waterfall: $cwd/skills → ~/.agents/skills → ~/.claude/skills)
# skills_dir = "/path/to/skills"
[defaults]
model = "gpt-4o"
max_turns = 15
[agents.my-agent]
model = "gpt-4o-mini"
max_turns = 5
# Optional: limit this agent to a subset of the global skill pool
# skills = ["git-master", "terraform"]
Override priority: main config defaults < per-agent overrides < frontmatter.
Architecture
sage/
agent.py # Core Agent class (run loop, delegation, hook emission)
config.py # Markdown frontmatter loading (Pydantic)
models.py # Message, ToolCall, ToolSchema, Usage, etc.
exceptions.py # SageError, ConfigError, ProviderError, ToolError
frontmatter.py # YAML frontmatter parser
main_config.py # TOML main config support
research.py # Pre-response research system
providers/ # ProviderProtocol + LiteLLMProvider
tools/ # @tool decorator, ToolRegistry, ToolDispatcher, builtins
skills/ # Skill loader (markdown-based reusable capabilities)
orchestrator/ # Orchestrator (parallel, race) + Pipeline (>>)
memory/ # MemoryProtocol, SQLiteMemory, FileMemory, compaction
hooks/ # HookRegistry, HookEvent, built-in hooks
coordination/ # MessageBus, CancellationScope, SessionManager
parsing/ # Multi-format tool call parser, JSON repair
mcp/ # MCPClient + MCPServer
permissions/ # PermissionProtocol, policy rules, interactive prompts
context/ # Token-aware context budget, fallback table
git/ # GitSnapshot (snapshot/restore capability)
cli/ # Click CLI commands + Textual TUI
main.py # sage agent / exec / eval / tool / init / tui commands
tui.py # Textual interactive TUI
exit_codes.py # SageExitCode IntEnum (exit codes 0–7)
output.py # OutputWriter — TextWriter, JSONLWriter, QuietWriter
eval/ # Built-in evaluation framework
suite.py # TestSuite, TestCase, EvalSettings, load_suite()
assertions.py # 11 assertion types + run_assertion()
runner.py # EvalRunner, CaseResult, EvalRunResult
history.py # EvalHistory — SQLite run history (~/.config/sage/eval_history.db)
report.py # Text/JSON/comparison formatters
Examples
examples/simple_agent/— Minimal agent with markdown configexamples/custom_tools/— Agent with@tool-decorated functionsexamples/parallel_agents/— Orchestrator with subagentsexamples/mcp_agent/— Agent with MCP filesystem serverexamples/memory_agent/— Semantic memory backend usageexamples/skills_agent/— Skills in actionexamples/skills_demo/— Complex skills demoexamples/permissions_agent/— Permission policiesexamples/safe_coder/— Code generation with safetyexamples/devtools_agent/— Developer toolsexamples/claude_agent/— Anthropic Claude model
Claude Code Skills
This repo includes Claude Code skills that help you create and optimize Sage agents through a guided, conversational workflow. The skills live in the skills/ directory.
Prerequisites
- Claude Code installed
sage-agentinstalled — theevaluate-sage-agentskill usessage evalwhich is built in
Available Skills
create-sage-agent — Create a new agent from a description
Walks you through generating a complete AGENTS.md from a natural language description. It infers permissions, model settings, and system prompt structure, and identifies opportunities for subagent decomposition.
Invoke it in Claude Code:
/skill create-sage-agent
What it does:
- Asks what the agent should do
- Infers permissions, model, and complexity from your description
- Identifies subagent opportunities (e.g., pipeline stages, distinct roles)
- Generates a complete
AGENTS.mdwith frontmatter config and system prompt - Validates the config
- Optionally hands off to
evaluate-sage-agentfor optimization
Example interaction:
You: /skill create-sage-agent
Claude: What should this agent do?
You: A code reviewer that reads files, checks git history, and flags security issues
Claude: [generates AGENTS.md with read + shell permissions, appropriate system prompt]
evaluate-sage-agent — Validate, benchmark, and optimize an agent
Runs a structured evaluation pipeline on an existing agent config: validation, suggestion generation, model benchmarking, and before/after comparison.
Invoke it in Claude Code:
/skill evaluate-sage-agent
What it does:
- Validate — Checks the agent config for errors
- Suggest — Analyzes the config and recommends prompt improvements, tool extractions, guardrails, and architectural changes
- Apply — Creates a versioned backup (
AGENTS.v1.md, etc.) and applies accepted suggestions - Benchmark — Tests the agent against one or more models, reporting quality, latency, token usage, and cost
- Compare — Runs a before/after comparison if changes were applied, showing deltas in each metric
Example interaction:
You: /skill evaluate-sage-agent
Claude: Which agent would you like to evaluate?
You: ./code-reviewer
Claude: [validates, suggests improvements, benchmarks, shows results]
Typical Workflow
create-sage-agent → evaluate-sage-agent
(describe it) (optimize it)
- Use
create-sage-agentto generate a new agent from a description - Use
evaluate-sage-agentto validate and optimize it - Iterate — the evaluate skill versions your config so you can compare changes
Requirements
- Python 3.10+
- See
pyproject.tomlfor full dependency list
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file sage_agent-1.2.0rc1.tar.gz.
File metadata
- Download URL: sage_agent-1.2.0rc1.tar.gz
- Upload date:
- Size: 111.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
ab848374395f2de3b6fc1d7ee982e8b9a5ca95508827c1f57bfa9efb9a2aaeb9
|
|
| MD5 |
b1594c924d94c6a8eced86cc148cf24d
|
|
| BLAKE2b-256 |
21943b0970e8136a15d86fb04ab787f3ec8db0ed44b7f1d2a772354897deb673
|
Provenance
The following attestation bundles were made for sage_agent-1.2.0rc1.tar.gz:
Publisher:
release.yml on sagebynature/sage-agent
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
sage_agent-1.2.0rc1.tar.gz -
Subject digest:
ab848374395f2de3b6fc1d7ee982e8b9a5ca95508827c1f57bfa9efb9a2aaeb9 - Sigstore transparency entry: 1007137308
- Sigstore integration time:
-
Permalink:
sagebynature/sage-agent@d489579e935ec3348bf2633682fb55363bc04a62 -
Branch / Tag:
refs/heads/dev - Owner: https://github.com/sagebynature
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@d489579e935ec3348bf2633682fb55363bc04a62 -
Trigger Event:
push
-
Statement type:
File details
Details for the file sage_agent-1.2.0rc1-py3-none-any.whl.
File metadata
- Download URL: sage_agent-1.2.0rc1-py3-none-any.whl
- Upload date:
- Size: 141.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
173d7f8fb6f9ec41ff24b38cce15b9a220c9fc517dfce9c6148d58dbdfd838a5
|
|
| MD5 |
6bf5146ad8dbc08938e8f20d1d08639f
|
|
| BLAKE2b-256 |
1ef7d5bb18165ea53cd529ff29cf155d28704f7e3a56503448eac108178a4499
|
Provenance
The following attestation bundles were made for sage_agent-1.2.0rc1-py3-none-any.whl:
Publisher:
release.yml on sagebynature/sage-agent
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
sage_agent-1.2.0rc1-py3-none-any.whl -
Subject digest:
173d7f8fb6f9ec41ff24b38cce15b9a220c9fc517dfce9c6148d58dbdfd838a5 - Sigstore transparency entry: 1007137322
- Sigstore integration time:
-
Permalink:
sagebynature/sage-agent@d489579e935ec3348bf2633682fb55363bc04a62 -
Branch / Tag:
refs/heads/dev - Owner: https://github.com/sagebynature
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@d489579e935ec3348bf2633682fb55363bc04a62 -
Trigger Event:
push
-
Statement type: