Skip to main content

A minimal, universal agent framework. Zero mandatory dependencies.

Project description

English | 中文

all-in-agents

A minimal, universal agent framework for Python. Zero mandatory dependencies.

PyPI version Python versions License GitHub Stars

pip install all-in-agents
pip install "all-in-agents[openai]"      # OpenAI GPT
pip install "all-in-agents[anthropic]"   # Anthropic Claude
pip install "all-in-agents[all]"         # all optional deps

Why all-in-agents

  • 🪶 Zero dependencies — pure stdlib core; adapters are opt-in extras
  • 🔌 Pluggable everything — swap LLM adapter, tools, history, or orchestration without touching other parts
  • 🔍 Transparent by default — append-only NDJSON event log; every run is replayable
  • 🛡️ Safe by default — dangerous tools require explicit approval; budget stops runaway agents

Quick Start

pip install "all-in-agents[openai]"      # or [anthropic]
from all_in_agents import Agent

agent = Agent.quick(model="gpt-4o", workspace=".")
result = agent.run_sync("Summarize README.md in three bullet points")
print(result.final_answer)

Or with full control:

from all_in_agents import Agent, OpenAIAdapter, ToolRegistry, BUILTIN_TOOLS, unsafe_defaults

llm = OpenAIAdapter(model="gpt-4o")     # reads OPENAI_API_KEY from env
tools = ToolRegistry(approval_callback=unsafe_defaults())
for t in BUILTIN_TOOLS:                  # read_file, write_file, bash, list_files, text_search
    tools.register(t)

agent = Agent(llm=llm, tools=tools, workspace_root=".")
result = agent.run_sync("Summarize README.md in three bullet points")
print(result.final_answer)

Jupyter Notebook or async framework? Use await agent.run(goal) directly.

CLI

# Single-shot
python -m all_in_agents "Summarize README.md" --model gpt-4o --unsafe

# Interactive REPL
python -m all_in_agents --model gpt-4o --unsafe

Core Concepts

Node / Flow

Everything is a node. A flow is a graph of nodes.

from all_in_agents import BaseNode, Flow

class MyNode(BaseNode):
    async def prep(self, shared: dict):
        return shared["input"]

    async def exec(self, prep_result):
        return prep_result.upper()

    async def post(self, shared: dict, exec_result) -> str:
        shared["output"] = exec_result
        return "default"   # action name → next node

node_a = MyNode()
node_b = MyNode()
node_a >> node_b           # default edge
# or: (node_a - "custom_action") >> node_b

flow = Flow()
await flow.run(shared={}, start=node_a)

State contract: all inter-node state lives in shared dict. Node instance fields hold only configuration.

Budget & Loop Detection

from all_in_agents import Budget

budget = Budget(
    max_llm_calls=40,
    max_tool_calls=80,
    max_wall_ms=1_800_000,       # 30 min wall-clock limit
    loop_same_action_limit=3,    # raise LoopDetectedError after 3 consecutive identical tool calls
)

agent = Agent(llm=llm, tools=tools, budget=budget)

Artifact Contracts

Use artifact contracts when a run must produce machine-checkable outputs. The agent can still work freely, but the framework marks the run incomplete if required artifacts are missing or invalid.

from all_in_agents import Agent, ArtifactContract

contract = ArtifactContract.files("research_plan.md", "observation.md")

agent = Agent.quick(
    model="gpt-4o",
    workspace=".",
    artifact_contract=contract,
)
result = agent.run_sync("Create the required research artifacts")

assert result.status == "success"

JSON artifacts can be schema-checked when the jsonschema extra is installed:

contract = ArtifactContract.json_files({
    "metrics.json": {
        "type": "object",
        "required": ["score"],
        "properties": {"score": {"type": "number"}},
    }
})

Tool Registry

from all_in_agents import Tool, ToolRegistry, SideEffectLevel, ToolResponse

async def my_tool(args: dict, run) -> ToolResponse:
    result = do_something(args["input"])
    return ToolResponse(status="success", content=result)

registry = ToolRegistry(
    approval_callback=my_approval_fn   # async (name, args) -> bool
)
registry.register(Tool(
    name="my_tool",
    description="Does something useful",
    input_schema={
        "type": "object",
        "properties": {"input": {"type": "string"}},
        "required": ["input"],
    },
    side_effect_level=SideEffectLevel.READ_ONLY,
    execute=my_tool,
))

DANGEROUS and WRITES_LOCAL tools call approval_callback before executing. By default, the callback denies all requests (safe by default). Use unsafe_defaults() for development or provide your own callback. Install jsonschema for automatic argument validation with type coercion.

Skills

Project skills are prompt bundles stored as SKILL.md files:

skills/
  reviewer/
    SKILL.md
.skills/
  local-debug/
    SKILL.md

Load selected skills by name:

agent = Agent.quick(
    model="gpt-4o",
    workspace=".",
    skills=["reviewer"],
)

Or load every discovered skill:

agent = Agent.quick(model="gpt-4o", workspace=".", skills="all")

CLI usage:

python -m all_in_agents --skill reviewer "Review this code"
python -m all_in_agents --all-skills "Use the relevant project skill"
python -m all_in_agents --project-context "Follow AGENTS.md and project context"

Hidden .skills/ entries take precedence over skills/ entries with the same name. Skills are injected into the system prompt; they do not automatically register Python tools.

History & Compression

HistoryManager compresses conversation history when it exceeds a soft threshold. By default, that threshold is 70% of the model's context window; override it with history_compress_threshold_tokens on Agent or Agent.quick. The built-in compactor targets that same soft threshold, keeps recent turns verbatim, summarizes older turns into structured JSON (facts / decisions / open_threads), and falls back to deterministic snipping if summarization fails.

agent = Agent.quick(
    model="gpt-4o",
    history_compress_threshold_tokens=18_000,
)

Custom compaction strategies can implement compact_turns(llm, turns, *, max_context_tokens, target_tokens=None) and return CompactionResult.

Event Store

Every run writes an append-only NDJSON log to ./runs/<run_id>/events.ndjson:

{"event_id": "...", "run_id": "...", "ts": "...", "type": "RUN_CREATED", "payload": {...}}
{"event_id": "...", "run_id": "...", "ts": "...", "type": "ASSISTANT_MESSAGE", "payload": {...}}
{"event_id": "...", "run_id": "...", "ts": "...", "type": "TOOL_RESULT", "payload": {...}}
{"event_id": "...", "run_id": "...", "ts": "...", "type": "RUN_STOPPED", "payload": {"reason": "goal_met"}}

Multi-Agent

from all_in_agents import MessageBus, TaskManager, MessageEnvelope, Task

bus = MessageBus(run_dir="./runs/session_1")
tm  = TaskManager(run_dir="./runs/session_1")

# coordinator creates tasks
task = await tm.create_task(goal="Analyze file X")

# worker claims and runs
available = await tm.get_available(agent_id="worker_1")
claimed   = await tm.claim_task(available[0].task_id, "worker_1")

# agents communicate
await bus.send(MessageEnvelope(
    msg_id="...", run_id="...",
    from_agent="worker_1", to_agent="coordinator",
    msg_type="TASK_DONE", payload={"result": "..."}, ts="...",
))

TaskManager uses file-based locking (fcntl on Unix, .lock file on Windows) for safe concurrent access. Tasks support dependency chains via dependencies: list[str].

LLM Adapters

Adapter Install extra Environment variable
OpenAIAdapter all-in-agents[openai] OPENAI_API_KEY
AnthropicAdapter all-in-agents[anthropic] ANTHROPIC_API_KEY

Both adapters classify errors (TRANSIENT, RATE_LIMITED, AUTH, INVALID_REQUEST, INTERNAL) and retry with exponential backoff. Rate-limited requests honor retry-after headers when available.

from all_in_agents import OpenAIAdapter, AnthropicAdapter

llm = OpenAIAdapter(model="gpt-4o-mini", max_retries=3)
llm = AnthropicAdapter(model="claude-sonnet-4-6", max_retries=3)

Architecture

📁 Directory Structure
all_in_agents/
├── cli.py       Lightweight CLI runner
├── core/
│   ├── node.py      BaseNode · Node · BatchNode
│   ├── flow.py      Flow (graph runner, auto-retry via exec_with_retry)
│   └── run.py       Run · RunResult · Budget · BudgetExceededError · LoopDetectedError
├── adapters/
│   ├── base.py      LLMAdapter · LLMResponse · ToolCall · LLMError · ErrorClass
│   ├── anthropic.py AnthropicAdapter (error classification, prompt caching)
│   └── openai.py    OpenAIAdapter (error classification, rate-limit tracking)
├── tools/
│   ├── registry.py  ToolRegistry (safe-by-default, approval callbacks, jsonschema)
│   ├── policy.py    ToolPolicy · SideEffectLevel
│   ├── coerce.py    Schema-driven argument type coercion
│   └── builtin.py   read_file · write_file · bash · list_files · text_search
├── history/
│   ├── manager.py   HistoryManager (dynamic threshold, LLM-based compression)
│   ├── compactor.py HistoryCompactor (micro-compact + summarize + fallback)
│   └── store.py     FileEventStore (append-only NDJSON, event callbacks)
└── agents/
    ├── base.py      Agent · Agent.quick() · LLMCallNode · ToolDispatchNode
    ├── harness.py   AGENTS.md / .context/ project context loader
    └── multi.py     MessageBus · TaskManager · MessageEnvelope · Task · TaskStatus

Package Naming

The PyPI package is all-in-agents, but the Python import name is all_in_agents:

pip install all-in-agents
from all_in_agents import Agent   # Python import name is 'all_in_agents'

The hyphen in the PyPI name can't be used in Python imports, so the module name uses underscores.

Design Goals

  • Zero mandatory deps — pure stdlib core; adapters opt-in
  • Small — ~120 LOC core loop, readable in one sitting
  • Composable — every piece (Node, Tool, Adapter, History) is replaceable
  • Safe by default — dangerous tools require approval; budget stops runaway agents

Requirements

Python 3.10+

Optional: anthropic, openai, jsonschema

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

all_in_agents-0.2.5.tar.gz (44.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

all_in_agents-0.2.5-py3-none-any.whl (48.5 kB view details)

Uploaded Python 3

File details

Details for the file all_in_agents-0.2.5.tar.gz.

File metadata

  • Download URL: all_in_agents-0.2.5.tar.gz
  • Upload date:
  • Size: 44.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for all_in_agents-0.2.5.tar.gz
Algorithm Hash digest
SHA256 c257fcfa13baaf6fa2a928df2f8a5396ac15774e09e15845cbbe91fa581ed7a3
MD5 7f605b236e88bedc62ab20001090d18d
BLAKE2b-256 7f0bc4e79615c64db29ce7c8534ee3d362bc2c7da082f5dd1e61959eef78a57a

See more details on using hashes here.

File details

Details for the file all_in_agents-0.2.5-py3-none-any.whl.

File metadata

  • Download URL: all_in_agents-0.2.5-py3-none-any.whl
  • Upload date:
  • Size: 48.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for all_in_agents-0.2.5-py3-none-any.whl
Algorithm Hash digest
SHA256 b2d6fb7272ccd7d20643b37c8a34771f6d98e37144b841f3951feb660e92e15b
MD5 85b5650aea8515e78cc4c64e8c50fb67
BLAKE2b-256 0f96c2d6bfdc7014837158caed5f87239610db93079f84bec34da232748b2b69

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page