Skip to main content

An easy to use framework for adding tool use to AI agents.

Project description

AI Agent Toolbox

AI Agent Toolbox Logo

AI Agent Toolbox makes AI tool usage across models and frameworks easy. For use with parsing single responses or in agent loops and workflow.

AI Agent Toolbox is meant to be stable, reliable, and easy to master.

18x faster than naive streaming parsers at scale - True O(n) streaming parser, optimized for RL and agent loops. See benchmarks

Features

  • Model provider-agnostic - supports Anthropic, OpenAI, Groq, NEAR AI, Ollama, Hyperbolic, NanoGPT, and more
  • Framework compatible - usable in anthropic-sdk-python, openai-python, ell, LangChain, etc
  • Supports protocols such as Anthropic Model Context Protocol(MCP)
  • Robust XML parsing
  • Streaming support
  • Support for read-write tools (feedback) as well as write-only tools

Documentation

See the full documentation at toolbox.255labs.xyz

Installation

pip install ai-agent-toolbox

Examples

See our examples folder for:

  • Simple tool usage
  • Streaming integration
  • Read-write tools with feedback loops
  • Agent workflow examples

Usage

Synchronous (Complete Response)

To parse a fully formed response:

from ai_agent_toolbox import Toolbox, XMLParser, XMLPromptFormatter
from examples.util import anthropic_llm_call

# Setup
toolbox = Toolbox()
parser = XMLParser(tag="use_tool")
formatter = XMLPromptFormatter(tag="use_tool")

# Add tools to your toolbox
def thinking(thoughts=""):
    print("I'm thinking:", thoughts)

toolbox.add_tool(
    name="thinking",
    fn=thinking,
    args={
        "thoughts": {
            "type": "string",
            "description": "Anything you want to think about"
        }
    },
    description="For thinking out loud"
)

system = "You are a thinking AI. You have interesting thoughts.\n"
prompt = "Think about something interesting."

# Add instructions on using the available tools to the AI system prompt
system += formatter.usage_prompt(toolbox)

response = anthropic_llm_call(system_prompt=system, prompt=prompt)
events = parser.parse(response)

for event in events:
    toolbox.use(event)

Structured arguments and validation

Tool schemas can describe composite inputs (lists, dicts, enums) and enforce validation constraints. The toolbox automatically parses JSON strings emitted by models and applies optional custom parsers before validation.

import json
from dataclasses import dataclass

from ai_agent_toolbox import Toolbox
from ai_agent_toolbox.parser_event import ParserEvent
from ai_agent_toolbox.tool_use import ToolUse


@dataclass
class Task:
    title: str
    estimate_hours: int


def pick_next_task(tasks, metadata, priority, limit):
    ranked = sorted(tasks, key=lambda task: task.estimate_hours)[:limit]
    return {
        "next_task": ranked[0].title,
        "tasks_considered": [task.title for task in ranked],
        "metadata": metadata,
        "priority": priority,
    }


toolbox = Toolbox()
toolbox.add_tool(
    name="pick_next_task",
    fn=pick_next_task,
    args={
        "tasks": {
            "type": "list",
            "parser": lambda payload: [Task(**task) for task in payload],
        },
        "metadata": {"type": "dict"},
        "priority": {"type": "enum", "choices": ["low", "medium", "high"]},
        "limit": {"type": "int", "min": 1, "max": 5},
    },
    description="Choose the next task to execute",
)

event = ParserEvent(
    type="tool",
    mode="close",
    id="tasks-1",
    tool=ToolUse(
        name="pick_next_task",
        args={
            "tasks": json.dumps(
                [
                    {"title": "Write docs", "estimate_hours": 2},
                    {"title": "Ship release", "estimate_hours": 1},
                ]
            ),
            "metadata": json.dumps({"owner": "core-team"}),
            "priority": "high",
            "limit": "2",
        },
    ),
    is_tool_call=True,
)

response = toolbox.use(event)
print(response.result)
# {
#   'tasks': [Task(title='Write docs', estimate_hours=2), ...],
#   'metadata': {'owner': 'core-team'},
#   'priority': 'high',
#   'limit': 2,
#   'next_task': 'Ship release',
#   'tasks_considered': ['Ship release', 'Write docs']
# }

Asynchronous (Streaming)

If you want to parse LLM responses as they come in:

import asyncio
from ai_agent_toolbox import Toolbox, XMLParser, XMLPromptFormatter
from examples.util import anthropic_stream

async def main():
    # Initialize components
    toolbox = Toolbox()
    parser = XMLParser(tag="tool")
    formatter = XMLPromptFormatter(tag="tool")
    
    # Register tools (add your actual tools here)
    toolbox.add_tool(
        name="search",
        fn=lambda query: f"Results for {query}",
        args={"query": {"type": "string"}},
        description="Web search tool"
    )
    # Set up the system and user prompt
    system = "You are a search agent.\n"

    # Add tool usage instructions
    system += formatter.usage_prompt(toolbox)
    prompt = "Search for ..."

    # Simulated streaming response from LLM
    async for chunk in anthropic_stream(system=system, prompt=prompt, ...):
        # Parse each chunk as it arrives
        for event in parser.parse_chunk(chunk):
            if event.is_tool_call:
                print(f"Executing tool: {event.tool.name}")
                await toolbox.use_async(event)  # Handle async tools

    # Call this at the end of output to handle any unclosed or invalid LLM outputs
    for event in parser.flush():
        if event.is_tool_call:
            print(f"Executing tool: {event.tool.name}")
            await toolbox.use_async(event)  # Handle async tools

if __name__ == "__main__":
    asyncio.run(main())

Local Tooling

This supports other providers and open source models. Here you parse the results yourself.

Retrieval and read-write tools

Some tools, such as search, may want to give the AI information for further action. This involves crafting a new prompt to the LLM includes any tool responses. We make this easy.

from ai_agent_toolbox import ToolResponse

def search_tool(query: str):
    return f"Results for {query}"

# In your agent loop:
responses = [r.result for r in tool_responses if r.result]
new_prompt = f"Previous tool outputs:\n{'\n'.join(responses)}\n{original_prompt}"

# Execute next LLM call with enriched prompt
next_response = llm_call(
    system_prompt=system,
    prompt=new_prompt
)
# Continue processing...

Running Tests

Streaming parsers in this project are validated against golden event traces so refactors keep their streaming semantics intact. Before opening a pull request, run the parser regression suite:

pytest tests/**/*.py

You can also execute the entire test suite with pytest tests to include any additional checks.

Benefits

Feature/Capability AI Agent Toolbox ✅ Naive Regex ❌ Standard XML Parsers ❌
Streaming Support ✅ Chunk-by-chunk processing ❌ Full text required ❌ DOM-based requires full document
Nested HTML/React ✅ Handles JSX-like fragments ❌ Fails on nesting ❌ Requires strict hierarchy
Flexible Tool Format ✅ XML tool format with streaming ❌ Brittle pattern matching ❌ Requires schema validation
Automatic Type Conversion ✅ String→int/float/bool ❌ Manual casting needed ❌ Returns only strings
Error Recovery ✅ Heals partial/malformed tags ❌ Fails on first mismatch ❌ Aborts on validation errors
Battle Tested ✅ Heavily tested ❌ Ad-hoc testing ❌ Generic XML cases only
Tool Schema Enforcement ✅ Args + types validation ❌ No validation ❌ Only structural validation
Mixed Content Handling ✅ Text + tools interleaved ❌ Captures block text ❌ Text nodes require special handling
Async Ready ✅ Native async/sync support ❌ Callback hell ❌ Sync-only typically
Memory Safety ✅ Guardrails against OOM ❌ Unbounded buffers ❌ DOM explosion risk
LLM Output Optimized ✅ Tolerates unclosed tags ❌ Fails on partials ❌ Strict tag matching
Tool Feedback Loops ✅ ToolResponse chaining ❌ Manual stitching ❌ No built-in concept

Agent loops

Workflows and agent loops involve multiple calls to a LLM provider.

Tips

  • Keep the system prompt the same across invocations when using multiple LLM calls
  • Stream when necessary, for example when a user is waiting for output
  • Native provider tooling can be used with local parsing
  • Start simple and expand. You can test with static strings to ensure your tools are working correctly.

Ecosystem

Used by

License

MIT

Credits

AI Agent Toolbox is created and maintained by 255labs.xyz

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ai_agent_toolbox-1.0.3.tar.gz (21.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

ai_agent_toolbox-1.0.3-py3-none-any.whl (19.4 kB view details)

Uploaded Python 3

File details

Details for the file ai_agent_toolbox-1.0.3.tar.gz.

File metadata

  • Download URL: ai_agent_toolbox-1.0.3.tar.gz
  • Upload date:
  • Size: 21.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.2

File hashes

Hashes for ai_agent_toolbox-1.0.3.tar.gz
Algorithm Hash digest
SHA256 9da216b7c33ed35a708fccfdce755003d9a020007111890ad5b53e370ba3f682
MD5 021af26ca61fd489837ecf98b85aac13
BLAKE2b-256 cfe25b56ffbc2c7119abbe32dad02d959319016eece53a9befef13c3dd775a7c

See more details on using hashes here.

File details

Details for the file ai_agent_toolbox-1.0.3-py3-none-any.whl.

File metadata

File hashes

Hashes for ai_agent_toolbox-1.0.3-py3-none-any.whl
Algorithm Hash digest
SHA256 775593c7738bf7e4099f1075404f4837cc91280bbeea486c4e297261f97ad274
MD5 94675654fbf7085456c1c265a8e40d7e
BLAKE2b-256 f697ca12c1ca2a54dfab53dc5e7f3204c64c22b1b261fad3004a985698457ebb

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page