Skip to main content

Compliance, governance & observability layer for AI agents

Project description

๐Ÿ›ก๏ธ AgentGuard

The governance layer that lets companies trust their AI agents enough to actually deploy them.

Python 3.10+ MIT License PyPI Tests


62% of production AI teams plan to improve observability in the next year. Over 40% of agentic AI projects will be canceled by 2027 due to inadequate risk controls. Humans still verify 69% of AI decisions because there are no guardrails they trust.

AgentGuard fixes this. One SDK. Full audit trail. Every LLM call and tool use โ€” intercepted, policy-checked, cost-tracked, and logged. 3 lines of code.


Quick Start

from openai import OpenAI
from agentguard import AgentGuard

client = OpenAI()

guard = AgentGuard(
    policies=["pii", "content_filter", "cost_limit"],
    audit_path="audit.jsonl",
    cost_limit=5.00,
)

safe_client = guard.wrap_openai(client)

# Use exactly like the original โ€” now with full protection
response = safe_client.chat.completions.create(
    model="gpt-4o",
    messages=[{"role": "user", "content": "Hello!"}],
)

Every call is now:

  • โœ… PII-scanned โ€” blocks emails, SSNs, credit cards, phone numbers
  • โœ… Policy-checked โ€” blocks prompt injections, enforces budget limits
  • โœ… Cost-tracked โ€” per-model, per-run, and daily spend tracking
  • โœ… Audit-logged โ€” immutable JSON-lines trail for compliance

Installation

# Core (OpenAI support included)
pip install agentaudit-sdk

# With Anthropic support
pip install "agentaudit-sdk[anthropics]"

๐Ÿค– Anthropic Claude Integration

Wrap Claude exactly like OpenAI โ€” 3 lines, full protection.

import anthropic
from agentguard import AgentGuard

client = anthropic.Anthropic()

guard = AgentGuard(
    policies=["pii", "content_filter", "cost_limit"],
    audit_path="audit.jsonl",
    cost_limit=5.00,
)

safe = guard.wrap_anthropic(client)

# Use exactly like the original โ€” now fully protected
response = safe.messages.create(
    model="claude-3-5-sonnet-20241022",
    max_tokens=1024,
    messages=[{"role": "user", "content": "Hello, Claude!"}],
)
print(response.content[0].text)

Every call is now:

  • โœ… PII-scanned โ€” both messages list AND the top-level system prompt
  • โœ… Policy-checked โ€” prompt injections blocked, budget enforced
  • โœ… Cost-tracked โ€” accurate per-model pricing for all Claude 3 variants
  • โœ… Audit-logged โ€” immutable JSON-lines trail

Async Claude

import anthropic
from agentguard import AgentGuard

async with AgentGuard(policies=["pii", "content_filter"]) as guard:
    client = anthropic.AsyncAnthropic()
    safe = guard.wrap_anthropic_async(client)

    response = await safe.messages.create(
        model="claude-3-5-haiku-20241022",
        max_tokens=512,
        system="You are a helpful assistant.",   # ๐Ÿ›ก๏ธ system prompt is PII-scanned too
        messages=[{"role": "user", "content": "Summarise this report."}],
    )
    print(response.content[0].text)

Supported Claude Models (with built-in pricing)

Model Input / 1M tokens Output / 1M tokens
claude-3-5-sonnet-20241022 $3.00 $15.00
claude-3-5-haiku-20241022 $0.80 $4.00
claude-3-opus-20240229 $15.00 $75.00
claude-3-sonnet-20240229 $3.00 $15.00
claude-3-haiku-20240307 $0.25 $1.25

Features

๐Ÿ›ก๏ธ Built-in Policies

Policy What It Does
pii Blocks PII (emails, SSN, credit cards, phones, IPs) in inputs & outputs
content_filter Blocks prompt injection attempts & system prompt extraction
cost_limit Enforces per-run, daily, and total budget limits
rate_limit Throttles calls per time window (sliding window)
tool_restriction Blocklist/allowlist for agent tool usage

๐Ÿ”ง Tool Guarding

Wrap any function โ€” sync or async. Policies are enforced before the tool runs.

def delete_database(db_name: str) -> str:
    ...

safe_delete = guard.wrap_tool(delete_database)
safe_delete(db_name="production")  # ๐Ÿ›ก๏ธ Blocked by tool_restriction policy
# PII is caught in tool arguments too
def send_email(to: str, body: str) -> str:
    ...

safe_send = guard.wrap_tool(send_email)
safe_send(to="john@example.com", body="Hi")  # ๐Ÿ›ก๏ธ Blocked: PII detected

โšก Full Async Support

Works with AsyncOpenAI and async tool functions โ€” zero changes to your logic.

from openai import AsyncOpenAI

async with AgentGuard(policies=["pii", "content_filter"]) as guard:
    client = AsyncOpenAI()
    safe = guard.wrap_openai_async(client)

    response = await safe.chat.completions.create(
        model="gpt-4o",
        messages=[{"role": "user", "content": "Hello!"}],
    )

    # Async tools โ€” auto-detected
    async def fetch_data(url: str) -> str:
        ...

    safe_fetch = guard.wrap_tool(fetch_data)  # auto-detects async
    result = await safe_fetch(url="https://api.example.com")

๐Ÿ’ฐ Cost Tracking

Real-time spend tracking with per-model pricing for GPT-4o, GPT-4o-mini, Claude, Gemini, o1, o3-mini, and more.

report = guard.get_report()
# {
#     'total_cost_usd': 0.0234,
#     'total_tokens_in': 1500,
#     'total_tokens_out': 800,
#     'daily_cost_usd': 0.0234,
#     'run_cost_usd': 0.0120,
#     'policies_active': ['pii', 'content_filter', 'cost_limit']
# }

๐ŸŽฌ Audit Reader & Replay

The killer feature. Prove exactly what your agent did, step by step.

Python API

from agentguard import AuditReader

reader = AuditReader("audit.jsonl")
run = reader.get_run("run_abc123")
run.print_trace()
โ•”โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•—
โ•‘  AGENTGUARD RUN TRACE                                   โ•‘
โ• โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•ฃ
โ•‘  Run ID:     run_abc123                                  โ•‘
โ•‘  Events:     3                                           โ•‘
โ•‘  Tokens:     1,500 in / 800 out                          โ•‘
โ•‘  Cost:       $0.0234                                     โ•‘
โ•‘  Violations: None                                        โ•‘
โ• โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•ฃ
โ•‘  Step 1  LLM  OK
โ•‘    Model:  gpt-4o
โ•‘    user: "Process the customer refund"
โ•‘    assistant: "I'll process that refund now."
โ•‘
โ•‘  Step 2  TOOL  OK
โ•‘    Tool:     process_refund
โ•‘    Args:     {"order_id": "ORD-12345", "amount": 49.99}
โ•‘    Duration: 230ms
โ•‘
โ•‘  Step 3  LLM  OK
โ•‘    Model:  gpt-4o
โ•‘    assistant: "The refund of $49.99 has been processed."
โ•šโ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•

CLI Tool

# List all runs with summary stats
agentguard --file audit.jsonl runs

# Step-by-step replay of any run
agentguard --file audit.jsonl replay <run_id>
agentguard --file audit.jsonl replay <run_id> --delay 0.5  # slow replay

# Show all policy violations (audit-ready)
agentguard --file audit.jsonl violations

# Dashboard โ€” costs, models, tools, violations
agentguard --file audit.jsonl stats

# Search events by any content
agentguard --file audit.jsonl search "delete_database"

# Export for compliance reports
agentguard --file audit.jsonl export --format json -o report.json
agentguard --file audit.jsonl export --format csv -o audit.csv

# Live tail โ€” watch events in real-time
agentguard --file audit.jsonl tail

๐Ÿ”Œ Custom Policies

Build your own โ€” just subclass Policy and implement evaluate().

from agentguard import Policy, PolicyResult, PolicyAction
from agentguard.core.events import LLMCallEvent

class NoProfanityPolicy(Policy):
    name = "no_profanity"
    supported_events = [LLMCallEvent]

    def evaluate(self, event):
        bad_words = ["damn", "hell"]
        content = str(event.messages).lower()
        if any(w in content for w in bad_words):
            return PolicyResult(
                action=PolicyAction.BLOCK,
                policy_name=self.name,
                reason="Profanity detected",
            )
        return PolicyResult(action=PolicyAction.ALLOW, policy_name=self.name)

guard = AgentGuard(policies=[NoProfanityPolicy(), "pii"])

๐Ÿ”” Human-in-the-Loop Escalation

def on_escalation(event):
    print(f"ALERT: {event.reason}")
    # Send to Slack, PagerDuty, email, etc.

guard = AgentGuard(
    policies=["pii", "content_filter"],
    on_escalation=on_escalation,  # supports async callbacks too
)

Run the Demo

python examples/basic_usage.py

Run Tests

pip install agentaudit-sdk[dev]
pytest tests/ -v
# 104 tests passing in <1 second

Architecture

Your App โ†’ AI Agent โ†’ ๐Ÿ›ก๏ธ AgentGuard SDK โ†’ Tool / LLM API
                            โ”‚
                     โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
                     โ”‚   Interceptor   โ”‚  โ† before/after hooks
                     โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
                     โ”‚  Policy Engine  โ”‚  โ† PII, Cost, Content, Rate, Tool
                     โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
                     โ”‚  PII Detector   โ”‚  โ† Regex (pluggable to ML/Presidio)
                     โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
                     โ”‚  Cost Tracker   โ”‚  โ† Per-model pricing, run/daily/total
                     โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
                     โ”‚  Audit Logger   โ”‚  โ† Thread-safe, JSON-lines, rotation
                     โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
                     โ”‚  Audit Reader   โ”‚  โ† Query, filter, replay, CLI
                     โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜
src/agentguard/
โ”œโ”€โ”€ core/
โ”‚   โ”œโ”€โ”€ events.py          # Pydantic event models (run_id grouping)
โ”‚   โ”œโ”€โ”€ interceptor.py     # Central before/after hooks
โ”‚   โ””โ”€โ”€ guard.py           # Main orchestrator (3-line API)
โ”œโ”€โ”€ policies/
โ”‚   โ”œโ”€โ”€ base.py            # Policy engine + event-type filtering
โ”‚   โ”œโ”€โ”€ pii_policy.py      # PII blocking
โ”‚   โ”œโ”€โ”€ cost_policy.py     # Budget enforcement
โ”‚   โ”œโ”€โ”€ tool_policy.py     # Tool blocklist/allowlist
โ”‚   โ”œโ”€โ”€ rate_limit_policy.py  # Sliding window rate limiter
โ”‚   โ””โ”€โ”€ content_policy.py  # Prompt injection detection
โ”œโ”€โ”€ detectors/
โ”‚   โ””โ”€โ”€ pii.py             # Regex PII detector (pluggable Protocol)
โ”œโ”€โ”€ tracking/
โ”‚   โ””โ”€โ”€ cost.py            # Token & cost tracking
โ”œโ”€โ”€ logging/
โ”‚   โ”œโ”€โ”€ audit.py           # Thread-safe JSON-lines logger
โ”‚   โ””โ”€โ”€ reader.py          # Audit reader + replay engine
โ”œโ”€โ”€ integrations/
โ”‚   โ””โ”€โ”€ openai.py          # Sync + Async OpenAI proxy
โ””โ”€โ”€ cli.py                 # CLI audit reader (7 commands)

Why AgentGuard?

Problem How AgentGuard Solves It
"Nobody knows what our agent is doing" Every LLM call and tool use is logged with full context
"We can't trace agent failures" Run-level audit trails with step-by-step replay
"Auditors want proof" JSON-lines logs + CSV export mapped to compliance frameworks
"Humans verify 69% of AI decisions" Policy guardrails let you reduce human review confidently
"Agents keep leaking PII" Automatic PII detection and blocking on all inputs & outputs
"AI costs are unpredictable" Per-run, daily, and total budget limits with real-time tracking
"Demo works, production doesn't" The missing operating system โ€” cost controls, guardrails, audit trails

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

agentaudit_sdk-0.1.2.tar.gz (44.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

agentaudit_sdk-0.1.2-py3-none-any.whl (42.8 kB view details)

Uploaded Python 3

File details

Details for the file agentaudit_sdk-0.1.2.tar.gz.

File metadata

  • Download URL: agentaudit_sdk-0.1.2.tar.gz
  • Upload date:
  • Size: 44.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.0

File hashes

Hashes for agentaudit_sdk-0.1.2.tar.gz
Algorithm Hash digest
SHA256 dccee4b67985ce5cfbbb5575883caf04bbb17a6606914b123df4ab144c5da007
MD5 fd57a3b74aabf074011fabbe68d61b51
BLAKE2b-256 60950cecc3bddf69eba41e34e4a4449a6b310e72ae85851199b278819a010321

See more details on using hashes here.

File details

Details for the file agentaudit_sdk-0.1.2-py3-none-any.whl.

File metadata

  • Download URL: agentaudit_sdk-0.1.2-py3-none-any.whl
  • Upload date:
  • Size: 42.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.0

File hashes

Hashes for agentaudit_sdk-0.1.2-py3-none-any.whl
Algorithm Hash digest
SHA256 1f0d8c795da014e68a342a73207243a6e1e8cbe2503a83eef37965d5232f734f
MD5 6023859d96baff2908683e99e20e3df9
BLAKE2b-256 a36dbfcda96a7559f69ef2f18af18827ec972f034359ad12753d9dccd59d4ecf

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page