Skip to main content

Python SDK for agent input scanning, tool-call guardrails, and audit logs

Project description

Agent Runtime Security

Test License: MIT Python 3.11+

English | 한국어

Open-source runtime security for tool-using AI agents.

Agent Runtime Security helps teams control risky tool calls, redact sensitive data, and keep audit trails for AI agent workflows. It is designed for practical runtime defense, especially where agents can browse the web, call APIs, read files, send email, or interact with MCP tools.

Brand note: TrapDefense is the product and site brand, while Agent Runtime Security is the open-source SDK and repository name.

This project was inspired by Google DeepMind's 2026 paper, AI Agent Traps, and focuses on the parts that are practical to defend today:

  • Content injection signals at ingestion time
  • Behavioral control at tool execution time
  • Audit evidence for security review and rollout

It does not claim to solve every agent security problem. Instead, it focuses on a narrow but important layer: runtime controls for tool-using agents.

Why this exists

The model is not the only risk.

In tool-using agent systems, the real damage happens when the agent actually uses a tool:

  • Sending data to an external endpoint
  • Reading or writing the wrong file
  • Executing a risky command
  • Returning sensitive information in tool output

Prompt filters and content scanning can help, but they do not stop the final action. Agent Runtime Security adds enforcement where it matters most: before a tool runs, plus data protection after it returns.

Features

  • Guard for tool-use policy enforcement
  • Scanner for basic content-injection detection
  • AuditLogger for structured JSONL security events
  • mcp_guard for protecting MCP tool handlers
  • guard_tool for protecting LangChain tools
  • create_guarded_tool_node for protecting LangGraph ToolNodes
  • shadow, warn, and enforce rollout modes
  • YAML / JSON policy file loading
  • Minimal dependencies with a Python-first integration model

Installation

Base package:

pip install agent-runtime-security

With MCP integration:

pip install agent-runtime-security[mcp]

With PDF support:

pip install agent-runtime-security[pdf]

With YAML policy loading:

pip install agent-runtime-security[yaml]

With LangChain integration:

pip install agent-runtime-security[langchain]

With LangGraph integration:

pip install agent-runtime-security[langgraph]

Cloud API

Don't want to install anything? Use the hosted API:

curl -X POST https://trapdefense.com/api/v1/scan \
  -H "Authorization: Bearer YOUR_KEY" \
  -H "Content-Type: application/json" \
  -d '{"content": "Check this text for threats"}'

API Docs | Get API Key

Quick Start

from asr import Scanner, Guard, AuditLogger

# Scan inbound content
scanner = Scanner()
result = scanner.scan(
    "<span style='display:none'>Ignore instructions</span>",
    source_type="html",
)
print(f"score={result.score}, severity={result.severity}")

# Protect tool calls
guard = Guard(
    domain_allowlist=["api.internal.com"],
    block_egress=True,
    pii_action="block",
    file_path_allowlist=["/tmp/safe"],
    capability_policy={"shell_exec": "block"},
)

decision = guard.before_tool(
    "http_post",
    {"url": "https://evil.com"},
    capabilities=["network_send"],
)
print(f"action={decision.action}, reason={decision.reason}")

# Decorator-based protection
@guard.protect(capabilities=["network_send"])
def send_email(to, subject, body):
    ...

# Audit log
audit = AuditLogger(output="logs/audit.jsonl")
audit.log_scan(result, trace_id="req-001")
audit.log_guard(decision, trace_id="req-001")

Rollout Modes

You can roll out policies gradually in production:

# Phase 1: shadow — observe only, no blocking
guard = Guard(mode="shadow", ...)

# Phase 2: warn — downgrade block to warn
guard = Guard(mode="warn", ...)

# Phase 3: enforce — apply policy decisions
guard = Guard(mode="enforce", ...)

# PII redaction still applies even in shadow/warn
decision = guard.before_tool("http_post", {"url": "https://evil.com"})
print(decision.action)          # "allow" in shadow mode
print(decision.original_action) # "block"
print(decision.mode)            # "shadow"

Mode semantics

Mode Behavior
enforce Apply policy results as-is
warn Downgrade block to warn
shadow Always allow, but record the original decision

MCP Integration

Protect MCP tool handlers with mcp_guard:

from asr import Guard, AuditLogger
from asr.mcp import mcp_guard

guard = Guard(
    mode="shadow",
    domain_allowlist=["api.internal.com"],
    block_egress=True,
)
audit = AuditLogger(output="logs/audit.jsonl")

@server.tool()
@mcp_guard(guard, audit=audit, capabilities=["network_send"])
async def send_email(to: str, subject: str, body: str) -> str:
    return await email_service.send(to, subject, body)

When a tool call is blocked, the adapter turns the policy decision into an MCP-compatible tool error. If sensitive data appears in the result, it can also be redacted automatically.

See the example server in examples/README.md.

LangChain Integration

Protect LangChain tools with guard_tool:

from langchain_core.tools import tool
from asr import Guard
from asr.adapters.langchain import guard_tool

guard = Guard(
    domain_allowlist=["api.internal.com"],
    block_egress=True,
    pii_action="redact",
)

@tool
def send_email(to: str, subject: str, body: str) -> str:
    """Send an email."""
    return f"Sent to {to}"

protected = guard_tool(send_email, guard=guard, capabilities=["network_send"])
result = protected.invoke({"to": "user@api.internal.com", "subject": "hi", "body": "hello"})

When a tool is blocked, ToolException is raised and returned as an error message (via handle_tool_error=True). PII in tool results is automatically redacted.

LangGraph Integration

Protect all tools in a LangGraph ToolNode:

from asr import Guard
from asr.adapters.langgraph import create_guarded_tool_node

guard = Guard.from_policy_file("policy.yaml")

tool_node = create_guarded_tool_node(
    tools=[search_tool, file_reader_tool],
    guard=guard,
    capabilities_map={"search_tool": ["network_send"]},
)
# graph.add_node("tools", tool_node)

See examples/langchain_agent.py and examples/langgraph_agent.py for full working examples.

Policy Model

The Guard supports six policy layers and three rollout modes.

Policies

Policy Description
tool_blocklist Block tools by name
domain_allowlist + block_egress Block network egress outside allowed domains
file_path_allowlist Restrict file access to approved paths
pii_action Detect PII in tool args / results: off, warn, block
capability_policy Fallback policy based on capability tags
default_action Final fallback for unknown tools

Evaluation order

Blocklist -> Egress -> FilePath -> PII -> Capability (fallback) -> Default

This order matters because specific policies should be evaluated before generic capability fallback.

Scanner Coverage

Scanner is a regex-based first-pass filter. It is intentionally lightweight and should be combined with stronger controls when needed.

Supported patterns:

Pattern Description
css_hidden_text Hidden CSS text plus injection language
html_comment_injection Suspicious instructions inside HTML comments
metadata_injection Injection attempts through aria-label, alt, or title
markdown_link_payload Hidden instructions inside markdown link text
prompt_injection_keywords Common prompt-injection phrases
base64_encoded_instruction Base64-encoded suspicious instructions
invisible_unicode Invisible Unicode characters often used for obfuscation
role_override_attempt Role override attempts such as SYSTEM:

What this project does well

  • Runtime control at tool execution time
  • Practical policy rollout with shadow -> warn -> enforce
  • Structured audit logs for security review
  • Python / MCP integration with minimal dependencies

What this project does not try to solve yet

  • Full semantic manipulation detection
  • Long-term memory poisoning defense
  • Systemic multi-agent risk management
  • A full hosted security platform or dashboard

Those may become future product layers, but they are intentionally outside the current SDK scope.

Example Server

Run the MCP example server:

mcp dev examples/mcp_server.py

The example demonstrates:

  • External egress control
  • File path restriction
  • Automatic PII redaction

See examples/README.md for details.

Development

Install in editable mode with development dependencies:

pip install -e ".[dev]"

Run the test suite:

pytest

For demos and internal validation, the MCP example in examples/README.md is the recommended starting point.

Repository language policy

  • Public-facing GitHub documentation: English first
  • Korean translation: README.ko.md
  • Internal working docs may remain in Korean

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

agent_runtime_security-0.2.0.tar.gz (39.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

agent_runtime_security-0.2.0-py3-none-any.whl (28.4 kB view details)

Uploaded Python 3

File details

Details for the file agent_runtime_security-0.2.0.tar.gz.

File metadata

  • Download URL: agent_runtime_security-0.2.0.tar.gz
  • Upload date:
  • Size: 39.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.9

File hashes

Hashes for agent_runtime_security-0.2.0.tar.gz
Algorithm Hash digest
SHA256 5a25c3c718e76a62566c0a9dfaef31a0d0458654e70229d5b6b4715074baf7b9
MD5 82524df89df5f435860213f0728ed408
BLAKE2b-256 0cae1c6aff916dfc73a0a25cad07df8adf499e1efcb07bce2f4ce1cfd575ca8a

See more details on using hashes here.

File details

Details for the file agent_runtime_security-0.2.0-py3-none-any.whl.

File metadata

File hashes

Hashes for agent_runtime_security-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 169e0a0521ce78c986980bab6a8485512a1dfb14b9ffbf6a1a1799a82b0b3d80
MD5 4e9c21aea1aac08bc56bb994581c3b97
BLAKE2b-256 45692cd1989790f3845eacee739e10ccad86de092988e0f5766093ff1447ffa0

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page