Guardrails for the OpenAI Agents SDK
Project description
HiddenLayer Guardrails for OpenAI Agents (Beta)
Drop-in replacement for the Agents SDK Agent that wires HiddenLayer guardrails into agent and tool execution. Agent input/output and every tool call are sent through HiddenLayer's analyze endpoint so prompt-injection and policy violations are caught automatically.
Note: OpenAI's native guardrails only support blocking content—they do not support redacting sensitive information from input or output. This library provides redaction capabilities through HiddenLayer's
REDACTaction, allowing you to sanitize content while still allowing the request to proceed.
Installation
pip install hiddenlayer-openai-guardrails
Configuration
Environment Variables
The following environment variables must be set for authentication:
HIDDENLAYER_CLIENT_ID- HiddenLayer API client ID (required)HIDDENLAYER_CLIENT_SECRET- HiddenLayer API client secret (required)
Optional environment variables:
HIDDENLAYER_PROJECT_ID- HiddenLayer project ID for policy routingHIDDENLAYER_REQUESTER_ID- Identifier for tracking requests (default: "hiddenlayer-openai-integration")
# Required
export HIDDENLAYER_CLIENT_ID="your-client-id"
export HIDDENLAYER_CLIENT_SECRET="your-client-secret"
# Optional
export HIDDENLAYER_PROJECT_ID="your-project-id"
export HIDDENLAYER_REQUESTER_ID="your-app-name"
HiddenLayerParams
Configure HiddenLayer behavior using the HiddenLayerParams object:
from hiddenlayer_openai_guardrails import HiddenLayerParams
params = HiddenLayerParams(
project_id="my-project", # Optional: HiddenLayer project ID for policy routing
model="gpt-4o-mini", # Optional: Model name for tracking (auto-detected from agent if not set)
requester_id="my-app-v1", # Optional: Identifier for tracking requests (default: "hiddenlayer-openai-integration")
)
All fields are optional. If model is not provided, it will be automatically detected from the agent's model configuration.
Usage
Basic Agent with Guardrails
The Agent class mirrors agents.Agent but adds HiddenLayer guardrails to the agent and all tools. Guardrails automatically block malicious content:
from agents import Runner, function_tool
from agents.run import RunConfig
from hiddenlayer_openai_guardrails import Agent, HiddenLayerParams
@function_tool
def get_weather(city: str) -> str:
"""returns weather info for the specified city."""
return f"The weather in {city} is sunny"
# Configure HiddenLayer parameters
params = HiddenLayerParams(
project_id="my-project", # optional: for policy routing
)
agent = Agent(
name="Haiku agent",
instructions="Always respond in haiku form",
model="gpt-4o-mini",
tools=[get_weather], # tool input/output are screened by HiddenLayer
hiddenlayer_params=params, # optional: defaults will be used if not provided
)
result = Runner.run_sync(
agent,
"What's the weather in Toronto",
run_config=RunConfig(tracing_disabled=True),
)
print(result.final_output)
Redacting Input and Output
Since OpenAI's guardrails can only block (not redact), this library provides helper functions for content redaction:
from agents import Runner
from hiddenlayer_openai_guardrails import (
Agent,
HiddenLayerParams,
redact_input,
redact_output,
InputBlockedError,
OutputBlockedError,
)
# Configure HiddenLayer parameters
params = HiddenLayerParams(project_id="my-project")
agent = Agent(
name="Assistant",
instructions="You are a helpful assistant.",
hiddenlayer_params=params,
)
try:
# Redact sensitive info from user input before processing
safe_input = await redact_input(
user_input,
hiddenlayer_params=params,
)
# Run agent (guardrails will block malicious content)
result = await Runner.run(agent, safe_input)
# Redact sensitive info from output before showing to user
safe_output = await redact_output(
result.final_output,
hiddenlayer_params=params,
)
print(safe_output)
except InputBlockedError:
print("Input was blocked by HiddenLayer")
except OutputBlockedError:
print("Output was blocked by HiddenLayer")
Safe Streaming Output
For streaming responses, use safe_stream to stream event objects while scanning the final output through HiddenLayer guardrails:
from agents import Runner
from hiddenlayer_openai_guardrails import Agent, HiddenLayerParams, safe_stream
# Configure HiddenLayer parameters
params = HiddenLayerParams(project_id="my-project")
agent = Agent(
name="Assistant",
instructions="Help users",
hiddenlayer_params=params,
)
result = Runner.run_streamed(agent, user_input)
async for event in safe_stream(result, hiddenlayer_params=params):
# event is an Agents SDK stream event (not plain text)
print(event)
If HiddenLayer returns a BLOCK action for the final streamed output, safe_stream raises OutputBlockedError after streaming completes.
MCP Server Tools
When using MCP servers with the Agents SDK, HiddenLayer guardrails are automatically applied to dynamically discovered MCP tools:
from agents import Runner
from agents.mcp import MCPServerStreamableHttp
from agents.run import RunConfig
from hiddenlayer_openai_guardrails import Agent, HiddenLayerParams
servers = [
MCPServerStreamableHttp(name="calculator", params={"url": "http://localhost:8000/mcp"}),
]
agent = Agent(
name="Math agent",
instructions="Use the calculator to answer math questions.",
model="gpt-4o-mini",
hiddenlayer_params=HiddenLayerParams(project_id="my-project"),
mcp_servers=servers,
)
result = await Runner.run(
agent,
"What is 2 + 2?",
run_config=RunConfig(tracing_disabled=True),
)
print(result.final_output)
MCP tool definitions are scanned through HiddenLayer at discovery time. Tools that violate policy are blocked and excluded from the agent (fail-closed). Scan results are cached per tool so repeated get_mcp_tools() calls don't re-scan the same definitions.
How it works
hiddenlayer_openai_guardrails.agents.Agentreturns a regularagents.Agentconfigured with:- Agent-level input/output guardrails that analyze user and assistant messages.
- Tool-level guardrails that inspect tool arguments before execution and tool output afterward.
- Guardrails rely on
AsyncHiddenLayer.interactions.analyzeand will raise when HiddenLayer signals a blocking action. - Input guardrails scan one message per request, in order, and skip already-seen messages for the same conversation thread.
- Tool guardrails currently enforce block-only behavior;
REDACTactions are not applied in tool hooks. - Strict phase mapping is used: model-produced tool arguments scan as
output, and tool results destined for the model scan asinput. - MCP tools are scanned at discovery time — tool definitions that violate policy are excluded (fail-closed). Allowed tools receive the same input/output guardrails as regular tools. Per-tool scan results are cached so repeated discovery calls are efficient.
Development
# Install dependencies (uses uv)
uv sync
# Run unit tests (no network, mocked dependencies)
pytest tests -m unit
# Run live integration tests (requires credentials and OPENAI_API_KEY)
RUN_LIVE_INTEGRATION_TESTS=1 pytest tests -m integration
# Run all tests
pytest tests
- Public API lives in
src/hiddenlayer_openai_guardrails/agents.py(facade); implementation is split across:src/hiddenlayer_openai_guardrails/_hiddenlayer.py(HiddenLayer client + analyze calls)src/hiddenlayer_openai_guardrails/_guardrails.py(agent/tool/MCP guardrail wiring)src/hiddenlayer_openai_guardrails/_normalize.py(payload normalization)src/hiddenlayer_openai_guardrails/_analysis.py(response parsing)src/hiddenlayer_openai_guardrails/_redaction.pyandsrc/hiddenlayer_openai_guardrails/_streaming.py
- Tests are in
tests/test_agents_unit.pyandtests/test_agents_integration.py.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters