Standalone framework to evaluate agent correctness based on portable OpenTelemetry traces
Project description
Ship Agents Reliably
Benchmark your agents before they hit production.
agentevals scores performance and inference quality from OpenTelemetry traces — no re-runs, no guesswork.
Install · Quick Start · Releases · Contributing · Discord
What is agentevals?
agentevals is a framework-agnostic evaluation solution that scores AI agent behavior directly from OpenTelemetry traces. Record your agent's actions once, then evaluate as many times as you want — no re-runs, no guesswork.
It works with any OTel-instrumented framework (LangChain, Strands, Google ADK, and others), supports Jaeger JSON and OTLP trace formats, and ships with built-in evaluators, custom evaluator support, and LLM-based judges.
- CLI for scripting and CI pipelines
- Web UI for visual inspection and local developer experience
- MCP server so MCP clients can run evaluations from a conversation
Why agentevals?
Most evaluation tools require you to re-execute your agent for every test — burning tokens, time, and money on duplicate LLM calls. agentevals takes a different approach:
- No re-execution — score agents from existing traces without replaying expensive LLM calls
- Framework-agnostic — works with any agent framework that emits OpenTelemetry spans
- Golden eval sets — compare actual behavior against defined expected behaviors for deterministic pass/fail gating
- Custom evaluators — write scoring logic in Python, JavaScript, or any language
- CI/CD ready — gate deployments on quality thresholds directly in your pipeline
- Local-first — no cloud dependency required; everything runs on your machine
How It Works
agentevals follows three simple steps:
- Collect traces — Instrument your agent with OpenTelemetry (or export traces from your tracing backend). Point the OTLP exporter at the agentevals receiver, or load trace files directly.
- Define eval sets — Create golden evaluation sets that describe expected agent behavior: which tools should be called, in what order, and what the output should look like.
- Run evaluations — Use the CLI, Web UI, or MCP server to score traces against your eval sets. Get per-metric scores, pass/fail results, and detailed span-level breakdowns.
[!IMPORTANT] This project is under active development. Expect breaking changes.
Contents
- Installation
- Quick Start
- Integration
- CLI
- Custom Evaluators
- Web UI
- REST API Reference
- MCP Server
- Claude Code Skills
- Docs
- Development
- FAQ
Installation
From PyPI (recommended): the published package includes the CLI, REST API, and embedded web UI.
pip install agentevals-cli
Optional extras:
pip install "agentevals-cli[live]" # MCP server support
pip install "agentevals-cli[openai]" # OpenAI Evals API graders
GitHub releases also ship core wheels (CLI and API only) and bundle wheels (with the embedded UI) if you need a specific version or offline pip install ./path/to.whl.
From source with uv or Nix:
uv sync
# or: nix develop .
See DEVELOPMENT.md for build instructions.
Quick Start
Examples use agentevals on your PATH after pip install agentevals-cli. If you are working from a clone of this repo, use uv run agentevals instead.
Run an evaluation against a sample trace:
agentevals run samples/helm.json \
--eval-set samples/eval_set_helm.json \
-m tool_trajectory_avg_score
List available evaluators:
agentevals evaluator list
Integration
Zero-Code (Recommended)
Point any OTel-instrumented agent at the receiver. No SDK, no code changes:
# Terminal 1
agentevals serve --dev
# Terminal 2
export OTEL_EXPORTER_OTLP_ENDPOINT=http://localhost:4318
export OTEL_RESOURCE_ATTRIBUTES="agentevals.session_name=my-agent"
python your_agent.py
Traces stream to the UI in real-time. Works with LangChain, Strands, Google ADK, or any framework that emits OTel spans (http/protobuf and http/json supported). Sessions are auto-created and grouped by agentevals.session_name. Set agentevals.eval_set_id to associate traces with an eval set.
See examples/zero-code-examples/ for working examples.
SDK
For programmatic session lifecycle and decorator API:
from agentevals import AgentEvals
app = AgentEvals()
with app.session(eval_set_id="my-eval"):
agent.invoke("Roll a 20-sided die for me")
Requires pip install "agentevals-cli[streaming]". See examples/sdk_example/ for framework-specific patterns.
CLI
# Single trace
agentevals run samples/helm.json \
--eval-set samples/eval_set_helm.json \
-m tool_trajectory_avg_score
# Multiple traces
agentevals run samples/helm.json samples/k8s.json \
--eval-set samples/eval_set_helm.json \
-m tool_trajectory_avg_score
# JSON output
agentevals run samples/helm.json \
--eval-set samples/eval_set_helm.json \
--output json
# List available evaluators (builtin + community)
agentevals evaluator list
# List only builtin evaluators
agentevals evaluator list --source builtin
Custom Evaluators
Beyond the built-in metrics, you can write your own evaluators in Python, JavaScript, or any language. An evaluator is any program that reads JSON from stdin and writes a score to stdout.
agentevals evaluator init my_evaluator
This scaffolds a directory with boilerplate and a manifest. You can also list supported runtimes and generate config snippets:
agentevals evaluator runtimes # show supported languages
agentevals evaluator config my_evaluator --path ./evaluators/my_evaluator.py
Implement your scoring logic, then reference it in an eval config:
# eval_config.yaml
evaluators:
- name: tool_trajectory_avg_score
type: builtin
- name: my_evaluator
type: code
path: ./evaluators/my_evaluator.py
threshold: 0.7
agentevals run trace.json --config eval_config.yaml --eval-set eval_set.json
Community evaluators can be referenced directly from a shared GitHub repository using type: remote. You can also delegate grading to the OpenAI Evals API using type: openai_eval (requires pip install "agentevals-cli[openai]" and OPENAI_API_KEY). See the Custom Evaluators guide for the full protocol reference, SDK usage, and how to contribute evaluators.
Web UI
Installed bundle (port 8001):
agentevals serve
From source (two terminals):
uv run agentevals serve --dev # Terminal 1
cd ui && npm install && npm run dev # Terminal 2 → http://localhost:5173
Upload traces and eval sets, select metrics, and view results with interactive span trees. Live-streamed traces appear in the "Local Dev" tab, grouped by session ID.
REST API Reference
While the server is running, interactive API documentation is available at:
| Endpoint | Description |
|---|---|
/docs |
Swagger UI with interactive request builder |
/redoc |
ReDoc reference documentation |
/openapi.json |
Raw OpenAPI 3.x schema (for code generation or CI) |
The OTLP receiver (port 4318) serves its own docs at http://localhost:4318/docs.
MCP Server
Exposes evaluation tools to MCP clients. A .mcp.json at the project root lets Claude Code pick it up automatically.
| Tool | Requires serve |
Description |
|---|---|---|
list_metrics |
yes | List available metrics |
evaluate_traces |
no | Evaluate local trace files (OTLP or Jaeger) |
list_sessions |
yes | List streaming sessions |
summarize_session |
yes | Structured summary of a session's tool calls |
evaluate_sessions |
yes | Evaluate sessions against a golden reference |
# Custom server URL (requires pip install "agentevals-cli[live]")
AGENTEVALS_SERVER_URL=http://localhost:9000 agentevals mcp
The React UI and MCP server share the same in-memory session state and can run simultaneously.
Claude Code Skills
Two slash-command workflows in .claude/skills/, available automatically in this repo:
| Skill | What it does |
|---|---|
/eval |
Score traces or compare sessions against a golden reference |
/inspect |
Turn-by-turn narrative of a live session with anomaly detection |
Docs
| Guide | Description |
|---|---|
| Eval Set Format | Schema, field reference, and examples for golden eval set JSON files |
| Custom Evaluators | Write your own scoring logic in Python, JavaScript, or any language |
| Live Streaming | Real-time trace streaming, dev server setup, and session management |
| OpenTelemetry Compatibility | Supported OTel conventions, message delivery mechanisms, and OTLP receiver |
Development
uv run pytest # run tests
uv run agentevals serve --dev # backend
cd ui && npm run dev # frontend (separate terminal)
See DEVELOPMENT.md for build tiers, Makefile targets, and Nix setup. To contribute, see CONTRIBUTING.md.
FAQ
How does this compare to ADK's evaluations? Unlike ADK's LocalEvalService, which couples agent execution with evaluation, agentevals only handles scoring: it takes pre-recorded traces and compares them against expected behavior using metrics like tool trajectory matching, response quality, and LLM-based judgments.
However, if you're iterating on your agents locally, you can point your agents to agentevals and you will see rich runtime information in your browser. For more details, use the bundled wheel and explore the Local Development option in the UI.
How does this compare to Bedrock AgentCore's evaluation?
AgentCore's evaluation integration (via strands-agents-evals) also couples agent execution with evaluation. It re-invokes the agent for each test case, converts the resulting OTel spans to AWS's ADOT format, and scores them against 4 built-in evaluators (Helpfulness, Accuracy, Harmfulness, Relevance) via a cloud API call. This means you need an AWS account, valid credentials, and network access for every evaluation.
agentevals takes a different approach: it scores pre-recorded traces locally without re-running anything. It works with standard Jaeger JSON and OTLP formats from any framework, supports open-ended metrics (tool trajectory matching, LLM-based judges, custom scorers), and ships with a CLI, web UI, and MCP server. No cloud dependency required, though we do include all ADK's GCP-based evals as of now.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file agentevals_cli-0.6.1.tar.gz.
File metadata
- Download URL: agentevals_cli-0.6.1.tar.gz
- Upload date:
- Size: 1.3 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.11.2 {"installer":{"name":"uv","version":"0.11.2","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
22a71e64c02b868d8a233ac70764338bf04ee13f26243d3a64ff7d55e59ce744
|
|
| MD5 |
6bfc58e872f4ab3fbb842f94a45a6c52
|
|
| BLAKE2b-256 |
9b4b722c2c82910d99f1d1fca5d1d68be11c13b190fd4a2fbb9a14a31aa6f13c
|
File details
Details for the file agentevals_cli-0.6.1-py3-none-any.whl.
File metadata
- Download URL: agentevals_cli-0.6.1-py3-none-any.whl
- Upload date:
- Size: 559.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.11.2 {"installer":{"name":"uv","version":"0.11.2","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
aac8ea19d01f5ce9d25984a6516a892612e1c223469806739b2b89a00e9af4b3
|
|
| MD5 |
2f0850a1ab11bedbc137998bb55264c6
|
|
| BLAKE2b-256 |
e96302ed30af6ccfb2239433b3afaa281b002a18527effa776c3cfc626d0d944
|