Skip to main content

Standalone framework to evaluate agent correctness based on portable OpenTelemetry traces

Project description

agentevals

Ship Agents Reliably

Benchmark your agents before they hit production.
agentevals scores performance and inference quality from OpenTelemetry traces. No re-runs, no guesswork.

GitHub Stars   Discord   Release   License   PyPI

Install · Quick Start · Releases · Contributing · Discord


What is agentevals?

agentevals is a framework-agnostic evaluation solution that scores AI agent behavior directly from OpenTelemetry traces. Record your agent's actions once, then evaluate as many times as you want. No re-runs, no guesswork.

It works with any OTel-instrumented framework (LangChain, Strands, Google ADK, OpenAI Agents SDK, and others), supports Jaeger JSON and native OTLP trace formats, and ships with built-in evaluators, custom evaluator support, and LLM-based judges.

  • CLI for scripting and CI pipelines
  • Web UI for visual inspection and local developer experience
  • Kubernetes and OTel support so you can deploy right next to your agents; works natively in your OpenTelemetry pipeline
  • MCP server so MCP clients can run evaluations from a conversation

Why agentevals?

Most evaluation tools require you to re-execute your agent for every test, burning tokens, time, and money on duplicate LLM calls. agentevals takes a different approach:

  • No re-execution: score agents from existing traces without replaying expensive LLM calls
  • Framework-agnostic: works with any agent framework that emits OpenTelemetry spans
  • Golden eval sets: compare actual behavior against defined expected behaviors for deterministic pass/fail gating
  • Custom evaluators: write scoring logic in Python, JavaScript, or any language, or offload scoring to OpenAI Eval API
  • CI/CD ready: gate deployments on quality thresholds directly in your pipeline
  • Local-first: no cloud dependency required; everything runs on your machine

How It Works

agentevals follows three simple steps:

  1. Collect traces: Instrument your agent with OpenTelemetry (or export traces from your tracing backend). Point the OTLP exporter at the agentevals receiver, or load trace files directly.
  2. Define eval sets: Create golden evaluation sets that describe expected agent behavior: which tools should be called, in what order, and what the output should look like.
  3. Run evaluations: Use the CLI, Web UI, or MCP server to score traces against your eval sets. Get per-metric scores, pass/fail results, and detailed span-level breakdowns.

[!IMPORTANT] This project is under active development. Expect breaking changes.

Contents

Installation

From PyPI (recommended): the published package includes the CLI, REST API, and embedded web UI.

pip install agentevals-cli

Optional extras:

pip install "agentevals-cli[live]"        # MCP server support
pip install "agentevals-cli[openai]"      # OpenAI Evals API graders

GitHub releases also ship core wheels (CLI and API only) and bundle wheels (with the embedded UI) if you need a specific version or offline pip install ./path/to.whl.

From source with uv or Nix:

uv sync
# or: nix develop .

See DEVELOPMENT.md for build instructions.

Quick Start

Examples use agentevals on your PATH after pip install agentevals-cli. If you are working from a clone of this repo, use uv run agentevals instead.

The samples/ directory includes real traces from a Kubernetes Helm agent and matching eval sets that define expected behavior (which tools should be called, what the response should contain).

Score a trace against an eval set:

agentevals run samples/helm.json \
  --eval-set samples/eval_set_helm.json \
  -m tool_trajectory_avg_score

The agent was asked to list Helm releases. The eval set expects a call to helm_list_releases. It matches:

Trace: 3e289017fe03ffd7c4145316d2eb3d0d
Invocations: 1
        Metric                       Score  Status      Per-Invocation  Time
------  -------------------------  -------  --------  ----------------  ------
[PASS]  tool_trajectory_avg_score        1  PASSED                   1  0ms

Catch a mismatch. Run a different trace against the same eval set:

agentevals run samples/k8s.json \
  --eval-set samples/eval_set_helm.json \
  -m tool_trajectory_avg_score

This trace is from a different agent session that never called the expected tool. The evaluation fails:

[FAIL]  tool_trajectory_avg_score        0  FAILED                   0  0ms
  Invocation 1 trajectory mismatch:
    Expected:
      - helm_list_releases({})
    Actual:
      (none)

Evaluate multiple dimensions at once:

agentevals run samples/helm_3.json \
  --eval-set samples/evalset_helm_3_2026-02-23.json \
  -m tool_trajectory_avg_score \
  -m response_match_score

tool_trajectory_avg_score checks whether the right tools were called. response_match_score checks whether the agent's final answer matches the expected response.

Explore visually. Launch the Web UI and upload traces from the browser:

agentevals serve
# opens http://localhost:8001

You can also point any OTel-instrumented agent directly at the built-in receiver (OTEL_EXPORTER_OTLP_ENDPOINT=http://localhost:4318). The UI streams tool calls, inputs, and outputs live as your agent runs. For production setups, the same receiver slots into a Kubernetes OTel Collector pipeline as an exporter destination. See Integration and the Kubernetes example for walkthroughs.

Next steps:

  • agentevals evaluator list to see all built-in and community evaluators
  • Custom Evaluators to write your own scoring logic

Use-cases and integrations

Zero-Code (Recommended)

Point any OTel-instrumented agent at the agentevals receiver. No SDK, no code changes:

# Terminal 1
agentevals serve --dev

# Terminal 2
export OTEL_EXPORTER_OTLP_ENDPOINT=http://localhost:4318
export OTEL_RESOURCE_ATTRIBUTES="agentevals.session_name=my-agent"
python your_agent.py

For OTLP/gRPC exporters, use:

export OTEL_EXPORTER_OTLP_ENDPOINT=localhost:4317
export OTEL_EXPORTER_OTLP_PROTOCOL=grpc

Traces stream to the UI in real-time. Works with LangChain, Strands, Google ADK, OpenAI Agents SDK, or any framework that emits OTel spans (http/protobuf, http/json, and OTLP/gRPC supported). Sessions are auto-created and grouped by agentevals.session_name. Set agentevals.eval_set_id to associate traces with an eval set.

See examples/zero-code-examples/ for working examples.

AgentEvals SDK

For programmatic session lifecycle and decorator API:

from agentevals import AgentEvals

app = AgentEvals()

with app.session(eval_set_id="my-eval"):
    agent.invoke("Roll a 20-sided die for me")

Requires pip install "agentevals-cli[streaming]". See examples/sdk_example/ for framework-specific patterns.

CLI for local testing, and CI pipelines

# Multiple traces, JSON output
agentevals run samples/helm.json samples/k8s.json \
  --eval-set samples/eval_set_helm.json \
  -m tool_trajectory_avg_score \
  --output json

# List available evaluators
agentevals evaluator list

# Flexible trajectory matching (EXACT | IN_ORDER | ANY_ORDER)
agentevals run trace.json \
  --eval-set eval_set.json \
  -m tool_trajectory_avg_score \
  --trajectory-match-type IN_ORDER

Run agentevals run --help for all options.

Custom Evaluators

Write scoring logic in Python, JavaScript, or any language. Scaffold a new evaluator with:

agentevals evaluator init my_evaluator

Reference it alongside built-in metrics in an eval config:

evaluators:
  - name: tool_trajectory_avg_score
    type: builtin
  - name: my_evaluator
    type: code
    path: ./evaluators/my_evaluator.py
    threshold: 0.7

Evaluators with a requirements.txt get automatic virtual environment management. You can also use type: remote for community evaluators from GitHub, or type: openai_eval to delegate grading to the OpenAI Evals API (requires pip install "agentevals-cli[openai]").

See the Custom Evaluators guide for the full protocol reference, SDK helpers, and how to contribute evaluators.

Web UI

agentevals serve            # bundled UI on http://localhost:8001

Upload traces and eval sets, select metrics, and view results with interactive span trees. Live-streamed traces appear in the "Local Dev" tab, grouped by session ID. For running from source, see DEVELOPMENT.md.

Interactive API docs are available at /docs (Swagger) and /redoc while the server is running. The OTLP receiver on port 4318 serves its own docs at http://localhost:4318/docs.

Deployment

Docker

A Dockerfile is included at the project root. The image bundles the API, web UI, and OTLP receiver:

docker build -t agentevals .
docker run -p 8001:8001 -p 4318:4318 agentevals
Port Purpose
8001 Web UI and REST API
4318 OTLP HTTP receiver (traces and logs)
8080 MCP (Streamable HTTP)

Helm

A Helm chart is available in charts/agentevals/:

helm install agentevals ./charts/agentevals

See the Kubernetes example for an end-to-end walkthrough deploying agentevals alongside kagent and an OTel Collector on Kubernetes.

MCP Server

Exposes evaluation tools to MCP clients. A .mcp.json at the project root lets Claude Code pick it up automatically.

Tool Requires serve Description
list_metrics yes List available metrics
evaluate_traces no Evaluate local trace files (OTLP or Jaeger)
list_sessions yes List streaming sessions
summarize_session yes Structured summary of a session's tool calls
evaluate_sessions yes Evaluate sessions against a golden reference
# Custom server URL (requires pip install "agentevals-cli[live]")
AGENTEVALS_SERVER_URL=http://localhost:9000 agentevals mcp

The React UI and MCP server share the same in-memory session state and can run simultaneously.

Claude Code Skills

Two slash-command workflows in .claude/skills/, available automatically in this repo:

Skill What it does
/eval Score traces or compare sessions against a golden reference
/inspect Turn-by-turn narrative of a live session with anomaly detection

Examples

Working examples are in the examples/ directory:

Example Description
ADK Google ADK agent with zero-code OTel export
LangChain LangChain agent with zero-code OTel export
Strands Strands SDK agent with zero-code OTel export
OpenAI Agents OpenAI Agents SDK with zero-code OTel export
Ollama LangChain + Ollama for local LLM evaluation
Kubernetes End-to-end deployment with kagent and OTel Collector

Docs

Guide Description
Eval Set Format Schema, field reference, and examples for golden eval set JSON files
Custom Evaluators Write your own scoring logic in Python, JavaScript, or any language
Live Streaming Real-time trace streaming, dev server setup, and session management
OpenTelemetry Compatibility Supported OTel conventions, message delivery mechanisms, and OTLP receiver

Development

uv run pytest                      # run tests
uv run agentevals serve --dev      # backend
cd ui && npm run dev               # frontend (separate terminal)

See DEVELOPMENT.md for build tiers, Makefile targets, and Nix setup. To contribute, see CONTRIBUTING.md.

FAQ

Do I need a database or any infrastructure to run agentevals?

No. agentevals is a single pip install with no database, no message queue, and no external services. The CLI evaluates trace files directly from disk. The web UI and live streaming use in-memory session state. You can go from zero to scored traces in under a minute.

Does the CLI require a running server?

No. agentevals run evaluates trace files entirely offline. The server (agentevals serve) is only needed for the web UI, live OTLP streaming, and server-dependent MCP tools like list_sessions.

Can I use agentevals in CI/CD?

Yes. The CLI is designed for pipeline use: pass trace files and an eval set, set a threshold, and let the exit code gate your deployment. Combine it with --output json for machine-readable results. No server process needed.

What if I switch agent frameworks?

Because agentevals uses OpenTelemetry as its universal interface, switching frameworks (e.g., from LangChain to Strands, or from ADK to OpenAI Agents) does not require changing your evaluation setup. As long as your new framework emits OTel spans, the same eval sets and metrics work as before.

Can I write evaluators in my own language?

Yes. A custom evaluator is any program that reads JSON from stdin and writes a score to stdout. Python and JavaScript have first-class scaffolding support (agentevals evaluator init), but any language works. If your evaluator has a requirements.txt, agentevals manages a cached virtual environment automatically.

Can I plug agentevals into an existing OTel pipeline?

Yes. The OTLP receiver on port 4318 accepts standard http/protobuf and http/json trace exports, so it slots into any OpenTelemetry pipeline as just another exporter destination. If your pipeline uses gRPC (port 4317), place an OTel Collector in front to bridge gRPC to HTTP. The Kubernetes example shows this exact pattern.

Can I deploy agentevals on Kubernetes?

Yes. A Dockerfile and a Helm chart are included. A single pod exposes the web UI (8001), OTLP receiver (4318), and MCP server (8080). See the Kubernetes example for a full walkthrough deploying agentevals alongside kagent and an OTel Collector.

How does this compare to ADK's evaluations?

Unlike ADK's eval method, which couples agent execution with evaluation, agentevals only handles scoring: it takes pre-recorded traces and compares them against expected behavior using metrics like tool trajectory matching, response quality, and LLM-based judgments.

However, if you're iterating on your agents locally, you can point your agents to agentevals and you will see rich runtime information in your browser. For more details, use the bundled wheel and explore the Local Development option in the UI.

How does this compare to Bedrock AgentCore's evaluation?

AgentCore's evaluation integration (via strands-agents-evals) also couples agent execution with evaluation. It re-invokes the agent for each test case, converts the resulting OTel spans to AWS's ADOT format, and scores them against 4 built-in evaluators (Helpfulness, Accuracy, Harmfulness, Relevance) via a cloud API call. This means you need an AWS account, valid credentials, and network access for every evaluation.

agentevals takes a different approach: it scores pre-recorded traces locally without re-running anything. It works with standard Jaeger JSON and OTLP formats from any framework, supports open-ended metrics (tool trajectory matching, LLM-based judges, custom scorers), and ships with a CLI, web UI, and MCP server. No cloud dependency required, though we do include all ADK's GCP-based evals as of now.

How does this compare to LangSmith?

LangSmith is a cloud platform (self-hosting requires an Enterprise plan) where offline evaluation re-executes your application against curated datasets. Its deepest integration is with LangChain/LangGraph, though it can work with other frameworks. agentevals scores pre-recorded OTel traces without re-execution, requires no cloud account or enterprise license, and uses OpenTelemetry as the universal interface rather than a proprietary SDK.

How does this compare to Langfuse?

Langfuse is a full observability platform (requires Postgres, ClickHouse, Redis, and S3 for self-hosting) that supports both offline experiments (re-execution) and online evaluation of ingested traces. Traces must be ingested into Langfuse first via its SDK or OTel integration before they can be scored. agentevals evaluates raw OTel trace files or live OTLP streams directly with no database or platform infrastructure required.

How does this compare to Opik?

Opik's primary evaluation path re-runs your application code against a dataset, incurring additional LLM costs per eval run. It also supports online evaluation rules that auto-score production traces. While Opik supports OpenTelemetry ingestion alongside its own SDK, its evaluation workflow still centers on re-execution against datasets. agentevals evaluates pre-recorded OTel traces from any framework without re-execution, and runs entirely locally with no cloud dependency.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

agentevals_cli-0.6.4.tar.gz (1.3 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

agentevals_cli-0.6.4-py3-none-any.whl (570.3 kB view details)

Uploaded Python 3

File details

Details for the file agentevals_cli-0.6.4.tar.gz.

File metadata

  • Download URL: agentevals_cli-0.6.4.tar.gz
  • Upload date:
  • Size: 1.3 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.11.3 {"installer":{"name":"uv","version":"0.11.3","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for agentevals_cli-0.6.4.tar.gz
Algorithm Hash digest
SHA256 ab5fa87da3395ab592a1b56fc73c51bb6c70ebc0114df4b7d61ef03f15b3ca4e
MD5 d4c2853e48914111f26fcc6f87f8f7e5
BLAKE2b-256 20c7cd9b27fd622c48732475b2ce907c70da32db7d58bf111429dc844b2ae6cc

See more details on using hashes here.

File details

Details for the file agentevals_cli-0.6.4-py3-none-any.whl.

File metadata

  • Download URL: agentevals_cli-0.6.4-py3-none-any.whl
  • Upload date:
  • Size: 570.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.11.3 {"installer":{"name":"uv","version":"0.11.3","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for agentevals_cli-0.6.4-py3-none-any.whl
Algorithm Hash digest
SHA256 3228eca474524ce2fa9e0ecb815844440b296d8ba9293d1de10465a7ced85ce4
MD5 5b35ed9266ced8e7b65d1019d01db7a7
BLAKE2b-256 5f3600a283e32f0b8f6533b84b94dffdd899420c1777eed119603a74b33b77b2

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page