Skip to main content

The official Python SDK for Eval Protocol (EP.) EP is an open protocol that standardizes how developers author evals for large language model (LLM) applications.

Project description

Eval Protocol (EP)

PyPI - Version Ask DeepWiki

Stop guessing which AI model to use. Build a data-driven model leaderboard.

With hundreds of models and configs, you need objective data to choose the right one for your use case. EP helps you evaluate real traces, compare models, and visualize results locally.

🚀 Features

  • Pytest authoring: @evaluation_test decorator to configure evaluations
  • Robust rollouts: Handles flaky LLM APIs and parallel execution
  • Integrations: Works with Langfuse, LangSmith, Braintrust, Responses API
  • Agent support: LangGraph and Pydantic AI
  • MCP RL envs: Build reinforcement learning environments with MCP
  • Built-in benchmarks: AIME, tau-bench
  • LLM judge: Stack-rank models using pairwise Arena-Hard-Auto
  • Local UI: Pivot/table views for real-time analysis

⚡ Quickstart (no labels needed)

Install with your tracing platform extras and set API keys:

pip install 'eval-protocol[langfuse]'

# Model API keys (set what you need)
export OPENAI_API_KEY=...
export FIREWORKS_API_KEY=...
export GEMINI_API_KEY=...

# Platform keys
export LANGFUSE_PUBLIC_KEY=...
export LANGFUSE_SECRET_KEY=...
export LANGFUSE_HOST=https://your-deployment.com  # optional

Minimal evaluation using the built-in AHA judge:

from datetime import datetime
import pytest

from eval_protocol import (
    evaluation_test,
    aha_judge,
    EvaluationRow,
    SingleTurnRolloutProcessor,
    DynamicDataLoader,
    create_langfuse_adapter,
)


def langfuse_data_generator() -> list[EvaluationRow]:
    adapter = create_langfuse_adapter()
    return adapter.get_evaluation_rows(
        to_timestamp=datetime.utcnow(),
        limit=20,
        sample_size=5,
    )


@pytest.mark.parametrize(
    "completion_params",
    [
        {"model": "openai/gpt-4.1"},
        {"model": "fireworks_ai/accounts/fireworks/models/gpt-oss-120b"},
    ],
)
@evaluation_test(
    data_loaders=DynamicDataLoader(generators=[langfuse_data_generator]),
    rollout_processor=SingleTurnRolloutProcessor(),
)
async def test_llm_judge(row: EvaluationRow) -> EvaluationRow:
    return await aha_judge(row)

Run it:

pytest -q -s

The pytest output includes local links for a leaderboard and row-level traces (pivot/table) at http://localhost:8000.

Installation

This library requires Python >= 3.10.

pip

pip install eval-protocol

uv (recommended)

# Install uv (if needed)
curl -LsSf https://astral.sh/uv/install.sh | sh

# Add to your project
uv add eval-protocol

📚 Resources

License

MIT

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

eval_protocol-0.2.54.tar.gz (2.0 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

eval_protocol-0.2.54-py3-none-any.whl (2.0 MB view details)

Uploaded Python 3

File details

Details for the file eval_protocol-0.2.54.tar.gz.

File metadata

  • Download URL: eval_protocol-0.2.54.tar.gz
  • Upload date:
  • Size: 2.0 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for eval_protocol-0.2.54.tar.gz
Algorithm Hash digest
SHA256 10fe3223873acfea20fbbc1b67b77b53f6f9338c242225de94d8fe51cab6e1a0
MD5 7dc4a6824a69bf6da091f154d89138ff
BLAKE2b-256 6a3c52324f2a55df060a03336c24ef02287df388bf8c3ee40c915712eda27c12

See more details on using hashes here.

Provenance

The following attestation bundles were made for eval_protocol-0.2.54.tar.gz:

Publisher: release.yml on eval-protocol/python-sdk

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file eval_protocol-0.2.54-py3-none-any.whl.

File metadata

  • Download URL: eval_protocol-0.2.54-py3-none-any.whl
  • Upload date:
  • Size: 2.0 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for eval_protocol-0.2.54-py3-none-any.whl
Algorithm Hash digest
SHA256 f46ab4f62a34f2bb2c40d4b22d4c151e4587bc70189f266c947d7a8a3de035a0
MD5 421167da643efc66daa4e7e2ddc970f6
BLAKE2b-256 226d7a8adb0b8913ce0306b4b912abd31630d2bde898efb1e6d0844937b25362

See more details on using hashes here.

Provenance

The following attestation bundles were made for eval_protocol-0.2.54-py3-none-any.whl:

Publisher: release.yml on eval-protocol/python-sdk

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page