OpenTelemetry GenAI Utils
Project description
This package plugs the deepeval metrics suite into the OpenTelemetry GenAI evaluation pipeline. When it is installed a Deepeval evaluator is registered automatically and, unless explicitly disabled, is executed for every LLM/agent invocation alongside the builtin metrics.
Installation
Install the evaluator (and its runtime dependencies) from PyPI:
pip install opentelemetry-util-genai-evals-deepeval
The command pulls in opentelemetry-util-genai, deepeval and openai automatically so the evaluator is ready to use right after installation.
Requirements
opentelemetry-util-genai together with deepeval and openai – these are installed automatically when you install this package.
An LLM provider supported by Deepeval. By default the evaluator uses OpenAI’s gpt-4o-mini model because it offers the best balance of latency and cost for judge workloads right now, so make sure OPENAI_API_KEY is available. To override the model, set DEEPEVAL_EVALUATION_MODEL (or DEEPEVAL_MODEL / OPENAI_MODEL) to a different deployment along with the corresponding provider credentials.
(Optional) DEEPEVAL_API_KEY if your Deepeval account requires it.
Configuration
Use OTEL_INSTRUMENTATION_GENAI_EVALS_EVALUATORS to select the metrics that should run. Leaving the variable unset enables every registered evaluator with its default metric set. Examples:
OTEL_INSTRUMENTATION_GENAI_EVALS_EVALUATORS=Deepeval – run the default Deepeval bundle (Bias, Toxicity, Answer Relevancy, Faithfulness).
OTEL_INSTRUMENTATION_GENAI_EVALS_EVALUATORS=Deepeval(LLMInvocation(bias(threshold=0.75))) – override the Bias threshold for LLM invocations and skip the remaining metrics.
OTEL_INSTRUMENTATION_GENAI_EVALS_EVALUATORS=none – disable the evaluator entirely.
Results are emitted through the standard GenAI evaluation emitters (events, metrics, spans). Each metric includes helper attributes such as deepeval.success, deepeval.threshold and any evaluation model metadata returned by Deepeval. Metrics that cannot run because required inputs are missing (for example Faithfulness without a retrieval_context) are marked as label="skipped" and carry a deepeval.error attribute so you can wire the necessary data or disable that metric explicitly.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file splunk_otel_genai_evals_deepeval-0.1.6.tar.gz.
File metadata
- Download URL: splunk_otel_genai_evals_deepeval-0.1.6.tar.gz
- Upload date:
- Size: 18.3 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: python-httpx/0.28.1
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
eb98700b2370cc66c9ce5ebf94e2c045e0599cf2c15909eaf333545bddcaab57
|
|
| MD5 |
f29837651e54d016dd93c1af2ffd0429
|
|
| BLAKE2b-256 |
4787bba883f6a3248b70258535e13b90bc48dd3fa5bd4537f5403da580d23210
|
File details
Details for the file splunk_otel_genai_evals_deepeval-0.1.6-py3-none-any.whl.
File metadata
- Download URL: splunk_otel_genai_evals_deepeval-0.1.6-py3-none-any.whl
- Upload date:
- Size: 18.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: python-httpx/0.28.1
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
5d325e5400f97d308b5e38e3ec06515a96fc5498b505f7d3dc6c5f05c5b84160
|
|
| MD5 |
b5873b1a4fbbf5472f50a8a7fe16af96
|
|
| BLAKE2b-256 |
3151507374c8c47a5cbb42424cd18f0407d9481ba8c31fee4a3a5c6d695b3dac
|