Zero-config observability for AI agents
Project description
cProfile tells you where CPU time went in your Python code. Peekr tells you where time, tokens, and money went in your agent — and what each step actually saw and returned.
# cProfile
function calls cumtime
search_results 1 3.8s
openai.create 2 0.9s
# peekr
tool.search_web 3800ms ← same bottleneck, now you can fix it
openai.chat 490ms 891tok ← plus token cost you'd never see in cProfile
But agents fail for reasons a profiler can't catch: a tool returned null, the LLM received a malformed prompt, history grew until it pushed the system prompt out of the context window. Peekr captures the semantics — inputs, outputs, LLM context — not just timing.
Two lines to add, no backend required.
pip install peekr
import peekr
peekr.instrument()
# Your existing agent code — zero changes needed
How to use it
Step 1 — Instrument
Call peekr.instrument() once, before any LLM calls. It patches the OpenAI and Anthropic SDKs automatically.
import peekr
peekr.instrument()
import openai
response = openai.chat.completions.create(
model="gpt-4o",
messages=[{"role": "user", "content": "Summarize this doc"}]
)
Every LLM call is now captured. Peekr writes spans to traces.jsonl and prints them live to the console.
Step 2 — Trace your tools
Decorate your tool functions with @trace so they appear in the tree alongside LLM calls:
from peekr import trace
@trace
def search_web(query: str) -> list[str]:
return fetch_results(query)
@trace(name="tool.calculator")
def calculate(expression: str) -> float:
return eval(expression)
@trace # async works too
async def fetch_user(user_id: int) -> dict:
return await db.get(user_id)
Decorated functions nest automatically under whatever called them — no wiring needed.
Step 3 — View the trace
peekr view traces.jsonl # tree view
peekr view --io traces.jsonl # include inputs and outputs
Trace a3f2b1c0 1243ms 891tok
────────────────────────────────────────────────
agent.run 1243ms
└─ tool.search_web 210ms
in: {"query": "climate policy"}
out: ["result1", "result2", ...]
└─ openai.chat.completions [gpt-4o] 1033ms 891tok
in: [{"role": "user", "content": "..."}]
out: "Based on recent research..."
Now you can see exactly what happened — what went in, what came out, how long each step took, how many tokens were used.
What it profiles
A CPU profiler tells you a function was slow. Peekr tells you a function was slow, returned bad data, and passed it to an LLM that had no idea.
Full examples with annotated traces → docs
Wrong answers
The exact prompt that was sent — not what you think was sent, what was actually sent. Spot bad tool output before it reaches the LLM.
agent.run 2100ms
└─ tool.fetch_user 12ms
out: null ← returned null, agent didn't check
└─ openai.chat [gpt-4o] 2088ms
in: "User profile: null..." ← LLM received garbage
Slow responses
agent.run 4300ms
└─ tool.search_web 3800ms ← 88% of time. Cache this, not swap models.
└─ openai.chat 490ms
High token costs
Trace 1: 18,432 tokens
Trace 2: 21,104 tokens
Trace 3: 24,891 tokens ← growing = unbounded history. Summarize after 5 turns.
Prod vs local bugs
# local: out: [{"id": 1, "qty": 42}]
# prod: out: [] ← data pipeline bug, not agent logic
What's in v0.2
| Capability | API |
|---|---|
| Session tracing | with peekr.session(user_id="u1"): |
| Alerts | instrument(alerts=[peekr.alert.ErrorRate(0.05)]) |
| LLM-as-judge eval | instrument(evaluators=[peekr.eval.Rubric("Be concise")]) |
| Feedback + fine-tuning export | peekr.feedback(trace_id, rating="good") |
| A/B experiments | @peekr.experiment(variants=["control", "test"]) |
| Trace replay | peekr replay <trace_id> |
Supported clients
| Provider | SDK | Install |
|---|---|---|
| OpenAI | openai |
pip install "peekr[openai]" |
| Anthropic | anthropic |
pip install "peekr[anthropic]" |
| AWS Bedrock | boto3 |
pip install "peekr[bedrock]" |
All three auto-instrument with the same two lines — peekr.instrument() detects whichever SDKs are installed and patches them. Streaming is supported for all three.
import peekr
peekr.instrument()
# OpenAI
import openai
openai.chat.completions.create(model="gpt-4o", messages=[...])
# Anthropic
import anthropic
anthropic.Anthropic().messages.create(model="claude-opus-4-5", messages=[...])
# Bedrock
import boto3
boto3.client("bedrock-runtime").converse(modelId="anthropic.claude-3-haiku-20240307-v1:0", messages=[...])
Installation
pip install peekr # base
pip install "peekr[openai]" # with OpenAI
pip install "peekr[anthropic]" # with Anthropic
pip install "peekr[bedrock]" # with AWS Bedrock
pip install "peekr[all]" # everything
Storage options
peekr.instrument() # JSONL — default, grep-able
peekr.instrument(storage="sqlite") # SQLite — queryable, multi-process safe
peekr.instrument(storage="both") # both at once
SQLite — query your traces with SQL
SQLite storage uses WAL mode so multiple processes (Docker, CI, parallel agents) can write safely at the same time. And because it's SQLite, you can query across runs:
# slowest tool calls
sqlite3 traces.db "
SELECT name, ROUND(AVG(duration_ms)) avg_ms
FROM spans GROUP BY name ORDER BY avg_ms DESC;"
# token spend by model
sqlite3 traces.db "
SELECT json_extract(attributes,'$.model') model,
SUM(json_extract(attributes,'$.tokens_total')) tokens
FROM spans GROUP BY model;"
# all errors
sqlite3 traces.db "
SELECT name, trace_id, json_extract(attributes,'$.error') msg
FROM spans WHERE status = 'error';"
# cost growth over time
sqlite3 traces.db "
SELECT trace_id,
SUM(json_extract(attributes,'$.tokens_total')) total
FROM spans GROUP BY trace_id ORDER BY start_time;"
View SQLite traces the same way as JSONL:
peekr view traces.db
peekr view --io traces.db
@trace options
@trace # auto-names from module.function, captures io
@trace(name="tool.search") # custom span name
@trace(capture_io=False) # skip capturing args/output (e.g. secrets)
Custom exporters
Ship spans to any backend by implementing a single method:
from peekr.exporters import add_exporter
class MyExporter:
def export(self, span):
requests.post("https://my-backend.com/spans", json=span.to_dict())
peekr.instrument()
add_exporter(MyExporter())
How it works
instrument() monkey-patches the OpenAI and Anthropic SDK methods before your code runs. Python resolves function references at call time, so every subsequent call hits the wrapper with zero changes to your code.
Parent/child span relationships are tracked via Python's contextvars.ContextVar, which propagates correctly across async/await without manual threading.
Contributing
git clone https://github.com/ashwanijha04/peekr
cd peekr
pip install -e ".[dev]"
pytest
Open an issue before large changes. PRs welcome.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file peekr-0.2.0.tar.gz.
File metadata
- Download URL: peekr-0.2.0.tar.gz
- Upload date:
- Size: 38.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
09df380b99a8c456b1cca9d13a7a7cdb3c631a9ec6aaa7c978002523b90e1140
|
|
| MD5 |
41a57d27bc51f75356302b418fe80d03
|
|
| BLAKE2b-256 |
67cf41a9cdedd28b3b5711a6832c798ca52b5d7da42f5b2cf4501ee5492ffb96
|
Provenance
The following attestation bundles were made for peekr-0.2.0.tar.gz:
Publisher:
publish.yml on ashwanijha04/peekr
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
peekr-0.2.0.tar.gz -
Subject digest:
09df380b99a8c456b1cca9d13a7a7cdb3c631a9ec6aaa7c978002523b90e1140 - Sigstore transparency entry: 1504558512
- Sigstore integration time:
-
Permalink:
ashwanijha04/peekr@eadef27b8257ebbf7ea08d748c186ab7c9aef505 -
Branch / Tag:
refs/tags/v0.2.0 - Owner: https://github.com/ashwanijha04
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@eadef27b8257ebbf7ea08d748c186ab7c9aef505 -
Trigger Event:
push
-
Statement type:
File details
Details for the file peekr-0.2.0-py3-none-any.whl.
File metadata
- Download URL: peekr-0.2.0-py3-none-any.whl
- Upload date:
- Size: 26.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
575371e03ccb40554c487d2ca59d97b34b4d49f136b5ec59a1cf0f1ca2e4e0fb
|
|
| MD5 |
0146848491ae392764f0f2061b40ccf0
|
|
| BLAKE2b-256 |
f4d7dbdcb9ae41033322a28239ccc18da00d96774937946e04786cfe7362b7c4
|
Provenance
The following attestation bundles were made for peekr-0.2.0-py3-none-any.whl:
Publisher:
publish.yml on ashwanijha04/peekr
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
peekr-0.2.0-py3-none-any.whl -
Subject digest:
575371e03ccb40554c487d2ca59d97b34b4d49f136b5ec59a1cf0f1ca2e4e0fb - Sigstore transparency entry: 1504558692
- Sigstore integration time:
-
Permalink:
ashwanijha04/peekr@eadef27b8257ebbf7ea08d748c186ab7c9aef505 -
Branch / Tag:
refs/tags/v0.2.0 - Owner: https://github.com/ashwanijha04
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@eadef27b8257ebbf7ea08d748c186ab7c9aef505 -
Trigger Event:
push
-
Statement type: