Raindrop integration for Deep Agents
Project description
raindrop-deep-agents
Raindrop integration for Deep Agents (Python). Pass a single LangChain callback handler and Raindrop captures every LLM call, tool invocation, chain step, and agent action — with full OpenTelemetry trace nesting linked to the resulting dashboard event.
Installation
pip install raindrop-deep-agents deepagents langchain-core langchain-openai
Works with both
wrapt < 2.0andwrapt >= 2.0. The upstreamopentelemetry-instrumentation-langchainpackage (latest 0.59.2) still callswrapt.wrap_function_wrapper(module="...", ...)with the legacymodule=keyword argument that wrapt 2.0 renamed totarget=. Without a workaround the instrumentor crashes on activation and Traceloop silently swallows the error — traces never reach the dashboard while events keep working (tracked upstream at traceloop/openllmetry#4009 and Arize-ai/openinference#2996). On import, this package installs a small backwards-compat shim (raindrop_deep_agents/_wrapt_compat.py) that translatesmodule=→target=for any caller using wrapt 2.x; on wrapt 1.x the shim is a no-op. We can drop the shim once the upstream instrumentors migrate.
Quick Start
from raindrop_deep_agents import RaindropDeepAgents
from deepagents import create_deep_agent
from langchain_openai import ChatOpenAI
from langchain_core.messages import HumanMessage
raindrop = RaindropDeepAgents(
api_key="rk_...",
user_id="user-123",
convo_id="session-abc",
)
llm = ChatOpenAI(model="gpt-4o-mini")
agent = create_deep_agent(model=llm, system_prompt="Be concise.")
result = agent.invoke(
{"messages": [HumanMessage(content="Hello!")]},
config={"callbacks": [raindrop.handler]},
)
raindrop.shutdown()
Factory Function (alternative)
from raindrop_deep_agents import create_raindrop_deep_agents
raindrop = create_raindrop_deep_agents(api_key="rk_...", user_id="user-123")
agent = create_deep_agent(model=llm)
agent.invoke(
{"messages": [HumanMessage(content="Hello")]},
config={"callbacks": [raindrop.handler]},
)
raindrop.shutdown()
What Gets Captured
- LLM calls — model name, input, output, prompt + completion tokens, finish reason
- Tool calls — tool name, input arguments, output, duration as nested
TOOL_CALLspans linked to the parent event (viainteraction.track_tool()) - Chains — root chain input/output captured as the canonical event (the user's question → final assistant reply)
- Agent actions — tool selection events captured during the agent loop
- Errors — exception type and message captured for both LLM and tool failures
- Extended token categories — cached tokens (
ai.usage.cached_tokens) and reasoning tokens (ai.usage.thoughts_tokens) when available from the provider (e.g. OpenAI, Anthropic) - Finish reason — captured as
ai.finish_reasonin event properties ("stop","length","tool_calls", etc.) - Tags & metadata — anything passed via LangChain
config={"tags": [...], "metadata": {...}}is forwarded to event properties
Debug Mode
Enable verbose logging with debug=True:
raindrop = RaindropDeepAgents(
api_key="rk_...",
debug=True,
)
Identify Users
Associate events with a user after initialization:
raindrop.identify("user-123", {"name": "Alice", "plan": "pro"})
Track Signals
Send feedback, edits, or custom signals attached to a specific event:
# After an agent.invoke(...), attach a feedback signal to the event
# that just shipped. ``raindrop.last_event_id`` is the canonical
# event_id the handler emitted for the root chain — using it (rather
# than a freshly generated UUID) ensures the signal links to the real
# dashboard row instead of creating an orphan signal.
if raindrop.last_event_id:
raindrop.track_signal(
event_id=raindrop.last_event_id,
name="thumbs_up",
signal_type="feedback",
sentiment="POSITIVE",
comment="user liked the answer",
)
Flushing and Shutdown
raindrop.flush() # flush pending events
raindrop.shutdown() # flush + release resources (always call before process exit)
API Reference
RaindropDeepAgents
| Parameter | Type | Default | Description |
|---|---|---|---|
api_key |
Optional[str] |
None |
Raindrop API key. If None, telemetry shipping is disabled (a UserWarning is issued). |
user_id |
Optional[str] |
None |
Stamped on every event. Falls back to "unknown" if not set. |
convo_id |
Optional[str] |
None |
Conversation/session ID — events sharing this value are grouped on the dashboard. |
tracing_enabled |
bool |
True |
Enable distributed tracing (activates the LangChain OTel instrumentor). |
bypass_otel_for_tools |
bool |
True |
Bypass OTEL for tool spans — emits OTLP directly to /v1/traces. |
debug |
bool |
False |
Enable DEBUG-level logging. |
Methods
| Method | Description |
|---|---|
handler |
Property — the LangChain callback handler. Pass into config={"callbacks": [...]} on agent.invoke(...). |
last_event_id |
Property — event_id of the most recently shipped root event (or None if none has landed yet). Use this to attach a track_signal to the agent invocation you just ran without having to poll the dashboard API. |
identify(user_id, traits) |
Identify a user with optional traits. |
track_signal(event_id, name, signal_type, ...) |
Track a feedback / edit / default signal against an event. |
flush() |
Flush all pending events to the Raindrop API. |
shutdown() |
Flush remaining events and release resources. |
Async Support
The handler inherits from LangChain's BaseCallbackHandler and works with both synchronous and asynchronous Deep Agents invocations:
result = await agent.ainvoke(
{"messages": [HumanMessage(content="Hello")]},
config={"callbacks": [raindrop.handler]},
)
LangGraph Internals
Deep Agents is built on LangGraph. The handler keeps the outer LangGraph chain as the canonical root (so OTel association context is set before any span exists), and filters internal LangGraph nodes (__start__, __end__, ChannelWrite, ChannelRead, Branch:*, dunder names) so they don't pollute the dashboard.
Known Limitations
- Beta status — API surface may change in minor releases.
- Streaming — Token-by-token streaming events are not captured individually; only the final aggregated response per LLM call is tracked.
- Subagent isolation — Each subagent invoked via the
tasktool fires its own callback chain. Events are correctly grouped byconvo_id, but the parent–child trace relationship across subagents is best-effort. - OTel context propagation — The LangChain instrumentor sometimes does not propagate
traceloop.association.properties.user_idbaggage to every span produced inside a LangGraph callback context. The dashboard still shows the trace under the correct event because the SDK recordstrace_idon the root event's properties — but filtering the global Traces page byuser_idmay not surface every relevant span. Tracked upstream at traceloop/openllmetry#2271. - Concurrent invocations on a shared handler — A single
RaindropDeepAgentsinstance keeps one in-flight OTel association context (_interaction) at a time. Running multipleagent.invoke(...)/agent.ainvoke(...)calls concurrently with the same handler (e.g.await asyncio.gather(...)inside one process) will let later invocations share or overwrite the earlier one's interaction, scrambling the linkage between events and traces. Workaround: instantiate oneRaindropDeepAgentsper concurrent request (cheap — it's just a wrapper around module-level SDK state). Sequential invocations on a single instance are fine and exercised by the e2e test suite. - Long chain inputs/outputs are truncated — Chain-level input and output captured on the root event are truncated to ~8 KB to stay within the SDK's payload limits. The full message history is still visible on the per-LLM child events; only the canonical-row summary is shortened.
- Python SDK feature surface — The Raindrop Python SDK uses module-level functions and does not expose the TypeScript-only
EventShipper/TraceShipperlow-level APIs. Equivalent capability is exposed as top-level methods (identify(),track_signal()) on this class.
Testing
cd packages/deep-agents-python
pip install -e ".[dev]"
python -m pytest tests/test_handler.py -v # unit tests (no external services)
# E2E suite — exercises real Deep Agents + LLM and verifies dashboard ingestion.
# Tests skip in environments without keys (CI), so set them locally:
RAINDROP_WRITE_KEY=... \
OPENAI_API_KEY=... \
RAINDROP_DASHBOARD_TOKEN=... \
python -m pytest tests/test_e2e.py -v
The dashboard token comes from app.raindrop.ai → DevTools → Network → any backend.raindrop.ai request → Authorization: Bearer ... (expires every ~30 minutes).
See examples/deep-agents-python-basic/ for a runnable end-to-end example with a tool call, identify(), and track_signal().
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file raindrop_deep_agents-0.0.1.tar.gz.
File metadata
- Download URL: raindrop_deep_agents-0.0.1.tar.gz
- Upload date:
- Size: 53.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
baf75563dbf0ef09998818279da759d342842f4663530351480250b46e691278
|
|
| MD5 |
05ebb600fb6f6a3610b9149aa9b193a9
|
|
| BLAKE2b-256 |
74248cdd813988203c80439c90c6d3a6900661f6eaa39eceb0033134daf812b1
|
File details
Details for the file raindrop_deep_agents-0.0.1-py3-none-any.whl.
File metadata
- Download URL: raindrop_deep_agents-0.0.1-py3-none-any.whl
- Upload date:
- Size: 21.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
c73f45c12e022fc56de3d2beed0f1b936474a9189225e129ed700159b1decd6d
|
|
| MD5 |
a1661d9fc8fcfbbcb08ad5524c6c7e21
|
|
| BLAKE2b-256 |
e82bc6c398751b08e5a9f19d01d575c035e721509f3939d64959129de80c75f9
|