Skip to main content

Real-time behavioral drift detection for AI systems

Project description

Drift//Monitor

Real-time behavioral drift detection for AI systems.

License: MIT Python 3.11+ Tests


The Problem

LLM outputs degrade over long sessions. Tone shifts. Goals wander. Instructions erode. Content diverges from source material. This happens silently -- no existing tool monitors for it in real time.

Drift//Monitor is the andon cord for language models. It sits as a transparent proxy between your application and the LLM API, measuring output deviation from an established baseline using embedding distance and statistical process control.

Quick Start

pip install driftmonitor
import anthropic
from driftmonitor import DriftAwareClient, SessionConfig

# Wrap your existing Anthropic client -- zero code changes needed
base = anthropic.Anthropic()
client = DriftAwareClient(base, config=SessionConfig(name="My Session"))

# Use exactly like the base client
response = client.messages.create(
    model="claude-sonnet-4-20250514",
    max_tokens=1024,
    messages=[{"role": "user", "content": "Hello"}],
)
# Response passes through unchanged. Drift scores logged to SQLite.

# Check drift
print(client.latest_score)        # DriftScore(persona=5.2, goal=3.1, ...)
print(client.latest_severity)     # Severity.NOMINAL
print(client.composite_history)   # [2.1, 3.4, 5.2, ...]

How It Works

Three-Layer Architecture

Layer Function Technology
Capture Intercept every LLM input/output via proxy SDK Drop-in Python client wrapper
Score Embed + measure statistical deviation from anchor sentence-transformers, SPC control charts
Act Alert, correct, gate, and log drift events FastAPI, SQLite, React dashboard

Four Drift Axes

Axis Measures Example Signal
Persona / Tone Voice, formality, character consistency Technical writer starts editorializing
Goal Alignment Adherence to original objective Code agent starts refactoring instead of fixing
Instruction Follow Respect for earlier constraints Agent told not to delete files begins deleting
Semantic Fidelity Alignment with source material Summary contradicts source document

Severity Scale (SPC-based)

Level Threshold Response
NOMINAL < mean + 1σ Log only
LOW ≥ mean + 1σ Dashboard flag
ELEVATED ≥ mean + 2σ Operator notification, correction generated
CRITICAL ≥ mean + 3σ Andon cord -- gate or halt

Operational Modes

from driftmonitor import DriftAwareClient, SessionConfig

# OBSERVE -- log everything, no intervention (default)
client = DriftAwareClient(base, config=SessionConfig(mode="observe"))

# ALERT -- notify operator on threshold breach
client = DriftAwareClient(base, config=SessionConfig(mode="alert"))

# GATE -- hold response until operator approves
client = DriftAwareClient(base, config=SessionConfig(mode="gate"))
response = client.messages.create(...)
if client.is_gated:
    client.gate_approve(operator="admin", reason="Acceptable drift")
    # or: client.gate_reject(operator="admin", reason="Too much drift")

REST API

# Install server dependencies
pip install driftmonitor[server]

# Launch the API
driftmonitor serve --port 8000

# Or with a custom database
driftmonitor serve --db /path/to/drift.db

Endpoints

Method Path Description
GET /health Service status
GET /sessions List all sessions
GET /sessions/{id} Session details
DELETE /sessions/{id} Delete session
GET /sessions/{id}/events Drift events (filterable)
GET /sessions/{id}/latest Most recent event
GET /sessions/{id}/summary Aggregate statistics
GET /audit Audit log entries
GET /audit/export/json Export audit as JSON
GET /audit/export/csv Export audit as CSV
WS /ws/{session_id} Live score streaming

CLI

# List sessions
driftmonitor sessions --db drift.db

# View scores for a session
driftmonitor scores <session-id> --limit 20

# Session summary
driftmonitor summary <session-id>

# Launch API server
driftmonitor serve --port 8000

# Version
driftmonitor version

Audit Log

Every drift event, mode change, gate decision, and correction is recorded in an immutable audit log with timestamps and operator attribution.

# Export for compliance
audit_json = client.audit.export_json(session_id="sess-001")
audit_csv = client.audit.export_csv(session_id="sess-001")

# Query entries
entries = client.audit.get_entries(action="gate_released", limit=50)

Custom Anchor

By default, the first assistant response becomes the drift baseline. You can provide a reference document instead:

client.set_anchor("Your reference text or system prompt here.")

Architecture

Your App
    |
    v
DriftAwareClient  ───>  EmbeddingEngine  ───>  SPCEngine
    |                    (sentence-transformers)  (control limits)
    |                         |                       |
    v                         v                       v
Anthropic API            DriftStore              DriftExplainer
    |                    (SQLite WAL)            (Claude Haiku)
    v                         |                       |
Response                      v                       v
(unchanged)              Audit Log              Corrections

Competitive Landscape

Category Tools Gap
LLM Observability LangSmith, Helicone, Arize Log tokens and latency. No behavioral drift scoring.
AI Testing PromptFoo, Braintrust Snapshot evaluation. Not live longitudinal monitoring.
AI Guardrails Guardrails AI, NeMo Input/output filters. Not drift detection over time.
AI Governance Various No operator-controlled correction with audit trail.

Development

# Clone and install
git clone https://github.com/AreteDriver/drift-monitor.git
cd drift-monitor
python -m venv .venv && source .venv/bin/activate
pip install -e ".[server,dev]"

# Run tests
pytest tests/ -v

# Lint
ruff check src/ tests/
ruff format src/ tests/

License

MIT


Built by AreteDriver -- standard work for AI behavior.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

driftmonitor_ai-0.1.0.tar.gz (157.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

driftmonitor_ai-0.1.0-py3-none-any.whl (32.8 kB view details)

Uploaded Python 3

File details

Details for the file driftmonitor_ai-0.1.0.tar.gz.

File metadata

  • Download URL: driftmonitor_ai-0.1.0.tar.gz
  • Upload date:
  • Size: 157.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for driftmonitor_ai-0.1.0.tar.gz
Algorithm Hash digest
SHA256 3ec9ded517759d26a13874437c27d5641cfd5ef2d004b8e1a67fdcd476d9add3
MD5 40b8f8b7542f9005c076450d85e68737
BLAKE2b-256 51aac23ab3a6cd3445fb13c220a0883f726d390498a98429bcae6836c08ac3d0

See more details on using hashes here.

Provenance

The following attestation bundles were made for driftmonitor_ai-0.1.0.tar.gz:

Publisher: publish.yml on AreteDriver/drift-monitor

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file driftmonitor_ai-0.1.0-py3-none-any.whl.

File metadata

File hashes

Hashes for driftmonitor_ai-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 72c19bdbe3fc9aecf8691f2dab29c4d72c14a2df631a476b8e928e65e21bce68
MD5 caf0f35f63edf49e9173935ecc226479
BLAKE2b-256 c4f275b295ad7f096780392ce82f85ba112528c71d188593df5f33b0c3de0df5

See more details on using hashes here.

Provenance

The following attestation bundles were made for driftmonitor_ai-0.1.0-py3-none-any.whl:

Publisher: publish.yml on AreteDriver/drift-monitor

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page