Official Python SDK for the Claiv Memory API (V6 - Catalog Memory + Deterministic Routing)
Project description
claiv-memory
Official Python SDK for the Claiv Memory API, aligned with the current V6.3 API.
Installation
pip install claiv-memory
Quick Start
from claiv import ClaivClient
client = ClaivClient(api_key="your-api-key")
# Store a memory
result = client.ingest({
"user_id": "user-123",
"conversation_id": "conv-session-456",
"type": "message",
"content": "User prefers dark mode and uses VS Code",
})
print(result["event_id"])
# Recall relevant memory
result = client.recall({
"user_id": "user-123",
"conversation_id": "conv-session-456",
"query": "Help the user configure their editor",
})
# result["llm_context"]["text"] → pre-synthesized context, inject as system prompt
# result["answer_facts"] → [{"fact_id", "subject", "predicate", "object_text", ...}]
system_prompt = result["llm_context"]["text"] or "No memory found."
# Forget memory for a user
result = client.forget({"user_id": "user-123"})
print(f"Deleted: {result['deleted_counts']}")
Async Support
from claiv import AsyncClaivClient
async with AsyncClaivClient(api_key="your-api-key") as client:
result = await client.ingest({
"user_id": "user-123",
"type": "message",
"content": "User prefers dark mode",
})
API Reference
ClaivClient(*, api_key, base_url, timeout, max_retries, http_client)
| Parameter | Type | Default | Description |
|---|---|---|---|
api_key |
str |
required | API key (sent as Bearer token) |
base_url |
str |
https://api.claiv.io |
API base URL |
timeout |
float |
30.0 |
Request timeout in seconds |
max_retries |
int |
2 |
Retries on 429/5xx (0 to disable) |
http_client |
httpx.Client |
None |
Custom httpx client |
AsyncClaivClient accepts the same parameters (with httpx.AsyncClient).
Core Methods
client.ingest(request) -> IngestResponse
result = client.ingest({
"user_id": "user-123", # required
"conversation_id": "conv-session-456", # required
"type": "message", # required: "message" | "tool_call" | "app_event"
"content": "The actual text", # required
"project_id": "project-xyz", # optional: for project-scoped facts
"scope": "global", # optional: "global" | "project" | "conversation"
"metadata": {"source": "chat"},# optional
"event_time": "2025-01-01T00:00:00Z", # optional: ISO 8601
"idempotency_key": "unique-1", # optional: prevents duplicates
})
# result: {"event_id": str, "deduped": bool}
client.recall(request) -> RecallResponse
Automatically includes matching document chunks in llm_context["text"] when documents exist for the user.
Document retrieval mode is auto-detected from the query:
- Semantic (default): top-K chunks by cosine similarity, controlled by
limits["document_chunks"]. - Working-set: triggered by structural references (
"chapter 3","the introduction") — returns the full section plus a document summary and summaries of other sections.
result = client.recall({
"user_id": "user-123", # required
"conversation_id": "conv-session-456", # required
"project_id": "project-xyz", # optional: include project-scoped facts
"document_id": "doc_uuid", # optional: restrict to a specific document
"query": "Help configure their editor", # required: natural-language question
"reference_time": None, # optional: ISO datetime for temporal anchoring
"limits": {
"answer_facts": 12, # default: 12
"document_chunks": 5, # default: 5 — max chunks in semantic mode (1–50)
},
})
# result["llm_context"]["text"] → synthesized narrative, inject as system prompt
# result["answer_facts"] → [{"fact_id", "subject", "predicate", "object_text", ...}]
# result["supporting_facts"] → corroborating facts
# result["background_context"] → broader context facts
Use result["llm_context"]["text"] directly as your LLM system prompt:
system_prompt = result["llm_context"]["text"] or "No memory found."
client.upload_document(request) -> DocumentUploadResponse
Upload and index a document for RAG retrieval. Chunking and embedding happen synchronously — the document is fully indexed when the response returns.
result = client.upload_document({
"user_id": "user-123", # required
"content": document_text, # required: full document text (up to 5 MB)
"document_name": "Product Manual", # required: shown as citation in llm_context
"document_id": "manual-v2", # optional: stable ID; re-uploading replaces all chunks
"conversation_id": "conv-session-456", # optional: scope to this conversation
"project_id": "project-xyz", # optional: scope to this project
"scope": "global", # optional: "global" | "project" | "conversation"
"chunk_size": 800, # optional: target chars per chunk (200–4000)
"chunk_overlap": 100, # optional: overlap between chunks (0–500)
})
# result["document_id"] → use on recall to target this document
# result["chunks_ingested"] → number of chunks stored
# result["chunk_ids"] → list of chunk UUIDs
# The document is now ready — no polling needed.
# Subsequent recalls will automatically include matching chunks in llm_context["text"].
# To target recall at this document only:
recall = client.recall({
"user_id": "user-123",
"conversation_id": "conv-session-456",
"query": "What does the manual say about installation?",
"document_id": result["document_id"],
})
# To delete the document later:
client.forget({
"user_id": "user-123",
"document_id": result["document_id"],
})
Async version:
result = await client.upload_document({
"user_id": "user-123",
"content": document_text,
"document_name": "Product Manual",
})
client.forget(request) -> ForgetResponse
result = client.forget({
"user_id": "user-123", # required
"conversation_id": "conv-session-456", # optional
"project_id": "project-xyz", # optional
"document_id": "manual-v2", # optional: removes all chunks for this document
"from_time": "2025-01-01T00:00:00Z", # optional
"to_time": "2025-06-01T00:00:00Z", # optional
})
# result: {"receipt_id": str, "deleted_counts": {...}}
Usage Methods
summary = client.get_usage_summary("30d") # "7d" | "30d" | "month" | "today"
breakdown = client.get_usage_breakdown("today")
limits = client.get_usage_limits()
Health Check
result = client.health_check() # no auth required
# {"ok": True}
Error Handling
All errors inherit from ClaivError.
from claiv import ClaivApiError, ClaivTimeoutError, ClaivNetworkError
try:
client.ingest({...})
except ClaivApiError as e:
print(e.status) # HTTP status code
print(e.code) # "invalid_request" | "unauthorized" | "quota_exceeded" | ...
print(e.request_id) # server request ID for support
print(e.details) # validation errors, quota info, etc.
except ClaivTimeoutError:
pass # request timed out
except ClaivNetworkError:
pass # DNS failure, connection refused, etc.
Retries
The SDK automatically retries on 429 (rate limited) and 5xx (server error) responses with exponential backoff and jitter. Client errors (400, 401, 403, 404) are never retried.
# Default: 2 retries (3 total attempts)
client = ClaivClient(api_key="key")
# Disable retries
client = ClaivClient(api_key="key", max_retries=0)
# More retries for critical paths
client = ClaivClient(api_key="key", max_retries=5)
Context Manager
Both clients support context managers for automatic cleanup:
with ClaivClient(api_key="key") as client:
client.ingest({...})
async with AsyncClaivClient(api_key="key") as client:
await client.ingest({...})
Type Hints
All request/response types are exported as TypedDicts:
from claiv import (
IngestRequest, IngestResponse,
RecallRequest, RecallResponse, RecallFact, V6LLMContext, ContextPack,
ForgetRequest, ForgetResponse, DeletedCounts,
DocumentUploadRequest, DocumentUploadResponse,
UsageSummaryResponse, UsageBreakdownResponse, UsageLimitsResponse,
)
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file claiv_memory-0.6.4.tar.gz.
File metadata
- Download URL: claiv_memory-0.6.4.tar.gz
- Upload date:
- Size: 18.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.5
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
765e2728e1da537982a0634afcd734b7b36c9d88925272bef01d97778acf2a3b
|
|
| MD5 |
e7b32bec7aee7a66d18de93b879cfc6f
|
|
| BLAKE2b-256 |
202bf99ed83789112250546e452e847a1c31a2f310df4f0a75f17c717ccdb5bf
|
File details
Details for the file claiv_memory-0.6.4-py3-none-any.whl.
File metadata
- Download URL: claiv_memory-0.6.4-py3-none-any.whl
- Upload date:
- Size: 15.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.5
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
739ac052f68d9a6c614f85c8a4c14202b3a79686be86b341e60abe5951d6e9dd
|
|
| MD5 |
02f3f7bfd69c3da81887df485283dc16
|
|
| BLAKE2b-256 |
5b8d523c42e533a196a53ff010e6c9d66f21757af7a7ea1edccbb92edbe879ae
|