Skip to main content

Official Python SDK for the Claiv Memory API (V6 - Catalog Memory + Deterministic Routing)

Project description

claiv-memory

Official Python SDK for the Claiv Memory API, aligned with the current V6.3 API.

Installation

pip install claiv-memory

Quick Start

from claiv import ClaivClient

client = ClaivClient(api_key="your-api-key")

# Store a memory
result = client.ingest({
    "user_id": "user-123",
    "conversation_id": "conv-session-456",
    "type": "message",
    "content": "User prefers dark mode and uses VS Code",
})
print(result["event_id"])

# Recall relevant memory
result = client.recall({
    "user_id": "user-123",
    "conversation_id": "conv-session-456",
    "query": "Help the user configure their editor",
})
# result["llm_context"]["text"] → pre-synthesized context, inject as system prompt
# result["answer_facts"]        → [{"fact_id", "subject", "predicate", "object_text", ...}]
system_prompt = result["llm_context"]["text"] or "No memory found."

# Forget memory for a user
result = client.forget({"user_id": "user-123"})
print(f"Deleted: {result['deleted_counts']}")

Async Support

from claiv import AsyncClaivClient

async with AsyncClaivClient(api_key="your-api-key") as client:
    result = await client.ingest({
        "user_id": "user-123",
        "type": "message",
        "content": "User prefers dark mode",
    })

API Reference

ClaivClient(*, api_key, base_url, timeout, max_retries, http_client)

Parameter Type Default Description
api_key str required API key (sent as Bearer token)
base_url str https://api.claiv.io API base URL
timeout float 30.0 Request timeout in seconds
max_retries int 2 Retries on 429/5xx (0 to disable)
http_client httpx.Client None Custom httpx client

AsyncClaivClient accepts the same parameters (with httpx.AsyncClient).

Core Methods

client.ingest(request) -> IngestResponse

result = client.ingest({
    "user_id": "user-123",         # required
    "conversation_id": "conv-session-456",  # required
    "type": "message",             # required: "message" | "tool_call" | "app_event"
    "content": "The actual text",  # required
    "project_id": "project-xyz",   # optional: for project-scoped facts
    "scope": "global",             # optional: "global" | "project" | "conversation"
    "metadata": {"source": "chat"},# optional
    "event_time": "2025-01-01T00:00:00Z",  # optional: ISO 8601
    "idempotency_key": "unique-1", # optional: prevents duplicates
})
# result: {"event_id": str, "deduped": bool}

client.recall(request) -> RecallResponse

Automatically includes matching document chunks in llm_context["text"] when documents exist for the user.

Document retrieval mode is auto-detected from the query:

  • Semantic (default): top-K chunks by cosine similarity, controlled by limits["document_chunks"].
  • Working-set: triggered by structural references ("chapter 3", "the introduction") — returns the full section plus a document summary and summaries of other sections.
result = client.recall({
    "user_id": "user-123",                    # required
    "conversation_id": "conv-session-456",    # required
    "project_id": "project-xyz",              # optional: include project-scoped facts
    "document_id": "doc_uuid",                # optional: restrict to a specific document
    "query": "Help configure their editor",   # required: natural-language question
    "reference_time": None,                   # optional: ISO datetime for temporal anchoring
    "limits": {
        "answer_facts": 12,                   # default: 12
        "document_chunks": 5,                 # default: 5 — max chunks in semantic mode (1–50)
    },
})
# result["llm_context"]["text"]  → synthesized narrative, inject as system prompt
# result["answer_facts"]         → [{"fact_id", "subject", "predicate", "object_text", ...}]
# result["supporting_facts"]     → corroborating facts
# result["background_context"]   → broader context facts

Use result["llm_context"]["text"] directly as your LLM system prompt:

system_prompt = result["llm_context"]["text"] or "No memory found."

client.upload_document(request) -> DocumentUploadResponse

Upload and index a document for RAG retrieval. Chunking and embedding happen synchronously — the document is fully indexed when the response returns.

result = client.upload_document({
    "user_id": "user-123",              # required
    "content": document_text,           # required: full document text (up to 5 MB)
    "document_name": "Product Manual",  # required: shown as citation in llm_context
    "document_id": "manual-v2",         # optional: stable ID; re-uploading replaces all chunks
    "conversation_id": "conv-session-456",  # optional: scope to this conversation
    "project_id": "project-xyz",        # optional: scope to this project
    "scope": "global",                  # optional: "global" | "project" | "conversation"
    "chunk_size": 800,                  # optional: target chars per chunk (200–4000)
    "chunk_overlap": 100,               # optional: overlap between chunks (0–500)
})
# result["document_id"]      → use on recall to target this document
# result["chunks_ingested"]  → number of chunks stored
# result["chunk_ids"]        → list of chunk UUIDs

# The document is now ready — no polling needed.
# Subsequent recalls will automatically include matching chunks in llm_context["text"].

# To target recall at this document only:
recall = client.recall({
    "user_id": "user-123",
    "conversation_id": "conv-session-456",
    "query": "What does the manual say about installation?",
    "document_id": result["document_id"],
})

# To delete the document later:
client.forget({
    "user_id": "user-123",
    "document_id": result["document_id"],
})

Async version:

result = await client.upload_document({
    "user_id": "user-123",
    "content": document_text,
    "document_name": "Product Manual",
})

client.forget(request) -> ForgetResponse

result = client.forget({
    "user_id": "user-123",                     # required
    "conversation_id": "conv-session-456",     # optional
    "project_id": "project-xyz",               # optional
    "document_id": "manual-v2",                # optional: removes all chunks for this document
    "from_time": "2025-01-01T00:00:00Z",       # optional
    "to_time": "2025-06-01T00:00:00Z",         # optional
})
# result: {"receipt_id": str, "deleted_counts": {...}}

Usage Methods

summary = client.get_usage_summary("30d")    # "7d" | "30d" | "month" | "today"
breakdown = client.get_usage_breakdown("today")
limits = client.get_usage_limits()

Health Check

result = client.health_check()  # no auth required
# {"ok": True}

Error Handling

All errors inherit from ClaivError.

from claiv import ClaivApiError, ClaivTimeoutError, ClaivNetworkError

try:
    client.ingest({...})
except ClaivApiError as e:
    print(e.status)      # HTTP status code
    print(e.code)        # "invalid_request" | "unauthorized" | "quota_exceeded" | ...
    print(e.request_id)  # server request ID for support
    print(e.details)     # validation errors, quota info, etc.
except ClaivTimeoutError:
    pass  # request timed out
except ClaivNetworkError:
    pass  # DNS failure, connection refused, etc.

Retries

The SDK automatically retries on 429 (rate limited) and 5xx (server error) responses with exponential backoff and jitter. Client errors (400, 401, 403, 404) are never retried.

# Default: 2 retries (3 total attempts)
client = ClaivClient(api_key="key")

# Disable retries
client = ClaivClient(api_key="key", max_retries=0)

# More retries for critical paths
client = ClaivClient(api_key="key", max_retries=5)

Context Manager

Both clients support context managers for automatic cleanup:

with ClaivClient(api_key="key") as client:
    client.ingest({...})

async with AsyncClaivClient(api_key="key") as client:
    await client.ingest({...})

Type Hints

All request/response types are exported as TypedDicts:

from claiv import (
    IngestRequest, IngestResponse,
    RecallRequest, RecallResponse, RecallFact, V6LLMContext, ContextPack,
    ForgetRequest, ForgetResponse, DeletedCounts,
    DocumentUploadRequest, DocumentUploadResponse,
    UsageSummaryResponse, UsageBreakdownResponse, UsageLimitsResponse,
)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

claiv_memory-0.6.4.tar.gz (18.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

claiv_memory-0.6.4-py3-none-any.whl (15.5 kB view details)

Uploaded Python 3

File details

Details for the file claiv_memory-0.6.4.tar.gz.

File metadata

  • Download URL: claiv_memory-0.6.4.tar.gz
  • Upload date:
  • Size: 18.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.5

File hashes

Hashes for claiv_memory-0.6.4.tar.gz
Algorithm Hash digest
SHA256 765e2728e1da537982a0634afcd734b7b36c9d88925272bef01d97778acf2a3b
MD5 e7b32bec7aee7a66d18de93b879cfc6f
BLAKE2b-256 202bf99ed83789112250546e452e847a1c31a2f310df4f0a75f17c717ccdb5bf

See more details on using hashes here.

File details

Details for the file claiv_memory-0.6.4-py3-none-any.whl.

File metadata

  • Download URL: claiv_memory-0.6.4-py3-none-any.whl
  • Upload date:
  • Size: 15.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.5

File hashes

Hashes for claiv_memory-0.6.4-py3-none-any.whl
Algorithm Hash digest
SHA256 739ac052f68d9a6c614f85c8a4c14202b3a79686be86b341e60abe5951d6e9dd
MD5 02f3f7bfd69c3da81887df485283dc16
BLAKE2b-256 5b8d523c42e533a196a53ff010e6c9d66f21757af7a7ea1edccbb92edbe879ae

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page