Skip to main content

Python SDK for Kagura Memory Cloud — AI-driven memory management

Project description

Kagura Ai
Memory SDK — Python client for Kagura Memory Cloud

PyPI version Python versions CI codecov License: MIT MCP Checked with pyright

What is this?

This SDK connects your Python code to Kagura Memory Cloud, giving AI assistants the ability to remember, search, and learn from past interactions. It provides three clients for different use cases:

Client Protocol Use Case
KaguraAgent MCP + LLM AI-powered — auto-decides what to remember/recall from conversations
KaguraClient MCP (JSON-RPC) Direct memory ops — remember, recall, explore, reference, forget
ResourceClient REST API External data ingestion — push data from Slack, CI/CD, CRM into Kagura

Installation

pip install kagura-memory
# or
uv add kagura-memory

Quick Start

Configuration

Copy the example and fill in your credentials:

cp .kagura.json.example .kagura.json
# Edit .kagura.json — set api_key and mcp_url

Used by the CLI (kagura commands) and load_config() in Python code:

{
  "api_key": "kagura_your_api_key",
  "mcp_url": "http://localhost:8080/mcp/w/{workspace_id}",
  "model": "gpt-5.4-nano",
  "context_id": "auto"
}

Or use environment variables: KAGURA_API_KEY, KAGURA_MCP_URL, KAGURA_MODEL, KAGURA_CONTEXT_ID

Get your API key from the Kagura Memory Cloud Web UI: Integrations > API Keys

KaguraAgent — AI-Powered Memory

Let the AI analyze conversations and automatically decide what to remember and recall:

from kagura_memory import KaguraAgent, Session, Message

agent = KaguraAgent(api_key="kagura_...", model="gpt-5.4-nano")

session = Session(messages=[
    Message(role="user", content="FastAPIでOAuth2を実装したい"),
    Message(role="assistant", content="Authlibを使うパターンが推奨です..."),
    Message(role="user", content="なるほど、これ覚えておいて"),
])

async with agent:
    result = await agent.process(session, deep=True, verbose=2)
    print(f"Remembered: {len(result.remembered)}, Recalled: {len(result.recalled)}")

Supports OpenAI, Claude, Gemini via LiteLLM, and Ollama for local models:

# Local LLM via Ollama (no cloud API key needed)
agent = KaguraAgent(api_key="kagura_...", model="ollama/qwen3:30b")

Ollama Local Model Requirements

Model Size Context Min VRAM Recommended GPU
qwen3:30b (recommended) 19 GB 256K 24 GB RTX 4090 or equivalent
qwen3:14b 9.3 GB 40K 16 GB RTX 4080 or equivalent

Recommended minimum: qwen3:30b on an RTX 4090 (24 GB VRAM) or equivalent.

Smaller models (< 30B parameters) may produce lower quality memory analysis — summaries may lack searchable keywords, and recall query generation may be less precise.

KaguraClient — Direct Memory Operations

For programmatic control without LLM:

from kagura_memory import KaguraClient

async with KaguraClient(api_key="kagura_...", mcp_url="https://...") as client:
    await client.remember(context_id="dev", summary="OAuth2 pattern", content="Use Authlib...")
    results = await client.recall(context_id="dev", query="OAuth2", k=5)
    await client.explore(context_id="dev", memory_id="uuid", depth=3)

    # Tag AND filter — match memories with ALL specified tags
    results = await client.recall(
        context_id="dev", query="budget",
        filters={"tags": ["予算", "2026"], "tags_match": "all"},
    )

    # Date range filter
    results = await client.recall(
        context_id="dev", query="recent decisions",
        filters={"created_after": "2026-03-01T00:00:00Z", "created_before": "2026-03-31T23:59:59Z"},
    )

    # Cross-context recall — search across multiple contexts at once
    results = await client.recall(
        query="authentication",
        context_ids=["ctx-uuid-1", "ctx-uuid-2"], k=10,
    )

    # Merge contexts — copy all memories from source to target
    result = await client.merge_contexts(source_id="old-ctx", target_id="new-ctx")
    print(f"Merged {result['merged']} memories")

    # Merge and delete the source context
    result = await client.merge_contexts(
        source_id="old-ctx", target_id="new-ctx", delete_source=True,
    )

    # Workspace usage — check quota limits
    usage = await client.get_usage()
    print(f"Plan: {usage.plan}, Memories: {usage.memories.used}/{usage.memories.limit}")

    # Context info — includes search_config
    info = await client.get_context_info(context_id="dev")
    print(f"Search config: {info.context.search_config}")

    # Embedding status — check for failures
    status = await client.get_embedding_status()
    print(f"Embeddings: {status.total}, Failed: {len(status.failed_memories)}")

    # Per-memory stats — recall frequency, access patterns
    stats = await client.get_memory_stats(context_id="dev", sort_by="use_count", limit=10)

    # Duplicate detection — find similar memory pairs
    dupes = await client.find_duplicates(context_id="dev", threshold=0.90)
    print(f"Found {dupes.total_pairs} duplicate pairs")

ResourceClient — External Data Ingestion

Push data from external systems into Kagura so AI can search it:

from kagura_memory import ResourceClient, ResourceEventRequest

async with ResourceClient.from_mcp_url(api_key="kagura_...", mcp_url="http://localhost:8080/mcp/w/...") as client:
    # One-call setup: create public context + set resource_id + create token
    token = await client.setup_resource(resource_id="products", summary="Product catalog")
    print(f"Save this token: {token.token}")  # Shown only once!

    event = ResourceEventRequest(
        op="upsert", doc_id="SKU-001", version=1,
        payload={"name": "Wireless Headphones", "price": 79.99},
    )
    await client.ingest_event("products", token.token, event)

    # Check ingestion stats
    stats = await client.get_resource_impact("products")
    print(f"Memories: {stats.memory_count}, Tokens: {stats.token_count}")

See examples/ for complete working examples.

CLI

# AI-powered (requires LLM API key)
kagura process -m "Remember: FastAPI uses Depends() for DI"

# Direct memory operations
kagura remember -s "FastAPI DI" --content "Use Depends()..." -c dev
kagura recall "dependency injection" -k 10
kagura explore -m "memory-uuid" --depth 3
kagura forget -m "memory-uuid"
kagura contexts

# Resource tokens
kagura resource tokens create -r products -d "Product sync"
kagura resource ingest -r products -k TOKEN --doc-id SKU-001 -V 1 -p '{"name":"Widget"}'
kagura resource ingest-batch -r products -k TOKEN -f events.json
kagura resource stats -r products
kagura resource schema -r products

# Sleep Maintenance — observability + rollback
kagura sleep history <context-id> --limit 5
kagura sleep report <context-id> <report-id>
kagura sleep rollback <context-id> <report-id> -y    # destructive: prompts unless --yes / -y is set

# Config
kagura config show

Claude Code Integration

Use Kagura Memory as an MCP server in Claude Code:

cp .mcp.json.example .mcp.json
# Edit .mcp.json — set workspace_id and API key

Or use the CLI directly:

kagura process -m "今日の学び:FastAPIのDIはDepends()を使う"

API Coverage

Operation SDK Client Protocol Auth
Memory (remember/recall/forget/explore/reference) KaguraClient MCP API Key
Context (create/update/list/get_context_info) KaguraClient MCP API Key
Workspace (get_usage) KaguraClient MCP API Key
Search config (update_search_config) KaguraClient MCP API Key
Embedding status (get_embedding_status) KaguraClient REST API Key
Memory stats (get_memory_stats) KaguraClient REST API Key
Duplicate detection (find_duplicates) KaguraClient REST API Key
Context delete Web UI only Session
Sleep Maintenance (history / report / rollback) KaguraClient MCP API Key
Resource Token (create/list/update/revoke) ResourceClient REST API API Key
Resource Event ingestion ResourceClient REST API Resource Token
Resource Impact (stats) ResourceClient REST API API Key
Resource Schema ResourceClient REST API API Key
Account erasure (GDPR Art.17 / APPI) Web UI only Session

Context deletion and account erasure are intentionally Web UI only — destructive operations require session authentication and confirmation. kagura sleep rollback runs over the MCP API Key but is itself destructive (reverses edge creation, merges, importance updates, promotions, and archives) and the CLI requires --yes to skip the interactive confirmation. The server commits per-action without a Saga, so a 5xx response after partial success means SOME actions may have been reversed before the error surfaced — re-run kagura sleep report to inspect the post-failure state.

Development

git clone https://github.com/kagura-ai/kagura-memory-python-sdk.git
cd kagura-memory-python-sdk
uv sync --dev
uv run ruff check src/ tests/   # Lint
uv run ruff format src/ tests/  # Format
uv run pyright src/              # Type check
uv run pytest tests/ -v          # Test

Development with Claude Code

This project is developed with Claude Code:

/onboarding      # Interactive setup — verify config, test connection
/workflow        # Check current state and next step
/quality         # Run all quality checks
/simplify        # Review for reuse, quality, efficiency
/self-review     # Pre-PR self-review
/self-maint      # Audit .claude/ config against codebase
/release <level> # Bump version, tag, push, create GitHub Release
/kagura-guide    # SDK usage reference

Typical flow: Issue → Branch → Implement → /quality/simplify/self-review → PR → Merge → /release

Links

License

MIT License — see LICENSE for details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

kagura_memory-0.13.0.tar.gz (371.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

kagura_memory-0.13.0-py3-none-any.whl (54.7 kB view details)

Uploaded Python 3

File details

Details for the file kagura_memory-0.13.0.tar.gz.

File metadata

  • Download URL: kagura_memory-0.13.0.tar.gz
  • Upload date:
  • Size: 371.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for kagura_memory-0.13.0.tar.gz
Algorithm Hash digest
SHA256 79d51dddfee954b5bc38738c8549ef8ed1919ecc5b6d9ab13f7b233bc08de33a
MD5 8f5eb7ec4bb08d331f390e31b5601b86
BLAKE2b-256 687e56202af8eff00c5037602c9bc1682251a9a0ee93839eb08db7e3698c55bb

See more details on using hashes here.

Provenance

The following attestation bundles were made for kagura_memory-0.13.0.tar.gz:

Publisher: publish.yml on kagura-ai/kagura-memory-python-sdk

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file kagura_memory-0.13.0-py3-none-any.whl.

File metadata

  • Download URL: kagura_memory-0.13.0-py3-none-any.whl
  • Upload date:
  • Size: 54.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for kagura_memory-0.13.0-py3-none-any.whl
Algorithm Hash digest
SHA256 b9b53e5ba907ea43f70e7867827461c7da2add983d4c6a53d8b76d876cd981dc
MD5 7aac7b44e17bd969847dce1d429f34c8
BLAKE2b-256 1083d7ecd813c12a2335e665b2a70180e85a4b46944bb82d98ce3b8e190b37f8

See more details on using hashes here.

Provenance

The following attestation bundles were made for kagura_memory-0.13.0-py3-none-any.whl:

Publisher: publish.yml on kagura-ai/kagura-memory-python-sdk

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page