Skip to main content

The First Database Built for Agents to Own, Not Just Query

Project description

KameleonDB Logo

PyPI version Python 3.11+ License: Apache 2.0

KameleonDB

Find the true color of your data.

The First Database Built for Agents to Own, Not Just Query

Most databases let agents query data that humans structured. KameleonDB goes further: agents own the entire data lifecycle—from schema design to data ingestion to continuous evolution. You provide the goals and policies, agents build and manage the database.

Built on PostgreSQL (JSONB) or SQLite (JSON1) with schema-as-data storage, agents can restructure information on the fly without migrations, DDL, or human intervention.

Philosophy: Agents as Data Engineers

In traditional databases, humans are the data engineers: they design schemas, write migrations, and structure data for agents to query.

KameleonDB makes agents the data engineers. Agents don't just consume data—they design the schema, ingest records, evolve structure, and reshape information as they reason about it. Humans shift from data architects to policy makers, defining what agents can do, not how to structure every field.

This is schema-on-reason: structure emerges from agent reasoning, not upfront human design. As agents learn more about the data, they adapt the schema to match their understanding.

First Principles:

  1. Radical Simplicity — Perfection achieved by removing, not adding
  2. Agent-First Design — APIs optimized for agent reasoning patterns
  3. Schema-on-Reason — Schema emerges from reasoning, not upfront design
  4. Provenance & Auditability — Every decision traceable
  5. Policy-Driven Governance — Autonomy bounded by declarative policies
  6. Security by Design — Zero-trust architecture
  7. Enterprise-Grade Reliability — ACID guarantees and multi-tenancy

See FIRST-PRINCIPLES.md for detailed explanations and AGENTS.md for the complete agent-native design philosophy.

Features

  • Dynamic Schema: Create and modify entity fields at runtime without migrations
  • Multi-Database: PostgreSQL (JSONB) and SQLite (JSON1) support
  • Agent-First Design: Every operation is a tool for AI agents with JSON-serializable I/O
  • Self-Describing: Agents can discover schema before querying
  • Idempotent Operations: Safe for agents to call repeatedly
  • Audit Trail: Track who made schema changes and why
  • Zero-Lock Evolution: Schema changes are metadata-only, no table locks

Installation

# Core only (SQLite works out of the box)
pip install kameleondb

# With PostgreSQL support
pip install kameleondb[postgresql]

# With MCP server
pip install kameleondb[mcp]

# For development
pip install kameleondb[dev,postgresql]

# Everything
pip install kameleondb[all]

Database Requirements:

  • SQLite: 3.9+ with JSON1 extension (included in Python stdlib)
  • PostgreSQL: 12+ with JSONB support

Quick Start

Option 1: MCP Server (Recommended for AI Agents)

The MCP (Model Context Protocol) server exposes KameleonDB as tools that AI agents can use directly.

Installation:

pip install kameleondb[mcp]

Start the MCP server:

# PostgreSQL
kameleondb-mcp --database postgresql://user:pass@localhost/kameleondb

# SQLite (for development)
kameleondb-mcp --database sqlite:///./kameleondb.db

Configure in Claude Desktop (~/Library/Application Support/Claude/claude_desktop_config.json):

{
  "mcpServers": {
    "kameleondb": {
      "command": "kameleondb-mcp",
      "args": ["--database", "postgresql://localhost/kameleondb"]
    }
  }
}

Available MCP Tools:

  • kameleondb_describe() - Discover database schema
  • kameleondb_create_entity() - Create new entity types
  • kameleondb_insert() - Add records
  • kameleondb_execute_sql() - Query with LLM-generated SQL
  • kameleondb_materialize_entity() - Optimize storage for performance
  • ...and 20+ more tools

See MCP Documentation for client setup.

Option 2: Command-Line Interface

Installation:

pip install kameleondb

Initialize and create your first entity:

# Initialize database
kameleondb init

# Create an entity
kameleondb schema create Contact \
  --field "name:string:required" \
  --field "email:string:unique" \
  --field "phone:string"

# Insert data (inline JSON)
kameleondb data insert Contact '{"name": "Alice", "email": "alice@example.com"}'

# Insert from file
kameleondb data insert Contact --from-file contact.json

# List records
kameleondb data list Contact

# Query with SQL
kameleondb query run "SELECT * FROM kdb_records WHERE entity_id='...' LIMIT 10"

JSON output for scripting:

kameleondb --json schema list | jq .
kameleondb --json data insert Contact '{"name":"Bob","email":"bob@example.com"}'

Available Commands:

  • schema - Create, list, describe, modify entities
  • data - Insert, get, update, delete, list records
  • query - Execute and validate SQL
  • storage - Materialize entities, check storage mode
  • admin - Initialize, info, changelog

See kameleondb --help for full command reference.

Option 3: Python API Integration

For developers integrating KameleonDB into Python applications:

from kameleondb import KameleonDB

# Initialize with PostgreSQL
db = KameleonDB("postgresql://user:pass@localhost/kameleondb")

# Or use SQLite for development/testing
# db = KameleonDB("sqlite:///./kameleondb.db")

# Create an entity with fields
contacts = db.create_entity(
    name="Contact",
    fields=[
        {"name": "first_name", "type": "string", "required": True},
        {"name": "email", "type": "string", "unique": True},
    ],
    created_by="my-agent",
    if_not_exists=True,  # Idempotent - safe to call multiple times
)

# Add a field later (with reasoning for audit)
contacts.add_field(
    name="linkedin_url",
    field_type="string",
    created_by="enrichment-agent",
    reason="Found LinkedIn profiles in documents",
    if_not_exists=True,
)

# Insert data
contact_id = contacts.insert({
    "first_name": "John",
    "email": "john@example.com",
})

# Retrieve by ID
contact = contacts.find_by_id(contact_id)
print(contact)  # {"id": "...", "first_name": "John", "email": "john@example.com", ...}

# For complex queries, use SQL generation via schema context
context = db.get_schema_context()
# Use context with an LLM to generate SQL, then:
# results = db.execute_sql("SELECT ... FROM kdb_records WHERE ...")

# Discover schema (agents call this first)
schema = db.describe()
print(schema)
# {
#     "entities": {
#         "Contact": {
#             "fields": ["first_name", "email", "linkedin_url"],
#             ...
#         }
#     }
# }

Tool Integration:

# Get all operations as tools for AI agents
tools = db.get_tools()

# Each tool has:
# - name: "kameleondb_create_entity"
# - description: Human-readable description
# - parameters: JSON Schema for inputs
# - function: Callable to execute

See docs/ARCHITECTURE.md for technical details.

Development

# Clone the repository
git clone https://github.com/marcosnataqs/kameleondb.git
cd kameleondb

# Install with dev dependencies
pip install -e ".[dev]"

# Run tests
pytest

# Run linting
ruff check src tests
mypy src/kameleondb

# Run pre-commit hooks
pre-commit install
pre-commit run --all-files

Roadmap

  • v0.1: Core schema engine ✅
  • v0.2: Relationships + Hybrid Storage + Query Intelligence ✅
    • Relationship metadata (many-to-one, one-to-many, many-to-many)
    • Schema context for SQL generation
    • Query validation and execution
    • SQLite support
    • Hybrid storage (shared/dedicated modes)
    • Storage migration (materialize/dematerialize)
    • Query metrics and materialization suggestions
  • v0.3: Relational queries + Many-to-many (planned)
    • Cross-entity queries with JOINs
    • Cascading operations
    • Many-to-many junction tables
  • v0.4: Natural language queries (planned)
    • LLM-powered query generation
    • Query caching and optimization

See docs/tasks/BACKLOG.md for detailed roadmap.

License

Apache 2.0 License - see LICENSE for details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

kameleondb-0.1.1.tar.gz (1.1 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

kameleondb-0.1.1-py3-none-any.whl (84.8 kB view details)

Uploaded Python 3

File details

Details for the file kameleondb-0.1.1.tar.gz.

File metadata

  • Download URL: kameleondb-0.1.1.tar.gz
  • Upload date:
  • Size: 1.1 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for kameleondb-0.1.1.tar.gz
Algorithm Hash digest
SHA256 1c5989f00308d0eeb4201f760997d492d081f74a737ff2dc5ce3031dd934a9f8
MD5 2a473ea243959b36d640a447e188bd8c
BLAKE2b-256 268b1026474e3c9b9eb99fd1abad7df21607f4ec0bb0a8ad98022e83f346fdf8

See more details on using hashes here.

Provenance

The following attestation bundles were made for kameleondb-0.1.1.tar.gz:

Publisher: publish.yml on marcosnataqs/kameleondb

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file kameleondb-0.1.1-py3-none-any.whl.

File metadata

  • Download URL: kameleondb-0.1.1-py3-none-any.whl
  • Upload date:
  • Size: 84.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for kameleondb-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 6a970323356a402bb5d9b3ba97d2f081b5704bfa94a38a5f0bb231755109d64d
MD5 6bc9d88a322170de3c3d886634b96803
BLAKE2b-256 2345b59be14baa79ff71f3ba8dc7d96930ee7aa1f1171ec5d64706a45fa9b61b

See more details on using hashes here.

Provenance

The following attestation bundles were made for kameleondb-0.1.1-py3-none-any.whl:

Publisher: publish.yml on marcosnataqs/kameleondb

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page