Skip to main content

VECTOR: N-dimensional coordinate database with O(1) lookups

Project description

VECTOR Logo

VECTOR - VECTOR Encodes Coordinates To Optimize Retrieval

A lightweight vector database library for Python that organizes data using mathematical coordinate systems. Built with domain-driven architecture and designed for single-file storage with O(1) lookup performance.

Project Philosophy

Vector embraces the "coordinate-based data organization" approach where every table must have an X-axis as the primary key, with other attributes representing relationships between dimensions. This creates a natural mathematical model for data organization:

  • X-axis (Central Axis): Primary key and coordinate system foundation
  • Y, Z, J... (Dimensional Spaces): Additional attributes that define relationships
  • Coordinate Mappings: Functions that map between dimensional spaces
  • Vector Points: Individual data records positioned in the coordinate space

Key Features

Vector Mathematics Foundation

  • Coordinate System Architecture: Data organized around mathematical coordinate principles
  • Dimensional Spaces: Scalable N-dimensional data representation without structural changes
  • O(1) Lookup Performance: Coordinate indexing for instant data retrieval
  • Value Deduplication: Automatic optimization of storage through value deduplication in dimensional spaces

Domain-Driven Architecture

  • Clean Architecture: Separation of domain logic, application services, and infrastructure
  • Coordinate Abstractions: Rich domain objects representing mathematical concepts
  • Immutable Value Objects: Thread-safe coordinate and mapping representations
  • Repository Patterns: Clean data access interfaces

Quick Start

Installation

# Clone the repository
git clone <repository-url>
cd vector

# Install in development mode
pip install -e .

# Or install from PyPI (when published)
pip install vector-datalib

Basic Usage

Pure Async Architecture - Modern Python Standards

import asyncio
from vector_datalib import VectorDB

async def main():
    # Create database with async context manager
    async with VectorDB("my_data.db") as db:
        # Insert data - all operations are async
        await db.insert(101, {"age": 25, "name": "Alice"})
        await db.insert(102, {"age": 30, "name": "Bob"})
        await db.insert(103, {"age": 25, "name": "Charlie"})  # age=25 deduplicated automatically

        # O(1) coordinate-based lookup
        name = await db.lookup(101, "name")
        print(f"Employee 101: {name}")  # Employee 101: Alice

        # Concurrent batch operations with asyncio.gather()
        await db.batch_insert([
            (104, {"name": "Diana", "age": 28}),
            (105, {"name": "Eve", "age": 32}),
            (106, {"name": "Frank", "age": 27})
        ])

        # Update operations
        await db.update(101, "age", 26)  # Alice's age updated

        # Auto-save with async context manager
        # Database automatically saved on __aexit__

# Run with asyncio
asyncio.run(main())

Advanced Async Patterns

async def advanced_usage():
    async with VectorDB("analytics.db") as db:
        # Concurrent lookups with asyncio.gather()
        user_queries = [(101, "name"), (102, "age"), (103, "department")]
        results = await db.batch_lookup(user_queries)

        # LRU caching automatically optimizes repeated lookups
        name1 = await db.lookup(101, "name")  # Database hit
        name2 = await db.lookup(101, "name")  # Cache hit (faster)

        # Concurrent updates
        updates = [
            (101, "status", "active"),
            (102, "status", "inactive"), 
            (103, "role", "manager")
        ]

        successful = await db.batch_update(updates)
        print(f"Updated {successful} records concurrently")

asyncio.run(advanced_usage())

Architecture

Vector follows clean architecture principles with mathematical domain modeling:

src/vector_datalib/
├── domain/
│   ├── coordinates/            # X-axis coordinate system (primary key)
│   ├── spaces/                 # Y, Z, J... dimensional spaces  
│   ├── mappings/               # Functions between dimensional spaces
│   └── __init__.py
├── application/
│   ├── main.py                 # Main database API
│   └── __init__.py
├── infrastructure/
│   ├── storage/                # .db file persistence
│   └── __init__.py
├── meta.py                     # Version and metadata
└── __init__.py                 # Public API exports

Domain Layer

  • CentralAxis: Manages X-coordinate system and primary key constraints
  • DimensionalSpace: Handles Y, Z, J... dimensions with value deduplication
  • CoordinateMapping: Maps relationships between dimensional spaces
  • VectorPoint: Represents individual data records as coordinate positions

Application Layer

  • VectorDB: Main database interface providing the scripting API
  • Coordinate Operations: Insert, lookup, update operations on coordinate system
  • Dimensional Management: Dynamic expansion and contraction of coordinate spaces

Infrastructure Layer

  • VectorFileStorage: Handles .db file format with JSON and gzip compression
  • Persistence Management: Atomic save/load operations with metadata

Mathematical Model

Coordinate System Design

All tables in Vector must follow the coordinate system principle:

  • X-axis (Primary Key): Central coordinate that uniquely identifies each vector point
  • Dimensional Relationships: Other attributes represent relationships between the X-coordinate and various dimensional spaces
# Mathematical representation:
# Point P at coordinate X has relationships to multiple dimensions
# P(x) = {Y: f_y(x), Z: f_z(x), J: f_j(x), ...}
# where f_axis represents the mapping function for each dimension

async with VectorDB("data.db") as db:
    await db.insert(101, {"age": 25, "name": "Alice", "city": "Boston"})
    # Creates: P(101) = {age: f_age(101)=25, name: f_name(101)="Alice", city: f_city(101)="Boston"}

Value Deduplication

Vector automatically optimizes storage by deduplicating values within dimensional spaces:

async with VectorDB("data.db") as db:
    await db.insert(101, {"age": 25, "name": "Alice"})
    await db.insert(102, {"age": 25, "name": "Bob"})     # age=25 stored once
    await db.insert(103, {"age": 25, "name": "Charlie"}) # age=25 referenced

    # Storage optimization: age=25 stored once, referenced by multiple coordinates

N-Dimensional Scalability

Add new dimensions without structural changes:

async with VectorDB("data.db") as db:
    # Start with 2 dimensions
    await db.insert(101, {"age": 25, "name": "Alice"})

    # Expand to 3 dimensions
    await db.insert(102, {"age": 30, "name": "Bob", "city": "Boston"})

    # Expand to N dimensions dynamically
    await db.insert(103, {"age": 25, "name": "Charlie", "city": "Boston", "department": "Engineering"})

Performance Characteristics

Time Complexity

  • Insert: O(1) average case with hash-based coordinate indexing
  • Lookup: O(1) direct coordinate access
  • Update: O(1) coordinate-based modification
  • Dimensional Expansion: O(1) addition of new coordinate relationships

Concurrency Benefits

  • Pure Async Architecture: Non-blocking I/O operations with asyncio
  • Concurrent Batch Operations: Multiple operations with asyncio.gather()
  • LRU Caching: In-memory caching for frequently accessed data
  • Async Context Managers: Automatic resource management and cleanup

Space Complexity

  • Value Deduplication: Automatic optimization reduces memory usage
  • Coordinate Indexing: Hash-based storage for constant-time access
  • Compression: Gzip compression for persistent storage efficiency

File Format

.db File Structure

{
  "metadata": {
    "version": "1.1.0-beta",
    "created_at": "2025-01-XX",
    "coordinate_count": 1000
  },
  "central_axis": {
    "coordinates": [1, 2, 3, ...]
  },
  "dimensional_spaces": {
    "age": {
      "values": [25, 30, 35],
      "coordinate_mappings": {"1": 0, "2": 1, "3": 0}
    },
    "name": {
      "values": ["Alice", "Bob", "Charlie"],
      "coordinate_mappings": {"1": 0, "2": 1, "3": 2}
    }
  }
}

Development

Requirements

  • Python 3.9+
  • No external dependencies (uses only standard library)

Coordinate System Examples

User Management System

async with VectorDB("users.db") as db:
    # X-coordinate: User ID, Y-dimension: Profile data
    await db.insert(1001, {"name": "Alice Johnson", "age": 28, "department": "Engineering"})
    await db.insert(1002, {"name": "Bob Smith", "age": 32, "department": "Sales"})  
    await db.insert(1003, {"name": "Charlie Brown", "age": 28, "department": "Engineering"})

    # O(1) user lookup
    name = await db.lookup(1001, "name")
    age = await db.lookup(1001, "age") 
    print(f"User: {name}, Age: {age}")

    # Dynamic expansion - add new dimensional relationships
    await db.update(1001, "salary", 75000)
    await db.update(1001, "location", "Boston")

Product Catalog

async with VectorDB("products.db") as db:
    # X-coordinate: Product ID, Y/Z dimensions: Product attributes
    await db.insert(2001, {"name": "Laptop", "price": 999.99, "category": "Electronics"})
    await db.insert(2002, {"name": "Mouse", "price": 29.99, "category": "Electronics"})
    await db.insert(2003, {"name": "Desk", "price": 299.99, "category": "Furniture"})

    # Value deduplication automatically optimizes "Electronics" category storage

Best Practices

Coordinate System Design

  • Always use X-axis as primary key: This maintains the mathematical foundation
  • Design dimensional relationships: Think about how attributes relate to coordinates
  • Leverage value deduplication: Repeated values in dimensions are automatically optimized
  • Plan for dimensional expansion: Design coordinate spaces that can grow dynamically

Performance Optimization

  • Use async context managers: Always use async with VectorDB() for resource management
  • Leverage concurrent operations: Use batch_insert(), batch_lookup(), batch_update() for multiple operations
  • LRU cache awareness: Repeated lookups are cached automatically
  • Appropriate coordinate ranges: Choose coordinate values that distribute well
  • Monitor dimensional growth: Large numbers of unique values reduce deduplication benefits

Data Organization

  • Logical coordinate grouping: Group related data with nearby coordinates when possible
  • Consistent dimensional naming: Use clear, consistent names for dimensional spaces
  • Document coordinate meanings: Maintain documentation of what each coordinate represents

Troubleshooting

Common Issues

Large file sizes with compressed storage:

  • Check for high dimensional diversity (many unique values)
  • Consider coordinate space reorganization for better deduplication

Performance degradation:

  • Monitor the number of unique values in dimensional spaces
  • Consider splitting large coordinate spaces into multiple databases

Contributing

  1. Fork the repository
  2. Create a feature branch following the coordinate system principles
  3. Implement changes with proper domain modeling
  4. Ensure mathematical consistency in coordinate operations
  5. Submit a pull request

License

This project is licensed under the Apache 2.0 License - see the LICENSE file for details.

Vector Mathematics

Vector database design is inspired by mathematical vector spaces where:

  • Coordinates define position: X-axis establishes the coordinate system foundation
  • Dimensions represent relationships: Each dimension shows how data relates to coordinates
  • Mappings preserve structure: Functions between dimensions maintain mathematical consistency
  • Scalability through expansion: N-dimensional growth without architectural changes

The name "Vector" reflects this mathematical foundation where data points exist as vectors in a coordinate space, with the X-axis serving as the primary coordinate system and other dimensions representing the vector's components in different spaces.


Organize your data with mathematical precision. Scale with coordinate clarity.

Built for developers who appreciate clean architecture and mathematical elegance.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

vector_datalib-1.1.0b0.tar.gz (24.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

vector_datalib-1.1.0b0-py3-none-any.whl (25.6 kB view details)

Uploaded Python 3

File details

Details for the file vector_datalib-1.1.0b0.tar.gz.

File metadata

  • Download URL: vector_datalib-1.1.0b0.tar.gz
  • Upload date:
  • Size: 24.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for vector_datalib-1.1.0b0.tar.gz
Algorithm Hash digest
SHA256 b292ae44365084e32ad8346d2931cf8a4d2edef456478b64ca810a18d4ef3c25
MD5 04eb0c9c4a5ebb140d5e99c631537978
BLAKE2b-256 e5806da231f0a7d5d072e86ad201d1fbee6e849c7ddd9a02289e9108007fc7a8

See more details on using hashes here.

File details

Details for the file vector_datalib-1.1.0b0-py3-none-any.whl.

File metadata

File hashes

Hashes for vector_datalib-1.1.0b0-py3-none-any.whl
Algorithm Hash digest
SHA256 d6b340fccec4ab8226298702b9b332f1f631a98ca84465ab1b60d3e75e2961dd
MD5 ec64858dabb0ae738762d838f5fc7af9
BLAKE2b-256 9ff6a0074a17fc1df0c87e9e177646e0f24374cdf38baf3aa9daf9ac6bc41812

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page