Skip to main content

High-performance HTAP embedded database with Rust core and Python API

Project description

ApexBase

High-performance HTAP embedded database with Rust core and Python API

ApexBase is an embedded columnar database designed for Hybrid Transactional/Analytical Processing (HTAP) workloads. It combines a high-throughput columnar storage engine written in Rust with an ergonomic Python API, delivering analytical query performance that surpasses DuckDB and SQLite on most benchmarks — all in a single .apex file with zero external dependencies.

Table of Contents


Features

  • HTAP architecture — V4 Row Group columnar storage with DeltaStore for cell-level updates; fast inserts and fast analytical scans in one engine
  • Multi-database support — multiple isolated databases in one directory; cross-database queries with standard db.table SQL syntax
  • Single-file storage — custom .apex format per table, no server process, no external dependencies
  • Comprehensive SQL — DDL, DML, JOINs (INNER/LEFT/RIGHT/FULL/CROSS), subqueries (IN/EXISTS/scalar), CTEs (WITH ... AS), UNION/UNION ALL/INTERSECT/EXCEPT, window functions, EXPLAIN/ANALYZE, multi-statement execution
  • 70+ built-in functions — math (ABS, SQRT, POWER, LOG, trig), string (UPPER, LOWER, SUBSTR, REPLACE, CONCAT, REGEXP_REPLACE, ...), date (YEAR, MONTH, DAY, DATEDIFF, DATE_ADD, ...), conditional (COALESCE, IFNULL, NULLIF, CASE WHEN, GREATEST, LEAST)
  • Aggregation and analytics — COUNT, SUM, AVG, MIN, MAX, COUNT(DISTINCT), GROUP BY, HAVING, ORDER BY with NULLS FIRST/LAST
  • Window functions — ROW_NUMBER, RANK, DENSE_RANK, NTILE, PERCENT_RANK, CUME_DIST, LAG, LEAD, FIRST_VALUE, LAST_VALUE, NTH_VALUE, RUNNING_SUM, and windowed SUM/AVG/COUNT/MIN/MAX with PARTITION BY and ORDER BY
  • Transactions — BEGIN / COMMIT / ROLLBACK with OCC (Optimistic Concurrency Control), SAVEPOINT / ROLLBACK TO / RELEASE, statement-level auto-rollback
  • MVCC — multi-version concurrency control with snapshot isolation, version store, and garbage collection
  • Indexing — B-Tree and Hash indexes with CREATE INDEX / DROP INDEX / REINDEX; automatic multi-index AND intersection for compound predicates
  • Full-text search — built-in NanoFTS integration with fuzzy matching
  • Vector search — SIMD-accelerated nearest-neighbour search with 6 distance metrics (L2, cosine, dot, L1, L∞, L2²); heap-based O(n log k) TopK; single-query topk_distance() and batch batch_topk_distance() Python APIs; SQL explode_rename(topk_distance(...)) syntax; current verified snapshot shows 7-14x faster shared TopK queries than DuckDB
  • Float16 vector storageFLOAT16_VECTOR column type stores embeddings as 16-bit floats (half the memory of float32); SIMD-accelerated f16 distance kernels via NEON fp16 on ARM (FCVTL/FCVTL2) and AVX2+F16C on x86_64; automatic runtime CPU dispatch; ≥2× faster than f32 on Apple Silicon; transparent API — query with float32, stored as f16
  • JIT compilation — Cranelift-based JIT for predicate evaluation and SIMD-vectorized aggregations
  • Zero-copy Python bridge — Arrow IPC between Rust and Python; direct conversion to Pandas, Polars, and PyArrow
  • Durability levels — configurable fast / safe / max with WAL support and crash recovery
  • Compact storage — dictionary encoding for low-cardinality strings, LZ4 and Zstd compression
  • File reading table functionsread_csv(), read_parquet(), read_json() directly in SQL FROM clauses; parallel mmap parsing; full SQL (WHERE / GROUP BY / JOIN / UNION) on top of any file
  • Temporary tables from filesregister_temp_table() parses CSV/JSON/Parquet once and materializes as a native .apex temp table; mmap-backed zero-copy reads, zone maps, bloom filters; order-of-magnitude faster than repeated read_* calls; auto-cleanup on close
  • Parquet interop — COPY TO / COPY FROM Parquet files
  • PostgreSQL wire protocol — built-in server for DBeaver, psql, DataGrip, pgAdmin, Navicat, and any PostgreSQL-compatible client; two distribution modes (Python CLI or standalone Rust binary)
  • Arrow Flight gRPC server — high-performance columnar data transfer over HTTP/2; streams Arrow IPC RecordBatch directly, 4–7× faster than PG wire for large result sets; accessible via pyarrow.flight, Go arrow, Java arrow, and any Arrow Flight client
  • Cross-platform — Linux, macOS, and Windows; x86_64 and ARM64; Python 3.9 -- 3.13

Installation

pip install apexbase

Build from source (requires Rust toolchain):

maturin develop --release

Quick Start

from apexbase import ApexClient

# Open (or create) a database directory
client = ApexClient("./data")

# Create a table
client.create_table("users")

# Store records
client.store({"name": "Alice", "age": 30, "city": "Beijing"})
client.store([
    {"name": "Bob", "age": 25, "city": "Shanghai"},
    {"name": "Charlie", "age": 35, "city": "Beijing"},
])

# SQL query
results = client.execute("SELECT * FROM users WHERE age > 28 ORDER BY age DESC")

# Convert to DataFrame
df = results.to_pandas()

client.close()

Usage Guide

Database Management

ApexBase supports multiple isolated databases within a single root directory. Each named database lives in its own subdirectory; the default database uses the root directory.

# Switch to a named database (creates it if needed)
client.use_database("analytics")

# Combined: switch database + select/create a table in one call
client.use(database="analytics", table="events")

# List all databases
dbs = client.list_databases()  # ["analytics", "default", "hr"]

# Current database
print(client.current_database)  # "analytics"

# Cross-database SQL — standard db.table syntax
client.execute("SELECT * FROM default.users")
client.execute("SELECT u.name, e.event FROM default.users u JOIN analytics.events e ON u.id = e.user_id")
client.execute("INSERT INTO analytics.events (name) VALUES ('click')")
client.execute("UPDATE default.users SET age = 31 WHERE name = 'Alice'")
client.execute("DELETE FROM default.users WHERE age < 18")

All SQL operations (SELECT, INSERT, UPDATE, DELETE, JOIN, CREATE TABLE, DROP TABLE, ALTER TABLE) support database.table qualified names, allowing cross-database queries in a single statement.

Table Management

Each table is stored as a separate .apex file. Tables must be created before use.

# Create with optional schema
client.create_table("orders", schema={
    "order_id": "int64",
    "product": "string",
    "price": "float64",
})

# Switch tables
client.use_table("users")

# List / drop
tables = client.list_tables()
client.drop_table("orders")

Data Ingestion

import pandas as pd
import polars as pl
import pyarrow as pa

# Columnar dict (fastest for bulk data)
client.store({
    "name": ["D", "E", "F"],
    "age": [22, 32, 42],
})

# From pandas / polars / PyArrow (auto-creates table when table_name given)
client.from_pandas(pd.DataFrame({"name": ["G"], "age": [28]}), table_name="users")
client.from_polars(pl.DataFrame({"name": ["H"], "age": [38]}), table_name="users")
client.from_pyarrow(pa.table({"name": ["I"], "age": [48]}), table_name="users")

SQL

ApexBase supports a broad SQL dialect. Examples:

# DDL
client.execute("CREATE TABLE IF NOT EXISTS products")
client.execute("ALTER TABLE products ADD COLUMN name STRING")
client.execute("DROP TABLE IF EXISTS products")

# DML
client.execute("INSERT INTO users (name, age) VALUES ('Zoe', 29)")
client.execute("UPDATE users SET age = 31 WHERE name = 'Alice'")
client.execute("DELETE FROM users WHERE age < 20")

# SELECT with full clause support
client.execute("""
    SELECT city, COUNT(*) AS cnt, AVG(age) AS avg_age
    FROM users
    WHERE age BETWEEN 20 AND 40
    GROUP BY city
    HAVING cnt > 1
    ORDER BY avg_age DESC
    LIMIT 10
""")

# JOINs
client.execute("""
    SELECT u.name, o.product
    FROM users u
    INNER JOIN orders o ON u._id = o.user_id
""")

# Subqueries
client.execute("SELECT * FROM users WHERE age > (SELECT AVG(age) FROM users)")
client.execute("SELECT * FROM users WHERE city IN (SELECT city FROM cities WHERE pop > 1000000)")

# CTEs
client.execute("""
    WITH seniors AS (SELECT * FROM users WHERE age >= 30)
    SELECT city, COUNT(*) FROM seniors GROUP BY city
""")

# Window functions
client.execute("""
    SELECT name, age,
           ROW_NUMBER() OVER (ORDER BY age DESC) AS rank,
           AVG(age) OVER (PARTITION BY city) AS city_avg
    FROM users
""")

# Set operations
client.execute("""
    SELECT name FROM users WHERE city = 'Beijing'
    UNION ALL
    SELECT name FROM users WHERE city = 'Shanghai'
""")
client.execute("""
    SELECT user_id FROM orders
    INTERSECT
    SELECT user_id FROM wishlist
""")
client.execute("""
    SELECT user_id FROM orders
    EXCEPT
    SELECT user_id FROM support_tickets WHERE status = 'open'
""")

# Multi-statement
client.execute("""
    INSERT INTO users (name, age) VALUES ('New1', 20);
    INSERT INTO users (name, age) VALUES ('New2', 21);
    SELECT COUNT(*) FROM users
""")

# INSERT ... ON CONFLICT (upsert)
client.execute("""
    INSERT INTO users (name, age) VALUES ('Alice', 31)
    ON CONFLICT (name) DO UPDATE SET age = 31
""")

# CREATE TABLE AS
client.execute("CREATE TABLE seniors AS SELECT * FROM users WHERE age >= 30")

# EXPLAIN / EXPLAIN ANALYZE
client.execute("EXPLAIN SELECT * FROM users WHERE age > 25")

# Parquet interop
client.execute("COPY users TO '/tmp/users.parquet'")
client.execute("COPY users FROM '/tmp/users.parquet'")

File Reading Table Functions

Read external files directly in a SQL FROM clause — no import step required. The full SQL engine runs on top: WHERE, GROUP BY, ORDER BY, JOIN, UNION, etc.

# CSV: schema inferred automatically, parallel mmap parser
df = client.execute("SELECT * FROM read_csv('/data/sales.csv')").to_pandas()

# TSV — specify delimiter
df = client.execute("SELECT * FROM read_csv('/data/data.tsv', delimiter='\t')").to_pandas()

# No header row
df = client.execute("SELECT * FROM read_csv('/data/raw.csv', header=false)").to_pandas()

# Parquet: schema from file metadata, parallel column decode
table = client.execute("SELECT * FROM read_parquet('/data/events.parquet')").to_arrow()

# JSON / NDJSON: auto-detects format (NDJSON or pandas column-oriented)
df = client.execute("SELECT * FROM read_json('/data/logs.ndjson')").to_pandas()

# Full SQL on top of a file
result = client.execute("""
    SELECT city, COUNT(*) AS cnt, AVG(price)
    FROM read_csv('/data/orders.csv')
    WHERE price > 100
    GROUP BY city
    ORDER BY cnt DESC
    LIMIT 10
""")

# JOIN a file with a stored table
result = client.execute("""
    SELECT u.name, f.score
    FROM users u
    JOIN read_csv('/data/scores.csv') f ON u.id = f.user_id
    WHERE f.score > 90
""")

# EXCEPT using a file as a blocklist
result = client.execute("""
    SELECT email FROM users
    EXCEPT
    SELECT email FROM read_csv('/data/unsubscribed.csv')
""")
Function Options Description
read_csv(path) header=true, delimiter=',' Read CSV/TSV; auto-infers schema
read_parquet(path) Read Parquet; schema from file metadata
read_json(path) Read NDJSON or pandas JSON; auto-detects format

See docs/API_REFERENCE.md for full details.

Temporary Tables from Files

Register CSV, JSON, or Parquet files as temporary native tables. The file is parsed once and stored in ApexBase's mmap-backed .apex format. Subsequent queries bypass file parsing entirely, leveraging zone maps, bloom filters, and zero-copy mmap reads — an order of magnitude faster than repeated read_csv() / read_json() / read_parquet() calls.

# Register a CSV file as a temp table
client.register_temp_table("orders", "/data/orders.csv")

# Query it as a regular table — lightning fast, near-zero memory
result = client.execute("SELECT city, COUNT(*) FROM orders GROUP BY city")

# Full SQL works (WHERE, JOIN, GROUP BY, UNION, window functions, etc.)
result = client.execute("""
    SELECT o.city, u.name
    FROM orders o
    JOIN users u ON o.user_id = u._id
    WHERE o.amount > 100
    ORDER BY o.amount DESC
    LIMIT 20
""")

# Drop when done (or just close the client — auto-cleanup)
client.drop_temp_table("orders")

Also supports SQL syntax:

client.execute("CREATE TEMP TABLE invoices AS SELECT * FROM read_csv('/data/invoices.csv')")

Supported formats:

  • CSV / TSV — auto-detected by .csv / .tsv extension
  • JSON / NDJSON — auto-detected by .json / .ndjson / .jsonl extension
  • Parquet — auto-detected by .parquet extension

Memory & performance:

  • Temp tables use memory-mapped I/O — data stays on disk, near-zero RAM footprint
  • Zone maps (min/max indexes) skip irrelevant row groups for filtered queries
  • Bloom filters accelerate point lookups
  • Cleaned up automatically when the client is closed or the database is dropped

Transactions

client.execute("BEGIN")
client.execute("INSERT INTO users (name, age) VALUES ('Tx1', 20)")
client.execute("SAVEPOINT sp1")
client.execute("INSERT INTO users (name, age) VALUES ('Tx2', 21)")
client.execute("ROLLBACK TO sp1")   # undo Tx2 only
client.execute("COMMIT")            # Tx1 persisted

Transactions use OCC validation — concurrent writes are detected at commit time.

Indexes

client.execute("CREATE INDEX idx_age ON users (age)")
client.execute("CREATE UNIQUE INDEX idx_name ON users (name)")

# Queries automatically use indexes when applicable
client.execute("SELECT * FROM users WHERE age = 30")  # index scan

client.execute("DROP INDEX idx_age ON users")
client.execute("REINDEX users")

Full-Text Search

ApexBase ships a native full-text search engine (NanoFTS) integrated directly into the SQL executor. FTS is available through all interfaces — Python API, PostgreSQL Wire, and Arrow Flight — without any Python-side middleware.

SQL interface (recommended)

# 1. Create the FTS index via SQL DDL
client.execute("CREATE FTS INDEX ON articles (title, content)")

# Optional: specify lazy loading and cache size
client.execute("CREATE FTS INDEX ON logs WITH (lazy_load=true, cache_size=50000)")

# 2. Query using MATCH() / FUZZY_MATCH() in WHERE
results = client.execute("SELECT * FROM articles WHERE MATCH('rust programming')")
results = client.execute("SELECT title, content FROM articles WHERE FUZZY_MATCH('pytohn')")

# Combine with other predicates
results = client.execute("""
    SELECT * FROM articles
    WHERE MATCH('machine learning') AND published_at > '2024-01-01'
    ORDER BY _id DESC LIMIT 20
""")

# FTS also works in aggregations
count = client.execute("SELECT COUNT(*) FROM articles WHERE MATCH('deep learning')")

# Manage indexes
client.execute("SHOW FTS INDEXES")           # list all FTS-enabled tables
client.execute("ALTER FTS INDEX ON articles DISABLE")  # disable, keep files
client.execute("DROP FTS INDEX ON articles") # remove index + delete files

Python API (alternative)

# Initialize FTS for current table
client.use_table("articles")
client.init_fts(index_fields=["title", "content"])

# Search
ids    = client.search_text("database")
fuzzy  = client.fuzzy_search_text("databse")   # tolerates typos
recs   = client.search_and_retrieve("python", limit=10)
top5   = client.search_and_retrieve_top("neural network", n=5)

# Lifecycle
client.get_fts_stats()
client.disable_fts()   # suspend without deleting files
client.drop_fts()      # remove index + delete files

Tip: The SQL interface (MATCH() / FUZZY_MATCH()) works over PG Wire and Arrow Flight without any extra setup; the Python API methods are Python-process-only.

Vector Search

ApexBase provides SIMD-accelerated nearest-neighbour search with a zero-copy mmap scan buffer. Supports 6 distance metrics and both single-query and batch modes.

import numpy as np

# Store vectors — numpy arrays are stored as FixedList columns (optimal)
client.create_table("items")
client.store({
    "label": ["a", "b", "c"],
    "vec":   [np.random.rand(128).astype(np.float32) for _ in range(3)],
})

query = np.random.rand(128).astype(np.float32)

# Single-query: returns ResultView with _id and dist columns
results = client.topk_distance('vec', query, k=10)
df = results.to_pandas()           # columns: _id, dist
top_ids = results.get_ids()        # numpy int64 array
records = client.retrieve_many(top_ids.tolist())  # full records

# Custom metric and column names
results = client.topk_distance('vec', query, k=5, metric='cosine',
                                id_col='item_id', dist_col='cosine_dist')

# Batch: N queries in one Rust call (scan_buf loaded once, Rayon parallel)
queries = np.random.rand(100, 128).astype(np.float32)
result  = client.batch_topk_distance('vec', queries, k=10)
# result.shape == (100, 10, 2)
ids   = result[:, :, 0].astype(np.int64)   # (100, 10)
dists = result[:, :, 1]                     # (100, 10)

# SQL: explode_rename(topk_distance(...)) — same query, SQL form
results = client.execute("""
    SELECT explode_rename(topk_distance(vec, [0.1, 0.2, 0.3], 10, 'l2'), '_id', 'dist')
    FROM items
""")

Supported metrics: 'l2' / 'euclidean', 'l2_squared', 'l1' / 'manhattan', 'linf' / 'chebyshev', 'cosine' / 'cosine_distance', 'dot' / 'inner_product'

Latest verified benchmark snapshot

  • Dataset: 200,000 vectors x dim=128, k=10
  • Method: 2 warmup + 3 timed iterations
  • Harness: integrated into benchmarks/bench_vs_sqlite_duckdb.py
  • SQLite note: stock sqlite3 in this harness has no native vector distance/top-k functions, so ranked vector comparisons are ApexBase vs DuckDB only

Single-query TopK

Metric ApexBase DuckDB Gap
L2 3.58 ms 26.46 ms 7.4x faster
Cosine 3.80 ms 31.89 ms 8.4x faster
Dot 3.48 ms 26.37 ms 7.6x faster

Batch TopK (10 queries)

Metric ApexBase DuckDB Gap
L2 23.18 ms 266.92 ms 11.5x faster
Cosine 23.24 ms 322.07 ms 13.9x faster
Dot 21.12 ms 268.44 ms 12.7x faster

ApexBase-only metrics

Metric ApexBase
L2 squared 3.56 ms
L1 3.34 ms
Linf 3.39 ms

See docs/API_REFERENCE.md#vector-search for full details.

Record-Level Operations

record = client.retrieve(1)               # by internal _id
records = client.retrieve_many([1, 2, 3])
all_data = client.retrieve_all()

client.replace(1, {"name": "Alice2", "age": 31})
client.delete(1)
client.delete([2, 3, 4])

Column Operations

client.add_column("email", "String")
client.rename_column("email", "email_addr")
client.drop_column("email_addr")
client.get_column_dtype("age")    # "Int64"
client.list_fields()              # ["name", "age", "city"]

ResultView

Query results are returned as ResultView objects with multiple output formats:

results = client.execute("SELECT * FROM users")

df = results.to_pandas()       # pandas DataFrame (zero-copy by default)
pl_df = results.to_polars()    # polars DataFrame
arrow = results.to_arrow()     # PyArrow Table
dicts = results.to_dict()      # list of dicts

results.shape                  # (rows, columns)
results.columns                # column names
len(results)                   # row count
results.first()                # first row as dict
results.scalar()               # single value (for aggregates)
results.get_ids()              # numpy array of _id values

Context Manager

with ApexClient("./data") as client:
    client.create_table("tmp")
    client.store({"key": "value"})
    # Automatically closed on exit

Performance

Latest Verified Snapshot

This section tracks the latest verified local benchmark snapshot rather than an old best-case run.

  • System: macOS 26.4.1, Apple arm (10 cores), 32 GB RAM
  • Stack: Python 3.12.4, ApexBase 1.17.0, SQLite 3.45.3, DuckDB 1.1.3, PyArrow 23.0.1
  • Dataset: 200,000 rows x 5 columns (name, age, score, city, category)
  • Vector dataset: 200,000 vectors x dim=128, k=10, batch size 10 queries
  • Method: 2 warmup iterations + 3 timed iterations
  • Layout: 92 named metrics total (37 OLAP, 46 OLTP, 9 vector)
  • Fairness rule: only the default fair OLAP/OLTP cross-engine tables count toward the 38/38 win/loss summary. The vector similarity module uses a separate vector dataset and has its own ApexBase-vs-DuckDB scoreboard; Apex-only buffered, memtable, materialization, and diagnostic paths are kept separate so semantics stay comparable.

Scoreboard

Scope Metrics Apex wins Ties Slower
Default fair (OLAP + OLTP) 38 38 0 0
OLAP fair 29 29 0 0
OLTP fair 9 9 0 0
Vector similarity (ApexBase vs DuckDB) 6 6 0 0

Stock SQLite is not ranked in the vector table because the built-in sqlite3 used here has no native vector distance/top-k functions in this harness.

Representative OLAP Gaps

These are the easiest rows to scan if you want the shape of the result set quickly.

Metric ApexBase SQLite DuckDB Gap to best other
COUNT(*) 0.106 ms 1.775 ms 0.397 ms 3.7x faster vs DuckDB
SELECT * LIMIT 100 (warm cache) 6 us 0.107 ms 0.236 ms 17.8x faster vs SQLite
Filtered LIMIT 100 (age>30) 0.050 ms 0.173 ms 0.603 ms 3.5x faster vs SQLite
GROUP BY city (10 groups) 0.060 ms 60.108 ms 2.399 ms 40.0x faster vs DuckDB
Window ROW_NUMBER PARTITION BY city 0.622 ms 99.086 ms 12.809 ms 20.6x faster vs DuckDB

Representative OLTP Gaps

Metric ApexBase SQLite DuckDB Gap to best other
Bulk Insert (N rows; default fair) 53.948 ms 197.464 ms 35.84 s 3.7x faster vs SQLite
Point Lookup (SQL by ID) 0.035 ms 0.067 ms 2.198 ms 1.9x faster vs SQLite
Retrieve Many (SQL, 100 IDs) 0.175 ms 0.317 ms 3.942 ms 1.8x faster vs SQLite
FTS Index Build (name,city,category) 103.738 ms 246.588 ms 790.703 ms 2.4x faster vs SQLite
FTS Search ('Electronics') 0.160 ms 5.700 ms 14.644 ms 35.6x faster vs SQLite

Representative Vector Gaps

SQLite is excluded here for one reason only: stock sqlite3 in this harness has no native vector distance/top-k support.

Metric ApexBase DuckDB Gap to DuckDB
TopK L2 3.58 ms 26.46 ms 7.4x faster
TopK Cosine 3.80 ms 31.89 ms 8.4x faster
TopK Dot 3.48 ms 26.37 ms 7.6x faster
Batch TopK L2 (10 queries) 23.18 ms 266.92 ms 11.5x faster
Batch TopK Cosine (10 queries) 23.24 ms 322.07 ms 13.9x faster
Batch TopK Dot (10 queries) 21.12 ms 268.44 ms 12.7x faster

Throughput Snapshot

Q/s uses a mixed analytical profile: COUNT(*), two GROUP BY scans, and Filtered LIMIT 100, all materialized to Python rows.

Throughput metric ApexBase SQLite DuckDB Gap to best other
OLAP Q/s (single thread) 123,700.3 34.8 942.2 131.3x higher vs DuckDB
OLAP Q/s (4 threads) 125,196.3 126.6 2,776.8 45.1x higher vs DuckDB

Hot-Path Latency Snapshot

These tables are not part of the 38/38 fair scoreboard. They answer a different question: how fast is the already-loaded hot path, and what happens when durability or transaction semantics are made explicit?

Default Microbenchmarks

Metric ApexBase SQLite DuckDB Gap to best other
COUNT(*) (direct API) 7.43 us 1.243 ms 0.132 ms 17.8x faster vs DuckDB
Point lookup (projected SQL) 2.12 us 2.99 us 1.722 ms 1.4x faster vs SQLite
Retrieve 100 IDs (projected SQL) 0.041 ms 0.099 ms 3.438 ms 2.4x faster vs SQLite
Insert 1 row (default fair) 0.010 ms 0.014 ms 0.297 ms 1.4x faster vs SQLite
UPDATE by ID 1.13 us 4.23 us 0.483 ms 3.7x faster vs SQLite
DELETE missing ID 2.72 us 3.81 us 1.160 ms 1.4x faster vs SQLite

Durable Fair Microbenchmarks

Metric ApexBase SQLite DuckDB Gap to best other
Insert 1 row (durable fair) 0.101 ms 0.126 ms 31.106 ms 1.2x faster vs SQLite
UPDATE by ID (durable fair) 2.02 us 6.53 us 4.469 ms 3.2x faster vs SQLite

Transaction Fair Microbenchmarks

Metric ApexBase SQLite DuckDB Gap to best other
TXN empty (BEGIN+COMMIT; durable sync) 3.29 us 4.08 us 0.148 ms 1.2x faster vs SQLite
TXN read COUNT(*) (COMMIT; durable sync) 0.018 ms 1.295 ms 0.294 ms 16.3x faster vs DuckDB
TXN backlog string miss (COMMIT; 1500 preseed; durable sync) 0.051 ms 8.011 ms 0.404 ms 7.9x faster vs DuckDB
TXN backlog COUNT(*) (COMMIT; 1500 preseed; durable sync) 0.030 ms 3.793 ms 0.305 ms 10.2x faster vs DuckDB
TXN backlog INSERT+read-own-name (COMMIT; 1500 preseed; durable sync) 0.262 ms 8.073 ms 33.522 ms 30.8x faster vs SQLite

Full Fair Tables

For readability, the README keeps the competitive summary, representative gaps, and hot-path snapshots above instead of embedding all 38 fair rows inline.

  • Run python benchmarks/bench_vs_sqlite_duckdb.py --rows 200000 --warmup 2 --iterations 3 to print the complete OLAP and OLTP fair tables.
  • Add --output FILE.json to export every raw metric, including the separate vector module, in machine-readable form.

OLTP Write Visibility

ApexBase exposes two fast single-row append paths, and the benchmark keeps them out of the fair scoreboard because their visibility rules are Apex-specific:

  • Memtable OLTP is the default fast single-row path for schema-stable store({...}) calls with durability="fast". The writing client can read the row immediately, managed clients in the same Python process share the storage instance, and flush() / close() persists pending rows. A separate process sees those rows only after the writer flushes, closes, or reaches the auto-flush threshold.
  • Buffered OLTP is explicit: call begin_buffered_writes(), issue many single-row store({...}) calls, then call flush_buffered_writes() or end_buffered_writes(flush=True). Buffered rows are not visible until flushed.

That separation is deliberate: the fair tables compare committed cross-engine behavior, while the Apex-only write modes remain visible as diagnostics instead of being mixed into the competitive summary.

Reproduce

Use the same command as the snapshot above:

python benchmarks/bench_vs_sqlite_duckdb.py --rows 200000 --warmup 2 --iterations 3

Add --skip-vector if you want a tabular-only rerun without the separate vector module.

For a larger stress run, increase --rows to 1000000.


Server Protocols

ApexBase ships two complementary server protocols for external access:

Protocol Port Best for Binary / CLI
PG Wire 5432 DBeaver, psql, DataGrip, BI tools apexbase-server
Arrow Flight 50051 Python (pyarrow), Go, Java, Spark apexbase-flight

Combined Launcher (Both Servers at Once)

# Start PG Wire + Arrow Flight simultaneously
apexbase-serve --dir /path/to/data

# Custom ports
apexbase-serve --dir /path/to/data --pg-port 5432 --flight-port 50051

# Disable one server
apexbase-serve --dir /path/to/data --no-flight   # PG Wire only
apexbase-serve --dir /path/to/data --no-pg       # Arrow Flight only
Flag Default Description
--dir, -d . Directory containing .apex database files
--host 127.0.0.1 Bind host for both servers
--pg-port 5432 PostgreSQL Wire port
--flight-port 50051 Arrow Flight gRPC port
--no-pg Disable PG Wire server
--no-flight Disable Arrow Flight server

PostgreSQL Wire Protocol Server

ApexBase includes a built-in PostgreSQL wire protocol server, allowing you to connect using DBeaver, psql, DataGrip, pgAdmin, Navicat, and any other tool that supports the PostgreSQL protocol.

Starting the Server

Method 1: Python CLI (after pip install apexbase)

apexbase-server --dir /path/to/data --port 5432

Options:

Flag Default Description
--dir, -d . Directory containing .apex database files
--host 127.0.0.1 Host to bind to (use 0.0.0.0 for remote access)
--port, -p 5432 Port to listen on

Method 2: Standalone Rust binary (no Python required)

# Build
cargo build --release --bin apexbase-server --no-default-features --features server

# Run
./target/release/apexbase-server --dir /path/to/data --port 5432

Connecting with Database Tools

The server emulates PostgreSQL 15.0, reports a pg_catalog and information_schema compatible metadata layer, and supports both SimpleQuery and Extended Query protocols (prepared statements; binary result format for psycopg3). No username or password is required (authentication is disabled).

DBeaver

  1. New Database Connection → choose PostgreSQL
  2. Fill in connection details:
    • Host: 127.0.0.1 (or the --host you specified)
    • Port: 5432 (or the --port you specified)
    • Database: apexbase (any value accepted)
    • Authentication: select No Authentication or leave username/password empty
  3. Click Test ConnectionFinish
  4. DBeaver will discover tables and columns automatically via pg_catalog / information_schema

psql

psql -h 127.0.0.1 -p 5432 -d apexbase

DataGrip / IntelliJ IDEA

  1. Database tool window → +Data SourcePostgreSQL
  2. Set Host, Port, Database as above; leave User and Password empty
  3. Click Test ConnectionOK

pgAdmin

  1. Add New ServerGeneral tab: give it a name
  2. Connection tab: set Host and Port; leave Username as postgres (ignored) and Password empty
  3. Save — tables appear under Databases > apexbase > Schemas > public > Tables

Navicat for PostgreSQL

  1. ConnectionPostgreSQL
  2. Set Host, Port; leave User and Password blank
  3. Test ConnectionOK

Other Compatible Tools

Any tool or library that speaks the PostgreSQL wire protocol (libpq) can connect, including:

  • TablePlus, Beekeeper Studio, Heidisql
  • Python: psycopg2 / asyncpg
  • Node.js: pg (node-postgres)
  • Go: pgx / lib/pq
  • Rust: tokio-postgres / sqlx
  • Java: JDBC PostgreSQL driver

Example with psycopg2:

import psycopg2

conn = psycopg2.connect(host="127.0.0.1", port=5432, dbname="apexbase")
cur = conn.cursor()
cur.execute("SELECT * FROM users LIMIT 10")
print(cur.fetchall())
conn.close()

Supported SQL over Wire Protocol

The wire protocol server passes SQL directly to the ApexBase query engine. All SQL features listed in Usage Guide are available, including JOINs, CTEs, window functions, transactions, and DDL.

Metadata Compatibility

The server implements a pg_catalog compatibility layer that responds to common catalog queries:

Catalog / View Purpose
pg_catalog.pg_namespace Schema listing
pg_catalog.pg_database Database listing
pg_catalog.pg_class Table discovery
pg_catalog.pg_attribute Column metadata
pg_catalog.pg_type Type information
pg_catalog.pg_settings Server settings
information_schema.tables Standard table listing
information_schema.columns Standard column listing
SET / SHOW statements Client configuration probes

This enables GUI tools to browse tables, inspect columns, and display data types without modification.

Supported Protocol Features

Feature Status
Simple Query Protocol ✅ Fully supported
Extended Query Protocol (prepared statements) ✅ Supported — schema cached, binary format for psycopg3
Cross-database SQL (db.table) ✅ Supported — USE dbname / \c dbname to switch context
pg_catalog / information_schema ✅ Compatible layer for GUI tools
All ApexBase SQL (JOINs, CTEs, window functions, DDL) ✅ Full pass-through to query engine

Limitations

  • Authentication is not implemented — the server accepts all connections regardless of username/password
  • SSL/TLS is not supported — use an SSH tunnel (ssh -L 5432:127.0.0.1:5432 user@host) for remote access

Arrow Flight gRPC Server

Arrow Flight sends Arrow IPC RecordBatch directly over gRPC (HTTP/2), bypassing per-row text serialization entirely. It is 4–7× faster than PG wire for large result sets (10K+ rows).

Query PG Wire Arrow Flight Speedup
SELECT 10K rows 5.1ms 0.7ms 7× faster
BETWEEN (~33K rows) 22ms 5.6ms 4× faster
Single row / point lookup ~7.5ms ~7.9ms equal

Starting the Flight Server

Python CLI:

apexbase-flight --dir /path/to/data --port 50051

Standalone Rust binary:

cargo build --release --bin apexbase-flight --no-default-features --features flight
./target/release/apexbase-flight --dir /path/to/data --port 50051

Python Client

import pyarrow.flight as fl
import pandas as pd

client = fl.connect("grpc://127.0.0.1:50051")

# SELECT — returns Arrow Table
table = client.do_get(fl.Ticket(b"SELECT * FROM users LIMIT 10000")).read_all()
df = table.to_pandas()              # zero-copy to pandas
pl_df = pl.from_arrow(table)        # zero-copy to polars

# DML / DDL
client.do_action(fl.Action("sql", b"INSERT INTO users (name, age) VALUES ('Alice', 30)"))
client.do_action(fl.Action("sql", b"CREATE TABLE logs (event STRING, ts INT64)"))

# List available actions
for action in client.list_actions():
    print(action.type, "—", action.description)

When to Use Arrow Flight vs PG Wire

Scenario Recommendation
DBeaver / Tableau / BI tools PG Wire (only option)
Python + small queries (<100 rows) Native API (fastest, in-process)
Python + large queries (10K+ rows, remote) Arrow Flight (4–7× faster than PG wire)
Go / Java / Spark workers Arrow Flight (native Arrow support)
Local Python (same machine) Native API (ApexClient.execute())

PyO3 Python API

Both servers are also accessible as blocking Python functions (released GIL):

import threading
from apexbase._core import start_pg_server, start_flight_server

t1 = threading.Thread(target=start_pg_server,     args=("/data", "0.0.0.0", 5432),  daemon=True)
t2 = threading.Thread(target=start_flight_server, args=("/data", "0.0.0.0", 50051), daemon=True)
t1.start()
t2.start()

Rust Native API

ApexBase can be used directly from Rust as a zero-overhead embedded database — no Python, no FFI, no server process required. The full SQL engine, Arrow-native query results, SIMD vector search, FTS, and transactions are all available from the same Rust API.

Cargo Dependency

[dependencies]
# Local checkout
apexbase = { path = "path/to/ApexBase", default-features = false }

# Git
apexbase = { git = "https://github.com/BirchKwok/ApexBase.git", default-features = false }

default-features = false disables PyO3/numpy and significantly reduces compile time. Add features = ["server"] or features = ["flight"] if you also need the wire protocol servers.

Rust Quick Start

use apexbase::embedded::{ApexDB, Row};
use apexbase::data::Value;
use apexbase::storage::DurabilityLevel;
use apexbase::storage::on_demand::ColumnType;
use std::collections::HashMap;

fn main() -> apexbase::Result<()> {
    // Open (or create) a database
    let db = ApexDB::builder("./data")
        .durability(DurabilityLevel::Fast)
        .build()?;

    // Create a table with a predefined schema
    let users = db.create_table_with_schema("users", &[
        ("name".to_string(),  ColumnType::String),
        ("age".to_string(),   ColumnType::Int64),
        ("score".to_string(), ColumnType::Float64),
        ("city".to_string(),  ColumnType::String),
    ])?;

    // Insert rows
    let id = users.insert([
        ("name".to_string(),  Value::String("Alice".to_string())),
        ("age".to_string(),   Value::Int64(30)),
        ("score".to_string(), Value::Float64(92.5)),
        ("city".to_string(),  Value::String("Beijing".to_string())),
    ].into_iter().collect())?;

    // Batch insert 1 000 rows
    let bulk: Vec<Row> = (0..1_000i64).map(|i| {
        [("name".to_string(),  Value::String(format!("user_{i}"))),
         ("age".to_string(),   Value::Int64(20 + i % 50)),
         ("score".to_string(), Value::Float64(50.0 + (i % 50) as f64)),
         ("city".to_string(),  Value::String(if i % 2 == 0 { "A".to_string() } else { "B".to_string() })),
        ].into_iter().collect()
    }).collect();
    users.insert_batch(&bulk)?;

    // Full SQL query → Arrow RecordBatch
    let rs = users.execute(
        "SELECT city, COUNT(*) AS n, AVG(score) AS avg
         FROM users GROUP BY city ORDER BY n DESC"
    )?;
    let batch = rs.to_record_batch()?;
    println!("{} rows × {} columns", batch.num_rows(), batch.num_columns());

    // Or Vec<HashMap<String, Value>>
    let rows = users.execute("SELECT * FROM users WHERE age > 28 LIMIT 5")?.to_rows()?;
    for row in &rows {
        println!("{:?}", row.get("name"));
    }

    // Point lookup by _id
    if let Some(row) = users.retrieve(id)? {
        println!("Retrieved: {:?}", row.get("name"));
    }

    // O(1) row count
    println!("Total rows: {}", users.count()?);

    // Schema changes
    users.add_column("active", apexbase::data::DataType::Bool)?;
    println!("Columns: {:?}", users.columns()?);

    Ok(())
}

Run the full working example:

cargo run --example embedded --no-default-features

Key Rust Types

Type Import path Description
ApexDB apexbase::embedded::ApexDB Database handle — Clone + Send + Sync
ApexDBBuilder apexbase::embedded::ApexDB (via ApexDB::builder) Builder with durability / drop options
Table apexbase::embedded::Table Table-scoped operations — Clone + Send + Sync
ResultSet apexbase::embedded::ResultSet Query result (Arrow RecordBatch or scalar)
Row apexbase::embedded::Row HashMap<String, Value>
Value apexbase::data::Value Int64 / Float64 / String / Bool / Binary / FixedList / Null
ColumnType apexbase::storage::on_demand::ColumnType Schema type for create_table_with_schema
DataType apexbase::data::DataType Schema type for add_column / schema()
DurabilityLevel apexbase::storage::DurabilityLevel Fast / Safe / Max

For the full Rust API reference — all methods, transactions, FTS, vector search, concurrency patterns, and performance notes — see docs/RUST_EMBEDDED_API.md.


Architecture

Python (ApexClient)
  |
  |-- Arrow IPC / columnar dict --------> ResultView (Pandas / Polars / PyArrow)
  |
Rust Core (PyO3 bindings)
  |
  +-- SQL Parser -----> Query Planner -----> Query Executor
  |                                              |
  |   +-- JIT Compiler (Cranelift)               |
  |   +-- Expression Evaluator (70+ functions)   |
  |   +-- Window Function Engine                 |
  |                                              |
  +-- Storage Engine                             |
  |     +-- V4 Row Group Format (.apex)          |
  |     +-- DeltaStore (cell-level updates)      |
  |     +-- WAL (write-ahead log)                |
  |     +-- Mmap on-demand reads                 |
  |     +-- LZ4 / Zstd compression              |
  |     +-- Dictionary encoding                  |
  |                                              |
  +-- Index Manager (B-Tree, Hash)               |
  +-- TxnManager (OCC + MVCC)                    |
  +-- NanoFTS (full-text search)                  |
  +-- PG Wire Protocol Server (pgwire)             |
  |   +-- DBeaver / psql / DataGrip / pgAdmin      |
  |   +-- pg_catalog & information_schema compat    |
  |                                                 |
  +-- Arrow Flight gRPC Server (tonic + HTTP/2)     |
      +-- pyarrow.flight / Go / Java / Spark        |
      +-- Arrow IPC — zero serialization overhead   |

Storage Format

ApexBase uses a custom V4 Row Group format:

  • Each table is a single .apex file containing a header, row groups, and a footer
  • Row groups store columns contiguously with per-column compression (LZ4 or Zstd)
  • Low-cardinality string columns are dictionary-encoded on disk
  • Null bitmaps are stored per column per row group
  • A DeltaStore file (.deltastore) holds cell-level updates that are merged on read and compacted automatically
  • WAL records provide crash recovery with idempotent replay

Query Execution

  • The SQL parser produces an AST that the query planner analyzes for optimization strategy
  • Fast paths bypass the full executor for common patterns (COUNT(*), SELECT * LIMIT N, point lookups, single-column GROUP BY)
  • Arrow RecordBatch is the internal data representation; results flow to Python via Arrow IPC with zero-copy when possible
  • Repeated identical read queries are served from an in-process result cache

API Reference

ApexClient

Constructor

ApexClient(
    dirpath="./data",           # data directory
    drop_if_exists=False,       # clear existing data on open
    batch_size=1000,            # batch size for operations
    enable_cache=True,          # enable query cache
    cache_size=10000,           # cache capacity
    prefer_arrow_format=True,   # prefer Arrow format for results
    durability="fast",          # "fast" | "safe" | "max"
)

Database Management

Method Description
use_database(database='default') Switch to a named database (creates it if needed)
use(database='default', table=None) Switch database and optionally select/create a table
list_databases() List all databases ('default' always included)
current_database Property: current database name

Table Management

Method Description
create_table(name, schema=None) Create a new table, optionally with pre-defined schema
drop_table(name) Drop a table
use_table(name) Switch active table
list_tables() List all tables in the current database
current_table Property: current table name

Temporary Tables

Method Description
register_temp_table(name, file_path) Parse a CSV/JSON/Parquet file and register as a native temp table
drop_temp_table(name) Drop a temp table

Data Storage

Method Description
store(data) Store data (dict, list, DataFrame, Arrow Table)
from_pandas(df, table_name=None) Import from pandas DataFrame
from_polars(df, table_name=None) Import from polars DataFrame
from_pyarrow(table, table_name=None) Import from PyArrow Table

Data Retrieval

Method Description
execute(sql) Execute SQL statement(s)
query(where, limit) Query with WHERE expression
retrieve(id) Get record by _id
retrieve_many(ids) Get multiple records by _id
retrieve_all() Get all records
count_rows(table) Count rows in table

Data Modification

Method Description
replace(id, data) Replace a record
batch_replace({id: data}) Batch replace records
delete(id) or delete([ids]) Delete record(s)

Column Operations

Method Description
add_column(name, type) Add a column
drop_column(name) Drop a column
rename_column(old, new) Rename a column
get_column_dtype(name) Get column data type
list_fields() List all fields

Full-Text Search

Method Description
init_fts(fields, lazy_load, cache_size) Initialize FTS
search_text(query) Search documents
fuzzy_search_text(query) Fuzzy search
search_and_retrieve(query, limit, offset) Search and return records
search_and_retrieve_top(query, n) Top N results
get_fts_stats() FTS statistics
disable_fts() / drop_fts() Disable or drop FTS

Vector Search

Method Description
topk_distance(col, query, k=10, metric='l2', id_col='_id', dist_col='dist') Single-query TopK: returns ResultView with id and distance columns
batch_topk_distance(col, queries, k=10, metric='l2') Batch TopK: ndarray of shape (N, k, 2) — ids and distances

Utility

Method Description
flush() Flush data to disk
set_auto_flush(rows, bytes) Set auto-flush thresholds
get_auto_flush() Get auto-flush config
estimate_memory_bytes() Estimate memory usage
close() Close the client

ResultView

Method / Property Description
to_pandas(zero_copy=True) Convert to pandas DataFrame
to_polars() Convert to polars DataFrame
to_arrow() Convert to PyArrow Table
to_dict() Convert to list of dicts
scalar() Get single scalar value
first() Get first row as dict
get_ids(return_list=False) Get record IDs
shape (rows, columns)
columns Column names
__len__() Row count
__iter__() Iterate over rows
__getitem__(idx) Index access

Documentation

Additional documentation is available in the docs/ directory.

License

Apache-2.0

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

apexbase-1.17.0.tar.gz (1.0 MB view details)

Uploaded Source

Built Distributions

If you're not sure about the file name format, learn more about wheel file names.

apexbase-1.17.0-cp313-cp313-win_amd64.whl (9.4 MB view details)

Uploaded CPython 3.13Windows x86-64

apexbase-1.17.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (9.2 MB view details)

Uploaded CPython 3.13manylinux: glibc 2.17+ x86-64

apexbase-1.17.0-cp313-cp313-macosx_11_0_arm64.whl (8.2 MB view details)

Uploaded CPython 3.13macOS 11.0+ ARM64

apexbase-1.17.0-cp312-cp312-win_amd64.whl (9.4 MB view details)

Uploaded CPython 3.12Windows x86-64

apexbase-1.17.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (9.2 MB view details)

Uploaded CPython 3.12manylinux: glibc 2.17+ x86-64

apexbase-1.17.0-cp312-cp312-macosx_11_0_arm64.whl (8.2 MB view details)

Uploaded CPython 3.12macOS 11.0+ ARM64

apexbase-1.17.0-cp311-cp311-win_amd64.whl (9.4 MB view details)

Uploaded CPython 3.11Windows x86-64

apexbase-1.17.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (9.2 MB view details)

Uploaded CPython 3.11manylinux: glibc 2.17+ x86-64

apexbase-1.17.0-cp311-cp311-macosx_11_0_arm64.whl (8.2 MB view details)

Uploaded CPython 3.11macOS 11.0+ ARM64

apexbase-1.17.0-cp310-cp310-win_amd64.whl (9.4 MB view details)

Uploaded CPython 3.10Windows x86-64

apexbase-1.17.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (9.2 MB view details)

Uploaded CPython 3.10manylinux: glibc 2.17+ x86-64

apexbase-1.17.0-cp310-cp310-macosx_11_0_arm64.whl (8.2 MB view details)

Uploaded CPython 3.10macOS 11.0+ ARM64

apexbase-1.17.0-cp39-cp39-win_amd64.whl (9.4 MB view details)

Uploaded CPython 3.9Windows x86-64

apexbase-1.17.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (9.2 MB view details)

Uploaded CPython 3.9manylinux: glibc 2.17+ x86-64

apexbase-1.17.0-cp39-cp39-macosx_11_0_arm64.whl (8.2 MB view details)

Uploaded CPython 3.9macOS 11.0+ ARM64

File details

Details for the file apexbase-1.17.0.tar.gz.

File metadata

  • Download URL: apexbase-1.17.0.tar.gz
  • Upload date:
  • Size: 1.0 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.15

File hashes

Hashes for apexbase-1.17.0.tar.gz
Algorithm Hash digest
SHA256 a1b6c29d1b3602feece36f648fbf0276b21424b29fd7abefaaf9ba8843e92878
MD5 3fd0450bb3a4992369ad09068177b18d
BLAKE2b-256 92671a083eb32cdcb9f8ca20ade7cb0ba9a5ac1905fdcfbb2d8b693ddacaafaf

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp313-cp313-win_amd64.whl.

File metadata

  • Download URL: apexbase-1.17.0-cp313-cp313-win_amd64.whl
  • Upload date:
  • Size: 9.4 MB
  • Tags: CPython 3.13, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.15

File hashes

Hashes for apexbase-1.17.0-cp313-cp313-win_amd64.whl
Algorithm Hash digest
SHA256 fe6f0508774c0e9d023c48fa99ee6132eed7aa51e5ccc5b796d6ff0509be8a5a
MD5 f25ae6969659fadedec21ceb09e6e0ae
BLAKE2b-256 8f405d3461169e84e87fa924e00f22e99889b2d3bba6ff373aa9b430aad8c997

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for apexbase-1.17.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 efb6ba0cdd447b58398e7cd2517b13895fd696123e6ac990672de8d45273824a
MD5 f73f516ba39a7bcd3de0cfd5387ef70d
BLAKE2b-256 efd5900294a2eb5453fda8722786c6b0b6a60bf1b3b9f404c1f426f9d6c2b3b7

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp313-cp313-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for apexbase-1.17.0-cp313-cp313-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 4eca0b5a484f26e8543323a092012a40021605390f622e88e3ea4ddbea2001c9
MD5 ea6022b8a813a3793e4d107e4d835fca
BLAKE2b-256 de7d8bfc0d08ca1bf0174d90b768652feef220d9ac1d5342cd5542f032e143fb

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp312-cp312-win_amd64.whl.

File metadata

  • Download URL: apexbase-1.17.0-cp312-cp312-win_amd64.whl
  • Upload date:
  • Size: 9.4 MB
  • Tags: CPython 3.12, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.15

File hashes

Hashes for apexbase-1.17.0-cp312-cp312-win_amd64.whl
Algorithm Hash digest
SHA256 b34f5f1a2c30fbb9f4a038372683515fa7639efd6488ec7d5335b38a7aa52821
MD5 1f8c2c46b5fccf3feeec5d7cac93d224
BLAKE2b-256 fdcb6885402208b65bb36d5081cb847985a983d1068adcfd85e7bbcae46a2268

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for apexbase-1.17.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 db623690dd36c25db7cff043d374cda12df3d59e670625cd3ce6419b7ff8106e
MD5 55a349384c1294606eadf84e972e7f76
BLAKE2b-256 f90a626275c969c475d204f0d4c0d9dba7e18c7a41934525a1911d18d47b63c2

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp312-cp312-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for apexbase-1.17.0-cp312-cp312-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 26f4e061b1b19662f505ecc0ddd4c68851cb6a3729106b5ce2cf5d18da1b9d00
MD5 b8520f2f2f32588153f66c5e5849091b
BLAKE2b-256 3729867cb7ef5ed3a7e421d7f877981c5a894e381131c21e62f0921ab472d7e4

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp311-cp311-win_amd64.whl.

File metadata

  • Download URL: apexbase-1.17.0-cp311-cp311-win_amd64.whl
  • Upload date:
  • Size: 9.4 MB
  • Tags: CPython 3.11, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.15

File hashes

Hashes for apexbase-1.17.0-cp311-cp311-win_amd64.whl
Algorithm Hash digest
SHA256 2d1d145d71ec7588850bac9f4190f926579ff71d61d19fbec5902ea8384a8baa
MD5 11eee052f8495282018cc3a8aecde2f6
BLAKE2b-256 6a1083439d9ba82a8b8d74fe6fcdc68eb17b1d4b7b7a2ab5a048ff80a8531b28

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for apexbase-1.17.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 1e767d3e08a247ce987bb9428ae546f5c76d19664b08a6f8f8b57f098667f28c
MD5 f51419208ab40074ecb6148c6619db82
BLAKE2b-256 409c8cce7c31ea77e4c9517e16f8fceeebac0710ddcb086865dd1c3cee7fba46

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp311-cp311-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for apexbase-1.17.0-cp311-cp311-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 5eeb7ac81b6b30b0a4e6a9feb9af4c09ca025ad9d6e6365afe9073fe894e553b
MD5 703e0b91a2b1f236e7e3a402faca8bde
BLAKE2b-256 702c1b73ac442fa610e7427bf0a8e40a06d0e280be6c6922d425ee8fd4a7e094

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp310-cp310-win_amd64.whl.

File metadata

  • Download URL: apexbase-1.17.0-cp310-cp310-win_amd64.whl
  • Upload date:
  • Size: 9.4 MB
  • Tags: CPython 3.10, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.15

File hashes

Hashes for apexbase-1.17.0-cp310-cp310-win_amd64.whl
Algorithm Hash digest
SHA256 4b2cd05a9e809ba692c82d99bc0c638d92de279a1a46e72c24e2d42ec342d3d0
MD5 2a6ecdc4ac4c874cb4acd258f5bab45c
BLAKE2b-256 35454c304b1b7466567c3fa677363ce0c88b0bd6fab06251de7890840966933b

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for apexbase-1.17.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 9cecbe3f63d85eb2762a2bce049f54fb22ffbf0515c3861b753a6e9d5be74622
MD5 254e008fd8e99e9125e04657c36c7502
BLAKE2b-256 78b30263c884feb115d918a1a8f3ca445ed7e5655c8c59c4b8419bb4fd78a982

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp310-cp310-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for apexbase-1.17.0-cp310-cp310-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 de0935315c932d276ac5eddbff821c921b34200824b7a72a565a2ddc1569fc60
MD5 583da06f15a192afd8d8b635e8117602
BLAKE2b-256 6928c94cc3514c909623f6c43034eda46a3e61fcadef57d0ef93d05c5ba6d6a5

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp39-cp39-win_amd64.whl.

File metadata

  • Download URL: apexbase-1.17.0-cp39-cp39-win_amd64.whl
  • Upload date:
  • Size: 9.4 MB
  • Tags: CPython 3.9, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.15

File hashes

Hashes for apexbase-1.17.0-cp39-cp39-win_amd64.whl
Algorithm Hash digest
SHA256 8290f16eb3c7a0f93e2018590c20a6aef4bfafe306ca0a1ba107a2d62882b975
MD5 bc970c1f9f9b612640a56d54f6ddff99
BLAKE2b-256 4970efc39673b3b60015cbcfcb3a4075dd3453b35706f73c3a7833c18e7183aa

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for apexbase-1.17.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 578c4d92eb509d93a909c481e38595c94bb60d4541bb92fbee9d5e25161a4ce3
MD5 5dadda55631daa7cbe2f38fb3721553d
BLAKE2b-256 71279a7641972118b34a476cf49c2420b06cc2079305e4a3cdbc349035ce1220

See more details on using hashes here.

File details

Details for the file apexbase-1.17.0-cp39-cp39-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for apexbase-1.17.0-cp39-cp39-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 31016fc7c9664c81001a7be3241bdc6eb055bb2262b9d0cbe9f0f137080f880c
MD5 371fffee575b1901e8649ec428258464
BLAKE2b-256 192947f2d78984e7aae10a2def0c6a74ad78b61138a6aab31134ad80afa02e12

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page