Skip to main content

AI-powered data exploration with natural language

Project description

datasight

Status: early and evolving. This project is in active development and the code is changing rapidly — APIs, CLI flags, and behavior may shift between commits. Feedback and bug reports from users are very welcome; please open an issue on GitHub.

AI-powered data exploration with natural language.

datasight connects an AI agent to your database and provides a web UI where you can ask questions in natural language. The agent writes SQL, runs queries, and generates interactive Plotly visualizations.

Supports DuckDB, PostgreSQL, SQLite, and Flight SQL databases. Also queries local CSV and Parquet files directly — no database setup required. Supports Anthropic Claude (default), GitHub Models (open source), and Ollama (local) as LLM backends.

Quick start

uv tool install datasight

# Create a new project
mkdir my-project && cd my-project
datasight init

# Edit .env with your API key and database path
# Edit schema_description.md to describe your data
# Edit queries.yaml with example questions

# Run the web UI
datasight run

Open http://localhost:8084 and start asking questions.

Explore CSV or Parquet files with no setup

# Launch the web UI with no project, then paste a file or directory path
# into "Explore Files" to create views automatically
datasight run

# Or inspect a file from the command line (schema, row count, column stats)
datasight inspect generation.parquet

# Or build a persistent project from CSV/Parquet inputs
datasight generate generation.csv plants.csv --db-path grid.duckdb

Or ask from the command line without starting a server:

datasight ask "What are the top 10 records?"
datasight ask "Show trends by year" --chart-format html -o chart.html
datasight profile
datasight quality --format markdown -o quality.md
datasight ask --file questions.txt --output-dir batch-output

Features

  • Natural language queries — ask questions in English, get SQL + results
  • Interactive charts — Plotly visualizations with chart-type switching
  • Multiple databases — DuckDB, PostgreSQL, SQLite, and Flight SQL
  • Query files directly — point at a local CSV or Parquet file (or directory) and start asking questions; datasight creates DuckDB views on the fly
  • Headless CLIdatasight ask runs queries without a web server
  • Deterministic CLI workflows — profile, quality, dimension, trend, and recipe commands that do not require an LLM
  • Schema browser — sidebar with tables, columns, and example queries
  • Schema auto-discovery — tables, columns, and types detected automatically
  • Domain context — describe your data in Markdown for better AI understanding
  • Example queries — seed the AI with question/SQL pairs
  • Reusable prompt recipes — project-specific analysis prompts derived from the schema
  • Multi-chart dashboard — pin results, filter cards, and configure layouts
  • Session export — export conversations as shareable HTML pages
  • Keyboard shortcuts? to see all shortcuts, / to focus input
  • Streaming responses — real-time SSE streaming from the LLM

Architecture

datasight pairs a FastAPI backend with a Svelte 5 + TypeScript + Tailwind CSS frontend built with Vite. It supports multiple LLM backends — Anthropic (default), OpenAI, GitHub Models, and Ollama — selectable via LLM_PROVIDER in .env.

datasight run / datasight ask / datasight profile / datasight quality
  → LLM provider (Anthropic / OpenAI / GitHub Models / Ollama)
    → DuckDB / PostgreSQL / SQLite / Flight SQL (or CSV/Parquet via DuckDB views)
    → Plotly chart generator
  → Web UI (SSE streaming) or CLI output

Documentation

uv sync --extra dev
. .venv/bin/activate
zensical serve
zensical build
python scripts/generate_cli_reference.py

Development Tests

# Build frontend assets for FastAPI serving after a clean checkout
bash scripts/build-frontend.sh

# Python test suite
pytest

# CI-safe Python test suite, excluding tests that need local Ollama
pytest -m "not integration"

# Frontend unit tests (Vitest)
cd frontend && npm test

# Frontend E2E tests (Playwright, requires datasight run)
cd frontend && npm run test:e2e

# Rebuild frontend for FastAPI serving after frontend changes
bash scripts/build-frontend.sh

Generated web assets under src/datasight/web/static/ and src/datasight/web/templates/index.html are not checked in. Run bash scripts/build-frontend.sh before using datasight run from a clean checkout when you want FastAPI to serve the production UI.

Ollama-backed CLI tests are marked integration because they require a running local Ollama server with the qwen2.5:7b model available. CI runs pytest -m "not integration"; run pytest -m integration locally when you want to exercise the live LLM path.

Software Record

datasight is developed under NLR Software Record SWR-26-045.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

datasight-0.5.0.tar.gz (3.3 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

datasight-0.5.0-py3-none-any.whl (1.9 MB view details)

Uploaded Python 3

File details

Details for the file datasight-0.5.0.tar.gz.

File metadata

  • Download URL: datasight-0.5.0.tar.gz
  • Upload date:
  • Size: 3.3 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for datasight-0.5.0.tar.gz
Algorithm Hash digest
SHA256 b020eda29227e86bb33b95400086ad8a4d1cd6763f2591454fc421fb57a00ec8
MD5 48b3b2f568058af34e1be0cc434043cf
BLAKE2b-256 90c79013339b10ac31293f829a4090984b535925c79ededf4caa5940780ee311

See more details on using hashes here.

Provenance

The following attestation bundles were made for datasight-0.5.0.tar.gz:

Publisher: release.yml on dsgrid/datasight

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file datasight-0.5.0-py3-none-any.whl.

File metadata

  • Download URL: datasight-0.5.0-py3-none-any.whl
  • Upload date:
  • Size: 1.9 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for datasight-0.5.0-py3-none-any.whl
Algorithm Hash digest
SHA256 d8869a5df555b7130bfd0bd2ad9f82f03c5174af2edb2c0d9ac0c146c1fe3be8
MD5 44f2abbd635a370957f0bac9f4cf46f2
BLAKE2b-256 a9f6845a4db68c23b5c7b834d64cb1b8a41dfae4a6a32e8a077ffd627d8c435d

See more details on using hashes here.

Provenance

The following attestation bundles were made for datasight-0.5.0-py3-none-any.whl:

Publisher: release.yml on dsgrid/datasight

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page