Skip to main content

Native Polars computation engine for Graphic Walker — translates GW workflow payloads directly to Polars operations

Project description

polars-gw

Native Polars computation engine for Graphic Walker.

Translates Graphic Walker IDataQueryPayload workflow steps directly into Polars operations — no DuckDB, no SQL intermediate.

Installation

pip install polars-gw              # core translator only
pip install 'polars-gw[viz]'       # core + built-in walk() UI

The core install only pulls in polars. The [viz] extra adds FastAPI, uvicorn and pydantic so you can launch a local Graphic Walker against a DataFrame in one line.

Interactive UI

import polars as pl
from polars_gw import walk

df = pl.read_parquet("sales.parquet")
handle = walk(df)          # opens http://127.0.0.1:<free-port> in your browser
# ...
handle.stop()

walk() starts a FastAPI server in a background daemon thread, serves the Graphic Walker UI bundled inside the wheel (no CDN, no network required), and wires its computation callback to execute_workflow.

Export / Import chart specs

Charts you build in the UI can be saved and restored across sessions:

# Export the current charts to a JSON file
handle.export("my_charts.json")

# Or just grab the spec as a Python list
spec = handle.export()

Next time, pass the file back to walk() to restore exactly where you left off:

handle = walk(df, spec_file="my_charts.json")

The spec file is a plain JSON array of Graphic Walker IChart objects — safe to version-control or share with collaborators.

Logs

By default walk() prints one line per compute call so you can see what GW is asking the backend for:

13:42:11  INFO    polars_gw.viz: Graphic Walker running on http://127.0.0.1:54221 — 12 000 000 rows x 19 cols, max_rows=1000000
13:42:18  INFO    polars_gw.viz: compute: 2 step(s) → 847 row(s) in 142.3 ms
13:42:24  INFO    polars_gw.viz: compute: 1 step(s) → 1000000 row(s) in 1284.7 ms [CAPPED]

[CAPPED] means the result hit max_rows and was truncated — Graphic Walker shows its "Data Limit Reached" toast in the UI when this happens. Default cap is 1 000 000 rows; tune it with max_rows= or disable with max_rows=None.

Quiet things down with log_level="warning", or get more detail with log_level="debug":

walk(df, log_level="warning")   # only warnings and errors
walk(df, max_rows=None)         # no row cap (use with care on big data)

If you've already configured Python logging yourself (logging.basicConfig, a framework, pytest's caplog, etc.), walk() won't overwrite it — it only attaches a console handler when nothing else owns logging.

Usage

import polars as pl
from polars_gw import execute_workflow, get_fields

df = pl.read_csv("data.csv")

# Get field definitions for Graphic Walker
fields = get_fields(df)
# [{"fid": "city", "name": "city", "semanticType": "nominal", "analyticType": "dimension"}, ...]

# Execute a Graphic Walker computation payload
payload = {
    "workflow": [
        {"type": "filter", "filters": [
            {"fid": "age", "rule": {"type": "range", "value": [18, 65]}}
        ]},
        {"type": "view", "query": [
            {"op": "aggregate", "groupBy": ["city"], "measures": [
                {"field": "salary", "agg": "mean", "asFieldKey": "avg_salary"}
            ]}
        ]},
        {"type": "sort", "sort": "descending", "by": ["avg_salary"]}
    ],
    "limit": 100
}
results = execute_workflow(df, payload)
# Returns list[dict] — ready to return as IRow[] to Graphic Walker

Supported Operations

Workflow Steps

Step Type Description
filter Range, temporal range, one of, not in, regexp
view/aggregate Group by + aggregation (sum, count, mean, median, min, max, variance, stdev, distinctCount)
view/fold Unpivot (wide to long)
view/bin Numeric binning
view/raw Column selection
sort Ascending/descending sort
transform Computed fields (bin, log/log2/log10, binCount, dateTimeDrill, dateTimeFeature, one, expr)

Field Inference

get_fields() maps Polars dtypes to Graphic Walker field types:

Polars Type Semantic Type Analytic Type
Float*, Decimal quantitative measure
Int*, UInt* quantitative dimension
Date, Datetime, Time, Duration temporal dimension
Utf8, Categorical, Boolean, etc. nominal dimension

How It Differs from PyGWalker

PyGWalker (and panel-graphic-walker) always route through DuckDB — even for Polars DataFrames:

DataFrame → DuckDB → SQL → execute → dicts

polars-gw translates directly to Polars operations:

DataFrame → Polars expressions → execute → dicts

No DuckDB dependency. No SQL intermediate. Just Polars.

polars-gw vs PyGWalker — which should you use?

Different tools, different jobs. polars-gw is a focused compute engine

  • a standalone browser UI. PyGWalker is a broader product with richer UX surface area.

Use polars-gw when

  • You're already in the Polars ecosystem (e.g. LazyFrames, streaming, or tools like Flowfile) and don't want a DuckDB round-trip in the middle.
  • You need type fidelity for Categorical, Decimal, Duration, List, Struct — they pass through as-is instead of degrading through SQL.
  • You want a lean install (~40 MB core, no DuckDB).
  • You're building your own frontend and just need the translator (execute_workflow + get_fields).
  • You want debuggable query plans (Polars expressions, not generated SQL).
  • You need a standalone browser UI (walk(df)) without notebook dependencies.

Use PyGWalker when

  • You want the Jupyter inline widget — PyGWalker renders the UI inside a notebook cell via anywidget. polars_gw.walk() pops a browser tab.
  • You need first-party framework integrations (Streamlit, Gradio, Dash). PyGWalker has these; polars-gw does not.
  • You rely on chart persistence — saving/loading chart specs, exporting HTML/PNG, vis_spec round-tripping.
  • You want heterogeneous input support (pandas + Polars + parquet + SQL tables under one DuckDB layer).
  • Your data is too large to collect into memory and you need DuckDB's battle-tested out-of-core execution.
  • You're using Kanaries cloud features (sharing, cloud chat, etc.).

At a glance

Concern polars-gw PyGWalker
Backend Polars expressions DuckDB SQL
Install weight Lean (polars only) DuckDB + heavier deps
Jupyter inline widget ❌ (browser tab)
Streamlit / Gradio / Dash
Standalone browser UI ✅ (walk())
Chart save/load/export ❌ (defers to GW client)
LazyFrame native via DuckDB
Polars Categorical/Decimal/Duration fidelity lossy through SQL
New GW payload ops Requires translator update Often works via SQL
Heterogeneous inputs (pandas/parquet/SQL) Polars-only

Short version: if you're all-in on Polars and want the fast, native path, use polars-gw. If you want inline-notebook, Streamlit, or chart persistence out of the box, use PyGWalker.

Development

Python

uv sync --extra viz           # runtime + viz + dev deps
uv run pytest                 # 60+ tests
uv run ruff check .

Bundling the viz assets

The walk() UI ships a pre-built JS/CSS bundle under polars_gw/viz_assets/ (committed to the repo) so end users don't need Node to pip install polars-gw[viz].

Maintainers rebuild when bumping Graphic Walker:

cd js
npm install
npm run build                 # one-shot production build

Or iterate with watch mode (rebuilds JS + CSS on save, source maps on):

npm run dev                   # in one shell
uv run python example/walk_demo.py   # in another — refresh browser to pick up changes

Bundle layout:

  • graphic-walker.js — Graphic Walker + React 19.2.0, minified IIFE (~4.4 MB)
  • graphic-walker.css — Tailwind-compiled stylesheet (~57 KB)
  • versions.json — pinned npm versions + build mode + timestamp

See js/README.md for details (including why React is pinned to exactly 19.2.0).

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

polars_gw-0.1.1.tar.gz (1.5 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

polars_gw-0.1.1-py3-none-any.whl (1.3 MB view details)

Uploaded Python 3

File details

Details for the file polars_gw-0.1.1.tar.gz.

File metadata

  • Download URL: polars_gw-0.1.1.tar.gz
  • Upload date:
  • Size: 1.5 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for polars_gw-0.1.1.tar.gz
Algorithm Hash digest
SHA256 f54973953684236723d95536a2fd9363265b058be1b688eb803345ad67d1fa4a
MD5 0604e1e54a879ff59785cd5a69314899
BLAKE2b-256 e46b215dd0bba5f33c0de20c0a2994f7d5927869d26e96ce32c827bae5d3c20b

See more details on using hashes here.

Provenance

The following attestation bundles were made for polars_gw-0.1.1.tar.gz:

Publisher: publish.yml on Edwardvaneechoud/polars-gw

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file polars_gw-0.1.1-py3-none-any.whl.

File metadata

  • Download URL: polars_gw-0.1.1-py3-none-any.whl
  • Upload date:
  • Size: 1.3 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for polars_gw-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 840c0f27d2baaf1a9686fce35522c76cef8d9e6fe87e5145ebd6ea62476c4c0f
MD5 db514bd3e4f27103a25f6d33952d408a
BLAKE2b-256 63b6f5dd609f3d2f52c744b4cc5c66be318a6e82092bee6410dcd98489b0426c

See more details on using hashes here.

Provenance

The following attestation bundles were made for polars_gw-0.1.1-py3-none-any.whl:

Publisher: publish.yml on Edwardvaneechoud/polars-gw

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page