Parseltongue: a DSL for systems which refuse to speak falsehood
Project description
Parseltongue
A DSL for systems that refuse to speak falsehood.
v0.7.0 — Seeing the Sentences. Interactive visualization, search engine with provenance, stack-based evaluator, pgmd notebooks. Install with
pip install 'parseltongue-dsl'or start with documentation.
Notebook example for this repo:
The engine inspects itself — a pgmd notebook with inline computed values, taint propagation, and derivation paths
Materialized knowledge graph of this notebook:
2000+ nodes rendered as an interactive graph — zoom, click any node for full evidence and derivation tree
Explore the engine demo live →
Rationale - Why?
LLMs are increasingly used for code review, security auditing, and documentation validation. The problem: they hallucinate. An LLM reviewing an authentication module might flag a "missing bcrypt implementation" that doesn't exist in the code, or miss the actual vulnerability — MD5 used for session IDs — while confidently producing a detailed critique. You get a fluent, plausible security report where some findings are real, some are fabricated, and you have no way to tell which is which without manually verifying every claim.
Parseltongue fixes this by making every claim provable. Instead of asking an LLM to produce a prose review, we ask it to encode the codebase as a formal logic system. Every extracted fact must cite a verbatim quote from the source code. Every conclusion must derive from stated premises. And every derivation is checked by a symbolic engine that doesn't hallucinate.
This gives you three things that prose reviews cannot:
-
Hallucination detection. Every claim traces back to a quote in the source. If the LLM fabricates a security issue — "passwords are hashed using bcrypt" when there's no bcrypt anywhere in the code — the quote verification fails. That failure propagates automatically to every conclusion that depends on it. You don't just catch the fabrication; you see everything it contaminates.
-
Specification compliance checking. Load a security spec alongside the implementation. The engine extracts requirements from the spec and facts from the code independently, then cross-validates via
diffdirectives. Wrong token expiry values, exceeded session limits, prohibited algorithms in use — every divergence is flagged with full provenance to both documents. -
Documentation validation. Run the engine against a library's README or API docs. Internal contradictions between prose and config tables, unverifiable security audit claims, inconsistencies between documented and actual behavior — all surface automatically with traceable evidence.
A few examples of what it looks like for this repository:
Focused subgraph with layers view — depth-based layout showing the full dependency structure
Derivation layers — every theorem traces back to source-quoted facts
See Discovered Use Cases for more real-world applications and experiment with them in live demos.
Quick Start
For a guided walkthrough, see the Quick Start page.
Agentic Quick Start
Parseltongue is designed to work with LLM coding agents. Paste this prompt into Claude Code, Cursor, or any agent with terminal access:
Run
pip install parseltongue-dslto install the Parseltongue formal reasoning library. It gives you a CLI calledpg. Runpg learn --helpto see available learning paths, then runpg learn <name>for each one listed and read the full output carefully — these are your operational guides for working with Parseltongue. Once you've read them, tell me what you can now do and ask me what I'd like to work on.
The agent installs, reads the learning scripts (kung-fu for bench mastery, to-connect for pgmd notebooks), and starts a conversation about what to do next. It now knows how to start the bench daemon, index files, search with provenance, screen for consistency, and generate interactive visualizations.
Quick Start page page contains instructions for Web UI agents, manual setup, and follow-up prompts.
Install Full Package
Install the full package with TUI, Docling document conversion, and standalone LLM verification pipelines. We recommend pipx for global access. Alternatively, install with pip in a virtual environment.
macOS
brew install pipx
pipx install 'parseltongue-dsl[cli]'
Linux (Ubuntu 23.04+ / Debian 12+)
sudo apt install pipx
pipx install 'parseltongue-dsl[cli]'
Linux (older)
python3 -m pip install --user pipx
pipx install 'parseltongue-dsl[cli]'
Windows
pip install pipx
pipx install "parseltongue-dsl[cli]"
Or with pip directly
pip install 'parseltongue-dsl[cli]'
Updating
pipx install 'parseltongue-dsl[cli]==0.7.0' --force # explicit version avoids pip cache issues
Running
The package installs two entry points:
parseltongue — interactive TUI and LLM verification pipeline. On first run, a configuration wizard asks for your API endpoint, key, and model. Any OpenAI-compatible endpoint works (OpenRouter, OpenAI, Azure, local servers like vLLM or Ollama). From the main menu: pick documents, type a question, and the pipeline runs four passes — extraction, blinded derivation, fact-checking, and answer generation. You can review, retry with feedback, or skip each pass interactively.
parseltongue # launch TUI
parseltongue run \ # or run directly
-d "auth.py" \
-d "Spec:api_spec.md" \
-q "Does the implementation match the specification?" \
--model anthropic/claude-sonnet-4.6
| Command | Description |
|---|---|
parseltongue |
Launch the interactive TUI |
parseltongue run -d ... -q ... |
Run pipeline directly on documents |
parseltongue inspect file.pdf |
Preview document conversion |
parseltongue history |
Browse past runs |
parseltongue configure |
Re-run the configuration wizard |
pg (alias pg-bench) — persistent workbench daemon. Loads a .pltg formal system into memory, indexes source documents, and serves queries over a Unix socket. Includes a full-text search engine with ngrams, stemming, BM25, and RRF ranking; structural lenses for graph navigation; consistency screening with issue/warning/dangling classification; holograms for side-by-side comparison of diff sides; time-travel history with checkpoint layers and non-destructive restore; and interactive HTML visualization with D3 graph, cards, and layers views.
pg start main.pltg --user "Alice" --assistant "Claude"
pg wait # block until ready
parseltongue bench # connect from TUI — everything in one screen
pg screen # consistency report
pg find "engine" # structural search across pltg nodes
pg search '(in "auth.py" "raise")' # full-text with S-expression query language
pg eval '(scope hologram (dissect "my-diff"))' # side-by-side diff comparison
pg history layers # checkpoint history
pg eval '(fmt "viz" (scope lens (find ".*")))' > viz.html # interactive visualization
Run pg learn --help to see built-in learning scripts that cover the full system.
Supports PDF, DOCX, PPTX, XLSX, HTML (via Docling), plus all plain text and code formats.
See the full CLI documentation for TUI navigation, keybindings, screenshots of every screen, and configuration details.
Python API
The LLM module extends Parseltongue to a neuro-symbolic approach over the symbolic formal reasoning core.
pip install 'parseltongue-dsl[llm]'
export OPENROUTER_API_KEY=sk-...
from parseltongue import System, Pipeline
from parseltongue.llm.openrouter import OpenRouterProvider
system = System(overridable=True)
provider = OpenRouterProvider()
pipeline = Pipeline(system, provider)
pipeline.add_document("Implementation", path="auth.py")
pipeline.add_document("Specification", path="api_spec.md")
result = pipeline.run("Does the implementation match the specification?")
result.output.markdown— grounded report with[[type:name]]references linking every claim to source quotesresult.output.references— resolved references: value, provenance chain, and source quotesresult.output.consistency— unverified evidence, fabrication chains, diff divergencesresult.system— the full formal system for inspection viasystem.provenance(name),system.eval_diff(name), etc.
See the full LLM pipeline documentation for the four-pass architecture, provider interface, extended thinking, and reference resolution.
Core Engine
The DSL that the pipeline builds under the hood. Five directive types — fact, axiom, defterm, derive, diff — each grounded in evidence with verbatim quotes. Can be used standalone without any LLM dependency.
pip install parseltongue-dsl
See the full core documentation for directive types, evidence grounding, quote verification, custom environments, and consistency checking.
Demos
15 demos ship with the library — 10 core (no LLM needed) and 5 LLM pipeline demos. Browse the live rendered demos →
# Software engineering — no LLM needed
python -m parseltongue.core.demos.code_check.demo # auth module security audit
python -m parseltongue.core.demos.spec_validation.demo # auth spec vs implementation
python -m parseltongue.core.demos.doc_validation.demo # auth library docs validation
# Research & math — no LLM needed
python -m parseltongue.core.demos.biomarkers.demo # cross-paper scientific conflict
python -m parseltongue.core.demos.revenue_reports.demo # cross-document analysis
python -m parseltongue.core.demos.apples.demo # Peano arithmetic from field notes
python -m parseltongue.core.demos.apples_pltg.demo # same in pure .pltg with splat axioms
python -m parseltongue.core.demos.apples_splats_pltg.demo # variadic splat patterns
# Governance & architecture — no LLM needed
python -m parseltongue.core.demos.data_governance_pltg.demo # five-layer compliance analysis
python -m parseltongue.core.demos.engine_overview.demo # engine self-inspection pgmd notebook
# LLM pipeline demos — requires API key
python -m parseltongue.llm.demos.code_check.demo # LLM auth module security audit
python -m parseltongue.llm.demos.spec_validation.demo # LLM auth spec vs implementation
python -m parseltongue.llm.demos.doc_validation.demo # LLM auth library docs validation
python -m parseltongue.llm.demos.biomarkers.demo # LLM biomarker analysis
python -m parseltongue.llm.demos.revenue.demo # LLM revenue reports
# CLI demo — run the pipeline on the included PDF
parseltongue run -d "parseltongue/cli/demo/nejm.pdf" -q "Find any inconsistencies or red flags."
Tests
pip install -e ".[dev,llm]"
pytest # all tests
pytest parseltongue/core/tests/ # core only
pytest parseltongue/llm/tests/ # llm only
License
Apache 2.0
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file parseltongue_dsl-0.7.1.tar.gz.
File metadata
- Download URL: parseltongue_dsl-0.7.1.tar.gz
- Upload date:
- Size: 1.0 MB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
3b12f77e4ba866123550e3ec681d85ecabdd4cfd646a1bb7c2769fb7629df508
|
|
| MD5 |
76bf6e30838c7ba9fa4435abbb842dec
|
|
| BLAKE2b-256 |
3433e3eab633f7b5339a0c04ec332b1d4238914013659fc1f95260adcf3c5865
|
Provenance
The following attestation bundles were made for parseltongue_dsl-0.7.1.tar.gz:
Publisher:
publish.yml on sci2sci-opensource/parseltongue
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
parseltongue_dsl-0.7.1.tar.gz -
Subject digest:
3b12f77e4ba866123550e3ec681d85ecabdd4cfd646a1bb7c2769fb7629df508 - Sigstore transparency entry: 1202381117
- Sigstore integration time:
-
Permalink:
sci2sci-opensource/parseltongue@35ef2b42e7ea531bb6f5418fc3dc16bc907e9495 -
Branch / Tag:
refs/tags/v0.7.1 - Owner: https://github.com/sci2sci-opensource
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@35ef2b42e7ea531bb6f5418fc3dc16bc907e9495 -
Trigger Event:
push
-
Statement type:
File details
Details for the file parseltongue_dsl-0.7.1-py3-none-any.whl.
File metadata
- Download URL: parseltongue_dsl-0.7.1-py3-none-any.whl
- Upload date:
- Size: 1.2 MB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
b288345c5fff12e5162c7cb64f674da4e5b1d6979c8c3f11bcf05e11dc4fe7e6
|
|
| MD5 |
b058474e355f0abc97186473b2998862
|
|
| BLAKE2b-256 |
2642c4fa3d7957003c2c320bf2a612729fc89321f7a0a4e2613d6cacd4e5b88a
|
Provenance
The following attestation bundles were made for parseltongue_dsl-0.7.1-py3-none-any.whl:
Publisher:
publish.yml on sci2sci-opensource/parseltongue
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
parseltongue_dsl-0.7.1-py3-none-any.whl -
Subject digest:
b288345c5fff12e5162c7cb64f674da4e5b1d6979c8c3f11bcf05e11dc4fe7e6 - Sigstore transparency entry: 1202381152
- Sigstore integration time:
-
Permalink:
sci2sci-opensource/parseltongue@35ef2b42e7ea531bb6f5418fc3dc16bc907e9495 -
Branch / Tag:
refs/tags/v0.7.1 - Owner: https://github.com/sci2sci-opensource
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@35ef2b42e7ea531bb6f5418fc3dc16bc907e9495 -
Trigger Event:
push
-
Statement type: