Skip to main content

Autonomous documentation-code misalignment detection using LLM agents

Project description

LyingDocs

LyingDocs

A trust layer for your repository.

Detect when your docs, code, configs, and examples stop agreeing with each other.


Modern repositories are read by more than humans.

They are read by teammates, new contributors, users, reviewers, downstream integrators — and increasingly by AI agents.

That only works if the repository can be trusted.

But trust quietly erodes over time:

  • documentation describes features that were never shipped
  • code behavior drifts away from the spec
  • examples stop matching reality
  • values claimed to be configurable are hardcoded deep in the codebase
  • papers and implementation tell different stories

LyingDocs is a trust layer for your repository.
It audits the gap between what your repo says and what your code actually does — before your users, contributors, or agents learn the wrong thing.


Why LyingDocs exists

Every codebase accumulates invisible trust debt.

In the age of fast iteration and LLM-assisted development, teams now ship code and documentation faster than ever — but not always in sync. A repo may still look polished while becoming progressively less reliable as a source of truth.

That is the problem LyingDocs is built to solve.

LyingDocs is not just a documentation checker. It is a system for surfacing repository misalignment:

  • docs that overclaim
  • code paths that are undocumented
  • specs that no longer match implementation
  • "configurable" behavior that is actually fixed
  • claims in papers or READMEs that cannot be supported by the code

The goal is simple:

Keep your repository trustworthy for humans and machines.


What LyingDocs does

LyingDocs deploys two autonomous agents against your repository:

  • Hermes reads your documentation, plans an audit strategy, and decides what needs to be verified
  • Argus investigates the actual codebase and reports what the code really does

Hermes then reconciles the two and writes a structured report of the mismatches it finds.

This lets you catch cases where your repository is no longer telling the truth about itself.


How it works

1. Hermes reads what the repo claims

Hermes traverses your documentation and extracts claims, assumptions, and implementation promises from sources such as:

  • docs/
  • README files
  • setup guides
  • usage examples
  • configuration references
  • papers and research writeups

It then plans an audit by turning those claims into targeted investigation tasks.

2. Argus checks what the code actually does

Argus executes each task against your real codebase.

You can choose the backend that best fits your setup:

  • codexOpenAI Codex CLI subprocess
  • claude_codeClaude Code CLI subprocess (claude -p)
  • local — built-in minimal agent loop using filesystem tools and any OpenAI-compatible API directly

3. LyingDocs reports the trust gaps

Hermes reconciles documented claims with observed implementation behavior and outputs a report of misalignments.

These findings can then be reviewed by maintainers, turned into issues, and eventually enforced in CI.


Positioning

LyingDocs is best thought of as:

  • a trust layer for your repo
  • a docs-to-code alignment guard
  • a pre-user warning system for misleading documentation
  • a future CI / GitHub Action quality gate for repository truthfulness

It is not meant to be a tool you manually open every day.

It is meant to become something your repository runs automatically:

  • on pull requests
  • before releases
  • during scheduled audits
  • before docs deployment
  • as part of your GitHub Actions workflow

Installation

pip install lyingdocs

Quick Start

export OPENAI_API_KEY="sk-..."

lyingdocs analyze --doc-path docs/ --code-path . -o output/audit

This performs a full audit of your repository and produces a report describing where documentation and implementation no longer align.


Documentation

Configuration Config file schema, environment variables, layer resolution
Argus Backends Setup for local, codex, and claude_code
CLI Reference All flags, commands, and output artifacts
GitHub Actions CI integration, authentication, triggers, and approval gates
GitHub Issues Using --gen-issue to draft and post issues

Example use cases

Use LyingDocs when you want to answer questions like:

  • Does the README still reflect the real behavior of the project?
  • Are our examples and quickstarts still valid?
  • Did code change without the docs changing with it?
  • Are we claiming configuration that does not actually exist?
  • Does our paper describe behavior the implementation does not support?
  • Can an AI agent trust this repository as a source of truth?

Misalignment categories

Category Description
LogicMismatch Code contradicts documentation
PhantomSpec Documentation describes non-existent features
ShadowLogic Important code behavior exists but is undocumented
HardcodedDrift Supposedly configurable values are actually hardcoded

These categories represent different ways repository trust breaks down.


GitHub Actions

LyingDocs runs natively in GitHub Actions as a trust gate for your CI pipeline.

pip install lyingdocs
lyingdocs init-ci --doc-path docs/ --backend claude_code --claude-oauth --trigger tag,manual

This generates a workflow that:

  • audits docs-vs-code alignment on every tag push (or on demand)
  • posts findings as a PR comment
  • optionally requires manual approval before merging

Supports all three backends (local, codex, claude_code) and both API key and OAuth token authentication for Claude Code.

See the full setup guide for trigger options, approval gates, and backend configuration.


Roadmap

  • Multi-harness support — Argus runs on Codex, Claude Code, or a built-in local agent
  • Issue generation--gen-issue drafts GitHub issues from findings
  • GitHub Action integrationlyingdocs init-ci generates a ready-to-use workflow with configurable triggers, backend selection, PR comments, and manual approval gates
  • One-session memory support — Argus backends retain state across tasks for deeper multi-step investigations
  • Deeper analysis — multi-hop reasoning across doc hierarchies and version-aware diffing to detect when code changed but docs did not
  • Paper mode — treat academic papers as documentation and detect paper-to-code misalignment
  • Auto-fix mode — Hermes proposes doc patches for human review and application

For researchers

A paper is also documentation.

It is a human-language description of code, behavior, claims, and expected results — often written under deadline, and often drifting away from the implementation over time.

If you want to know whether:

  • your repo matches your paper
  • your claims are supported by the code
  • another researcher can trust your implementation

then LyingDocs can help.

The problem is the same. Paper is documentation for code. LyingDocs is for papers too.


Why “trust layer”

Because the problem is bigger than stale docs.

A repository becomes untrustworthy whenever its outward description and inward behavior drift apart.

That harms:

  • users trying to adopt the project
  • contributors trying to extend it
  • maintainers trying to review changes
  • researchers trying to reproduce results
  • AI agents trying to understand the repo

LyingDocs exists to make that gap visible.

Not after users complain. Before.


License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lyingdocs-0.1.4.tar.gz (36.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lyingdocs-0.1.4-py3-none-any.whl (46.5 kB view details)

Uploaded Python 3

File details

Details for the file lyingdocs-0.1.4.tar.gz.

File metadata

  • Download URL: lyingdocs-0.1.4.tar.gz
  • Upload date:
  • Size: 36.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for lyingdocs-0.1.4.tar.gz
Algorithm Hash digest
SHA256 6f915b92de866a4ae7bedca6d9ceae59e70b8d4e6db3dca43686f4936283d3d2
MD5 2839a70e8e08465a7cd96ac7764b271a
BLAKE2b-256 2cf0cbce6dddbe821e18e73b6e4f3961124ac53557802b128a577fbfc8a4b69f

See more details on using hashes here.

Provenance

The following attestation bundles were made for lyingdocs-0.1.4.tar.gz:

Publisher: publish.yml on KMing-L/lying-docs

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file lyingdocs-0.1.4-py3-none-any.whl.

File metadata

  • Download URL: lyingdocs-0.1.4-py3-none-any.whl
  • Upload date:
  • Size: 46.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for lyingdocs-0.1.4-py3-none-any.whl
Algorithm Hash digest
SHA256 b48f01675544a1a1c42d9f8a532c0eac78c4cbd782b938c7fb7efaf782304a38
MD5 fdea80c46271d7bbe988a2d159d000dc
BLAKE2b-256 4e25158db7732a0a3d34904240cf92b53daaf2261c3017fafdeafde1633589b1

See more details on using hashes here.

Provenance

The following attestation bundles were made for lyingdocs-0.1.4-py3-none-any.whl:

Publisher: publish.yml on KMing-L/lying-docs

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page