Skip to main content

Keep spec, tests, and code in sync during AI-assisted development

Project description

Plumb

Plumb is proof of concept software demonstrating how we can manage the "Spec Driven Development Triange". It is very much alpha software, use cautiously.

A tool to keep things true.

plumb keeps your spec, tests, and code in sync during AI-assisted development.

When you work with Claude Code, decisions get made — a caching strategy is chosen, an API contract changes, a behavior is refined. These decisions live in conversation history and staged diffs, but they never make it back to the spec or tests. Over time, the spec drifts from reality, tests cover the wrong behavior, and the codebase becomes its own undocumented source of truth.

Plumb fixes this by intercepting git commit via a pre-commit hook. It analyzes your staged changes and Claude Code conversation, extracts the decisions that were made, and gates the commit on your review. Approved decisions are automatically synced back to the spec and tests. Rejected decisions trigger code modifications to undo them. The result: every committed state has a spec and test suite that could reconstruct the program.

Install

pip install plumb-dev

or

uv add plumb-dev

Quick Start

cd your-project
plumb init

This will:

  1. Ask for paths to your spec markdown and test directory
  2. Create a .plumb/ directory for state (commit this to version control)
  3. Install a git pre-commit hook
  4. Install a Claude Code skill file at .claude/skills/plumb/SKILL.md
  5. Add a Plumb block to CLAUDE.md
  6. Create a .plumbignore file for excluding irrelevant files from analysis
  7. Parse your spec into requirements

From here, just work normally. Plumb activates when you commit.

How It Works

Committing inside Claude Code

  1. You run git commit (or Claude Code does)
  2. The pre-commit hook fires, analyzes the staged diff and conversation log
  3. It writes pending decisions and exits non-zero, aborting the commit
  4. Claude Code's skill reads the output and presents each decision:

    Question: Should we cache API responses in memory or on disk? Decision made: In-memory cache using a dict. Approve, reject, or edit?

  5. You respond in chat. The skill calls plumb approve, plumb reject, or plumb edit
  6. Rejected decisions trigger plumb modify, which rewrites the staged code
  7. Once all decisions are resolved, git commit runs again and lands

Committing from the terminal

Same flow, but you drive it with plumb review instead of the skill.

Commands

Command What it does
plumb init Initialize Plumb in a git repo
plumb status Show spec files, requirements, pending decisions, coverage
plumb diff Preview what decisions Plumb would extract from staged changes
plumb review Interactively review pending decisions in the terminal
plumb approve <id> Approve a decision and sync it to spec/tests
plumb approve --all Approve all pending decisions at once
plumb reject <id> --reason "..." Reject a decision
plumb edit <id> "new text" Amend a decision's text and approve it
plumb modify <id> Auto-modify staged code to satisfy a rejected decision
plumb sync Sync all unsynced approved/edited decisions to spec and tests
plumb parse-spec Re-parse spec files into requirements
plumb coverage Report code coverage, spec-to-test, and spec-to-code coverage

Coverage

Plumb tracks three dimensions of coverage:

  • Code coverage — pytest line coverage via pytest --cov
  • Spec-to-test — which requirements have corresponding tests
  • Spec-to-code — which requirements have corresponding implementations

Run plumb coverage to see all three.

.plumbignore

Plumb uses a .plumbignore file with gitignore-style patterns to exclude files from analysis. This keeps noise out of the decision extraction process — lock files, generated code, and other irrelevant diffs won't produce spurious decisions.

Project State

All Plumb state lives in .plumb/ at the repo root:

.plumb/
├── config.json          # Spec paths, test paths, settings
├── coverage.json        # Cached coverage data
├── decisions.jsonl      # Append-only log of all decisions
└── requirements.json    # Parsed requirements from the spec

Commit this directory to version control.

Requirements

  • Python 3.10+
  • A git repository
  • An ANTHROPIC_API_KEY environment variable or .env file (for LLM-powered analysis)

Note: plumb init, plumb status, and plumb review work without an API key. The key is only needed when the hook analyzes diffs or when syncing decisions to spec/tests.

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

plumb_dev-0.2.3.tar.gz (249.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

plumb_dev-0.2.3-py3-none-any.whl (49.3 kB view details)

Uploaded Python 3

File details

Details for the file plumb_dev-0.2.3.tar.gz.

File metadata

  • Download URL: plumb_dev-0.2.3.tar.gz
  • Upload date:
  • Size: 249.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.4

File hashes

Hashes for plumb_dev-0.2.3.tar.gz
Algorithm Hash digest
SHA256 4d5c84ee3788a121fbadb2c031524ed3100ffd3bf6c2a27105705cf921682e4d
MD5 ab87624a7939aa8222275f222e0b6a30
BLAKE2b-256 b570740b118913119c75749481ac74c1c65c526b30b7cad441f86e4b5715885e

See more details on using hashes here.

File details

Details for the file plumb_dev-0.2.3-py3-none-any.whl.

File metadata

  • Download URL: plumb_dev-0.2.3-py3-none-any.whl
  • Upload date:
  • Size: 49.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.4

File hashes

Hashes for plumb_dev-0.2.3-py3-none-any.whl
Algorithm Hash digest
SHA256 cb49c8a57da9679a2def8a1bea800ab091fbb3c36479127cce7f05bc5d7c656d
MD5 3b23d31efbb6b12b3d2e7e143bc2c78b
BLAKE2b-256 6cc9e6e2fcbc08a32e0f70fd3cc433303e9a0041253ef301ae3758de9f3ba089

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page