Skip to main content

Keep spec, tests, and code in sync during AI-assisted development

Project description

Plumb

Plumb is proof of concept software demonstrating how we can manage the "Spec Driven Development Triange". It is very much alpha software, use cautiously.

A tool to keep things true.

plumb keeps your spec, tests, and code in sync during AI-assisted development.

When you work with Claude Code, decisions get made — a caching strategy is chosen, an API contract changes, a behavior is refined. These decisions live in conversation history and staged diffs, but they never make it back to the spec or tests. Over time, the spec drifts from reality, tests cover the wrong behavior, and the codebase becomes its own undocumented source of truth.

Plumb fixes this by intercepting git commit via a pre-commit hook. It analyzes your staged changes and Claude Code conversation, extracts the decisions that were made, and gates the commit on your review. Approved decisions are automatically synced back to the spec and tests. Rejected decisions trigger code modifications to undo them. The result: every committed state has a spec and test suite that could reconstruct the program.

Install

pip install plumb-dev

or

uv add plumb-dev

Quick Start

cd your-project
plumb init

This will:

  1. Ask for paths to your spec markdown and test directory
  2. Create a .plumb/ directory for state (commit this to version control)
  3. Install a git pre-commit hook
  4. Install a Claude Code skill file at .claude/skills/plumb/SKILL.md
  5. Add a Plumb block to CLAUDE.md
  6. Create a .plumbignore file for excluding irrelevant files from analysis
  7. Parse your spec into requirements

From here, just work normally. Plumb activates when you commit.

How It Works

Committing inside Claude Code

  1. You run git commit (or Claude Code does)
  2. The pre-commit hook fires, analyzes the staged diff and conversation log
  3. It writes pending decisions and exits non-zero, aborting the commit
  4. Claude Code's skill reads the output and presents each decision:

    Question: Should we cache API responses in memory or on disk? Decision made: In-memory cache using a dict. Approve, reject, or edit?

  5. You respond in chat. The skill calls plumb approve, plumb reject, or plumb edit
  6. Rejected decisions trigger plumb modify, which rewrites the staged code
  7. Once all decisions are resolved, git commit runs again and lands

Committing from the terminal

Same flow, but you drive it with plumb review instead of the skill.

Commands

Command What it does
plumb init Initialize Plumb in a git repo
plumb status Show spec files, requirements, pending decisions, coverage
plumb diff Preview what decisions Plumb would extract from staged changes
plumb review Interactively review pending decisions in the terminal
plumb approve <id> Approve a decision and sync it to spec/tests
plumb approve --all Approve all pending decisions at once
plumb reject <id> --reason "..." Reject a decision
plumb edit <id> "new text" Amend a decision's text and approve it
plumb modify <id> Auto-modify staged code to satisfy a rejected decision
plumb sync Sync all unsynced approved/edited decisions to spec and tests
plumb parse-spec Re-parse spec files into requirements
plumb coverage Report code coverage, spec-to-test, and spec-to-code coverage

Coverage

Plumb tracks three dimensions of coverage:

  • Code coverage — pytest line coverage via pytest --cov
  • Spec-to-test — which requirements have corresponding tests
  • Spec-to-code — which requirements have corresponding implementations

Run plumb coverage to see all three.

.plumbignore

Plumb uses a .plumbignore file with gitignore-style patterns to exclude files from analysis. This keeps noise out of the decision extraction process — lock files, generated code, and other irrelevant diffs won't produce spurious decisions.

Project State

All Plumb state lives in .plumb/ at the repo root:

.plumb/
├── config.json          # Spec paths, test paths, settings
├── coverage.json        # Cached coverage data
├── decisions.jsonl      # Append-only log of all decisions
└── requirements.json    # Parsed requirements from the spec

Commit this directory to version control.

Requirements

  • Python 3.10+
  • A git repository
  • An ANTHROPIC_API_KEY environment variable or .env file (for LLM-powered analysis)

Note: plumb init, plumb status, and plumb review work without an API key. The key is only needed when the hook analyzes diffs or when syncing decisions to spec/tests.

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

plumb_dev-0.2.1.tar.gz (232.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

plumb_dev-0.2.1-py3-none-any.whl (49.0 kB view details)

Uploaded Python 3

File details

Details for the file plumb_dev-0.2.1.tar.gz.

File metadata

  • Download URL: plumb_dev-0.2.1.tar.gz
  • Upload date:
  • Size: 232.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.4

File hashes

Hashes for plumb_dev-0.2.1.tar.gz
Algorithm Hash digest
SHA256 2099ba87ef310fe85420db5c3d0197b53c00a1f689ddb87a7bce37314410b258
MD5 6dad236e783e307771b15d747a6a902f
BLAKE2b-256 2e5ec88d97a9115b240d53bdc4f6506072315e5544c8abfb8184c67917d3c263

See more details on using hashes here.

File details

Details for the file plumb_dev-0.2.1-py3-none-any.whl.

File metadata

  • Download URL: plumb_dev-0.2.1-py3-none-any.whl
  • Upload date:
  • Size: 49.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.4

File hashes

Hashes for plumb_dev-0.2.1-py3-none-any.whl
Algorithm Hash digest
SHA256 6c2ed4bea487417950c76068c261e132f49460b2d3ea552a68303a73f7470aac
MD5 3cc04853868b6839d33940b6d2d493bf
BLAKE2b-256 2c73954bb3fac6730f13f2e5302b1f7e13ce29a5b9cb264d8b098ab03ba0be01

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page