Skip to main content

Cost-aware multi-LLM adversarial review engine with deterministic governance

Project description

dvad-color-white

Devil's Advocate

Cost-aware multi-LLM adversarial review engine with deterministic governance.

Do you have an implementation plan, codebase, or spec created by Claude, GPT, Gemini, Grok, etc and you want the flagship model from competing frontier providers to rip it apart, exposing the holes in logic and potential coding landmines, before a single line of code gets written?

Devil's Advocate pits multiple LLM reviewers against an LLM author in a structured 2-round adversarial protocol. A deterministic governance engine (straight python, no LLM calls, no probability) resolves every finding into a machine-readable outcome. The result is a vetted artifact where every finding has been accepted, defended, challenged, or escalated (to you for final decision) with full traceability from the first objection through final resolution.

DVAD is borne from my frustrations with code generated inside an AI echo chamber (single source provider). Often I'll post a script to an LLM that didn't write it to see if the other providers could spot issues or problems. They usually do. I decided to turn it into an easy to use UI.

Screenshot from 2026-03-03 06-57-08 Screenshot from 2026-03-03 06-57-36 Screenshot from 2026-03-03 06-59-36 Screenshot from 2026-03-03 07-44-41

Requirements

  • Python 3.12+
  • Bare Minimum: 1 API key from 1 provider, 3 models - an author and two reviewers. (The other roles: dedup, integration, normalization, revision, can use the same models the author or reviewers use).
  • Comfortably Confrontational: 3 API keys from 3 different providers - one for the author, one per reviewer. (Different providers mean different blind spots and that's where the friction comes from. Friction is good.)
  • Supported providers: Every frontier provider that uses OpenAI-compatible, Anthropic, or Minimax prompt formatting. (Google Gemini, ChatGPT, Claude, DeepSeek, xAI/Grok, Minimax, Kimi, etc.)

Quick Install (or update to latest)

curl -fsSL https://raw.githubusercontent.com/briankelley/devils-advocate/main/install.sh | bash

Installs dvad, initializes config, sets up the systemd service, and launches the web GUI on port 8411.

Manual Install

python3 -m venv ~/.local/share/devils-advocate/venv
~/.local/share/devils-advocate/venv/bin/pip install devils-advocate
ln -sf ~/.local/share/devils-advocate/venv/bin/dvad ~/.local/bin/dvad
dvad install
1. Configure your models

Run dvad config --init to generate ~/.config/devils-advocate/models.yaml, then edit it. Each model needs a provider, model ID, and an environment variable name for its API key. See examples/models.yaml.example for a fully annotated template.

2. Set your API keys

API keys are resolved from environment variables — never stored in the config file. Set them in your shell or in ~/.config/devils-advocate/.env (auto-loaded, won't override existing env vars).

3. Validate
dvad config --show
4. Web GUI

The primary way to use Devil's Advocate. Covers the full workflow - submitting reviews, monitoring progress in real time, resolving escalated findings, and generating revised artifacts.

dvad gui

Opens at http://127.0.0.1:8411. The GUI includes a dashboard for submitting and browsing reviews, real-time review progress via SSE with per-model cost tracking, governance override controls, revision generation, and visual model/role configuration with a raw YAML editor.

By default the GUI refuses to bind to non-localhost interfaces. --allow-nonlocal overrides this and requires a CSRF token header on all mutating requests.

How It Works

  1. Independent review - Multiple reviewer models analyze the input in parallel, producing findings with severity, category, location, and recommendation.
  2. Deduplication - A dedup model groups overlapping findings into consolidated review groups, preserving source attribution.
  3. Author response - The author model responds to each group: ACCEPTED, REJECTED, or PARTIAL with a rationale.
  4. Rebuttal - Reviewers issue rebuttals on contested groups only. Each votes CONCUR or CHALLENGE.
  5. Final position - For challenged groups, the author provides a final position.
  6. Governance - A deterministic engine (no LLM calls, pure rule-based logic) maps every group to an outcome: AUTO_ACCEPTED, AUTO_DISMISSED, or ESCALATED. No finding passes through without the author demonstrating engagement - implicit and rote acceptance both escalate to human review.

Escalated findings are resolved through the GUI's override controls or dvad override. After governance, dvad revise generates the final revised artifact.

Review Modes

Mode Protocol Input Output
plan Adversarial Plan file + optional reference files revised-plan.md
code Adversarial Exactly one code file, optional spec revised-diff.patch
spec Collaborative Spec file(s) revised-spec-suggestions.md
integration Adversarial Input files or .dvad/manifest.json, optional spec remediation-plan.md

spec is non-adversarial - no author, no rebuttals, no governance. Findings are grouped by theme and compiled into a suggestion report. All other modes use the full 2-round adversarial protocol.

CLI Quick Start

dvad config --show
dvad history --project <project name>
dvad review --mode plan --input plan.md --input ref.py --project myproject
dvad review --mode code --input src/app.py --spec spec.md --project myproject
dvad review --mode plan --input plan.md --project myproject --max-cost 0.50
dvad review --mode plan --input plan.md --project myproject --dry-run

Design Notes

  • No vendor SDKs. All provider calls use httpx directly - full control over request shape and retry behavior.
  • Deterministic governance. Zero LLM calls. Every outcome is reproducible from the same inputs.
  • Atomic operations. File writes use mkstemp + os.replace. Locking uses O_CREAT | O_EXCL.
  • XDG-compliant. Config and data paths follow the XDG Base Directory specification.

Full CLI reference, configuration schema, governance rules, and cost tracking details are available in the documentation.

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

devils_advocate-0.9.27.tar.gz (4.9 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

devils_advocate-0.9.27-py3-none-any.whl (2.9 MB view details)

Uploaded Python 3

File details

Details for the file devils_advocate-0.9.27.tar.gz.

File metadata

  • Download URL: devils_advocate-0.9.27.tar.gz
  • Upload date:
  • Size: 4.9 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.12

File hashes

Hashes for devils_advocate-0.9.27.tar.gz
Algorithm Hash digest
SHA256 c50711d219320dac1678e77390b9812f2346ded9b1bccd84fae0130d4ed63b7f
MD5 6fcfb36b3d7d3159060bb405a2112461
BLAKE2b-256 00fbf6525cfd65f8cb4514f9441a772f9a3b8f29727a6f550ac584266c89b566

See more details on using hashes here.

File details

Details for the file devils_advocate-0.9.27-py3-none-any.whl.

File metadata

File hashes

Hashes for devils_advocate-0.9.27-py3-none-any.whl
Algorithm Hash digest
SHA256 0f62139687981833b6ba6d84dcca0b1e218405cfb1bc9e3c44f79b59cd784a57
MD5 2b70347701a534b613c596a0b9fce497
BLAKE2b-256 a46ae64080601f1ca2c3cfc7dfa0a536f871983d737a48d2d373d00bf12c7f51

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page