Skip to main content

Cost-aware multi-LLM adversarial review engine with deterministic governance

Project description

dvad-color-white

Devil's Advocate

Cost-aware multi-LLM adversarial review engine with deterministic governance.

Do you have an implementation plan, codebase, or spec created by Claude, GPT, Gemini, Grok, etc and you want the flagship model from competing frontier providers to rip it apart, exposing the holes in logic and potential coding landmines, before a single line of code gets written?

Devil's Advocate pits multiple LLM reviewers against an LLM author in a structured 2-round adversarial protocol. A deterministic governance engine (straight python, no LLM calls, no probability) resolves every finding into a machine-readable outcome. The result is a vetted artifact where every finding has been accepted, defended, challenged, or escalated (to you for final decision) with full traceability from the first objection through final resolution.

DVAD is borne from my frustrations with code generated inside an AI echo chamber (single source provider). Often I'll post a script to an LLM that didn't write it to see if the other providers could spot issues or problems. They usually do. I decided to turn it into an easy to use UI.

Screenshot from 2026-03-03 06-57-08 Screenshot from 2026-03-03 06-57-36 Screenshot from 2026-03-03 06-59-36 Screenshot from 2026-03-03 07-44-41

Requirements

  • Python 3.12+
  • Bare Minimum: 1 API key from 1 provider, 3 models - an author and two reviewers. (The other roles: dedup, integration, normalization, revision, can use the same models the author or reviewers use).
  • Comfortably Confrontational: 3 API keys from 3 different providers - one for the author, one per reviewer. (Different providers mean different blind spots and that's where the friction comes from. Friction is good.)
  • Supported providers: Every frontier provider that uses OpenAI-compatible, Anthropic, or Minimax prompt formatting. (Google Gemini, ChatGPT, Claude, DeepSeek, xAI/Grok, Minimax, Kimi, etc.)

Quick Install & Update

curl -fsSL https://raw.githubusercontent.com/briankelley/devils-advocate/main/install.sh | bash

Installs dvad, initializes config, sets up the systemd service, and launches the web GUI on port 8411.

Manual Install

python3 -m venv ~/.local/share/devils-advocate/venv
~/.local/share/devils-advocate/venv/bin/pip install devils-advocate
ln -sf ~/.local/share/devils-advocate/venv/bin/dvad ~/.local/bin/dvad
dvad install

Getting Started

1. Configure your models

Run dvad config --init to generate ~/.config/devils-advocate/models.yaml, then edit it. Each model needs a provider, model ID, and an environment variable name for its API key. See examples/models.yaml.example for a fully annotated template.

2. Set your API keys

API keys are resolved from environment variables — never stored in the config file. Set them in your shell or in ~/.config/devils-advocate/.env (auto-loaded, won't override existing env vars).

3. Validate

dvad config --show

Web GUI

The primary way to use Devil's Advocate. Covers the full workflow - submitting reviews, monitoring progress in real time, resolving escalated findings, and generating revised artifacts.

dvad gui

Opens at http://127.0.0.1:8411. The GUI includes a dashboard for submitting and browsing reviews, real-time review progress via SSE with per-model cost tracking, governance override controls, revision generation, and visual model/role configuration with a raw YAML editor.

By default the GUI refuses to bind to non-localhost interfaces. --allow-nonlocal overrides this and requires a CSRF token header on all mutating requests.

How It Works

  1. Independent review - Multiple reviewer models analyze the input in parallel, producing findings with severity, category, location, and recommendation.
  2. Deduplication - A dedup model groups overlapping findings into consolidated review groups, preserving source attribution.
  3. Author response - The author model responds to each group: ACCEPTED, REJECTED, or PARTIAL with a rationale.
  4. Rebuttal - Reviewers issue rebuttals on contested groups only. Each votes CONCUR or CHALLENGE.
  5. Final position - For challenged groups, the author provides a final position.
  6. Governance - A deterministic engine (no LLM calls, pure rule-based logic) maps every group to an outcome: AUTO_ACCEPTED, AUTO_DISMISSED, or ESCALATED. No finding passes through without the author demonstrating engagement - implicit and rote acceptance both escalate to human review.

Escalated findings are resolved through the GUI's override controls or dvad override. After governance, dvad revise generates the final revised artifact.

Review Modes

Mode Protocol Input Output
plan Adversarial Plan file + optional reference files revised-plan.md
code Adversarial Exactly one code file, optional spec revised-diff.patch
spec Collaborative Spec file(s) revised-spec-suggestions.md
integration Adversarial Input files or .dvad/manifest.json, optional spec remediation-plan.md

spec is non-adversarial - no author, no rebuttals, no governance. Findings are grouped by theme and compiled into a suggestion report. All other modes use the full 2-round adversarial protocol.

CLI Quick Start

dvad review --mode plan --input plan.md --input ref.py --project myproject
dvad review --mode code --input src/app.py --spec spec.md --project myproject
dvad review --mode plan --input plan.md --project myproject --max-cost 0.50
dvad review --mode plan --input plan.md --project myproject --dry-run

Design Notes

  • No vendor SDKs. All provider calls use httpx directly - full control over request shape and retry behavior.
  • Deterministic governance. Zero LLM calls. Every outcome is reproducible from the same inputs.
  • Atomic operations. File writes use mkstemp + os.replace. Locking uses O_CREAT | O_EXCL.
  • XDG-compliant. Config and data paths follow the XDG Base Directory specification.

Full CLI reference, configuration schema, governance rules, and cost tracking details are available in the documentation.

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

devils_advocate-0.9.14.tar.gz (4.8 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

devils_advocate-0.9.14-py3-none-any.whl (2.9 MB view details)

Uploaded Python 3

File details

Details for the file devils_advocate-0.9.14.tar.gz.

File metadata

  • Download URL: devils_advocate-0.9.14.tar.gz
  • Upload date:
  • Size: 4.8 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.12

File hashes

Hashes for devils_advocate-0.9.14.tar.gz
Algorithm Hash digest
SHA256 6f1e630f9916f161add64c18c3ac1945b05644326722b5dceb37526e6b5b32f4
MD5 e396d63f479d22887848bc5ae9238c00
BLAKE2b-256 67eb41adbb08438481e8a7a519000fc44fe753782a06b377c35ce391ddd39594

See more details on using hashes here.

File details

Details for the file devils_advocate-0.9.14-py3-none-any.whl.

File metadata

File hashes

Hashes for devils_advocate-0.9.14-py3-none-any.whl
Algorithm Hash digest
SHA256 e5304989b16d8c59571104c835daa9f0c20a2140fe6cade048d93e1933e1c4bf
MD5 f907fa6e61ece457664e4a9829e63726
BLAKE2b-256 b12b417ffd8d6cf6b5956b2cae66a153a938d208bb193fcc9dc4c2545ab49993

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page