Skip to main content

Diff visualizer for Databricks bundle plans — resources and job DAGs

Project description

Dagshund

Dagshund renders databricks bundle plan -o json output as a colored terminal diff, a markdown summary, or a self-contained HTML report with an interactive resource graph and per-job task DAGs. It is read-only: it does not connect to Databricks and does not deploy. The CLI is pure Python with no runtime dependencies.

Requires the direct deployment engine for Declarative Automation Bundles (formerly Databricks Asset Bundles).

Dagshund resource graph

Try it live: here. Renders the HTML report for each golden fixture.

Where it fits

Dagshund sits between databricks bundle plan and databricks bundle deploy. Two places you would run it:

Locally, while working on a bundle. The terminal output gives you a quick glance at what a deploy would do, filterable by resource type, status, or changed field. The HTML report is for detailed inspection: browse the full resource graph, open a node to see its field-level diff, and switch to the Jobs tab to walk task dependencies.

In CI, for example on every PR against the target workspace. Pipe a plan into dagshund to post a markdown summary on the PR, upload the HTML report as a pipeline artifact for deeper inspection, and use -e exit codes to gate the pipeline: 0 for no changes, 2 for safe changes, 3 for dangerous actions or manual edits, 1 for errors.

Install

Requires Python 3.12+. Tested against Databricks CLI 0.296 (the plan JSON shape is what dagshund reads).

# Install (recommended for regular use)
uv tool install dagshund

# Or run one-off without installing
uvx dagshund plan.json

Local usage

By default, dagshund reads a plan file (or stdin) and prints a colored diff to the terminal:

dagshund plan.json
databricks bundle plan -o json | dagshund    # pipe straight from the Databricks CLI

Terminal output

Dagshund works anywhere, it just needs a plan JSON file. You don't need to run it from inside your bundle directory.

Filter to specific change types:

dagshund plan.json -c              # changed resources only (hides unchanged)
dagshund plan.json -a              # added only
dagshund plan.json -m              # modified only
dagshund plan.json -a -r           # added and removed

The filter flags (-a, -m, -r) compose freely. -c is shorthand for -a -m -r.

Use -f to filter by resource type, name, diff status, or changed field:

dagshund plan.json -f 'type:jobs'                # only jobs
dagshund plan.json -f 'status:added'             # only new resources
dagshund plan.json -f '"etl_pipeline"'           # exact name match
dagshund plan.json -f 'type:jobs pipeline'       # jobs matching "pipeline"
dagshund plan.json -f 'field:email_notifications' # resources whose diff touches that field
dagshund plan.json -c -f 'type:volumes'          # changed volumes only

All tokens in a filter expression AND together. -f composes with -c/-a/-m/-r: both must match. field: is a substring match against changed field names.

Export the HTML report with -o for detailed inspection in a browser:

dagshund plan.json -o report.html           # write HTML, keep terminal output
dagshund plan.json -o report.html -b        # also open in the default browser
dagshund plan.json -q -o report.html        # HTML only, suppress terminal output

Or emit a markdown summary to stdout with --format md:

dagshund plan.json --format md

Interactive Visualization

The HTML report shows your resources as an interactive graph with diff highlighting. Resources are organized into visual groups:

  • Workspace, jobs, alerts, experiments, pipelines, and other bundle resources
  • Unity Catalog, catalogs, schemas, volumes, and registered models in their hierarchy
  • Lakebase, database instances and synced tables. When your plan includes any, this group appears and the non-hierarchical Workspace resources move into a sibling Other Resources group.

Click any node to open a detail panel with per-field structural diffs, old values in red, new values in green, unchanged fields for context.

Structural diff detail panel

Jobs with task dependencies get their own DAG view. Switch between the Resources and Jobs tabs to navigate between them.

Job task DAG

Phantom Nodes

When a resource in your bundle references something that isn't in the bundle itself, dagshund infers the missing piece and adds it to the graph as a phantom node (shown with a dashed border). Two kinds: hierarchy phantoms (for example a parent catalog above a schema) always display to preserve the hierarchy's structure. Inferred leaf phantoms (for example a warehouse referenced by an alert) are off by default; toggle them with the Inferred leaf nodes button in the toolbar.

Phantom nodes

Lateral Dependencies

Many resources reference each other across hierarchies, an alert might target a SQL warehouse, or a serving endpoint might bind to a registered model. These relationships are off by default to keep the graph clean; toggle them with the Lateral dependencies button in the toolbar to see how your resources connect across group boundaries.

Lateral dependencies

Search

The search bar dims non-matching nodes so matches stand out.

  • Type a name to filter: warehouse, analytics
  • Wrap in quotes for exact match: "analytics" finds only that node, not analytics_pipeline
  • Prefix with type: to filter by badge: type:wheel highlights all wheel tasks
  • Prefix with status: to filter by diff status: status:added, status:modified, status:removed
  • Press Escape to clear

The diff filter buttons (Added, Modified, Removed) compose with search, when both are active, only nodes matching both criteria stay highlighted. When exactly one node matches, the viewport auto-centers on it.

Manual Edit Detection

When someone edits a job directly in the Databricks UI (break glass), the bundle doesn't know about it. On the next deploy, those manual changes will be silently overwritten. Dagshund detects this by comparing the plan's expected state against the actual server state, and warns you when they diverge.

In the terminal:

Manual edit detection in terminal

In the HTML report:

Manual edit detection in browser

The warning surfaces in both views whenever the plan's old and remote states differ.

CI usage

Run dagshund in a CI pipeline to surface plan changes on every PR. One command produces all three CI outputs: markdown for the PR comment, HTML for a pipeline artifact, and an exit code for gating.

databricks bundle plan -t "$TARGET" -o json | \
  dagshund --format md -o report.html -e > summary.md
status=$?

Then in your pipeline step:

  • Post summary.md as a PR comment with gh pr comment, glab mr note, or your provider's API.
  • Upload report.html as a pipeline artifact for deeper inspection.
  • Branch on $status to gate the deploy step.

PR comment rendering

Exit codes from -e:

Code Meaning
0 Plan parsed, no changes detected
1 Error (bad input, missing file, etc.)
2 Plan parsed, changes detected
3 Plan parsed, changes detected AND dangerous actions or manual edits present

Dangerous actions are deletes or recreates of stateful resources (catalogs, schemas, volumes, registered models). Manual edits are detected whenever the plan's old and remote states differ.

Without -e, dagshund always exits 0 on success.

Agent Skill

Dagshund ships as a Claude Code plugin with an agent skill that teaches AI coding agents how to use it. Once installed, your agent can answer questions like "what's changing in my deploy?" by running dagshund automatically.

Install as a Claude Code plugin:

/plugin marketplace add https://github.com/chinchyisbored/dagshund.git
/plugin install dagshund@chinchy-dagshund

Or use the CLI to install the skill into any agent harness:

dagshund --install-skill .claude/skills     # Claude Code
dagshund --install-skill .cursor/skills     # Cursor
dagshund --install-skill .agents/skills     # Codex / Gemini CLI

Re-running --install-skill overwrites any existing SKILL.md at the target path without prompting.

Contributing

Dagshund is a solo project and I'm not accepting pull requests at this time. If you run into a bug or have a feature request, please open an issue, I'm happy to hear what you need.

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

dagshund-0.8.0.tar.gz (1.9 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

dagshund-0.8.0-py3-none-any.whl (782.8 kB view details)

Uploaded Python 3

File details

Details for the file dagshund-0.8.0.tar.gz.

File metadata

  • Download URL: dagshund-0.8.0.tar.gz
  • Upload date:
  • Size: 1.9 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: uv/0.11.6 {"installer":{"name":"uv","version":"0.11.6","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Debian GNU/Linux","version":"13","id":"trixie","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for dagshund-0.8.0.tar.gz
Algorithm Hash digest
SHA256 3a8a330b10e17681210e3101ae7de2afc846c17803f638f9882d6431587078b4
MD5 40f35cab71929f4b97c6e2aa522cc36d
BLAKE2b-256 0d633412d1cc234f6e895cc630ac2245f9a7b3e78307a9c6e124799ff2ac22f2

See more details on using hashes here.

File details

Details for the file dagshund-0.8.0-py3-none-any.whl.

File metadata

  • Download URL: dagshund-0.8.0-py3-none-any.whl
  • Upload date:
  • Size: 782.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: uv/0.11.6 {"installer":{"name":"uv","version":"0.11.6","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Debian GNU/Linux","version":"13","id":"trixie","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for dagshund-0.8.0-py3-none-any.whl
Algorithm Hash digest
SHA256 141c5dbd6008e22729c603d9dd2b31af1ecb77b9a19e5f0e94ec81246bd9ef9a
MD5 864ee0bdd7d3ab8ab16d3cf84536a59d
BLAKE2b-256 16118c89e8862255c9625d8332e0415a11c6d23fbb4fbb7b02abe8ad85b4c78d

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page