Skip to main content

Diff visualizer for Databricks bundle plans — resources and job DAGs

Project description

Dagshund

Dagshund renders databricks bundle plan -o json output as a colored terminal diff, a markdown summary, or a self-contained HTML report with an interactive resource graph and per-job task DAGs. It is read-only: it does not connect to Databricks and does not deploy. The CLI is pure Python with no runtime dependencies.

Requires the direct deployment engine for Declarative Automation Bundles (formerly Databricks Asset Bundles).

Dagshund resource graph

Try it live: here. Renders the HTML report for each golden fixture.

Where it fits

Dagshund sits between databricks bundle plan and databricks bundle deploy. Two places you would run it:

Locally, while working on a bundle. The terminal output gives you a quick glance at what a deploy would do, filterable by resource type, status, or changed field. The HTML report is for detailed inspection: browse the full resource graph, open a node to see its field-level diff, and switch to the Jobs tab to walk task dependencies.

In CI, for example on every PR against the target workspace. Pipe a plan into dagshund to post a markdown summary on the PR, upload the HTML report as a pipeline artifact for deeper inspection, and use -e exit codes to gate the pipeline: 0 for no changes, 2 for safe changes, 3 for dangerous actions or manual edits, 1 for errors.

Install

Requires Python 3.12+. Tested against Databricks CLI 0.298 (the plan JSON shape is what dagshund reads).

# Install (recommended for regular use)
uv tool install dagshund

# Or run one-off without installing
uvx dagshund plan.json

Local usage

By default, dagshund reads a plan file (or stdin) and prints a colored diff to the terminal:

dagshund plan.json
databricks bundle plan -o json | dagshund    # pipe straight from the Databricks CLI

Terminal output

Dagshund works anywhere, it just needs a plan JSON file. You don't need to run it from inside your bundle directory.

Filter to specific change types:

dagshund plan.json -c              # changed resources only (hides unchanged)
dagshund plan.json -a              # added only
dagshund plan.json -m              # modified only
dagshund plan.json -a -r           # added and removed

The filter flags (-a, -m, -r) compose freely. -c is shorthand for -a -m -r.

Use -f to filter by resource type, name, diff status, or changed field:

dagshund plan.json -f 'type:jobs'                # only jobs
dagshund plan.json -f 'status:added'             # only new resources
dagshund plan.json -f '"etl_pipeline"'           # exact name match
dagshund plan.json -f 'type:jobs pipeline'       # jobs matching "pipeline"
dagshund plan.json -f 'field:email_notifications' # resources whose diff touches that field
dagshund plan.json -c -f 'type:volumes'          # changed volumes only

All tokens in a filter expression AND together. -f composes with -c/-a/-m/-r: both must match. field: is a substring match against changed field names.

Export the HTML report with -o for detailed inspection in a browser:

dagshund plan.json -o report.html           # write HTML, keep terminal output
dagshund plan.json -o report.html -b        # also open in the default browser
dagshund plan.json -q -o report.html        # HTML only, suppress terminal output

Or emit a markdown summary to stdout with --format md:

dagshund plan.json --format md

Interactive Visualization

The HTML report shows your resources as an interactive graph with diff highlighting. Resources are organized into visual groups:

  • Workspace, jobs, alerts, experiments, pipelines, and other bundle resources
  • Unity Catalog, catalogs, schemas, volumes, and registered models in their hierarchy
  • Lakebase, database instances and synced tables. When your plan includes any, this group appears and the non-hierarchical Workspace resources move into a sibling Other Resources group.

Click any node to open a detail panel with per-field structural diffs, old values in red, new values in green, unchanged fields for context.

Structural diff detail panel

Jobs with task dependencies get their own DAG view. Switch between the Resources and Jobs tabs to navigate between them.

Job task DAG

Phantom Nodes

When a resource in your bundle references something that isn't in the bundle itself, dagshund infers the missing piece and adds it to the graph as a phantom node (shown with a dashed border). Two kinds: hierarchy phantoms (for example a parent catalog above a schema) always display to preserve the hierarchy's structure. Inferred leaf phantoms (for example a warehouse referenced by an alert) are off by default; toggle them with the Inferred leaf nodes button in the toolbar.

Phantom nodes

Lateral Dependencies

Many resources reference each other across hierarchies, an alert might target a SQL warehouse, or a serving endpoint might bind to a registered model. These relationships are off by default to keep the graph clean; toggle them with the Lateral dependencies button in the toolbar to see how your resources connect across group boundaries.

Lateral dependencies

Search

The search bar dims non-matching nodes so matches stand out.

  • Type a name to filter: warehouse, analytics
  • Wrap in quotes for exact match: "analytics" finds only that node, not analytics_pipeline
  • Prefix with type: to filter by badge: type:wheel highlights all wheel tasks
  • Prefix with status: to filter by diff status: status:added, status:modified, status:removed
  • Press Escape to clear

The diff filter buttons (Added, Modified, Removed) compose with search, when both are active, only nodes matching both criteria stay highlighted. When exactly one node matches, the viewport auto-centers on it.

Manual Edit Detection

When someone edits a job directly in the Databricks UI (break glass), the bundle doesn't know about it. On the next deploy, those manual changes will be silently overwritten. Dagshund detects this by comparing the plan's expected state against the actual server state, and warns you when they diverge.

In the terminal:

Manual edit detection in terminal

In the HTML report:

Manual edit detection in browser

The warning surfaces in both views whenever the plan's old and remote states differ.

CI usage

Run dagshund in a CI pipeline to surface plan changes on every PR. One command produces all three CI outputs: markdown for the PR comment, HTML for a pipeline artifact, and an exit code for gating.

databricks bundle plan -t "$TARGET" -o json | \
  dagshund --format md -o report.html -e > summary.md
status=$?

Then in your pipeline step:

  • Post summary.md as a PR comment with gh pr comment, glab mr note, or your provider's API.
  • Upload report.html as a pipeline artifact for deeper inspection.
  • Branch on $status to gate the deploy step.

PR comment rendering

Exit codes from -e:

Code Meaning
0 Plan parsed, no changes detected
1 Error (bad input, missing file, etc.)
2 Plan parsed, changes detected
3 Plan parsed, changes detected AND dangerous actions or manual edits present

Dangerous actions are deletes or recreates of stateful resources (catalogs, schemas, volumes, registered models). Manual edits are detected whenever the plan's old and remote states differ.

Without -e, dagshund always exits 0 on success.

Agent Skill

Dagshund ships as a Claude Code plugin with an agent skill that teaches AI coding agents how to use it. Once installed, your agent can answer questions like "what's changing in my deploy?" by running dagshund automatically.

Install as a Claude Code plugin:

/plugin marketplace add https://github.com/chinchyisbored/dagshund.git
/plugin install dagshund@chinchy-dagshund

Or use the CLI to install the skill into any agent harness:

dagshund --install-skill .claude/skills     # Claude Code
dagshund --install-skill .cursor/skills     # Cursor
dagshund --install-skill .agents/skills     # Codex / Gemini CLI

Re-running --install-skill overwrites any existing SKILL.md at the target path without prompting.

Contributing

Dagshund is a solo project and I'm not accepting pull requests at this time. If you run into a bug or have a feature request, please open an issue, I'm happy to hear what you need.

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

dagshund-0.10.0.tar.gz (2.0 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

dagshund-0.10.0-py3-none-any.whl (789.2 kB view details)

Uploaded Python 3

File details

Details for the file dagshund-0.10.0.tar.gz.

File metadata

  • Download URL: dagshund-0.10.0.tar.gz
  • Upload date:
  • Size: 2.0 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: uv/0.11.7 {"installer":{"name":"uv","version":"0.11.7","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Debian GNU/Linux","version":"13","id":"trixie","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for dagshund-0.10.0.tar.gz
Algorithm Hash digest
SHA256 387c24715f1b0eadb3dfbb8473eb07dcbf5de39ce1743c2b27a644f7a70f46df
MD5 8b41223a89864c395c653939163d4a27
BLAKE2b-256 2de0c600021c8113f88ad7fe77e49dc0d0662761350fbc7ad36f9a6c2f54650a

See more details on using hashes here.

File details

Details for the file dagshund-0.10.0-py3-none-any.whl.

File metadata

  • Download URL: dagshund-0.10.0-py3-none-any.whl
  • Upload date:
  • Size: 789.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: uv/0.11.7 {"installer":{"name":"uv","version":"0.11.7","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Debian GNU/Linux","version":"13","id":"trixie","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for dagshund-0.10.0-py3-none-any.whl
Algorithm Hash digest
SHA256 4407d93a92c574a4d8c0c2fe7666757f95b1110d23cd31a7c9d93384add87d65
MD5 8273e96838d8b9ba0ce00a0bfc2ac660
BLAKE2b-256 b557f88daae6def28739d2cf1bf6c4de9334950484ace5ca7102ecee9f2d3192

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page