Skip to main content

Monitor GitHub fork constellations with AI-powered analysis

Project description

ForkHub

PyPI version

Monitor GitHub fork constellations with AI-powered analysis.

ForkHub watches the forks around your GitHub repositories, uses a Claude AI agent to classify what changed and why, and surfaces interesting divergences through digest notifications. Think of it as a satellite view of all the gardens growing from your code — whether the gardeners sent a letter or not.

Quickstart

Prerequisites

Install

# Core install (tracking, syncing, digests, clustering)
uv tool install forkhub

# With Claude-powered features (AI analysis + agentic backfill test-fixer)
uv tool install 'forkhub[claude]'

# Or with pip
pip install forkhub              # core
pip install 'forkhub[claude]'    # with Claude integration

The [claude] extra enables AI-powered features that use Anthropic's claude-agent-sdk: fork change classification during sync, and the agentic test-fixer for backfill --auto-fix-tests. ForkHub's core tracking, syncing, and digest features work without it — and you can plug in your own TestFixer implementation (OpenAI, local models, rule-based) via the Python API or drive backfill from external agents via the CLI primitives.

For development:

git clone https://github.com/joshuaoliphant/forkhub.git
cd forkhub
uv sync

Configure

The easiest way is to create a .env file in your project directory:

cp env.example .env
# Edit .env with your tokens

ForkHub automatically loads .env files at startup. You can also export environment variables directly:

export GITHUB_TOKEN="ghp_..."
export ANTHROPIC_API_KEY="sk-ant-..."

Or use a TOML config file:

mkdir -p ~/.config/forkhub
cp forkhub.toml.example ~/.config/forkhub/forkhub.toml

Authentication options for Anthropic:

  • ANTHROPIC_API_KEY — standard API key
  • CLAUDE_ACCESS_TOKEN — OAuth token from claude set-token (used by Claude Code)

Either works. If both are set, the API key takes precedence.

First run

# Discover and track your repos
uv run forkhub init --user your-github-username

# See what's tracked
uv run forkhub repos

# Sync fork data from GitHub (and classify changes with the AI analyzer)
uv run forkhub sync

# Skip analysis — discovery only, no Anthropic API cost
uv run forkhub sync --no-analyze

# View forks for a specific repo
uv run forkhub forks owner/repo

# Generate a digest of interesting changes
uv run forkhub digest

CLI Commands

Command Description
forkhub init --user <username> Discover and track your GitHub repos
forkhub track <owner> <repo> Track a repository you don't own
forkhub untrack <owner> <repo> Stop tracking a repository
forkhub exclude <owner> <repo> Exclude a repo from tracking
forkhub include <owner> <repo> Re-include an excluded repo
forkhub repos List tracked repositories
forkhub forks <owner> <repo> List forks of a tracked repo
forkhub inspect <owner> <repo> Detailed view of a single fork
forkhub clusters <owner> <repo> Show signal clusters (similar changes across forks)
forkhub sync Sync fork data from GitHub and run AI analysis on changed forks
forkhub sync --no-analyze Sync without running the AI analyzer (no Anthropic API cost)
forkhub digest Generate and deliver a change digest
forkhub backfill run Run the autonomous backfill loop
forkhub backfill list List previous backfill attempts
forkhub backfill candidates List signals eligible for backfill
forkhub backfill apply <signal-id> Apply a patch to a candidate branch, run tests
forkhub backfill status <attempt-id> Inspect a backfill attempt
forkhub backfill record <attempt-id> Record the outcome of an attempt
forkhub backfill cleanup <attempt-id> Delete candidate branch, return to original
forkhub backfill read-failures Run tests, return failing test file contents
forkhub backfill write-test <path> Safety-gated test file write (stdin or --content)
forkhub backfill run-tests Run the configured test command
forkhub config show Show current configuration

Breaking change in v0.3.0: forkhub backfill and forkhub backfill-list are now forkhub backfill run and forkhub backfill list. The backfill subcommands (candidates/apply/status/record/cleanup/read-failures/write-test/run-tests) let any external agent — Claude Code, Cursor, Aider, local models, shell scripts, humans — drive the backfill test-fix loop without requiring the [claude] extra.

Library Usage

ForkHub is a library first. The CLI is a thin consumer.

import asyncio
from forkhub import ForkHub

async def main():
    async with ForkHub() as hub:
        # Discover your repos
        repos = await hub.init("your-username")

        # Sync fork data
        result = await hub.sync()
        print(f"Synced {result.repos_synced} repos, {result.total_changed_forks} changed forks")

        # Get forks for a repo
        forks = await hub.get_forks("owner", "repo", active_only=True)

        # Generate and deliver a digest
        digest = await hub.generate_digest()
        await hub.deliver_digest(digest)

        # Backfill valuable fork changes into your local repo
        result = await hub.backfill("owner/repo", dry_run=True)
        print(f"Evaluated {result.total_evaluated}, accepted {result.accepted}")

asyncio.run(main())

Custom providers

All extension points use Python Protocol classes, so you can inject your own implementations:

from forkhub import ForkHub

hub = ForkHub(
    git_provider=my_custom_provider,        # implements GitProvider protocol
    notification_backends=[my_slack_backend], # implements NotificationBackend protocol
    embedding_provider=my_embeddings,        # implements EmbeddingProvider protocol
)

How It Works

Tracking modes

Mode Description
owned Your repos, auto-discovered via init. Forks are monitored.
watched Repos you don't own but want to observe via track.
upstream Repos you've forked. Tracks upstream changes you might want.

Signals

When ForkHub syncs, a Claude AI agent analyzes what changed in each fork and produces signals — classified changes with a significance score (1-10):

Category Description
feature New functionality added
fix Bug fix not yet in upstream
refactor Structural/architectural change
config Configuration or deployment change
dependency Dependency swap or version change
removal Feature or code removed
adaptation Platform or environment adaptation
release A new tagged release

Clusters

When multiple forks independently make similar changes, ForkHub detects these as clusters using vector similarity of signal embeddings. Clusters reveal community-wide trends — if three forks all swap the same dependency, that's a signal worth knowing about.

Data flow

forkhub sync   ->  Discover forks (GitHub API)
               ->  Compare HEAD SHAs (skip unchanged)
               ->  AI agent classifies changes -> store signals
               ->  Update clusters via embedding similarity

forkhub digest ->  Query recent signals
               ->  AI agent composes readable summary
               ->  Deliver via notification backends

forkhub backfill run -> Rank high-significance signals
                     -> Fetch diffs, apply patches to candidate branches
                     -> Run test suite to score results
                     -> Accept or reject based on test outcome

External-agent backfill flow

The forkhub backfill sub-app exposes composable primitives so any agent can drive the test-fix loop manually. Example shell-driven flow:

# 1. Find a candidate
SIG=$(forkhub backfill candidates --json | jq -r '.candidates[0].signal_id')

# Optional: preview without touching git or running tests
forkhub backfill apply "$SIG" --dry-run --json

# 2. Apply — preserves candidate branch on test failure
forkhub backfill apply "$SIG" --json
# Exit codes: 0=passed, 1=tests failed, 2=conflict/patch failed,
#             3=fetch error, 4=signal not found

# 3. Inspect failing tests (runs the test command fresh each call)
forkhub backfill read-failures --json
# Exit codes: 0=tests passed, 1=tests failed, 124=timeout/spawn failure

# 4. Agent (any tool) produces fixed test content and pipes it in
cat fixed_test.py | forkhub backfill write-test tests/test_foo.py

# 5. Re-run tests (returncode propagated; -1 timeout maps to 124)
forkhub backfill run-tests --json

# 6. Record outcome (score is validated against 0.0-1.0)
ATTEMPT=$(forkhub backfill list --json | jq -r '.[0].id')
forkhub backfill record "$ATTEMPT" --status=accepted --score=0.9

# 7. Or discard the attempt — exit 2 if any git op was swallowed
forkhub backfill cleanup "$ATTEMPT"

Every primitive outputs JSON when --json is passed. The Python service layer enforces safety invariants (test-files-only, path traversal defense) regardless of which agent is driving.

Configuration

ForkHub looks for forkhub.toml in ~/.config/forkhub/ or the current directory. Environment variables override TOML values.

Setting Env Var Default
GitHub token GITHUB_TOKEN
Anthropic API key ANTHROPIC_API_KEY
OAuth token CLAUDE_ACCESS_TOKEN
Analysis budget $0.50 per sync
Analysis model sonnet
Digest model haiku
Sync interval 6h
Digest frequency weekly
Min significance 5
DB path ~/.local/share/forkhub/forkhub.db

ForkHub loads .env files automatically. See env.example for all supported variables.

See forkhub.toml.example for all options.

Architecture

ForkHub is a library first — the CLI is a thin consumer. The core library (src/forkhub/) exposes the ForkHub class as its public API.

Extension points (Protocol-based, swappable at runtime):

  • GitProvider — fetches repo/fork data (default: GitHub via githubkit)
  • NotificationBackend — delivers digests (default: Rich console output)
  • EmbeddingProvider — text embeddings for clustering (default: local sentence-transformers)

AI analysis uses the Claude Agent SDK with a coordinator + subagent pattern:

  • Coordinator agent gets tools to explore forks (list, summarize, diff)
  • diff-analyst subagent deep-dives individual forks
  • digest-writer subagent composes human-readable summaries

Storage: SQLite + sqlite-vec for vector similarity search.

Development

# Install with dev dependencies
uv sync

# Run tests (155 tests)
uv run pytest

# Lint and format
uv run ruff check src/ tests/
uv run ruff format src/ tests/

# Type check
uv run ty check

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

forkhub-0.4.0.tar.gz (314.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

forkhub-0.4.0-py3-none-any.whl (86.3 kB view details)

Uploaded Python 3

File details

Details for the file forkhub-0.4.0.tar.gz.

File metadata

  • Download URL: forkhub-0.4.0.tar.gz
  • Upload date:
  • Size: 314.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for forkhub-0.4.0.tar.gz
Algorithm Hash digest
SHA256 ecdea412a56880bd1ca7f06ce4431df77b9143392a26ef1a9ae206e68672c0c6
MD5 f8cc18f7de3b69ab11b9fe350ed54bf3
BLAKE2b-256 8ba90343ff4d0e20b31d679624ea6124b60df4b9655729273e75a0175c13afa3

See more details on using hashes here.

Provenance

The following attestation bundles were made for forkhub-0.4.0.tar.gz:

Publisher: publish.yml on JoshuaOliphant/forkhub

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file forkhub-0.4.0-py3-none-any.whl.

File metadata

  • Download URL: forkhub-0.4.0-py3-none-any.whl
  • Upload date:
  • Size: 86.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for forkhub-0.4.0-py3-none-any.whl
Algorithm Hash digest
SHA256 237346f91093f054f4c72fe38ec317fed2723484cb1226625f45583edfc6070e
MD5 4eec43bd61df659319c3069c91e20f45
BLAKE2b-256 8367d6d41a24c4d1561a237e326a6f053283269f3e1030d1ea5398a985b10d36

See more details on using hashes here.

Provenance

The following attestation bundles were made for forkhub-0.4.0-py3-none-any.whl:

Publisher: publish.yml on JoshuaOliphant/forkhub

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page