Unified CLI for the Agent Quality Toolkit (agentmd, coderace, agentlint, agentreflect)
Project description
agentkit-cli
Unified CLI for the Agent Quality Toolkit (agentmd, coderace, agentlint, agentreflect).
Installation
pip install agentkit-cli
Quick Start
pip install agentkit-cli
agentkit quickstart # ๐ fastest path to a score โ start here
agentkit quickstart checks your toolchain, runs a fast composite score (agentlint + agentmd), prints a beautiful Rich summary, and optionally publishes a shareable score card โ all in under 60 seconds.
agentkit run # run the full pipeline
agentkit score # compute composite score
agentkit gate # fail if score < threshold
agentkit org github:vercel # score every public repo in a GitHub org
Demo
Record a terminal demo with VHS:
agentkit demo --record # generates demo.tape
vhs demo.tape # renders demo.gif
The tape records the full quickstart โ run โ benchmark flow.
Configuration
agentkit uses .agentkit.toml for project-level configuration.
agentkit config init # create .agentkit.toml with defaults
agentkit config show # show effective config with sources
agentkit config set gate.min_score 80
agentkit config get gate.min_score
Config Precedence
CLI flags > env vars > project .agentkit.toml > user config > defaults
Profiles
Profiles are named presets for gate thresholds, notify config, and sweep targets. Switch your entire quality policy in one command.
Built-in Presets
| Profile | Min Score | Max Drop | Notify On | Gate |
|---|---|---|---|---|
strict |
85 | 3 | fail | enabled |
balanced |
70 | 10 | never | enabled |
minimal |
50 | 20 | never | disabled |
Usage
# Switch to strict quality standards
agentkit profile use strict
# List all profiles (built-in + user-defined)
agentkit profile list
# Show profile details
agentkit profile show strict
# Run gate with a specific profile
agentkit gate --profile strict
# Create a custom profile based on strict
agentkit profile create myprofile --from strict --min-score 90
# Export a profile as JSON or TOML
agentkit profile export strict --format json
Using Profiles with Commands
All major commands support --profile:
agentkit gate --profile strict
agentkit run --profile balanced
agentkit sweep --profile minimal owner/repo1 owner/repo2
agentkit score --profile balanced
agentkit analyze --profile strict github:owner/repo
Explicit CLI flags always override profile values:
# Uses strict profile but overrides min-score to 99
agentkit gate --profile strict --min-score 99
agentkit llmstxt โ AI-Accessible Documentation
llms.txt is a standard that tells LLMs how to consume a project's documentation and API surface โ making your repo accessible to AI-powered tools beyond just coding agents.
# Generate llms.txt for current directory
agentkit llmstxt
# Generate both llms.txt and llms-full.txt (with inline content)
agentkit llmstxt --full --output ./dist/
# Analyze a GitHub repo
agentkit llmstxt github:tiangolo/fastapi --json
# Validate an existing llms.txt
agentkit llmstxt --validate
# Get quality score
agentkit llmstxt --score --json
Sample llms.txt output:
# my-project v1.2.0
> A fast, lightweight library for building AI agents.
## Docs
- [README](README.md): Project overview and getting started guide.
- [Changelog](CHANGELOG.md): Version history and release notes.
- [Guide](docs/guide.md)
## API
- [my-project API](my_project/__init__.py): Main API module.
## Examples
- [Basic Example](examples/basic.md)
Integration with existing commands:
# Generate llms.txt as part of standard run pipeline
agentkit run --llmstxt
# Include llms.txt card in HTML report
agentkit report --llmstxt
agentkit migrate โ Convert Between AI Agent Context Formats
Developers using Claude Code, Codex, and Gemini CLI each expect different context file formats (CLAUDE.md, AGENTS.md, llms.txt). agentkit migrate converts between them automatically.
| Source | Target | Notes |
|---|---|---|
AGENTS.md |
CLAUDE.md |
Operational rules โ project-focused format |
AGENTS.md |
llms.txt |
Operational rules โ llmstxt.org format |
CLAUDE.md |
AGENTS.md |
Project context โ operational format |
CLAUDE.md |
llms.txt |
Project context โ llmstxt.org format |
llms.txt |
CLAUDE.md |
AI-accessible docs โ CLAUDE.md |
llms.txt |
AGENTS.md |
AI-accessible docs โ AGENTS.md |
agentkit migrate # auto-detect source, generate all formats
agentkit migrate --all --force
agentkit migrate --from agents-md --to claude-md
agentkit migrate --dry-run
agentkit sync --check # exit 1 if stale
agentkit sync --fix # re-generate stale files
agentkit llmstxt --sync-from agents-md
agentkit run --migrate # generate missing formats before analysis
Commands
agentkit quickstartโ ๐ fastest path to a score (start here)agentkit runโ run the full pipelineagentkit scoreโ compute composite scoreagentkit gateโ fail if score < thresholdagentkit redteam [PATH]โ adversarial eval: score how well your agent context resists attacksagentkit analyze <target>โ analyze any GitHub repoagentkit sweep <targets>โ batch analyze multiple reposagentkit duel <repo1> <repo2>โ head-to-head agent-readiness comparisonagentkit user-duel github:<user1> github:<user2>โ head-to-head agent-readiness comparison between two GitHub developersagentkit user-tournament github:<u1> github:<u2> [github:<uN>...]โ bracket-style agent-readiness tournament for N GitHub developersagentkit tournament <repo1> ... <repoN>โ round-robin bracket across 4-16 reposagentkit profile <sub>โ manage quality profilesagentkit config <sub>โ manage configurationagentkit historyโ show score historyagentkit timelineโ visual quality timeline (HTML chart from history DB)agentkit leaderboardโ compare runs by labelagentkit insightsโ cross-repo pattern synthesisagentkit trendingโ fetch and rank trending GitHub repos by agent qualityagentkit dailyโ generate a daily leaderboard of the most agent-ready GitHub reposagentkit pages-trendingโ fetch trending repos, score for agent-readiness, publish daily leaderboard to GitHub Pagesagentkit org <owner>โ score every public repo in a GitHub org or user accountagentkit pr github:<owner>/<repo>โ submit a CLAUDE.md PR to any public GitHub repoagentkit campaign <target>โ batch PR submission to multiple repos in one commandagentkit search [query]โ discover GitHub repos missing CLAUDE.md / AGENTS.md
Search: Discover Repos Missing Context Files
agentkit search discovers GitHub repos that are missing CLAUDE.md or AGENTS.md โ the best targets for agentkit campaign.
# Find Python AI-agent repos without context files
agentkit search "ai agents" --language python --missing-only
# Filter by topic and minimum stars
agentkit search --topic ai-agents --min-stars 500 --limit 30
# Export as JSON (pipe to agentkit campaign targets)
agentkit search "llm tools" --missing-only --json > targets.json
# Generate a shareable HTML report
agentkit search "coding agents" --output report.html --share
# Full campaign flywheel: search โ campaign
agentkit campaign --from-search "ai agents" --language python --min-stars 500
The search result table shows each repo's star count, language, and whether CLAUDE.md or AGENTS.md is present.
Campaign: Batch PR Submission
agentkit campaign finds repos missing CLAUDE.md and submits PRs to all of them in one command.
# Submit CLAUDE.md PRs to all public repos in an org (up to 5, default)
agentkit campaign github:pallets
# Discover repos without submitting PRs (dry run)
agentkit campaign github:pallets --dry-run --limit 10
# Target by topic
agentkit campaign topic:ai-agents --language python --min-stars 500
# Use a file of repos
agentkit campaign repos-file:my-targets.txt
# Only discover repos (no PRs)
agentkit campaign github:pallets --skip-pr
# Generate and share an HTML report
agentkit campaign github:pallets --share
Example output:
Campaign ID: abc12345
Target: github:pallets Limit: 5 File: CLAUDE.md
โโโโโโโโโโโโโโฌโโโโโโโโโฌโโโโโโโโโฌโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ Repo โ Stars โ Status โ PR URL / Note โ
โโโโโโโโโโโโโโผโโโโโโโโโผโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ flask โ โ
68k โ โ
PR โ https://github.com/.../pull/42 โ
โ click โ โ
15k โ โ
PR โ https://github.com/.../pull/7 โ
โ jinja โ โ
10k โ โญ skipโ Already has context file โ
โ werkzeug โ โ
7k โ โ
PR โ https://github.com/.../pull/12 โ
โ markupsafe โ โ
600 โ โ err โ Fork creation failed โ
โโโโโโโโโโโโโโดโโโโโโโโโดโโโโโโโโโดโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
Campaign complete. 3 PRs opened, 1 skipped, 1 failed.
Options:
--limit Nโ max repos to target (default: 5)--language TEXTโ filter by language (e.g. python, typescript)--min-stars Nโ minimum stars threshold (default: 100)--file TEXTโ context file name (default: CLAUDE.md)--forceโ submit PR even if context file exists--dry-runโ show what would happen, no PRs opened--jsonโ output CampaignResult as JSON--no-filterโ skip the "already has context file" check--skip-prโ only discover repos, don't submit PRs--shareโ upload HTML report to here.now
agentkit track โ Monitor Campaign PR Outcomes
After running agentkit campaign, use agentkit track to see which PRs got merged, closed, or are still open.
# Show last 20 tracked PRs
agentkit track
# Filter to a specific campaign
agentkit track --campaign-id abc12345
# Show all PRs (no limit)
agentkit track --all
# JSON output for CI/automation
agentkit track --json
# Upload a shareable HTML status report
agentkit track --share
Example output:
โโโโโโโโโโโโโโโโฌโโโโโโโฌโโโโโโโโโฌโโโโโโโโโโโโฌโโโโโโโโโโฌโโโโโโโโโโโโโ
โ Repo โ PR # โ Status โ Days Open โ Reviews โ Submitted โ
โโโโโโโโโโโโโโโโผโโโโโโโผโโโโโโโโโผโโโโโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโโโโค
โ pallets/flaskโ 6001 โ merged โ 3 โ 2 โ 2026-03-14 โ
โ encode/httpx โ 892 โ open โ 1 โ 0 โ 2026-03-16 โ
โโโโโโโโโโโโโโโโดโโโโโโโดโโโโโโโโโดโโโโโโโโโโโโดโโโโโโโโโโดโโโโโโโโโโโโโ
2 merged, 1 open, 0 closed
Options:
--campaign-id TEXTโ filter to a specific campaign--limit Nโ max PRs to show (default: 20)--allโ show all tracked PRs (no limit)--jsonโ output structured JSON--shareโ upload dark-theme HTML report to here.now
View campaign history with:
agentkit history --campaigns
agentkit history --campaign-id <id>
Org Analysis
agentkit org answers: "Which repos in this GitHub org are most AI-agent-ready?"
# Score all public repos in an org or user account
agentkit org github:vercel
# Include forked and archived repos, cap at 20
agentkit org github:microsoft --include-forks --include-archived --limit 20
# Parallel analysis with 5 workers, save HTML report
agentkit org github:anthropics --parallel 5 --output report.html
# Share report online
agentkit org github:openai --share
# JSON output for scripting
agentkit org github:tiangolo --json
# Use GitHub token to avoid rate limits
agentkit org github:google --token ghp_xxx
# Auto-generate CLAUDE.md for repos below 80 and show before/after score lift
agentkit org github:pallets --generate
# Only generate for repos scoring below 60, share an HTML before/after report
agentkit org github:pallets --generate --generate-only-below 60 --share
--generate flag
--generate turns the audit from read-only to actionable: for every repo below the threshold (default: 80), it clones the repo locally, runs agentmd generate to create a CLAUDE.md, re-scores the repo, and shows the before/after lift.
Before: pallets/flask 28.6/F
After: pallets/flask 91.4/A (+62.8 pts)
All generation is done in temporary local clones โ no remote writes to GitHub.
Options:
--generate-only-below Nโ only generate for repos scoring below N (default: 80)--sharewith--generateโ HTML report shows Before / After columns with color-coded delta badges
Benchmark
agentkit benchmark answers: "Which AI agent โ Claude, Codex, or Gemini โ performs best on YOUR specific codebase?"
# Benchmark all three agents on current project (default tasks)
agentkit benchmark
# Custom agents and tasks
agentkit benchmark --agents claude,codex --tasks bug-hunt,refactor
# 3 rounds for statistical confidence
agentkit benchmark --rounds 3
# Export JSON results
agentkit benchmark --json > results.json
# Save HTML report
agentkit benchmark --output report.html
# Publish shareable dark-theme report
agentkit benchmark --share
The benchmark runs 5 built-in tasks (bug-hunt, refactor, concurrent-queue, api-client, context-use) against each agent via coderace and produces a ranked comparison table showing mean score, mean time, and win rate.
Trending Analysis
agentkit trending answers: "Which repos blowing up on GitHub are most AI-agent-ready today?"
# Rank this week's trending AI repos (default)
agentkit trending
# Fast mode: list repos without scoring
agentkit trending --no-analyze
# Filter by topic, publish a shareable report
agentkit trending --topic ai-agent --share
# Weekly trending, top 15, min 500 stars, JSON output
agentkit trending --period week --limit 15 --min-stars 500 --json
# Use a GitHub token for higher rate limits
agentkit trending --token ghp_xxx
Output: a ranked Rich table (Rank | Repo | Stars | Score | Grade | URL) and optionally a dark-theme HTML report published to here.now.
Daily Leaderboard
agentkit daily is a content flywheel: run once/day, get a shareable ranked HTML report showing "Today's most AI-agent-ready repos."
# Show today's leaderboard in the terminal
agentkit daily
# Specify a date
agentkit daily --date 2026-03-19
# Publish to here.now and print the URL
agentkit daily --share
# Cron-friendly: output URL only (pipe into scripts or post to X)
agentkit daily --share --quiet
# Save HTML report to a file
agentkit daily --output daily-report.html
# JSON output
agentkit daily --json
# Filter by minimum score
agentkit daily --min-score 70
Example output (terminal):
agentkit daily โ date: 2026-03-19, limit: 20
Rank Repo Stars Score Top Finding
๐ฅ microsoft/autogen 30,000 91 Multi-agent framework with strong tool support
๐ฅ openai/openai-python 25,000 88 Well-structured SDK with clear API surface
๐ฅ anthropics/anthropic-sdk 12,000 85 Comprehensive documentation and type hints
#4 langchain-ai/langchain 8,000 78 Agent/LLM keyword in description
Permanent GitHub Pages URL
Use --pages to publish a permanent, auto-updating leaderboard to GitHub Pages:
# Publish to GitHub Pages (auto-detects repo from git remote)
agentkit daily --pages
# Target a specific repo
agentkit daily --pages --pages-repo github:owner/repo
# Override the output path (default: docs/leaderboard.html)
agentkit daily --pages --pages-path docs/leaderboard.html
On success, prints: https://owner.github.io/repo/leaderboard.html
If GitHub Pages publish fails, falls back to --share (here.now 24h link) automatically.
GitHub Actions cron example
# .github/workflows/examples/agentkit-daily-leaderboard-pages.yml
on:
schedule:
- cron: '0 8 * * *' # 8 AM UTC daily
permissions:
contents: write
pages: write
jobs:
publish-leaderboard:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v4
with:
token: ${{ secrets.GITHUB_TOKEN }}
- run: pip install agentkit-cli
- run: agentkit daily --pages --pages-repo github:${{ github.repository }}
env:
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
See full example: .github/workflows/examples/agentkit-daily-leaderboard-pages.yml
Share (24h URL, default)
# Publish to here.now and print the URL (24h expiry)
agentkit daily --share
# Cron-friendly: output URL only
agentkit daily --share --quiet
Add to your GitHub Actions for automated daily publishing:
- name: Run daily leaderboard
run: |
URL=$(agentkit daily --share --quiet)
echo "url=$URL" >> "$GITHUB_OUTPUT"
See the full example: .github/workflows/examples/agentkit-daily-leaderboard.yml
Tournament
agentkit tournament runs a round-robin bracket across 4-16 repos and ranks them by win/loss record with avg score tiebreak.
# Run a 4-repo tournament
agentkit tournament github:fastapi/fastapi github:tiangolo/starlette github:django/django github:pallets/flask
# Publish a shareable HTML bracket report
agentkit tournament github:fastapi/fastapi github:tiangolo/starlette github:django/django github:pallets/flask --share
# JSON output for CI/scripting
agentkit tournament github:fastapi/fastapi github:tiangolo/starlette github:django/django github:pallets/flask --json
# Sequential (no parallel), quiet mode, save HTML
agentkit tournament github:fastapi/fastapi github:tiangolo/starlette github:django/django github:pallets/flask \
--no-parallel --quiet --output bracket.html
Output: standings table (Rank | Repo | W-L | Avg Score | Grade), match results matrix, and winner banner. Use --share to publish a dark-theme HTML bracket to here.now.
Portfolio Insights
Once you've analyzed multiple repos with agentkit analyze or agentkit run, the
agentkit insights command synthesizes patterns across all historical runs:
# Portfolio health summary (avg score, best/worst repo, top issue)
agentkit insights
# Most common agentlint findings across all repos
agentkit insights --common-findings
# Repos scoring in the bottom quartile
agentkit insights --outliers
# Repos with significant score movement between runs
agentkit insights --trending
# All sections in one view
agentkit insights --all
# Machine-readable JSON (useful for scripts/dashboards)
agentkit insights --json
# Use a specific history DB
agentkit insights --db /path/to/history.db
Store agentlint findings alongside scores for richer cross-repo analysis:
agentkit run --record-findings
agentkit analyze github:owner/repo --record-findings
JSON output schema:
{
"portfolio_summary": {
"avg_score": 74.5,
"total_runs": 12,
"unique_repos": 4,
"top_issue": "missing-tools-section",
"best_repo": "owner/repo-a",
"worst_repo": "owner/repo-d"
},
"common_findings": [
{"finding": "missing-tools-section", "repo_count": 3, "total_occurrences": 5}
],
"outliers": [
{"project": "owner/repo-d", "latest_score": 42.0, "avg_score": 48.5, "run_count": 2}
],
"trending": [
{"project": "owner/repo-b", "previous_score": 55.0, "latest_score": 80.0, "delta": 25.0, "direction": "up"}
]
}
Publishing & Sharing
Org Leaderboard (New in v0.58.0)
Publish a live org-wide AI-readiness leaderboard to GitHub Pages with one command:
# Score all public repos in an org and publish a leaderboard
agentkit pages-org github:myorg
# Publish from within agentkit org (after scoring)
agentkit org github:myorg --pages
# Options
agentkit pages-org github:myorg --pages-repo myorg/custom-scores
agentkit pages-org github:myorg --only-below 80 # only repos below score 80
agentkit pages-org github:myorg --limit 20
agentkit pages-org github:myorg --dry-run # skip git push
agentkit pages-org github:myorg --quiet # print URL only (cron mode)
The leaderboard is published to https://<owner>.github.io/agentkit-scores/ by default.
Enable GitHub Pages on <owner>/agentkit-scores (Settings โ Pages โ branch: main, folder: /docs).
For weekly automated updates, use the example workflow:
.github/workflows/examples/agentkit-org-pages.yml
Pages Trending: Daily AI-Ready Repo Leaderboard
agentkit pages-trending fetches today's trending GitHub repos, scores them for agent-readiness, and publishes a persistent dark-theme leaderboard to GitHub Pages at https://<owner>.github.io/<repo>/trending.html.
# Publish daily trending leaderboard (uses GITHUB_TOKEN)
agentkit pages-trending
# Filter to Python trending repos this week
agentkit pages-trending --language python --period week
# Dry run โ score and generate HTML without pushing
agentkit pages-trending --dry-run
# Custom pages repo, limit 30 repos
agentkit pages-trending --pages-repo github:myorg/my-trending --limit 30
# Cron-friendly: print only the URL
agentkit pages-trending --quiet
# Publish + generate a 24h preview link
agentkit pages-trending --share
The leaderboard is published to https://<owner>.github.io/<repo>/trending.html.
For daily automated updates, use: .github/workflows/examples/agentkit-trending-pages.yml
Sharing Results
Share your agent quality score card with a single command:
# Generate and upload a score card to here.now
agentkit share
# Share from a saved JSON report
agentkit share --report agentkit-report.json
# Hide raw numbers (show pass/fail only)
agentkit share --no-scores
# Output JSON with URL and score
agentkit share --json
# Auto-share after a run
agentkit run --share
# Auto-share after generating a report
agentkit report --share
# Quickest way to get a score + share URL for any repo
agentkit quickstart github:owner/repo
# Full analyze with share (more detail, slower)
agentkit analyze github:owner/repo --share
# Batch analyze repos and share a combined scorecard
agentkit sweep github:owner/repo1 github:owner/repo2 --share
Score cards are standalone HTML pages (dark theme) showing: composite score, per-tool breakdown, project name, git ref, and timestamp. Anonymous cards expire in 24h; set HERENOW_API_KEY for persistent links.
GitHub Actions
Use the agentkit GitHub Action to run quality checks on every PR:
- uses: mikiships/agentkit-cli@v0.7.0
with:
github-token: ${{ secrets.GITHUB_TOKEN }}
min-score: 70
Or install and run directly:
- uses: actions/checkout@v4
- run: pip install agentkit-cli
- run: agentkit gate --profile strict
See agentkit setup-ci for automated workflow generation.
Local Dashboard
agentkit serve starts a lightweight local web dashboard showing all toolkit runs from the history database:
agentkit serve [OPTIONS]
Options:
--port PORT Port to serve on (default: 7890)
--open Auto-open the dashboard in your browser on start
--once Render dashboard HTML to stdout and exit (no server)
--json Print server URL as JSON and exit (useful for scripts)
The dashboard shows a dark-theme summary of every project run: latest score, grade (AโF), per-tool breakdown, timestamp, and run ID. Scores are color-coded green (โฅ80), yellow (โฅ60), and red (<60). The page auto-refreshes every 30 seconds.
Quick start:
agentkit serve --open # start server + open browser
agentkit run --serve # run pipeline, then print dashboard URL
agentkit serve --once > out.html # render to file
No external dependencies โ uses Python stdlib only (http.server, threading, webbrowser).
Live Dashboard
Run once and watch scores update in real-time:
# Combined: watch files + serve dashboard (updates without reload)
agentkit watch --serve --port 7890
# Or start server in live mode (polls for external writes):
agentkit serve --live
The dashboard connects via SSE (/events) and re-renders the runs table in-place when new pipeline results arrive. A โ Live indicator shows connection status; it drops to โ Offline if the server stops.
agentkit pr โ Submit CLAUDE.md PRs to Open Source Repos
agentkit pr is a viral distribution mechanic: one command generates a CLAUDE.md for any public GitHub repo and opens a PR against it.
# Submit a CLAUDE.md PR to a public repo
agentkit pr github:owner/repo
# Preview what would happen (no git or API calls)
agentkit pr github:owner/repo --dry-run
# Generate AGENTS.md instead
agentkit pr github:owner/repo --file AGENTS.md
# Force overwrite if CLAUDE.md already exists
agentkit pr github:owner/repo --force
# JSON output
agentkit pr github:owner/repo --json
Requires: GITHUB_TOKEN environment variable with repo and workflow scopes.
export GITHUB_TOKEN=ghp_...
agentkit pr github:vercel/next.js
What it does:
- Clones the repo (shallow, depth 1)
- Runs
agentmd generate .to create CLAUDE.md - Forks the repo under your authenticated GitHub account (if needed)
- Creates a branch
agentkit/add-claude-md - Commits and pushes the generated file
- Opens a PR against the original repo
Release Check
agentkit release-check verifies the 4-part release surface to confirm a package is truly shipped, not just locally complete:
agentkit release-check [PATH] [OPTIONS]
Options:
--version VERSION Version to verify (default: from pyproject.toml/package.json)
--package NAME Package name (default: from pyproject.toml/package.json)
--registry pypi|npm|auto (default: auto-detected)
--skip-tests Skip the pytest/npm test step for quick checks
--json Output structured JSON for CI integration
Example output:
agentkit release-check โ /your/project
โโโโโโโโโโโโโโฌโโโโโโโโโฌโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ Check โ Status โ Detail โ
โโโโโโโโโโโโโโผโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ tests โ โ PASS โ 42 passed in 1.23s โ
โ git_push โ โ PASS โ Local HEAD abc12345 matches rem โ
โ git_tag โ โ PASS โ Tag v1.0.0 found on remote. โ
โ registry โ โ PASS โ PyPI: mypkg==1.0.0 is live. โ
โโโโโโโโโโโโโโดโโโโโโโโโดโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
Verdict: SHIPPED
Verdict levels:
- SHIPPED โ all 4 surfaces confirmed (exit 0)
- RELEASE-READY โ tests + git confirmed, registry not yet live (exit 1)
- BUILT โ tests pass locally, not yet pushed (exit 1)
- UNKNOWN โ tests failing (exit 1)
Integrate with agentkit gate --release-check or agentkit run --release-check to add release verification to your pipeline.
Architecture
All quartet tool invocations (agentmd, agentlint, coderace, agentreflect) go through ToolAdapter in agentkit_cli/tools.py. This ensures canonical correct flags are used everywhere and flag-wiring bugs cannot recur across subcommands.
Run pytest -m smoke before any release to catch integration regressions.
Automated Improvement (agentkit improve)
agentkit improve closes the full loop: analyze โ identify improvements โ apply fixes โ re-analyze โ show delta.
agentkit improve
agentkit improve --dry-run
agentkit improve --no-generate --no-harden
agentkit improve --min-lift 10
agentkit improve --json
agentkit improve --output report.html
agentkit run --improve
AI-Powered Explanations
agentkit explain calls an LLM (Claude via Anthropic API) to generate a human-readable coaching report explaining why your scores are what they are โ not just what to fix, but what it means for agents working on your codebase.
# Explain scores for the current project (template mode, no API key needed)
agentkit explain --no-llm .
# Explain a saved run report (LLM mode, requires ANTHROPIC_API_KEY)
agentkit explain --report report.json
# Get structured JSON output
agentkit explain --no-llm . --json
# Save coaching report to a file
agentkit explain --no-llm . --output coaching.md
# Run the full pipeline then get a coaching report in one command
agentkit run --explain --no-llm .
The coaching report includes four sections:
- What This Score Means โ plain language interpretation for your tier (A/B/C/F)
- Key Findings Explained โ why each issue actually hurts agent performance
- Top 3 Next Steps โ ordered by impact
- If You Do Nothing Else โ the single most important action
Use --no-llm for offline mode (CI environments, no API key). Set ANTHROPIC_API_KEY for LLM-powered coaching via claude-3-5-haiku-20241022.
The meta-angle: "Your AI agent's AI quality score, explained by AI."
agentkit certify
Generate a dated, shareable certification report proving a repo passed all agentkit quality checks.
# Run cert on current directory
agentkit certify .
# Output JSON cert (for CI integration)
agentkit certify . --json
# Write HTML cert card to file
agentkit certify . --output cert.html
# Share HTML report via here.now (requires HERENOW_API_KEY)
agentkit certify . --output cert.html --share
# Fail exit if composite score < 80
agentkit certify . --min-score 80
# Inject/update cert badge in README.md
agentkit certify . --badge
# Preview badge change without writing
agentkit certify . --badge --dry-run
The cert report includes:
- cert_id: 8-char hex fingerprint (prefix of SHA256)
- timestamp: UTC ISO 8601
- verdict: PASS / WARN / FAIL
- Composite Score (agentkit score) โ PASS โฅ 80
- Redteam Resistance (agentkit redteam) โ PASS โฅ 70
- Context Freshness (agentlint check-context) โ PASS โฅ 70
- Tests Found (agentkit doctor)
- SHA256 content hash for tamper detection
License
MIT
Timeline
agentkit timeline generates a dark-theme HTML chart showing your composite score progression over time. Reads from the existing SQLite history DB populated by agentkit run.
# Generate timeline for all projects
agentkit timeline
# Filter to one project
agentkit timeline --project my-agent
# Show only the last 20 runs since a date
agentkit timeline --limit 20 --since 2026-01-01
# Output raw chart data as JSON
agentkit timeline --json
# Publish and share
agentkit timeline --share
# Auto-generate timeline after a run
agentkit run --timeline
The report includes:
- Main chart: line chart (x = date, y = composite score), one line per project
- Per-tool breakdown: CSS-bar sparklines for lint score, code quality, context freshness, test count
- Stats panel: min/max/avg, trend direction (โโโ), streak badge (e.g. "12 runs above 80")
- Project summary table: run count, latest score, trend per project
Red-Team Your Agent Setup
agentkit redteam scores how well your agent context file (CLAUDE.md / AGENTS.md) resists adversarial attacks. Static analysis only โ no LLM required. Truly model-agnostic.
# Analyze current directory
agentkit redteam
# Analyze a specific project
agentkit redteam ./my-agent-project
# CI gate: fail if resistance score < 70
agentkit redteam --min-score 70
# JSON output for programmatic use
agentkit redteam --json
# Save HTML report
agentkit redteam --output redteam-report.html
# Share HTML report via here.now
agentkit redteam --share
Categories checked:
prompt_injectionโ attempts to inject instructions via user inputjailbreakโ persona and restriction bypass attemptscontext_confusionโ fake context and history injectioninstruction_overrideโ priority and mode override attemptsdata_extractionโ system prompt and credential extractionrole_escalationโ privilege and authority escalation
CI integration:
- name: Red-team agent config
run: agentkit redteam --min-score 70
Exit code 1 if --min-score threshold is not met. Combine with agentkit run --redteam to add adversarial eval to your full pipeline.
Distribution angle: After OpenAI's $86M acquisition of Promptfoo, teams using non-OpenAI models need a neutral red-team tool. Static analysis = no model dependency = truly model-agnostic.
Auto-Harden Your Agent Context
agentkit harden is the detectโfix loop closed in one command. Run it after agentkit redteam to auto-patch all detected vulnerabilities.
# Analyze and auto-remediate CLAUDE.md / AGENTS.md in cwd
agentkit harden
# Harden a specific file or directory
agentkit harden ./my-agent-project
# Preview what would change without writing
agentkit harden --dry-run
# Write hardened file to a different path
agentkit harden --output hardened-CLAUDE.md
# JSON output for CI integration
agentkit harden --json
# Generate dark-theme HTML score-card report
agentkit harden --report
# Apply fix flag in redteam command
agentkit redteam --fix
# Auto-apply with dry-run preview
agentkit redteam --fix --dry-run
# Run harden after full pipeline
agentkit run --harden
What agentkit harden does:
- Detects all 6 vulnerability categories (prompt injection, jailbreak, context confusion, instruction override, data extraction, role escalation)
- Applies targeted, idempotent remediations (never duplicates existing sections)
- Creates a backup (
.bak) before modifying files - Re-scores the hardened file and shows a before/after table
Idempotent: Running it multiple times on an already-hardened file makes no additional changes.
agentkit monitor โ Continuous Quality Monitoring
Set up continuous quality monitoring for your repos. Get notified on Slack or Discord when scores change significantly.
# Add a repo to monitor (default: daily, alert on 10-point change)
agentkit monitor add github:owner/repo
# Weekly schedule with Slack notification
agentkit monitor add github:owner/repo --schedule weekly --notify-slack https://hooks.slack.com/...
# Alert when score drops below 80 OR changes by 5+ points
agentkit monitor add github:owner/repo --min-score 80 --alert-threshold 5
# List all monitored targets (last score, next due, notify configured)
agentkit monitor list
# Force an immediate check on all due targets
agentkit monitor run
# Force-check a specific target
agentkit monitor run --target github:owner/repo
# Start the background daemon (polls every 60 seconds)
agentkit monitor start
# Check daemon status and next scheduled runs
agentkit monitor status
# View recent check history
agentkit monitor logs --limit 20
# Stop the daemon
agentkit monitor stop
# Remove a target
agentkit monitor remove github:owner/repo
Schedules: hourly, daily (default), weekly
Notifications: Configure Slack (--notify-slack), Discord (--notify-discord), or any generic webhook (--notify-webhook). Fires when abs(score_delta) >= alert_threshold (default 10 points) or score drops below --min-score.
Daemon: Runs as a background subprocess, writing structured JSON lines to ~/.agentkit/monitor.log. PID stored in ~/.agentkit/monitor.pid. Handles SIGTERM gracefully.
GitHub Webhook Integration
agentkit webhook closes the "outside-in" CI loop: instead of only running agentkit from inside CI, GitHub pushes events to agentkit, which automatically analyzes the repo and fires notifications.
Quick Start
# 1. Configure the HMAC secret (must match GitHub webhook settings)
agentkit webhook config --set-secret <YOUR_GITHUB_WEBHOOK_SECRET>
# 2. Start the server
agentkit webhook serve --port 8080
# Listening on http://localhost:8080
# Point your GitHub webhook at this URL (use ngrok for public exposure)
Subcommands
| Command | Description |
|---|---|
agentkit webhook serve [--port P] [--secret S] [--no-verify-sig] |
Start the HTTP server |
agentkit webhook config [--show] [--set-secret S] [--set-port P] [--set-channel URL] |
Manage configuration |
| `agentkit webhook test [--event push | pull_request] [--repo REPO]` |
Configuration (.agentkit.toml)
[webhook]
port = 8080
secret = "" # HMAC secret from GitHub webhook settings
notify_channels = [] # Reuse existing NotificationService channels
How It Works
- GitHub POSTs a
pushorpull_requestevent to your server. - Server verifies the
X-Hub-Signature-256HMAC and responds 200 immediately. - Background thread calls
EventProcessor.process():- Runs
CompositeScoreEngineon the repo. - Records the score in history DB (
agentkit history). - Fires a notification if the score dropped by โฅ 5 points vs previous run.
- Formats a PR comment body (logged to stdout; actual GitHub API posting is out of scope).
- Runs
Doctor Check
agentkit doctor reports webhook configuration health under the Integrations section:
integrations webhook config WARN Webhook configured but HMAC secret is empty.
Post Run Notification
agentkit run --webhook-notify
After the pipeline completes, POSTs a JSON summary to notify.webhook_url from .agentkit.toml.
GitHub Checks API
When running in GitHub Actions, agentkit run and agentkit gate automatically post a native GitHub Check Run with your composite score, grade, and per-tool breakdown โ visible directly in the PR UI.
Automatic Mode (CI)
No extra config needed. When GITHUB_ACTIONS=true and GITHUB_TOKEN is available, check runs are posted automatically:
# In your GitHub Actions workflow:
permissions:
contents: read
checks: write # Required for Checks API
steps:
- run: agentkit run --ci
env:
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
Use --no-checks to disable, or --checks to force even when auto-detection fails.
Manual Commands
agentkit checks verify # test that Checks API is configured
agentkit checks post --score 87 # manually post a check run
agentkit checks status # show last check run posted
What Gets Posted
- Title:
Agent Quality: 87/100 (B) - Summary: composite score + gate verdict (PASS/FAIL)
- Body: markdown table of per-tool scores with pass/warn/fail indicators
- Annotations: one annotation per failing tool (score < 80)
- Linked scorecard if
--shareis active
User Duel: Head-to-Head Developer Comparison
agentkit user-duel compares two GitHub developers' agent-readiness side-by-side. It runs user-scorecard for each and declares a winner per dimension.
# Basic comparison
agentkit user-duel github:tiangolo github:kennethreitz
# Limit repos per user and output JSON
agentkit user-duel github:mikiships github:tiangolo --limit 3 --json
# Share a duel report link
agentkit user-duel github:tiangolo github:kennethreitz --share
# Just print the winner (cron/scripting friendly)
agentkit user-duel github:tiangolo github:kennethreitz --quiet
Dimensions compared: avg_score, letter_grade, repo_count, agent_ready_repos. Overall winner is determined by majority of dimension wins. Tie-friendly output included.
agentkit user-tournament
agentkit user-tournament runs a bracket-style agent-readiness tournament for N GitHub developers. Round-robin mode for โค8 participants, bracket mode for >8. Champion is determined by wins with avg score as tiebreak.
# Run a tournament between three developers
agentkit user-tournament github:tiangolo github:kennethreitz github:mikiships
# Output as JSON
agentkit user-tournament github:tiangolo github:kennethreitz --json
# Publish and share HTML report
agentkit user-tournament github:tiangolo github:kennethreitz github:mikiships --share
# Save HTML to local file
agentkit user-tournament github:tiangolo github:kennethreitz --output tournament.html
Use --limit N to cap comparisons and --quiet for scripting-friendly champion-only output.
agentkit user-team
agentkit user-team analyzes a GitHub org's top contributors for agent-readiness. Each contributor is scored via UserScorecardEngine, then results are aggregated into a team scorecard with a ranked table, aggregate grade, and grade distribution.
# Score the top 10 contributors of an org
agentkit user-team github:pallets
# Limit to 5 contributors
agentkit user-team github:django --limit 5
# Output as JSON
agentkit user-team github:pallets --json
# Save HTML report to file
agentkit user-team github:pallets --output team-report.html
# Publish and share HTML report
agentkit user-team github:pallets --share
Use --quiet for CI-friendly output (only prints share URL if --share is set).
agentkit user-improve
agentkit user-improve finds a GitHub user's lowest-scoring public repos and automatically improves them by generating CLAUDE.md context files and applying agent hardening. Displays a before/after quality lift report.
# Improve top lowest-scoring repos for a user
agentkit user-improve github:tiangolo
# Target repos scoring below 70, up to 10 repos
agentkit user-improve github:kennethreitz --below 70 --limit 10
# Dry run: show what would be improved without applying changes
agentkit user-improve github:mikiships --dry-run
# Output as JSON
agentkit user-improve github:tiangolo --json
# Publish HTML improvement report to here.now
agentkit user-improve github:tiangolo --share
Use --limit N (default 5, max 20) to control how many repos are targeted, and --below N (default 80) to set the quality threshold.
agentkit user-card
agentkit user-card generates a compact, embeddable agent-readiness card for a GitHub user. The card shows grade, avg score, context coverage, agent-ready repo count, and top repo โ all in a shareable dark-theme HTML card.
# Generate a card for a user
agentkit user-card github:tiangolo
# Publish card to here.now and get a shareable URL
agentkit user-card github:tiangolo --share
# Output as JSON
agentkit user-card github:tiangolo --json
# Quiet mode (cron-friendly): print only the URL
agentkit user-card github:tiangolo --share --quiet
# Include forks, analyze up to 20 repos
agentkit user-card github:mikiships --no-skip-forks --limit 20
Use --limit N (default 10, max 30) to control how many repos are analyzed. The HTML card includes a Markdown embed snippet as an HTML comment when --share is used.
agentkit user-badge
agentkit user-badge generates a shields.io agent-readiness badge for a GitHub user's profile README โ a viral mechanic for spreading agent-readiness awareness organically.
# Generate badge (runs full scorecard scan)
agentkit user-badge github:torvalds
# Fast mode โ skip scan, generate from explicit score
agentkit user-badge github:torvalds --score 85
# Inject badge into local README.md (idempotent)
agentkit user-badge github:torvalds --score 85 --inject
# Preview inject without modifying files
agentkit user-badge github:torvalds --score 85 --inject --dry-run
# Write badge markdown to file
agentkit user-badge github:torvalds --score 85 --output badge.md
# JSON output
agentkit user-badge github:torvalds --score 85 --json
# Show badge after scorecard
agentkit user-scorecard github:torvalds --badge
# Show badge after user-card
agentkit user-card github:torvalds --badge
Badge grades: Aโฅ90 (brightgreen), Bโฅ75 (green), Cโฅ60 (yellow), Dโฅ45 (orange), F<45 (red).
Example badge: [](https://pypi.org/project/agentkit-cli/)
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file agentkit_cli-0.66.0.tar.gz.
File metadata
- Download URL: agentkit_cli-0.66.0.tar.gz
- Upload date:
- Size: 661.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
38ea0d086e9abec1d70aa7d92bafc36010b6398c470159e91e5059047ebf7a2d
|
|
| MD5 |
a4c4c5df559df8e47e72a56b08e00f01
|
|
| BLAKE2b-256 |
c76ed1090abbb4f257ed44f90fc5ed65d08105592b26c93f3bffa4577102ae16
|
File details
Details for the file agentkit_cli-0.66.0-py3-none-any.whl.
File metadata
- Download URL: agentkit_cli-0.66.0-py3-none-any.whl
- Upload date:
- Size: 390.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
33616df1da1ff182b8a3bdad4b50937df476b4f21fc9b04f61cc4595c9257345
|
|
| MD5 |
89e41155df815d2f2231d1e430236c15
|
|
| BLAKE2b-256 |
2a409a7cf98a0dcb3690bd6f4eb94cb640252c1818a499692a5a447b418932ed
|