Skip to main content

Security-first open-source coding agent with parallel tool execution, multimodal input, 4-tier permissions, audit trails, and 200+ LLM provider support

Project description

Godspeed

Trusted open-source coding agent for production code.

CI Python 3.11+ LiteLLM Ruff License: MIT

Project Status: Alpha — Seeking Testers

Godspeed is under active development by a solo maintainer. Breaking changes may occur between releases. We are looking for early adopters to test real-world workflows and report rough edges. See ROADMAP.md for the path to v1.0 stable.

A coding agent you can point at a production codebase and trust. Built-in permissions, audit trails, schema-validated tool calls, and automatic retries — so the agent writes safe code without babysitting.

Getting Started | Features | Architecture | Configuration | Contributing


What's new in v0.4.0

Five additions on top of the security-first core. Every one shipped with tests + CI-green across Python 3.11 / 3.12 / 3.13.

Area What changed Why
/remember command Persist permission rules from the TUI without editing YAML. Friction-free rule management — approve, deny, ask with --project scoping.
Model routing Task-aware routing (plan/edit/read/shell) to different models. Use cheap models for edits, frontier models for planning. Zero config with cheap_model/strong_model shortcuts.
Auto-load env files ~/.godspeed/.env.local and project .env files loaded on startup. No more $env:NVIDIA_NIM_API_KEY = ... every shell session.
Post-edit syntax gate .py / .pyi / .json edits that break parse are rejected before write. Catches indentation and syntax errors from fuzzy matching.
Windows UTF-8 stdio CLI auto-wraps stdout/stderr with errors='replace' at startup. Fixes UnicodeEncodeError on default cp1252 consoles.

Platform docs: Windows users should read docs/quickstart_windows.md for platform-specific setup (Miniconda, PYTHONIOENCODING, WSL for SWE-bench).

Why

Every open-source coding agent gives an LLM the ability to read files, write code, and run shell commands. Most of them assume the model will behave — and when it doesn't, you get broken files, leaked secrets, or worse. Godspeed is built for the reality that LLMs make mistakes, hallucinate tool calls, and sometimes go off-track.

Godspeed pairs full coding capability (30+ built-in tools, 200+ LLM providers, MCP support, parallel execution) with a trust model that fails safe by default. Every tool call is validated against its JSON Schema before execution. Transient failures are automatically retried with exponential backoff. Every action passes through a configurable permission engine (strict, normal, or yolo mode) and is recorded in a hash-chained audit log you can cryptographically verify. Secrets are caught at multiple layers before they reach the model or the log file.

If you want a coding agent you can actually point at a production codebase — one that catches its own mistakes, retries on transient failures, and leaves a verifiable trail of every action — this is it.

Features

Trust & Reliability

  • Schema-validated tool calls — every tool argument is validated against its JSON Schema before execution. Missing required fields, wrong types, and invalid values are caught with clear error messages — no silent failures.
  • Automatic retry on transient failures — network timeouts, connection resets, and rate limits trigger automatic retries with exponential backoff. Logic errors and permission denials fail immediately.
  • 3-tier permission modesstrict (deny most, ask for everything), normal (deny-first with allow rules), yolo (no permission checks, maximum speed). Configure via CLI (--permission-mode) or settings.yaml.
  • 4-tier permission engine -- deny-first evaluation with pattern matching, dangerous command detection (71 patterns), and fail-closed defaults. No tool call executes without explicit permission in strict/normal mode.
  • Hash-chained audit trail -- SHA-256 JSONL log where each entry chains to the previous. Tamper-evident, compressible, and verifiable with godspeed audit verify. Writes fail closed: any I/O error raises AuditWriteError and the chain state does not advance.
  • Secret protection -- 4 layers of defense: file deny-listing, context cleaning, output filtering, and audit redaction. 27 regex patterns plus Shannon entropy analysis catch API keys, tokens, and credentials before they leak.
  • Post-edit syntax gate -- .py / .pyi / .json edits that break parse are rejected before write. Lint-fix retry loop up to 3 rounds auto-corrects common issues.
  • Diff approve-before-write -- file_edit / file_write / diff_apply prompt with a side-by-side diff before hitting disk. (y)es · (n)o · (a)lways keys. Two independent axes of consent: permission engine ("may this tool run?") + diff reviewer ("apply THIS specific diff?").

Capability

  • 200+ LLM providers -- Claude, GPT, Gemini, Ollama, and everything else LiteLLM supports. Configure fallback chains so work never stops.
  • 30+ built-in tools -- file_read (images, PDFs, notebooks), file_write, file_edit (fuzzy matching), notebook_edit, file_move, image_read, pdf_read, shell (foreground + background), glob, grep, git, github (PR/issue workflow via gh), diff_apply (unified diffs), verify (6 languages), test_runner (5 frameworks), web_search, web_fetch, repo_map, code_search, tasks, background_check, coverage, complexity, dep_audit, security_scan, generate_tests, traceback_analyzer, system_optimizer, stock_price, db_query, and ollama_manager.
  • Parallel tool execution -- when the LLM returns multiple tool calls, they execute concurrently via asyncio.gather(). 3-phase dispatch: parse → permission check (sequential) → execute (parallel). READ_ONLY tools always parallel, write tools always serial.
  • Speculative tool dispatch -- during streaming, READ_ONLY tool calls are dispatched as background asyncio.Tasks before the full response completes. The main loop awaits cached results instead of re-dispatching, eliminating dispatch latency for reads.
  • Extended thinking -- pass thinking parameter to Anthropic/Claude models with configurable token budget. /think [budget] slash command. Thinking blocks displayed in collapsed dim panel.
  • Architect mode -- /architect toggles a two-phase pipeline. Phase 1 uses read-only tools to produce a plan. Phase 2 uses full tools guided by the plan. Configurable architect model.
  • Cost budget enforcement -- hard cost limit via max_cost_usd config or /budget command. Agent stops when exceeded. Ollama always free.
  • Self-evolution -- learn from execution traces to improve tool descriptions, system prompt sections, and permission configs. GEPA-style LLM-guided mutations scored by A/B testing with LLM-as-judge. Safety gate prevents regressions (size limits, semantic drift caps, human review). Runs entirely on Ollama for $0 with hardware-aware model selection (RTX 5070 Ti down to Jetson Orin Nano). /evolve command.
  • Sub-agent coordinator -- spawn isolated sub-agents for parallel tasks, each with their own conversation context. Depth limit 3, reuses the same async agent loop.
  • MCP client -- connect to Model Context Protocol servers via stdio or SSE transport. Remote tools are auto-adapted to Godspeed's Tool ABC with HIGH risk level.
  • Model routing -- route LLM calls by task type (plan/edit/chat) to different models. Use a cheap model for edits and a frontier model for planning.
  • Human-in-the-loop -- /pause stops the agent at the next iteration, /guidance <msg> injects mid-conversation correction and resumes.
  • Conversation compaction -- model-aware summarization when approaching the token limit. Small models (Ollama defaults) get aggressive compaction to preserve context slots; frontier models (Claude, GPT-4) get detailed preservation with structured summaries. Uses the cheapest model in the fallback chain to minimize cost. Compaction preserves: tool call history, file paths mentioned, key decisions, and user guidance. Summarized content is marked with [compacted: N messages → M messages] so the agent knows context was condensed.
  • Background commands -- shell tool gains background: true parameter. BackgroundRegistry tracks processes. background_check tool for status/output/kill.
  • Checkpoint save/restore -- /checkpoint name saves conversation state, /restore name loads it back. Never lose context again.
  • Memory -- SQLite-backed persistent preferences, session event logging, and automatic correction tracking across sessions.
  • Cross-agent project instructions -- loads GODSPEED.md, AGENTS.md (Linux Foundation standard), CLAUDE.md, and .cursorrules. Zero-friction migration from any agent.
  • Token cost tracking -- real-time token usage and estimated cost per session. /stats command. Supports 20+ model pricing tiers. Local models always show "free".
  • Prompt caching -- system prompt marked with cache_control for Anthropic/OpenAI. ~50% cost reduction on repeated prefixes.
  • Headless/CI mode -- godspeed run for non-interactive execution. Task from positional arg, --prompt-file, or stdin. --timeout N wall-clock cap. Differentiated exit codes (0 success, 1 tool error, 2 max iterations, 3 budget, 4 LLM error, 5 invalid input, 6 timeout, 130 interrupt) for pipeline orchestration. JSON output includes exit_reason, iterations_used, tool_calls, cost_usd, duration_seconds, audit_log_path. Audit trail is written by default.
  • Web tools -- web_search (DuckDuckGo, no API key) and web_fetch (HTML-to-text extraction) let the agent look up documentation and error messages.
  • Multi-language verify -- auto-verification after edits supports Python (ruff), JS/TS (biome/eslint), Go (go vet), Rust (cargo check), C/C++ (clang-tidy). Lint-fix retry loop up to 3 rounds.
  • Test runner -- auto-detect pytest, jest, vitest, go test, cargo test. Run targeted or full test suites. Agent-accessible for edit-test-fix loops.
  • Conversation export -- /export saves the full session as formatted markdown for sharing or review.
  • Session hooks -- run shell commands at lifecycle events (session_start, session_end, turn_end). Configure in ~/.godspeed/hooks.yaml for custom notifications, logging, or automation.
  • Skills -- markdown prompt files in ~/.godspeed/skills/ or .godspeed/skills/ that extend Godspeed's capabilities. Each skill is a self-contained prompt with optional trigger patterns.
  • Rich TUI -- syntax highlighting, unified diff rendering, streaming output, and slash commands via Rich and prompt-toolkit.

Training & Fine-Tuning

  • Conversation logger -- automatically persists every conversation message (user, assistant, tool calls, tool results, compaction summaries) to per-session JSONL at ~/.godspeed/training/. Captures the full conversation flow that the audit trail misses. Gated on log_conversations config (default: on).
  • Training data exporter -- godspeed export-training converts conversation logs to openai, chatml, or sharegpt fine-tuning formats. Filtering by tool count, success rate, turn count, and tool whitelist. Designed for Qwen/Mistral/Llama fine-tuning via Unsloth + TRL.
  • Per-step reward annotations -- automatic reward signals for GRPO/DPO: success (+1.0), verify passed (+0.5), dangerous command (-1.0), efficient tool sequence (+0.5). Session-level summarization for training pipeline integration.
  • Benchmark suite -- 20 hand-crafted tasks (easy/medium/hard) with Jaccard tool selection scoring and LCS sequence quality scoring for evaluating fine-tuned models against base models.
  • Enhanced tool descriptions -- all tools include inline usage examples and JSON Schema examples fields, improving both live agent performance and training data quality.

Benchmarks

SWE-Bench Lite — single-run results (dev-23, free-tier)

Same 23-instance dev subset across rows. All free-tier (NVIDIA NIM R&D), $0 API spend. Numbers are from sb-cli; report JSONs live in experiments/swebench_lite/reports/.

Godspeed Method Resolved Rate
v0.2.11 Qwen3.5-397B single-shot 6 / 23 26.1%
v0.2.12 Kimi K2.5 single-shot (driver swap) 8 / 23 34.8%
v0.3.1 Kimi K2.5 + agent-in-loop (single seed) 7 / 23 30.4%

Single-run performance is what you get when you run Godspeed once against a task. The agent-in-loop result (30.4%) underperformed the single-shot baseline (34.8%) in this early experiment — we publish this null result honestly rather than hiding it.

For context: published SOTA on full SWE-Bench Lite (April 2026) is 62.7%; top open-source agents with paid frontier drivers sit in the 40–50% band on the same benchmark.

Ensemble / research results

We also ran an oracle_best_of_5 ensemble (published ensemble methodology) combining five constituent runs. This is not what you get in a single run — it is a research ceiling showing what is possible with model diversity and post-hoc selection:

Method Resolved Rate
Oracle-selector best-of-5 (free-tier ensemble) 12 / 23 52.2%

Constituent runs: Kimi K2.5 single-shot, GPT-OSS-120B, Qwen3.5-397B iter1, Qwen3.5-397B seed3, Kimi K2.5 + agent-in-loop. oracle_merge.py picks per instance the patch from whichever run resolved, preferring shortest among resolvers. The agent-in-loop run contributed one unique resolve (marshmallow-code__marshmallow-1359) that no single-shot run landed.

Full methodology, per-instance resolution map, constituent-run numbers, null-result discussion, and limitations: experiments/swebench_lite/findings_2026_04_21.md.

Reproduce:

./experiments/swebench_lite/reproduce_v3_1.sh   # uses committed predictions + sb-cli

Prior release notes: findings_2026_04_20.md (v2.12.0 driver shootout), baseline_2026_04_19.md (v2.11.0 first honest result).

Internal 20-task suite — 2026-04-19 model shootout

Real numbers from the 20-task suite in benchmarks/tasks.jsonl, run against deterministic fixtures in benchmarks/fixtures/. Each fixture is isolated in a temp workspace per run; 13 of the 20 tasks have a verify.py hook that mechanically checks whether the agent actually completed the work.

Shootout (all NIM runs on the free R&D tier; local via Ollama):

Model Overall Pass (J>=0.6) Easy Medium Hard Mech
nvidia_nim/qwen/qwen3.5-397b-a17b 0.608 11/20 0.840 0.831 0.189 7/13
nvidia_nim/moonshotai/kimi-k2.5 0.548 9/20 0.840 0.727 0.135 6/13
nvidia_nim/mistralai/devstral-2-123b-instruct-2512 0.446 5/20 0.450 0.473 0.413 2/13
nvidia_nim/qwen/qwen3-coder-480b-a35b-instruct 0.333 5/20 0.870 0.138 0.174 2/13
ollama/qwen3-coder:latest (local) 0.107 1/20 0.150 0.125 0.057 1/13

Recommended production driver: nvidia_nim/qwen/qwen3.5-397b-a17b with ollama/qwen3-coder:latest as local fallback. Devstral-2 is the only contender that doesn't collapse on hard tasks (0.413 vs Qwen3.5's 0.189) — worth considering if your workload skews hard.

Full run outputs in experiments/bench_*/ and the aggregated table in experiments/benchmark_shootout_2026_04.md. Reproduce with scripts/run_benchmark.py --model <id>.

Architecture

flowchart LR
    User([User]) --> TUI["TUI\n(Rich + prompt-toolkit)"]
    TUI --> Loop["Agent Loop\n(loop.py)"]
    Loop --> LLM["LLM\n(LiteLLM + ModelRouter)"]
    LLM -->|streaming| Spec["Speculative\nDispatch"]
    Spec -->|READ_ONLY| Tools
    LLM -->|tool calls| Perm["Permission\nEngine"]
    Perm -->|allowed| Dispatch["Parallel/Serial\nDispatch"]
    Dispatch --> Tools["Tools\n(30+ built-in + MCP)"]
    Perm -->|denied| Deny[Deny + Log]
    Tools --> Audit["Audit Trail\n(SHA-256 JSONL)"]
    Deny --> Audit
    Audit --> Loop
    Loop -->|sub-agents| Loop
    Loop -->|response| TUI
    Loop --> Memory["Memory\n(SQLite)"]
    Audit -->|traces| Evo["Self-Evolution\n(Ollama, $0)"]
    Evo -->|improved descriptions| Tools
    Loop -->|messages| Train["Training Logger\n(JSONL)"]
    Train -->|export| FT["Fine-Tuning\n(openai/chatml/sharegpt)"]

    style Perm fill:#e74c3c,color:#fff
    style Audit fill:#2ecc71,color:#fff
    style Memory fill:#3498db,color:#fff
    style Spec fill:#9b59b6,color:#fff
    style Evo fill:#e67e22,color:#fff
    style Train fill:#1abc9c,color:#fff

How it works:

The agent loop is hand-rolled (no framework) following the same pattern proven by top-performing coding agents. The LLM decides when to stop. On each turn, the LLM either responds with text (done) or requests tool calls. During streaming, speculative dispatch starts READ_ONLY tool calls as background asyncio.Tasks before the full response completes — the main loop awaits cached results instead of re-dispatching. Every tool call passes through the permission engine before execution: deny rules are evaluated first and always win, then dangerous command detection (71 regex patterns) blocks destructive operations, then session grants and allow rules, and finally the tool's risk level determines the default. If anything is ambiguous, it fails closed. Permitted calls are split by risk level: READ_ONLY tools run in parallel via asyncio.gather(), write tools run sequentially. After execution, the tool call, its result, and the permission decision are recorded in the audit trail -- a hash-chained JSONL file where each record includes the SHA-256 hash of the previous record. Secrets are redacted at four layers: file access deny rules, context cleaning before the LLM sees content, output filtering on LLM responses, and audit log redaction. The loop also includes stuck-loop detection (3 identical errors triggers a replan), auto-verification (linter check after file edits in 6 languages with retry), auto-stash (git stash after 3+ consecutive writes), cost budget enforcement, and pause/resume for human-in-the-loop intervention. The self-evolution system mines audit trails for failure patterns and uses LLM-guided mutations to improve tool descriptions and prompts over time. The training logger captures the full conversation flow (user messages, assistant reasoning, tool results) to JSONL for fine-tuning tool-calling LLMs.

Key modules:

Module Path Purpose
Agent loop src/godspeed/agent/ Conversation management, LLM interaction, parallel + speculative dispatch, sub-agent coordinator
Security src/godspeed/security/ Permission engine, dangerous command detection, secret scanning
Audit src/godspeed/audit/ Hash-chained event logging, redaction, verification, compression
Tools src/godspeed/tools/ 30+ built-in tools with JSON schemas
LLM src/godspeed/llm/ LiteLLM client wrapper, model routing, token counting, cost tracking
Context src/godspeed/context/ Project instructions, compaction, checkpoints, repo map
MCP src/godspeed/mcp/ Model Context Protocol client (stdio + SSE) and tool adapter
Memory src/godspeed/memory/ SQLite-backed preferences, session events, correction tracking
Evolution src/godspeed/evolution/ Trace analysis, GEPA mutations, LLM-as-judge fitness, safety gate, registry
Training src/godspeed/training/ Conversation logger, fine-tuning exporter (openai/chatml/sharegpt), reward annotations, benchmark suite
Hooks src/godspeed/hooks/ Lifecycle event hooks (session_start, session_end, turn_end) for custom automation
Skills src/godspeed/skills/ Markdown-based skill system for extending agent capabilities
TUI src/godspeed/tui/ Terminal UI, rich output, permission prompts, slash commands

Getting Started

Install

pip install godspeed-coding-agent

Or with uv:

uv tool install godspeed-coding-agent     # installs globally — run 'godspeed' from anywhere

Setup

# One-time setup — creates ~/.godspeed/ and default settings
godspeed init

# Pull a free local model (default, no API key needed)
ollama pull qwen3:4b

Run

# Launch in any project directory — uses free local model by default
cd your-project/
godspeed

Or use a paid cloud model:

export ANTHROPIC_API_KEY="sk-..."
godspeed -m claude-sonnet-4-20250514

Switch models at any time with /model <name> inside the TUI, or run godspeed models to see all options.

Godspeed auto-upgrades ollama/ to ollama_chat/ for tool-capable models (Qwen, Llama, Gemma, Mistral, etc.).

Godspeed reads GODSPEED.md, AGENTS.md, CLAUDE.md, and .cursorrules from the project root for persistent instructions. Bring your existing config from any agent.

First session

$ godspeed
godspeed> Explain the authentication flow in this codebase

The agent will read your code, answer questions, write files, and run commands -- all gated by the permission engine.

Slash commands

Command Description
/help Show available commands
/model [name] Show or switch the active model
/clear Clear conversation history
/undo Undo last git commit (git reset --soft HEAD~1)
/audit Show audit trail stats and verify chain integrity
/permissions Show current permission rules and session grants
/remember [action] [pattern] Persist a permission rule (approve/deny/ask)
/extend [N] Set max iterations per turn (default: 50)
/context Show context window usage (tokens, percentage)
/plan Toggle plan mode (read-only, explore and plan only)
/architect Toggle architect mode (plan with read-only tools, then execute)
/think [budget] Toggle extended thinking for Claude models
/budget [amount] Show or set cost budget for the session
/autocommit Toggle auto-commit after successful edits
/evolve [cmd] Self-evolution: status, run, history, rollback, review
/checkpoint [name] Save conversation checkpoint, or list if no name
/restore <name> Restore a saved checkpoint
/pause Pause the agent loop at next iteration
/resume Resume a paused agent loop
/guidance <msg> Inject guidance and resume paused agent
/stats Show token usage and estimated cost
/export [name] Export conversation as markdown
/quit Exit Godspeed

Configuration

Project-level: GODSPEED.md

Drop a GODSPEED.md in your project root. The agent loads it as system context on every session. Use it for coding standards, architecture notes, or constraints. See GODSPEED.md.example for a template.

Global: ~/.godspeed/settings.yaml

model: claude-sonnet-4-20250514
fallback_models:
  - gpt-4o
  - gemini-2.0-flash

# Route different task types to different models
routing:
  plan: claude-sonnet-4-20250514
  edit: ollama/qwen3:4b
  chat: claude-sonnet-4-20250514

permissions:
  deny:
    - "FileRead(.env)"
    - "FileRead(*.pem)"
    - "FileRead(.ssh/*)"
  allow:
    - "shell(git *)"
    - "shell(ruff *)"
    - "shell(pytest *)"
    - "shell(make *)"
  ask:
    - "shell(*)"

audit:
  enabled: true
  retention_days: 30

memory_enabled: true

# MCP servers (optional)
mcp_servers:
  - name: github
    command: npx
    args: ["-y", "@modelcontextprotocol/server-github"]

Permission rules use glob-style matching against ToolName(argument) strings. Deny rules are additive across config levels -- a project config cannot weaken global denies.

Environment variables

Variable Purpose
ANTHROPIC_API_KEY Claude access
OPENAI_API_KEY GPT access
GEMINI_API_KEY Gemini access
GODSPEED_MODEL Override default model

Inspiration & Attribution

Godspeed stands on the shoulders of excellent prior work:

  • Agent loop pattern — Inspired by hand-rolled ReAct loops from open-source agent research and the proven design that the LLM decides when to stop. Permission gating at every tool call is an original extension.
  • Dangerous command detection — Inspired by Hermes Agent's Tirith security scanner. The regex-based approach to blocking destructive shell commands follows their design.
  • LiteLLM — Unified provider access via the LiteLLM library. Godspeed would not support 200+ providers without it.
  • Prompt-toolkit + Rich — The TUI is built on prompt-toolkit for input handling and Rich for output rendering.
  • SWE-Bench — Benchmark methodology and harness from SWE-bench. All published numbers use their evaluation protocol.
  • AGENTS.md / CLAUDE.md — Cross-agent config file idea from the Linux Foundation's AGENTS.md proposal and Anthropic's CLAUDE.md convention.

Security-first design, speculative tool dispatch, self-evolution, and multi-language verification are original contributions.

Development

# Clone and install
git clone https://github.com/omnipotence-eth/godspeed-coding-agent.git
cd godspeed
uv sync --all-extras

# Lint and format
ruff check . --fix && ruff format .

# Run tests
pytest --cov

# Verify audit trail integrity
godspeed audit verify

License

MIT


Built by Tremayne Timms

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

godspeed_coding_agent-0.4.0.tar.gz (815.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

godspeed_coding_agent-0.4.0-py3-none-any.whl (310.6 kB view details)

Uploaded Python 3

File details

Details for the file godspeed_coding_agent-0.4.0.tar.gz.

File metadata

  • Download URL: godspeed_coding_agent-0.4.0.tar.gz
  • Upload date:
  • Size: 815.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for godspeed_coding_agent-0.4.0.tar.gz
Algorithm Hash digest
SHA256 a8000a9670c62960ace9b557905cc72d6e422dab083d508d04da3c71d8c3dfde
MD5 cb4ec43c82720b07420b9718170e2d46
BLAKE2b-256 40d1d0d1841502e9213bd87eb4c49045dc03640e2a51dc7406a099cfe46d6a76

See more details on using hashes here.

Provenance

The following attestation bundles were made for godspeed_coding_agent-0.4.0.tar.gz:

Publisher: release.yml on omnipotence-eth/godspeed-coding-agent

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file godspeed_coding_agent-0.4.0-py3-none-any.whl.

File metadata

File hashes

Hashes for godspeed_coding_agent-0.4.0-py3-none-any.whl
Algorithm Hash digest
SHA256 a9e72acbe62e9281e778ce60e9da4d6d93329fcb8244f57a0d210ee7aa381868
MD5 e3cfb63a2f7efb8aa3800254d9d6fafe
BLAKE2b-256 c9c13555936be594d67e77f6278a89c0b083b0fa8f7f6942bc7cc99e6b6bcad2

See more details on using hashes here.

Provenance

The following attestation bundles were made for godspeed_coding_agent-0.4.0-py3-none-any.whl:

Publisher: release.yml on omnipotence-eth/godspeed-coding-agent

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page