Skip to main content

AI-Powered Infrastructure Copilot: The Self-Healing SRE.

Project description

ResponseIQ

CI GitHub Release PyPI License Checked with mypy Code style: black

"Don't just debug. Fix."

ResponseIQ is an AI-Native Self-Healing Infrastructure Copilot. Unlike traditional parsers that match regex strings, ResponseIQ reads your application logs, loads your actual source code into an LLM context, and generates surgical, context-aware remediation patches for incidents.


📸 See It In Action

ResponseIQ CLI Demo

Above: ResponseIQ scanning a crash log, reading the service.py file mentioned in the stack trace, and proposing a specific code patch.


✨ Key Features

  • 🧠 AI-Native Analysis: Uses Generic AI reasoning instead of fragile regex parsing rules.
  • 👁️ Context-Aware: Reads the local source files referenced in logs to understand why the crash happened.
  • ⚡ Self-Healing: Can generate Pull Requests or apply patches directly (CLI mode).
  • 🛡️ Battle-Tested: Includes "Sandbox Mode" to safely test remediation logic.

🚀 Quick Start (CLI Tool)

For developers who want to fix bugs in their local environment or CI pipeline.

1. Install

pip install responseiq

2. Configure Credentials

ResponseIQ requires access to an LLM provider to reason about your code.

Currently supported: OpenAI

export OPENAI_API_KEY="sk-..."

3. Usage Examples

Example A: Analyze Local Logs Scan a directory of log files and get a report of active incidents.

responseiq --mode scan --target ./var/log/app/

Example B: The "Magic Fix" Analyze logs AND the current source code to generate a patch.

# Finds errors in logs, locates the source file, and explains the fix
responseiq --mode fix --target ./logs/error.log

Example C: CI/CD Pipeline Integration Run ResponseIQ as a step in your GitHub Action to auto-triage build failures.

# In your workflow
responseiq --mode scan --target ./build_logs.txt >> summary.md

🏢 Platform Server (Self-Hosted)

For Platform Engineers who want to host a centralized incident response API (Webhook receiver for Datadog, PagerDuty, etc.).

Prerequisites

  • Docker & Docker Compose
  • OpenAI API Key configured in .env

Running with Docker

# 1. Start the API and Database
docker-compose up -d

# 2. The API is now available at http://localhost:8000
curl http://localhost:8000/health

Development Setup (Local)

We use UV for lightning-fast dependency management.

# Install dependencies
uv sync

# Run the API server with hot-reload
uv run uvicorn src.app:app --reload

🧪 Development & Contributing

Workflow

  1. Linting: make lint
  2. Testing: make test
  3. Format: make format

Project Structure

  • src/cli.py: Entry point for the CLI tool.
  • src/app.py: Entry point for the API Server.
  • src/services/remediation_service.py: The core "Brain" that interfaces with the LLM.

License

MIT


⚠️ Disclaimer & Liability

This tool uses Generative AI to suggest infrastructure and code fixes. By using ResponseIQ, you acknowledge that:

  1. AI Can Hallucinate: The suggestions provided may be syntactically correct but functionally wrong or insecure.
  2. Human Review is Mandatory: You must strictly review all Pull Requests or patches generated by this tool before deploying them.
  3. No Warranty: As per the MIT License, the authors assume no liability for system outages, data loss, or security vulnerabilities resulting from the use of this software.

For security reporting instructions, please see SECURITY.md.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

responseiq-2.2.0.tar.gz (64.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

responseiq-2.2.0-py3-none-any.whl (76.6 kB view details)

Uploaded Python 3

File details

Details for the file responseiq-2.2.0.tar.gz.

File metadata

  • Download URL: responseiq-2.2.0.tar.gz
  • Upload date:
  • Size: 64.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for responseiq-2.2.0.tar.gz
Algorithm Hash digest
SHA256 18e10083d627833c450db2c1f9394ab4f9668fd32a610c3be6205f49f519d90d
MD5 e27f14daaf33b5015a5409456d3bc0cb
BLAKE2b-256 c7d40d5f6f0cc1a8d5d00a36ab0e0d61fa379be63dfe5c498973609dded5af5b

See more details on using hashes here.

File details

Details for the file responseiq-2.2.0-py3-none-any.whl.

File metadata

  • Download URL: responseiq-2.2.0-py3-none-any.whl
  • Upload date:
  • Size: 76.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for responseiq-2.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 dc8ffe7976280c07057ad8c8d05ab6cb5f0ba240867c7d92ae6053059da15a5a
MD5 db7b4495ce90c8f241cfac3cf3784e8e
BLAKE2b-256 38118ac6afb3afd86cfa86afee9028eac85f67d3a2292147d3832dcb9f926e4f

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page