Skip to main content

Hybrid static + LLM codebase bug detector for multi-language projects.

Project description

testx

testx is a hybrid codebase auditing engine for teams that need deeper bug discovery than linting alone.
It combines deterministic static analysis with optional AI reasoning, then emits structured, actionable reports.

For a deep system breakdown, see ARCHITECTURE.md.

Table of Contents

Why testx

Traditional static tools often miss contextual or cross-language risks. testx is designed to:

  • detect high-signal bug/security/reliability risks quickly
  • operate fully offline (--ai none) for restricted environments
  • scale analysis depth with AI providers when allowed
  • track cost and avoid repeated spend through chunk-level caching
  • produce readable audits for engineers, leads, and reviewers

Core Capabilities

  • Hybrid analysis pipeline
    • Python AST rules for semantic checks
    • Generic analyzer for cross-language heuristics
    • Optional AI deep analysis over code chunks
  • One-library integration SDK
    • importable API for embedding audits into your app/CI
    • severity gating for release policies
    • extension and directory scan controls
  • Multi-language scanning
    • Python, JS/TS/React, Java, Go, Rust, C/C++, C#, Ruby, PHP, SQL, shell, YAML, JSON
  • Cost and throughput controls
    • rate limiter for AI calls
    • hard cap with --max-cost
    • SQLite cache keyed by provider/model/file/content hash
  • Audit-focused reporting
    • text, JSON, and HTML outputs
    • severity/type/source counts and risk hotspots
    • concrete fix guidance per issue

Architecture at a Glance

Execution pipeline:

  1. CLI parses command options and loads config.
  2. Scanner enumerates supported source files.
  3. Static analyzers execute per file.
  4. If AI is enabled, files are chunked and analyzed with provider client.
  5. Results are deduplicated and aggregated.
  6. Reporter renders final output.

Primary modules:

  • bugfinder/cli.py - command interface
  • bugfinder/config.py - environment and TOML config merge
  • bugfinder/scanner.py - file discovery + chunking
  • bugfinder/analyzer/ast_analyzer.py - Python semantic checks
  • bugfinder/analyzer/generic_analyzer.py - non-Python heuristics
  • bugfinder/analyzer/hybrid_analyzer.py - orchestration layer
  • bugfinder/ai/*_client.py - provider adapters
  • bugfinder/cache/cache_manager.py - SQLite cache
  • bugfinder/reporters.py - renderers
  • bugfinder/models.py - issue/report data models

Detailed architecture and data flow are documented in ARCHITECTURE.md.

Installation

Runtime install

pip install testx

Local development install

pip install -e ".[dev]"

Quick Start

1) Static-only audit (fastest and offline)

testx scan . --ai none --output text

1b) Programmatic integration (all in one library)

from bugfinder import AuditOptions, render_report, run_audit, should_fail_ci

report = run_audit(
    ".",
    AuditOptions(
        ai_provider="none",
        include_extensions={".py", ".ts", ".tsx"},
        exclude_dirs={"dist", "build"},
        min_severity="medium",
    ),
)

print(render_report(report, output="text"))
if should_fail_ci(report, "high"):
    raise SystemExit("High severity issues detected")

2) Deep audit with OpenAI

testx scan . \
  --ai openai \
  --model gpt-4o-mini \
  --max-cost 2.50 \
  --output html \
  --output-file audit.html

3) Deep audit with Claude

testx scan . \
  --ai claude \
  --model claude-3-5-sonnet-latest \
  --max-cost 2.50 \
  --output json \
  --output-file audit.json

Configuration

Configuration resolution order:

  1. Environment variables
  2. .bugfinder.toml (or --config path)

Supported environment variables:

  • OPENAI_API_KEY
  • ANTHROPIC_API_KEY

Example .bugfinder.toml:

[ai]
openai_api_key = "sk-..."
anthropic_api_key = "sk-ant-..."
default_provider = "none"
default_model = "gpt-4o-mini"
max_cost = 5.0
rate_limit_per_minute = 30

CLI Reference

Command:

testx scan <path> [options]

Options:

  • --ai: openai | claude | none
  • --model: provider model override
  • --max-cost: max estimated AI spend in USD
  • --output: text | json | html
  • --output-file: write report to file path
  • --config: explicit TOML config path
  • --cache-db: SQLite cache path (default .bugfinder_cache.sqlite3)
  • --exclude-dir: directory name to exclude (repeatable)
  • --include-ext: only analyze listed extensions (repeatable)
  • --min-severity: only include findings at or above threshold
  • --fail-on-severity: non-zero exit for CI policy enforcement
  • --fix: apply safe auto-fixes for supported issue types
  • --dry-run: preview safe fixes without editing files
  • --retest-command: post-fix validation command (default pytest -q)
  • --force: also apply medium-confidence fixes (default applies high only)

MCP Server Integration

testx ships an MCP stdio server so AI clients (Cursor, Claude Desktop, other MCP hosts) can call scan/fix tools directly.

Start command:

testx-mcp

Exposed MCP tools:

  • scan_codebase
    • runs the analyzer and returns report output (text/json/html)
  • fix_codebase
    • runs detect -> safe fix pipeline (supports dry_run and force)
    • returns applied/skipped/planned summary and remaining issue count
  • enterprise_audit
    • runs deep audit + dry-run fix simulation + risk scoring
    • returns prioritized remediation buckets (P0/P1/P2)
  • remediation_plan
    • builds a prioritized remediation plan from the latest audit in server memory

Enterprise MCP improvements:

  • full Content-Length framed stdio protocol support (not just newline JSON)
  • structured responses for dashboards, governance tools, and CI orchestrators
  • risk scoring model based on severity distribution for executive reporting
  • remediation planning output for security and platform teams

Example Cursor MCP config snippet:

{
  "mcpServers": {
    "testx": {
      "command": "testx-mcp"
    }
  }
}

Example Claude Desktop MCP config snippet:

{
  "mcpServers": {
    "testx": {
      "command": "testx-mcp"
    }
  }
}

Detect -> suggest -> apply safe fixes -> re-test -> human review

# Preview what would be auto-fixed
testx scan . --ai none --dry-run --output text

# Apply safe fixes, re-run tests, then print remaining issues
testx scan . --ai none --fix --retest-command "pytest -q" --output text

# Include medium-confidence fixes (for example inferred unused imports)
testx scan . --ai none --fix --force --retest-command "pytest -q" --output text

Safe auto-fix rules currently include:

  • remove console.log(...) / debugger; lines
  • remove Python print(...) debug statements
  • replace bare except: with except Exception:
  • remove inferred unused single-line imports (medium confidence, requires --force)
  • trim trailing whitespace and add missing newline at end of file

Output Model

Each issue includes:

  • type: bug | security | performance | code_smell
  • severity: low | medium | high
  • description: actionable risk statement
  • file and optional line
  • fix: concrete remediation guidance
  • source: static, ai:<provider>, or system

Report summary includes:

  • total issue count
  • severity/type/source distribution
  • top risky files by issue density
  • scan metadata (files_scanned, chunks_analyzed, provider/model, estimated cost)

How Analysis Works

Static analysis layer

  • AST analyzer (.py)

    • dynamic execution risks (eval, exec)
    • broad exception handling (except, except Exception)
    • mutable defaults
    • unreachable code patterns
    • unsafe subprocess usage (shell=True)
    • unsafe YAML loading patterns
    • runtime assert misuse warnings
  • Generic analyzer (other languages + text patterns)

    • TODO/FIXME/HACK risk markers
    • debug statements (console.log, debugger)
    • hardcoded secrets/tokens/password-like assignments
    • private key marker detection

AI analysis layer (optional)

  • file chunks are generated
  • prompt includes language + line range context
  • provider returns strict JSON
  • parsed issues are normalized into shared model
  • responses are cached by file content hash + provider + model

Merge and dedup

Issues are deduplicated on type, severity, description, file path, and line.

Operational Guidance

  • Start with --ai none in CI for fast deterministic checks.
  • Run AI scans on main branch nightly or pre-release for deeper coverage.
  • Use --max-cost aggressively in shared environments.
  • Persist report artifacts as CI artifacts, not committed source files.
  • Rotate API keys immediately if accidentally exposed.

Testing

Run unit tests:

pytest -q

Tests cover:

  • scanner chunking behavior
  • cache read/write semantics
  • issue deduplication
  • deep-audit regression rules (security detection + report summary structure)

Release / Publish

python -m pip install --upgrade build twine
python -m build
twine check dist/*
twine upload --repository testpypi dist/*
twine upload dist/*

Trusted Publishing with GitHub Actions (recommended)

testx includes .github/workflows/publish.yml for OIDC-based PyPI publishing.

PyPI Trusted Publisher settings:

  • Project name: testx
  • Owner: Najeebullah3124
  • Repository: testx
  • Workflow: publish.yml
  • Environment: pypi

Workflow behavior:

  • triggers automatically when a GitHub Release is published
  • also supports manual run via workflow_dispatch
  • builds wheel + sdist, then publishes via pypa/gh-action-pypi-publish

Security Practices

  • never commit secrets or API keys
  • prefer environment variables for credentials
  • keep AI optional in sensitive environments
  • use dedicated CI environments for publishing permissions
  • rotate credentials after any accidental exposure

Roadmap

  • incremental and diff-only scans
  • pluggable custom rule packs
  • SARIF output for security tooling integration
  • confidence scoring calibration and false-positive suppression

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

testx-1.0.4.tar.gz (30.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

testx-1.0.4-py3-none-any.whl (30.6 kB view details)

Uploaded Python 3

File details

Details for the file testx-1.0.4.tar.gz.

File metadata

  • Download URL: testx-1.0.4.tar.gz
  • Upload date:
  • Size: 30.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for testx-1.0.4.tar.gz
Algorithm Hash digest
SHA256 ce29e7d2fd40aeab9df4eff8ec410368a294e42e53e0e125cf85b4503cc5efaa
MD5 8d086e0b8c0080bf70df95d238991854
BLAKE2b-256 2f45144a909ec9b1a1d74ab70b4ee6a04084a4c29c749c79db2372a076bbbc73

See more details on using hashes here.

Provenance

The following attestation bundles were made for testx-1.0.4.tar.gz:

Publisher: publish.yml on Najeebullah3124/testx

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file testx-1.0.4-py3-none-any.whl.

File metadata

  • Download URL: testx-1.0.4-py3-none-any.whl
  • Upload date:
  • Size: 30.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for testx-1.0.4-py3-none-any.whl
Algorithm Hash digest
SHA256 1fbbcef38fd7d039c9f6ec1a33a54c2320e5dad06bbb6a6c0f81f6e6c11b9742
MD5 c6e01c3c610900d33955d1210113bfd8
BLAKE2b-256 b7ac75a617df1930e6e44d8ea5530d002a51e0bc40083fed18f594dc18f5085c

See more details on using hashes here.

Provenance

The following attestation bundles were made for testx-1.0.4-py3-none-any.whl:

Publisher: publish.yml on Najeebullah3124/testx

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page