Skip to main content

Automatic Python refactoring toolkit – detect, fix, and validate common code issues introduced by LLMs and humans alike.

Project description

prefact

Automatic Python refactoring toolkit — detect, fix, and validate common code issues introduced by LLMs and humans alike.

The Problem

When using LLMs for code generation, they often silently change import paths from absolute to deep relative:

# ❌ LLM introduces this
from ....llm.generator import generate_strategy
from ....loaders.yaml_loader import save_strategy_yaml

# ✅ You wanted this
from planfile.llm.generator import generate_strategy
from planfile.loaders.yaml_loader import save_strategy_yaml

prefact automatically detects, fixes, and validates such issues in a three-phase pipeline.

Features

Rule ID Auto-fix Description
Relative → Absolute imports relative-imports Converts from ....x import y to from pkg.x import y
Unused imports unused-imports Removes imports never referenced in the module
Duplicate imports duplicate-imports Removes the same name imported twice
Wildcard imports wildcard-imports 🔍 Flags from x import *
Unsorted imports sorted-imports 🔍 Flags import blocks not ordered stdlib→3rd-party→local
String concatenation string-concat 🔍 Flags "Hello " + name → suggests f-strings
Print statements print-statements 🔍 Flags debug print() calls
Missing return types missing-return-type 🔍 Flags public functions without return type hints

✅ = auto-fix · 🔍 = scan-only (report)

Installation

pip install -e .

# with dev dependencies (pytest)
pip install -e ".[dev]"

Quick Start

# Generate config file
prefact init

# List all available rules
prefact rules

# Scan only (no changes)
prefact scan --path ./my_project --package mypackage

# Fix + validate (with backups)
prefact fix --path ./my_project --package mypackage

# Dry-run (show what would change)
prefact fix --path ./my_project --package mypackage --dry-run

# Check a single file
prefact check ./my_project/src/mypackage/core/service.py --package mypackage

# JSON output for CI
prefact fix --path . --format json -o report.json

Pipeline Architecture

┌─────────┐      ┌─────────┐      ┌────────────┐
│  SCAN   │ ──→  │   FIX   │ ──→  │  VALIDATE  │
│         │      │         │      │            │
│ Detect  │      │ Apply   │      │ Syntax OK? │
│ issues  │      │ fixes   │      │ Regressions│
│ per rule│      │ + backup│      │ preserved? │
└─────────┘      └─────────┘      └────────────┘
  1. Scan — each rule walks the AST / CST and emits Issue objects
  2. Fix — rules with auto-fix transform the source (via libcst for formatting-safe changes)
  3. Validate — post-fix checks: syntax valid, no regressions, import counts preserved

Configuration

Create prefact.yaml (auto-generated via prefact init):

package_name: planfile

include:
  - "**/*.py"

exclude:
  - "**/venv/**"
  - "**/build/**"

rules:
  relative-imports:
    enabled: true
    severity: warning
  unused-imports:
    enabled: true
    severity: info
  duplicate-imports:
    enabled: true
  wildcard-imports:
    enabled: true
    severity: error
  sorted-imports:
    enabled: false
  string-concat:
    enabled: true
  print-statements:
    enabled: true
    options:
      ignore_patterns: ["cli.py", "scripts/"]
  missing-return-type:
    enabled: false

Python API

from pathlib import Path
from prefact.config import Config
from prefact.engine import RefactoringEngine

config = Config(
    project_root=Path("./my_project"),
    package_name="planfile",
    dry_run=False,
    backup=True,
)

engine = RefactoringEngine(config)
result = engine.run()

print(f"Found {result.total_issues} issues")
print(f"Fixed {result.total_fixed}")
print(f"All valid: {result.all_valid}")

Writing Custom Rules

Extend BaseRule and use the @register decorator:

from prefact.rules import BaseRule, register
from prefact.models import Issue, Fix, ValidationResult

@register
class MyCustomRule(BaseRule):
    rule_id = "my-custom-rule"
    description = "Does something useful."

    def scan_file(self, path, source):
        # Return list[Issue]
        ...

    def fix(self, path, source, issues):
        # Return (fixed_source, list[Fix])
        ...

    def validate(self, path, original, fixed):
        # Return ValidationResult
        ...

CI/CD Integration

# GitHub Actions
- name: prefact check
  run: |
    pip install ./prefact
    prefact scan --path . --format json -o prefact-report.json
    prefact fix --path . --dry-run

Running Tests

pip install -e ".[dev]"
pytest -v

License

Apache License 2.0 - see LICENSE for details.

Author

Created by Tom Sapletta - tom@sapletta.com

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

prefact-0.1.1.tar.gz (25.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

prefact-0.1.1-py3-none-any.whl (28.3 kB view details)

Uploaded Python 3

File details

Details for the file prefact-0.1.1.tar.gz.

File metadata

  • Download URL: prefact-0.1.1.tar.gz
  • Upload date:
  • Size: 25.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.7

File hashes

Hashes for prefact-0.1.1.tar.gz
Algorithm Hash digest
SHA256 b12f7812031bbe5e6d3dbd29203d3ea51d7bb81028f07f0d59d63593d9875dbd
MD5 2c8a4741ce26c9c31697af177283e944
BLAKE2b-256 6bdc000b66f751cd0c0c8a53344046e593574cfac75eeceb929001db47de912c

See more details on using hashes here.

File details

Details for the file prefact-0.1.1-py3-none-any.whl.

File metadata

  • Download URL: prefact-0.1.1-py3-none-any.whl
  • Upload date:
  • Size: 28.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.7

File hashes

Hashes for prefact-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 0e3e05fda9011878535698496478569a236930ec5f0b82ba56fe614a20936567
MD5 525d677fc6096f108251bf9eb94c04de
BLAKE2b-256 fe8a7883175a6be711f3d39653419ecdd88d7a8197131855751dedaea7f1af3b

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page