Skip to main content

A high-performance, security-focused static analysis tool for Python, powered by Rust.

Project description

image

High-Performance Python/Rust Graph-Based SAST Framework

POWERED BY Total PyPI Downloads PyPI Downloads latest release PyPI version Python version Rust version

PySpector is a State-of-the-Art Static Analysis Security Testing (SAST) framework, built in Rust for next-gen performances, made for modern Python projects and large codebases. Unlike traditional linters, PySpector utilizes a Flow-Sensitive, Inter-Procedural Taint Engine to track untrusted data across complex function boundaries and control flow structures.

By compiling the core analysis engine to a native binary, PySpector avoids the performance limitations of traditional Python-only tools. This makes it well-suited for CI/CD pipelines and local development environments where speed and scalability matter.

PySpector is designed to be both comprehensive and intuitive, offering a multi-layered analysis approach that goes beyond simple pattern matching to understand the structure and data flow of your Python application.

Table of Contents

Getting Started

Prerequisites

  • Python: Python 3.9 – 3.12 supported (Python 3.9 or newer, up to 3.12).
  • Rust: The Rust compiler (rustc) and Cargo package manager are required. You can easily install the Rust toolchain via rustup and verify your installation by running cargo --version.

Installation

It is highly recommended to install PySpector in a dedicated Python 3.12 venv.

Create a Virtual Environment:

  • Linux (Bash):

    # Download Python 3.12
    # Download Python 3.12
    python3.12 -m venv venv
    source venv/bin/activate
    
  • Windows (PowerShell):

    # Download Python 3.12 from the Microsoft Store and run:
    python3.12 -m venv venv
    .\venv\Scripts\Activate.ps1
    # or, depending on the Python 3.12 installation source:
    .\venv\bin\Activate.ps1
    

With PySpector now officially on PyPI🎉, installation is as simple as running:

pip install pyspector

Key Features

  • Flow-Sensitive Analysis: Utilizes a Control Flow Graph (CFG) to track variable states sequentially, accurately distinguishing between safe and vulnerable code paths.

  • Inter-Procedural Taint Tracking: Propagates untrusted data across function boundaries using global fixed-point iteration and function summaries.

  • Context-Aware Summaries: Sophisticated mapping of which function parameters flow to return values, allowing for high-precision tracking through complex utility functions.

  • Multi-Engine Hybrid Scanning:

    • Regex Engine: High-speed scanning for secrets, hardcoded credentials, and configuration errors.

    • AST Engine: Deep structural pattern matching to find Python-specific anti-patterns.

    • Graph Engine: Advanced CFG and Call-Graph-based data flow analysis for complex vulnerability chains.

  • Fastest Market Performances: Core analysis engine implemented in Rust with Rayon for multi-threaded parallelization (allowing PySpector to scan 71% faster than Bandit, and 16.6x faster than Semgrep).

  • AI-Agent Security: Specialized rulesets designed to identify prompt injection, insecure tool use, and data leakage in LLM-integrated Python applications.

Core Engine Architecture

PySpector v0.1.5 represents a shift from partially-static pattern matching, to a full graph-based analysis engine:

  1. AST Parsing: Python source is converted into a structured JSON AST, for semantic analysis.
  2. Call Graph Construction: PySpector builds a project-wide map of function definitions, and call sites to enable cross-file analysis.
  3. CFG Generation: Each function is decomposed into a Control Flow Graph (CFG), allowing the engine to understand the order of operations and conditional Python logic.
  4. Fixed-Point Taint Propagation: Using a Worklist Algorithm, the engine propagates "taint" from defined Sources to Sinks, while respecting Sanitizers that clean the data along the way.

How It Works

PySpector's hybrid architecture is key to its performance and effectiveness.

  • Python CLI Orchestration: The process begins with the Python-based CLI. It handles command-line arguments, loads the configuration and rules, and prepares the target files for analysis. For each Python file, it uses the native ast module to generate an Abstract Syntax Tree, which is then serialized to JSON.

  • Invocation of the Rust Core: The serialized ASTs, along with the ruleset and configuration, are passed to the compiled Rust core. The handoff from Python to Rust is managed by the pyo3 library.

  • Parallel Analysis in Rust: The Rust engine takes over and performs the heavy lifting. It leverages the rayon crate to execute file scans and analysis in parallel, maximizing the use of available CPU cores. It builds a complete call graph of the application to understand inter-file function calls, which is essential for the taint analysis module.

  • Results and Reporting: Once the analysis is complete, the Rust core returns a structured list of findings to the Python CLI. The Python wrapper then handles the final steps of filtering the results based on the severity threshold and the baseline file, and generating the report in the user-specified format.

This architecture combines the best of both worlds: a flexible, user-friendly interface in Python and a high-performance, memory-safe analysis engine in Rust :)

Performance Benchmarks

Performance benchmarks demonstrate PySpector's competitive advantages in SAST scanning speed while maintaining comprehensive security analysis.

Performance benchmarks were executed in a deterministic and controlled environment using automated stress-testing scripts, ensuring repeatable and unbiased measurements

Benchmark Results

speed_benchmark_charts

Comparative analysis across major Python codebases (Django, Flask, Pandas, Scikit-learn, Requests) shows:

Metric PySpector Bandit Semgrep
Throughput 25,607 lines/sec 14,927 lines/sec 1,538 lines/sec
Performance Advantage 71% faster than Bandit Baseline 16.6x slower
Memory Usage 1.4 GB average 111 MB average 277 MB average
CPU Utilization 120% (multi-core) 100% (single-core) 40%

Key Performance Characteristics

  • Speed: Delivers 71% faster scanning than traditional tools through Rust-powered parallel analysis
  • Scalability: Maintains high throughput on large codebases (500k+ lines of code)
  • Resource Profile: Optimized for modern multi-core environments with adequate memory allocation
  • Consistency: Stable performance across different project types and sizes

System Requirements for Optimal Performance

  • Minimum: 2 CPU cores, 2 GB RAM
  • Recommended: 4+ CPU cores, 4+ GB RAM for large codebases
  • Storage: SSD recommended for large repository scanning

Benchmark Methodology

Performance testing conducted on:

  • Test Environment: Debian-based Linux VM (2 cores, 4GB RAM)
  • Test Projects: 5 major Python repositories (13k-530k lines of code)
  • Measurement: Average of multiple runs with CPU settling periods
  • Comparison: Head-to-head against Bandit and Semgrep using identical configurations

Benchmark data available in the project repository for transparency and reproducibility.

Usage

PySpector is operated through a straightforward command-line interface.

Running a Scan

The primary command is scan, which can target a local file, a directory, or even a remote Git repository.

pyspector scan [PATH or --url REPO_URL] [OPTIONS]

Examples:

  • Scan a single file
pyspector scan /path/to/your/project
  • Scan a local directory and save the report as HTML:
pyspector scan /path/to/your/project -o report.html -f html
  • Scan a public GitHub repository:
pyspector scan --url https://github.com/username/repo.git

Wizard Mode for Beginners (NEW FEATURE🚀)

image
  • Use the --wizard flag to enter the guided scan mode, perfect for 1st time users and beginners or students:
pyspector scan --wizard

Scan for AI and LLM Vulnerabilities

image
  • Use the --ai flag to enable a specialized ruleset, for projects using Large Language Models:
pyspector scan /path/to/your/project --ai

Plugin System

image PySpector ships with an extensible plugin architecture that lets you post-process findings, generate custom artefacts, or orchestrate follow-up actions after every scan. Plugins run in-process once the Rust core returns the final issue list, so they see exactly the same normalized data that drives the built-in reports.

Lifecycle Overview

  1. Discovery - Plugin files live in the repository's plugins directory (PySpector/plugins) and are discovered automatically.
  2. Registration - Trusted plugins are recorded in PySpector/plugins/plugin_registry.json together with their checksum and metadata.
  3. Validation - Before execution PySpector validates plugin configuration, statically inspects the source for dangerous APIs, and checks the on-disk checksum.
  4. Execution - The plugin is initialized, receives the full findings list, and can emit additional files or data. cleanup() is always called at the end.

Managing Plugins from the CLI

The CLI exposes helper commands for maintaining your local catalogue:

pyspector plugin list               # Show discovered plugins, trust status, version, author
pyspector plugin trust plugin_name     # Validate, checksum, and mark a plugin as trusted
pyspector plugin info plugin_name     # Display stored metadata and checksum verification
pyspector plugin install path/to/plugin.py --trust
pyspector plugin remove legacy_plugin

Only trusted plugins are executed automatically. When you trust a plugin PySpector calculates its SHA256 checksum and stores the version, author, and description that the plugin declares via PluginMetadata. If the file is modified later you will be warned before it runs again. To trust a plugin:

pyspector plugin install ./PySpector/plugins/aipocgen.py --trust

Running Plugins During a Scan

Use one or more --plugin flags during pyspector scan and provide a JSON configuration file if the plugin expects custom settings:

pyspector scan vulnerableapp.py --plugin aipocgen --plugin-config ./PySpector/pluginconfig/aipocgen.json

The configuration file must be a JSON object whose keys match plugin names, for example:

{
  "aipocgen": {
    "api_key": "YOUR-GROQ-KEY",
    "model": "llama-3.3-70b",
    "severity_filter": ["HIGH", "CRITICAL"],
    "max_pocs": 5,
    "output_dir": "pocs",
    "dry_run": false
  }
}

Each plugin receives only its own configuration block. Results are printed in the CLI, and any paths returned in the output_files list are shown under “Generated files”.

Authoring a Plugin

Create a new Python file in ~/.pyspector/plugins/<name>.py and subclass PySpectorPlugin:

from pathlib import Path
from typing import Any, Dict, List

from pyspector.plugin_system import PySpectorPlugin, PluginMetadata


class MyPlugin(PySpectorPlugin):
    @property
    def metadata(self) -> PluginMetadata:
        return PluginMetadata(
            name="my_plugin",
            version="0.1.0",
            author="Your Name",
            description="Summarises HIGH severity findings",
            category="reporting",
        )

    def validate_config(self, config: Dict[str, Any]) -> tuple[bool, str]:
        if "output_file" not in config:
            return False, "output_file is required"
        return True, ""

    def initialize(self, config: Dict[str, Any]) -> bool:
        self.output = Path(config["output_file"]).resolve()
        return True

    def process_findings(
        self,
        findings: List[Dict[str, Any]],
        scan_path: Path,
        **kwargs,
    ) -> Dict[str, Any]:
        highs = [f for f in findings if f.get("severity") == "HIGH"]
        self.output.write_text(f"{len(highs)} HIGH findings\n", encoding="utf-8")
        return {
            "success": True,
            "message": f"Summarised {len(highs)} HIGH findings",
            "output_files": [str(self.output)],
        }

Your plugin must implement the following:

  • metadata – Return a PluginMetadata instance describing the plugin.
  • validate_config(config) (optional but recommended) – Abort gracefully when required settings are missing by returning (False, "reason").
  • initialize(config) – Prepare state or dependencies; return False to skip execution.
  • process_findings(findings, scan_path, **kwargs) – Receive every finding as a dictionary and return a result object containing:
    • success: boolean status
    • message: short summary for the CLI
    • data: optional serializable payload
    • output_files: optional list of generated file paths
  • cleanup() (optional) – Release resources; called even if an exception occurs.

Tip: Plugins are plain Python modules, so you can run python my_plugin.py while developing to perform quick checks before trusting them through the CLI.

Configuration Tips and Best Practices

  • Store API keys or long-lived secrets in environment variables and read them during initialize. Provide helpful error messages when credentials are missing.
  • Keep side-effects inside the scan directory. When PySpector scans a single file scan_path is that file, so the reference plugins switch to scan_path.parent before writing outputs.
  • Validate configuration early using validate_config; PySpector surfaces the error message in the CLI without executing the plugin.
  • Return meaningful message values and populate output_files so automation can pick up generated artifacts.
  • Document optional switches such as dry_run (see the bundled aipocgen plugin for an example) to support air-gapped testing.

Security Model

The plugin manager enforces several safeguards:

  • AST-based static inspection blocks dangerous constructs (eval, exec, subprocess.*, etc.) and prints warnings when sensitive but acceptable calls (e.g., open) are used.
  • Trust workflow – you must explicitly trust a plugin before it can run; the CLI informs you about any warnings produced during validation.
  • Checksum verification – each trusted plugin has a stored SHA256 hash; changes are flagged before execution.
  • Argument isolation – the runner resets sys.argv to a minimal value so Click-based plugins cannot consume the parent CLI arguments accidentally.
  • Structured error handling – exceptions are caught, traced, and reported without aborting the main scan, and cleanup() still runs.

Together these measures let you extend PySpector confidently while maintaining a secure supply chain for third-party automation.

Triaging and Baselining Findings

image

PySpector includes an interactive triage mode to help manage and baseline findings. This allows you to review issues and mark them as "ignored" so they don't appear in future scans.

  • Generate a JSON report:
pyspector scan /path/to/your/project -o report.json -f json
  • Start the triage TUI:
pyspector triage report.json

Inside the TUI, you can navigate with the arrow keys, press i to toggle the "ignored" status of an issue, and s to save your changes to a .pyspector_baseline.json file. This baseline file will be automatically loaded on subsequent scans.

Automation and Integration

PySpector includes Shell helper scripts to integrate security scanning directly into your development and operational workflows.

Git Pre-Commit Hook

To ensure that no new high-severity issues are introduced into the codebase, you can set up a Git pre-commit hook. This hook will automatically scan staged Python files before each commit and block the commit if any HIGH or CRITICAL issues are found.

To set up the hook, run the following script from the root of your Git repository:

./scripts/setup_hooks.sh

This script creates an executable .git/hooks/pre-commit file that performs the check. You can bypass the hook for a specific commit by using the --no-verify flag with your git commit command.

Scheduled Scans with Cron

For continuous monitoring, you can schedule regular scans of your projects using a cron job. PySpector provides an interactive script to help you generate the correct crontab entry.

To generate your cron job command, run:

./scripts/setup_cron.sh

The script will prompt you for the project path, desired scan frequency (daily, weekly, monthly), and a location to store the JSON reports. It will then output the command to add to your crontab, automating your security scanning and reporting process.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pyspector-0.1.5.tar.gz (69.3 kB view details)

Uploaded Source

Built Distributions

If you're not sure about the file name format, learn more about wheel file names.

pyspector-0.1.5-cp312-cp312-win_amd64.whl (1.2 MB view details)

Uploaded CPython 3.12Windows x86-64

pyspector-0.1.5-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (1.4 MB view details)

Uploaded CPython 3.12manylinux: glibc 2.17+ x86-64

pyspector-0.1.5-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (1.4 MB view details)

Uploaded CPython 3.11manylinux: glibc 2.17+ x86-64

pyspector-0.1.5-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (1.4 MB view details)

Uploaded CPython 3.10manylinux: glibc 2.17+ x86-64

pyspector-0.1.5-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (1.4 MB view details)

Uploaded CPython 3.9manylinux: glibc 2.17+ x86-64

File details

Details for the file pyspector-0.1.5.tar.gz.

File metadata

  • Download URL: pyspector-0.1.5.tar.gz
  • Upload date:
  • Size: 69.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.6

File hashes

Hashes for pyspector-0.1.5.tar.gz
Algorithm Hash digest
SHA256 9795527b26cd54b59174db28f85b67de4f4aee025c2cce9004744a7565f46461
MD5 a023a1b4e158217a337469438be0b9b7
BLAKE2b-256 0852c65b326f553da3556f35dc503a33d5740fd12950c6ff64a5ecad7bc8d270

See more details on using hashes here.

File details

Details for the file pyspector-0.1.5-cp312-cp312-win_amd64.whl.

File metadata

  • Download URL: pyspector-0.1.5-cp312-cp312-win_amd64.whl
  • Upload date:
  • Size: 1.2 MB
  • Tags: CPython 3.12, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.6

File hashes

Hashes for pyspector-0.1.5-cp312-cp312-win_amd64.whl
Algorithm Hash digest
SHA256 e6b1d4dbc80151e47fa94e264ae5a70729ac8d1e08ae7bb0dba05fb315511e79
MD5 049f86a9aeb8df5f50e59446312c33c5
BLAKE2b-256 f456bf240b197000405d3248cfe3d32b9ef2dbf8c9a0b48a9082d2612524e6f9

See more details on using hashes here.

File details

Details for the file pyspector-0.1.5-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.

File metadata

File hashes

Hashes for pyspector-0.1.5-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
Algorithm Hash digest
SHA256 821cd701977e1e5d7b6ec63d838b73b890699611580d462e5af90da4ab087665
MD5 a4f986aa95e49a5786007576fbbee8f0
BLAKE2b-256 7a68fef1f589ea74dd784273b592e4bb645b97f00902f62b08b1507a00b5272a

See more details on using hashes here.

File details

Details for the file pyspector-0.1.5-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.

File metadata

File hashes

Hashes for pyspector-0.1.5-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
Algorithm Hash digest
SHA256 9d9148f685c0bba796ff2fa83e2b369c954bee6e54430a2432482eb3fb86a383
MD5 95d5fb0de0df3d4d5ffaf214768b034d
BLAKE2b-256 6c7e4d0909783d7196ea7896c56a413cf828383910f3021685c3c9426c92b0ec

See more details on using hashes here.

File details

Details for the file pyspector-0.1.5-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.

File metadata

File hashes

Hashes for pyspector-0.1.5-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
Algorithm Hash digest
SHA256 f068bb42adb77db3ecbe40b87430cff02c0e86e1355a8a16fee10503c141e395
MD5 5fed27f395953d80716f3c09ba50ff8f
BLAKE2b-256 c9eb214367fcc1b8d2a8eb8c87c8a9a493dfbefd1f4490e04f974bf33b9b3beb

See more details on using hashes here.

File details

Details for the file pyspector-0.1.5-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.

File metadata

File hashes

Hashes for pyspector-0.1.5-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
Algorithm Hash digest
SHA256 e04443acbb04ad6ced10e3b284bf9dc90ad833fcd4fdaddd242b0c4fe5a6d43b
MD5 2fc186441aca31741d69d4a5a21bd038
BLAKE2b-256 6817881cd52c3329b34d2e3ee1dadb7ecc2cd4b42ab3bb1067bdf55ce84b2c86

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page