Skip to main content

OCR-driven anonymization pipeline for medical reports and endoscopy frames

Project description

LX Anonymizer

LX Anonymizer is a comprehensive toolkit for de-identifying endoscopy frames and medical reports. It combines advanced OCR pipelines, spaCy-based NER, heuristic sanitizers, and report-specific rules to redact or pseudonymize sensitive information while preserving clinical context.

Core Components

ReportReader

Specialized for medical report anonymization with support for:

  • Multi-format processing: PDFs and images with automatic OCR fallback
  • Advanced metadata extraction: LLM-powered extraction using DeepSeek, MedLLaMA, or Llama3
  • Ensemble OCR: Combines Tesseract and TrOCR for improved accuracy
  • PDF anonymization: Creates blackened PDFs with sensitive regions automatically masked
  • Batch processing: Handles multiple reports with comprehensive error handling

FrameCleaner

Designed for real-time video frame anonymization featuring:

  • Hardware-accelerated processing: NVIDIA NVENC support with CPU fallback
  • Streaming video processing: Processes videos without full re-encoding when possible
  • Adaptive frame sampling: Optimizes performance for long videos (>10,000 frames)
  • Multiple anonymization strategies: Frame removal or mask overlay techniques
  • ROI-based masking: Device-specific region masking for endoscopic equipment

Default Return Format

LX Anonymizer will return a sensitive meta compliant dict when running either of the main client functions above.

Highlights

  • End-to-end anonymization of PDFs and video sequences using OCR, NER, and pseudonymization helpers.
  • Modular pipeline that lets you choose between Tesseract, TrOCR, ensemble OCR, and multiple metadata extractors.
  • Hardware optimization with NVENC acceleration for real-time video processing and streaming capabilities.
  • Human-in-the-loop ready outputs: original/anonymized text side by side, metadata JSON, and validation artefacts.
  • Extensible ruleset covering device-specific renderers, fuzzy name matching, and language-specific replacements.

Requirements

  • Python 3.12+
  • Linux or macOS (Windows support is experimental)
  • NVIDIA GPU recommended for real-time video anonymization (CUDA 12.x). CPU-only processing works but is slower.
  • Optional extras:
    • spaCy de_core_news_lg model (download after installation)
    • Torch vision/audio for video OCR workloads
    • Ollama-compatible LLMs for advanced metadata extraction

Installation

From PyPI (upcoming release)

pip install lx-anonymizer

Install extras to tailor the footprint:

pip install "lx-anonymizer[gpu,ocr,llm,dev]"

From source

git clone https://github.com/wg-lux/lx-anonymizer.git
cd lx-anonymizer
uv sync

Nix development shell

direnv allow
nix develop

This loads GPU, OCR, and tooling dependencies declared in devenv.nix.

Model downloads

After installation, fetch the German spaCy model:

python -m spacy download de_core_news_lg

First CLI runs also download OCR checkpoints (EAST, TrOCR, etc.). For air-gapped deployments, grab the archives listed in lx_anonymizer/settings.py and place them in ~/.cache/lx-anonymizer.

Quickstart

CLI Usage

Report Processing

# Process a single medical report
python -m cli.report_reader process report.pdf --ensemble --output-dir ./anonymized

# Use LLM for enhanced metadata extraction
python -m cli.report_reader process report.pdf --llm-extractor deepseek --use-ocr

# Batch process multiple reports
python -m cli.report_reader batch /path/to/reports/ --output-dir ./output --max-files 10

Video Frame Cleaning

# Clean a single video file
python -m lx_anonymizer.cli.frame_cleaner_cli clean video.mp4 --output-dir ./cleaned

# Batch clean multiple videos
python -m lx_anonymizer.cli.frame_cleaner_cli batch /path/to/videos/ --output-dir ./output

Useful CLI options:

  • --llm-extractor {deepseek,medllama,llama3} for LLM-powered metadata extraction.
  • --use-ocr and --ensemble to switch OCR strategies.
  • batch and extract sub-commands for folder processing or metadata-only runs.

Python API

ReportReader API

from lx_anonymizer import ReportReader

# Basic usage
reader = ReportReader(locale="de_DE")
original, anonymized, meta, pdf_path = reader.process_report(
    pdf_path="/path/to/report.pdf",
    use_ensemble=True,
    use_llm_extractor="deepseek",
)

# Create anonymized PDF with blackened sensitive regions
original, anonymized, meta, anonymized_pdf = reader.process_report(
    pdf_path="/path/to/report.pdf",
    create_anonymized_pdf=True,
    anonymized_pdf_output_path="/path/to/output.pdf"
)

# Advanced processing with region cropping
original, anonymized, meta, cropped_regions, pdf_path = reader.process_report_with_cropping(
    pdf_path="/path/to/report.pdf",
    crop_output_dir="/path/to/cropped_regions",
    crop_sensitive_regions=True,
    use_llm_extractor="deepseek"
)

FrameCleaner API

from lx_anonymizer.frame_cleaner import FrameCleaner
from pathlib import Path

# Initialize with hardware acceleration
cleaner = FrameCleaner(use_llm=True)

# Clean video with mask overlay (preserves all frames)
cleaned_path, metadata = cleaner.clean_video(
    video_path=Path("endoscopy.mp4"),
    endoscope_image_roi={"x": 550, "y": 0, "width": 1350, "height": 1080},
    endoscope_data_roi_nested={"patient_info": {"x": 10, "y": 10, "width": 300, "height": 50}},
    technique="mask_overlay"
)

# Remove sensitive frames entirely
cleaned_path, metadata = cleaner.clean_video(
    video_path=Path("endoscopy.mp4"),
    endoscope_image_roi=roi_config,
    endoscope_data_roi_nested=data_roi_config,
    technique="remove_frames"
)

See tests/test_cli_integration.py and tests/test_frame_cleaner.py for more examples.

Advanced Features

ReportReader Capabilities

  • Intelligent OCR Fallback: Automatically switches to OCR when PDF text extraction yields poor results
  • Multi-LLM Support: DeepSeek, MedLLaMA, and Llama3 integration for enhanced medical entity extraction
  • Ensemble OCR: Combines multiple OCR engines (Tesseract + TrOCR) for improved accuracy
  • PDF Anonymization: Creates masked PDFs with sensitive regions automatically blackened
  • Batch Processing: Processes multiple reports with error recovery and progress tracking
  • Metadata Validation: Cross-validates extracted information using multiple extraction methods

FrameCleaner Capabilities

  • Adaptive Sampling: Automatically samples frames for long videos (>10,000 frames) to optimize performance
  • Hardware Acceleration: NVIDIA NVENC support with automatic CPU fallback for unsupported systems
  • Streaming Processing: Uses FFmpeg streaming and named pipes to minimize memory usage and processing time
  • ROI-based Processing: Device-specific region configurations for endoscopic equipment (Olympus CV-1500, etc.)
  • Multiple Anonymization Strategies:
    • Mask Overlay: Blacks out sensitive regions while preserving video timeline
    • Frame Removal: Completely removes sensitive frames from the video stream
  • Quality Optimization: Automatic pixel format conversion and codec selection for minimal quality loss

Performance Optimizations

  • Stream Copy Operations: Avoids re-encoding when possible, using FFmpeg's -c copy for maximum speed
  • Named Pipe Support: In-memory video streaming for frame removal operations
  • Batch Metadata Extraction: Processes multiple frames simultaneously for improved efficiency
  • Hardware Detection: Automatically detects and uses available hardware acceleration (NVENC, QuickSync)

Data directories

By default, outputs live in ~/etc/lx-anonymizer/{data,temp}. Adjust them in lx_anonymizer/directory_setup.py. Clean temp regularly to avoid large intermediate artefacts.

Development workflow

  • Code quality: uv run flake8 for linting and formatting
  • Testing:
    • CPU-friendly tests: uv run pytest -m "not gpu"
    • GPU-accelerated tests: uv run pytest -m gpu (requires CUDA-capable hardware)
    • Integration tests: uv run pytest tests/test_cli_integration.py
    • Frame processing tests: uv run pytest tests/test_frame_cleaner.py
  • Performance profiling: Use --log-level DEBUG for detailed timing information
  • Build: uv run python -m build for wheel creation
  • Full validation: scripts/run_checks.sh for comprehensive local testing

Testing Medical Workflows

  • ReportReader: Test with sample medical PDFs in German and English
  • FrameCleaner: Validate with endoscopic video files (MP4, AVI formats supported)
  • Integration: Use example_anonymize_pdf.py for end-to-end testing scenarios

Project roadmap

  1. Release Management:
    • Publish CPU-only wheel to TestPyPI
    • Add optional extras for GPU/LLM workloads and slim default install
    • Automate release workflow (wheel + sdist upload, GitHub release notes)
  2. API Enhancement:
    • Expose REST/gRPC service with validation UI
    • WebSocket support for real-time video processing
    • Enhanced batch processing APIs
  3. Performance & Scalability:
    • Distributed processing support for large video collections
    • Advanced caching mechanisms for repeated processing
    • Multi-GPU support for FrameCleaner operations
  4. Medical Workflow Integration:
    • DICOM support for medical imaging workflows
    • HL7 FHIR integration for healthcare systems
    • Advanced medical entity recognition models

Contributing

See CONTRIBUTING.md for contribution guidelines, testing instructions, and communication channels.

License

Released under the MIT License.

Contact

Questions? Email lux@coloreg.de .

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lx_anonymizer-0.8.9.13.tar.gz (532.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lx_anonymizer-0.8.9.13-py3-none-any.whl (691.5 kB view details)

Uploaded Python 3

File details

Details for the file lx_anonymizer-0.8.9.13.tar.gz.

File metadata

  • Download URL: lx_anonymizer-0.8.9.13.tar.gz
  • Upload date:
  • Size: 532.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.11

File hashes

Hashes for lx_anonymizer-0.8.9.13.tar.gz
Algorithm Hash digest
SHA256 58bcb0cb234b05ce0444c8c567453ef3fc8de8a6cab2b2be5af091ee3d91ebe1
MD5 33b3abaf839a65041156c56f17f62a09
BLAKE2b-256 466f72fe84f0611ee5fd5fba0556984a3256231a364e8c138835d3723faae395

See more details on using hashes here.

File details

Details for the file lx_anonymizer-0.8.9.13-py3-none-any.whl.

File metadata

File hashes

Hashes for lx_anonymizer-0.8.9.13-py3-none-any.whl
Algorithm Hash digest
SHA256 30e351244160ab9c4a4f7ff407502eb9445f8c4a605a9cfb92468367ed1f6f14
MD5 732c37937cc02b4abdc5e6ed376d31e6
BLAKE2b-256 20597eca1b12f2944dc538db5a6c08902e034cb777bd6a04acb04ac27ce1d000

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page