Skip to main content

OCR-driven anonymization pipeline for medical reports and endoscopy frames

Project description

LX Anonymizer

LX Anonymizer is a comprehensive toolkit for de-identifying endoscopy frames and medical reports. It combines advanced OCR pipelines, spaCy-based NER, heuristic sanitizers, and report-specific rules to redact or pseudonymize sensitive information while preserving clinical context.

Core Components

ReportReader

Specialized for medical report anonymization with support for:

  • Multi-format processing: PDFs and images with automatic OCR fallback
  • Advanced metadata extraction: LLM-powered extraction using DeepSeek, MedLLaMA, or Llama3
  • Ensemble OCR: Combines Tesseract and TrOCR for improved accuracy
  • PDF anonymization: Creates blackened PDFs with sensitive regions automatically masked
  • Batch processing: Handles multiple reports with comprehensive error handling

FrameCleaner

Designed for real-time video frame anonymization featuring:

  • Hardware-accelerated processing: NVIDIA NVENC support with CPU fallback
  • Streaming video processing: Processes videos without full re-encoding when possible
  • Adaptive frame sampling: Optimizes performance for long videos (>10,000 frames)
  • Multiple anonymization strategies: Frame removal or mask overlay techniques
  • ROI-based masking: Device-specific region masking for endoscopic equipment

Default Return Format

LX Anonymizer will return a sensitive meta compliant dict when running either of the main client functions above.

Highlights

  • End-to-end anonymization of PDFs and video sequences using OCR, NER, and pseudonymization helpers.
  • Modular pipeline that lets you choose between Tesseract, TrOCR, ensemble OCR, and multiple metadata extractors.
  • Hardware optimization with NVENC acceleration for real-time video processing and streaming capabilities.
  • Human-in-the-loop ready outputs: original/anonymized text side by side, metadata JSON, and validation artefacts.
  • Extensible ruleset covering device-specific renderers, fuzzy name matching, and language-specific replacements.

Requirements

  • Python 3.12+
  • Linux or macOS (Windows support is experimental)
  • NVIDIA GPU recommended for real-time video anonymization (CUDA 12.x). CPU-only processing works but is slower.
  • Optional extras:
    • spaCy de_core_news_lg model (download after installation)
    • Torch vision/audio for video OCR workloads
    • Ollama-compatible LLMs for advanced metadata extraction

Installation

From PyPI (upcoming release)

pip install lx-anonymizer

Install extras to tailor the footprint:

pip install "lx-anonymizer[gpu,ocr,llm,dev]"

From source

git clone https://github.com/wg-lux/lx-anonymizer.git
cd lx-anonymizer
uv sync

Nix development shell

direnv allow
nix develop

This loads GPU, OCR, and tooling dependencies declared in devenv.nix.

Model downloads

After installation, fetch the German spaCy model:

python -m spacy download de_core_news_lg

First CLI runs also download OCR checkpoints (EAST, TrOCR, etc.). For air-gapped deployments, grab the archives listed in lx_anonymizer/settings.py and place them in ~/.cache/lx-anonymizer.

Quickstart

CLI Usage

Report Processing

# Process a single medical report
python -m cli.report_reader process report.pdf --ensemble --output-dir ./anonymized

# Use LLM for enhanced metadata extraction
python -m cli.report_reader process report.pdf --llm-extractor deepseek --use-ocr

# Batch process multiple reports
python -m cli.report_reader batch /path/to/reports/ --output-dir ./output --max-files 10

Video Frame Cleaning

# Clean a single video file
python -m lx_anonymizer.cli.frame_cleaner_cli clean video.mp4 --output-dir ./cleaned

# Batch clean multiple videos
python -m lx_anonymizer.cli.frame_cleaner_cli batch /path/to/videos/ --output-dir ./output

Useful CLI options:

  • --llm-extractor {deepseek,medllama,llama3} for LLM-powered metadata extraction.
  • --use-ocr and --ensemble to switch OCR strategies.
  • batch and extract sub-commands for folder processing or metadata-only runs.

Python API

ReportReader API

from lx_anonymizer import ReportReader

# Basic usage
reader = ReportReader(locale="de_DE")
original, anonymized, meta, pdf_path = reader.process_report(
    pdf_path="/path/to/report.pdf",
    use_ensemble=True,
    use_llm_extractor="deepseek",
)

# Create anonymized PDF with blackened sensitive regions
original, anonymized, meta, anonymized_pdf = reader.process_report(
    pdf_path="/path/to/report.pdf",
    create_anonymized_pdf=True,
    anonymized_pdf_output_path="/path/to/output.pdf"
)

# Advanced processing with region cropping
original, anonymized, meta, cropped_regions, pdf_path = reader.process_report_with_cropping(
    pdf_path="/path/to/report.pdf",
    crop_output_dir="/path/to/cropped_regions",
    crop_sensitive_regions=True,
    use_llm_extractor="deepseek"
)

FrameCleaner API

from lx_anonymizer.frame_cleaner import FrameCleaner
from pathlib import Path

# Initialize with hardware acceleration
cleaner = FrameCleaner(use_llm=True)

# Clean video with mask overlay (preserves all frames)
cleaned_path, metadata = cleaner.clean_video(
    video_path=Path("endoscopy.mp4"),
    endoscope_image_roi={"x": 550, "y": 0, "width": 1350, "height": 1080},
    endoscope_data_roi_nested={"patient_info": {"x": 10, "y": 10, "width": 300, "height": 50}},
    technique="mask_overlay"
)

# Remove sensitive frames entirely
cleaned_path, metadata = cleaner.clean_video(
    video_path=Path("endoscopy.mp4"),
    endoscope_image_roi=roi_config,
    endoscope_data_roi_nested=data_roi_config,
    technique="remove_frames"
)

See tests/test_cli_integration.py and tests/test_frame_cleaner.py for more examples.

Advanced Features

ReportReader Capabilities

  • Intelligent OCR Fallback: Automatically switches to OCR when PDF text extraction yields poor results
  • Multi-LLM Support: DeepSeek, MedLLaMA, and Llama3 integration for enhanced medical entity extraction
  • Ensemble OCR: Combines multiple OCR engines (Tesseract + TrOCR) for improved accuracy
  • PDF Anonymization: Creates masked PDFs with sensitive regions automatically blackened
  • Batch Processing: Processes multiple reports with error recovery and progress tracking
  • Metadata Validation: Cross-validates extracted information using multiple extraction methods

FrameCleaner Capabilities

  • Adaptive Sampling: Automatically samples frames for long videos (>10,000 frames) to optimize performance
  • Hardware Acceleration: NVIDIA NVENC support with automatic CPU fallback for unsupported systems
  • Streaming Processing: Uses FFmpeg streaming and named pipes to minimize memory usage and processing time
  • ROI-based Processing: Device-specific region configurations for endoscopic equipment (Olympus CV-1500, etc.)
  • Multiple Anonymization Strategies:
    • Mask Overlay: Blacks out sensitive regions while preserving video timeline
    • Frame Removal: Completely removes sensitive frames from the video stream
  • Quality Optimization: Automatic pixel format conversion and codec selection for minimal quality loss

Performance Optimizations

  • Stream Copy Operations: Avoids re-encoding when possible, using FFmpeg's -c copy for maximum speed
  • Named Pipe Support: In-memory video streaming for frame removal operations
  • Batch Metadata Extraction: Processes multiple frames simultaneously for improved efficiency
  • Hardware Detection: Automatically detects and uses available hardware acceleration (NVENC, QuickSync)

Data directories

By default, outputs live in ~/etc/lx-anonymizer/{data,temp}. Adjust them in lx_anonymizer/directory_setup.py. Clean temp regularly to avoid large intermediate artefacts.

Development workflow

  • Code quality: uv run flake8 for linting and formatting
  • Testing:
    • CPU-friendly tests: uv run pytest -m "not gpu"
    • GPU-accelerated tests: uv run pytest -m gpu (requires CUDA-capable hardware)
    • Integration tests: uv run pytest tests/test_cli_integration.py
    • Frame processing tests: uv run pytest tests/test_frame_cleaner.py
  • Performance profiling: Use --log-level DEBUG for detailed timing information
  • Build: uv run python -m build for wheel creation
  • Full validation: scripts/run_checks.sh for comprehensive local testing

Testing Medical Workflows

  • ReportReader: Test with sample medical PDFs in German and English
  • FrameCleaner: Validate with endoscopic video files (MP4, AVI formats supported)
  • Integration: Use example_anonymize_pdf.py for end-to-end testing scenarios

Project roadmap

  1. Release Management:
    • Publish CPU-only wheel to TestPyPI
    • Add optional extras for GPU/LLM workloads and slim default install
    • Automate release workflow (wheel + sdist upload, GitHub release notes)
  2. API Enhancement:
    • Expose REST/gRPC service with validation UI
    • WebSocket support for real-time video processing
    • Enhanced batch processing APIs
  3. Performance & Scalability:
    • Distributed processing support for large video collections
    • Advanced caching mechanisms for repeated processing
    • Multi-GPU support for FrameCleaner operations
  4. Medical Workflow Integration:
    • DICOM support for medical imaging workflows
    • HL7 FHIR integration for healthcare systems
    • Advanced medical entity recognition models

Contributing

See CONTRIBUTING.md for contribution guidelines, testing instructions, and communication channels.

License

Released under the MIT License.

Contact

Questions? Email lux@coloreg.de .

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lx_anonymizer-0.8.9.10.tar.gz (532.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lx_anonymizer-0.8.9.10-py3-none-any.whl (692.0 kB view details)

Uploaded Python 3

File details

Details for the file lx_anonymizer-0.8.9.10.tar.gz.

File metadata

  • Download URL: lx_anonymizer-0.8.9.10.tar.gz
  • Upload date:
  • Size: 532.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.11

File hashes

Hashes for lx_anonymizer-0.8.9.10.tar.gz
Algorithm Hash digest
SHA256 466a0574e70e6f2561dc8bde21cfb3c4669ed00e8372e87467c7e97074307d10
MD5 fa6adc249993585363c2294c0b383351
BLAKE2b-256 1280faa4710d6f2fe7b3813c93b0f3c869972bb96491f1f2a78ececace1ecf3f

See more details on using hashes here.

File details

Details for the file lx_anonymizer-0.8.9.10-py3-none-any.whl.

File metadata

File hashes

Hashes for lx_anonymizer-0.8.9.10-py3-none-any.whl
Algorithm Hash digest
SHA256 d2e46285ef4b21488ee8f89a62c2b512eeef77eac393daaaa042a8dc743b4b71
MD5 f724738e8b85cd494724aa66f4d196bb
BLAKE2b-256 5e7e7bcfb0827f6df5faf2c737d04596ec5c72dad70a3f50aee6ba56e8810d3a

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page