Skip to main content

Production-grade Document Ingestion & Canonicalization Engine

Project description

Document IR - Production Document Ingestion Engine

An IR-first, extensible document compiler for AI systems.

This is NOT a PDF-to-Markdown script. It is a production-grade document ingestion and canonicalization engine designed with compiler-like architecture: Input → IR → Backends.

Architecture

Design Philosophy

Think like a compiler engineer:

  • Input Layer: Format-specific parsers (currently PDF via Docling)
  • AST/IR: Canonical intermediate representation with strict schema
  • Backends: Multiple export formats (Markdown, Text, Parquet)

Layer Separation (Non-Negotiable)

flowchart TD
    A[Input Adapter Layer<br/>Format-specific parsing only]
    B[Extraction Layer<br/>Extract raw structural elements]
    C[Normalization Layer<br/>Convert to canonical IR with hashing]
    D[Canonical IR Layer<br/>Typed schema, stable IDs, relationships]
    E[Export Layer<br/>Markdown, Text, Parquet, Assets]

    A --> B
    B --> C
    C --> D
    D --> E

Key Features

✅ Deterministic & Idempotent

  • Hash-based stable IDs (document, block, table, image, chunk)
  • Running pipeline twice produces identical output
  • No UUIDs, no randomness

✅ Canonical IR Schema

Document
├── document_id: str (hash-based)
├── schema_version: str
├── parser_version: str
├── metadata: DocumentMetadata
├── blocks: List[Block]
   ├── block_id: str (deterministic)
   ├── type: BlockType (heading, paragraph, table, image, etc.)
   ├── content: str
   ├── page_number: int
   ├── bbox: BoundingBox
   └── metadata: dict
└── relationships: List[Relationship]

✅ Pluggable Chunking

  • SemanticSectionChunker: Section-based (headings)
  • TokenWindowChunker: Fixed token windows with overlap
  • LayoutAwareChunker: Layout-aware (stub)

All chunking operates on IR, not raw text.

✅ Multiple Export Formats

  • Markdown: Human-readable with formatting
  • Plain Text: Simple text extraction
  • Parquet: Efficient structured storage for tables/blocks
  • Assets: Extracted images (PNG) and tables (CSV)

✅ Structured Output

/<document_id>/
    manifest.json       # Processing metadata
    ir.json            # Canonical IR
    chunks.json        # Chunk definitions
    /assets/
        /images/       # Extracted images
        /tables/       # Tables as CSV
    /exports/
        /markdown/     # Markdown output
        /text/         # Plain text output
        /parquet/      # Parquet datasets
    /logs/             # Processing logs

Installation

IMPORTANT: LayoutIR requires PyTorch with CUDA 13.0 support for GPU acceleration. Install PyTorch first:

# Step 1: Install PyTorch with CUDA 13.0 (REQUIRED)
pip3 install torch torchvision --index-url https://download.pytorch.org/whl/cu130

# Step 2: Install LayoutIR
pip install layoutir

Alternative Installation Methods

# Install from source
git clone https://github.com/RahulPatnaik/layoutir.git
cd layoutir
pip3 install torch torchvision --index-url https://download.pytorch.org/whl/cu130
pip install -e .

Note: The package intentionally does not include PyTorch in its base dependencies to ensure you get the correct CUDA version. Any existing PyTorch installation will be overwritten by the CUDA 13.0 version.

Usage

Basic Usage

# Using the CLI
layoutir --input file.pdf --output ./out

# Or using Python directly
python -m layoutir.cli --input file.pdf --output ./out

Advanced Options

# Semantic chunking (default)
layoutir --input file.pdf --output ./out --chunk-strategy semantic

# Token-based chunking with custom size
layoutir --input file.pdf --output ./out \
  --chunk-strategy token \
  --chunk-size 1024 \
  --chunk-overlap 128

# Enable GPU acceleration
layoutir --input file.pdf --output ./out --use-gpu

# Debug mode with structured logging
layoutir --input file.pdf --output ./out \
  --log-level DEBUG \
  --structured-logs

Python API

from pathlib import Path
from layoutir import Pipeline
from layoutir.adapters import DoclingAdapter
from layoutir.chunking import SemanticSectionChunker

# Create pipeline
adapter = DoclingAdapter(use_gpu=True)
chunker = SemanticSectionChunker(max_heading_level=2)
pipeline = Pipeline(adapter=adapter, chunk_strategy=chunker)

# Process document
document = pipeline.process(
    input_path=Path("document.pdf"),
    output_dir=Path("./output")
)

# Access results
print(f"Extracted {len(document.blocks)} blocks")
print(f"Document ID: {document.document_id}")

Project Structure

src/layoutir/
├── schema.py              # Canonical IR schema (Pydantic)
├── pipeline.py            # Main orchestrator
│
├── adapters/              # Input adapters
│   ├── base.py           # Abstract interface
│   └── docling_adapter.py # PDF via Docling
│
├── extraction/            # Raw element extraction
│   └── docling_extractor.py
│
├── normalization/         # IR normalization
│   └── normalizer.py
│
├── chunking/              # Chunking strategies
│   └── strategies.py
│
├── exporters/             # Export backends
│   ├── markdown_exporter.py
│   ├── text_exporter.py
│   ├── parquet_exporter.py
│   └── asset_writer.py
│
└── utils/
    ├── hashing.py        # Deterministic ID generation
    └── logging_config.py  # Structured logging

ingest.py                  # CLI entrypoint
benchmark.py               # Performance benchmark
test_pipeline.py           # Integration test

Design Constraints

✅ What We DO

  • Strict layer separation
  • Deterministic processing
  • Schema validation
  • Pluggable strategies
  • Observability/timing
  • Efficient storage (Parquet)

❌ What We DON'T DO

  • Mix business logic into adapters
  • Hardcode paths or configurations
  • Use non-deterministic IDs (UUIDs)
  • Combine IR and export logic
  • Skip schema validation
  • Load entire files into memory unnecessarily

Extensibility

Adding New Input Formats

  1. Implement InputAdapter interface:
class DocxAdapter(InputAdapter):
    def parse(self, file_path: Path) -> Any: ...
    def supports_format(self, file_path: Path) -> bool: ...
    def get_parser_version(self) -> str: ...
  1. Implement corresponding extractor
  2. Update pipeline to use new adapter

Adding New Chunk Strategies

class CustomChunker(ChunkStrategy):
    def chunk(self, document: Document) -> List[Chunk]:
        # Operate on IR blocks
        ...

Adding New Export Formats

class JsonExporter(Exporter):
    def export(self, document: Document, output_dir: Path, chunks: List[Chunk]):
        # Export from canonical IR
        ...

Performance

Designed to handle 200+ page PDFs efficiently:

  • Streaming processing where possible
  • Lazy loading of heavy dependencies
  • GPU acceleration support
  • Parallel export operations
  • Efficient Parquet storage for tables

Observability

  • Structured JSON logging
  • Stage-level timing metrics
  • Extraction statistics
  • Deterministic output for debugging

Schema Versioning

Current schema version: 1.0.0

Future schema changes will be tracked via semantic versioning:

  • Major: Breaking changes to IR structure
  • Minor: Backwards-compatible additions
  • Patch: Bug fixes

Future Enhancements

  • DOCX input adapter
  • HTML input adapter
  • Advanced layout-aware chunking
  • Parallel page processing
  • Incremental updates (only reprocess changed pages)
  • Vector embeddings export
  • OCR fallback for scanned PDFs

License

See project root for license information.

Contributing

This is a research/prototype phase project. See main project README for contribution guidelines.

layoutir

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

layoutir-1.0.3.tar.gz (31.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

layoutir-1.0.3-py3-none-any.whl (35.7 kB view details)

Uploaded Python 3

File details

Details for the file layoutir-1.0.3.tar.gz.

File metadata

  • Download URL: layoutir-1.0.3.tar.gz
  • Upload date:
  • Size: 31.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for layoutir-1.0.3.tar.gz
Algorithm Hash digest
SHA256 fc8d8dc61d7fc7595a8239ef58e7055fe96ad59284f0dc39b537d5f30503d33d
MD5 fc2ca03e0b4c1871d5be39b1202cba3b
BLAKE2b-256 753b5cb9d95a62d0493cadf3179ce055cb8ff3805a25381b6264576a80e087d8

See more details on using hashes here.

Provenance

The following attestation bundles were made for layoutir-1.0.3.tar.gz:

Publisher: ci.yml on RahulPatnaik/layoutir

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file layoutir-1.0.3-py3-none-any.whl.

File metadata

  • Download URL: layoutir-1.0.3-py3-none-any.whl
  • Upload date:
  • Size: 35.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for layoutir-1.0.3-py3-none-any.whl
Algorithm Hash digest
SHA256 ccfca964bb4120fc87c6141b3de88cf6a342a65e2c7d9627eb225cb3389731fe
MD5 6ccbe355aae5e2badc9e64ac88ae1707
BLAKE2b-256 57a4f854d37c5764c6e92c7ba07fe31d30b4d9ee8d2c996d8c6741db0c81dac5

See more details on using hashes here.

Provenance

The following attestation bundles were made for layoutir-1.0.3-py3-none-any.whl:

Publisher: ci.yml on RahulPatnaik/layoutir

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page