Skip to main content

Agentic and single-pass Key Information Extraction (KIE) from documents using LLMs

Project description

Agentic KIE

Structured key information extraction from PDF documents, powered by LLMs.

CI CD codecov PyPI License: MIT


A document enters the system as a file path. It leaves as a validated Pydantic instance. Everything in between (text-layer detection, OCR routing, image rendering, LLM orchestration, output parsing, retry logic) is the library's responsibility.

Contents


The problem

Extracting structured data from PDFs is deceptively hard. The file format is a rendering instruction set, not a data container. Text layers may be missing, malformed, or absent entirely in scanned documents. Layout carries semantic meaning that raw text extraction destroys. And once you have the content, you still need an orchestration layer that let a LLM reason over it, produce typed output, and handle the inevitable failures.


The idea

A document enters the system as a file path. It leaves as a validated Pydantic instance. Everything in between (text-layer detection, OCR routing, image rendering, LLM orchestration, output parsing, retry logic) is the library's responsibility.

Two extraction strategies are available:

  • Single-pass: One structured LLM call over the full document text. Fastest and cheapest option. Matches or outperforms agentic in most configurations (especially with smaller models).
  • Agentic: A ReAct agent loop with multimodal document tools. More resilient to document length, but only justifies its cost with standard-tier models on long or complex documents.

Both strategies satisfy the same protocol and return the same type. Swap one for the other without changing downstream code.

from pathlib import Path
from pydantic import BaseModel
from langchain_anthropic import ChatGoogleGenerativeAI
from agentic_kie import PDFLoader, SinglePassExtractor, AgenticExtractor

class Invoice(BaseModel):
    vendor: str
    total: float
    currency: str
    due_date: str | None

model = ChatGoogleGenerativeAI(model="gemini-3.1-flash-lite-preview")
document = PDFLoader().load(Path("invoice.pdf"))

# Single LLM call
single = SinglePassExtractor(model=model, schema=Invoice)
result = single.extract(document)

# Or let an agent reason over the document
agent = AgenticExtractor(model=model, schema=Invoice)
result = agent.extract(document)

agentic-kie packages that entire workflow into a typed, tested library with a clear separation of concerns: document ingestion, content representation, and structured extraction.


Installation

[!IMPORTANT] Requires Python 3.13 or later.

uv add agentic-kie

Install with a model provider:

uv add "agentic-kie[anthropic]"   # Claude
uv add "agentic-kie[google]"      # Gemini
uv add "agentic-kie[openai]"      # GPT
uv add "agentic-kie[bedrock]"     # AWS Bedrock
uv add "agentic-kie[all]"         # All of the above

[!TIP] Any LangChain chat model works. The extras above are provided for convenience.


Core abstractions

The library is organized around four concepts: a loader that absorbs PDF complexity, an immutable document that exposes content, a protocol for pluggable OCR, and extractors that produce structured output.

PDFLoader

The ingestion boundary. Takes a file path, detects whether the document has a native text layer (using a characters-per-page heuristic), routes to OCR when needed, and returns a validated PDFDocument.

from pathlib import Path
from agentic_kie import PDFLoader

loader = PDFLoader()
document = loader.load(Path("contract.pdf"))

For scanned documents, pass an OCR provider:

loader = PDFLoader(ocr_provider=MyOCRBackend())
document = loader.load(Path("scanned_contract.pdf"))

PDFDocument

An immutable representation of the loaded document. Exposes text and rendered page images (the two modalities that LLMs can reason over). Images are rendered lazily and cached on first access.

Attribute / Method Description
page_count Total number of pages
is_ocr True if text was extracted via OCR
full_text All pages joined with double newlines
read_text(start, end=None) Text slice over a page range (zero-indexed, half-open)
all_images All pages as base64-encoded PNGs (cached)
load_images(start, end=None) Image slice over a page range

OCRProvider

A structural protocol. Any object with an extract_text(image: bytes) -> str method qualifies.

from agentic_kie import OCRProvider

class TextractProvider:
    """Wraps AWS Textract as an OCR backend."""

    def extract_text(self, image: bytes) -> str:
        # call Textract, return plain text
        ...

# TextractProvider satisfies OCRProvider by structure alone
loader = PDFLoader(ocr_provider=TextractProvider())

Extractors

Both extraction strategies satisfy the Extractor protocol: a single extract(document) -> T method that takes a PDFDocument and returns a validated instance of a Pydantic schema. This enables type-safe dispatch without coupling strategies through inheritance.


Extraction strategies

Single-pass extraction

SinglePassExtractor sends the full document content to the model in one call, with structured output bound to the target schema. The chain is built once at construction time and reused across documents.

from pydantic import BaseModel
from langchain_openai import ChatOpenAI
from agentic_kie import PDFLoader, SinglePassExtractor

class Invoice(BaseModel):
    vendor: str
    total: float
    currency: str
    due_date: str | None

document = PDFLoader().load(Path("invoice.pdf"))

extractor = SinglePassExtractor(
    model=ChatOpenAI(model="gpt-5.4-mini"),
    schema=Invoice,
    modality="multimodal",
    max_retries=3,
)

result = extractor.extract(document)
Parameter Type Default Description
model BaseChatModel required Any LangChain chat model
schema type[T] required Pydantic model defining the extraction target
modality "text" | "image" | "multimodal" "text" Document representation sent to the model
system_prompt str | None None Custom system prompt (uses a sensible default when omitted)
max_retries int 3 Retry attempts with exponential backoff and jitter

Agentic extraction

AgenticExtractor builds a ReAct agent equipped with document tools (get_page_count, read_text, and load_images) scoped to the document being extracted. The agent decides which pages to inspect, in what order, and stops when it has enough information to produce the target schema.

from pydantic import BaseModel
from langchain_google_genai import ChatAnthropic
from agentic_kie import PDFLoader, AgenticExtractor

class Contract(BaseModel):
    parties: list[str]
    effective_date: str
    governing_law: str | None
    termination_clause: str | None

document = PDFLoader().load(Path("contract.pdf"))

extractor = AgenticExtractor(
    model=ChatAnthropic(model="claude-haiku-4-5"),
    schema=Contract,
    modality="text",
    max_iterations=50,
)

result = extractor.extract(document)
Parameter Type Default Description
model BaseChatModel required Any LangChain chat model
schema type[T] required Pydantic model defining the extraction target
modality "text" | "image" | "multimodal" "text" Controls which document tools the agent can use
system_prompt str (built-in) Custom system prompt for the agent
max_iterations int 50 Maximum agent steps before raising ExtractionError
max_retries int 3 Retry attempts on transient model failures

Modalities

Both extractors accept a modality parameter that controls how document content is presented to the model:

Modality What the model sees When to use
"text" Extracted text only Reliable text layer, cost-sensitive, fast
"image" Rendered page images (base64 PNG) Visually rich documents, layout matters
"multimodal" Text followed by images Maximum signal, when accuracy justifies cost

[!NOTE] For the agentic extractor, modality controls which tools are exposed: "text" provides read_text, "image" provides load_images, and "multimodal" provides both. get_page_count is always available.


Error handling

All document-level failures derive from DocumentLoadError, making them easy to catch together or individually. Extraction failures raise ExtractionError.

from agentic_kie import (
    DocumentLoadError,
    CorruptDocumentError,
    PasswordProtectedError,
    EmptyDocumentError,
    OCRNotConfiguredError,
    ExtractionError,
)

try:
    doc = loader.load(path)
    result = extractor.extract(doc)
except PasswordProtectedError:
    ...  # encrypted PDF
except OCRNotConfiguredError:
    ...  # scanned document, no OCR provider
except EmptyDocumentError:
    ...  # zero pages or no extractable text
except CorruptDocumentError:
    ...  # unparseable file
except DocumentLoadError:
    ...  # catch-all for loading failures
except ExtractionError:
    ...  # agent exceeded iteration limit

Examples

The examples/ directory contains runnable scripts demonstrating both extraction strategies across different providers, using the Kleister NDA preparation package.

Before running any example, fetch the dataset:

uv run nda ./examples/data

This processes the Kleister NDA dataset into examples/data/, which the scripts expect. Then run a script from the project root:

uv run examples/agent/text-only.py

Contributing

See CONTRIBUTING.md for development setup, available make targets, and the CI/CD pipeline.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

agentic_kie-0.5.1.tar.gz (848.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

agentic_kie-0.5.1-py3-none-any.whl (18.6 kB view details)

Uploaded Python 3

File details

Details for the file agentic_kie-0.5.1.tar.gz.

File metadata

  • Download URL: agentic_kie-0.5.1.tar.gz
  • Upload date:
  • Size: 848.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for agentic_kie-0.5.1.tar.gz
Algorithm Hash digest
SHA256 3a177a1463b4858ef5f1ebe91516e01664edfdbb480c615712a573eaff6bdb81
MD5 33d622096fe311554147c78390589c19
BLAKE2b-256 dec138f6e9d5b852656fc1ff95c4999b0c5ef1de0ef7e0687fac13b88c65c7ab

See more details on using hashes here.

Provenance

The following attestation bundles were made for agentic_kie-0.5.1.tar.gz:

Publisher: cd.yml on gafnts/agentic-kie

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file agentic_kie-0.5.1-py3-none-any.whl.

File metadata

  • Download URL: agentic_kie-0.5.1-py3-none-any.whl
  • Upload date:
  • Size: 18.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for agentic_kie-0.5.1-py3-none-any.whl
Algorithm Hash digest
SHA256 aba755f76f4bc06cab815d331cca1a2ace06fbf43382833654ed26457ad487ec
MD5 eaa1e7980522d62a134301316e67410e
BLAKE2b-256 846e2d3d44082b7e3324b8e04b7deec35f8417d4b8f131e1a7a02b13dc330943

See more details on using hashes here.

Provenance

The following attestation bundles were made for agentic_kie-0.5.1-py3-none-any.whl:

Publisher: cd.yml on gafnts/agentic-kie

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page