Agentic and single-pass Key Information Extraction (KIE) from documents using LLMs
Project description
Agentic KIE
Structured key information extraction from PDF documents, powered by LLMs.
A document enters the system as a file path. It leaves as a validated Pydantic instance. Everything in between (text-layer detection, OCR routing, image rendering, LLM orchestration, output parsing, retry logic) is the library's responsibility.
Contents
- The problem
- The idea
- Installation
- Core abstractions
- Extraction strategies
- Modalities
- Error handling
- Examples
- Contributing
The problem
Extracting structured data from PDFs is deceptively hard. The file format is a rendering instruction set, not a data container. Text layers may be missing, malformed, or absent entirely in scanned documents. Layout carries semantic meaning that raw text extraction destroys. And once you have the content, you still need an orchestration layer that let a LLM reason over it, produce typed output, and handle the inevitable failures.
The idea
A document enters the system as a file path. It leaves as a validated Pydantic instance. Everything in between (text-layer detection, OCR routing, image rendering, LLM orchestration, output parsing, retry logic) is the library's responsibility.
Two extraction strategies are available:
- Single-pass: One structured LLM call over the full document text. Fastest and cheapest option. Matches or outperforms agentic in most configurations (especially with smaller models).
- Agentic: A ReAct agent loop with multimodal document tools. More resilient to document length, but only justifies its cost with standard-tier models on long or complex documents.
Both strategies satisfy the same protocol and return the same type. Swap one for the other without changing downstream code.
from pathlib import Path
from pydantic import BaseModel
from langchain_anthropic import ChatGoogleGenerativeAI
from agentic_kie import PDFLoader, SinglePassExtractor, AgenticExtractor
class Invoice(BaseModel):
vendor: str
total: float
currency: str
due_date: str | None
model = ChatGoogleGenerativeAI(model="gemini-3.1-flash-lite-preview")
document = PDFLoader().load(Path("invoice.pdf"))
# Single LLM call
single = SinglePassExtractor(model=model, schema=Invoice)
result = single.extract(document)
# Or let an agent reason over the document
agent = AgenticExtractor(model=model, schema=Invoice)
result = agent.extract(document)
agentic-kie packages that entire workflow into a typed, tested library with a clear separation of concerns: document ingestion, content representation, and structured extraction.
Installation
[!IMPORTANT] Requires Python 3.13 or later.
uv add agentic-kie
Install with a model provider:
uv add "agentic-kie[anthropic]" # Claude
uv add "agentic-kie[google]" # Gemini
uv add "agentic-kie[openai]" # GPT
uv add "agentic-kie[bedrock]" # AWS Bedrock
uv add "agentic-kie[all]" # All of the above
[!TIP] Any LangChain chat model works. The extras above are provided for convenience.
Core abstractions
The library is organized around four concepts: a loader that absorbs PDF complexity, an immutable document that exposes content, a protocol for pluggable OCR, and extractors that produce structured output.
PDFLoader
The ingestion boundary. Takes a file path, detects whether the document has a native text layer (using a characters-per-page heuristic), routes to OCR when needed, and returns a validated PDFDocument.
from pathlib import Path
from agentic_kie import PDFLoader
loader = PDFLoader()
document = loader.load(Path("contract.pdf"))
For scanned documents, pass an OCR provider:
loader = PDFLoader(ocr_provider=MyOCRBackend())
document = loader.load(Path("scanned_contract.pdf"))
PDFDocument
An immutable representation of the loaded document. Exposes text and rendered page images (the two modalities that LLMs can reason over). Images are rendered lazily and cached on first access.
| Attribute / Method | Description |
|---|---|
page_count |
Total number of pages |
is_ocr |
True if text was extracted via OCR |
full_text |
All pages joined with double newlines |
read_text(start, end=None) |
Text slice over a page range (zero-indexed, half-open) |
all_images |
All pages as base64-encoded PNGs (cached) |
load_images(start, end=None) |
Image slice over a page range |
OCRProvider
A structural protocol. Any object with an extract_text(image: bytes) -> str method qualifies.
from agentic_kie import OCRProvider
class TextractProvider:
"""Wraps AWS Textract as an OCR backend."""
def extract_text(self, image: bytes) -> str:
# call Textract, return plain text
...
# TextractProvider satisfies OCRProvider by structure alone
loader = PDFLoader(ocr_provider=TextractProvider())
Extractors
Both extraction strategies satisfy the Extractor protocol: a single extract(document) -> T method that takes a PDFDocument and returns a validated instance of a Pydantic schema. This enables type-safe dispatch without coupling strategies through inheritance.
Extraction strategies
Single-pass extraction
SinglePassExtractor sends the full document content to the model in one call, with structured output bound to the target schema. The chain is built once at construction time and reused across documents.
from pydantic import BaseModel
from langchain_openai import ChatOpenAI
from agentic_kie import PDFLoader, SinglePassExtractor
class Invoice(BaseModel):
vendor: str
total: float
currency: str
due_date: str | None
document = PDFLoader().load(Path("invoice.pdf"))
extractor = SinglePassExtractor(
model=ChatOpenAI(model="gpt-5.4-mini"),
schema=Invoice,
modality="multimodal",
max_retries=3,
)
result = extractor.extract(document)
| Parameter | Type | Default | Description |
|---|---|---|---|
model |
BaseChatModel |
required | Any LangChain chat model |
schema |
type[T] |
required | Pydantic model defining the extraction target |
modality |
"text" | "image" | "multimodal" |
"text" |
Document representation sent to the model |
system_prompt |
str | None |
None |
Custom system prompt (uses a sensible default when omitted) |
max_retries |
int |
3 |
Retry attempts with exponential backoff and jitter |
Agentic extraction
AgenticExtractor builds a ReAct agent equipped with document tools (get_page_count, read_text, and load_images) scoped to the document being extracted. The agent decides which pages to inspect, in what order, and stops when it has enough information to produce the target schema.
from pydantic import BaseModel
from langchain_google_genai import ChatAnthropic
from agentic_kie import PDFLoader, AgenticExtractor
class Contract(BaseModel):
parties: list[str]
effective_date: str
governing_law: str | None
termination_clause: str | None
document = PDFLoader().load(Path("contract.pdf"))
extractor = AgenticExtractor(
model=ChatAnthropic(model="claude-haiku-4-5"),
schema=Contract,
modality="text",
max_iterations=50,
)
result = extractor.extract(document)
| Parameter | Type | Default | Description |
|---|---|---|---|
model |
BaseChatModel |
required | Any LangChain chat model |
schema |
type[T] |
required | Pydantic model defining the extraction target |
modality |
"text" | "image" | "multimodal" |
"text" |
Controls which document tools the agent can use |
system_prompt |
str |
(built-in) | Custom system prompt for the agent |
max_iterations |
int |
50 |
Maximum agent steps before raising ExtractionError |
max_retries |
int |
3 |
Retry attempts on transient model failures |
Modalities
Both extractors accept a modality parameter that controls how document content is presented to the model:
| Modality | What the model sees | When to use |
|---|---|---|
"text" |
Extracted text only | Reliable text layer, cost-sensitive, fast |
"image" |
Rendered page images (base64 PNG) | Visually rich documents, layout matters |
"multimodal" |
Text followed by images | Maximum signal, when accuracy justifies cost |
[!NOTE] For the agentic extractor, modality controls which tools are exposed:
"text"providesread_text,"image"providesload_images, and"multimodal"provides both.get_page_countis always available.
Error handling
All document-level failures derive from DocumentLoadError, making them easy to catch together or individually. Extraction failures raise ExtractionError.
from agentic_kie import (
DocumentLoadError,
CorruptDocumentError,
PasswordProtectedError,
EmptyDocumentError,
OCRNotConfiguredError,
ExtractionError,
)
try:
doc = loader.load(path)
result = extractor.extract(doc)
except PasswordProtectedError:
... # encrypted PDF
except OCRNotConfiguredError:
... # scanned document, no OCR provider
except EmptyDocumentError:
... # zero pages or no extractable text
except CorruptDocumentError:
... # unparseable file
except DocumentLoadError:
... # catch-all for loading failures
except ExtractionError:
... # agent exceeded iteration limit
Examples
The examples/ directory contains runnable scripts demonstrating both extraction strategies across different providers, using the Kleister NDA preparation package.
Before running any example, fetch the dataset:
uv run nda ./examples/data
This processes the Kleister NDA dataset into examples/data/, which the scripts expect. Then run a script from the project root:
uv run examples/agent/text-only.py
Contributing
See CONTRIBUTING.md for development setup, available make targets, and the CI/CD pipeline.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file agentic_kie-0.5.1.tar.gz.
File metadata
- Download URL: agentic_kie-0.5.1.tar.gz
- Upload date:
- Size: 848.3 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
3a177a1463b4858ef5f1ebe91516e01664edfdbb480c615712a573eaff6bdb81
|
|
| MD5 |
33d622096fe311554147c78390589c19
|
|
| BLAKE2b-256 |
dec138f6e9d5b852656fc1ff95c4999b0c5ef1de0ef7e0687fac13b88c65c7ab
|
Provenance
The following attestation bundles were made for agentic_kie-0.5.1.tar.gz:
Publisher:
cd.yml on gafnts/agentic-kie
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
agentic_kie-0.5.1.tar.gz -
Subject digest:
3a177a1463b4858ef5f1ebe91516e01664edfdbb480c615712a573eaff6bdb81 - Sigstore transparency entry: 1281004500
- Sigstore integration time:
-
Permalink:
gafnts/agentic-kie@335f2c22837e73e865963e151a6d76c54e31e6a7 -
Branch / Tag:
refs/tags/v0.5.1 - Owner: https://github.com/gafnts
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
cd.yml@335f2c22837e73e865963e151a6d76c54e31e6a7 -
Trigger Event:
push
-
Statement type:
File details
Details for the file agentic_kie-0.5.1-py3-none-any.whl.
File metadata
- Download URL: agentic_kie-0.5.1-py3-none-any.whl
- Upload date:
- Size: 18.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
aba755f76f4bc06cab815d331cca1a2ace06fbf43382833654ed26457ad487ec
|
|
| MD5 |
eaa1e7980522d62a134301316e67410e
|
|
| BLAKE2b-256 |
846e2d3d44082b7e3324b8e04b7deec35f8417d4b8f131e1a7a02b13dc330943
|
Provenance
The following attestation bundles were made for agentic_kie-0.5.1-py3-none-any.whl:
Publisher:
cd.yml on gafnts/agentic-kie
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
agentic_kie-0.5.1-py3-none-any.whl -
Subject digest:
aba755f76f4bc06cab815d331cca1a2ace06fbf43382833654ed26457ad487ec - Sigstore transparency entry: 1281004522
- Sigstore integration time:
-
Permalink:
gafnts/agentic-kie@335f2c22837e73e865963e151a6d76c54e31e6a7 -
Branch / Tag:
refs/tags/v0.5.1 - Owner: https://github.com/gafnts
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
cd.yml@335f2c22837e73e865963e151a6d76c54e31e6a7 -
Trigger Event:
push
-
Statement type: