Skip to main content

A library for performing inference using trained models.

Project description

Open-Source Pre-Processing Tools for Unstructured Data

The unstructured-inference repo contains hosted model inference code for layout parsing models. These models are invoked via API as part of the partitioning bricks in the unstructured package.

Requires Python 3.12+.

Installation

Package

pip install unstructured-inference

Detectron2

Detectron2 is required for using models from the layoutparser model zoo but is not automatically installed with this package. For MacOS and Linux, build from source with:

pip install 'git+https://github.com/facebookresearch/detectron2.git@57bdb21249d5418c130d54e2ebdc94dda7a4c01a'

Other install options can be found in the Detectron2 installation guide.

Windows is not officially supported by Detectron2, but some users are able to install it anyway. See discussion here for tips on installing Detectron2 on Windows.

Development Setup

This project uses uv for dependency management.

# Clone and install all dependencies (including dev/test/lint groups)
git clone https://github.com/Unstructured-IO/unstructured-inference.git
cd unstructured-inference
make install

Run make help for a full list of available targets.

Getting Started

To get started with the layout parsing model, use the following commands:

from unstructured_inference.inference.layout import DocumentLayout

layout = DocumentLayout.from_file("sample-docs/loremipsum.pdf")

print(layout.pages[0].elements)

Once the model has detected the layout and OCR'd the document, the text extracted from the first page of the sample document will be displayed. You can convert a given element to a dict by running the .to_dict() method.

Models

The inference pipeline operates by finding text elements in a document page using a detection model, then extracting the contents of the elements using direct extraction (if available), OCR, and optionally table inference models.

We offer several detection models including Detectron2 and YOLOX.

Using a non-default model

When doing inference, an alternate model can be used by passing the model object to the ingestion method via the model parameter. The get_model function can be used to construct one of our out-of-the-box models from a keyword, e.g.:

from unstructured_inference.models.base import get_model
from unstructured_inference.inference.layout import DocumentLayout

model = get_model("yolox")
layout = DocumentLayout.from_file("sample-docs/layout-parser-paper.pdf", detection_model=model)

Using your own model

Any detection model can be used for in the unstructured_inference pipeline by wrapping the model in the UnstructuredObjectDetectionModel class. To integrate with the DocumentLayout class, a subclass of UnstructuredObjectDetectionModel must have a predict method that accepts a PIL.Image.Image and returns a list of LayoutElements, and an initialize method, which loads the model and prepares it for inference.

Security Policy

See our security policy for information on how to report security vulnerabilities.

Learn more

Section Description
Unstructured Community Github Information about Unstructured.io community projects
Unstructured Github Unstructured.io open source repositories
Company Website Unstructured.io product and company info

Project details


Release history Release notifications | RSS feed

This version

1.6.0

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

unstructured_inference-1.6.0.tar.gz (46.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

unstructured_inference-1.6.0-py3-none-any.whl (53.5 kB view details)

Uploaded Python 3

File details

Details for the file unstructured_inference-1.6.0.tar.gz.

File metadata

  • Download URL: unstructured_inference-1.6.0.tar.gz
  • Upload date:
  • Size: 46.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for unstructured_inference-1.6.0.tar.gz
Algorithm Hash digest
SHA256 00f8f38c796c9d3cbe83a691d2980e62296d4805df1c5207b9128f5aa6ff8caa
MD5 0d64b45374e3de2912a0ae8836b77a38
BLAKE2b-256 d548d3cc604d1fe2d706e0d1343bc970c80c6d91bf5d39a76ea1dd1c12c99e07

See more details on using hashes here.

Provenance

The following attestation bundles were made for unstructured_inference-1.6.0.tar.gz:

Publisher: release.yml on Unstructured-IO/unstructured-inference

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file unstructured_inference-1.6.0-py3-none-any.whl.

File metadata

File hashes

Hashes for unstructured_inference-1.6.0-py3-none-any.whl
Algorithm Hash digest
SHA256 299df847dd7ddefc27e0c67b3f901c317789c32ce7a4cf2fd3b46d3729ef7ae3
MD5 af9485ab30fcdc710bf0a17ef57d9b8d
BLAKE2b-256 4d69aac631695bcfec159b7edced467ee2fa22f5dc33966174332c1a9bcc3e37

See more details on using hashes here.

Provenance

The following attestation bundles were made for unstructured_inference-1.6.0-py3-none-any.whl:

Publisher: release.yml on Unstructured-IO/unstructured-inference

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page