Skip to main content

Automate information extraction for multimodal LLMs.

Project description

Pipeline Illustration The Pipe

English | 中文

codecov python-gh-action Website get API Join discord

Feed PDFs, URLs, Slides, YouTube videos, Word docs and more into Vision-Language models with one line of code ⚡

The Pipe is a multimodal-first tool for feeding files and web pages into vision-language models such as GPT-4V. It is best for LLM and RAG applications that want to support comprehensive textual and visual understanding across a wide range of data sources. The Pipe is available as a 24/7 hosted API at thepi.pe, or it can be set up locally to let you run the compute.

Science assistant demo

Features 🌟

  • Extracts text and visuals from files or web pages 📚
  • Outputs chunks optimized for multimodal LLMs and RAG frameworks 🖼️
  • Interpret complex PDFs, web pages, docs, videos, data, and more 🧠
  • Auto-compress prompts exceeding your chosen token limit 📦
  • Works even with missing file extensions, in-memory data streams 💾
  • Works with codebases, git repos, and custom integrations 🌐
  • Multi-threaded ⚡️

Getting Started 🚀

The Pipe handles a wide array of complex filetypes, and thus has many dependencies that must be installed separately. It also requires a strong machine for good response times. For this reason, we host it as an API that works out-of-the-box.

First, install The Pipe.

pip install thepipe_api

The Pipe is available as a hosted API, or it can be set up locally. An API key is recommended for out-of-the-box functionality (alternatively, see the local installation section). Ensure the THEPIPE_API_KEY environment variable is set. Don't have a key yet? Get one here.

Now you can extract comprehensive text and visuals from any file:

from thepipe_api import thepipe
messages = thepipe.extract("example.pdf")

Or websites:

messages = thepipe.extract("https://example.com")

Then feed it into GPT-4-Vision:

response = client.chat.completions.create(
    model="gpt-4-vision-preview",
    messages = messages,
)

Just call OpenAI

You can also use The Pipe from the command line. Here's how to recursively extract from a directory, matching only files containing a substring (in this example, typescript files) and ignore files containing other substrings (in this example, anything in the "tests" folder):

thepipe path/to/folder --match tsx --ignore tests

Supported File Types 📚

Source Type Input types Token Compression 🗜️ Image Extraction 👁️ Notes 📌
Directory Any /path/to/directory ✔️ ✔️ Extracts from all files in directory, supports match and ignore patterns
Code .py, .tsx, .js, .html, .css, .cpp, etc ✔️ (varies) Combines all code files. .c, .cpp, .py are compressible with ctags, others are not
Plaintext .txt, .md, .rtf, etc ✔️ Regular text files
PDF .pdf ✔️ ✔️ Extracts text and images of each page; can use AI for extraction of table data and images within pages
Image .jpg, .jpeg, .png ✔️ Extracts images, uses OCR if text_only
Data Table .csv, .xls, .xlsx ✔️ Extracts data from spreadsheets; converts to text representation. For very large datasets, will only extract column names and types
Jupyter Notebook .ipynb ✔️ Extracts code, markdown, and images from Jupyter notebooks
Microsoft Word Document .docx ✔️ ✔️ Extracts text and images from Word documents
Microsoft PowerPoint Presentation .pptx ✔️ ✔️ Extracts text and images from PowerPoint presentations
Video .mp4, .avi, .mov, .wmv ✔️ ✔️ Extracts frames from video files; supports frame extraction and OCR for text extraction from frames
Audio .mp3, .wav ✔️ Extracts text from audio files; supports speech-to-text conversion
Website URLs (inputs starting with http, https, ftp) ✔️ ✔️ Extracts text from web page along with image (or images if scrollable); text-only extraction available
GitHub Repository GitHub repo URLs (inputs starting with https://github.com or https://www.github.com) ✔️ ✔️ Extracts from GitHub repositories; supports branch specification
YouTube Video YouTube video URLs (inputs starting with https://youtube.com or https://www.youtube.com) ✔️ ✔️ Extracts frames and transcript from YouTube videos in per-minute chunks
ZIP File .zip ✔️ ✔️ Extracts contents of ZIP files; supports nested directory extraction

How it works 🛠️

The input source is either a file path, a URL, or a directory. The pipe will extract information from the source and process it for downstream use with language models, vision transformers, or vision-language models. The output from the pipe is a sensible list of multimodal messages representing chunks of the extracted information, carefully crafted to fit within context windows for any models from gemma-7b to GPT-4. The messages returned should look like this:

[
  {
    "role": "user",
    "content": [
      {
        "type": "text",
        "text": "..."
      },
      {
        "type": "image_url",
        "image_url": {
          "url": "data:image/jpeg;base64,..."
        }
      }
    ]
  }
]

If you want to feed these messages directly into the model, it is important to be mindful of the token limit. OpenAI does not allow too many images in the prompt (see discussion here), so long files should be extracted with text_only=True to avoid this issue, while long text files should either be compressed or embedded in a RAG framework.

The text and images from these messages may also be prepared for a vector database with thepipe.core.create_chunks_from_messages or for downstream use with RAG frameworks. LiteLLM can be used to easily integrate The Pipe with any LLM provider.

It uses a variety of heuristics for optimal performance with vision-language models, including AI filetype detection with filetype detection, opt-in AI table, equation, and figure extraction, efficient token compression, automatic image encoding, reranking for lost-in-the-middle effects, and more, all pre-built to work out-of-the-box.

Demo

Local Installation 🛠️

If you do not wish to use our API, you are welcome host The Pipe for yourself locally. If you choose to do this, you must install a number of dependencies for the code to function correctly, some of which may incur compute costs and/or require a GPU for reasonable performance. Additional installed dependencies are required: pytorch, universal-ctags, playwright, pytesseract, llmlingua, moviepy, and pytube. This installation process will depend on your system and compute capabilities. After installing them, follow these steps for a local setup:

Arguments are:

  • source (required): can be a file path, a URL, or a directory path.
  • local (optional): Use the local version of The Pipe instead of the hosted API.
  • match (optional): Substring to match files in the directory. Regex is not yet supported.
  • ignore (optional): Substring to ignore files in the directory. Regex is not yet supported.
  • limit (optional): The token limit for the output prompt, defaults to 100K. Prompts exceeding the limit will be compressed. This may not work as expected with the API, as it is in active development.
  • ai_extraction (optional): Extract tables, figures, and math from PDFs using our extractor. Incurs extra costs.
  • text_only (optional): Do not extract images from documents or websites. Additionally, image files will be represented with OCR instead of as images.

Sponsors

Book us with Cal.com

Thank you to Cal.com for sponsoring this project. Contact emmett@thepi.pe for sponsorship information.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

thepipe_api-0.3.4.tar.gz (21.9 kB view details)

Uploaded Source

Built Distribution

thepipe_api-0.3.4-py3-none-any.whl (19.9 kB view details)

Uploaded Python 3

File details

Details for the file thepipe_api-0.3.4.tar.gz.

File metadata

  • Download URL: thepipe_api-0.3.4.tar.gz
  • Upload date:
  • Size: 21.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.8.0 pkginfo/1.9.6 readme-renderer/37.3 requests/2.31.0 requests-toolbelt/0.10.1 urllib3/1.26.18 tqdm/4.66.2 importlib-metadata/6.11.0 keyring/24.3.0 rfc3986/1.5.0 colorama/0.4.6 CPython/3.10.8

File hashes

Hashes for thepipe_api-0.3.4.tar.gz
Algorithm Hash digest
SHA256 36bbfc398d882cd256ec14bb4bd276727089bb5fb662a2df811bc86ac90077cb
MD5 a4d69c1651fd619542f2a144ff3e7aea
BLAKE2b-256 d6d4593ddc5326963b7c9973385f2817c1cf2dfec81a6d41ff9eddec44d77e86

See more details on using hashes here.

Provenance

File details

Details for the file thepipe_api-0.3.4-py3-none-any.whl.

File metadata

  • Download URL: thepipe_api-0.3.4-py3-none-any.whl
  • Upload date:
  • Size: 19.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.8.0 pkginfo/1.9.6 readme-renderer/37.3 requests/2.31.0 requests-toolbelt/0.10.1 urllib3/1.26.18 tqdm/4.66.2 importlib-metadata/6.11.0 keyring/24.3.0 rfc3986/1.5.0 colorama/0.4.6 CPython/3.10.8

File hashes

Hashes for thepipe_api-0.3.4-py3-none-any.whl
Algorithm Hash digest
SHA256 81bbcf38ca3e44ea55fe85437e020e685ccbc5dbb78d49c2ac95b7c564480027
MD5 8b9a6a42e090c7210196ca5b34bd475b
BLAKE2b-256 7f925d4be4348e91cd24110637da602d40542207b462597b7806b7924fadb53a

See more details on using hashes here.

Provenance

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page