Fast PaddleOCR MCP server - Extract text from images using PaddleOCR with optimized performance
Project description
PaddleOCR-MCP
PaddleOCR MCP (Model Context Protocol) server and CLI tool that extracts text from images and outputs results in markdown format. Optimized for fast inference with GPU auto-detection.
MCP Server Configuration
The MCP (Model Context Protocol) server allows integration with MCP clients like Cursor, Claude Desktop, etc.
Use uvx directly (no installation required, automatically downloads from PyPI):
{
"mcpServers": {
"fast-paddleocr-mcp": {
"command": "uvx",
"args": ["fast-paddleocr-mcp"]
}
}
}
MCP Tool: ocr_image
The server provides a single tool called ocr_image that:
- Input:
image_path(string) - Path to the input image file - Output: Returns the path to the generated markdown file containing OCR results
Example: When called with image_path: "photo.png", it returns "photo.png.md" containing the recognized text.
See MCP_README.md for detailed MCP server documentation.
Usage
Basic Usage
The tool is optimized for speed by default with the following settings:
- Fast mode enabled (disables preprocessing for maximum speed)
- PP-OCRv4 (faster mobile models)
- 640px image size limit (faster processing)
- Auto GPU detection (uses GPU if available, falls back to CPU)
# Output will be saved as <image_name>.png.md
# Uses: fast mode + PP-OCRv4 + 640px + auto GPU detection
uvx --from . paddleocr-md image.png
# Specify custom output path
uvx --from . paddleocr-md image.png -o result.md
# Force CPU mode
uvx --from . paddleocr-md image.png --cpu
# Disable fast mode for better accuracy on rotated text
uvx --from . paddleocr-md image.png --no-fast
# Use PP-OCRv5 for better accuracy (slower)
uvx --from . paddleocr-md image.png --ocr-version PP-OCRv5
Default Optimization Settings
The tool is optimized for speed by default with these settings:
- ✅ Fast mode enabled: Disables textline orientation classification (skips one model)
- ✅ PP-OCRv4: Uses faster mobile models (PP-OCRv4_mobile_det, PP-OCRv4_mobile_rec)
- ✅ 640px image size limit: Faster processing (vs default 960px)
- ✅ Auto GPU detection: Automatically uses GPU if available, falls back to CPU
- ✅ Document preprocessing disabled: Skips unnecessary preprocessing steps
Customization Options
-
--no-fast: Disable fast mode for better accuracy- Enables textline orientation classification
- Better accuracy on rotated text, but slower
-
--cpu: Force CPU mode- Overrides auto GPU detection
- Explicitly use CPU
-
--gpu: Force GPU mode- Will fail if GPU not available
- Use when you want to ensure GPU usage
-
--ocr-version PP-OCRv5: Use better accuracy version- PP-OCRv5 has better accuracy but slower than PP-OCRv4 (default)
- Uses server models
-
--max-size <pixels>: Adjust image processing size- Default: 640px
- Larger values (e.g., 960, 1280) = better accuracy, slower
- Smaller values (e.g., 480) = faster, may reduce accuracy
-
--hpi: High-Performance Inference- Automatically selects best inference backend (Paddle Inference, OpenVINO, ONNX Runtime, TensorRT)
- Requires HPI dependencies:
paddleocr install_hpi_deps cpu/gpu - Best performance but requires additional setup
Examples
# Basic usage (uses all optimizations by default: fast + PP-OCRv4 + 640px + auto GPU)
uvx --from . paddleocr-md photo.jpg
# Process with custom output
uvx --from . paddleocr-md document.png -o extracted_text.md
# Better accuracy (slower) - disable fast mode and use PP-OCRv5
uvx --from . paddleocr-md image.png --no-fast --ocr-version PP-OCRv5 --max-size 960
# Force CPU mode
uvx --from . paddleocr-md image.png --cpu
# Use High-Performance Inference (requires HPI dependencies)
uvx --from . paddleocr-md image.png --hpi
Output Format
The tool generates a markdown file containing:
- Source image path
- List of detected text (one per line)
Example output (test_image.png.md):
# OCR Result
**Source Image:** `test_image.png`
---
- HelloPaddleOcR
- 10000C
License
MIT
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file fast_paddleocr_mcp-0.2.2.tar.gz.
File metadata
- Download URL: fast_paddleocr_mcp-0.2.2.tar.gz
- Upload date:
- Size: 10.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
f19d8268058db427cfd6954ea5bdc94c1ef6f276667ddb894b1880e2ed8d7ce9
|
|
| MD5 |
a9268c67285b5bf2c7f6fdfd3bb46d21
|
|
| BLAKE2b-256 |
5e66449e1455eb3fef70abe7f3edb2caec95cd647d2f08a272a17f65fb5a0338
|
File details
Details for the file fast_paddleocr_mcp-0.2.2-py3-none-any.whl.
File metadata
- Download URL: fast_paddleocr_mcp-0.2.2-py3-none-any.whl
- Upload date:
- Size: 9.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
1cbc95462d9420eb90629e2dce4b58d06688832fc46d970a0b039fb369e38077
|
|
| MD5 |
837441ccdc2cbab56d3b85265ba1a1a0
|
|
| BLAKE2b-256 |
f3b7e1197049d0dd6000c985d7d198ef60ddd190883c99cf03bbd21cb112b04c
|