Skip to main content

Optical character recognition model for Tensorflow based on Visual Attention.

Project description

Visual attention-based OCR model for image recognition with additional tools for creating TFRecords datasets and exporting the trained model with weights as a SavedModel or a frozen graph.

Acknowledgements

This project is based on a model by Qi Guo and Yuntian Deng. You can find the original model in the da03/Attention-OCR repository.

The model

Authors: Qi Guo and Yuntian Deng.

The model first runs a sliding CNN on the image (images are resized to height 32 while preserving aspect ratio). Then an LSTM is stacked on top of the CNN. Finally, an attention model is used as a decoder for producing the final outputs.

OCR example

OCR example

Installation

pip install aocr

Note: Tensorflow 1.2 and Numpy will be installed as dependencies. Additional dependencies are PIL/Pillow, distance, and six.

Usage

Create a dataset

To build a TFRecords dataset, you need a collection of images and an annotation file with their respective labels.

aocr dataset datasets/annotations-training.txt datasets/training.tfrecords
aocr dataset datasets/annotations-testing.txt datasets/testing.tfrecords

Annotations are simple text files containing the image paths (either absolute or relative to your working dir) and their corresponding labels:

datasets/images/hello.jpg hello
datasets/images/world.jpg world

Train

aocr train datasets/training.tfrecords

A new model will be created, and the training will start. Note that it takes quite a long time to reach convergence, since we are training the CNN and attention model simultaneously.

The --steps-per-checkpoint parameter determines how often the model checkpoints will be saved (the default output dir is checkpoints/).

Important: there is a lot of available training options. See the CLI help or the parameters section of this README.

Test and visualize

aocr test datasets/testing.tfrecords

Additionally, you can visualize the attention results during testing (saved to results/ by default):

aocr test --visualize datasets/testing.tfrecords

Example output images in results/correct:

Image 0 (j/j):

example image 0

example image 0

Image 1 (u/u):

example image 1

example image 1

Image 2 (n/n):

example image 2

example image 2

Image 3 (g/g):

example image 3

example image 3

Image 4 (l/l):

example image 4

example image 4

Image 5 (e/e):

example image 5

example image 5

Export

aocr export exported-model

Load weights from the latest checkpoints and export the model into the ./exported-model directory.

Google Cloud ML Engine

To train the model in the Google Cloud Machine Learning Engine, upload the training dataset into a Google Cloud Storage bucket and start a training job with the gcloud tool.

  1. Set the environment variables:

# Prefix for the job name.
export JOB_PREFIX="aocr"

# Region to launch the training job in.
# Should be the same as the storage bucket region.
export REGION="us-central1"

# Your storage bucket.
export GS_BUCKET="gs://aocr-bucket"

# Path to store your training dataset in the bucket.
export DATASET_UPLOAD_PATH="training.tfrecords"
  1. Upload the training dataset:

gsutil cp datasets/training.tfrecords $GS_BUCKET/$DATASET_UPLOAD_PATH
  1. Launch the ML Engine job:

export NOW=$(date +"%Y%m%d_%H%M%S")
export JOB_NAME="$JOB_PREFIX$NOW"
export JOB_DIR="$GS_BUCKET/$JOB_NAME"

gcloud ml-engine jobs submit training $JOB_NAME \
    --job-dir=$JOB_DIR \
    --module-name=aocr \
    --package-path=aocr \
    --region=$REGION \
    --scale-tier=BASIC_GPU \
    --runtime-version=1.2 \
    -- \
    --model-dir=$JOB_DIR \
    train $GS_BUCKET/$DATASET_UPLOAD_PATH \
    --steps-per-checkpoint=500 \
    --batch-size=512 \
    --num-epoch=20

Parameters

Global

  • log-path: Path for the log file.

Testing

  • visualize: Output the attention maps on the original image.

Exporting

  • format: Format for the export (either savedmodel or frozengraph).

Training

  • steps-per-checkpoint: Checkpointing (print perplexity, save model) per how many steps

  • num-epoch: The number of whole data passes.

  • batch-size: Batch size.

  • initial-learning-rate: Initial learning rate, note the we use AdaDelta, so the initial value does not matter much.

  • target-embedding-size: Embedding dimension for each target.

  • attn-use-lstm: Whether or not use LSTM attention decoder cell.

  • attn-num-hidden: Number of hidden units in attention decoder cell.

  • attn-num-layers: Number of layers in attention decoder cell. (Encoder number of hidden units will be attn-num-hidden*attn-num-layers).

  • target-vocab-size: Target vocabulary size. Default is = 26+10+3 # 0: PADDING, 1: GO, 2: EOS, >2: 0-9, a-z

  • no-resume: Create new weights even if there are checkpoints present.

  • max-gradient-norm: Clip gradients to this norm.

  • no-gradient-clipping: Do not perform gradient clipping.

  • gpu-id: GPU to use.

  • use-gru: Use GRU cells.

References

Convert a formula to its LaTex source

What You Get Is What You See: A Visual Markup Decompiler

Torch attention OCR

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

aocr-0.0.3.tar.gz (29.5 kB view details)

Uploaded Source

File details

Details for the file aocr-0.0.3.tar.gz.

File metadata

  • Download URL: aocr-0.0.3.tar.gz
  • Upload date:
  • Size: 29.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No

File hashes

Hashes for aocr-0.0.3.tar.gz
Algorithm Hash digest
SHA256 5eff55bf8ef390a59b4dec461a199bf396059b4613ed9ae46514e315afefe00e
MD5 7217af105da0c5f687e4aa0bcdfb7962
BLAKE2b-256 672ecf86ad88b59926c84040502bda6582d7c7085a53e39171f092ac3ad2b4b4

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page