Skip to main content

Automatically process scanned forms

Project description

fast-form

PyPI Latest Release

Project from UniHack hackathon, helps to extract data from forms.

Installation

To use first install anaconda on your computer https://www.anaconda.com/ then open anaconda terminal and run:

conda create -n fast-form python=3.7.10
conda activate fast-form
pip install fast-form

if the installation was successful you can start using the tool.

first, you can test if the CLI works correctly, by running

fast-form --version

Run

To run fast-form on your data, first, initialize data folder.

fast-form init

This will create a folder containing test data and all the config files.

fast-form-data
│   config.json    
│   path_config.json
│   template.pdf
└───documents
    │   document.jpg
    │   ...

All you have to do is replace files in documents folder with scanned pdfs you want to process. To test it out, you can run it as is. Simply call following command without changing the directory:

fast-form extract

This will create validation_excel.xlsx, which you can open in office suite and manually fix all the results that were not recognized correctly.

  • -1 is used for answers that were recognized as having multiple fields crossed
  • -2 in case no crossed filed was recognized You can replace those, in case picture recognition did a poor job.

Once you are happy with the result, save the validation_excel and run

fast-form finalize

This will compile validated data in to format "one questionare per line" and append it in to excel configured in final_excel_path in path_config.json.

Advanced configuration

To configure fast-form for custom questionare, one has to provide

  • template - empty questionare with perfect quality. Use original pdf file used for printing the questionares
  • field configuration - description of all the fields in the questionare and their exact position on the page those has to be done manually for the time being and there is no visual tool to do it. We have jupyter notebook for this in repository for this task, but workflow is far from smooth for the time being.

Configure path to these files in path_config.json

{
    "template_path": "/pathtotemplate/template.jpg",
    "form_structure_path": "/pathtoconfig/config.json",
    "folder_with_documents_path": "/pathtodocumentfolder/documents",
    "final_excel_path": "/pathtofinalexcel/final_excel.xlsx"
}

Development

Before development please run make setup Installing new dependencies added on remote

pip-sync

Adding dependencies:

add line to requirements.in and then run

pip-compile

Retrain model

We are using preparsed emnist dataset. https://www.kaggle.com/crawford/emnist

To retrain the model, you need to download it to training_data folder

fast-form
 - training_data
    - emnist-balanced-mapping.txt
    - emnist-balanced-train.csv
    - emnist-balanced-test.csv
    ...
 ...

and then first run field_recognizer/enrich_dataset.ipynb in jupyter notebook and after that use field_recognizer/emnist-neural-network.ipynb twice, once for type numbers, once for type letters (see inside of notebook )

Example form configuration - JSON

tests/form_for_test/config.json

Road map

priorities till 15.11.2020

Character Recognition

  • retrain model on capital letters only
  • enrich emnist dataset by different letter positioning
  • use this notebook to improve the model https://www.kaggle.com/tarunkr/digit-recognition-tutorial-cnn-99-67-accuracy
  • Add model for numbers
  • Solve czech characters, either
    • find different dataset with czech diacritics
    • enrich emnist dataset by letters with diacritics
    • Add second clasifier, that would recognize diacritics / accute(') / carron(ˇ) / none /
    • Create our own, small dataset (maybe adding the diacritics to current dataset) and use methods here: https://arxiv.org/pdf/1904.08095.pdf
  • (depends on something in section Model results processing) Create a model distinguishing X or filled circle or nothing or anomaly. Improve true/false computation for boxes

Refactoring

  • Restructure repository to separate code from data
  • Refactor to original emnist dataset structure instead of preparsed csv data from kaggle

Preprocessing

  • smarter thresholding
  • smarter corner classification
  • completely drop corners, orient page based on the original file provided (fitting text on text)
  • Improve just config so it centers boxes better and reponse 159 in PID1 is correctly parsed
  • Change code to handle multiple pages somehow
  • Better noise cancellation on the whole file
  • Better noise cancellation on the level of the cropped boxes (important to do before cropping to the symbol area as it is broken by noise at the moment sometimes) many options how to approach, choose any
  • Possibly (but maybe hard to do): remove bounding lines from figures.
  • Speedup: the current matching teplate logic is simly tooo slow. Think what to do
  • use other feature extraction method than SIFT (I have tried ORB which did not work well enough, but maybe after some tuning it would. it is way faster)
  • do whathever else

Model results processing

  • Implement logic around single choice forms return only single choice if there is only 1 true choice found, otherwise -1 (if no question answered) or -2 (if more than one question answered).
  • Add to single choice logic also logic for filled in and anomaly circles.

Specific Questionnaire processing

  • Prepare empty template that will have only header and some repeating thigs for PID, test how it works, potentially alter PID a bit, so it works for such empty template
  • Prepare at the top of the first page space where one can fill in the patient code and add this to config as only number field

Configuration

  • location of text via top left corner, bottom right corner plus number of letters
  • Create a tool for config file creation

Packaging, proper output

  • prepare for pdf with multiple questionnaires
  • create config for the current questionnaire we are working on
  • For the name of the patient use the name from questionnaire of question called patient_id. In case such question is not present, use the name of the pdf (as it is done now)
  • Prepare python script that will process the scanned pdfs:

The program should get the following parameters from config file that will have to be existing in the folder:

- folder with scanned files (how should be new questionnaire recongized? new file for each?
- path to template document as pdf or jpg or something
- config file with location of answers in the template document via coordinates in pixels (see configuration above)

And it should output:

- excel file with the answers and their accuracies and also the parts of the original scan that were used for the recognition. 
  • Prepare python script that will process the excel document with all the answers into an excel sheet. One questionnaire per line. params:

    • path to the excel with answers
    • path to the excel to put it in (if none, create new)
    • name of the sheet (if none create new) if the sheet is non empty put the data below the data already there
  • Write readme

  • Write nicer readme

  • Create from this a package on pypi

  • Test and especially let someone test on mac

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

fast-form-0.3.0.tar.gz (11.8 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

fast_form-0.3.0-py3-none-any.whl (11.8 MB view details)

Uploaded Python 3

File details

Details for the file fast-form-0.3.0.tar.gz.

File metadata

  • Download URL: fast-form-0.3.0.tar.gz
  • Upload date:
  • Size: 11.8 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.3.0 pkginfo/1.7.0 requests/2.25.1 setuptools/47.1.0 requests-toolbelt/0.9.1 tqdm/4.58.0 CPython/3.7.9

File hashes

Hashes for fast-form-0.3.0.tar.gz
Algorithm Hash digest
SHA256 82a492ff37ffd5095539a3cf577813b8633c03a5bbff9bf2aab6ef1d1bab0036
MD5 c39c431498841c607d6efee6e3e8103f
BLAKE2b-256 4c82dc5be4285fff2636a241c065730e74c38542f6d52e8eb771eb8626ae8122

See more details on using hashes here.

File details

Details for the file fast_form-0.3.0-py3-none-any.whl.

File metadata

  • Download URL: fast_form-0.3.0-py3-none-any.whl
  • Upload date:
  • Size: 11.8 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.3.0 pkginfo/1.7.0 requests/2.25.1 setuptools/47.1.0 requests-toolbelt/0.9.1 tqdm/4.58.0 CPython/3.7.9

File hashes

Hashes for fast_form-0.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 bb63f192d14bf6b081fc6b1cfb19508b585677a6bebc40464d0d678217a8da25
MD5 ca2e35df8e89af697c7d69ca86f813ff
BLAKE2b-256 bbbd56a8affba1512fa2ab35bad0982d9000a58a203d5e9b1290eb371f09d126

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page