Skip to main content

PySin is a toolbox for text retrieval in unstructured documents datasets. It contains both a multi-type text extractor and a search engine. To test them, you can use the medical prescriptions generator that is also provided.

Project description

PySin

PySin is a toolbox for text retrieval in unstructured documents datasets. It contains both a multi-type text extractor and a search engine. To test them, you can use the medical documents generator that is also provided.

OS Dependencies

You will need geckodriver to run the generator. Download it and copy it to your PATH (eg: /usr/local/bin)

Debian, Ubuntu, and friends

sudo apt-get install build-essential libpoppler-cpp-dev pkg-config python-dev

Fedora, Red Hat, and friends

sudo yum install gcc-c++ pkgconfig poppler-cpp-devel python-devel redhat-rpm-config

macOS

brew install pkg-config poppler

Conda users may also need libgcc:

conda install -c anaconda libgcc

Windows

Currently tested only when using conda:

  • Install the Microsoft Visual C++ Build Tools
  • Install poppler through conda:
conda install -c conda-forge poppler

Install

pip install pysin

Search engine

Arguments

The function search takes 5 arguments.

Positionnal arguments :

  • query : your query
  • input_path : the path to the directory to search in
  • output_path : the path to the directory to put the results in

Keyword arguments:

  • scale : can take the values row or doc depending on if the query should be satisfied by a single row or by a whole document. The row scale is more precise whereas the doc scale is faster. The scale defaults to row.
  • update_cache : True to update the cached files (for example if some files have been added to the folder since the last search), else False. Defaults to True. If you're working with a huge amount of data that doesn't change, you should set update_cache to False.

To search the word 'word' within the files of the folder 'path/to/data/' by writing the results in the folder '/path/to/results/', just run the following command :

from pysin import search
search('word', 'path/to/data/', 'path/to/results/')

Queries

To search one word beyond multiple ones, just write them side to side in the query.

search('word1 word2 word3', 'path/to/data/', 'path/to/results/')

To search the files where 'mandatory' is and where 'foo' or 'bar' is also (but not necessarily both at the same time), just type the following command :

search('+mandatory foo bar', 'path/to/data/', 'path/to/results/', scale='doc')

The same query holds for the row scale. The previous command might return a document that contains 'mandatory' at the first row and 'foo' at the last one whereas in the row scale, only the occurrences where 'mandatory' AND 'foo' (and/or 'bar') are in the same row are returned.

To search the rows where 'mandatory' is but 'forbidden' isn't, type the following command :

search('mandatory -forbidden', 'path/to/data/', 'path/to/results/')

To search an expression with several words, use quotes :

search('"complex expression"', 'path/to/data/', 'path/to/results/')

You can obviously combine everything into a single query :

search('+mandatory choice1 choic2 "choice3" -"not this one" +"another mandatory"', 'path/to/data/', 'path/to/results/')

Results

When a research is launched, a folder is made at output_path in which are two files :

  • results.csv : in row scale, one row correspond to one occurrence and contains the path to the file, the occurrence row number and the context of the occurrence. In doc scale, there are only the paths to the corresponding files.
  • folders.json : returns the number of occurrences in each folder using a tree structure

Extractor

The extractor preprocesses all the files to enable the research by converting the handled files into txt cached files. The handled types are csv, doc, docx, html, md, pdf, rtf, txt, xml.

To extract all the files within a folder at path 'path', just run :

extract('path/to/data')

To erase all the cached files, just run :

reset_cache('path/to/data')

Medical prescriptions generator

The generator is based on the data of the faker module. It can generate both medical prescriptions and medical report. To generate 19 fake medical documents in the folder 'data', just run the following command :

generate(19, 'path/to/data')

Soft mode

The search engine and the extractor can also by used as softs. For the search engine, just run the following command :

$ python src/search.py +mandatory choice1 choic2 "choice3" -"not this one" +"another mandatory" --input_path path/to/data/ --output_path path/to/results/

To search at the doc scale, just add the argument --d.

The extractor can be used like this :

$ python src/extractor.py path/to/data/

To clear the cached files, just add the argument --reset :

$ python src/extractor.py --reset path/to/data/

Trick

If you have to do lots of researchs in one folder, let's say absolute/path/to/data/, by putting the results always in the same folder, let's say absolute/path/to/results/, and always at the same scale, let's say the row one, then you can create a shortcut to search more easily by running the following commands :

$ echo alias search=\'python /absolute/path/to/search.py --input_path /absolute/path/to/data/ --output_path /absolute/path/to/results\' >> ~/.bashrc
$ source ~/.bashrc

Then, you're able to do a research from any location by typing :

$ search +mandatory choice1 choic2 "choice3" -"not this one" +"another mandatory"

WARNING : before doing this, make sure that the search alias doesn't exist yet, for example by running the command search and checking that shell returns the following message :

ModuleNotFoundError: No module named 'apt_pkg'

Example

You can test this module using the example.py script.

TODO

  • multithreaded research
  • improve medication notation
  • new document types
  • adapt .doc extraction to windows environment

Publish

First, you need to have twine installedd

pip install --user --upgrade twine

Make sure you have bumped the version number in setup.py, then run the following:

python setup.py sdist bdist_wheel
python -m twine upload dist/*

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pysin-1.5.2.tar.gz (190.1 kB view details)

Uploaded Source

Built Distribution

pysin-1.5.2-py3-none-any.whl (191.5 kB view details)

Uploaded Python 3

File details

Details for the file pysin-1.5.2.tar.gz.

File metadata

  • Download URL: pysin-1.5.2.tar.gz
  • Upload date:
  • Size: 190.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.1.1 pkginfo/1.5.0.1 requests/2.23.0 setuptools/42.0.0 requests-toolbelt/0.9.1 tqdm/4.36.1 CPython/3.7.0

File hashes

Hashes for pysin-1.5.2.tar.gz
Algorithm Hash digest
SHA256 ca55da22a5188ae76768f53fa99845f61961c14860b659028ee5a87a73d3fe59
MD5 fc3da06f16af71b77af69fda1cabed10
BLAKE2b-256 30e64ddda68fd9ae42e9f419858946b135057b2b17fa9274702ef1cdfb8f77b5

See more details on using hashes here.

File details

Details for the file pysin-1.5.2-py3-none-any.whl.

File metadata

  • Download URL: pysin-1.5.2-py3-none-any.whl
  • Upload date:
  • Size: 191.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.1.1 pkginfo/1.5.0.1 requests/2.23.0 setuptools/42.0.0 requests-toolbelt/0.9.1 tqdm/4.36.1 CPython/3.7.0

File hashes

Hashes for pysin-1.5.2-py3-none-any.whl
Algorithm Hash digest
SHA256 189447775ae9dd7e4cf158e468d351f59cf54c43eca8995373b81fa5e76063bd
MD5 c042bd438cd3e6b748a54af4e1b36d87
BLAKE2b-256 b9ea13907b39ef782a45a1ecca4a3e7787a7b4b29c64cd29b1c4d8eacc5a7191

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page