A NLP library that simplifies pattern finding in strings
Project description
Simple pattern finding in strings and natural language processing.
There has been an error identified within the toolkit-related functions, which will be fixed during the next update.
Features
Automatic pattern detection ( semantic and literal )
Multiple parsers ( implementations of nltk, pattern, and nlpnet )
Keyword searching to find specific phrases
Roadmap
Some features I plan to implement in the future:
Machine Learning. This will allow the parsers to learn multiple grammatical “styles” and be able to successfully parse a much wider selection of strings
Additional parsers
Continued improvements on the current parsers. This will be a long-term project, and additional details can be found on the main Github page
If you have feature requests, feel free to add an issue to the Github issue tracker. All contributions and requests are appreciated!
Usage
regex4dummies is very easy to use. Simply import the library, get some strings, and compare them!
from regex4dummies import regex4dummies
# Creating strings
strings = [ "This is the first test string.", "This is the second test string." ]
regex = regex4dummies()
# Identifying literal patterns in strings
print regex.compare_strings( 'default', True, strings )
# Identifying semantic patterns in strings using the nltk parser
print regex.compare_strings( 'nltk', False, strings )
Above is regex4dummies in its simplest form. It allows for additional features as well, including:
# Display the version of regex4dummies you are using
print regex.__version__
# To use the other parsers, replace the above line of code with either of the following:
# print regex.compare_strings( 'pattern', False, strings )
# print regex.compare_strings( 'nlpnet', False, strings )
# To call all of the parsers, replace the above line of code with the following:
# print regex.compare_strings( '', False, strings )
# To get the topics of the strings, call the get_pattern_topics function
print regex.get_pattern_topics()
# Printing pattern information
pattern_information = regex.get_sentence_information()
for objects in pattern_information:
print "[ Pattern ] : " + objects.pattern
print "[ Subject ] : " + objects.subject
print "[ Verb ] : " + objects.verb
print "[ Object ] : " + objects.object[0]
print "[ Prep Phrases ] : " + str( objects.prepositional_phrases )
print "[ Reliability Score ] : " + str( objects.reliability_score )
print "[ Applicability Score ] : " + str( objects.applicability_score )
print ""
A newly released set of features include a tokenizer and dependency finder function. To use them, simply give a string as the first parameter, and the name of the parser you would like to use as the second for both functions.
# Testing the toolkit functions
tool_tester = Toolkit()
# Testing the tokenizer functions
print tool_tester.tokenize( "This is a test string.", "pattern" )
# Testing the dependency functions
print tool_tester.find_dependencies( "This is a test string.", "pattern" )
Installation
To install this library, use pip.
$ pip install regex4dummies
In addition to the library, wget is a required command-line command to use the nlpnet parser. If you do not have wget or cannot get it, follow the below directions to still get the functionality of the nlpnet parser.
Instructions to install the required dependency for nlpnet:
Download the nlpnet_dependency file on the most recent release found in Github ( please not, when uncompressed, this file is over 350 MB large ).
Place this directory into the same directory that nltk-data is located ( if you don’t have that installed, just run the library and go through the GUI downloader )
That’s it! The nlpnet parser should now be able to be used.
Patch Notes
( Latest ) v1.4.0: Feature additions and some minor updates
Two new functions have been added! The tokenizer function will return the tokenized version of the string passed using the given parser ( pattern, nlpnet, or nltk ). The dependency function will return the string with the added dependency information using a specific parser( pattern, nltk, or nlpnet )
FuzzyWuzzy is the new applicability score. In the past, it has been pretty hard to find a good way to implement the applicability score, and this has been reflected in the quality of the score. Hopefully, this update will greatly improve the accuracy of this score
Small topic finder update
Contributing
Contributors are welcome and much needed! regex4dummies is still under heavy development, and needs all of the help it can get. If you have any feature ideas, feel free to create an issue on the github repository ( https://github.com/darkmattervale/regex4dummies/issues ) or fork the repository and create your addition.
Any help you can give is much appreciated. The more help we get, the better regex4dummies will perform. Thanks for contributing!
License
Please see LICENSE.txt for information about the MIT license
Citations
nlpnet:
Fonseca, E. R. and Rosa, J.L.G. Mac-Morpho Revisited: Towards Robust Part-of-Speech Tagging. Proceedings of the 9th Brazilian Symposium in Information and Human Language Technology, 2013. p. 98-107 [PDF]
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for regex4dummies-1.4.0-py2.py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 2a00430841bb4ec0bd70e48c9049ad488e77ac96b2a300ca3be038df9f3e637b |
|
MD5 | 565e41330fc9237aefaebcff1a4067bb |
|
BLAKE2b-256 | 78cf52b54689e8f1da7d56005c8058d93706fcf9456f7987c23653d8d217e12c |