Parse natural language time expressions in python
Project description
ctparse - Parse natural language time expressions in python
Free software: MIT license
Documentation: https://ctparse.readthedocs.io.
Background
The package ctparse is a pure python package to parse time expressions from natural language (i.e. strings). In many ways it builds on similar concepts as Facebook’s duckling package (https://github.com/facebook/duckling). However, for the time being it only targets times and only German and English text.
In principle ctparse can be used to detect time expressions in a text, however its main use case is the semantic interpretation of such expressions. Detecting time expressions in the first place can - to our experience - be done more efficiently (and precisely) using e.g. CRFs or other models targeted at this specific task.
ctparse is designed with the use case in mind where interpretation of time expressions is done under the following assumptions:
All expressions are relative to some pre-defined reference times
Unless explicitly specified in the time expression, valid resolutions are in the future relative to the reference time (i.e. 12.5. will be the next 12th of May, but 12.5.2012 should correctly resolve to the 12th of May 2012).
If in doubt, resolutions in the near future are more likely than resolutions in the far future (not implemented yet, but any resolution more than i.e. 3 month in the future is extremely unlikely).
The specific comtravo use-case is resolving time expressions in booking requests which almost always refer to some point in time within the next 4-8 weeks.
ctparse currently is language agnostic and supports German and English expressions. This might get an extension in the future. The main reason is that in real world communication more often than not people write in one language (their business language) but use constructs to express times that are based on their mother tongue and/or what they believe to be the way to express dates in the target language. This leads to text in German with English time expressions and vice-versa. Using a language detection upfront on the complete original text is for obvious no solution - rather it would make the problem worse.
Example
from ctparse import ctparse
from datetime import datetime
# Set reference time
ts = datetime(2018, 3, 12, 14, 30)
ctparse('May 5th 2:30 in the afternoon', ts=ts)
This should return a Time object represented as Time[0-29]{2018-05-05 14:30 (X/X)}, indicating that characters 0-29 were used in the resolution, that the resolved date time is the 5th of May 2018 at 14:30 and that this resolution is neither based on a day of week (first X) nor a part of day (second X).
Latent time
Normally, ctparse will anchor time expressions to the reference time. For example, when parsing the time expression 8:00 pm, ctparse will resolve the expression to 8 pm after the reference time as follows
parse = ctparse("8:00 pm", ts=datetime(2020, 1, 1, 7, 0), latent_time=True) # default
# parse.resolution -> Time(2020, 1, 1, 20, 00)
This behavior can be customized using the option latent_time=False, which will return a time resolution not anchored to a particular date
parse = ctparse("8:00 pm", ts=datetime(2020, 1, 1, 7, 0), latent_time=False)
# parse.resolution -> Time(None, None, None, 20, 00)
Implementation
ctparse - as duckling - is a mixture of a rule and regular expression based system + some probabilistic modeling. In this sense it resembles a PCFG.
Rules
At the core ctparse is a collection of production rules over sequences of regular expressions and (intermediate) productions.
Productions are either of type Time, Interval or Duration and can have certain predicates (e.g. whether a Time is a part of day like 'afternoon').
A typical rule than looks like this:
@rule(predicate('isDate'), dimension(Interval))
I.e. this rule is applicable when the intermediate production resulted in something that has a date, followed by something that is in interval (like e.g. in 'May 5th 9-10').
The actual production is a python function with the following signature:
@rule(predicate('isDate'), dimension(Interval))
def ruleDateInterval(ts, d, i):
"""
param ts: datetime - the current refenrence time
d: Time - a time that contains at least a full date
i: Interval - some Interval
"""
if not (i.t_from.isTOD and i.t_to.isTOD):
return None
return Interval(
t_from=Time(year=d.year, month=d.month, day=d.day,
hour=i.t_from.hour, minute=i.t_from.minute),
t_to=Time(year=d.year, month=d.month, day=d.day,
hour=i.t_to.hour, minute=i.t_to.minute))
This production will return a new interval at the date of predicate('isDate') spanning the time coded in dimension(Interval). If the latter does code for something else than a time of day (TOD), no production is returned, e.g. the rule matched but failed.
Technical Background
Some observations on the problem:
Each rule is a combination of regular expressions and productions.
Consequently, each production must originate in a sequence of regular expressions that must have matched (parts of) the text.
Hence, only subsequence of all regular expressions in all rules can lead to a successful production.
To this end the algorithm proceeds as follows:
Input a string and a reference time
Find all matches of all regular expressions from all rules in the input strings. Each regular expression is assigned an identifier.
Find all distinct sequences of these matches where two matches do not overlap nor have a gap inbetween
To each such subsequence apply all rules at all possible positions until no further rules can be applied - in which case one solution is produced
Obviously, not all sequences of matching expressions and not all sequences of rules applied on top lead to meaningful results. Here the PCFG kicks in:
Based on example data (corpus.py) a model is calibrated to predict how likely a production is to lead to a/the correct result. Instead of doing a breadth first search, the most promising productions are applied first.
Resolutions are produced until there are no more resolutions or a timeout is hit.
Based on the same model from all resolutions the highest scoring is returned.
Credits
This package was created with Cookiecutter and the audreyr/cookiecutter-pypackage project template.
History
0.3.4 (2022-07-28)
Add fuzzy matching on longer literals
[internal] De-tangle corpus tests into isolated test cases
Allow spaces around separators in ruleDDMMYYYY and ruleYYYYMMDD
0.3.3 (2022-07-18)
Add rule for straight forward US formatted dates (ruleYYYYMMDD)
Added rule ruleYearMonth
Added corpus cases for some open issues that now pass
Changed all internal imports to be absolute (i.e. from ctparse.x instead of from .x)
Dropped tox (now using github actions)
0.3.2 (2022-07-18)
Drop support for python 3.6, update dev requirements
0.3.1 (2021-07-07)
Add support for python 3.9 on travis and in manifest; update build config
0.3.0 (2021-02-01)
Removed latent rules regarding times (latent rules regarding dates are still present)
Added latent_time option to customize the new behavior, defauld behavior is backwards-compatible
0.2.1 (2020-05-27)
Update development dependencies
Add flake8-bugbear and fixed issues
0.2.0 (2020-04-23)
Implemented new type Duration, to handle lengths of time
Adapted the dataset to include Duration
Implemented basic rule to merge Duration, Time and Interval in simple cases.
Created a make target to train the model make train
0.1.0 (2020-03-20)
Major refactor of code underlying predictive model
Based on a contribution from @bharathi-srini: replace naive bayes from sklearn by own implementation
Thus remove dependencies on numpy, scipy, scikit-learn
Predictions are much faster: 97/s in the old vs. 239/s in the new code base
Performance identical
Deprecate support for python 3.5, add 3.8
Add more strict type checking rules (mypy.ini)
Force black code formatting, make this a linter step, “black” all code
0.0.47 (2020-02-28)
Allow overlapping matches of regular expression when generating inital stack of “tokens”
0.0.46 (2020-02-26)
Implemented heuristics to detect (albeit imperfectly) military times
0.0.44 (2019-11-05)
Released time corpus
Implemented training model using ctparse corpus
0.0.43 (2019-11-01)
Added slash as a general separator
Added ruleTODTOD (to support expression like afternoon/evening)
0.0.42 (2019-10-30)
Removed nb module
Fix for two digit years
Freshly retrained model binary file
0.0.41 (2019-10-29)
Fix run_corpus refactoring bug
Implemented retraining utilities
0.0.40 (2019-10-25)
update develop dependencies
remove unused Protocol import from typing_extensions
0.0.39 (2019-10-24)
split ctparse file into several different modules
added types to public interface
introduced the Scorer abstraction to implement richer scoring strategies
0.0.38 (2018-11-05)
Added python 3.7 to supported versions (fix on travis available)
0.0.8 (2018-06-07)
First release on PyPI.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file ctparse-0.3.6.tar.gz
.
File metadata
- Download URL: ctparse-0.3.6.tar.gz
- Upload date:
- Size: 128.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.1 CPython/3.9.15
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 35ff7f87843ebed34de162142d0f7909bbc0f9366fd14b6584742700abef3f62 |
|
MD5 | 8d428f20ec13aef27acbb958b54ef7b9 |
|
BLAKE2b-256 | 7c9bc43cffc46999471ab611ec3078bf198e4b0fca88cb5d144716a80c2c30eb |
File details
Details for the file ctparse-0.3.6-py2.py3-none-any.whl
.
File metadata
- Download URL: ctparse-0.3.6-py2.py3-none-any.whl
- Upload date:
- Size: 112.2 kB
- Tags: Python 2, Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.1 CPython/3.9.15
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 9014e2c6a64e97a9baa236f3677b958699d6e4713df823a2439b7a003f0d6afa |
|
MD5 | 2d285700e5d6d02ba4a0edc7e3a1be8b |
|
BLAKE2b-256 | ef1147cc8dd49aba2992b47b41a282c195fc750672becf59310d3859a5253c99 |