Linguistic Identification of Morphosyntactic and Expressive Snags
Project description
LIMES: Linguistic Identification of Morphosyntactic and Expressive Snags
LIMES is a library for performing linguistic analyses on provided texts regarding their complexity. The goal of this project is to create a tool that provides actionable insights on how to make written texts easier to comprehend.
Refer to the project documentation for in-depth information about concepts, API, and more.
Please note that the actual logic for identifying language barriers is completely language-specific. Because it is a lot of work to develop these heuristics, the library currently only ships with implemented analyzers for German texts. However, we encourage you to build your own analyzers based on the provided class templates, either for your own use or to contribute to the project.
Installation
You can install this package via pip by running:
pip install lang-limes
Additional Dependencies
The library requires use of a Parser. Currently, we only ship a parser based on spaCy's excellent NLP pipeline. This means that you need to install a spaCy model that supports the language you are working with.
Example Usage
You must use a string container to wrap the text you want to analyze. As our analysis work on a sentence level, you can either manually sentencize and create separate Sentence objects or just throw your whole text into a Text object that takes care of sentencization for you.
We will do the latter for the purpose of this example.
from limes import Text
from limes.parsers import SpacyParser
from limes.analyzers.de import GermanAnalyzer
analyzer = GermanAnalyzer()
# You can also pass a spacy NLP object instead of the model name.
# Make sure the model you want to use is installed.
parser = SpacyParser(model="de_core_news_sm")
text = Text(
raw="Das hier ist ein Text. Dieser Text hat mehrere Sätze.",
analyzer=analyzer,
parser=parser,
)
For parsing, we recommend using the most powerful model that your system can
reasonably run, as barrier detection is very sensitive to errors during morpho-
syntactic analysis. We've used the small model in the example above but if you
can, try using the transformer-based de-dep-news-trf for improved results at
the cost of compute time.
Identifying Barriers
Barriers are detected lazily, and results are cached to avoid redundant
computations. Barriers themselves are a property of the Text object.
# You can iterate over the all barriers in the entire text if you want.
for barrier in text.barriers:
print(barrier.title)
# Print the actual string of the token.
print(barrier.affected_tokens)
# Print the position of the token in the source text.
if barrier.affected_tokens is not None:
print([token.i for token in barrier.affected_tokens])
# You can also iterate over each sentence.
for sentence in text:
print(sentence.barriers)
# Alternatively, you can also inspect a specific sentence by index.
print(text[1].barriers)
Please note that barriers are also language-specific (because different languages also differ in how they make comprehension "difficult").
Calculating Complexities
There are multiple ways in which you can try to approximate language complexity (see our documentation for more information).
from limes import ComplexityAlgorithm
# Get the average complexity of the text. You can manually set the heuristic.
avg_complexity = text.average_complexity(
heuristic=ComplexityAlgorithm.AGGREGATED_LOCAL,
)
print(avg_complexity)
# Alternatively, you can get phrase-level complexities.
# These are also lazily computed and cached.
for phrase, complexity in text.local_complexities:
print(phrase)
print(complexity)
# You could also iterate over all sentences in the text and get each sentence's
# global complexity.
for sentence in text:
complexity = sentence.global_complexity(
heuristic=ComplexityAlgorithm.AGGREGATED_LOCAL,
)
print(sentence)
print(complexity)
Finding Words
If you want to find specific tokens in a Text or Sentence (e.g. if you have a list of words that are allowlisted or disallowlisted for purposes of language simplification), you can utilize LIMES to perform more refined search than simple string matching.
You can search for tokens by providing the word and its Part-of-Speech tag.
This allows more robust word identification as we can normalize capitalization and decrease the risk of matching homographs.
See the example below for an example where LIMES provides improved locating of words over naive string matching.
text = Text(
raw="ich esse gerne essen.",
analyzer=analyzer,
parser=parser,
)
results = text.find("Essen", pos="NOUN")
print(results) # Will print reference to last token in sequence.
results = text.find("essen", pos="NOUN")
print(results) # Will print reference to second token in sequence.
Please note that incorrectly cased input texts usually decrease accuracy of automated text parsing tools, meaning that even with LIMES' word matching, all-lowercase texts may still yield accurate results.
Next Steps
A good place to start is to get an overview of the concepts used to build and configure the whole processing pipeline.
Currently Supported Languages
| Language | Contributors |
|---|---|
| DE | Katja Grosch & Susanne Wagner (IFTO GmbH), Jannik Schmitt (deepsight GmbH) |
Additional Resources
Word Frequency Lists
German
The frequency list for German words was kindly provided by Projekt Deutscher Wortschatz
of the Universität Leipzig. The unprocessed list included in this repository
(data/deu_words_2024.txt) is based on [1]. Please note that it is not based on
the publicly available "Normgrößenkorpora" but was provided on request by the
Leipzig Corpora team under a CC BY 4.0 license.
References
[1] Leipzig Corpora Collection (2024). German news corpus based on material from 2024. Leipzig Corpora Collection. Dataset. https://corpora.uni-leipzig.de/en?corpusId=deu_news_2024
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file lang_limes-0.1.4.tar.gz.
File metadata
- Download URL: lang_limes-0.1.4.tar.gz
- Upload date:
- Size: 53.2 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.0 {"installer":{"name":"uv","version":"0.10.0","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"CachyOS Linux","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
321d7c013ceb1304eff99f9d00afe4b83d7c205615012ab19e59f93075dbbdb4
|
|
| MD5 |
96f1332644f9f25daf87ecf73656a0ea
|
|
| BLAKE2b-256 |
33fe06f73345a810bb5e6d81eb3a42e9f610a1e4fd4d0caee05dcea603688dd0
|
File details
Details for the file lang_limes-0.1.4-py3-none-any.whl.
File metadata
- Download URL: lang_limes-0.1.4-py3-none-any.whl
- Upload date:
- Size: 19.8 MB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.0 {"installer":{"name":"uv","version":"0.10.0","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"CachyOS Linux","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
1d93d90473e03ce733411af9f8d2d6382fd7e72a4e8ebe595c01cea8d4c6a86b
|
|
| MD5 |
abe18e741db81cb99aa8d519ee25342f
|
|
| BLAKE2b-256 |
a48267a810376349ddfe1a951a4cbd8798a67cc81c43f61f455c552e39fbc7f8
|