Skip to main content

Rule-based morphological analysis for Ossetic (Iron)

Project description

Ossetic (Iron) morphological analyzer

This is a rule-based morphological analyzer for Ossetic (oss). It is based on a formalized description of the morphology of literary Ossetic, which is based on the Iron dialect, and uses uniparser-morph for parsing. It performs full morphological analysis of Ossetic words (lemmatization, POS tagging, grammatical tagging).

How to use

Python package

The analyzer is available as a Python package. If you want to analyze Ossetic texts in Python, install the module:

pip3 install uniparser-ossetic

Import the module and create an instance of OsseticAnalyzer class. Set mode='strict' if you are going to process text in standard orthography, or mode='nodiacritics' if you expect the ӕ character to be misrepresented in some words (either as an identically looking Latin character or as ае). After that, you can either parse tokens or lists of tokens with analyze_words(), or parse a frequency list with analyze_wordlist(). Here is a simple example:

from uniparser_ossetic import OsseticAnalyzer
a = OsseticAnalyzer(mode='strict')

analyses = a.analyze_words('ӕвзаджы')
# The parser is initialized before first use, so expect
# some delay here (usually several seconds)

# You will get a list of Wordform objects
# The analysis attributes are stored in its properties
# as string values, e.g.:
for ana in analyses:
        print(ana.wf, ana.lemma, ana.gramm)

# You can also pass lists (even nested lists) and specify
# output format ('xml', 'json' or 'conll')
# If you pass a list, you will get a list of analyses
# with the same structure
analyses = a.analyze_words([['Фӕлӕ'], ['Æз', 'дæ', 'уарзын', '.']],
	                       format='xml')
analyses = a.analyze_words([['Фӕлӕ'], ['Æз', 'дæ', 'уарзын', '.']],
	                       format='conll')
analyses = a.analyze_words(['ӕвзаджы', [['Фӕлӕ'], ['Æз', 'дæ', 'уарзын', '.']]],
	                       format='json')

Refer to the uniparser-morph documentation for the full list of options.

Word lists

Alternatively, you can use a preprocessed word list. The wordlists directory contains a list of words from a 12-million-word Ossetic National Corpus (wordlist.csv) with 438,000 unique tokens, list of analyzed tokens (wordlist_analyzed.txt; each line contains all possible analyses for one word in an XML format), and list of tokens the parser could not analyze (wordlist_unanalyzed.txt). The recall of the analyzer on the corpus texts is about 90%.

Description format

The description is carried out in the uniparser-morph format and involves a description of the inflection (oss_paradigms.txt), a grammatical dictionary (oss_lexemes.txt), a list of productive lemma-changing derivations (derivations.txt), and a short list of analyses that should be avoided (bad_analyses.txt). The dictionary contains descriptions of individual lexemes, each of which is accompanied by information about its stem, its part-of-speech tag and some other grammatical information, its inflectional type (paradigm), and Russian and/or English translation. See more about the format in the uniparser-morph documentation.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

uniparser-ossetic-2.0.0.tar.gz (1.3 MB view hashes)

Uploaded Source

Built Distribution

uniparser_ossetic-2.0.0-py3-none-any.whl (1.4 MB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page