Skip to main content

A flexible sentence segmentation library using CRF model and regex rules

Project description

sentsplit

A flexible sentence segmentation library using CRF model and regex rules

This library allows splitting of text paragraphs into sentences. It is built with the following desiderata:

  • Be able to extend to new languages or "types" of sentences from data alone by learning a conditional random field (CRF) model.
  • Also provide functionality to segment (or not to segment) lines based on regular expression rules (referred as segment_regexes and prevent_regexes, respectively).
  • Be able to reconstruct the exact original text paragraphs from joining the segmented sentences.

All in all, the library aims to benefit from the best of both worlds: data-driven and rule-based approaches.

Installation

Supports Python 3.6+

# stable
pip install sentsplit

# bleeding-edge
pip install git+https://github.com/zaemyung/sentsplit

Uses python-crfsuite, which, in turn, is built upon CRFsuite.

Segmentation

CLI

$ sentsplit segment -l lang_code -i /path/to/input_file  # outputs to /path/to/input_file.segment
$ sentsplit segment -l lang_code -i /path/to/input_file -o /path/to/output_file

$ sentsplit segment -h  # prints out the detailed usage

Python Library

from sentsplit.segment import SentSplit

# use default setting
sent_splitter = SentSplit(lang_code)

# override default setting - see "Features" for detail
sent_splitter = SentSplit(lang_code, **overriding_kwargs)

# segment a single line
sentences = sent_splitter.segment(line)

# can also segment a list of lines
sentences = sent_splitter.segment([lines])

Features

The behavior of segmentation can be adjusted by the following arguments:

  • mincut: a line is not segmented if its character-level length is smaller than mincut, preventing too short sentences.
  • maxcut: a line is segmented if its character-level length is greater or equal to maxcut, preventing too long sentences.
  • strip_spaces: trim any white spaces in front and end of a sentence; does not guarantee exact reconstruction of original passages.
  • handle_multiple_spaces: substitute multiple spaces with a single space, perform segmentation, and recover the original spaces.
  • segment_regexes: segment at either start or end index of the matched group defined by the regex patterns.
  • prevent_regexes: a line is not segmented at characters that fall within the matching group(s) captured by the regex patterns.
  • prevent_word_split: a line is not segmented at characters that are within a word where the word boundary is denoted by white spaces around it or a punctuation; may not be suitable for languages (e.g. Chinese, Japanese, Thai) that do not use spaces to differentiate words.

Segmentation is performed by first applying a trained CRF model to a line, where each character in the line is labelled as either O or EOS. EOS label indicates the position for segmentation.

Note that prevent_regexes is applied after segment_regexes, meaning that the segmentation positions captured by segment_regexes can be overridden by prevent_regexes.

An Example

Let's suppose we want to segment sentences that end with a tilde (~ or ) which is often used in some East Asian countries to convey a sense of friendliness, silliness, whimsy or flirtatiousness. We can devise a regex that looks something like this: (?<=[다요])~+(?= ), where and are the most common characters that finish the sentences in the polite/formal form. This regex can be added to segment_regexes to take effect:

from copy import deepcopy
from sentsplit.config import ko_config
from sentsplit.segment import SentSplit

my_config = deepcopy(ko_config)
my_config['segment_regexes'].append({'name': 'tilde_ending', 'regex': r'(?<=[다요])~+(?= )', 'at': 'end'})
sent_splitter = SentSplit('ko', **my_config)

sent_splitter.segment('안녕하세요~ 만나서 정말 반갑습니다~~ 잘 부탁드립니다!')

# results with the regex: ['안녕하세요~', ' 만나서 정말 반갑습니다~~', ' 잘 부탁드립니다!']
# results without the regex: ['안녕하세요~ 만나서 정말 반갑습니다~~ 잘 부탁드립니다!']

To learn more about the regular expressions, this website provides a good tutorial.

Creating a New SentSplit Model

Creating a new model involves first training a CRF model on a dataset of clean sentences, followed by (optionally) adding or modifying the feature arguments for better performance.

Training a CRF Model

First, prepare a corpus file where a single line corresponds to a single sentence. Then, a CRF model can be trained by running a command:

sentsplit train -l lang_code -c corpus_file_path  # outputs to {corpus_file_path}.{lang_code}-{ngram}-gram-{YearMonthDate}.model

sentsplit train -h  # prints out the detailed usage

The following arguments are used to set the training setting:

  • ngram: maximum ngram features used for CRF model; default is 5.
  • crf_max_iteration: maximum number of CRF iteration for training; default is 50.
  • sample_min_length: when preparing an input sample for CRF model, gold sentences are concatenated to form a longer sample with a length greater than sample_min_length; default is 450.
  • depunctuation_ratio: ratio of training samples with no punctuation inbetween the sentences. May only be suitable for certain languages (e.g. "ko", "ja") that have specific endings for sentences. The top-num_depunctuation_endings most common endings are computed from corpus. 1.0 means 100% of the training samples are depunctuated.
  • num_depunctuation_endings: number of most common sentence endings to extract and use.
  • ending_length: length of sentence endings counted from reverse, exclusing any punctuation.
  • despace_ratio: ratio of training samples without whitespaces inbetween the sentences. 1.0 means 100% of the training samples are despaced. For languages that do not often use whitespaces, set this to a high value ~1.0.

Setting Configuration

Refer to the base_config in config.py. Append a new config to the file, adjusting the arguments accordingly if needed.

A newly created model can also be called directly in codes by passing the kwargs accordingly:

from sentsplit.segment import SentSplit

sent_splitter = SentSplit(lang_code, model='path/to/model', ...)

Supported Languages

Currently supported languages are:

  • English (en)
  • French (fr)
  • German (de)
  • Italian (it)
  • Japanese (ja)
  • Korean (ko)
  • Lithuanian (lt)
  • Polish (pl)
  • Portuguese (pt)
  • Russian (ru)
  • Simplified Chinese (zh)
  • Turkish (tr)

Please note that many of these languages are trained with openly available sentences gathered from bilingual corpora for machine translations. The training sentences for European languages are mostly from the Europarl corpora, so the default models may not handle colloquial sentences effectively. We can either train a new CRF model with more gold sentences from the target domain, or devise a set of domain-specific regex rules if need be.

License

sentsplit is licensed under MIT license, as found in LICENSE file.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

sentsplit-1.0.3.tar.gz (1.7 MB view details)

Uploaded Source

Built Distribution

sentsplit-1.0.3-py3-none-any.whl (2.6 MB view details)

Uploaded Python 3

File details

Details for the file sentsplit-1.0.3.tar.gz.

File metadata

  • Download URL: sentsplit-1.0.3.tar.gz
  • Upload date:
  • Size: 1.7 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.1 importlib_metadata/3.7.3 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.59.0 CPython/3.6.8

File hashes

Hashes for sentsplit-1.0.3.tar.gz
Algorithm Hash digest
SHA256 8782a7c09d5b52b4e65109165b3aa729ed236f9ea11ed02623e5859a31e83d7a
MD5 af77cc1af74345fbac29dff0b3bc0255
BLAKE2b-256 fa1ade8bca23138608e6062eff46f74bf5df50d7fa45e27fa17990deecf7a046

See more details on using hashes here.

File details

Details for the file sentsplit-1.0.3-py3-none-any.whl.

File metadata

  • Download URL: sentsplit-1.0.3-py3-none-any.whl
  • Upload date:
  • Size: 2.6 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.1 importlib_metadata/3.7.3 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.59.0 CPython/3.6.8

File hashes

Hashes for sentsplit-1.0.3-py3-none-any.whl
Algorithm Hash digest
SHA256 379777c538d85b78de832e53dfccdf99645b2a4bf14ae35308597298390633dc
MD5 d6680f07964729c2e67b14b7866b4c83
BLAKE2b-256 1ecb5c4402154dd8ccc776976d2edefe7868583e13c9c6f4a0a3846c61e07d6a

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page