Skip to main content

A flexible sentence segmentation library using CRF model and regex rules

Project description

sentsplit

A flexible sentence segmentation library using CRF model and regex rules

This library allows splitting of text paragraphs into sentences. It is built with the following desiderata:

  • Be able to extend to new languages or "types" of sentences from data alone by learning a conditional random field (CRF) model.
  • Also provide functionality to segment (or not to segment) lines based on regular expression rules (referred as segment_regexes and prevent_regexes, respectively).
  • Be able to reconstruct the exact original text paragraphs from joining the segmented sentences.

All in all, the library aims to benefit from the best of both worlds: data-driven and rule-based approaches.

Installation

Supports Python 3.6+

# stable
pip install sentsplit

# bleeding-edge
pip install git+https://github.com/zaemyung/sentsplit

Uses python-crfsuite, which, in turn, is built upon CRFsuite.

Segmentation

CLI

$ sentsplit segment -l lang_code -i /path/to/input_file  # outputs to /path/to/input_file.segment
$ sentsplit segment -l lang_code -i /path/to/input_file -o /path/to/output_file

$ sentsplit segment -h  # prints out the detailed usage

Python Library

from sentsplit.segment import SentSplit

# use default setting
sent_splitter = SentSplit(lang_code)

# override default setting - see "Features" for detail
sent_splitter = SentSplit(lang_code, **overriding_kwargs)

# segment a single line
sentences = sent_splitter.segment(line)

# can also segment a list of lines
sentences = sent_splitter.segment([lines])

Features

The behavior of segmentation can be adjusted by the following arguments:

  • mincut: a line is not segmented if its character-level length is smaller than mincut, preventing too short sentences.
  • maxcut: a line is segmented if its character-level length is greater or equal to maxcut, preventing too long sentences.
  • strip_spaces: trim any white spaces in front and end of a sentence; does not guarantee exact reconstruction of original passages.
  • handle_multiple_spaces: substitute multiple spaces with a single space, perform segmentation, and recover the original spaces.
  • segment_regexes: segment at either start or end index of the matched group defined by the regex patterns.
  • prevent_regexes: a line is not segmented at characters that fall within the matching group(s) captured by the regex patterns.
  • prevent_word_split: a line is not segmented at characters that are within a word where the word boundary is denoted by white spaces around it or a punctuation; may not be suitable for languages (e.g. Chinese, Japanese, Thai) that do not use spaces to differentiate words.

Segmentation is performed by first applying a trained CRF model to a line, where each character in the line is labelled as either O or EOS. EOS label indicates the position for segmentation.

Note that prevent_regexes is applied after segment_regexes, meaning that the segmentation positions captured by segment_regexes can be overridden by prevent_regexes.

An Example

Let's suppose we want to segment sentences that end with a tilde (~ or ) which is often used in some East Asian countries to convey a sense of friendliness, silliness, whimsy or flirtatiousness. We can devise a regex that looks something like this: (?<=[다요])~+(?= ), where and are the most common characters that finish the sentences in the polite/formal form. This regex can be added to segment_regexes to take effect:

from copy import deepcopy
from sentsplit.config import ko_config
from sentsplit.segment import SentSplit

my_config = deepcopy(ko_config)
my_config['segment_regexes'].append({'name': 'tilde_ending', 'regex': r'(?<=[다요])~+(?= )', 'at': 'end'})
sent_splitter = SentSplit('ko', **my_config)

sent_splitter.segment('안녕하세요~ 만나서 정말 반갑습니다~~ 잘 부탁드립니다!')

# results with the regex: ['안녕하세요~', ' 만나서 정말 반갑습니다~~', ' 잘 부탁드립니다!']
# results without the regex: ['안녕하세요~ 만나서 정말 반갑습니다~~ 잘 부탁드립니다!']

To learn more about the regular expressions, this website provides a good tutorial.

Creating a New SentSplit Model

Creating a new model involves first training a CRF model on a dataset of clean sentences, followed by (optionally) adding or modifying the feature arguments for better performance.

Training a CRF Model

First, prepare a corpus file where a single line corresponds to a single sentence. Then, a CRF model can be trained by running a command:

sentsplit train -l lang_code -c corpus_file_path  # outputs to {corpus_file_path}.{lang_code}-{ngram}-gram-{YearMonthDate}.model

sentsplit train -h  # prints out the detailed usage

The following arguments are used to set the training setting:

  • ngram: maximum ngram features used for CRF model; default is 5.
  • crf_max_iteration: maximum number of CRF iteration for training; default is 50.
  • sample_min_length: when preparing an input sample for CRF model, gold sentences are concatenated to form a longer sample with a length greater than sample_min_length; default is 450.
  • add_depunctuated_samples: when set to True, randomly (30% chance) remove the punctuation of the current sentence before concatenation; default is False. May only be suitable for languages (e.g. Korean, Japanese) that have specific endings for sentences.
  • add_despaced_samples: when set to True, with 35% chance, current sentence is concatenated to input sample without a prepending white space; default is False.

Setting Configuration

Refer to the base_config in config.py. Append a new config to the file, adjusting the arguments accordingly if needed.

A newly created model can also be called directly in codes by passing the kwargs accordingly:

from sentsplit.segment import SentSplit

sent_splitter = SentSplit(lang_code, model='path/to/model', ...)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

sentsplit-1.0.2.tar.gz (1.3 MB view details)

Uploaded Source

Built Distribution

sentsplit-1.0.2-py3-none-any.whl (1.5 MB view details)

Uploaded Python 3

File details

Details for the file sentsplit-1.0.2.tar.gz.

File metadata

  • Download URL: sentsplit-1.0.2.tar.gz
  • Upload date:
  • Size: 1.3 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.1 importlib_metadata/3.7.3 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.59.0 CPython/3.6.8

File hashes

Hashes for sentsplit-1.0.2.tar.gz
Algorithm Hash digest
SHA256 066c5bf3c9ac0f3f5b79f2903e328badc68f57eaaf1520d0fd7f0199a5696fa8
MD5 1cc4d5939a53b45fa2f969b173d3b178
BLAKE2b-256 f62a90372f1a99f3ce8d4ac97a90db67a458eae88b1c836314a0bf90ec082bc2

See more details on using hashes here.

File details

Details for the file sentsplit-1.0.2-py3-none-any.whl.

File metadata

  • Download URL: sentsplit-1.0.2-py3-none-any.whl
  • Upload date:
  • Size: 1.5 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.1 importlib_metadata/3.7.3 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.59.0 CPython/3.6.8

File hashes

Hashes for sentsplit-1.0.2-py3-none-any.whl
Algorithm Hash digest
SHA256 08dc7a43c2a3d99693798aaa1623591172d836bcd24bcae19cc551f5b49a2319
MD5 f577a9c588de45e1875af1fab0e50134
BLAKE2b-256 f96273f8b3a2b5a2f9dfda0c8c1af7af2d00de84efcf67def8655a309411a325

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page