Light-weight sentence tokenizer for Chinese languages.
Project description
A light-weight sentence tokenizer for Chinese languages.
Sample Code:
from zh_sentence.tokenizer import tokenize
paragraph_str = “你好吗?你快乐吗?”
sentence_list = tokenize(input_str)
- for sentence in sentence_list:
print(sentence)
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
zh_sentence-0.0.2.tar.gz
(2.0 kB
view hashes)