Light-weight sentence tokenizer for Chinese languages.
Project description
A light-weight sentence tokenizer for Chinese languages.
Sample Code:
from zh_sentence.tokenizer import tokenize
paragraph_str = "你好吗?你快乐吗?"
sentence_list = tokenize(paragraph_str)
for sentence in sentence_list: print(sentence)
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
zh_sentence-0.0.4.tar.gz
(2.4 kB
view hashes)
Built Distribution
Close
Hashes for zh_sentence-0.0.4-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 39806c9921a5ae06558f20a6e108976a7982b72267e6064447d477f8541e145b |
|
MD5 | c9f15c553ac7bb0d2286bb2fd8de2bbc |
|
BLAKE2b-256 | 34b5ab180bbcd0ac9681c64b4747bab79d040f9d5cdb95376734ee0cb62eea91 |