NLP tools
Project description
word_segmentation
Chinese word segmentation algorithm without corpus
Usage
from word_segmentation import get_words
content = '北京比武汉的人口多,但是北京的天气没有武汉的热,武汉有热干面,北京有北京烤鸭'
words = get_words(content, max_word_len=2, min_aggregation=1, min_entropy=0.5)
print(words)
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
hellonlp-0.2.41.tar.gz
(1.5 MB
view details)
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file hellonlp-0.2.41.tar.gz.
File metadata
- Download URL: hellonlp-0.2.41.tar.gz
- Upload date:
- Size: 1.5 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.6.0 importlib_metadata/4.8.2 pkginfo/1.5.0.1 requests/2.24.0 requests-toolbelt/0.9.1 tqdm/4.23.0 CPython/3.6.1
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
8e3d4d308c6e88f99e1c65ee88fd49d851e708c5006c0ece96fbd4df078140fe
|
|
| MD5 |
cd2969b17503cff0be8e6f778f473f2f
|
|
| BLAKE2b-256 |
c1df64624955b0cd328bc9c37034353ceed386ad5e5827f5ee9cb9a9f869621e
|
File details
Details for the file hellonlp-0.2.41-py3-none-any.whl.
File metadata
- Download URL: hellonlp-0.2.41-py3-none-any.whl
- Upload date:
- Size: 1.5 MB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.6.0 importlib_metadata/4.8.2 pkginfo/1.5.0.1 requests/2.24.0 requests-toolbelt/0.9.1 tqdm/4.23.0 CPython/3.6.1
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
2eb82dee5c098456105098ebad56bd731ea9d0dc992315f973991446a81b7905
|
|
| MD5 |
0d9290ad5a475d8809d209c1596a7e77
|
|
| BLAKE2b-256 |
ac3a93dfc2fdffda7899117bd48ad995ff46d4160f99647df0f67a3606c351cc
|