Skip to main content

Package to detokenize the stream of tokens

Project description

De-Tokenize


Package PyPI version Author License made-with-python Python 3.7 Downloads

What is it?

De-Tokenize is a Python package which provides fast, accurate structuring of tokens back to original sentence form.

Contributor

Littin Rajan

Main Features

  • Light-weight package
  • No more dependencies
  • Powerful, flexible

Where to get it?

The source code is currently hosted on GitHub at: https://github.com/littinrajan/detokenize

Binary installers for the latest released version are available at the Python Package Index (PyPI) and on Conda.

# conda
conda install detokenize
# or PyPI
pip install detokenize

Usage

from detokenize.detokenizer import detokenize

sample_tokens = ['These', 'are', 'some', 'tokens', '.']
sentence = detokenize(sample_tokens)

License

MIT

Contributing to De-Tokenize

All contributions, bug reports, bug fixes, documentation improvements, enhancements, and ideas are welcome.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

detokenize-0.0.4.tar.gz (3.0 kB view hashes)

Uploaded Source

Built Distribution

detokenize-0.0.4-py3-none-any.whl (3.5 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page