Skip to main content

Module encoding and encrypting text by key

Project description


Module encoding and encrypting text by key

Usage example

from toxicityclassifier import *

classifier = ToxicityClassificator()

print(classifier.predict(text))          # (0 or 1, probability)
print(classifier.get_probability(text))  # probability
print(classifier.classify(text))         # 0 or 1


Weight for classification (if probability >= weight => 1 else 0)

classifier.weight = 0.5

Weight for language detection (English or Russian)

if the percentage of the Russian language >= language_weight, then the Russian model is used, otherwise the English one

classifier.language_weight = 0.5

Project details

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

toxicityclassifier-0.1.9.tar.gz (3.6 MB view hashes)

Uploaded Source

Built Distribution

toxicityclassifier-0.1.9-py3-none-any.whl (3.6 MB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page