Skip to main content

A fast, robust library to check for offensive language in strings. This version of "profanity-check" is used for Bleepy.

Project description

bleepy-profanity-check

pip install bleepy-profanity-check

bleepy-profanity-check

profanity-check version used for Bleepy

A fast, robust Python library to check for profanity or offensive language in strings. Read more about how and why profanity-check was built in this blog post. You can also test out profanity-check in your browser.

Notice: We created a library for this project, see: https://pypi.org/project/alt-profanity-check/ and https://gitlab.com/dimitrios/alt-profanity-check/

Disclaimer

Profanity-check is originally created by Victor Zhou, and updated by other contributors. This version of profanity check will be use by Bleepy to detect and bleep tagalog and english profanity from videos. Therefore this version is called taglish-profanity-check.

How It Works

profanity-check uses a linear SVM model trained on 200k human-labeled samples of clean and profane text strings. Its model is simple but surprisingly effective, meaning profanity-check is both robust and extremely performant.

Why Use profanity-check?

No Explicit Blacklist

Many profanity detection libraries use a hard-coded list of bad words to detect and filter profanity. For example, profanity uses this wordlist, and even better-profanity still uses a wordlist. There are obviously glaring issues with this approach, and, while they might be performant, these libraries are not accurate at all.

A simple example for which profanity-check is better is the phrase "You cocksucker" - profanity thinks this is clean because it doesn't have "cocksucker" in its wordlist.

Performance

Other libraries like profanity-filter use more sophisticated methods that are much more accurate but at the cost of performance. A benchmark (performed December 2018 on a new 2018 Macbook Pro) using a Kaggle dataset of Wikipedia comments yielded roughly the following results:

Package 1 Prediction (ms) 10 Predictions (ms) 100 Predictions (ms)
profanity-check 0.2 0.5 3.5
profanity-filter 60 1200 13000
profanity 0.3 1.2 24

profanity-check is anywhere from 300 - 4000 times faster than profanity-filter in this benchmark!

Accuracy

This table speaks for itself:

Package Test Accuracy Balanced Test Accuracy Precision Recall F1 Score
profanity-check 95.0% 93.0% 86.1% 89.6% 0.88
profanity-filter 91.8% 83.6% 85.4% 70.2% 0.77
profanity 85.6% 65.1% 91.7% 30.8% 0.46

See the How section below for more details on the dataset used for these results.

Installation

$ pip install profanity-check

Usage

from profanity_check import predict, predict_prob

predict(['predict() takes an array and returns a 1 for each string if it is offensive, else 0.'])
# [0]

predict(['fuck you'])
# [1]

predict_prob(['predict_prob() takes an array and returns the probability each string is offensive'])
# [0.08686173]

predict_prob(['go to hell, you scum'])
# [0.7618861]

Note that both predict() and predict_prob return numpy arrays.

More on How/Why It Works

How

Special thanks to the authors of the datasets used in this project. profanity-check was trained on a combined dataset from 2 sources:

profanity-check relies heavily on the excellent scikit-learn library. It's mostly powered by scikit-learn classes CountVectorizer, LinearSVC, and CalibratedClassifierCV. It uses a Bag-of-words model to vectorize input strings before feeding them to a linear classifier.

Why

One simplified way you could think about why profanity-check works is this: during the training process, the model learns which words are "bad" and how "bad" they are because those words will appear more often in offensive texts. Thus, it's as if the training process is picking out the "bad" words out of all possible words and using those to make future predictions. This is better than just relying on arbitrary word blacklists chosen by humans!

Caveats

This library is far from perfect. For example, it has a hard time picking up on less common variants of swear words like "f4ck you" or "you b1tch" because they don't appear often enough in the training corpus. Never treat any prediction from this library as unquestionable truth, because it does and will make mistakes. Instead, use this library as a heuristic.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

bleepy-profanity-check-0.1.8.tar.gz (1.9 MB view details)

Uploaded Source

Built Distribution

bleepy_profanity_check-0.1.8-py3-none-any.whl (1.9 MB view details)

Uploaded Python 3

File details

Details for the file bleepy-profanity-check-0.1.8.tar.gz.

File metadata

File hashes

Hashes for bleepy-profanity-check-0.1.8.tar.gz
Algorithm Hash digest
SHA256 8bd5222d2da5412147d6d64e972e32de9c9a5f6297a1a5f8b1d768270a627b98
MD5 b6c612d324e1f45ac4f02cadd047cffe
BLAKE2b-256 d637299351895115cd3ca1e3f9647bab3f84a3e7b7e790e1acd493952f277f6b

See more details on using hashes here.

Provenance

File details

Details for the file bleepy_profanity_check-0.1.8-py3-none-any.whl.

File metadata

File hashes

Hashes for bleepy_profanity_check-0.1.8-py3-none-any.whl
Algorithm Hash digest
SHA256 3468543645b0d4facc87c1b6cebf7c5459a212a66296d834b71319dd7d60cfbb
MD5 655ecdea2b096a73fa4eba6600060c11
BLAKE2b-256 398bb95f5cd2320b3671239915c0eee1e323e1212f5033ecbd525d7e76619a63

See more details on using hashes here.

Provenance

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page