A package for detecting and filtering spam messages using Machine Learning models.
Project description
Spam-Detector-AI
Spam-Detector-AI is a Python package for detecting and filtering spam messages using Machine Learning models. The package integrates with Django or any other project that uses python and offers different types of classifiers: Naive Bayes, Random Forest, and Support Vector Machine (SVM). Since version 2.1.0, two new classifiers have been added: Logistic Regression and XGBClassifier.
⚠️ Warning: No significant breaking changes were added to the version 2.x.x in terms of usage. On the other hand,
the fine-tuning of the models has been moved to a separate module (tuning
) and the tests have been moved to a
separate module (tests
).
⚠️
Table of Contents
Installation
You can install the spam detection package via pip:
pip install spam-detector-ai
Make sure you have the following dependencies installed:
- scikit-learn
- nltk
- pandas
- numpy
- joblib
- xgboost
Additionally, you'll need to download the NLTK data and to do so, use the python interpreter to run the following commands:
import nltk
nltk.download('wordnet')
nltk.download('stopwords')
Usage
Training the Models
Before using the classifiers, you must train the models. Training data is loaded from a CSV file. You can find the
training data in the data
directory in the GitHub's page of the project. The CSV file must have 3 columns: label
,
text
and label_num
. The text
column contains the content of the message to analyze and the label
column
contains the labels ham
or spam
and label_num
contains the number 0
(not spam) or 1
(spam).
The more data you have, the better the models will perform.
To train the models, run the following command:
python3 spam_detector_ai/trainer.py
⚠️ Warning: A module not found error may occur ⚠️
If this happens, use an IDE to run the trainer.py
file until a fix is implemented.
This will train all the models and save them as .joblib
files in the models directory. For now, there is 3 models:
naive_bayes_model.joblib
random_forest_model.joblib
svm_model.joblib
logistic_regression_model.joblib
xgb_model.joblib
Tests
The test results are shown below:
Model: NAIVE_BAYES
Confusion Matrix:
Predicted: Ham | Predicted: Spam | |
---|---|---|
Actual: Ham | 1935 (True Negative) | 170 (False Positive) |
Actual: Spam | 221 (False Negative) | 633 (True Positive) |
- True Negative (TN): 1935 messages were correctly identified as ham (non-spam).
- False Positive (FP): 170 ham messages were incorrectly identified as spam.
- False Negative (FN): 221 spam messages were incorrectly identified as ham.
- True Positive (TP): 633 messages were correctly identified as spam.
Performance Metrics:
Precision | Recall | F1-Score | Support | |
---|---|---|---|---|
Ham | 0.90 | 0.92 | 0.91 | 2105 |
Spam | 0.79 | 0.74 | 0.76 | 854 |
Accuracy | 0.87 | 2959 | ||
Macro Avg | 0.84 | 0.83 | 0.84 | 2959 |
Weighted Avg | 0.87 | 0.87 | 0.87 | 2959 |
Accuracy: 0.8678607637715444
Model: RANDOM_FOREST
Confusion Matrix:
Predicted: Ham | Predicted: Spam | |
---|---|---|
Actual: Ham | 2067 (True Negative) | 38 (False Positive) |
Actual: Spam | 36 (False Negative) | 818 (True Positive) |
- True Negative (TN): 2067 messages were correctly identified as ham (non-spam).
- False Positive (FP): 38 ham messages were incorrectly identified as spam.
- False Negative (FN): 36 spam messages were incorrectly identified as ham.
- True Positive (TP): 818 messages were correctly identified as spam.
Performance Metrics:
Precision | Recall | F1-Score | Support | |
---|---|---|---|---|
Ham | 0.98 | 0.98 | 0.98 | 2105 |
Spam | 0.96 | 0.96 | 0.96 | 854 |
Accuracy | 0.97 | 2959 | ||
Macro Avg | 0.97 | 0.97 | 0.97 | 2959 |
Weighted Avg | 0.98 | 0.97 | 0.98 | 2959 |
Accuracy: 0.9749915511997297
Model: SVM
Confusion Matrix:
Predicted: Ham | Predicted: Spam | |
---|---|---|
Actual: Ham | 2080 (True Negative) | 25 (False Positive) |
Actual: Spam | 41 (False Negative) | 813 (True Positive) |
- True Negative (TN): 2080 messages were correctly identified as ham (non-spam).
- False Positive (FP): 25 ham messages were incorrectly identified as spam.
- False Negative (FN): 41 spam messages were incorrectly identified as ham.
- True Positive (TP): 813 messages were correctly identified as spam.
Performance Metrics:
Precision | Recall | F1-Score | Support | |
---|---|---|---|---|
Ham | 0.98 | 0.99 | 0.98 | 2105 |
Spam | 0.97 | 0.95 | 0.96 | 854 |
Accuracy | 0.98 | 2959 | ||
Macro Avg | 0.98 | 0.97 | 0.97 | 2959 |
Weighted Avg | 0.98 | 0.98 | 0.98 | 2959 |
Accuracy: 0.9773572152754308
Model: LOGISTIC_REGRESSION
Confusion Matrix:
Predicted: Ham | Predicted: Spam | |
---|---|---|
Actual: Ham | 2065 (True Negative) | 48 (False Positive) |
Actual: Spam | 46 (False Negative) | 989 (True Positive) |
- True Negative (TN): 2065 messages were correctly identified as ham (non-spam).
- False Positive (FP): 48 ham messages were incorrectly identified as spam.
- False Negative (FN): 46 spam messages were incorrectly identified as ham.
- True Positive (TP): 989 messages were correctly identified as spam.
Performance Metrics:
Precision | Recall | F1-Score | Support | |
---|---|---|---|---|
Ham | 0.98 | 0.98 | 0.98 | 2113 |
Spam | 0.95 | 0.96 | 0.95 | 1035 |
Accuracy | 0.97 | 3148 | ||
Macro Avg | 0.97 | 0.97 | 0.97 | 3148 |
Weighted Avg | 0.97 | 0.97 | 0.97 | 3148 |
Accuracy: 0.9707680491551459
Model: XGB
Confusion Matrix:
Predicted: Ham | Predicted: Spam | |
---|---|---|
Actual: Ham | 2050 (True Negative) | 63 (False Positive) |
Actual: Spam | 28 (False Negative) | 1007 (True Positive) |
- True Negative (TN): 2050 messages were correctly identified as ham (non-spam).
- False Positive (FP): 63 ham messages were incorrectly identified as spam.
- False Negative (FN): 28 spam messages were incorrectly identified as ham.
- True Positive (TP): 1007 messages were correctly identified as spam.
Performance Metrics:
Precision | Recall | F1-Score | Support | |
---|---|---|---|---|
Ham | 0.99 | 0.97 | 0.98 | 2113 |
Spam | 0.94 | 0.97 | 0.96 | 1035 |
Accuracy | 0.97 | 3148 | ||
Macro Avg | 0.96 | 0.97 | 0.97 | 3148 |
Weighted Avg | 0.97 | 0.97 | 0.97 | 3148 |
Accuracy: 0.9710927573062261
The models that performed the best are the SVM and Logistic Regression, with the SVM model achieving slightly higher accuracy than Logistic Regression. Given that no single model achieved perfect accuracy, I have decided to implement a voting classifier. This classifier will combine the predictions of the five models (Naive Bayes, Random Forest, SVM, Logistic Regression, and XGB) using a majority vote system to make the final prediction. This approach aims to leverage the strengths of each model to improve overall prediction accuracy.
Weighted Voting System
To enhance the decision-making process, I've refined our approach to a weighted voting system. This new system assigns different weights to each model's vote based on their respective accuracies. The weights are proportional to the accuracy of each model relative to the sum of the accuracies of all models. The models with higher accuracy have a greater influence on the final decision.
The models and their respective proportional weights are as follows:
- Naive Bayes: Weight = 0.1822
- Random Forest: Weight = 0.2047
- SVM (Support Vector Machine): Weight = 0.2052
- Logistic Regression: Weight = 0.2039
- XGBoost (XGB): Weight = 0.2039
These weights were calculated based on the accuracy of each model as a proportion of the total accuracy of all models. The final decision whether a message is spam or not is determined by the weighted spam score. Each model casts a vote (spam or not spam), and this vote is multiplied by the model's weight. The weighted spam scores from all models are then summed up. If this total weighted spam score exceeds 50% of the total possible weight, the message is classified as spam. Otherwise, it's classified as not spam (ham).
This approach ensures that the more accurate models have a larger say in the final decision, thereby increasing the reliability of spam detection. It combines the strengths of each model, compensating for individual weaknesses and provides a more nuanced classification.
System Output
The system provides a detailed output for each message, showing the vote (spam or ham) from each model, along with its weight. It also displays the total weighted spam score and the final classification decision (Spam or Not Spam). This transparency in the voting process allows for easier understanding and debugging of the model's performance on different messages.
If you have trained the models on new data, you can test them by running the following command:
python tests/test.py
⚠️ Warning: A module not found error may occur ⚠️
If this happens, use an IDE to run the test.py
file until a fix is implemented.
Making Predictions
To use the spam detector in your Django project:
- Import the
VotingSpamDetector
from theprediction
module. - Create an instance of the detector.
- Use the
is_spam
method to check if a message is spam.
from spam_detector_ai.prediction.predict import VotingSpamDetector
# Create the spam detector
spam_detector = VotingSpamDetector()
# Check if a message is spam
message = "Enter the message here"
is_spam = spam_detector.is_spam(message)
print(f"Is spam: {is_spam}")
Project Structure
classifiers/
: Contains the different classifiers (Naive Bayes, Random Forest, SVM, XGB & Logistic Regression).data/
: Contains the sample dataset for training the classifiers.loading_and_processing/
: Contains utility functions for loading and preprocessing data.models/
: Contains the trained models and their vectorizers.prediction/
: Contains the main spam detector class.tests/
: Contains scripts for testingtuning/
: Contains scripts for tuning the classifiers.training/
: Contains scripts for training the classifiers.
Contributing
Contributions are welcome! Please read the contributing guidelines before getting started.
License
This project is licensed under the MIT License.
Notes
The project contains 5 pre-trained models that can be used directly if you want to skip the training step. If you don't want to use the package, you can use the API that I have deployed here.
The API is built with Django, and the following is an example of how I use it in a personal project:
The code:
import requests
def check_website_contact_form(request):
message_ = request.POST.get('message')
subject = request.POST.get('subject')
# etc...
# Concatenate subject and message
message_with_subject = f'subject: {subject}. {message_}'
# Call the spam detection API
response = requests.post(
"https://spam-detection-api.adamspierredavid.com/v1/check-spam/",
json={'text': message_with_subject} # Use json parameter instead of data
# You can also do `json={'message': message_with_subject}
)
is_spam = False
# Check if the API request was successful
if response.status_code == 200:
# Parse the JSON response
json_response = response.json()
is_spam = json_response.get('is_spam')
if is_spam:
# Do something
pass
else:
# Do something else
pass
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for spam_detector_ai-2.1.2-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 2e94723bcdd661dccea897a9fdb4207593fee2e85ba3c8fe820a39f0bf769a53 |
|
MD5 | cc6d0f0292a50ecdd0de75258f5a7dd4 |
|
BLAKE2b-256 | a17a7dccc8c00cd0ef30643dd7f00135095afd3cfa58655bd25bbc9e09d3c379 |