A Python package for simultaneous regression and binary classification for educational analytics.
Project description
dualPredictor
by D
dualPredictor
is can provide simultaneous regression and binary classification results for tabular dataset. Utilizing well-established regression models such as Lasso, Ridge, and OLS (Ordinary Least Squares) from scikit-learn, dualPredictor
goes a step further by introducing a cutoff-based binary classification. This dual approach allows users not only to predict student grades but also to identify at-risk students efficiently, bridging the gap between traditional regression and classification methods.
Features
- Simultaneous Predictions: Seamlessly perform regression and binary classification in a single step.
- Regressor Selection (choose one): Choose from Lasso, Ridge, or LinearRegression(OLS) as the base regression model.
- Dynamic Cutoff Tuning (choose one): Automatically tunes the cutoff value to maximize the Youden index, F1 score, or F2 score, user choose metric type.
1. Youden Index (J)
https://miro.medium.com/v2/resize:fit:842/1*LVilqC3cy4AgyC1wD4RH-A.png
$$J= Recall + Specificity - 1$$ J is a measure of the overall performance of a binary classifier. It is calculated as the sum of the recall and specificity minus 1. A high J statistic indicates that the classifier is performing well on both positive and negative cases.
- Recall measures the ability of a classifier to correctly identify positive cases. A high recall means that the classifier is avoiding miss detects.
- Specificity measures the ability of a classifier to correctly identify negative cases. A high specificity means that the classifier is avoiding false alarms.
2. F scores (Option F1, F2 in Package)
$$F1 = 2 \cdot \frac{precision \cdot recall}{precision + recall}$$
F1 score is another measure of the overall performance of a binary classifier. It is calculated as the harmonic mean of the precision and recall. A high F1 score indicates that the classifier is performing well on both positive and negative cases.
$$F_\beta = (1 + \beta^2) \cdot \frac{precision \cdot recall}{\beta^2 \cdot precision + recall}$$ F-score with factor beta is a generalization of the F1 score that allows for different weights to be given to precision and recall. A beta value less than 1 indicates that the F-score is prone to precision, while a beta value greater than 1 indicates that the F-score is prone to recall.
In educational settings
In educational settings, it is important to avoid miss detects (i.e., failing to identify at-risk students). However, it is also important to avoid false alarms (i.e., identifying students as at-risk when they are not). Therefore, it is often desirable to use a measure that is prone to recall, such as the F1 score with beta > 1. Youden's J statistic and the F1 score are both measures that balance the avoidance of miss detects and the avoidance of false alarms. However, the Youden's J statistic is less sensitive to false alarms (Specificity is less sensitive to false alarms compares to Precision) than the F1 score.
Installation
Install dualPredictor
directly from PyPI using pip:
pip install dualPredictor
Example Usage
from sklearn.datasets import fetch_california_housing
from dualPredictor.dual_model import DualModel
from dualPredictor.model_plot import plot_scatter,plot_feature_coefficients,plot_cm
# Fetching a dataset from scikit-learn for demonstration purposes
housing = fetch_california_housing(as_frame=True)
y = housing.target # Target variable (e.g., housing prices)
X = housing.data # Feature matrix
Train Model based on X and y
choose model_type and metric
# Initializing and fitting the DualModel, 'ols' for Ordinary Least Squares, a default cut-off value is provided
dual_clf = DualModel(model_type='ols', default_cut_off=2.5)
# The metric parameter specifies the method to tune the optimal cut-off
dual_clf.fit(X, y, metric='youden_index')
The DualModel
object has the following attributes:
alpha_
: The alpha value of the model. This value is only available if the model is a Lasso or Ridge regression model. (OLS do not have alpha)coef_
: The coefficients of the model.intercept_
: The intercept of the model.feature_names_in_
: The names of the features that were used to train the model.optimal_cut_off
: The optimal cut-off value that was determined by the specified metric.y_label_true_
: The true binary labels that were generated using the optimal cut-off value.
print(dual_clf.coef_)
[ 4.36693293e-01 9.43577803e-03 -1.07322041e-01 6.45065694e-01
-3.97638942e-06 -3.78654265e-03 -4.21314378e-01 -4.34513755e-01]
print(dual_clf.intercept_)
-36.94192020718441
print(dual_clf.feature_names_in_)
['MedInc' 'HouseAge' 'AveRooms' 'AveBedrms' 'Population' 'AveOccup'
'Latitude' 'Longitude']
print(dual_clf.optimal_cut_off)
2.5679
print(dual_clf.y_label_true_)
0 0
1 0
2 0
3 0
4 0
..
20635 1
20636 1
20637 1
20638 1
20639 1
Name: MedHouseVal, Length: 20640, dtype: int64
# Accessing the true binary labels generated based on the default cut-off
y_label_true = dual_clf.y_label_true_
# Retrieving the optimal cut-off value tuned based on the Youden Index
optimal_cut_off = dual_clf.optimal_cut_off
model predict on new data
# Predicting grades and binary classification (at-risk or not) based on the optimal cut-off
y_pred, y_label_pred = dual_clf.predict(X)
y_pred (regression result)
array([4.13164983, 3.97660644, 3.67657094, ..., 0.17125141, 0.31910524,
0.51580363])
y_label_pred (binary classification result)
array([0, 0, 0, ..., 1, 1, 1])
Exmaples of Model Performances Plot
# Visualizations
# Plotting the actual vs. predicted values to assess regression performance
scatter_plot_fig = plot_scatter(y_pred, y)
# Plotting the confusion matrix to evaluate binary classification performance
cm_plot = plot_cm(y_label_true, y_label_pred)
# Plotting the non-zero coefficients of the regression model to interpret feature importance
feature_plot = plot_feature_coefficients(coef=dual_clf.coef_, feature_names=dual_clf.feature_names_in_)
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file dualPredictor-0.0.4.tar.gz
.
File metadata
- Download URL: dualPredictor-0.0.4.tar.gz
- Upload date:
- Size: 7.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.0.0 CPython/3.12.2
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 90ab704279831cabfdaafe7c28cd78732af409259b623c31b32e35b9351a43c9 |
|
MD5 | fcf1d8fb79ece54049222266a25675e2 |
|
BLAKE2b-256 | 882596d1169431808c55cd03d0288fc39038b2a8a298ed23e93fdf36b43aa453 |
File details
Details for the file dualPredictor-0.0.4-py3-none-any.whl
.
File metadata
- Download URL: dualPredictor-0.0.4-py3-none-any.whl
- Upload date:
- Size: 7.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.0.0 CPython/3.12.2
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 18f22106c70395fe1589dba8d0434c45a12900d1de4190a60364ddb96be9f03c |
|
MD5 | d18b6064d799ca5b4ae16efa00516727 |
|
BLAKE2b-256 | e1172062a48e95106c8a7c418c5669d0bd8647e19d105f673695c2ccb09f9057 |