Parameter free cost-sensitive adaptive boosting
Project description
AdaCC: Adaptive Cost-sensitive Boosting for Imbalanced Data
Overview
AdaCC is a novel cost-sensitive boosting approach designed to address the challenge of class imbalance in machine learning. Traditional supervised learning models often exhibit bias towards the majority class, leading to under-performance in minority classes. Cost-sensitive learning attempts to mitigate this issue by treating classes differently, employing a fixed misclassification cost matrix provided by users. However, manually tuning these parameters can be daunting and might lead to suboptimal results if not done accurately.
In this work, we introduce AdaCC, a groundbreaking method that dynamically adjusts misclassification costs over boosting rounds based on the model's performance. Unlike conventional approaches, AdaCC eliminates the need for fixed misclassification cost matrices, offering a parameter-free solution. By leveraging the cumulative behavior of the boosting model, AdaCC automatically adapts misclassification costs for subsequent boosting rounds, ensuring optimal balance between classes and enhancing predictive accuracy.
The following example showcases how the weighting strategy of AdaCC differs from AdaBoost. Feel free to also read the technical paper of this approach "AdaCC: cumulative cost-sensitive boosting for imbalanced classification"
Key Features
-
Dynamic Cost Adjustment: AdaCC dynamically modifies misclassification costs in response to the boosting model's performance, optimizing class balance without the need for manual parameter tuning.
-
Parameter-Free Solution: Eliminates the complexity of defining fixed misclassification cost matrices, providing a hassle-free experience for users without requiring domain knowledge.
-
Theoretical Guarantees: AdaCC comes with theoretical guarantees regarding training error, ensuring the reliability and robustness of the boosting model across various datasets.
How to Use
- Installation
- pip install cumulative-cost-boosting
- Usage
from cumulative_cost_boosting import AdaCC
clf = AdaCC(n_estimators=100, algorithm='AdaCC1')
clf.fit(X_train, y_train)
predictions = clf.predict(X_test)
Example
A detailed example is displayed in the run_example.py
file.
Contributions and Issues
Contributions and feedback are welcome. If you encounter any issues or have suggestions for improvement, please feel free to create an issue in the repository or submit a pull request.
Note: AdaCC is a cutting-edge solution for handling class imbalance, ensuring accurate and fair predictions in machine learning tasks. Thank you for considering AdaCC for your imbalanced data challenges. Let's empower your models together!
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file cumulative_cost_boosting-0.1.2.tar.gz
.
File metadata
- Download URL: cumulative_cost_boosting-0.1.2.tar.gz
- Upload date:
- Size: 8.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.7.0 CPython/3.8.2 Windows/10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 47d2f97a7a3902687b468da728db959976f452d557835c1105d6ddb0bf487e90 |
|
MD5 | 12583286116955f19f678e429fa1fc71 |
|
BLAKE2b-256 | a3549a33ecd98df9ee61a71b5b2a0bc2193d1328c9d2e0e33fe93adde2d505f4 |
File details
Details for the file cumulative_cost_boosting-0.1.2-py3-none-any.whl
.
File metadata
- Download URL: cumulative_cost_boosting-0.1.2-py3-none-any.whl
- Upload date:
- Size: 8.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.7.0 CPython/3.8.2 Windows/10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 1656371017f9fb3c42232cd7c735e8b9294dc41c2fad0d39d71c8af50540c8dc |
|
MD5 | f276b0e8910ace6bda99bb3f8e8cbd62 |
|
BLAKE2b-256 | 90458e7eeae01920c88f1d8c826f35196cd6c3803a81e7ac21934b1a270a0902 |